Comments (4)
Unfortunately, flash attention v2 does not support P100 (nor V100). You may need to uninstall the related packages in the image (pip uninstall flash_attn dropout_layer_norm
) or build the image from scratch and set environment variable BUNDLE_FLASH_ATTENTION
to false
.
from qwen.
If you are using the provided the docker image with tag qwenllm/qwen(:latest), it is based on CUDA 11.7 and bundles the layer_norm module from flash attention v2, where that invalid device function (cudaOccupancyMaxActiveBlocksPerMultiprocessor which is a CUDA runtime API) is called.
It is likely your nvidia driver is too old to support CUDA 11.7 (and later versions). Please run nvidia-smi
and provide the result.
from qwen.
If you are using the provided the docker image with tag qwenllm/qwen(:latest), it is based on CUDA 11.7 and bundles the layer_norm module from flash attention v2, where that invalid device function (cudaOccupancyMaxActiveBlocksPerMultiprocessor which is a CUDA runtime API) is called.
It is likely your nvidia driver is too old to support CUDA 11.7 (and later versions). Please run
nvidia-smi
and provide the result.
Wed Apr 10 06:16:11 2024
nvidia-smi驱动查询结果如下,感觉应该能支持CUDA11.7,有没有可能是别的什么原因呢
+---------------------------------------------------------------------------------------+
| NVIDIA-SMI 545.23.06 Driver Version: 545.23.06 CUDA Version: 12.3 |
|-----------------------------------------+----------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+======================+======================|
| 0 Tesla P100-PCIE-16GB Off | 00000000:44:00.0 Off | 0 |
| N/A 27C P0 29W / 250W | 0MiB / 16384MiB | 0% Default |
| | | N/A |
+-----------------------------------------+----------------------+----------------------+
| 1 Tesla P100-PCIE-16GB Off | 00000000:87:00.0 Off | 0 |
| N/A 27C P0 28W / 250W | 0MiB / 16384MiB | 0% Default |
| | | N/A |
+-----------------------------------------+----------------------+----------------------+
| 2 Tesla P100-PCIE-16GB Off | 00000000:C1:00.0 Off | 0 |
| N/A 26C P0 30W / 250W | 0MiB / 16384MiB | 0% Default |
| | | N/A |
+-----------------------------------------+----------------------+----------------------+
| 3 Tesla P100-PCIE-16GB Off | 00000000:C4:00.0 Off | 0 |
| N/A 26C P0 29W / 250W | 0MiB / 16384MiB | 0% Default |
| | | N/A |
+-----------------------------------------+----------------------+----------------------+
+---------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=======================================================================================|
| No running processes found |
+---------------------------------------------------------------------------------------+
from qwen.
Unfortunately, flash attention v2 does not support P100 (nor V100). You may need to uninstall the related packages in the image (
pip uninstall flash_attn dropout_layer_norm
) or build the image from scratch and set environment variableBUNDLE_FLASH_ATTENTION
tofalse
.
谢谢。
from qwen.
Related Issues (20)
- [BUG] <title> 如何用vllm部署qlora后的模型 HOT 1
- [BUG] .CalledProcessError: Command '['/usr/bin/gcc', '/tmp/tmpecd6su1w/main.c' HOT 3
- how to convert qwen.tiktoken to tokenzier.model HOT 1
- Run Qwen /openai_api.py, Error :Input should be a valid string, body.messages[3].function_call,请问Qwen1.5不支持了么? HOT 1
- pip install csrc/layer_norm 不成功 HOT 1
- [BUG] <title> wrong system prompt check? HOT 2
- [BUG] <title>batch_infer报错:'tuple' object has no attribute 'dtype' HOT 2
- 如何添加`LogitsProcessor`控制结果输出?
- [BUG] <title>lora微调loss异常? HOT 5
- tokenizer.decoder 抛出'utf-8' codec can't decode bytes in position 1-2: unexpected end of data异常 HOT 2
- [BUG] lora微调后,合并成一个模型。这种方式如何加载且推理 HOT 4
- [BUG] Qwen/Qwen-72B-Chat-Int8,不能多GPU并行计算 HOT 1
- Qwen/eval中的评测CEval和CMMLU,开大推理的batchsize评测指标会显著降低 HOT 1
- 请问基于qwen-72b-chat,基于怎样的配置可以在一台4090上训练起来? HOT 4
- 💡 [REQUEST] - <title> 关于lora 模型合并的几个问题 HOT 3
- [BUG] <关于model.generate时发现的源码错误> HOT 2
- [BUG] <Qwen-14B-Chat 输入长文本时无输出结果> HOT 5
- Qwen1.5量化结果不一致
- [BUG] Function Calling 示例有错误,最新的 openai sdk 运行时提示 api 已经废弃 HOT 2
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from qwen.