Comments (6)
Package Version
accelerate 0.18.0
aiohttp 3.8.4
aiosignal 1.3.1
async-timeout 4.0.2
asynctest 0.13.0
attrs 22.2.0
cachetools 5.3.0
certifi 2022.12.7
charset-normalizer 3.1.0
cpm-kernels 1.0.11
datasets 2.10.1
DBUtils 3.0.2
dill 0.3.6
filelock 3.10.7
frozenlist 1.3.3
fsspec 2023.1.0
gpt-db-tools 0.0.1
huggingface-hub 0.13.3
icetk 0.0.4
idna 3.4
importlib-metadata 6.1.0
multidict 6.0.4
multiprocess 0.70.14
numpy 1.21.6
nvidia-cublas-cu11 11.10.3.66
nvidia-cuda-nvrtc-cu11 11.7.99
nvidia-cuda-runtime-cu11 11.7.99
nvidia-cudnn-cu11 8.5.0.96
nvidia-ml-py 11.525.84
nvitop 1.0.0
packaging 23.0
pandas 1.3.5
peft 0.2.0
Pillow 9.4.0
pip 22.0.4
protobuf 3.20.0
psutil 5.9.4
psycopg2-binary 2.9.5
pyarrow 11.0.0
python-dateutil 2.8.2
pytz 2023.2
PyYAML 6.0
regex 2022.10.31
requests 2.28.2
responses 0.18.0
sentencepiece 0.1.97
setuptools 47.1.0
six 1.16.0
termcolor 2.2.0
tokenizers 0.13.2
torch 1.13.1
torchvision 0.14.1
tqdm 4.65.0
transformers 4.26.1
typing_extensions 4.5.0
urllib3 1.26.15
wheel 0.40.0
xxhash 3.2.0
yarl 1.8.2
zipp 3.15.0
我这个是跑的微调chatglm-6b 的,成功运行能微调
from zero_nlp.
cuda 11.7
from zero_nlp.
Package Version
accelerate 0.18.0 aiohttp 3.8.4 aiosignal 1.3.1 async-timeout 4.0.2 asynctest 0.13.0 attrs 22.2.0 cachetools 5.3.0 certifi 2022.12.7 charset-normalizer 3.1.0 cpm-kernels 1.0.11 datasets 2.10.1 DBUtils 3.0.2 dill 0.3.6 filelock 3.10.7 frozenlist 1.3.3 fsspec 2023.1.0 gpt-db-tools 0.0.1 huggingface-hub 0.13.3 icetk 0.0.4 idna 3.4 importlib-metadata 6.1.0 multidict 6.0.4 multiprocess 0.70.14 numpy 1.21.6 nvidia-cublas-cu11 11.10.3.66 nvidia-cuda-nvrtc-cu11 11.7.99 nvidia-cuda-runtime-cu11 11.7.99 nvidia-cudnn-cu11 8.5.0.96 nvidia-ml-py 11.525.84 nvitop 1.0.0 packaging 23.0 pandas 1.3.5 peft 0.2.0 Pillow 9.4.0 pip 22.0.4 protobuf 3.20.0 psutil 5.9.4 psycopg2-binary 2.9.5 pyarrow 11.0.0 python-dateutil 2.8.2 pytz 2023.2 PyYAML 6.0 regex 2022.10.31 requests 2.28.2 responses 0.18.0 sentencepiece 0.1.97 setuptools 47.1.0 six 1.16.0 termcolor 2.2.0 tokenizers 0.13.2 torch 1.13.1 torchvision 0.14.1 tqdm 4.65.0 transformers 4.26.1 typing_extensions 4.5.0 urllib3 1.26.15 wheel 0.40.0 xxhash 3.2.0 yarl 1.8.2 zipp 3.15.0
我这个是跑的微调chatglm-6b 的,成功运行能微调
感谢感谢,transformers 4.26.1这个版本可以跑chatgml吗?
from zero_nlp.
Package Version
accelerate 0.18.0 aiohttp 3.8.4 aiosignal 1.3.1 async-timeout 4.0.2 asynctest 0.13.0 attrs 22.2.0 cachetools 5.3.0 certifi 2022.12.7 charset-normalizer 3.1.0 cpm-kernels 1.0.11 datasets 2.10.1 DBUtils 3.0.2 dill 0.3.6 filelock 3.10.7 frozenlist 1.3.3 fsspec 2023.1.0 gpt-db-tools 0.0.1 huggingface-hub 0.13.3 icetk 0.0.4 idna 3.4 importlib-metadata 6.1.0 multidict 6.0.4 multiprocess 0.70.14 numpy 1.21.6 nvidia-cublas-cu11 11.10.3.66 nvidia-cuda-nvrtc-cu11 11.7.99 nvidia-cuda-runtime-cu11 11.7.99 nvidia-cudnn-cu11 8.5.0.96 nvidia-ml-py 11.525.84 nvitop 1.0.0 packaging 23.0 pandas 1.3.5 peft 0.2.0 Pillow 9.4.0 pip 22.0.4 protobuf 3.20.0 psutil 5.9.4 psycopg2-binary 2.9.5 pyarrow 11.0.0 python-dateutil 2.8.2 pytz 2023.2 PyYAML 6.0 regex 2022.10.31 requests 2.28.2 responses 0.18.0 sentencepiece 0.1.97 setuptools 47.1.0 six 1.16.0 termcolor 2.2.0 tokenizers 0.13.2 torch 1.13.1 torchvision 0.14.1 tqdm 4.65.0 transformers 4.26.1 typing_extensions 4.5.0 urllib3 1.26.15 wheel 0.40.0 xxhash 3.2.0 yarl 1.8.2 zipp 3.15.0
我这个是跑的微调chatglm-6b 的,成功运行能微调感谢感谢,transformers 4.26.1这个版本可以跑chatgml吗?
必须可以啊,我就是跑的这个模型
from zero_nlp.
感谢大哥,不过我的cuda最高只能是11.3,用不了11.7的
from zero_nlp.
有没有用cuda11.3的跑通的大哥,发一下conda环境参考一下,十分感谢!!!
from zero_nlp.
Related Issues (20)
- chatGLMv2-6b p-tuning 和 LoRA数据预处理的方法是一样的吗 ?
- ChatGLM2 lora finetuning 加载 lora 参数:RuntimeError: Expected 4-dimensional input for 4-dimensional weight [3072, 32, 1, 1], but got 3-dimensional input of size [1, 64, 4096] instead HOT 4
- 4张3080ti跑chatglm2-6b-lora报oom HOT 5
- 求助:chatglm2 lora训练error:RuntimeError: Expected is_sm80 to be true, but got false. HOT 2
- 训练的时候报错ValueError: The current `device_map` had weights offloaded to the disk. HOT 11
- 训练出错
- 两张4090单机多卡跑,咋感觉越跑越慢了,比单卡慢 HOT 2
- 请问有部署或者运行的文档吗?在哪里可以看?
- 实时微调可以通过加入传统RL实现吗
- 请问如果单纯使用zeroth-order向前优化少量batch(只要体现出一定的优化效果)的话要怎么实现 HOT 2
- lora推理中只能指定一个输入吗?有办法实现batch_size的推理吗
- 救命!!ChatGlm-v2-6b_Lora该怎么设置epoch?? HOT 1
- 大佬,可以多个多个lora叠加使用吗?
- chatglm_v2_6b_lora多卡如何设置,没有找到 HOT 2
- 能出一个ChatGLM
- 能出一个ChatGLM的教程吗
- Segment Fault 是哪的问题?
- 大佬 chinese_llama 还可以用吗 HOT 1
- 出个chatglm3的吧 微调后 推理老是出问题 HOT 1
- internlm-sft 单机多卡微调 GPU 利用率低 HOT 5
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from zero_nlp.