Comments (17)
Try training on multiple GPUs, or increase the number of gradient_accumulation_steps
.
from alpaca-cot.
这很奇怪,你是单卡训练的吗?
from alpaca-cot.
是的,单卡3090
from alpaca-cot.
已私聊
from alpaca-cot.
I’m encountering the same problem, can you help me?
from alpaca-cot.
will try. But how about tweaking the parameters? do you know what parameter I should adjust to make the response not repetitive and truthful? I tried changing the temperature to 0.8 and setting the greedy decoding to True
from alpaca-cot.
or should I just train it longer by adding the epoch?
from alpaca-cot.
- Raising the temperature has a slight effect on reducing repetition: higher temperatures have a smoother vocabulary prediction distribution, and a greater chance of picking up words that are not usually picked up.
- If you set
greedy
to true, there is no need to configure the temperature. - The main reason is that the model is not well trained. More data can be used. If the amount of data is small (50k), it is recommended to train for 3 epoch. If the amount is large (0.5M+), you can only train for 1 epoch.
from alpaca-cot.
by 50k do you mean the number of instructions? my data size is currently 26MB
from alpaca-cot.
yes. 26MB (i mean the number of instructions) data is fairly large. Is the data quality not good?
from alpaca-cot.
well my data size is 26MB and has 50k instructions. The quality should be similar to the alpaca data, as I only translate it to other language using chatgpt.
from alpaca-cot.
I see. This may be due to Bloom's weak ability in your target language. You can collect more data, try larger bloom models, or try the LLM performing well on the target language.
from alpaca-cot.
do you think adding more epochs will solve the problem?
from alpaca-cot.
Yes, it should be better.
from alpaca-cot.
How many epoches do you set? 3 epochs is suitable for 50k instructions.
from alpaca-cot.
currently only one, since I have a limitation on my GPU usage. Three epochs would cost about 10 hours of training (in A100 80GB)
from alpaca-cot.
I'll switch the model to llama and see how it goes
from alpaca-cot.
Related Issues (20)
- ChatGLM的Finetune推荐命令,使用3090 24G会OOM,代码默认使用8Bit量化同样会导致OOM HOT 1
- 请问如何修改模型的自我认知
- GPTeacher Code-Instruct HOT 1
- 8卡V100跑moss OOV HOT 2
- Prompt设置 HOT 1
- About the tokenizer
- The text meaning in zh_helpfulness_context.json in Alpaca-CoT / MOSS / moss-002-sft
- DataCollatorForLanguageModeling uses the unmasked labels
- web.py中缺少--size参数 HOT 1
- inference结果差异比较大,请问是什么原因 HOT 2
- 是否可以提供一个Gdrive和百度云的下载方式 HOT 2
- 是否可以支持qlora
- 你好,群二维码过期了 HOT 1
- About the source of the dataset
- What is the relationship between the data and the link you provided?
- 你好,能更新下群信息么 HOT 2
- Adding Contributors Section In readme.md
- 你好,群二维码过期了,能更新一下么~ HOT 6
- main分支下的readme顺序,以及base模型能否提供huggingface的链接 HOT 4
- 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from alpaca-cot.