Comments (8)
There is a comment in the code about needing authorization by Meta and passing a token to Hugging Face.
model_name = "NousResearch/llama-2-7b-chat-hf" # use this if you have access to the official LLaMA 2 model "meta-llama/Llama-2-7b-chat-hf", though keep in mind you'll need to pass a Hugging Face key argument
QUESTION 1: Does this mean that the default code assumes that you have HF/Meta access to 2-7b-chat-hf? And if you don't have it, does that explain the memory error?
QUESTION 2: Has anyone plugged anything else in here with success? If so, what?
from gpt-llm-trainer.
Has anyone found a solution to this? Same her.
from gpt-llm-trainer.
I am still unable to merge the model because I am getting the same error as @smilinrobin. I have Colab Pro and am running on a V100. I have implemented the recommendations that are readily available on the Internet, such as setting max_split_size_mb to 250, and none of them have made much difference.
Can we have a team effort to improve this situation? If you have llm-trainer working on Colab, can you please share your configuration?
from gpt-llm-trainer.
Yes I am having the same issue as others at the Merge section, it gets to about 50% before running out of memory (even though I am using a V100 GPU on Colab Pro. Hope this gets fixed soon as this is a very useful project
from gpt-llm-trainer.
UPDATE: I went through the sign-up and have verified that I have access to Meta, to NousResearch, to Hugging Face, and am successfully passing the login.
I also set max_split_per_mb to 32 and experimented with various values of memory fraction.
I updated to Google Colab Pro+. No benefit, since they still only issue one 16GB GPU.
At this point I have to conclude that the model can't be run on standard Google Colab Pro accounts, unless I am missing something. Can anyone prove me wrong?
from gpt-llm-trainer.
UPDATE: I went through the sign-up and have verified that I have access to Meta, to NousResearch, to Hugging Face, and am successfully passing the login.
I also set max_split_per_mb to 32 and experimented with various values of memory fraction.
I updated to Google Colab Pro+. No benefit, since they still only issue one 16GB GPU.
At this point I have to conclude that the model can't be run on standard Google Cloud Pro accounts, unless I am missing something. Can anyone prove me wrong?
same conclusion here, I got access to the llama 2 from Meta , and still getting "CUDA out of memory" error.
if anyone plugged anything else other then the llama 2 it would be very helpful.
from gpt-llm-trainer.
Yes the issue continues, the gpt 3.5 llm trainer does work well on Colab though. But would love to have a model that doesn't incur suc h outrageous costs once it is spun into a production environment
from gpt-llm-trainer.
Related Issues (18)
- Token generation limit HOT 5
- The model `gpt-4` does not exist or you do not have access to it
- Logging into wandb.ai HOT 1
- NousResearch/llama-2-7b-chat-hf NOT AVAILABLE HOT 1
- ㅂㅂ
- Cost estimate? HOT 3
- Merge the model and store in Google Drive (Section) HOT 3
- the model before lora load and after lora load is diff HOT 1
- llm
- API not working even after upgrading to gpt 4 HOT 1
- Problem with workflow
- hello, would you have time for a chat? HOT 1
- error :You tried to access openai.ChatCompletion, but this is no longer supported in openai>=1.0.0 HOT 1
- without openai !!! HOT 1
- Add 'LLM Knowledge Distillation' to Readme or Topic Tags
- which GPU? HOT 1
- Can we use GPT3.5? HOT 6
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from gpt-llm-trainer.