Comments (6)
I tried 3.5 today on two small runs, one of 10 and the other of 30 examples. They yielded 2 and 25 valids, respectively. I had to adjust to gpt-3.5-turbo-16k.
The reason I was doing this is because I was trying to isolate the out of memory problem that is preventing me from successfully using gpt-llm-trainer, and was tired of spending money on gpt-4 runs. ;-)
from gpt-llm-trainer.
I used model="gpt-3.5-turbo"
, and it ran correctly. I can't comment on the quality of the GTP output as I actually don't need it, I just tried as a test.
from gpt-llm-trainer.
@fredzannarbor @Afo92 Thank your for the confirmation.
from gpt-llm-trainer.
I am facing this issue, with the default model
OSError: NousResearch/llama-2-7b-chat-hf is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models'
If this is a private repository, make sure to pass a token having permission to this repo with use_auth_token
or log in with huggingface-cli login
and pass use_auth_token=True
.
any idea what is to be done?
I tried my own sharded model but got me some other error
from gpt-llm-trainer.
I used model_name = "meta-llama/Llama-2-7b-chat-hf"
and then
access_token = "hf_YOUR TOKEN HER!"
model = AutoModelForCausalLM.from_pretrained(
model_name,
use_auth_token=access_token,
quantization_config=bnb_config,
device_map=device_map
)
from gpt-llm-trainer.
@Afo92 Gotcha thank you!
from gpt-llm-trainer.
Related Issues (18)
- Token generation limit HOT 5
- The model `gpt-4` does not exist or you do not have access to it
- Logging into wandb.ai HOT 1
- NousResearch/llama-2-7b-chat-hf NOT AVAILABLE HOT 1
- ㅂㅂ
- Cost estimate? HOT 3
- Merge the model and store in Google Drive (Section) HOT 3
- the model before lora load and after lora load is diff HOT 1
- llm
- API not working even after upgrading to gpt 4 HOT 1
- Problem with workflow
- hello, would you have time for a chat? HOT 1
- error :You tried to access openai.ChatCompletion, but this is no longer supported in openai>=1.0.0 HOT 1
- Running into CUDA out of memory on Colab HOT 8
- without openai !!! HOT 1
- Add 'LLM Knowledge Distillation' to Readme or Topic Tags
- which GPU? HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from gpt-llm-trainer.