Comments (6)
It'd be really cool if the minimum requirements of the model (size on disk for data set, vram requirements) on the readme, that would save a lot of people some time.
from openchatkit.
You ran out of GPU memory. Describe more on your setup like what you are using and what command you ran to resolve.
from openchatkit.
It'd be really cool if the minimum requirements of the model (size on disk for data set, vram requirements) on the readme, that would save a lot of people some time.
That's a great idea. I'll put up a PR soon to document this.
from openchatkit.
You ran out of GPU memory. Describe more on your setup like what you are using and what command you ran to resolve.
from openchatkit.
(OpenChatKit) root@aca2869c8358:~/OpenChatKit-main# python inference/bot.py
Loading /root/OpenChatKit-main/inference/../huggingface_models/GPT-NeoXT-Chat-Base-20B to cuda:0...
Traceback (most recent call last):
File "/root/OpenChatKit-main/inference/bot.py", line 185, in
main()
File "/root/OpenChatKit-main/inference/bot.py", line 181, in main
).cmdloop()
File "/root/anaconda3/envs/OpenChatKit/lib/python3.10/cmd.py", line 105, in cmdloop
self.preloop()
File "/root/OpenChatKit-main/inference/bot.py", line 64, in preloop
self._model = ChatModel(self._model_name_or_path, self._gpu_id)
File "/root/OpenChatKit-main/inference/bot.py", line 24, in init
self._model.to(device)
File "/root/anaconda3/envs/OpenChatKit/lib/python3.10/site-packages/torch/nn/modules/module.py", line 989, in to
return self._apply(convert)
File "/root/anaconda3/envs/OpenChatKit/lib/python3.10/site-packages/torch/nn/modules/module.py", line 641, in _apply
module._apply(fn)
File "/root/anaconda3/envs/OpenChatKit/lib/python3.10/site-packages/torch/nn/modules/module.py", line 641, in _apply
module._apply(fn)
File "/root/anaconda3/envs/OpenChatKit/lib/python3.10/site-packages/torch/nn/modules/module.py", line 641, in _apply
module._apply(fn)
[Previous line repeated 2 more times]
File "/root/anaconda3/envs/OpenChatKit/lib/python3.10/site-packages/torch/nn/modules/module.py", line 664, in _apply
param_applied = fn(param)
File "/root/anaconda3/envs/OpenChatKit/lib/python3.10/site-packages/torch/nn/modules/module.py", line 987, in convert
return t.to(device, dtype if t.is_floating_point() or t.is_complex() else None, non_blocking)
torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 288.00 MiB (GPU 0; 14.56 GiB total capacity; 13.86 GiB already allocated; 90.44 MiB free; 13.88 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
from openchatkit.
some problem, any idea how much memory it needs? or any solution to reduce the memory use? Thanks.
from openchatkit.
Related Issues (20)
- We couldn't connect to 'https://huggingface.co' HOT 1
- Add CodeAlpaca-20k dataset to improve coding skills.
- -
- When use one Gpu do model training, met one issue. HOT 3
- Environment Issues On Mac HOT 1
- Example script for continued pre-training? HOT 2
- How to disable AWS_ACCESS_KEY_ID when fine tuning? HOT 2
- LOST in the MIDDLE
- how many card days to Fine-tuning Llama-2-7B-32K-beta
- An error occurred while fine-tuning the model. HOT 3
- Cannot setup environment HOT 1
- Training on BookSum HOT 1
- how to train Fine-tuning Llama-2-7B-32K-beta?
- How to start the combined server/ send commands over HTTP?
- API is not working when inferenced with streamlit
- NotImplementedError: Loading a streaming dataset cached in a LocalFileSystem is not supported yet.
- ModuleNotFoundError: No module named 'flash_attn' HOT 1
- What is minimum resource requirement to fine-tuning Llama-2-7B-32K-beta model.
- H
- Newbie question. ยฟCan I use this to build a chat assistant? HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. ๐๐๐
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google โค๏ธ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from openchatkit.