Comments (14)
According to my machine, the program takes up so much memory that my 16 gigabytes of ram overflows. (My computer freezes for a second) The problem you are facing to could be somewhat similar to mine
This is exactly the problem. I increased it to 24GB but still it eats away memory, chews through the swap file (8GB) and then kills itself.
How much memory does this take? We really need a smaller model. 16GB is pretty common on desktops.
from localgpt.
For wsl you have to edit your .wslconfig file. Search online for the options.
https://learn.microsoft.com/en-us/windows/wsl/wsl-config
For default linux there's a few steps to take to unlink your swapfile, change the size etc..
from localgpt.
same problem - no solution yet
from localgpt.
@Anas-Dew @achillez Can you share your hardware configuration and memory utilization while the code is running?
from localgpt.
Sure it's a i7 11700K (8 phy core, 16 virt core), 32GB physical mem (16GB in WSL). GTX 1660 Ti 6GB mem. Not sure GPU config matters since I see the exact same issue in CPU mode.
from localgpt.
It was 3gb ram and single core cpu I guess, I was running it on cloud computer. (neverinstall)
from localgpt.
from localgpt.
I do this usually when trying out new things.
from localgpt.
Here's the python trace of the error. Appears it fails somewhere in torch
init.py(101): if nonlinearity in linear_fns or nonlinearity == 'sigmoid':
init.py(103): elif nonlinearity == 'tanh':
init.py(105): elif nonlinearity == 'relu':
init.py(107): elif nonlinearity == 'leaky_relu':
init.py(108): if param is None:
init.py(110): elif not isinstance(param, bool) and isinstance(param, int) or isinstance(param, float):
init.py(112): negative_slope = param
init.py(115): return math.sqrt(2.0 / (1 + negative_slope ** 2))
init.py(409): std = gain / math.sqrt(fan)
init.py(410): bound = math.sqrt(3.0) * std # Calculate uniform bounds from standard deviation
init.py(411): with torch.no_grad():
--- modulename: grad_mode, funcname: init
grad_mode.py(49): if not torch._jit_internal.is_scripting():
--- modulename: _jit_internal, funcname: is_scripting
_jit_internal.py(1121): return False
grad_mode.py(50): super().init()
grad_mode.py(51): self.prev = False
--- modulename: grad_mode, funcname: enter
grad_mode.py(54): self.prev = torch.is_grad_enabled()
grad_mode.py(55): torch.set_grad_enabled(False)
--- modulename: grad_mode, funcname: init
grad_mode.py(150): self.prev = torch.is_grad_enabled()
grad_mode.py(151): torch._C.set_grad_enabled(mode)
grad_mode.py(152): self.mode = mode
init.py(412): return tensor.uniform(-bound, bound)
Killed
from localgpt.
According to my machine, the program takes up so much memory that my 16 gigabytes of ram overflows. (My computer freezes for a second)
The problem you are facing to could be somewhat similar to mine
from localgpt.
same here!
from localgpt.
I'm having the same issue when running on CPU
$ python run_localGPT.py --device_type cpu
Running on: cpu
load INSTRUCTOR_Transformer
max_seq_length 512
Using embedded DuckDB with persistence: data will be stored in: ....
Killed
from localgpt.
Fixed this by increasing the swap file size. Appears you need around 40GB of memory (mem+swap) to avoid the app crashing. Now it runs and asks for a prompt. However, I can't get it to respond with a legitimate answer.
Not sure why - trying a rerun with ingest + run_LocalGPT
from localgpt.
@achillez what command did you use to set the swap file size?
from localgpt.
Related Issues (20)
- How to show document sources as images instead of text in the streamlit dashboard? HOT 1
- Error in running this run_localGPT.py HOT 4
- 500 Internal Server Error HOT 2
- SSL issues HOT 1
- /lib64/libgcc_s.so.1: version `GCC_7.0.0' not found libarrow.so.1400 HOT 1
- Phi-2 support HOT 1
- Terminal VS Interface GPU problem HOT 5
- IndexError: list index out of range HOT 2
- AutoGPTQ must be updated to 0.6 for Mistral HOT 2
- Compare the response time differences between 4xA100 and 8xH100 HOT 5
- Type=Value Error whenever we load any other LLM model rather than Llama and Mistral HOT 1
- PowerInfer integration
- Embdedings & Model files Location in local system. HOT 4
- Local Gpt
- Requested tokens exceed context window of 4096 HOT 3
- KeyError: 'Cache only has 0 layers, attempted to access layer with index 0' HOT 3
- Suggestion to improve or fine tune the model with custom documents HOT 1
- No module named click HOT 1
- run_localGPT.py fail: python run_localGPT.py --device_type cpu
- multi-user or async prompt requests crashes the app HOT 6
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from localgpt.