Comments (7)
Having the exact same issue with Ubuntu 22 LTS. 32GB of ram, Nvidia Tesla m40 20gb.
from localgpt.
I have a 32GB system running Ubuntu 20.04 on which am running the "python3 run_localGPT.py". After the execution, the RAM went to 0 in some sec and then I got the same issue. I am uncertain how much memory is needed to run the model.
from localgpt.
I am getting the same error on Ubuntu 22.04.2 .
load INSTRUCTOR_Transformer
max_seq_length 512
Using embedded DuckDB with persistence: data will be stored in: /home/ailocal/localGPT/DB
Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s]
Killed
from localgpt.
+1
from localgpt.
I am having the same issue. Just installed on Ubuntu 22.04.
16GB of Memory; 25GB of Swap. the process fails when Swap reaches 100% utilization.
Running on: cuda
load INSTRUCTOR_Transformer
max_seq_length 512
Using embedded DuckDB with persistence: data will be stored in: /home/MYNAME/snap/localgpt/localGPT/DB
Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s]Killed
from localgpt.
from localgpt.
While not an ideal solution, with having 32gb of physical ram, I expanded my swap file to be an additional 32gb, for a total of 64gb of combined physical and digital ram. With that, I am able to run the program, enter queries, and receive expected answers. It should be noted that this solution is incredibly slow, even with a PCI4 nvme ssd being used for the swap file.
from localgpt.
Related Issues (20)
- Why its sharing questions and data from different browser sessions?
- How do I add memory to chat-zero-shot-react-description?
- Autoawq HOT 2
- Mistral not supported HOT 2
- cpp-llama-python not found. HOT 1
- problem when ingesting (just CPU) HOT 1
- auto-gptq and auto awq is breaking in requiremetns.txt HOT 1
- I encountered a mistake when I executed run_localGPT_API.py HOT 1
- Extra Options with run_localGPT_API.py?
- error in /opt/nvidia/nvidia_entrypoint.sh HOT 1
- run_localGPT_API HOT 8
- Support llama-3 HOT 10
- If I want to improve the Recall access the reranker model ,how can I do it?
- llama3 support: ERROR: byte not found in vocab, segmentation fault HOT 1
- Unable to execute 'python run_localGPT.py --device_type cpu'
- Can I reuse the models which I have running locally via ollama service ?
- Error response from daemon: could not select device driver "" with capabilities: [[gpu]]. HOT 1
- How to authenticate to huggingface.co, from the run_localGPT.py script, using Docker? HOT 1
- Cannot access gated repo HOT 2
- Improved metadata at ingest
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from localgpt.