samansangini / gpu_poor Goto Github PK
View Code? Open in Web Editor NEWThis project forked from rahulschand/gpu_poor
Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization
Home Page: https://rahulschand.github.io/gpu_poor/