Comments (10)
I suggest you set fp16 to false, as setting it to true often results in unexpected outcomes.
from easyedit.
Sure thanks. One more thing. Is the R-ROME implementation supporting llama2-7b and gpt-j as in hparams or basically all the models as stated in README?
from easyedit.
Yes, all are supported. You can just modify the parameters a bit.
from easyedit.
Thnaks a lot! Now I changed fp16 to false. I'm exceeding 50gb of memory (2 x RTX 4090) for Mistral-7b. It barely worked on llama2-7b taking over 46gb. Should it be as memory consuming to run ROME?
from easyedit.
Same happened for KN model. Just running out of memory with 50gb in paralell setup on llama-7b.
from easyedit.
I don't think there's any problem; this is the normal memory usage. You can try setting it to fp16 at
EasyEdit/easyeditor/editors/editor.py
Line 63 in 38c5c34
from easyedit.
Hi, do you have any further questions?
from easyedit.
Hi, thank you for your answer. Yes, maybe one more.
How exactly in current implementation for example when I use editor with ROME locality, probability, rewrite and rephrase accuracies are calculated?
from easyedit.
You can check the file at easyeditor/evaluate/evaluate_utils.py. Except for locality, we calculate the number of matching tokens and then take the average for everything else. Locality calculates whether unrelated inputs have changed before and after the edits.
from easyedit.
Ok, thank you very much :)
from easyedit.
Related Issues (20)
- error when cache file exist HOT 6
- When will the WISE pipeline implementation be publicized? HOT 4
- Can i edit quantized model? HOT 1
- LoRA does not work HOT 4
- 关于portability和locality的评估具体应该怎么设置呢 HOT 6
- 使用mend预训练llama时出现的报错 HOT 2
- 数据集问题 HOT 2
- counterfact数据结果 HOT 11
- ROME Question HOT 1
- [Speed issue]: How would you recommend running EasyEdit faster HOT 3
- [Evaluation Issues]: T5 Results are really strange HOT 1
- Evaluation Question HOT 4
- locality and portability evaluation HOT 8
- GRACE sequential edit result HOT 8
- IKE fluency HOT 2
- ccks gpt2-xl 模型 为什么用roberta模型加载? HOT 2
- R-ROME has poor performance when using GPT2-xl HOT 2
- WISE CONTEXT_TEMPLATES_CACHE HOT 7
- WISE tokenize HOT 7
- what is the meaning of archive HOT 2
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from easyedit.