Code Monkey home page Code Monkey logo

Comments (11)

collant avatar collant commented on August 27, 2024 9

Hello, you can find this 13B one here: https://huggingface.co/samwit/alpaca13B-lora

Otherwise, there is the 7B one here: https://huggingface.co/tloen/alpaca-lora-7b

Please note these are LoRA models they need the base model to work.

And here is the base model for the 7B: https://huggingface.co/decapoda-research/llama-7b-hf

from alpaca-lora.

gururise avatar gururise commented on August 27, 2024 4

Any links for models trained w/3-epochs on the new cleaned dataset?

from alpaca-lora.

mattreid1 avatar mattreid1 commented on August 27, 2024 4

Any links for models trained w/3-epochs on the new cleaned dataset?

I just finished training this 13B one but haven't got it to work yet (I'm using multiple GPUs so maybe that's the issue) https://huggingface.co/mattreid/alpaca-lora-13b

from alpaca-lora.

felri avatar felri commented on August 27, 2024 3

@collant can you help me understand how can I load the Lora model trained with the 52k dataset and use it to train on another data.json?

In finetune.py I can find the loading of the llama 7b model

model = LlamaForCausalLM.from_pretrained(
    "decapoda-research/llama-7b-hf",
    load_in_8bit=True,
    device_map=device_map,
)
tokenizer = LlamaTokenizer.from_pretrained(
    "decapoda-research/llama-7b-hf", add_eos_token=True
)

and after the lora config obj is being created

config = LoraConfig(
    r=LORA_R,
    lora_alpha=LORA_ALPHA,
    target_modules=TARGET_MODULES,
    lora_dropout=LORA_DROPOUT,
    bias="none",
    task_type="CAUSAL_LM",
)
model = get_peft_model(model, config)

does loading the Lora model from hf involves calling another function and loading that checkpoint? I can see that there is a save_pretrained function, maybe I need to load the Lora model via this? Sorry if this sounds confusing

edit: after a little bit more google I found this load_attn_procs function, maybe it's something around here

edit2: it seems that it was inside generate.py all along

    model = LlamaForCausalLM.from_pretrained(
        "decapoda-research/llama-7b-hf",
        load_in_8bit=True,
        torch_dtype=torch.float16,
        device_map="auto",
    )
    model = PeftModel.from_pretrained(
        model, "tloen/alpaca-lora-7b",
        torch_dtype=torch.float16
    )

from alpaca-lora.

ttio2tech avatar ttio2tech commented on August 27, 2024 1

Hello, you can find this 13B one here: https://huggingface.co/samwit/alpaca13B-lora

Otherwise, there is the 7B one here: https://huggingface.co/tloen/alpaca-lora-7b

Please note these are LoRA models they need the base model to work.

And here is the base model for the 7B: https://huggingface.co/decapoda-research/llama-7b-hf

can the original LLaMA-7B weights (consolidated.00.pth) be used? or can I convert it to hf?

from alpaca-lora.

HCBlackFox avatar HCBlackFox commented on August 27, 2024

Thank you

Hello, you can find this 13B one here: https://huggingface.co/samwit/alpaca13B-lora

Otherwise, there is the 7B one here: https://huggingface.co/tloen/alpaca-lora-7b

Please note these are LoRA models they need the base model to work.

And here is the base model for the 7B: https://huggingface.co/decapoda-research/llama-7b-hf

Thank you

from alpaca-lora.

0xbitches avatar 0xbitches commented on August 27, 2024

Is there a 30B-4bit lora out there? I think I read somewhere that finetuning in 4bit might not be supported?

from alpaca-lora.

aspctu avatar aspctu commented on August 27, 2024

30B LoRa adapters here https://huggingface.co/baseten/alpaca-30b

from alpaca-lora.

T-Atlas avatar T-Atlas commented on August 27, 2024

@collant can you help me understand how can I load the Lora model trained with the 52k dataset and use it to train on another data.json?

In finetune.py I can find the loading of the llama 7b model

model = LlamaForCausalLM.from_pretrained(
    "decapoda-research/llama-7b-hf",
    load_in_8bit=True,
    device_map=device_map,
)
tokenizer = LlamaTokenizer.from_pretrained(
    "decapoda-research/llama-7b-hf", add_eos_token=True
)

and after the lora config obj is being created

config = LoraConfig(
    r=LORA_R,
    lora_alpha=LORA_ALPHA,
    target_modules=TARGET_MODULES,
    lora_dropout=LORA_DROPOUT,
    bias="none",
    task_type="CAUSAL_LM",
)
model = get_peft_model(model, config)

does loading the Lora model from hf involves calling another function and loading that checkpoint? I can see that there is a save_pretrained function, maybe I need to load the Lora model via this? Sorry if this sounds confusing

edit: after a little bit more google I found this load_attn_procs function, maybe it's something around here

edit2: it seems that it was inside generate.py all along

    model = LlamaForCausalLM.from_pretrained(
        "decapoda-research/llama-7b-hf",
        load_in_8bit=True,
        torch_dtype=torch.float16,
        device_map="auto",
    )
    model = PeftModel.from_pretrained(
        model, "tloen/alpaca-lora-7b",
        torch_dtype=torch.float16
    )

Have you found solution? #44 I found this may help? But I still confuse with what is

from alpaca-lora.

diegolondrina avatar diegolondrina commented on August 27, 2024

Any links for models trained w/3-epochs on the new cleaned dataset?

+1

from alpaca-lora.

wafflecomposite avatar wafflecomposite commented on August 27, 2024

Please, report @larasatistevany for spamming.

https://support.github.com/contact/report-abuse?category=report-abuse&report=larasatistevany

-> I want to report abusive content or behavior.
-> I want to report SPAM, a user that is disrupting me or my organization's experience on GitHub, or a user who is using my personal information without my permission
-> A user is disrupting me or my organization's experience and productivity by posting SPAM off-topic or other types of disruptive content in projects they do not own.

Put this in the form:

spamming in issue comments
https://github.com/tloen/alpaca-lora/issues/52#issuecomment-1570561693
https://github.com/tloen/alpaca-lora/issues/52#issuecomment-1571059071

Thanks!

from alpaca-lora.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.