Code Monkey home page Code Monkey logo

llama2-accessory's Introduction

LLaMA2-Accessory: An Open-source Toolkit for LLM Development πŸš€


πŸ“– Document

πŸ€— HF Repo β€’ πŸ‘‹ join our WeChat β€’ πŸš€ Demo

πŸš€LLaMA2-Accessory is an open-source toolkit for pretraining, finetuning and deployment of Large Language Models (LLMs) and multimodal LLMs. This repo is mainly inherited from LLaMA-Adapter with more advanced features.🧠

✨Within this toolkit, we present SPHINX, a versatile multimodal large language model (MLLM) that combines a diverse array of training tasks, data domains, and visual embeddings.

News

  • [2024-3-7] We release the demos and codebase of Large-DiT-T2I πŸŽ‰.
  • [2024-2-17] We release a 3 and 7 Billion Large-DiT trained on ImageNet. Pretrained checkpoints and full training codebase are released πŸŽ‰.
  • [2024-1-27] SPHINX-MoE achieves 29.57% and 29.33% accuracy results on CMMMU-test and CMMMU-val respectively.
  • [2024-1-24] SPHINX-MoE achieves new SOTA performance (49.33%) on MMVP, higher than GPT-4V! πŸ”₯πŸ”₯πŸ”₯
  • [2024-1-20] SPHINX-MoE achieves SOTA performance on AesBench! πŸ”₯πŸ”₯πŸ”₯
  • [2024-1-18] LLaMA-Adapter is accepted by ICLR 2024!πŸŽ‰
  • [2024-1-12] We release SPHINX-Tiny built on the compact 1.1B TinyLlama that everyone can play with! πŸ”₯πŸ”₯πŸ”₯
  • [2024-1-5] OpenCompass now supports seamless evaluation of all LLaMA2-Accessory models. πŸ”₯πŸ”₯πŸ”—Doc
  • [2024-1-2] We release the SPHINX-MoE, a MLLM based on Mixtral-8x7B-MoE. πŸ”₯πŸ”₯πŸ”₯
  • [2023-12-12] SPHINX-V2 achieve outstanding results in InfiMM-Eval, ranking just below GPT4-V! πŸ”₯πŸ”₯πŸ”₯
  • [2023-12-11] We now support mixtral-8x7b inference and finetuning! πŸ”₯πŸ”₯πŸ”₯
  • [2023-12-08] We release OneLLM which aligns eight modalities to language using a unified framework!πŸ”₯πŸ”₯πŸ”₯
  • [2023-11-17] We release SPHINX-V2, the same architecture but enhanced capabilities! πŸ”₯πŸ”₯πŸ”₯
  • [2023.10.17] We release the demo, code, and model of SPHINX!πŸ”₯πŸ”₯
  • [2023.09.15] We now support Falcon 180B!πŸ”₯πŸ”₯
  • [2023.09.14] WeMix-LLaMA2-70B shows excellent performance on the OpenCompass benchmark!πŸ”₯πŸ”₯
  • [2023.09.02] We now support InternLMπŸ”₯
  • [2023.08.28] We release quantized LLM with OmniQuant, which is an efficient, accurate, and omnibearing (even extremely low bit) quantization algorithm. Multimodal version is coming soon
  • [2023.08.27] We now support CodeLLaMA and instruction finetuning on evol-code-alpaca
  • [2023.08.27] We release our documentation in a webbook format πŸ”—Check it out here
  • [2023.08.21] We release the Quantization codes and Evaluation result
  • [2023.08.05] We release the multimodel finetuning codes and checkpoints
  • [2023.07.23] Initial release πŸ“Œ

Features

Setup

βš™οΈ For environment installation, please refer to Environment Setup.

Model Usage

πŸ€– Instructions for model pretraining, finetuning, inference, and other related topics are all available in the document.

Frequently Asked Questions (FAQ)

❓ Encountering issues or have further questions? Find answers to common inquiries here. We're here to assist you!

Demos

πŸ’‘ Now, our model SPHINX supports generating high-quality bounding boxes and then present masks created by SAM for all objects within an image driven by input prompts. Give it a try here! πŸš€

Core Contributors

Chris Liu, Ziyi Lin, Guian Fang, Jiaming Han, Yijiang Liu, Renrui Zhang, Longtian Qiu, Yichi Zhang, Siyuan Huang

Project Leader

Peng Gao, Wenqi Shao, Shanghang Zhang

Hiring Announcement

πŸ”₯ We are hiring interns, postdocs, and full-time researchers at the General Vision Group, Shanghai AI Lab, with a focus on multi-modality and vision foundation models. If you are interested, please contact [email protected].

Citation

If you find our code and paper useful, please kindly cite:

@article{zhang2023llamaadapter,
  title = {LLaMA-Adapter: Efficient Finetuning of Language Models with Zero-init Attention},
  author={Zhang, Renrui and Han, Jiaming and Liu, Chris and Gao, Peng and Zhou, Aojun and Hu, Xiangfei and Yan, Shilin and Lu, Pan and Li, Hongsheng and Qiao, Yu},
  journal={arXiv preprint arXiv:2303.16199},
  year={2023}
}
@article{gao2023llamaadapterv2,
  title = {LLaMA-Adapter V2: Parameter-Efficient Visual Instruction Model},
  author={Gao, Peng and Han, Jiaming and Zhang, Renrui and Lin, Ziyi and Geng, Shijie and Zhou, Aojun and Zhang, Wei and Lu, Pan and He, Conghui and Yue, Xiangyu and Li, Hongsheng and Qiao, Yu},
  journal={arXiv preprint arXiv:2304.15010},
  year={2023}
}

Acknowledgement

Show More

License

Llama 2 is licensed under the LLAMA 2 Community License, Copyright (c) Meta Platforms, Inc. All Rights Reserved.

llama2-accessory's People

Contributors

chrisliu6 avatar enderfga avatar kriskrisliu avatar zrrskywalker avatar void721 avatar linziyi96 avatar csuhan avatar jiangzhengkai avatar artanic30 avatar lloongx avatar gaopengpjlab avatar helanhu avatar zhaoshitian avatar tmm1 avatar eltociear avatar zhuole1025 avatar lupantech avatar theia-4869 avatar dependabot[bot] avatar

Watchers

marsggbo avatar Kostas Georgiou avatar

Forkers

zsk97

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    πŸ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. πŸ“ŠπŸ“ˆπŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❀️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.