Topic: mixture-of-experts Goto Github
Some thing interesting about mixture-of-experts
Some thing interesting about mixture-of-experts
mixture-of-experts,A sumary of MoE experimental setups across a number of different papers.
User: adamg012
mixture-of-experts,Hierarchical Mixture of Experts,Mixture Density Neural Network
User: amazaspshumik
mixture-of-experts,Multi-Task Learning package built with tensorflow 2 (Multi-Gate Mixture of Experts, Cross-Stitch, Ucertainty Weighting)
User: amazaspshumik
mixture-of-experts,From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)
User: avisoori1x
mixture-of-experts,PyTorch Implementation of the Multi-gate Mixture-of-Experts with Exclusivity (MMoEEx)
Organization: borealisai
mixture-of-experts,PyTorch implementation of "From Sparse to Soft Mixtures of Experts"
User: bwconrad
mixture-of-experts,MoLE: Cross-Domain Label-Adaptive Stance Detection
Organization: checkstep
Home Page: https://arxiv.org/abs/2104.07467
mixture-of-experts,PyTorch Re-Implementation of "The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer et al. https://arxiv.org/abs/1701.06538
User: davidmrau
mixture-of-experts,pytorch implementation of grok
User: dominiquegarmier
mixture-of-experts,A TensorFlow Keras implementation of "Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts" (KDD 2018)
Organization: drawbridge
mixture-of-experts,Tools for Analyzing Finite Mixture Models
User: dsy109
mixture-of-experts,Run Mixtral-8x7B models in Colab or consumer desktops
User: dvmazur
mixture-of-experts,Repository for our paper "See More Details: Efficient Image Super-Resolution by Experts Mining"
User: eduardzamfir
mixture-of-experts,Fast Inference of MoE Models with CPU-GPU Orchestration
Organization: efeslab
Home Page: https://arxiv.org/abs/2402.07033
mixture-of-experts,PyTorch implementation of Soft MoE by Google Brain in "From Sparse to Soft Mixtures of Experts" (https://arxiv.org/pdf/2308.00951.pdf)
User: fkodom
mixture-of-experts,The official code repo for the paper "Mixture of Stochastic Experts for Modeling Aleatoric Uncertainty in Segmentation". (ICLR 2023)
User: gaozhitong
Home Page: https://arxiv.org/pdf/2212.07328.pdf
mixture-of-experts,A toolbox for inference of mixture models
User: hanyas
mixture-of-experts,MoEL: Mixture of Empathetic Listeners
Organization: hltchkust
mixture-of-experts,Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).
User: inferflow
mixture-of-experts,A modular implementation of product of experts VAEs for multimodal data
User: jackgoffinet
mixture-of-experts,[arXiv'24] Multilinear Mixture of Experts: Scalable Expert Specialization through Factorization
User: james-oldfield
Home Page: http://james-oldfield.github.io/MMoE
mixture-of-experts,A curated reading list of research in Adaptive Computation, Dynamic Compute & Mixture of Experts (MoE).
User: koayon
mixture-of-experts,Implementation of the "the first large-scale multimodal mixture of experts models." from the paper: "Multimodal Contrastive Learning with LIMoE: the Language-Image Mixture of Experts"
User: kyegomez
Home Page: https://discord.gg/47ENfJQjMq
mixture-of-experts,Implementation of Switch Transformers from the paper: "Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity"
User: kyegomez
Home Page: https://discord.gg/GYbXvDGevY
mixture-of-experts,Decentralized deep learning in PyTorch. Built to train models on thousands of volunteers across the world.
Organization: learning-at-home
mixture-of-experts,The implementation of "Leeroo Orchestrator: Elevating LLMs Performance Through Model Integration"
Organization: leeroo-ai
Home Page: https://www.leeroo.com
mixture-of-experts,A library for easily merging multiple LLM experts, and efficiently train the merged LLM.
Organization: leeroo-ai
Home Page: https://www.leeroo.com/
mixture-of-experts,Some personal experiments around routing tokens to different autoregressive attention, akin to mixture-of-experts
User: lucidrains
mixture-of-experts,A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models
User: lucidrains
mixture-of-experts,Implementation of Soft MoE, proposed by Brain's Vision team, in Pytorch
User: lucidrains
mixture-of-experts,Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch
User: lucidrains
mixture-of-experts,GMoE could be the next backbone model for many kinds of generalization task.
User: luodian
mixture-of-experts,DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Organization: microsoft
Home Page: https://www.deepspeed.ai/
mixture-of-experts,Tutel MoE: An Optimized Mixture-of-Experts Implementation
Organization: microsoft
mixture-of-experts,"Towards Crowdsourced Training of Large Neural Networks using Decentralized Mixture-of-Experts" (NeurIPS 2020), original PyTorch implementation
User: mryab
Home Page: https://learning-at-home.github.io
mixture-of-experts,⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training
Organization: pjlab-sys4nlp
mixture-of-experts,Mixture-of-Experts for Large Vision-Language Models
Organization: pku-yuangroup
Home Page: https://arxiv.org/abs/2401.15947
mixture-of-experts,Efficient global optimization toolbox in Rust: bayesian optimization, mixture of gaussian processes, sampling methods
User: relf
Home Page: https://joss.theoj.org/papers/10.21105/joss.04737
mixture-of-experts,Mixture of Decision Trees for Interpretable Machine Learning
User: simsal0r
mixture-of-experts,Surrogate Modeling Toolbox
Organization: smtorg
Home Page: https://smt.readthedocs.io/en/stable
mixture-of-experts,PyTorch library for cost-effective, fast and easy serving of MoE models.
Organization: torchmoe
mixture-of-experts,Some recent state-of-the-art generative models in ONE notebook: (MIX-)?(GAN|WGAN|BigGAN|MHingeGAN|AMGAN|StyleGAN|StyleGAN2)(\+ADA|\+CR|\+EMA|\+GP|\+R1|\+SA|\+SN)*
User: tsc2017
Home Page: https://arxiv.org/abs/2007.06418
mixture-of-experts,This is the official repository of the papers "Parameter-Efficient Transfer Learning of Audio Spectrogram Transformers" and "Efficient Fine-tuning of Audio Spectrogram Transformers via Soft Mixture of Adapters".
User: umbertocappellazzo
Home Page: https://arxiv.org/abs/2312.03694
mixture-of-experts,[ICLR 2024 Spotlight] Code for the paper "Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing Policy"
Organization: unites-lab
Home Page: https://arxiv.org/abs/2310.01334
mixture-of-experts,[ICML 2022] "Neural Implicit Dictionary via Mixture-of-Expert Training" by Peihao Wang, Zhiwen Fan, Tianlong Chen, Zhangyang Wang
Organization: vita-group
mixture-of-experts,Machine learning code, derivatives calculation and optimization algorithms developed during the Machine Learning course at Universidade de Sao Paulo. All codes in Python, NumPy and Matplotlib with example in the end of file.
User: vivamoto
mixture-of-experts,Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*
User: xrsrke
mixture-of-experts,RealCompo: Dynamic Equilibrium between Realism and Compositionality Improves Text-to-Image Diffusion Models
User: yangling0818
mixture-of-experts,PyTorch implementation of moe, which stands for mixture of experts
User: yeonwoosung
mixture-of-experts,中文Mixtral混合专家大模型(Chinese Mixtral MoE LLMs)
User: ymcui
Home Page: https://arxiv.org/abs/2403.01851
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.