Comments (2)
@win10ogod I also have a similar question. When we are using the pass-through method, is there any logical way where we can select layers from each model? Can we use something like task arithmetic values to pick the most useful layers?
from mergekit.
I am pretty sure you are asking the same question as what I am looking at (if not sorry to hijack this post).
I read the paper Model soup mentioned by @cg123 here https://arxiv.org/pdf/2203.05482.pdf
When reading the section 4, we see they try to compare "soups" and "Ensembling"
If am not mistaken, my understanding is that Soups is well suited for models sharing the same initialization weights (seed) otherwise models take a completely different path and averaging weights is either irrelevant OR require a post training (finetuning) that may or may not be beneficial. On the other hand, Ensembling is suited for different models since it acts at logits level hence taking the best path mentioned in the title.
Ensembling is de facto superior to "Soup" (as they refer in the paper).
So the question is, do other methods than Linear emulate better ensembling for models that do not share the same initialization.
Am I correct ?
from mergekit.
Related Issues (20)
- Typo in your evolutionary merging tutorial causes error HOT 1
- what is the default model download path ? HOT 3
- Idea: Scaling the Down-Projection Matrix in 'Mixture of Experts' Models HOT 7
- The format of the inference result after model merging is not consistent with the base model inference format HOT 1
- DoRA HOT 1
- Training at Lower Context and Merging Large
- View size is not compatible with input tensor's size and stride/You are trying to save a non contiguous tensor Fix HOT 1
- RuntimeError: "svd_cuda_gesvdj" not implemented for 'BFloat16' HOT 4
- Mergekit-Evolve with vLLM enabled causes error if merge_method is linear HOT 2
- How are dimensions of multiple layer's matched?
- Merge only the transformer parts (including the input embedding layer) HOT 5
- Implementation of AdaMerging: Adaptive Model Merging for Multi-Task Learning
- Existing Mergekit algorithms to merge VLM with LLM?
- Qwen/Qwen1.5-1.8B MoE Merging fails HOT 3
- Mixed Precision Merging HOT 1
- Add support for `subfolder` loading
- How to merge a VLM and LLM with different model type.
- Merge of hidden_size
- Relax dependency versions
- EvoMerge Genome Bug HOT 2
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from mergekit.