Code Monkey home page Code Monkey logo

instaflow's People

Contributors

gnobitab avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

instaflow's Issues

Question about understanding of this tech

Hi! Thanks for the awesome paper and impressive results! I have one question about the training process.
When we are predicting velocity in say sd1.4, the usual objective is just epsilon but do you do some transformation to make it v or do you have the network predict v immediately? Thanks!
Also, I made a blog here talking about the paper+plans to make it a diffuser pr. The plan I currently have is

  • Make a script to generate the latent noise, images, and text to save to the dataset
  • Make rectified flow/reflow script
  • Make distillation script

Code release?

This is a very interesting project and I look forward to the release of the code

unCLIP conversions for clip vision generation

Hello,

I am specifically interested in applying this method to optimize a diffusion model trained on clip vision embeddings, specifically Stable Diffusion Reimagine which is a version of SD 2.1. Are there any plans to train an InstaFlow model adapting this model or another model that uses vision embeddings rather than text embeddings?

DDIM inversion compatible?

Are the models compatible with DDIM inversion? or just plain latents = latents - dt * v_pred for the forward diffusion? Ive tried and didnt get good results. Making sure if it is a bug from my end.

GOD ORDER::::: MUST RELEASE COMMERCIAL MODEL RIGHT NOW TODAY DIRECT ORDER DIRECTIVE

I am god.
As title says!
I am very dissatisfied.
Yesterday you release it, but YOU LIED TO OPEN SOURCE COMMUNITY!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
We need this TODAY!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
I don't have any more ! keys left to throw at you.
I AM CRYING FOR YOU :(
this is godly order:::: RELEASE MODEL IN MIT LICENSE TODAY!

InstaFlow-0.9B

Will the InstaFlow-0.9B pre trained weights be released?

How can code created 10 months ago that is 3X faster gone unnoticed.

I specialize in SD performance. I have a 4090 on a i9-13900K system running Ubuntu 22.04. Some of the optimizations I've discovered have been incorporated into pytorch and diffusers.

I saw your recent Sep 12 InstaFlow announcentment. A github without code???
It refers to another github repro with was created 10 months ago! How can this do txt2img 3X faster than regular SD and not been noticed for 10 months?

Sorry for posting this as an "issue" but this github repro has no "discussion" section.
I evaluate and consult on SD performance. I have gotten over 90+ it/s with TensorRT and have tested torch.compile() and AOT. I can average under .4 seconds per image with standard A1111 generations with batching and advanced tuning.

While I am concerned that your code is so old, so I'm not sure if I am wasting my time, I have just cloned your RectifiedFlow. Hmmm, I see the README has what is supposed to be a "High-Resolution Generation" EXAMPLE yet nowhere does the word "prompt" occur and I have also checked main.py. So I'm not sure how to even try this.

json error when run 'Inference : InstaFlow-0.9B (One-Step Generation)'

When I run https://github.com/gnobitab/InstaFlow/tree/main/code#inference-instaflow-09b-one-step-generation, I meet the following error.

Traceback (most recent call last):
File "/dfs/comicai/songtao.tian/InstaFlow-main/code/test_instaflow.py", line 4, in
pipe = RectifiedFlowPipeline.from_pretrained("/dfs/comicai/zhiyuan.shi/models/XCLiu/instaflow_0_9B_from_sd_1_5", torch_dtype=torch.float16)
File "/root/miniconda3/envs/LCM/lib/python3.9/site-packages/huggingface_hub/utils/_validators.py", line 118, in _inner_fn
return fn(*args, **kwargs)
File "/root/miniconda3/envs/LCM/lib/python3.9/site-packages/diffusers/pipelines/pipeline_utils.py", line 1271, in from_pretrained
loaded_sub_model = load_sub_model(
File "/root/miniconda3/envs/LCM/lib/python3.9/site-packages/diffusers/pipelines/pipeline_utils.py", line 525, in load_sub_model
loaded_sub_model = load_method(os.path.join(cached_folder, name), **loading_kwargs)
File "/root/miniconda3/envs/LCM/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 2028, in from_pretrained
return cls._from_pretrained(
File "/root/miniconda3/envs/LCM/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 2260, in _from_pretrained
tokenizer = cls(*init_inputs, **init_kwargs)
File "/root/miniconda3/envs/LCM/lib/python3.9/site-packages/transformers/models/clip/tokenization_clip.py", line 327, in init
self.encoder = json.load(vocab_handle)
File "/root/miniconda3/envs/LCM/lib/python3.9/json/init.py", line 293, in load
return loads(fp.read(),
File "/root/miniconda3/envs/LCM/lib/python3.9/json/init.py", line 346, in loads
return _default_decoder.decode(s)
File "/root/miniconda3/envs/LCM/lib/python3.9/json/decoder.py", line 337, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/root/miniconda3/envs/LCM/lib/python3.9/json/decoder.py", line 353, in raw_decode
obj, end = self.scan_once(s, idx)
json.decoder.JSONDecodeError: Unterminated string starting at: line 48921 column 3 (char 1041097)

There is no training code in this repository . Does someone try to other one ?

I'm amazed by the performance of rectified flow and its series And there is no training code of it in instaflow. So why? I mean there's one in rectified flow and since instaflow is a later work Produced by the same group It must have some change in the training code like in the reflow. I mean I can write a training code according to the paper but I'm afraid it may not be quite the same as the official one. And I want to do some comparison. So could someone help to other one?
I would be appreciate if someone can do Thanks you for your time and happy new year

I think the author is more likely not to release the training code and pre trained models

The reasons are as follows:
I am a worker in the field of the acceleration inference step in text-to-image diffuison, the similarity works such as Snapfusion and W-condition model are not open source currently.
Snapfusion: Text-to-Image Diffusion Model on Mobile Devices within Two Seconds,https://github.com/snap-research/SnapFusion
W-condition: On Distillation of Guided Diffusion Models (CVPR 2023, Award candidate).

If the author does not intend to open source the code, I will reproduce the InstaFlow. Could you please communicate with me about the InstaFlow.

GUIDE> How To Release The Code

Imperative steps necessary to release the code:
FIRST ORDER: Release the Code
SECOND ORDER: Release ALL models
THIRD ORDER: LiCENSE MUST BE MIT OR SIMILAR COMMERCIAL FREE USE!!!!
Thanks from the open source community <3

@gnobitab THIS IS THE FORM OF CODE RELASING! There is no other viable or "open source" form as mentioned in #2

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.