Code Monkey home page Code Monkey logo

Comments (4)

Kim-Dongjun avatar Kim-Dongjun commented on July 19, 2024

It seems that I forgot to include the entropy term for computing the "lossless" bpd.
The entropy for eps=1e-5 with VPSDE is -7.92, and the offset term is 7, so we should add 0.92 value to the reported bpds (in my paper) to calculate the lossless bpd with entropy term, which is unacceptably bad. Please check the newly commited "evaluation.py" and "likelihood.py" for the update. We remove the reported bpd value at paperswithcode to prevent giving a misguided bpd report to colleagues. Also, I recognize my paper misleadingly write down the formula (in Eq. (4)), and I will modify the formula to include the entropy term. Thank you for your contribution.

The original purpose of the lossless bpd is to compute the true bpd value because the uniform dequantized bpd is too weak. I believe that the variational dequantization is not a good approach to develop further because it is highly unstable to train (I almost spent 3 months to train the flow network for the variational dequantization, but with "almost all" released flow networks, I failed for its unstable and suboptimal training. I believe that many people also fail to obtain the variational bpd). Therefore, any method is strongly needed to compute the true bpd of the checkpoint.

Any help of that? I think my computation is wrong because the bpd is too high, and I would appreciate if anyone could help me computing the lossless bpd correctly..

from soft-truncation.

baofff avatar baofff commented on July 19, 2024

I will try figure it out.

from soft-truncation.

LuChengTHU avatar LuChengTHU commented on July 19, 2024

How's it going?

from soft-truncation.

Kim-Dongjun avatar Kim-Dongjun commented on July 19, 2024

@LuChengTHU
In contrast to the discrete case, the continuous diffusion model seems to perform worse when we apply the lossless compression. Thus, up to now, variational dequantization is the best way to improve NLL. We have used the variational network suggested by DenseFlow (implemented with PyTorch), and we trained this variational network "after" the training of the score network. We used the training loss of ScoreFlow to train the variational network. See #3 for details.

from soft-truncation.

Related Issues (5)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.