Code Monkey home page Code Monkey logo

Comments (8)

ashawkey avatar ashawkey commented on August 16, 2024 1

@xmk2222 there is a pure-pytorch implementation here.

from torch-ngp.

xmk2222 avatar xmk2222 commented on August 16, 2024

NVlabs/tiny-cuda-nn#63

Sadly, the --tcnn does not support multi-GPU training for sure.

from torch-ngp.

ashawkey avatar ashawkey commented on August 16, 2024

Yes, multi-GPU training for both CMD and GUI mode has not been supported now (the local_rank in trainer is not tested in this repo since it was borrowed from old projects).
In most cases the training speed should be fast enough on one GPU, may I know your using case for multi-GPU?

from torch-ngp.

slash-under avatar slash-under commented on August 16, 2024

(coming from the linked issue above)

I've been evaluating torch-ngp as a way to scale NeRF models as my dataset (750 4K images) no longer fits in any available GPU's VRAM -- only about 300 frames fit on a 40 GB A100. I saw there appeared to be some changes to only load portions of the training data at the time, could that support such a use-case?

from torch-ngp.

ashawkey avatar ashawkey commented on August 16, 2024

@slash-under that's strange, dataset size should not affect the runtime GPU memory, since it is only load one image at a time by default. Are you using the latest version, and not using the flat --preload for NeRFDataset?

from torch-ngp.

xmk2222 avatar xmk2222 commented on August 16, 2024

Thanks! Just looking for ways to speed up training, the training time of a single GPU is bearable.

Well, I'm still wondering if it is possible to implement everything in python including the hash-encoding? I think that would enable the training process to be distributed on multiple GPUs. Or perhaps the speedup gained by parallel computing would be cancelled out by the slowdown caused by the python implementation...

from torch-ngp.

hturki avatar hturki commented on August 16, 2024

hey @slash-under - apologies for tooting my own horn a bit, but I've run into similar issues where I've wanted to train NeRFs against thousands of high-resolution images. The best I could come up with is writing randomized batches of rgbs and rays across all of my images to disk as a preprocessing step, and then loading those from disk in chunks. You can take a look at https://github.com/cmusatyalab/mega-nerf/blob/main/mega_nerf/datasets/filesystem_dataset.py for inspiration and would be happy to hear of better suggestions.

from torch-ngp.

ashawkey avatar ashawkey commented on August 16, 2024

closed for now.

from torch-ngp.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.