Code Monkey home page Code Monkey logo

Comments (5)

subeeshvasu avatar subeeshvasu commented on July 26, 2024

I am also trying to do the same. After inspecting the code, I figured out the following solution, to train with more than one data volumes.
Let's assume that we have two training volumes at hand.

  • First, generate the coordinates for both volumes, but with different tags. i.e., while using 'build_coordinates.py' we could use "validation1" as the tag for volume 1 and "validation2" as the tag for volume 2.
    eg. python build_coordinates.py --partition_volumes validation1:address_to_volume1_af.h5:af_file,validation2:address_to_volume2_af.h5:af_file --coordinate_output ..
  • Next, change the arguments corresponding to input volume and corresponding labels:
    data_volumes
    from
    validation1:address_to_volume_1.h5:raw
    to
    validation1:address_to_volume_1.h5:raw,validation2:address_to_volume_2.h5:raw

Repeat the above step for
label_volumes
too

  • Now run the training code as usual.

from ffn.

mk123qwe avatar mk123qwe commented on July 26, 2024

How to save the model according to the set conditions, not periodically.
I noticed that MonitoredSession is a very inefficient method.

from ffn.

mk123qwe avatar mk123qwe commented on July 26, 2024

I am also trying to do the same. After inspecting the code, I figured out the following solution, to train with more than one data volumes.
Let's assume that we have two training volumes at hand.

  • First, generate the coordinates for both volumes, but with different tags. i.e., while using 'build_coordinates.py' we could use "validation1" as the tag for volume 1 and "validation2" as the tag for volume 2.
    eg. python build_coordinates.py --partition_volumes validation1:address_to_volume1_af.h5:af_file,validation2:address_to_volume2_af.h5:af_file --coordinate_output ..
  • Next, change the arguments corresponding to input volume and corresponding labels:
    data_volumes
    from
    validation1:address_to_volume_1.h5:raw
    to
    validation1:address_to_volume_1.h5:raw,validation2:address_to_volume_2.h5:raw

Repeat the above step for
label_volumes
too

  • Now run the training code as usual.

Hello, I'm confused that this program has no validation set.

from ffn.

subeeshvasu avatar subeeshvasu commented on July 26, 2024

I am also trying to do the same. After inspecting the code, I figured out the following solution, to train with more than one data volumes.
Let's assume that we have two training volumes at hand.

  • First, generate the coordinates for both volumes, but with different tags. i.e., while using 'build_coordinates.py' we could use "validation1" as the tag for volume 1 and "validation2" as the tag for volume 2.
    eg. python build_coordinates.py --partition_volumes validation1:address_to_volume1_af.h5:af_file,validation2:address_to_volume2_af.h5:af_file --coordinate_output ..
  • Next, change the arguments corresponding to input volume and corresponding labels:
    data_volumes
    from
    validation1:address_to_volume_1.h5:raw
    to
    validation1:address_to_volume_1.h5:raw,validation2:address_to_volume_2.h5:raw

Repeat the above step for
label_volumes
too

  • Now run the training code as usual.

Hello, I'm confused that this program has no verification set.
I didn't understand. You meant: "No setup/data to check monitor the improvement in segmentation?". If yes, I agree. I would also like to figure out a way to save the network weights based on some sort of quantitative evaluation. I haven't explored that direction yet. Let me know if you figure out a way to do it. Thanks.

from ffn.

mk123qwe avatar mk123qwe commented on July 26, 2024

I am also trying to do the same. After inspecting the code, I figured out the following solution, to train with more than one data volumes.
Let's assume that we have two training volumes at hand.

  • First, generate the coordinates for both volumes, but with different tags. i.e., while using 'build_coordinates.py' we could use "validation1" as the tag for volume 1 and "validation2" as the tag for volume 2.
    eg. python build_coordinates.py --partition_volumes validation1:address_to_volume1_af.h5:af_file,validation2:address_to_volume2_af.h5:af_file --coordinate_output ..
  • Next, change the arguments corresponding to input volume and corresponding labels:
    data_volumes
    from
    validation1:address_to_volume_1.h5:raw
    to
    validation1:address_to_volume_1.h5:raw,validation2:address_to_volume_2.h5:raw

Repeat the above step for
label_volumes
too

  • Now run the training code as usual.

Hello, I'm confused that this program has no verification set.
I didn't understand. You meant: "No setup/data to check monitor the improvement in segmentation?". If yes, I agree. I would also like to figure out a way to save the network weights based on some sort of quantitative evaluation. I haven't explored that direction yet. Let me know if you figure out a way to do it. Thanks.

I don't understand enough about FFN training process, which is different from what I know about segmentation tasks.
"MonitoredTrainingSession" is different from Session.You can try "Supervisor".

from ffn.

Related Issues (19)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.