Code Monkey home page Code Monkey logo

neuralturingmachine's People

Contributors

joeran avatar markpkcollier avatar reneraab avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

neuralturingmachine's Issues

replicating results

Hi!

I have been trying to replicate the copy-task results from your paper; however, after running python3 run_tasks.py --experiment_name=trial01 the copy task does not seem to converge.

I obtained an error of 31.89375 per sequence and it gave the following results:
trial01_20_0

I ran it now multiple times but I always got similar results (I believe that you ran the experiments 10 times and took the mean per plot).

I checked the defaults parameters but everything seems to be in accordance to your paper.

Is there something that I am missing?

By the way, great job with the project :-)!

recurrence on all state items?

NeuralTuringMachine/ntm.py

Lines 106 to 108 in 5644e3b

return NTM_output, NTMControllerState(
controller_state=controller_state, read_vector_list=read_vector_list, w_list=w_list, M=M)

It looks there is a recurrence happening on all elements in the state. Wouldn't you need
to do a tf.stop_gradient() on each of these items (except controller_state) to be consistent with the paper? At the very least there should be no recurrence on the memory M, right?

''tuple unpacking is not supported in Python3'' error

Hi Dear @MarkPKCollier

I wanted to run the code using python 3.5 on PyCharm, an error occured in file 'Ncross_analysis.py'

''''
C:\Users\Eduline\PycharmProjects\python\venv\Scripts\python.exe C:/Users/Eduline/Desktop/GEOMETRİ1/Ncross_analysis.py
File "", line 15
def getLine_pixelPoint((x1, y1), (x2, y2)):
^
SyntaxError: invalid syntax
''''
and it seems like the error is because 'tuple unpacking is not supported in Python3'.
321

Could you please help me to solve this problem?
How can I change the written code (in the photo), so that it will not be problematic ?

Have a good day

Error 'tuple unpacking is not supported in Python3'

Hi @MarkPKCollier

I wanted to run the code using python 3.6, an error occurs in file 'generate_data.py', Line: 263

File "D:\...\NeuralTuringMachine-master\generate_data.py", line 263
    outputs = map(lambda (source, dest, edge):
                         ^
SyntaxError: invalid syntax

and it seems that the error is because 'tuple unpacking is not supported in Python3'.

What can I do? switch to Python 2?

Thank you.

Sorting Task

Hi @MarkPKCollier

Thanks for your awesome implementation and making the code available in such a well packaged form!

You mentioned in the paper that you tried 3 of the 5 experiments. I was wondering whether you were also able to reproduce the other 2 experiments (in particular the sorting experiment), and whether you can share some of your thoughts on those as well.

Thanks

Could you add some examples?

I am really struggling to understand how to integrate your code in my tf/keras projects, could you provide some basic examples showing how to compile and train a network?

Implementing details of copy and repeated copy tasks

Say we are considering the copy task and we want to give a sequence of length 10 to our model. So, following the code, we actually give our controller a sequence of length 21, where the 11th element is an end of sequence marker (a vector of 1s) and the last 10 elements are vectors of 0s. Where does the code specify that the output should start at the end of sequence marker?

Now, for the repeated copy task, it doesn't seem like we should provide the target output size the same way we do in the copy task. For example, given a sequence of length 10 that is to be copied 2 times, it wouldn't make sense to give as input a sequence of 31 vectors since we want the model to learn when to stop copying. So how should the input be structured in this case?

Problem on associative recall

Using ntm as the default setting works fine on copy task, but it doesn't work on associative recall task.

I run for more than 20000 training steps with 5 hours and 1 GPU, and the sequence error doesn't drop.

Does associative recall task needs another setting?

where I can find the formula that generate head parameters

sorry for the silly question. I am an engineer, not a researcher(I don't have a strong backgroud on theory and math).
I try to apply this external memory idea on my work. I read the original NTM paper: 1410.5401 and your paper, I can't find the formula that generate these head parameters:
k, beta, g, s gamma.

Do I miss something or there are some "common knowledge" for this kind of parameter?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.