I'll attempt to explain what I understand from the paper. This will be higher level, and abstracted because that's the best I can do as an undergraduate currently.
The DNC essentially gives the neural network external memory. This memory is controlled by read heads, and write heads. These heads are actually just part of the vector output of the neural network. Each of them has a key vector, which is used to find where in memory to read or write. This is done by finding the most similar vector in memory, using cosine distance. The neural network can then use the memory. It also has mechanisms to remember the order that the memory vectors were written in, to allow for sequential recall.
The network itself is trained by reinforcement learning generally. I'm not entirely sure what is done with the parts of the vectors that are not used at a given timestep, when only input and not output is needed.
Again, this is what I believe I was able to understand from the paper. Please let me know if I got anything wrong.
The DNC essentially gives the neural network external memory. This memory is controlled by read heads, and write heads. These heads are actually just part of the vector output of the neural network. Each of them has a key vector, which is used to find where in memory to read or write. This is done by finding the most similar vector in memory, using cosine distance. The neural network can then use the memory. It also has mechanisms to remember the order that the memory vectors were written in, to allow for sequential recall.
The network itself is trained by reinforcement learning generally. I'm not entirely sure what is done with the parts of the vectors that are not used at a given timestep, when only input and not output is needed.
Again, this is what I believe I was able to understand from the paper. Please let me know if I got anything wrong.