Dr. Takayuki Osogami |
The team taught seven artificial neurons the word “SCIENCE” (one artificial neuron per bit) in a form of a bitmap image. So, the image of “SCIENCE” becomes:
00111001100111011110100010011001111
01000010010010010000100010100101000
01000010000010010000110010100001000
00110010000010011100101010100001110
00001010000010010000100110100001000
00001010010010010000100010100101000
01110001100111011110100010011001111.
The “1s” equate to the lines making up the letters, while the “0s” translate to the white space around the letters.
What these seven neurons can do all at once is read and write 7-bit information. The word “SCIENCE” is expressed as 7-bit x 35-bit of pattern sequences that equals
a 245 bits monochrome bitmap image. These seven neurons read and memorized each piece of the 7-bit information in the image. For example, “0100010” is the
tenth column of the whole image according to the learning order, and the neurons recall those pieces in the order they learned them. By memorizing the
word from left to right and from right to left, the neurons could recognize “SCIENCE” forward and backward, or in any order – like how we might solve a
jumble word puzzle or crossword puzzle.
![]() |
Figure 1: The DyBM successfully learns two target sequences and retrieves a particular sequence when an associated cue is presented. (Credit: Scientific Reports) |
More neurons. More memories.
Things get more complicated (and interesting) when these artificial neurons learn about different topics in different formats, such as the human evolution image, below. Takayuki’s team put 20 artificial neurons to the task of learning this image, which shows how we humans have evolved, from left to right. Why they used 20 artificial neurons this time? One column of the image showing the human evolution consists of 20 bits, and they made it consistent with it.
These neurons learned how the pieces of the image line up in the correct order of evolution – from apes to Homo sapiens. As Takayuki runs simulation, these neurons learn more over time, while detecting the mistakes they make – and making corrections per simulation. With each simulation, the neurons generate an image to show their progress in re-creating the image. It took just 19 seconds for the 20 neurons to learn the image correctly, as mapped out below.
Things get more complicated (and interesting) when these artificial neurons learn about different topics in different formats, such as the human evolution image, below. Takayuki’s team put 20 artificial neurons to the task of learning this image, which shows how we humans have evolved, from left to right. Why they used 20 artificial neurons this time? One column of the image showing the human evolution consists of 20 bits, and they made it consistent with it.
These neurons learned how the pieces of the image line up in the correct order of evolution – from apes to Homo sapiens. As Takayuki runs simulation, these neurons learn more over time, while detecting the mistakes they make – and making corrections per simulation. With each simulation, the neurons generate an image to show their progress in re-creating the image. It took just 19 seconds for the 20 neurons to learn the image correctly, as mapped out below.
![]() |
Figure 2: The DyBM learned the target sequence of human evolution. (Credit: Scientific Reports) |
Images and text are one thing. But neurons encompass all senses. So, Takayuki’s team put 12 of their artificial neurons to work learning music. Using a simplified version of the German folk song, Ich bin ein Musikante, each neuron was assigned to one of the 12 notes (Fa, So, Ra, Ti, Do, Re, Mi, Fa, So, Ra, Ti, Do). After 900,000 training sessions, they learned the sequential patterns of tones to the point of being able to generate a simplified version of the song.
The neurons learn
music much like we might: repetition from beginning to end to the point of
memorization. Currently, the 12 neurons only generate quarter notes, but simply
doubling the neurons to 24 gives the system the ability to comprehend
half-notes.
![]() |
Figure 3: The DyBM learned the target music. (Credit: Scientific Reports) |
Takayuki’s DyBM
not only memorizes and recalls sequential patterns, but it can also detect
anomalies in sequential patterns, and make predictions about future patterns.
It could be used to predict driving risks through car-mounted cameras, or
generate new music, or even detect and correct grammatical errors in text.
Takayuki’s work, currently funded by the Japan Science and Technology Agency’s
Core Research for Evolutionary Science and Technology (CREST), hopes to advance
his DyBM by integrating it with reinforcement learning techniques to optimally
act on the basis of such anomaly detection and prediction.
Speaking from a mental capacity of just a Machine Learning enthusiast, I found your accomplishment quite remarkable. To be frank it was hard for me to chew on all that you have done, however i did spend an hour or so to absorb what you are trying to accomplish.
ReplyDeleteIs there any scholarly article or paper that might help a person of my mental caliber understand it from the basics ?