"A new “end-to-end” deep learning method for the prediction of behavioral states uses whole-cortex functional imaging that do not require preprocessing or pre-specified features. Developed by medical student AJIOKA Takehiro and a team led by Kobe University’s TAKUMI Toru, their approach also allows them to identify which brain regions are most relevant for the algorithm (pictured). The ability to extract this information lays the foundation for future developments of brain-machine interfaces. Credit: Ajioka Takehiro". (ScitechDaily, Breakthrough AI Predicts Mouse Movement With 95% Accuracy Using Brain Data)
The breakthrough in AI is to predict what mice will do next by measuring brain activity. This means the AI and morphing neural networks can read the mind. The system must take the EEG from the mice. Then that system will follow the mice's activity using cameras. After that, it will connect that EEG with the activity that it sees from images. In laboratories, the morphing neural network can collect data from brain activity in simulated situations.
The system can show images to mice about things like predators like falcons and cats. Then it can see how brain activity changes when the animal sees those things. So that thing is the thing that can open many secrets from nature.
Many times people talk about the morphing neural networks. Those networks are so-called sensing systems. They will collect data from multiple points and then they combine data. The morphing neural networks can be networked workstations. The input system can be regular keyboards. The network operates backstage hidden from the users. It connects and combines data that those people input into a new entirety.
The system can also use some sensors to take data. The morphing neural networks can observe which keys a person pushes when they use their workstations using surveillance cameras. That allows to input of data to the system using keyboards that are not physically connected. Or the person can use the keyboard images.
The idea is that the system follows how fingers move on the keyboard image, and then the person can make gestures to enter. Voice commands and keyboard combination is the most effective way to input data to computers manually. The BCI- (Brain-Computer interface) can collect data from a microphone and Wernicke's areas.
Wernicke lobes produce speech and the system can connect the EEG image to words that people use while they command computers. The system can collect data using the electrodes that are behind the ears. That is the area where the Wernicke's areas are. The system can look like sunglasses. And that allows them to give orders to computers and robots using the low sound levels.
Or it can use keyloggers or screen catchers. That means the morphing neural network must not integrated into the system. It can be a separate system that follows the surveillance camera images.
Or it can collect data from brain implants. That thing makes it possible to use animals as the tools that collect data from nature. The AI that can decode EEG and introduce that data on a computer screen can collect data from natural animals' lives. That system can also turn animals into intelligence systems. That we cannot hide.
The AI is a tool that can make many things better than humans. Morphing neural networks can connect information about the environment. And then it can connect that thing with the actions that it sees. When we think about the higher class AI and quantum computing. We can hear statements. Maybe that technology will transfer the morphing neural networks into history. The fact is that quantum computer-driven AI requires morphing neural networks for collecting information.
https://scitechdaily.com/breakthrough-ai-predicts-mouse-movement-with-95-accuracy-using-brain-data/
https://en.wikipedia.org/wiki/Wernicke%27s_area
Comments
Post a Comment