The idea of the BCI (Brain-Computer Interface) is taken from the biomechanic prosthesis. That gives people the ability to control prostheses by using neural signals. The modern versions of that prosthesis are using WLAN or Bluetooth communication.
And that means the electrodes of those advanced systems are under the skin. There are no external wounds in that kind of system. And the transmitter gets its electricity from similar batteries that are used in pacemakers. Or the more futuristic systems can use neuro-electricity as an energy source.
Paralyzed people can control their wheelchairs by using the BCI (Brain-Computer Interface). The key in the brainwave control robots is simple. The user of the robots uses those systems through a computer.
BMI (Brain-Machine Interface) requires the BCI to operate. And teaching the BCI can happen by simulation. The key for those systems is that the computer must recognize the EEG that the brain sends at certain moments.
And, the simulations can teach the BCI to connect the BCI to a certain action. The requirement for the perfect BCI system is that the computers know what kind of brain waves the brain sends when it wants to do something.
Those BCI systems development is not as difficult as people normally think. The developers can use simulators as an example, a driving simulator where the person drives by using a virtual device. The computer will record those brainwaves while the person uses this machine.
There are three stages in that process.
*The driver will control the virtual device by using regular controls. And the system will record the EEG in those moments. The system connects that EEG to the movements of the control systems.
When as an example driver turns the wheel or presses the gas pedal the system connects that action with the EEG. And then that record can use for making the BCI system.
*In the next stage the controller will not use physical control systems. That means the controller can make movements but the system compiles the EEG to the records that are made in the first moment.
*And in the last stage, that data can connect to the other vehicles. The BCI might require that person moves hands or feet.
But the goal of that kind of development process is that the user of that kind of system requires only a helmet where the electrodes are. The system can interact with users by using HUD glasses.
Or in the most advanced visions, the system interacts with brains by stimulating the cerebral cortex or brain core. That kind of simulation makes the robot the external body of the user. The user cannot make difference between the senses. That is coming from robots and the user's senses.
But how to eliminate the need to move hands or feet if the person uses BCI to control things like jet fighters? In some visions, the person just is paralyzed by using chemicals. Or who cares how the person who is remote-controlling some robot moves their hands or feet?
https://www.psychologytoday.com/us/basics/brain-computer-interface
https://scitechdaily.com/not-science-fiction-paralyzed-people-can-navigate-using-mind-controlled-wheelchairs/
https://en.wikipedia.org/wiki/Brain%E2%80%93computer_interface
https://designandinnovationtales.blogspot.com/
Comments
Post a Comment