Skip to main content

Researchers use brain-machine interface to create a sixth sense in rats

a rat up close and personal
Jean-Jacques Boujot/Flickr

In the medical field, most brain-machine interface research focuses on trying to replace lost sensory information, such as restoring a sense of touch to people with spinal cord injuries. However, a recent study has taken a different approach by using a brain-machine interface to augment existing sensory systems and create a “sixth sense” in rats.

“This constitutes an important step in the direction of ‘Cyber-Physical’ systems, which meld computers with the living brain,” senior author Dr. Tim Lucas, Assistant Professor of Neurosurgery at the University of Pennsylvania, told Digital Trends. He said the technology could be developed in the future to restore sensory experiences to people suffering from paralysis.

Brain-computer interfaces can be used to control everything from drones to bionic arms, and they’ve become a hot topic in emerging technology. Elon Musk is working on the Neuralink project to use cybernetic implants to allow people to interface with gadgets or software, and Facebook is working on its own brain-reading computer system. These projects are a long way from creating usable prototypes, however. Before humans can interface neurally with computers, researchers need to find a way to integrate incoming information from a computer into the brain.

The new study from Penn Medicine does just that, by implanting tiny electrodes into the brains of rats and feeding them information in the form of sensory feedback. The researchers began by surgically implanting the electrodes into the rats’ brains. Then they put the animals in a water maze which was painted black inside, with a platform hidden beneath the water that they needed to reach in order to escape.

The rats couldn’t see the platform, so they received no visual information about how to navigate the maze. But they did have information from the interface. The electrodes stimulated their brains to inform the rats where the platform was located relative to their current position, and the rats were able to use this information to reach the platform even in the darkness.

The researchers used a technique called intracortical microstimulation, which is much more precise than other kinds of brain stimulation (such as transcutaneous direct current stimulation). These other methods activate thousands or millions of neurons and other neural elements, while intracortical microstimulation only activates around ten elements. This means that the stimulation applied to the brain can be precisely targeted, giving researchers the ability to create a single, discrete perception instead of activating a whole brain area.

With this more precise stimulation, the researchers could target very specific brain areas to convey information. However, there’s a challenge. It’s not enough to simply stimulate a brain area and assume that the animal will be able to understand that information. One of the breakthroughs the team made was to show that the “Rat-Robot” could assimilate the information, processing the externally produced signals just as successfully as if it was using its natural-born senses.

There have been previous attempts to create a “sixth sense” for directions using external tools like a vibrating belt which can help visually impaired people navigate around their environment. However, there are limitations on who can use these external tools — they can’t be used by people with paralysis, for example, who cannot experience sensory feedback.

“One eventual application of this brain-computer device is to restore sensation to individuals who have suffered from spinal cord injury,” Lucas said. “A patient like Christopher Reeve can neither lift his finger, nor feel a needle jabbed into his finger. Christopher Reeve would have little use for a vibrating belt.”

Before the researchers could consider implanting a brain stimulation device into a human, they would need to conduct many more trials in animals to ensure the technology is safe. Eventually, though, they believe they can use a brain-computer devices to integrate computers into human brains.

That opens the door for applications which connect devices in the brains to devices elsewhere in the body. “Our long term vision is to link this system with implantable sensors in paralyzed limbs to provide a complete sensory experience for paralyzed patients,” Lucas said.

And this research isn’t only of interest in terms of helping people with disabilities. It could potentially open up a whole new field of brain-computer devices, such as biorobots which can perform search and rescue operations.

The findings are published in the journal PNAS.

Editors' Recommendations

Georgina Torbet
Georgina is the Digital Trends space writer, covering human space exploration, planetary science, and cosmology. She…
Digital Trends’ Top Tech of CES 2023 Awards
Best of CES 2023 Awards Our Top Tech from the Show Feature

Let there be no doubt: CES isn’t just alive in 2023; it’s thriving. Take one glance at the taxi gridlock outside the Las Vegas Convention Center and it’s evident that two quiet COVID years didn’t kill the world’s desire for an overcrowded in-person tech extravaganza -- they just built up a ravenous demand.

From VR to AI, eVTOLs and QD-OLED, the acronyms were flying and fresh technologies populated every corner of the show floor, and even the parking lot. So naturally, we poked, prodded, and tried on everything we could. They weren’t all revolutionary. But they didn’t have to be. We’ve watched enough waves of “game-changing” technologies that never quite arrive to know that sometimes it’s the little tweaks that really count.

Read more
Digital Trends’ Tech For Change CES 2023 Awards
Digital Trends CES 2023 Tech For Change Award Winners Feature

CES is more than just a neon-drenched show-and-tell session for the world’s biggest tech manufacturers. More and more, it’s also a place where companies showcase innovations that could truly make the world a better place — and at CES 2023, this type of tech was on full display. We saw everything from accessibility-minded PS5 controllers to pedal-powered smart desks. But of all the amazing innovations on display this year, these three impressed us the most:

Samsung's Relumino Mode
Across the globe, roughly 300 million people suffer from moderate to severe vision loss, and generally speaking, most TVs don’t take that into account. So in an effort to make television more accessible and enjoyable for those millions of people suffering from impaired vision, Samsung is adding a new picture mode to many of its new TVs.
[CES 2023] Relumino Mode: Innovation for every need | Samsung
Relumino Mode, as it’s called, works by adding a bunch of different visual filters to the picture simultaneously. Outlines of people and objects on screen are highlighted, the contrast and brightness of the overall picture are cranked up, and extra sharpness is applied to everything. The resulting video would likely look strange to people with normal vision, but for folks with low vision, it should look clearer and closer to "normal" than it otherwise would.
Excitingly, since Relumino Mode is ultimately just a clever software trick, this technology could theoretically be pushed out via a software update and installed on millions of existing Samsung TVs -- not just new and recently purchased ones.

Read more
AI turned Breaking Bad into an anime — and it’s terrifying
Split image of Breaking Bad anime characters.

These days, it seems like there's nothing AI programs can't do. Thanks to advancements in artificial intelligence, deepfakes have done digital "face-offs" with Hollywood celebrities in films and TV shows, VFX artists can de-age actors almost instantly, and ChatGPT has learned how to write big-budget screenplays in the blink of an eye. Pretty soon, AI will probably decide who wins at the Oscars.

Within the past year, AI has also been used to generate beautiful works of art in seconds, creating a viral new trend and causing a boon for fan artists everywhere. TikTok user @cyborgism recently broke the internet by posting a clip featuring many AI-generated pictures of Breaking Bad. The theme here is that the characters are depicted as anime characters straight out of the 1980s, and the result is concerning to say the least. Depending on your viewpoint, Breaking Bad AI (my unofficial name for it) shows how technology can either threaten the integrity of original works of art or nurture artistic expression.
What if AI created Breaking Bad as a 1980s anime?
Playing over Metro Boomin's rap remix of the famous "I am the one who knocks" monologue, the video features images of the cast that range from shockingly realistic to full-on exaggerated. The clip currently has over 65,000 likes on TikTok alone, and many other users have shared their thoughts on the art. One user wrote, "Regardless of the repercussions on the entertainment industry, I can't wait for AI to be advanced enough to animate the whole show like this."

Read more