AI’s memory-forming mechanism found to be strikingly similar to the brain’s

Researchers at the Institute for Basic Science (IBS) in South Korea have discovered a striking similarity between AI memory processing of transformer models and the hippocampus of the human brain.

The finding provides a novel perspective on memory consolidation, a process that transforms short-term memories into long-term ones in AI systems.

Hippocampus role in memory

The team focused on memory consolidation through the NMDA receptor in the hippocampus, which facilitates learning and memory formation. When a brain chemical called glutamate is present, the nerve cell undergoes excitation, but a magnesium ion acts as a small gatekeeper blocking the door. When this ionic gatekeeper steps aside, substances are allowed to flow into the cell. This is the process that allows the brain to create and keep memories.

Low-cost, high-performance AI systems

The team discovered that large learning models, such as ChatGPT, seem to use a gatekeeping process similar to the brain’s NMDA receptor.

In the animal brain, a low magnesium level is known to weaken memory function. The researchers found that long-term memory in the transformer can be improved by mimicking the NMDA receptor. Similar to how changing magnesium levels affect memory strength, tweaking the transformer’s parameters to reflect the gating action of the NMDA receptor led to enhanced memory in the AI model.

CHA Meeyoung, a data scientist in the team, notes that “the human brain is remarkable in how it operates with minimal energy, unlike the large AI models that need immense resources. Our work opens up new possibilities for low-cost, high-performance AI systems that learn and remember information like humans.”

Let us know your thoughts! Sign up for a Mindplex account now, join our Telegram, or follow us on Twitter

Deep neural networks mimic human hearing

A new study from MIT has found that computational models derived from machine learning are moving closer to mimicking the structure and function of the human auditory system.

This could help researchers design better hearing aids, cochlear implants, and brain-machine interfaces, the researchers suggest.

In this largest study yet of deep neural networks trained to perform auditory tasks, the MIT team showed that most of these models share properties seen in the human brain when people are listening to the same sounds.

Mimicking the human auditory system

When trained on auditory inputs, including background noise, the models closely mimic the activation patterns of the human auditory cortex, matching those seen in fMRI scans of people listening to the same sounds.

“This is the most comprehensive comparison of these kinds of models to the auditory system so far,” says Josh McDermott, an associate professor of brain and cognitive sciences at MIT, a member of MIT’s McGovern Institute for Brain Research and Center for Brains, Minds, and Machines, and the senior author of the study, which appears in PLOS Biology.

Auditory models

For this study, the researchers analyzed nine publicly available deep neural network models that had been trained to perform auditory tasks. They also created 14 models of their own, based on two different architectures.

Most of these models were trained to perform a single task: recognizing words, identifying the speaker, recognizing environmental sounds, and identifying musical genre, while two of them were trained to perform multiple tasks.

The researchers found that models that had been trained on different tasks were better at replicating different aspects of audition. For example, models trained on a speech-related task more closely resembled speech-selective areas.

Better models

McDermott’s lab now plans to develop models that are even more successful at reproducing human brain responses. Such models could help scientists learn more about how the brain may be organized, and could also be used to help develop better hearing aids, cochlear implants, and brain-machine interfaces.

“A goal of our field is to end up with a computer model that can predict brain responses and behavior. We think that if we are successful in reaching that goal, it will open a lot of doors,” McDermott says.

The research was funded by the National Institutes of Health, an Amazon Fellowship from the Science Hub, an International Doctoral Fellowship from the American Association of University Women, an MIT Friends of McGovern Institute Fellowship, a fellowship from the K. Lisa Yang Integrative Computational Neuroscience (ICoN) Center at MITand a Department of Energy Computational Science Graduate Fellowship.

Citation: Greta Tuckute, Jenelle Feather, Dana Boebinger, Josh H. McDermott. December 13, 2023, PLOS Biology, (open-access)

Let us know your thoughts! Sign up for a Mindplex account now, join our Telegram, or follow us on Twitter

Scientists unveil complete cell map of a whole mammalian brain

An NIH-funded international team of researchers has created a complete cell atlas of a whole mammalian brain, a first. A map for the mouse brain, it describes the type, location, and molecular information of more than 32 million cells and provides information on connectivity between these cells.

The mouse is the most commonly used vertebrate experimental model in neuroscience research, and this cellular map paves the way for a greater understanding of the human brain. The cell atlas also lays the foundation for developing a new generation of precision therapeutics for people with mental and neurological disorders of the brain.

The findings were funded by the National Institutes of Health’s Brain Research Through Advancing Innovative Neurotechnologies® Initiative, or The BRAIN Initiative®, and appear in a collection of 10 papers published in Nature.

“The mouse atlas has brought the intricate network of mammalian brain cells into unprecedented focus, giving researchers the details needed to understand human brain function and diseases,” said Joshua A. Gordon, M.D., Ph.D., Director of the National Institute of Mental Health, part of the National Institutes of Health.

Unprecedented map of cellular organization and diversity

The cell atlas describes the types of cells in each region of the mouse brain and their organization within those regions. It also provides a detailed catalog of the cell’s transcriptome—the complete set of gene readouts in a cell, which contains instructions for making proteins and other cellular products.

The atlas also characterizes the cell epigenome—chemical modifications to a cell’s DNA and chromosomes that alter the way the cell’s genetic information is expressed—detailing thousands of epigenomic cell types and millions of candidate genetic regulation elements for different brain cell types.

A blueprint of transmitted chemical signals

The atlas also provides an accounting of the neurotransmitters and neuropeptides used by different cells and the relationship among cell types within the brain. This information can be used as a detailed blueprint for how chemical signals are initiated and transmitted in different parts of the brain. Those electrical signals are the basis for how brain circuits operate and how the brain functions overall.

“This product is a testament to the power of this unprecedented, cross-cutting collaboration and paves our path for more precision brain treatments,” said John Ngai, Ph.D., Director of the NIH BRAIN Initiative.

Of the 10 studies included in this collection, seven are funded through the NIH BRAIN Initiative Cell Census Network (BICCN), and two are funded through the larger NIH BRAIN Initiative.

The core aim of the BICCN, a groundbreaking, cross-collaborative effort to understand the brain’s cellular makeup, is to develop a comprehensive inventory of the cells in the brain—where they are, how they develop, how they work together, and how they regulate their activity—to better understand how brain disorders develop, progress, and are best treated.

Next big step: completing the cell maps of the human brain and the nonhuman primate brain

“By leveraging the unique nature of its multi-disciplinary and international collaboration, the BICCN was able to accomplish what no other team of scientists has been able to before,” said Ngai. “Now we are ready to take the next big step—completing the cell maps of the human brain and the nonhuman primate brain.”

The BRAIN Initiative Cell Atlas Network (BICAN) is the next stage in the NIH BRAIN Initiative’s effort to understand the cell and cellular functions of the mammalian brain.

Aim: revolutionize neuroscience research

Acording to NIH, BICAN is a transformative project that, together with two other large-scale projects—the BRAIN Initiative Connectivity Across Scales and the Armamentarium for Precision Brain Cell Access—aim to revolutionize neuroscience research by illuminating foundational principles governing the circuit basis of behavior and informing new approaches to treating human brain disorders.

Citation: Yao, Z., van Velthoven, C. T. J., Kunst, M., Zhang, M., McMillen, D., Lee, C., Jung, W., Goldy, J., Abdelhak, A., Aitken, M., Baker, K., Baker, P., Barkan, E., Bertagnolli, D., Bhandiwad, A., Bielstein, C., Bishwakarma, P., Campos, J., Carey, D., … Zeng, H. (2023). A high-resolution transcriptomic and spatial atlas of cell types in the whole mouse brain. Nature (open-access)

Let us know your thoughts! Sign up for a Mindplex account now, join our Telegram, or follow us on Twitter

A brain-scale neuromorphic supercomputer

Would it be possible one day to create a supercomputer capable of emulating networks at the scale of the human brain?

Yes. It’s called Deep South. And researchers from the International Centre for Neuromorphic Systems (ICNS) at Western Sydney University plan to have it operational by April 2024, they announced at this week’s 2023 ICNS NeuroEng Workshop.

Mimics biological processes, so less power required

DeepSouth uses a neuromorphic system that mimics biological processes, using hardware to efficiently emulate large networks of spiking neurons at 228 trillion synaptic operations per second—rivaling the estimated rate of operations in the human brain.

DeepSouth is purpose-built to operate like networks of neurons, so it will require less power and enable greater efficiencies, says CNS Director Professor André van Schaik. This contrasts with supercomputers optimized for more traditional computing loads, which are power-hungry.

“Progress in our understanding of how brains compute using neurons is hampered by our inability to simulate brain like networks at scale,” he said. “Simulating spiking neural networks on standard computers using graphics processing units (GPUs) and multicore central processing units (CPUs) is just too slow and power intensive. Our system will change that.”

Large-scale AI and other applications

“This platform will progress our understanding of the brain and develop brain-scale computing applications in diverse fields, including sensing, biomedical, robotics, space, and large-scale AI applications.”

Practically, this will lead to advances in smart devices, such as mobile phones, sensors for manufacturing and agriculture, and less power-hungry and smarter AI applications, he said. It will also enable a better understanding of how a healthy or diseased human brain works.

Western Sydney University’s ICNS team collaborated with partners across the neuromorphic field in developing this ground-breaking project, with researchers from the University of Sydney, University of Melbourne, and University of Aachen, Germany.

The supercomputer is aptly named DeepSouth, paying homage to IBM’s TrueNorth system, which initiated efforts to build machines simulating large networks of spiking neurons, and Deep Blue, which was the first computer to become a world chess champion. The name is also a nod to its geographical location.

Let us know your thoughts! Sign up for a Mindplex account now, join our Telegram, or follow us on Twitter

Mind-reading AI turns thoughts into text

Researchers at the University of Technology Sydney (UTS) have developed a portable, non-invasive system that can decode silent thoughts and turn them into text.

The technology could aid communication for people who are unable to speak due to illness or injury, including stroke or paralysis. It could also enable seamless communication between humans and machines, such as operating a bionic arm or robot.

The research at the GrapheneX-UTS Human-centric Artificial Intelligence Centre was led by Distinguished Professor CT Lin, Director of the GrapheneX-UTS HAI Centre

AI learning process

In the study, participants silently read passages of text while wearing a cap that recorded electrical brain activity through their scalp, using an electroencephalogram (EEG). An AI model called DeWave, developed by the researchers, translates these EEG signals into words and sentences . 

UNIVERSITY OF TECHNOLOGY SYDNEY MEETING Neural Information Processing Systems, FUNDER; Australian Research Council, GrapheneX

“It is the first to incorporate discrete encoding techniques in the brain-to-text translation process, introducing an innovative approach to neural decoding,” Lin said.

No need for invasive Neuralink-type implants or an MRI machine

He also said the new system eliminates the need to translate brain signals to language, which has required surgery to implant electrodes in the brain, such as Elon Musk’s Neuralink, or scanning in an MRI machine, which is large, expensive, and difficult to use in daily life.

The UTS research was carried out with 29 participants, so it’s likely to be more robust and adaptable than previous decoding technology that has only been tested on one or two individuals at a time.

Accuracy to be improved

The translation accuracy score is currently around 40% on BLEU-1 (a number between zero and one that measures the similarity of the machine-translated text to a set of high-quality reference translations). The researchers hope to see this improve to a level that is comparable to traditional language translation or speech recognition programs, which is closer to 90%.

The research follows on from previous brain-computer interface technology developed by UTS in association with the Australian Defence Force. It uses brainwaves to command a quadruped robot, demonstrated in this ADF video.

A paper will be presented at the Dec. 12 at the NeurIPS conference, to be held in New Orleans on December 12, 2023, and will be linked here.

EurekAlert! The Global Source for Science News
AAAS - American Association for the Advancement of Science

Let us know your thoughts! Sign up for a Mindplex account now, join our Telegram, or follow us on Twitter

Ultrasound microrobots guide medications to treat brain tumors, other conditions

Brain tumors, brain hemorrhages and neurological and psychological conditions are often hard to treat with medication. Even when effective drugs are available, they tend to have severe side effects because they circulate throughout the brain, not just the area they are meant to treat.

So researchers at ETH Zurich, the University of Zurich and University Hospital Zurich have now developed mini-​transporters that can be guided through the dense maze of blood vessels in the brain of an animal, using ultrasound. Compared to magnetic fields and other methods, “ultrasound is widely used in the medical field, is safe, and penetrates deep into the body,” said Daniel Ahmed, Professor of Acoustic Robotics at ETH Zurich and supervisor of the study.

Ultrasound-guided microbubbles “likely to be approved”

The gas-​filled 1.5 micrometer microbubbles are coated in lipids, the substances that biological cell membranes are made of. The bubbles are used as contrast materials and are guided through blood vessels.

“Since these bubbles, or vesicles, are already approved for use in humans, it’s likely that our technology will be approved and used in treatments for humans more quickly than other types of microvehicles currently in development,” says Ahmed, who was awarded a Starting Grant by the European Research Council ERC in 2019 for his project to research and develop this technology.

Ultrasound-​guided microbubbles also dissolve in the body once they’ve done their job, unlike magnetic-based microvehicles, and are small and smooth. “This makes it easy for us to guide them along narrow capillaries,” says Alexia Del Campo Fonseca, a doctoral student in Ahmed’s group and lead author of the study.

Tests in mice

Ahmed’s team has tested this method on blood vessels in the brains of mice. They also used ultrasound to hold the vesicles in place and guide them through the brain vessels against the direction of blood flow, including convoluted blood vessels.

The researchers attached four small transducers to the outside of each mouse’s skull. These devices generate vibrations in the ultrasonic range, which spread through the brain as waves.

To create the imaging for this study, the researchers used two-​photon microscopy. In the future, they want to use ultrasound itself for imaging. The researchers next plan to use these medical devices for treating cancer, stroke and psychological conditions. The researchers’ next step will be to attach drug molecules to the outside of the bubble casing for transport. They want to enhance the entire method to the point at which it can be used in humans, hoping it will one day provide the basis for the development of new treatments.

Citation: Alexia Del Campo FonsecaChaim GlückJeanne DrouxYann FerryCarole FreiSusanne WegenerBruno WeberMohamad El Amki,Daniel Ahmed I2023). Ultrasound trapping and navigation of microrobots in the mouse brain vasculature. Nature Communications volume 14, Article number: 5889 (21-Sep-2023). ticles/s41467-023-41557-3 (open-access)

Let us know your thoughts! Sign up for a Mindplex account now, join our Telegram, or follow us on Twitter

Blue light exposure can damage neuronal function and induce aging

In a study on fruit flies, daily low-intensity blue-light exposure—similar to that experienced daily by billions of humans in the form of LED lighting and device screens—changed flies at the sub-cellular level, affecting processes related to aging and circadian rhythms.

Xiaoyun Wang and colleagues exposed fruit flies (Drosophila melanogaster) to different durations of daily low-intensity blue light, and then analyzed the consequences to the cellular makeup of the insects, compared to flies raised in darkness.

Blue light’s effects on flies

The authors measured blue light’s effects on N6-Methyladenosine (m6A), an RNA modification that is common across the tree of life and plays a role in a wide range of processes.

The authors found that blue light induced genetic transcriptomic** (m6A epitranscriptomic) and metabolomic reprogramming.

Damage to neuronal function

The RNA profiles of 25-day-old male fly heads exposed to blue light were significantly different from 25-day-old male dark-raised flies’ heads, say the authors, and the types of genes that were up- and down-regulated* suggest that blue light exposure can damage neuronal function.

In comparisons of whole-body transcriptomes (the protein-coding part of an organism’s genome) of RNA molecules that are present in cells, however, age created bigger differences between groups than light exposure, suggesting that eye and brain tissues of Drosophila flies are the major tissues affected by blue light.

Differences in mRNA m6A levels and other m6A-related markers between light- and dark-raised flies also indicate that m6A methylation (which regulates reproductive system development) was involved in the impacts of blue light on Drosophila.

Hazards with blue light

According to the authors, attention should be paid to the potential hazards of cumulative blue light exposure in humans because blue light can influence the circadian rhythm and processes mediated by melanopsin-expressing intrinsically photosensitive retinal ganglion cells.

* The process by which a cell increases and decreases (respectively) the production and quantities of its cellular components, such as RNA and proteins, in response to an external stimulus

** The set of all the ribonucleic acid (RNA) molecules (called “transcripts”) expressed in some given entity, such as a cell, tissue, or organism

Citation: Jia Huang, Fan Zhou, Huanchan Zhou, Xiaoqi Zheng, Zhengyi Huo, Meng Yang, Zihe Xu, Runzhou Liu, Luoluo Wang, Xiaoyun Wang, Systematic assessment of transcriptomic and metabolic reprogramming by blue light exposure coupled with aging, PNAS Nexus, Volume 2, Issue 12, December 2023, pgad390, (open-access)

Let us know your thoughts! Sign up for a Mindplex account now, join our Telegram, or follow us on Twitter

Solar-powered atmospheric water harvesting

Researchers from Shanghai Jiao Tong University in China have developed a new solar-powered atmospheric water harvesting technology that could help provide enough drinking water for people to survive in dryland areas in water-stressed countries.

The United Nations estimates that 3.5 million die every year from water-related diseases and the areas most in need of improved drinking water are also located in some of the sunniest places in the world.

Multiple uses

“This atmospheric water harvesting technology can be used to increase the daily water supply needs, as household drinking water, industrial water, and water for personal hygiene,” said author Ruzhu Wang.

The researchers synthesized a super-hygroscopic gel using plant derivatives and hygroscopic salts that were capable of absorbing and retaining an unparalleled amount of water. One kilogram of dry gel could adsorb 1.18 kilograms of water in arid atmospheric environments and up to 6.4 kilograms in humid atmospheric environments, they estimate. This hygroscopic gel was simple and inexpensive to prepare and would consequently be suitable for large-scale preparation.

In addition to daily water production, these sorbent materials that harvest atmosphere water could also play an important role in future applications such as dehumidification, agriculture irrigation, and thermal management for electronic devices, according to the researchers.

In an outdoor prototype demonstration, the team found it released adsorbed water even in the morning or afternoon when the sun is weak. The system could also achieve simultaneous adsorption and desorption during the daytime.

The team now plans to work to achieve simultaneous adsorption and desorption, using renewable energy to maximize daily water yield per unit mass of adsorbent.

Citation: Chengjie Xiang, Xinge Yang, Fangfang Deng, Zhihui Chen, and Ruzhu Wang, Dec. 5, 2023, Applied Physics Reviews (open-access)

Let us know your thoughts! Sign up for a Mindplex account now, join our Telegram, or follow us on Twitter

Wearable helmet with magnetic sensors records brain functions while patient is moving

A new wearable brain scanner can accurately record magnetic fields generated by brain activity while people are in motion. It will enable researchers to learn more about brain development and neurological conditions that can affect movement, including autism, epilepsy, stroke, concussion, and Parkinson’s disease, say researchers. 

“This advance could also make it easier to conduct brain scans in young children and individuals with neurological disorders, who can’t always remain still in conventional scanners,” said Niall Holmes, Ph.D., a Mansfield Research Fellow in the School of Physics and Astronomy at the University of Nottingham, who led the research. “Unconstrained movement during a scan opens a wealth of possibilities for clinical investigation and allows a fundamentally new range of neuroscientific experiments.”

How magnetic fields are recorded

When neurons interact, they generate a small electric current. This current produces a magnetic field that can be detected, recorded, and analyzed by sensitive magnetic sensors, using a technique called magnetoencephalography (MEG).

MEG technology can record brain signals every millisecond. By overlaying the neuronal sources of these magnetic fields onto an anatomical image of the brain, clinicians can visualize where and when specific brain activities originate. However, current MEG systems are bulky and rigid, like an old-fashioned hair dryer (where you must keep your head still for a while). Also, these sensors require cooling at or below freezing temperatures, so they can’t be placed directly on your scalp.

Now researchers at the University of Nottingham have used a new generation of magnetic field sensors called “optically pumped magnetometers” (OPMs), which operate at room temperature and can be placed close to the head, enhancing data quality. They are also flexible, allowing children and adults to move during scanning. 

The researchers also designed a magnetic shielding system that would cancel out or compensate for external magnetic fields (see “Technical background: Designing a matrix coil system” below).

Research applications

A company co-founded by Holmes and his colleagues is selling the OPM-MEG systems (which include a magnetically shielded room) to research centers in North America and Europe to conduct a variety of neuroscientific experiments.

One of the U.S. centers to use the MEG-OPM system is Virginia Polytechnique Institute and State University, which collaborated with the Nottingham team on another study to determine how well the OPM-MEG helmet worked when two individuals each wore one and then interacted.  To conduct this proof-of-concept study, two experiments involving social interaction were conducted.

“To really study how the human brain works, we have to embed people in their favorite natural environment—that’s a social setting,” said Read Montague, Ph.D., the principal investigator of the Virginia Tech team and director of the university’s Center for Human Neuroscience Research. The research was published this year in Sensors

For social interactions, two participants stroked each other’s hands and then played a game of ping pong.  Both experiments showed that despite large and unpredictable motions by participants, each person’s brain activity was clearly recorded.

The company is collecting data to obtain approval from regulatory bodies, including the FDA, to deploy the system in clinical populations, which can take up to five years.

Technical background: Designing a matrix coil system

The Nottingham research team constructed a system of electromagnetic coils to shield against the background magnetic-field noise and positioned them on two panels around the participant. Prior research published in Nature shows that eight large coils cancelled the background magnetic fields, but at a fixed position that only allowed small head movements.

Holmes and his team instead designed a new matrix coil system that features 48 smaller coils on two panels positioned around the participant. The coils can be individually controlled and continually recalibrate to compensate for the magnetic field changes experienced by the moving sensors, ensuring high-quality MEG data are recorded.

“This enables magnetic field compensation in any position, which makes OPM-MEG scans more comfortable for everyone and allows people to walk around,” said Holmes.

The researchers demonstrated the capabilities of the new matrix coil system with four experiments. They first wanted to show that the stationary helmet (not worn by anyone) placed inside the two coil panels could reduce background magnetic fields, which it did. Then a healthy participant wore the helmet, demonstrating that the OPMs recorded his brain function when he moved his head and that the coils cancelled the magnetic fields.

A third experiment used a wire coil as a proxy for brain cell activity because it produces magnetic fields when electric currents are applied. The wire coil attached to the helmet with OPM sensors showed that the matrix coil compensated for motion-related changes, ensuring accurate measurements. The last experiment showed that the helmet worn by a second healthy participant could produce a high-quality recording of brain activity when walking around.

“By taking advantage of recent OPM-MEG technology and designing a new magnetic shielding system, this helmet represents a novel magnetoencephalography approach that could help reveal more about how the brain works,” said Shumin Wang, Ph.D., a program director in the U.S. National Institute of Biomedical Imaging and Bioengineering (NIBIB) Division of Applied Science & Technology (Bioimaging).

Citations: Holmes, N, et al. Enabling ambulatory movement in wearable magnetoencephalography with matrix coil active magnetic shielding. NeuroImage. (2023), (open-access); and Holmes N, et al. Naturalistic Hyperscanning with Wearable Magnetoencephalography. Sensors. (2023), (open-access)

Let us know your thoughts! Sign up for a Mindplex account now, join our Telegram, or follow us on Twitter

Scientists build tiny healing biological robots from human cells

These multicellular bots move around and help heal “wounds” created in cultured neurons (and other possible uses)

Researchers at Tufts University and Harvard University’s Wyss Institute have created tiny biological robots that they call Anthrobots from human tracheal cells. They can move across a surface and encourage the growth of neurons across a region of damage in a lab dish.

The self-assembling, multicellular robots, which range in size from the width of a human hair to the point of a sharpened pencil, are shown to have a remarkable healing effect on other cells. The discovery is a starting point for the researchers’ vision to use patient-derived biobots as new therapeutic tools for regeneration, healing, and treatment of disease.

The work follows from earlier research in the laboratories of Michael Levin, Vannevar Bush Professor of Biology at Tufts University School of Arts & Sciences, and Josh Bongard at the University of Vermont in which they created multicellular biological robots (biobots) from frog embryo cells called Xenobots. These are capable of navigating passageways, collecting material, recording information, healing themselves from injury, and even replicating for a few cycles on their own.

In the current study, published in Advanced Science, Levin, along with PhD student Gizem Gumuskaya, discovered that such bots can be created from adult human cells without any genetic modification and demonstrate some capabilities beyond what was observed with the Xenobots.

Repairing neurons

The researchers used human cells in the trachea to reboot to create new multicellular shapes and move over a surface of human neurons grown in a lab dish to encourage new growth to fill in gaps caused by scratching the layer of cells. (We all experience the work of ciliated cells when we take the final step of expelling the particles and excess fluid by coughing or clearing our throats.)

Exactly how the Anthrobots encourage growth of neurons is not yet clear, but the researchers confirmed that neurons grew under the area covered by a clustered assembly of Anthrobots, which they called a “superbot.”

“The cellular assemblies we construct in the lab can have capabilities that go beyond what they do in the body,” said Levin, who also serves as the director of the Allen Discovery Center at Tufts and is an associate faculty member of the Wyss Institute.

“It is fascinating and completely unexpected that normal patient tracheal cells, without modifying their DNA, can move on their own and encourage neuron growth across a region of damage,” he said. “We’re now looking at how the healing mechanism works, and asking what else these constructs can do.”

The advantages of using human cells include the ability to construct bots from a patient’s own cells to perform therapeutic work without the risk of triggering an immune response or requiring immunosuppressants.

How to make Anthrobots

Each Anthrobot starts out as a single celld, derived from an adult donor. The cells come from the surface of the trachea and are covered with hairlike projections called cilia that wave back and forth. The cilia help the tracheal cells push out tiny particles that find their way into air passages of the lung.

Earlier studies by others had shown that when the cells are grown in the lab, they spontaneously form tiny multicellular spheres called organoids.

Potential other vital uses

Levin says that if other features could be added to the Anthrobots (for example, contributed by different cells), they could be designed to respond to their environment, travel to and perform functions in the body, or help build engineered tissues in the lab.

Other uses of the bots could include clearing plaque buildup in the arteries of atherosclerosis patients, repairing spinal cord or retinal nerve damage, recognizing bacteria or cancer cells, or delivering drugs to targeted tissues.

A swarm of Anthrobots. Credit: Gizem Gumuskaya, Tufts University

Citation: Gumuskaya, G., Srivastava, P., Cooper, B. G., Lesser, H., Semegran, B., Garnier, S., & Levin, M. Motile Living Biobots Self-Construct from Adult Human Somatic Progenitor Seed Cells. Advanced Science, 2303575. (open access)

Let us know your thoughts! Sign up for a Mindplex account now, join our Telegram, or follow us on Twitter