MIT Latest News

Subscribe to MIT Latest News feed
MIT News is dedicated to communicating to the media and the public the news and achievements of the students, faculty, staff and the greater MIT community.
Updated: 10 hours 15 min ago

Dopamine signals when a fear can be forgotten

Wed, 05/07/2025 - 9:50am

Dangers come but dangers also go, and when they do, the brain has an “all-clear” signal that teaches it to extinguish its fear. A new study in mice by MIT neuroscientists shows that the signal is the release of dopamine along a specific interregional brain circuit. The research therefore pinpoints a potentially critical mechanism of mental health, restoring calm when it works, but prolonging anxiety or even post-traumatic stress disorder when it doesn’t.

“Dopamine is essential to initiate fear extinction,” says Michele Pignatelli di Spinazzola, co-author of the new study from the lab of senior author Susumu Tonegawa, Picower Professor of biology and neuroscience at the RIKEN-MIT Laboratory for Neural Circuit Genetics within The Picower Institute for Learning and Memory at MIT, and a Howard Hughes Medical Institute (HHMI) investigator.

In 2020, Tonegawa’s lab showed that learning to be afraid, and then learning when that’s no longer necessary, result from a competition between populations of cells in the brain’s amygdala region. When a mouse learns that a place is “dangerous” (because it gets a little foot shock there), the fear memory is encoded by neurons in the anterior of the basolateral amygdala (aBLA) that express the gene Rspo2. When the mouse then learns that a place is no longer associated with danger (because they wait there and the zap doesn’t recur), neurons in the posterior basolateral amygdala (pBLA) that express the gene Ppp1r1b encode a new fear extinction memory that overcomes the original dread. Notably, those same neurons encode feelings of reward, helping to explain why it feels so good when we realize that an expected danger has dwindled.

In the new study, the lab, led by former members Xiangyu Zhang and Katelyn Flick, sought to determine what prompts these amygdala neurons to encode these memories. The rigorous set of experiments the team reports in the Proceedings of the National Academy of Sciences show that it’s dopamine sent to the different amygdala populations from distinct groups of neurons in the ventral tegmental area (VTA).

“Our study uncovers a precise mechanism by which dopamine helps the brain unlearn fear,” says Zhang, who also led the 2020 study and is now a senior associate at Orbimed, a health care investment firm. “We found that dopamine activates specific amygdala neurons tied to reward, which in turn drive fear extinction. We now see that unlearning fear isn’t just about suppressing it — it’s a positive learning process powered by the brain’s reward machinery. This opens up new avenues for understanding and potentially treating fear-related disorders, like PTSD.”

Forgetting fear

The VTA was the lab’s prime suspect to be the source of the signal because the region is well known for encoding surprising experiences and instructing the brain, with dopamine, to learn from them. The first set of experiments in the paper used multiple methods for tracing neural circuits to see whether and how cells in the VTA and the amygdala connect. They found a clear pattern: Rspo2 neurons were targeted by dopaminergic neurons in the anterior and left and right sides of the VTA. Ppp1r1b neurons received dopaminergic input from neurons in the center and posterior sections of the VTA. The density of connections was greater on the Ppp1r1b neurons than for the Rspo2 ones.

The circuit tracing showed that dopamine is available to amygdala neurons that encode fear and its extinction, but do those neurons care about dopamine? The team showed that indeed they express “D1” receptors for the neuromodulator. Commensurate with the degree of dopamine connectivity, Ppp1r1b cells had more receptors than Rspo2 neurons.

Dopamine does a lot of things, so the next question was whether its activity in the amygdala actually correlated with fear encoding and extinction. Using a method to track and visualize it in the brain, the team watched dopamine in the amygdala as mice underwent a three-day experiment. On Day One, they went to an enclosure where they experienced three mild shocks on the feet. On Day Two, they went back to the enclosure for 45 minutes, where they didn’t experience any new shocks — at first, the mice froze in anticipation of a shock, but then relaxed after about 15 minutes. On Day Three they returned again to test whether they had indeed extinguished the fear they showed at the beginning of Day Two.

The dopamine activity tracking revealed that during the shocks on Day One, Rspo2 neurons had the larger response to dopamine, but in the early moments of Day Two, when the anticipated shocks didn’t come and the mice eased up on freezing, the Ppp1r1b neurons showed the stronger dopamine activity. More strikingly, the mice that learned to extinguish their fear most strongly also showed the greatest dopamine signal at those neurons.

Causal connections

The final sets of experiments sought to show that dopamine is not just available and associated with fear encoding and extinction, but also actually causes them. In one set, they turned to optogenetics, a technology that enables scientists to activate or quiet neurons with different colors of light. Sure enough, when they quieted VTA dopaminergic inputs in the pBLA, doing so impaired fear extinction. When they activated those inputs, it accelerated fear extinction. The researchers were surprised that when they activated VTA dopaminergic inputs into the aBLA they could reinstate fear even without any new foot shocks, impairing fear extinction.

The other way they confirmed a causal role for dopamine in fear encoding and extinction was to manipulate the amygdala neurons’ dopamine receptors. In Ppp1r1b neurons, over-expressing dopamine receptors impaired fear recall and promoted extinction, whereas knocking the receptors down impaired fear extinction. Meanwhile in the Rspo2 cells, knocking down receptors reduced the freezing behavior.

“We showed that fear extinction requires VTA dopaminergic activity in the pBLA Ppp1r1b neurons by using optogenetic inhibition of VTA terminals and cell-type-specific knockdown of D1 receptors in these neurons,” the authors wrote.

The scientists are careful in the study to note that while they’ve identified the “teaching signal” for fear extinction learning, the broader phenomenon of fear extinction occurs brainwide, rather than in just this single circuit.

But the circuit seems to be a key node to consider as drug developers and psychiatrists work to combat anxiety and PTSD, Pignatelli di Spinazzola says.

“Fear learning and fear extinction provide a strong framework to study generalized anxiety and PTSD,” he says. “Our study investigates the underlying mechanisms suggesting multiple targets for a translational approach, such as pBLA and use of dopaminergic modulation.”

Marianna Rizzo is also a co-author of the study. Support for the research came from the RIKEN Center for Brain Science, the HHMI, the Freedom Together Foundation, and The Picower Institute.

Using AI to explore the 3D structure of the genome

Wed, 05/07/2025 - 12:00am

Inside every human cell, 2 meters of DNA is crammed into a nucleus that is only one-hundredth of a millimeter in diameter.

To fit inside that tiny space, the genome must fold into a complex structure known as chromatin, made up of DNA and proteins. The structure of that chromatin, in turn, helps to determine which of the genes will be expressed in a given cell. Neurons, skin cells, and immune cells each express different genes depending on which of their genes are accessible to be transcribed.

Deciphering those structures experimentally is a time-consuming process, making it difficult to compare the 3D genome structures found in different cell types. MIT Professor Bin Zhang is taking a computational approach to this challenge, using computer simulations and generative artificial intelligence to determine these structures.

“Regulation of gene expression relies on the 3D genome structure, so the hope is that if we can fully understand those structures, then we could understand where this cellular diversity comes from,” says Zhang, an associate professor of chemistry.

From the farm to the lab

Zhang first became interested in chemistry when his brother, who was four years older, bought some lab equipment and started performing experiments at home.

“He would bring test tubes and some reagents home and do the experiment there. I didn’t really know what he was doing back then, but I was really fascinated with all the bright colors and the smoke and the odors that could come from the reactions. That really captivated my attention,” Zhang says.

His brother later became the first person from Zhang’s rural village to go to college. That was the first time Zhang had an inkling that it might be possible to pursue a future other than following in the footsteps of his parents, who were farmers in China’s Anhui province.

“Growing up, I would have never imagined doing science or working as a faculty member in America,” Zhang says. “When my brother went to college, that really opened up my perspective, and I realized I didn’t have to follow my parents’ path and become a farmer. That led me to think that I could go to college and study more chemistry.”

Zhang attended the University of Science and Technology in Hefei, China, where he majored in chemical physics. He enjoyed his studies and discovered computational chemistry and computational research, which became his new fascination.

“Computational chemistry combines chemistry with other subjects I love — math and physics — and brings a sense of rigor and reasoning to the otherwise more empirical rules,” he says. “I could use programming to solve interesting chemistry problems and test my own ideas very quickly.”

After graduating from college, he decided to continue his studies in the United States, which he recalled thinking was “the pinnacle of academics.” At Caltech, he worked with Thomas Miller, a professor of chemistry who used computational methods to understand molecular processes such as protein folding.

For Zhang’s PhD research, he studied a transmembrane protein that acts as a channel to allow other proteins to pass through the cell membrane. This protein, called translocon, can also open a side gate within the membrane, so that proteins that are meant to be embedded in the membrane can exit directly into the membrane.

“It’s really a remarkable protein, but it wasn’t clear how it worked,” Zhang says. “I built a computational model to understand the molecular mechanisms that dictate what are the molecular features that allow certain proteins to go into the membrane, while other proteins get secreted.”

Turning to the genome

After finishing grad school, Zhang’s research focus shifted from proteins to the genome. At Rice University, he did a postdoc with Peter Wolynes, a professor of chemistry who had made many key discoveries in the dynamics of protein folding. Around the time that Zhang joined the lab, Wolynes turned his attention to the structure of the genome, and Zhang decided to do the same.

Unlike proteins, which tend to have highly structured regions that can be studied using X-ray crystallography or cryo-EM, DNA is a very globular molecule that doesn’t lend itself to those types of analysis.

A few years earlier, in 2009, researchers at the Broad Institute, the University of Massachusetts Medical School, MIT, and Harvard University had developed a technique for studying the genome’s structure by cross-linking DNA in a cell’s nucleus. Researchers can then determine which segments are located near each other by shredding the DNA into many tiny pieces and sequencing it.

Zhang and Wolynes used data generated by this technique, known as Hi-C, to explore the question of whether DNA forms knots when it’s condensed in the nucleus, similar to how a strand of Christmas lights may become tangled when crammed into a box for storage.

“If DNA was just like a regular polymer, you would expect that it will become tangled and form knots. But that could be very detrimental for biology, because the genome is not just sitting there passively. It has to go through cell division, and also all this molecular machinery has to interact with the genome and transcribe it into RNA, and having knots will create a lot of unnecessary barriers,” Zhang says.

They found that, unlike Christmas lights, DNA does not form any knots even when packed into the cell nucleus, and they built a computational model allowing them to test hypotheses for how the genome is able to avoid those entanglements.

Since joining the MIT faculty in 2016, Zhang has continued developing models of how the genome behaves in 3D space, using molecular dynamic simulations. In one area of research, his lab is studying how differences between the genome structures of neurons and other brain cells give rise to their unique functions, and they are also exploring how misfolding of the genome may lead to diseases such as Alzheimer’s.

When it comes to connecting genome structure and function, Zhang believes that generative AI methods will also be essential. In a recent study, he and his students reported a new computational model, ChromoGen, that uses generative AI to predict the 3D structures of genomic regions, based on their DNA sequences.

“I think that in the future, we will have both components: generative AI and also theoretical chemistry-based approaches,” he says. “They nicely complement each other and allow us to both build accurate 3D structures and understand how those structures arise from the underlying physical forces.” 

How can India decarbonize its coal-dependent electric power system?

Tue, 05/06/2025 - 5:00pm

As the world struggles to reduce climate-warming carbon emissions, India has pledged to do its part, and its success is critical: In 2023, India was the third-largest carbon emitter worldwide. The Indian government has committed to having net-zero carbon emissions by 2070.

To fulfill that promise, India will need to decarbonize its electric power system, and that will be a challenge: Fully 60 percent of India’s electricity comes from coal-burning power plants that are extremely inefficient. To make matters worse, the demand for electricity in India is projected to more than double in the coming decade due to population growth and increased use of air conditioning, electric cars, and so on.

Despite having set an ambitious target, the Indian government has not proposed a plan for getting there. Indeed, as in other countries, in India the government continues to permit new coal-fired power plants to be built, and aging plants to be renovated and their retirement postponed.

To help India define an effective — and realistic — plan for decarbonizing its power system, key questions must be addressed. For example, India is already rapidly developing carbon-free solar and wind power generators. What opportunities remain for further deployment of renewable generation? Are there ways to retrofit or repurpose India’s existing coal plants that can substantially and affordably reduce their greenhouse gas emissions? And do the responses to those questions differ by region?

With funding from IHI Corp. through the MIT Energy Initiative (MITEI), Yifu Ding, a postdoc at MITEI, and her colleagues set out to answer those questions by first using machine learning to determine the efficiency of each of India’s current 806 coal plants, and then investigating the impacts that different decarbonization approaches would have on the mix of power plants and the price of electricity in 2035 under increasingly stringent caps on emissions.

First step: Develop the needed dataset

An important challenge in developing a decarbonization plan for India has been the lack of a complete dataset describing the current power plants in India. While other studies have generated plans, they haven’t taken into account the wide variation in the coal-fired power plants in different regions of the country. “So, we first needed to create a dataset covering and characterizing all of the operating coal plants in India. Such a dataset was not available in the existing literature,” says Ding.

Making a cost-effective plan for expanding the capacity of a power system requires knowing the efficiencies of all the power plants operating in the system. For this study, the researchers used as their metric the “station heat rate,” a standard measurement of the overall fuel efficiency of a given power plant. The station heat rate of each plant is needed in order to calculate the fuel consumption and power output of that plant as plans for capacity expansion are being developed.

Some of the Indian coal plants’ efficiencies were recorded before 2022, so Ding and her team used machine-learning models to predict the efficiencies of all the Indian coal plants operating now. In 2024, they created and posted online the first comprehensive, open-sourced dataset for all 806 power plants in 30 regions of India. The work won the 2024 MIT Open Data Prize. This dataset includes each plant’s power capacity, efficiency, age, load factor (a measure indicating how much of the time it operates), water stress, and more.

In addition, they categorized each plant according to its boiler design. A “supercritical” plant operates at a relatively high temperature and pressure, which makes it thermodynamically efficient, so it produces a lot of electricity for each unit of heat in the fuel. A “subcritical” plant runs at a lower temperature and pressure, so it’s less thermodynamically efficient. Most of the Indian coal plants are still subcritical plants running at low efficiency.

Next step: Investigate decarbonization options

Equipped with their detailed dataset covering all the coal power plants in India, the researchers were ready to investigate options for responding to tightening limits on carbon emissions. For that analysis, they turned to GenX, a modeling platform that was developed at MITEI to help guide decision-makers as they make investments and other plans for the future of their power systems.

Ding built a GenX model based on India’s power system in 2020, including details about each power plant and transmission network across 30 regions of the country. She also entered the coal price, potential resources for wind and solar power installations, and other attributes of each region. Based on the parameters given, the GenX model would calculate the lowest-cost combination of equipment and operating conditions that can fulfill a defined future level of demand while also meeting specified policy constraints, including limits on carbon emissions. The model and all data sources were also released as open-source tools for all viewers to use.

Ding and her colleagues — Dharik Mallapragada, a former principal research scientist at MITEI who is now an assistant professor of chemical and biomolecular energy at NYU Tandon School of Engineering and a MITEI visiting scientist; and Robert J. Stoner, the founding director of the MIT Tata Center for Technology and Design and former deputy director of MITEI for science and technology — then used the model to explore options for meeting demands in 2035 under progressively tighter carbon emissions caps, taking into account region-to-region variations in the efficiencies of the coal plants, the price of coal, and other factors. They describe their methods and their findings in a paper published in the journal Energy for Sustainable Development.

In separate runs, they explored plans involving various combinations of current coal plants, possible new renewable plants, and more, to see their outcome in 2035. Specifically, they assumed the following four “grid-evolution scenarios:”

Baseline: The baseline scenario assumes limited onshore wind and solar photovoltaics development and excludes retrofitting options, representing a business-as-usual pathway.

High renewable capacity: This scenario calls for the development of onshore wind and solar power without any supply chain constraints.

Biomass co-firing: This scenario assumes the baseline limits on renewables, but here all coal plants — both subcritical and supercritical — can be retrofitted for “co-firing” with biomass, an approach in which clean-burning biomass replaces some of the coal fuel. Certain coal power plants in India already co-fire coal and biomass, so the technology is known.

Carbon capture and sequestration plus biomass co-firing: This scenario is based on the same assumptions as the biomass co-firing scenario with one addition: All of the high-efficiency supercritical plants are also retrofitted for carbon capture and sequestration (CCS), a technology that captures and removes carbon from a power plant’s exhaust stream and prepares it for permanent disposal. Thus far, CCS has not been used in India. This study specifies that 90 percent of all carbon in the power plant exhaust is captured.

Ding and her team investigated power system planning under each of those grid-evolution scenarios and four assumptions about carbon caps: no cap, which is the current situation; 1,000 million tons (Mt) of carbon dioxide (CO2) emissions, which reflects India’s announced targets for 2035; and two more-ambitious targets, namely 800 Mt and 500 Mt. For context, CO2 emissions from India’s power sector totaled about 1,100 Mt in 2021. (Note that transmission network expansion is allowed in all scenarios.)

Key findings

Assuming the adoption of carbon caps under the four scenarios generated a vast array of detailed numerical results. But taken together, the results show interesting trends in the cost-optimal mix of generating capacity and the cost of electricity under the different scenarios.

Even without any limits on carbon emissions, most new capacity additions will be wind and solar generators — the lowest-cost option for expanding India’s electricity-generation capacity. Indeed, this is observed to be the case now in India. However, the increasing demand for electricity will still require some new coal plants to be built. Model results show a 10 to 20 percent increase in coal plant capacity by 2035 relative to 2020.

Under the baseline scenario, renewables are expanded up to the maximum allowed under the assumptions, implying that more deployment would be economical. More coal capacity is built, and as the cap on emissions tightens, there is also investment in natural gas power plants, as well as batteries to help compensate for the now-large amount of intermittent solar and wind generation. When a 500 Mt cap on carbon is imposed, the cost of electricity generation is twice as high as it was with no cap.

The high renewable capacity scenario reduces the development of new coal capacity and produces the lowest electricity cost of the four scenarios. Under the most stringent cap — 500 Mt — onshore wind farms play an important role in bringing the cost down. “Otherwise, it’ll be very expensive to reach such stringent carbon constraints,” notes Ding. “Certain coal plants that remain run only a few hours per year, so are inefficient as well as financially unviable. But they still need to be there to support wind and solar.” She explains that other backup sources of electricity, such as batteries, are even more costly. 

The biomass co-firing scenario assumes the same capacity limit on renewables as in the baseline scenario, and the results are much the same, in part because the biomass replaces such a low fraction — just 20 percent — of the coal in the fuel feedstock. “This scenario would be most similar to the current situation in India,” says Ding. “It won’t bring down the cost of electricity, so we’re basically saying that adding this technology doesn’t contribute effectively to decarbonization.”

But CCS plus biomass co-firing is a different story. It also assumes the limits on renewables development, yet it is the second-best option in terms of reducing costs. Under the 500 Mt cap on CO2 emissions, retrofitting for both CCS and biomass co-firing produces a 22 percent reduction in the cost of electricity compared to the baseline scenario. In addition, as the carbon cap tightens, this option reduces the extent of deployment of natural gas plants and significantly improves overall coal plant utilization. That increased utilization “means that coal plants have switched from just meeting the peak demand to supplying part of the baseline load, which will lower the cost of coal generation,” explains Ding.

Some concerns

While those trends are enlightening, the analyses also uncovered some concerns for India to consider, in particular, with the two approaches that yielded the lowest electricity costs.

The high renewables scenario is, Ding notes, “very ideal.” It assumes that there will be little limiting the development of wind and solar capacity, so there won’t be any issues with supply chains, which is unrealistic. More importantly, the analyses showed that implementing the high renewables approach would create uneven investment in renewables across the 30 regions. Resources for onshore and offshore wind farms are mainly concentrated in a few regions in western and southern India. “So all the wind farms would be put in those regions, near where the rich cities are,” says Ding. “The poorer cities on the eastern side, where the coal power plants are, will have little renewable investment.”

So the approach that’s best in terms of cost is not best in terms of social welfare, because it tends to benefit the rich regions more than the poor ones. “It’s like [the government will] need to consider the trade-off between energy justice and cost,” says Ding. Enacting state-level renewable generation targets could encourage a more even distribution of renewable capacity installation. Also, as transmission expansion is planned, coordination among power system operators and renewable energy investors in different regions could help in achieving the best outcome.

CCS plus biomass co-firing — the second-best option for reducing prices — solves the equity problem posed by high renewables, and it assumes a more realistic level of renewable power adoption. However, CCS hasn’t been used in India, so there is no precedent in terms of costs. The researchers therefore based their cost estimates on the cost of CCS in China and then increased the required investment by 10 percent, the “first-of-a-kind” index developed by the U.S. Energy Information Administration. Based on those costs and other assumptions, the researchers conclude that coal plants with CCS could come into use by 2035 when the carbon cap for power generation is less than 1,000 Mt.

But will CCS actually be implemented in India? While there’s been discussion about using CCS in heavy industry, the Indian government has not announced any plans for implementing the technology in coal-fired power plants. Indeed, India is currently “very conservative about CCS,” says Ding. “Some researchers say CCS won’t happen because it’s so expensive, and as long as there’s no direct use for the captured carbon, the only thing you can do is put it in the ground.” She adds, "It’s really controversial to talk about whether CCS will be implemented in India in the next 10 years.”

Ding and her colleagues hope that other researchers and policymakers — especially those working in developing countries — may benefit from gaining access to their datasets and learning about their methods. Based on their findings for India, she stresses the importance of understanding the detailed geographical situation in a country in order to design plans and policies that are both realistic and equitable.

Philip Khoury to step down as vice provost for the arts

Tue, 05/06/2025 - 12:50pm

MIT Provost Cynthia Barnhart has announced that Vice Provost for the Arts Philip S. Khoury will step down from the position on Aug. 31. Khoury, the Ford International Professor of History, served in the role for 19 years. After a sabbatical, he will rejoin the faculty in the School of Humanities, Arts, and Social Sciences (SHASS).

“Since arriving at MIT in 1981, Philip has championed what he calls the Institute’s ‘artistic ecosystem,’ which sits at the intersection of technology, science, the humanities, and the arts. Thanks to Philip’s vision, this ecosystem is now a foundational element of MIT’s educational and research missions and a critical component of how we advance knowledge, understanding, and discovery in service to the world,” says Barnhart.

Khoury was appointed associate provost in 2006 by then-MIT president Susan Hockfield, with a double portfolio enhancing the Institute’s nonacademic arts programs and beginning a review of MIT’s international activities. Those programs include the List Visual Arts Center, the MIT Museum, the Center for Art, Science and Technology (CAST), and the Council for the Arts at MIT (CAMIT). After five years, the latter half of this portfolio evolved into the Office of the Vice Provost for International Activities. 

Khoury devoted most of his tenure to expanding the Institute’s arts infrastructure, promoting the visibility of its stellar arts faculty, and guiding the growth of student participation in the arts. Today, more than 50 percent of MIT undergraduates take arts classes, with more than 1,500 studying music.

“Philip has been a remarkable leader at MIT over decades. He has ensured that the arts are a prominent part of the MIT ‘mens-et-manus’ [‘mind-and-hand’] experience and that our community has the opportunity to admire, learn from, and participate in creative thinking in all realms,” says L. Rafael Reif, the Ray and Maria Stata Professor of Electrical Engineering and Computer Science and MIT president emeritus. “A historian — and a humanist at heart — Philip also played a crucial role in helping MIT develop a thoughtful international strategy in research and education."

“I will miss my colleagues first and foremost as I leave this position behind,” says Khoury. “But I have been proud to see the quality of the faculty grow and the student interest in the arts grow almost exponentially, along with an awareness of how the arts are prospering at MIT.”

Stream of creativity

During his time as vice provost, he partnered with then-School of Architecture and Planning (SAP) dean Adèle Santos and SHASS dean Deborah Fitzgerald to establish the CAST in 2012. The center encourages artistic collaborations and provides seed funds and research grants to students and faculty. 

Khoury also helped oversee a significant expansion of the Institute’s art facilities, including the unique multipurpose design of the Theater Arts Building, the new MIT Museum, and the Edward and Joyce Linde Music Building. Along with the List Visual Arts Center, which will celebrate its 40th anniversary this year, these vibrant spaces “offer an opportunity for our students to do something different from what they came to MIT to do in science and engineering,” Khoury suggests. “It gives them an outlet to do other kinds of experimentation.”

“What makes the arts so successful here is that they are very much in the stream of creativity, which science and technology are all about,” he adds.

One of Khoury’s other long-standing goals has been to elevate the recognition of the arts faculty, “to show that the quality of what we do in those areas matches the quality of what we do in engineering and science,” he says.

“I will always remember Philip Khoury’s leadership and advocacy as dean of the School of Humanities and Social Sciences for changing the definition of the ‘A’ in SHASS from ‘and’ to ‘Arts.’ That small change had large implications for professional careers for artists, enrollments, and subject options that remain a source of renewal and strength to this day,” says Institute Professor Marcus Thompson.

Most recently, Khoury and his team, in collaboration with faculty, students, and staff from across the Institute, oversaw the development and production of MIT’s new festival of the arts, known as Artfinity. Launched in February and open to the public, the Institute-sponsored, campus-wide festival featured a series of 80 performing and visual arts events.

International activities

Khoury joined the faculty as an assistant professor in 1981 and later served as dean of SHASS between 1991 and 2006. In 2002, he was appointed the inaugural Kenan Sahin Dean of SHASS.

His academic focus made him a natural choice for the first coordinator of MIT international activities, a role he served in from 2006 to 2011. During that time, he traveled widely to learn more about the ways MIT faculty were engaged abroad, and he led the production of an influential report on the state of MIT’s international activities.

“We wanted to create a strategy, but not a foreign policy,” Khoury said of the report.

Khoury’s time in the international role led him to consider ways that collaborations with other countries should be balanced so as not to diminish MIT’s offerings at home, he says. He also looked for ways to encourage more collaborations with countries in sub-Saharan Africa, South America, and parts of the Middle East.

Future plans

Khoury was instrumental in establishing the Future of the Arts at MIT Committee, which was charged by Provost Barnhart in June 2024 in collaboration with Dean Hashim Sarkis of the School of Architecture and Planning and Dean Agustín Rayo of SHASS. The committee aims to find new ways to envision the place of arts at the Institute — a task that was last undertaken in 1987, he says. The committee submitted a draft report to Provost Barnhart in April. 

“I think it will hit the real sweet spot of where arts meet science and technology, but not where art is controlled by science and technology,” Khoury says. “I think the promotion of that, and the emphasis on that, among other connections with art, are really what we should be pushing for and developing.”

After he steps down as vice provost, Khoury plans to devote more time to writing two books: a personal memoir and a book about the Middle East. And he is looking forward to seeing how the arts at MIT will flourish in the near future. “I feel elated about where we’ve landed and where we’ll continue to go,” he says.

As Barnhart noted in her letter to the community, the Future of the Arts at MIT Committee's efforts combined with Khoury staying on through the end of the summer, provides President Kornbluth, the incoming provost, and Khoury with the opportunity to reflect on the Institute’s path forward in this critical space.

Hybrid AI model crafts smooth, high-quality videos in seconds

Tue, 05/06/2025 - 12:15pm

What would a behind-the-scenes look at a video generated by an artificial intelligence model be like? You might think the process is similar to stop-motion animation, where many images are created and stitched together, but that’s not quite the case for “diffusion models” like OpenAl's SORA and Google's VEO 2.

Instead of producing a video frame-by-frame (or “autoregressively”), these systems process the entire sequence at once. The resulting clip is often photorealistic, but the process is slow and doesn’t allow for on-the-fly changes. 

Scientists from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) and Adobe Research have now developed a hybrid approach, called “CausVid,” to create videos in seconds. Much like a quick-witted student learning from a well-versed teacher, a full-sequence diffusion model trains an autoregressive system to swiftly predict the next frame while ensuring high quality and consistency. CausVid’s student model can then generate clips from a simple text prompt, turning a photo into a moving scene, extending a video, or altering its creations with new inputs mid-generation.

This dynamic tool enables fast, interactive content creation, cutting a 50-step process into just a few actions. It can craft many imaginative and artistic scenes, such as a paper airplane morphing into a swan, woolly mammoths venturing through snow, or a child jumping in a puddle. Users can also make an initial prompt, like “generate a man crossing the street,” and then make follow-up inputs to add new elements to the scene, like “he writes in his notebook when he gets to the opposite sidewalk.”

The CSAIL researchers say that the model could be used for different video editing tasks, like helping viewers understand a livestream in a different language by generating a video that syncs with an audio translation. It could also help render new content in a video game or quickly produce training simulations to teach robots new tasks.

Tianwei Yin SM ’25, PhD ’25, a recently graduated student in electrical engineering and computer science and CSAIL affiliate, attributes the model’s strength to its mixed approach.

“CausVid combines a pre-trained diffusion-based model with autoregressive architecture that’s typically found in text generation models,” says Yin, co-lead author of a new paper about the tool. “This AI-powered teacher model can envision future steps to train a frame-by-frame system to avoid making rendering errors.”

Yin’s co-lead author, Qiang Zhang, is a research scientist at xAI and a former CSAIL visiting researcher. They worked on the project with Adobe Research scientists Richard Zhang, Eli Shechtman, and Xun Huang, and two CSAIL principal investigators: MIT professors Bill Freeman and Frédo Durand.

Caus(Vid) and effect

Many autoregressive models can create a video that’s initially smooth, but the quality tends to drop off later in the sequence. A clip of a person running might seem lifelike at first, but their legs begin to flail in unnatural directions, indicating frame-to-frame inconsistencies (also called “error accumulation”).

Error-prone video generation was common in prior causal approaches, which learned to predict frames one by one on their own. CausVid instead uses a high-powered diffusion model to teach a simpler system its general video expertise, enabling it to create smooth visuals, but much faster.

CausVid displayed its video-making aptitude when researchers tested its ability to make high-resolution, 10-second-long videos. It outperformed baselines like “OpenSORA” and “MovieGen,” working up to 100 times faster than its competition while producing the most stable, high-quality clips.

Then, Yin and his colleagues tested CausVid’s ability to put out stable 30-second videos, where it also topped comparable models on quality and consistency. These results indicate that CausVid may eventually produce stable, hours-long videos, or even an indefinite duration.

A subsequent study revealed that users preferred the videos generated by CausVid’s student model over its diffusion-based teacher.

“The speed of the autoregressive model really makes a difference,” says Yin. “Its videos look just as good as the teacher’s ones, but with less time to produce, the trade-off is that its visuals are less diverse.”

CausVid also excelled when tested on over 900 prompts using a text-to-video dataset, receiving the top overall score of 84.27. It boasted the best metrics in categories like imaging quality and realistic human actions, eclipsing state-of-the-art video generation models like “Vchitect” and “Gen-3.

While an efficient step forward in AI video generation, CausVid may soon be able to design visuals even faster — perhaps instantly — with a smaller causal architecture. Yin says that if the model is trained on domain-specific datasets, it will likely create higher-quality clips for robotics and gaming.

Experts say that this hybrid system is a promising upgrade from diffusion models, which are currently bogged down by processing speeds. “[Diffusion models] are way slower than LLMs [large language models] or generative image models,” says Carnegie Mellon University Assistant Professor Jun-Yan Zhu, who was not involved in the paper. “This new work changes that, making video generation much more efficient. That means better streaming speed, more interactive applications, and lower carbon footprints.”

The team’s work was supported, in part, by the Amazon Science Hub, the Gwangju Institute of Science and Technology, Adobe, Google, the U.S. Air Force Research Laboratory, and the U.S. Air Force Artificial Intelligence Accelerator. CausVid will be presented at the Conference on Computer Vision and Pattern Recognition in June.

How J-WAFS Solutions grants bring research to market

Tue, 05/06/2025 - 11:55am

For the Abdul Latif Jameel Water and Food Systems Lab (J-WAFS), 2025 marks a decade of translating groundbreaking research into tangible solutions for global challenges. Few examples illustrate that mission better than NONA Technologies. With support from a J-WAFS Solutions grant, MIT electrical engineering and biological engineering Professor Jongyoon Han and his team developed a portable desalination device that transforms seawater into clean drinking water without filters or high-pressure pumps. 

The device stands apart from traditional systems because conventional desalination technologies, like reverse osmosis, are energy-intensive, prone to fouling, and typically deployed at large, centralized plants. In contrast, the device developed in Han’s lab employs ion concentration polarization technology to remove salts and particles from seawater, producing potable water that exceeds World Health Organization standards. It is compact, solar-powered, and operable at the push of a button — making it an ideal solution for off-grid and disaster-stricken areas.

This research laid the foundation for spinning out NONA Technologies along with co-founders Junghyo Yoon PhD ’21 from Han’s lab and Bruce Crawford MBA ’22, to commercialize the technology and address pressing water-scarcity issues worldwide. “This is really the culmination of a 10-year journey that I and my group have been on,” said Han in an earlier MIT News article. “We worked for years on the physics behind individual desalination processes, but pushing all those advances into a box, building a system, and demonstrating it in the ocean ... that was a really meaningful and rewarding experience for me.” You can watch this video showcasing the device in action.

Moving breakthrough research out of the lab and into the world is a well-known challenge. While traditional “seed” grants typically support early-stage research at Technology Readiness Level (TRL) 1-2, few funding sources exist to help academic teams navigate to the next phase of technology development. The J-WAFS Solutions Program is strategically designed to address this critical gap by supporting technologies in the high-risk, early-commercialization phase that is often neglected by traditional research, corporate, and venture funding. By supporting technologies at TRLs 3-5, the program increases the likelihood that promising innovations will survive beyond the university setting, advancing sufficiently to attract follow-on funding.

Equally important, the program gives academic researchers the time, resources, and flexibility to de-risk their technology, explore customer need and potential real-world applications, and determine whether and how they want to pursue commercialization. For faculty-led teams like Han’s, the J-WAFS Solutions Program provided the critical financial runway and entrepreneurial guidance needed to refine the technology, test assumptions about market fit, and lay the foundation for a startup team. While still in the MIT innovation ecosystem, Nona secured over $200,000 in non-dilutive funding through competitions and accelerators, including the prestigious MIT delta v Educational Accelerator. These early wins laid the groundwork for further investment and technical advancement.

Since spinning out of MIT, NONA has made major strides in both technology development and business viability. What started as a device capable of producing just over half-a-liter of clean drinking water per hour has evolved into a system that now delivers 10 times that capacity, at 5 liters per hour. The company successfully raised a $3.5 million seed round to advance its portable desalination device, and entered into a collaboration with the U.S. Army Natick Soldier Systems Center, where it co-developed early prototypes and began generating revenue while validating the technology. Most recently, NONA was awarded two SBIR Phase I grants totaling $575,000, one from the National Science Foundation and another from the National Institute of Environmental Health Sciences.

Now operating out of Greentown Labs in Somerville, Massachusetts, NONA has grown to a dedicated team of five and is preparing to launch its nona5 product later this year, with a wait list of over 1,000 customers. It is also kicking off its first industrial pilot, marking a key step toward commercial scale-up. “Starting a business as a postdoc was challenging, especially with limited funding and industry knowledge,” says Yoon, who currently serves as CTO of NONA. “J-WAFS gave me the financial freedom to pursue my venture, and the mentorship pushed me to hit key milestones. Thanks to J-WAFS, I successfully transitioned from an academic researcher to an entrepreneur in the water industry.”

NONA is one of several J-WAFS-funded technologies that have moved from the lab to market, part of a growing portfolio of water and food solutions advancing through MIT’s innovation pipeline. As J-WAFS marks a decade of catalyzing innovation in water and food, NONA exemplifies what is possible when mission-driven research is paired with targeted early-stage support and mentorship.

To learn more or get involved in supporting startups through the J-WAFS Solutions Program, please contact jwafs@mit.edu.

If time is money, here’s one way consumers value it

Tue, 05/06/2025 - 12:00am

As the saying goes, time is money. That’s certainly evident in the transportation sector, where people will pay more for direct flights, express trains, and other ways to get somewhere quickly.

Still, it is difficult to measure precisely how much people value their time. Now, a paper co-authored by an MIT economist uses ride-sharing data to reveal multiple implications of personalized pricing.

By focusing on a European ride-sharing platform that auctions its rides, the researchers found that people are more responsive to prices than to wait times. They also found that people pay more to save time during the workday, and that when people pay more to avoid waiting, it notably increases business revenues. And some segments of consumers are distinctly more willing than others to pay higher prices.

Specifically, when people can bid for rides that arrive sooner, the amount above the minimum price the platform can charge increases by 5.2 percent. Meanwhile, the gap between offered prices and the maximum that consumers are willing to pay decreases by 2.5 percent. In economics terms, this creates additional “surplus” value for firms, while lowering the “consumer surplus” in these transactions.

“One of the important quantities in transportation is the value of time,” says MIT economist Tobias Salz, co-author of a new paper detailing the study’s findings. “We came across a setting that offered a very clean way of examining this quantity, where the value of time is revealed by people’s transportation choices.”

The paper, “Personalized Pricing and the Value of Time: Evidence from Auctioned Cab Rides,” is being published in Econometrica. The authors are Nicholas Buchholz, an assistant professor of economics at Princeton University; Laura Doval, a professor at Columbia Business School; Jakub Kastl, a professor of economics at Princeton University; Filip Matejka, a professor at Charles University in Prague; and Salz, the Castle Krob Career Development Associate Professor of Economics in MIT’s Department of Economics.

It is not easy to study how much money people will spend to save time — and time alone. Transportation is one sector where it is possible to do so, though not the only one. People will also pay more for, say, an express pass to avoid long lines at an amusement park. But data for those scenarios, even when available, may contain complicating factors. (Also, the value of time shouldn’t be confused with how much people pay for services charged by the hour, from accountants to tuba lessons.)

In this case, however, the researchers were provided data from Liftago, a ride-sharing platform in Prague with a distinctive feature: It lets drivers bid on a passenger’s business, with the wait time until the auto arrives as one of the factors involved. Drivers can also indicate when they will be available. In studying how passengers compare offers with different wait times and prices, the researchers see exactly how much people are paying not to wait, other things being equal. All told, they examined 1.9 million ride requests and 5.2 million bids.

“It’s like an eBay for taxis,” Salz says. “Instead of assigning the driver to you, drivers bid for the passengers’ business. With this, we can very directly observe how people make their choices. How they value time is revealed by the wait and the prices attached to that. In many settings we don’t observe that directly, so it’s a very clean comparison that rids the data of a lot of confounds.”

The data set allows the researchers to examine many aspects of personalized pricing and the way it affects the transportation market in this setting. That produces a set of insights on its own, along with the findings on time valuation. 

Ultimately, the researchers found that the elasticity of prices — how much they change — ranged from four to 10 times as much as the elasticity of wait times, meaning people are more keen on avoiding high prices.

The team found the overall value of time in this context is $13.21 per hour for users of the ride-share platform, though the researchers note that is not a universal measure of the value of time and is dependent on this setting. The study also shows that bids increase during work hours.

Additionally, the research reveals a split among consumers: The top quartile of bids placed a value on time that is 3.5 times higher than the value of the bids in the bottom quartile.

Then there is still the question of how much personalized pricing benefits consumers, providers, or both. The numbers, again, show that the overall surplus increases — meaning business benefits — while the consumer surplus is reduced. However, the data show an even more nuanced picture. Because the top quartile of bidders are paying substantially more to avoid longer waits, they are the ones who absorb the brunt of the costs in this kind of system.

“The majority of consumers still benefit,” Salz says. “The consumers hurt by this have a very high willingness to pay. The source of welfare gains is that most consumers can be brought into the market. But the flip side is that the firm, by knowing every consumer’s choke point, can extract the surplus. Welfare goes up, the ride-sharing platform captures most of that, and drivers — interestingly — also benefit from the system, although they do not have access to the data.”

Economic theory and other transportation studies alone would not necessarily have predicted the study’s results and various nuances.

“It was not clear a priori whether consumers benefit,” Salz observes. “That is not something you would know without going to the data.”

While this study might hold particular interest for firms and others interested in transportation, mobility, and ride-sharing, it also fits into a larger body of economics research about information in markets and how its presence, or absence, influences consumer behavior, consumer welfare, and the functioning of markets.

“The [research] umbrella here is really information about where to find trading partners and what their willingness to pay is,” Salz says. “What I’m broadly interested in is these types of information frictions and how they determine market outcomes, how they might impact consumers, and be used by firms.”

The research was supported, in part, by the National Bureau of Economic Research, the U.S. Department of Transportation, and the National Science Foundation. 

Startup helps farmers grow plant-based feed and fertilizer using wastewater

Tue, 05/06/2025 - 12:00am

Farmers today face a number of challenges, from supply chain stability to nutrient and waste management. But hanging over everything is the need to maintain profitability amid changing markets and increased uncertainty.

Fyto, founded by former MIT staff member Jason Prapas, is offering a highly automated cultivation system to address several of farmers’ biggest problems at once.

At the heart of Fyto’s system is Lemna, a genus of small aquatic plants otherwise known as duckweed. Most people have probably seen thick green mats of Lemna lying on top of ponds and swamps. But Lemna is also rich in protein and capable of doubling in biomass every two days. Fyto has built an automated cropping system that uses nitrogen-rich wastewater from dairy farms to grow Lemna in shallow pools on otherwise less productive farmland. On top of the pools, the company has built what it believes are the largest agricultural robots in the world, which monitor plant health and harvest the Lemna sustainably. The Lemna can then be used on farms as a high-protein cattle feed or fertilizer supplement.

Fyto’s systems are designed to rely on minimal land, water, and labor while creating a more sustainable, profitable food system.

“We developed from scratch a robotic system that takes the guesswork out of farming this crop,” says Prapas, who previously led the translational research program of MIT’s Tata Center. “It looks at the crop on a daily basis, takes inventory to know how many plants there are, how much should be harvested to have healthy growth the next day, can detect if the color is slightly off or there are nutrient deficiencies, and can suggest different interventions based on all that data.”

From kiddie pools to cow farms

Prapas’ first job out of college was with an MIT spinout called Green Fuel that harvested algae to make biofuel. He went back to school for a master’s and then a PhD in mechanical engineering, but he continued working with startups. Following his PhD at Colorado State University, he co-founded Factor[e] Ventures to fund and incubate startups focused on improving energy access in emerging markets.

Through that work, Prapas was introduced to MIT’s Tata Center for Technology and Design.

“We were really interested in the new technologies being developed at the MIT Tata Center, and in funding new startups taking on some of these global climate challenges in emerging markets,” Prapas recalls. “The Tata Center was interested in making sure these technologies get put into practice rather than patented and put on a shelf somewhere. It was a good synergy.”

One of the people Prapas got to know was Rob Stoner, the founding director of the Tata Center, who encouraged Prapas to get more directly involved with commercializing new technologies. In 2017, Prapas joined the Tata Center as the translational research director. During that time, Prapas worked with MIT students, faculty, and staff to test their inventions in the real world. Much of that work involved innovations in agriculture.

“Farming is a fact of life for a lot of folks around the world — both subsistence farming but also producing food for the community and beyond,” Prapas says. “That has huge implications for water usage, electricity consumption, labor. For years, I’d been thinking about how we make farming a more attractive endeavor for people: How do we make it less back-breaking, more efficient, and more economical?”

Between his work at MIT and Factor[e], Prapas visited hundreds of farms around the world, where he started to think about the lack of good choices for farming inputs like animal feed and fertilizers. The problem represented a business opportunity.

Fyto began with kiddie pools. Prapas started growing aquatic plants in his backyard, using them as a fertilizer source for vegetables. The experience taught him how difficult it would be to train people to grow and harvest Lemna at large scales on farms.

“I realized we’d have to invent both the farming method — the agronomy — and the equipment and processes to grow it at scale cost effectively,” Prapas explains.

Prapas started discussing his ideas with others around 2019.

“The MIT and Boston ecosystems are great for pitching somewhat crazy ideas to willing audiences and seeing what sticks,” Prapas says. “There’s an intangible benefit of being at MIT, where you just can’t help but think of bold ideas and try putting them into practice.”

Prapas, who left MIT to lead Fyto in 2019, partnered with Valerie Peng ’17, SM ’19, then a graduate student at MIT who became his first hire.

“Farmers work so hard, and I have so much respect for what they do,” says Peng, who serves as Fyto’s head of engineering. “People talk about the political divide, but there’s a lot of alignment around using less, doing more with what you have, and making our food systems more resilient to drought, supply chain disruptions, and everything else. There’s more in common with everyone than you’d expect.”

A new farming method

Lemna can produce much more protein per acre than soy, another common source of protein on farms, but it requires a lot of nitrogen to grow. Fortunately, many types of farmers, especially large dairy farmers, have abundant nitrogen sources in the waste streams that come from washing out cow manure.

“These waste streams are a big problem: In California it’s believed to be one of the largest source of greenhouse gas emissions in the agriculture sector despite the fact that hundreds of crops are grown in California,” Prapas says.

For the last few years, Fyto has run its systems in pilots on farms, trialing the crop as feed and fertilizer before delivering to its customers. The systems Fyto has deployed so far are about 50 feet wide, but it is actively commissioning its newest version that’s 160 feet wide. Eventually, Fyto plans to sell the systems directly to farmers.

Fyto is currently awaiting California’s approval for use in feed, but Lemna has already been approved in Europe. Fyto has also been granted a fertilizer license on its plant-based fertilizer, with promising early results in trials, and plans to sell new fertilizer products this year.

Although Fyto is focused on dairy farms for its early deployments, it has also grown Lemna using manure from chicken, and Prapas notes that even people like cheese producers have a nitrogen waste problem that Fyto could solve.

“Think of us like a polishing step you could put on the end of any system that has an organic waste stream,” Prapas says. “In that situation, we’re interested in growing our crops on it. We’ve had very few things that the plant can’t grow on. Globally, we see this as a new farming method, and that means it’s got a lot of potential applications.”

Q&A: A roadmap for revolutionizing health care through data-driven innovation

Mon, 05/05/2025 - 4:15pm

What if data could help predict a patient’s prognosis, streamline hospital operations, or optimize human resources in medicine? A book fresh off the shelves, “The Analytics Edge in Healthcare,” shows that this is already happening, and demonstrates how to scale it. 

Authored by Dimitris Bertsimas, MIT’s vice provost for open learning, along with two of Bertsimas’ former students — Agni Orfanoudaki PhD ’21, associate professor of operations management at University of Oxford’s Saïd Business School, and Holly Wiberg PhD ’22, assistant professor of public policy and operations research at Carnegie Mellon University — the book provides a practical introduction to the field of health care analytics. With an emphasis on real-world applications, the first part of the book establishes technical foundations — spanning machine learning and optimization — while the second part of the book presents integrated case studies that cover various clinical specialties and problem types using descriptive, predictive, and prescriptive analytics. 

Part of a broader series, “The Analytics Edge in Healthcare” demonstrates how to leverage data and models to make better decisions within the health care sector, while its predecessor, “The Analytics Edge,” dives into the science of using data to build models, improve decisions, and add value to institutions and individuals. 

Bertsimas, who is also the associate dean of business analytics and the Boeing Leaders for Global Operations Professor of Management at the MIT Sloan School of Management, is the innovator behind 15.071 (The Analytics Edge), a course on MIT Open Learning’s MITx that has attracted hundreds of thousands of online learners and served as the inspiration behind the book series. Bertsimas took a break from research and his work at MIT Open Learning to discuss how the field of analytics is transforming the health care system and share some surprising ways analytics are already being used in hospitals. 

Q: How is the field of analytics changing the way hospitals provide care and manage their operations?

A: As an academic, I’ve always aspired to educate, write publications, and utilize what we do in practice. Therefore, I founded Holistic Hospital Optimization (H20) with the goal of optimizing hospital operations with machine learning to improve patient care. We have developed a variety of tools at MIT and implemented them at hospitals around the world. For example, we manage patients’ length of stay and their deterioration indexes (a computerized tool that predicts a patient’s risk of clinical deterioration); we manage nurse optimization and how hospitals can allocate human resources appropriately; and we optimize blocks for surgeries. This is the beginning of a change where analytics and AI methods are now being utilized quite widely. My hope would be that this work and this book will accelerate the effect of using these tools. 

Additionally, I have taught a nine-lecture course twice with Agni and Holly at the Hartford Hospital System, where I realized that these analytics methods — which are typically not taught in medical schools — can be demonstrated for health care practitioners, including physicians, nurses, and administrators. To have an impact, you need to have appropriate methods, implement them, and apply them, but you also need to educate people on how to use them. This links well with my role at Open Learning, where our objective is to educate learners globally. In fact, Open Learning is launching this fall Universal AI, a dynamic online learning experience that provides comprehensive knowledge on artificial intelligence, preparing a global audience of learners for employment in our rapidly evolving job market. 

Q: What are some surprising ways analytics are being used in health care that most people wouldn’t expect?

A: Using analytics, we have reduced patients’ length of stay at Hartford Hospital from 5.67 days to five days. We have an algorithm that predicts patients’ probability of being released; therefore, doctors prioritize the patients with the highest probability, preparing them for discharge. This means that the hospital can treat far more patients, and the patients stay in the hospital less time.

Furthermore, when hospitals saw an increase in nurse turnover during the Covid-19 pandemic, we developed an analytics system that takes into account equity and fairness and decreases overtime costs, giving preferred slots to nurses and decreasing overall turnover substantially. These are just two examples; there are many others where an analytical perspective to health care and medicine has made a material difference. 

Q: Looking ahead, how do you see artificial intelligence shaping the future of health care?

A: In a very significant way — we use machine learning to make better predictions, but generative AI can explain them. I already see a movement in that direction. It’s really the evolution of AI that made this possible, and it is exciting. It’s also important for the world, because of its capabilities to improve care and save lives. 

For example, through our program at the Hartford Hospital System, we discovered that a patient was getting worse and predicted through analytics that they would get even worse. After our prediction, the doctors examined the patient more closely and discovered the patient had an early case of sepsis, a life-threatening condition in which the body responds improperly to an infection. If we hadn’t detected sepsis earlier, the patient might have died. This made an actual difference in saving a person’s life. 

Q: If you had to describe “The Analytics Edge in Healthcare” in one or two words, what would they be, and why? 

A: The book is a phased transition in health care because it is capable of affecting the health care sector in a way that has not been done before. The book really outlines my work in health care and its applications in the last decade.

New tool evaluates progress in reinforcement learning

Mon, 05/05/2025 - 4:00pm

If there’s one thing that characterizes driving in any major city, it’s the constant stop-and-go as traffic lights change and as cars and trucks merge and separate and turn and park. This constant stopping and starting is extremely inefficient, driving up the amount of pollution, including greenhouse gases, that gets emitted per mile of driving. 

One approach to counter this is known as eco-driving, which can be installed as a control system in autonomous vehicles to improve their efficiency.

How much of a difference could that make? Would the impact of such systems in reducing emissions be worth the investment in the technology? Addressing such questions is one of a broad category of optimization problems that have been difficult for researchers to address, and it has been difficult to test the solutions they come up with. These are problems that involve many different agents, such as the many different kinds of vehicles in a city, and different factors that influence their emissions, including speed, weather, road conditions, and traffic light timing.

“We got interested a few years ago in the question: Is there something that automated vehicles could do here in terms of mitigating emissions?” says Cathy Wu, the Thomas D. and Virginia W. Cabot Career Development Associate Professor in the Department of Civil and Environmental Engineering and the Institute for Data, Systems, and Society (IDSS) at MIT, and a principal investigator in the Laboratory for Information and Decision Systems. “Is it a drop in the bucket, or is it something to think about?,” she wondered.

To address such a question involving so many components, the first requirement is to gather all available data about the system, from many sources. One is the layout of the network’s topology, Wu says, in this case a map of all the intersections in each city. Then there are U.S. Geological Survey data showing the elevations, to determine the grade of the roads. There are also data on temperature and humidity, data on the mix of vehicle types and ages, and on the mix of fuel types.

Eco-driving involves making small adjustments to minimize unnecessary fuel consumption. For example, as cars approach a traffic light that has turned red, “there’s no point in me driving as fast as possible to the red light,” she says. By just coasting, “I am not burning gas or electricity in the meantime.” If one car, such as an automated vehicle, slows down at the approach to an intersection, then the conventional, non-automated cars behind it will also be forced to slow down, so the impact of such efficient driving can extend far beyond just the car that is doing it.

That’s the basic idea behind eco-driving, Wu says. But to figure out the impact of such measures, “these are challenging optimization problems” involving many different factors and parameters, “so there is a wave of interest right now in how to solve hard control problems using AI.” 

The new benchmark system that Wu and her collaborators developed based on urban eco-driving, which they call “IntersectionZoo,” is intended to help address part of that need. The benchmark was described in detail in a paper presented at the 2025 International Conference on Learning Representation in Singapore.

Looking at approaches that have been used to address such complex problems, Wu says an important category of methods is multi-agent deep reinforcement learning (DRL), but a lack of adequate standard benchmarks to evaluate the results of such methods has hampered progress in the field.

The new benchmark is intended to address an important issue that Wu and her team identified two years ago, which is that with most existing deep reinforcement learning algorithms, when trained for one specific situation (e.g., one particular intersection), the result does not remain relevant when even small modifications are made, such as adding a bike lane or changing the timing of a traffic light, even when they are allowed to train for the modified scenario.

In fact, Wu points out, this problem of non-generalizability “is not unique to traffic,” she says. “It goes back down all the way to canonical tasks that the community uses to evaluate progress in algorithm design.” But because most such canonical tasks do not involve making modifications, “it’s hard to know if your algorithm is making progress on this kind of robustness issue, if we don’t evaluate for that.”

While there are many benchmarks that are currently used to evaluate algorithmic progress in DRL, she says, “this eco-driving problem features a rich set of characteristics that are important in solving real-world problems, especially from the generalizability point of view, and that no other benchmark satisfies.” This is why the 1 million data-driven traffic scenarios in IntersectionZoo uniquely position it to advance the progress in DRL generalizability.  As a result, “this benchmark adds to the richness of ways to evaluate deep RL algorithms and progress.”

And as for the initial question about city traffic, one focus of ongoing work will be applying this newly developed benchmarking tool to address the particular case of how much impact on emissions would come from implementing eco-driving in automated vehicles in a city, depending on what percentage of such vehicles are actually deployed.

But Wu adds that “rather than making something that can deploy eco-driving at a city scale, the main goal of this study is to support the development of general-purpose deep reinforcement learning algorithms, that can be applied to this application, but also to all these other applications — autonomous driving, video games, security problems, robotics problems, warehousing, classical control problems.”

Wu adds that “the project’s goal is to provide this as a tool for researchers, that’s openly available.” IntersectionZoo, and the documentation on how to use it, are freely available at GitHub.

Wu is joined on the paper by lead authors Vindula Jayawardana, a graduate student in MIT’s Department of Electrical Engineering and Computer Science (EECS); Baptiste Freydt, a graduate student from ETH Zurich; and co-authors Ao Qu, a graduate student in transportation; Cameron Hickert, an IDSS graduate student; and Zhongxia Yan PhD ’24. 

New molecular label could lead to simpler, faster tuberculosis tests

Mon, 05/05/2025 - 3:00pm

Tuberculosis, the world’s deadliest infectious disease, is estimated to infect around 10 million people each year, and kills more than 1 million annually. Once established in the lungs, the bacteria’s thick cell wall helps it to fight off the host immune system.

Much of that cell wall is made from complex sugar molecules known as glycans, but it’s not well-understood how those glycans help to defend the bacteria. One reason for that is that there hasn’t been an easy way to label them inside cells.

MIT chemists have now overcome that obstacle, demonstrating that they can label a glycan called ManLAM using an organic molecule that reacts with specific sulfur-containing sugars. These sugars are found in only three bacterial species, the most notorious and prevalent of which is Mycobacterium tuberculosis, the microbe that causes TB.

After labeling the glycan, the researchers were able to visualize where it is located within the bacterial cell wall, and to study what happens to it throughout the first few days of tuberculosis infection of host immune cells.

The researchers now hope to use this approach to develop a diagnostic that could detect TB-associated glycans, either in culture or in a urine sample, which could offer a cheaper and faster alternative to existing diagnostics. Chest X-rays and molecular diagnostics are very accurate but are not always available in developing nations where TB rates are high. In those countries, TB is often diagnosed by culturing microbes from a sputum sample, but that test has a high false negative rate, and it can be difficult for some patients, especially children, to provide a sputum sample. This test also requires many weeks for the bacteria to grow, delaying diagnosis.

“There aren’t a lot of good diagnostic options, and there are some patient populations, including children, who have a hard time giving samples that can be analyzed. There’s a lot of impetus to develop very simple, fast tests,” says Laura Kiessling, the Novartis Professor of Chemistry at MIT and the senior author of the study.

MIT graduate student Stephanie Smelyansky is the lead author of the paper, which appears this week in the Proceedings of the National Academy of Sciences. Other authors include Chi-Wang Ma, an MIT postdoc; Victoria Marando PhD ’23; Gregory Babunovic, a postdoc at the Harvard T.H. Chan School of Public Health; So Young Lee, an MIT graduate student; and Bryan Bryson, an associate professor of biological engineering at MIT.

Labeling glycans

Glycans are found on the surfaces of most cells, where they perform critical functions such as mediating communication between cells.In bacteria, glycans help the microbes to enter host cells, and they also appear to communicate with the host immune system, in some cases blocking the immune response.

Mycobacterium tuberculosis has a really elaborate cell envelope compared to other bacteria, and it’s a rich structure that’s composed of a lot of different glycans,” Smelyansky says. “Something that’s often underappreciated is the fact that these glycans can also interact with our host cells. When our immune cells recognize these glycans, instead of sending out a danger signal, it can send the opposite message, that there’s no danger.”

Glycans are notoriously difficult to tag with any kind of probe, because unlike proteins or DNA, they don’t have distinctive sequences or chemical reactivities that can be targeted. And unlike proteins, they are not genetically encoded, so cells can’t be genetically engineered to produce sugars labeled with fluorescent tags such as green fluorescent protein.

One of the key glycans in M. tuberculosis, known as ManLAM, contains a rare sugar known as MTX, which is unusual in that it has a thioether — a sulfur atom sandwiched between two carbon atoms. This chemical group presented an opportunity to use a small-molecule tag that had been previously developed for labeling methionine, an  amino acid that contains a similar group.

The researchers showed that they could use this tag, known as an oxaziridine, to label ManLAM in M. tuberculosis. The researchers linked the oxaziridine to a fluorescent probe and showed that in M. tuberculosis, this tag showed up in the outer layer of the cell wall. When the researchers exposed the label to Mycobacterium smegmatis, a related bacterium that does not cause disease and does not have the sugar MTX, they saw no fluorescent signal.

“This is the first approach that really selectively allows us to visualize one glycan in particular,” Smelyansky says.

Better diagnostics

The researchers also showed that after labeling ManLAM in M. tuberculosis cells, they could track the cells as they infected immune cells called macrophages. Some tuberculosis researchers had hypothesized that the bacterial cells shed ManLAM once inside a host cell, and that those free glycans then interact with the host immune system. However, the MIT team found that the glycan appears to remain in the bacterial cell walls for at least the first few days of infection.

“The bacteria still have their cell walls attached to them. So it may be that some glycan is being released, but the majority of it is retained on the bacterial cell surface, which has never been shown before,” Smelyansky says.

The researchers now plan to use this approach to study what happens to the bacteria following treatment with different antibiotics, or immune stimulation of the macrophages. It could also be used to study in more detail how the bacterial cell wall is assembled, and how ManLAM helps bacteria get into macrophages and other cells.

“Having a handle to follow the bacteria is really valuable, and it will allow you to visualize processes, both in cells and in animal models, that were previously invisible,” Kiessling says.

She also hopes to use this approach to create new diagnostics for tuberculosis. There is currently a diagnostic in development that uses antibodies to detect ManLAM in a urine sample. However, this test only works well in patients with very active cases of TB, especially people who are immunosuppressed because of HIV or other conditions.

Using their small-molecule sensor instead of antibodies, the MIT team hopes to develop a more sensitive test that could detect ManLAM in the urine even when only small quantities are present.

“This is a beautifully elegant approach to selectively label the surface of mycobacteria, enabling real-time monitoring of cell wall dynamics in this important bacterial family. Such investigations will inform the development of novel strategies to diagnose, prevent, and treat mycobacterial disease, most notably tuberculosis, which remains a global health challenge,” says Todd Lowary, a distinguished research fellow at the Institute of Biological Chemistry, Academia Sinica, Taipei Taiwan, who was not involved in the research.

The research was funded by the National Institute of Allergy and Infectious Disease, the National Institutes of Health, the National Science Foundation, and the Croucher Fellowship.

MIT physicists snap the first images of “free-range” atoms

Mon, 05/05/2025 - 12:00am

MIT physicists have captured the first images of individual atoms freely interacting in space. The pictures reveal correlations among the “free-range” particles that until now were predicted but never directly observed. Their findings, appearing today in the journal Physical Review Letters, will help scientists visualize never-before-seen quantum phenomena in real space.

The images were taken using a technique developed by the team that first allows a cloud of atoms to move and interact freely. The researchers then turn on a lattice of light that briefly freezes the atoms in their tracks, and apply finely tuned lasers to quickly illuminate the suspended atoms, creating a picture of their positions before the atoms naturally dissipate.

The physicists applied the technique to visualize clouds of different types of atoms, and snapped a number of imaging firsts. The researchers directly observed atoms known as “bosons,” which bunched up in a quantum phenomenon to form a wave. They also captured atoms known as “fermions” in the act of pairing up in free space — a key mechanism that enables superconductivity.

“We are able to see single atoms in these interesting clouds of atoms and what they are doing in relation to each other, which is beautiful,” says Martin Zwierlein, the Thomas A. Frank Professor of Physics at MIT.

In the same journal issue, two other groups report using similar imaging techniques, including a team led by Nobel laureate Wolfgang Ketterle, the John D. MacArthur Professor of Physics at MIT. Ketterle’s group visualized enhanced pair correlations among bosons, while the other group, from École Normale Supérieure in Paris, led by Tarik Yefsah, a former postdoc in Zwierlein’s lab, imaged a cloud of noninteracting fermions.

The study by Zwierlein and his colleagues is co-authored by MIT graduate students Ruixiao Yao, Sungjae Chi, and Mingxuan Wang, and MIT assistant professor of physics Richard Fletcher.

Inside the cloud

A single atom is about one-tenth of a nanometer in diameter, which is one-millionth of the thickness of a strand of human hair. Unlike hair, atoms behave and interact according to the rules of quantum mechanics; it is their quantum nature that makes atoms difficult to understand. For example, we cannot simultaneously know precisely where an atom is and how fast it is moving.

Scientists can apply various methods to image individual atoms, including absorption imaging, where laser light shines onto the atom cloud and casts its shadow onto a camera screen.

“These techniques allow you to see the overall shape and structure of a cloud of atoms, but not the individual atoms themselves,” Zwierlein notes. “It’s like seeing a cloud in the sky, but not the individual water molecules that make up the cloud.”

He and his colleagues took a very different approach in order to directly image atoms interacting in free space. Their technique, called “atom-resolved microscopy,” involves first corralling a cloud of atoms in a loose trap formed by a laser beam. This trap contains the atoms in one place where they can freely interact. The researchers then flash on a lattice of light, which freezes the atoms in their positions. Then, a second laser illuminates the suspended atoms, whose fluorescence reveals their individual positions.

“The hardest part was to gather the light from the atoms without boiling them out of the optical lattice,” Zwierlein says. “You can imagine if you took a flamethrower to these atoms, they would not like that. So, we’ve learned some tricks through the years on how to do this. And it’s the first time we do it in-situ, where we can suddenly freeze the motion of the atoms when they’re strongly interacting, and see them, one after the other. That’s what makes this technique more powerful than what was done before.”

Bunches and pairs

The team applied the imaging technique to directly observe interactions among both bosons and fermions. Photons are an example of a boson, while electrons are a type of fermion. Atoms can be bosons or fermions, depending on their total spin, which is determined by whether the total number of their protons, neutrons, and electrons is even or odd. In general, bosons attract, whereas fermions repel.

Zwierlein and his colleagues first imaged a cloud of bosons made up of sodium atoms. At low temperatures, a cloud of bosons forms what’s known as a Bose-Einstein condensate — a state of matter where all bosons share one and the same quantum state. MIT’s Ketterle was one of the first to produce a Bose-Einstein condensate, of sodium atoms, for which he shared the 2001 Nobel Prize in Physics.

Zwierlein’s group now is able to image the individual sodium atoms within the cloud, to observe their quantum interactions. It has long been predicted that bosons should “bunch” together, having an increased probability to be near each other. This bunching is a direct consequence of their ability to share one and the same quantum mechanical wave. This wave-like character was first predicted by physicist Louis de Broglie. It is the “de Broglie wave” hypothesis that in part sparked the beginning of modern quantum mechanics.

“We understand so much more about the world from this wave-like nature,” Zwierlein says. “But it’s really tough to observe these quantum, wave-like effects. However, in our new microscope, we can visualize this wave directly.”

In their imaging experiments, the MIT team were able to see, for the first time in situ, bosons bunch together as they shared one quantum, correlated de Broglie wave. The team also imaged a cloud of two types of lithium atoms. Each type of atom is a fermion, that naturally repels its own kind, but that can strongly interact with other particular fermion types. As they imaged the cloud, the researchers observed that indeed, the opposite fermion types did interact, and formed fermion pairs — a coupling that they could directly see for the first time.

“This kind of pairing is the basis of a mathematical construction people came up with to explain experiments. But when you see pictures like these, it’s showing in a photograph, an object that was discovered in the mathematical world,” says study co-author Richard Fletcher. “So it’s a very nice reminder that physics is about physical things. It’s real.”

Going forward, the team will apply their imaging technique to visualize more exotic and less understood phenomena, such as “quantum Hall physics” — situations when interacting electrons display novel correlated behaviors in the presence of a magnetic field.

“That’s where theory gets really hairy — where people start drawing pictures instead of being able to write down a full-fledged theory because they can’t fully solve it,” Zwierlein says. “Now we can verify whether these cartoons of quantum Hall states are actually real. Because they are pretty bizarre states.”

This work was supported, in part, by National Science Foundation through the MIT-Harvard Center for Ultracold Atoms, as well as by the Air Force Office of Scientific Research, the Army Research Office, the Department of Energy, the Defense Advanced Projects Research Agency, a Vannevar Bush Faculty Fellowship, and the David and Lucile Packard Foundation.

The age-old problem of long-term care

Mon, 05/05/2025 - 12:00am

Caring well for the elderly is a familiar challenge. Some elderly people need close medical attention in facilities; others struggle with reduced capabilities while not wanting to leave their homes. For families, finding good care is hard and expensive, and already-burdened family members often pick up the slack.

The problem is expanding as birthrates drop while some segments of the population live longer, meaning that a growing portion of the population is elderly. In the U.S., there are currently three states currently where at least 20 percent of the population is 65 and older. (Yes, Florida is one.) But by 2050, demographic trends suggest, there will be 43 states with that profile.

In age terms, “America is becoming Florida,” quips MIT economist Jonathan Gruber. “And it’s not just America. The whole world is aging rapidly. The share of the population over 65 is growing rapidly everywhere, and within that, the share of the elderly that are over 85 is growing rapidly.”

In a new edited volume, Gruber and several other scholars explore the subject from a global perspective. The book, “Long-Term Care around the World,” is published this month by the University of Chicago Press. The co-editors are Gruber, the Ford Professor of Economics and chair of the Department of Economics at MIT; and Kathleen McGarry, a professor of economics at Stony Brook University.

The book looks at 10 relatively wealthy countries and how they approach the problem of long-term care. In their chapter about the U.S., Gruber and McGarry emphasize a remarkable fact: About one-third of long-term care for the elderly in the U.S. is informal, provided by family and friends, despite limited time and resources. Overall, long-term care is 2 percent of U.S. GDP.

“We have two fundamental long-term care problems in the U.S.,” Gruber says. “Too much informal care at home, and, relatedly, not enough options for elders to live with effective care in ‘congregate housing’ [or elder communities], even if they’re not sick enough for a nursing facility.”

The nature of the problem

The needs of the elderly sit in plain sight. In the U.S., about 30 percent of people 65 and over, and 60 percent of people 85 and over report limitations in basic activities. Getting dressed and taking baths are among the most common daily problems; shopping for groceries and managing money are also widely reported issues. Additionally, these limitations have mental health implications. About 10 percent of the elderly report depression, rising to 30 percent among those who struggle with three or more types of basic daily tasks.

Even so, the U.S. is not actually heavily dotted with nursing homes. In a country of about 330 million people, with 62 million being 65 and over, it’s unusual for an elderly person to be in one.

“We all think of nursing homes as where you go when you’re old, but there are only about 1.2 million people in nursing homes in America,” Gruber observes. “Which is a lot, but tiny compared to the share of people who are elderly in the U.S. and who have needs. Most people who have needs get them met at home.”

And while nursing homes can be costly, home care is too. Given an average U.S. salary of $23 per hour for a home health care aide, annual costs can reach six figures even with half-time care. As a result, many families simply help their elderly relatives as best they can.

Therefore, Gruber has found, we must account for the informal costs of elder care, too. Ultimately, Gruber says, informal help represents “an inefficient system of people taking care of their elderly parents at home, which is a stress on the family, and the elders don’t get enough care.”

To be sure, some people buy private long-term care insurance to defray these costs. But this is a tricky market, where insurers are concerned about “adverse selection,” people buying policies with a distinct need for them (beyond what insurers can detect). Rates therefore can seem high, and for limited, conditional benefits. Research by MIT economist Amy Finkelstein has shown that only 18 percent of long-term insurance policies are used.

“Private long-term care insurance is a market that just hasn’t worked well,” Gruber says. “It’s basically a fixed amount of money, should you meet certain conditions. And people are surprised by that, and it doesn’t meet their needs, and it’s expensive. We need a public solution.”

Congregate housing, a possible solution

Looking at long-term care internationally helps identify what those solutions might be. The U.S. does not neglect elder care, but could clearly broaden its affordable options.

“On the one hand, what jumped out at me is how normal the U.S. is,” Gruber says. “We’re in the middle of the pack in terms of the share of GDP we spend on long-term care.” However, some European countries that spend a similar share and also rely heavily on informal elder care, including Italy and Spain, have notably lower levels of GDP per capita.

Some other European countries with income levels closer to the U.S., including Germany and the Netherlands, do spend more on long-term elder care. The Netherlands tops the list by devoting about 4 percent of its GDP to this area.

However, in the U.S., the issue is not so much drastically changing how much it spends on long-term elder care, but how it spends. The Dutch have a relatively more extensive system of elder communities — the “congregate housing” for the elderly who are not desperately unwell, but simply find self-reliance increasingly hard.

“That’s the huge missing hole in the U.S. long-term care system, what do we do with people who aren’t sick enough for a nursing home, but probably shouldn’t be at home,” Gruber says. “Right now they stay at home, they’re lonely, they’re not getting services, their kids are super-stressed out, and they’re pulling millions of people out of the labor force, especially women. Everyone is unhappy about it, and they’re not growing GDP, so it’s hurting our economy and our well-being.”

Overall, then, Gruber thinks further investment in elder-care communities would be an example of effective government spending that can address the brewing crisis in long-term care — although it would require new federal legislation in a highly polarized political environment.

Could that happen? Could the U.S. invest more now and realize long-term financial benefits, while allowing working-age employees to spend more time at their jobs rather than acting as home caregivers? Making people more aware of the issue, Gruber thinks, is a necessary starting point.

“If anything might be bipartisan, it could be long-term care,” Gruber says. “Everybody has parents. A solution has to be bipartisan. Long-term care may be one of those areas where it’s possible.”

Support for the research was provided, in part, by the National Institute on Aging.

Radar and communications system extends signal range at millimeter-wave frequencies

Fri, 05/02/2025 - 4:00pm

A team from MIT Lincoln Laboratory has built and demonstrated the wideband selective propagation radar (WiSPR), a system capable of seeing out various distances at millimeter-wave (mmWave or MMW) frequencies. Typically, these high frequencies, which range from 30 to 300 gigahertz (GHz), are employed for only short-range operations. Using transmit-and-receive electronically scanned arrays of many antenna elements each, WiSPR produces narrow beams capable of quickly scanning around an area to detect objects of interest. The narrow beams can also be manipulated into broader beams for communications.

"Building a system with sufficient sensitivity to operate over long distances at these frequencies for radar and communications functions is challenging," says Greg Lyons, a senior staff member in the Airborne Radar Systems and Techniques Group, part of Lincoln Laboratory's ISR Systems and Technology R&D area. "We have many radar experts in our group, and we all debated whether such a system was even feasible. Much innovation is happening in the commercial sector, and we leveraged those advances to develop this multifunctional system."

The high signal bandwidth available at mmWave makes these frequencies appealing. Available licensed frequencies are quickly becoming overloaded, and harnessing mmWave frequencies frees up considerable bandwidth and reduces interference between systems. A high signal bandwidth is useful in a communications system to transmit more information, and in a radar system to improve range resolution (i.e., ability of radar to distinguish between objects in the same angular direction but at different distances from the radar).

The phases for success

In 2019, the laboratory team set out to assess the feasibility of their mmWave radar concept. Using commercial off-the-shelf radio-frequency integrated circuits (RFICs), which are chips that send and receive radio waves, they built a fixed-beam system (only capable of staring in one direction, not scanning) with horn antennas. During a demonstration on a foggy day at Joint Base Cape Cod, the proof-of-concept system successfully detected calibration objects at unprecedented ranges.  

"How do you build a prototype for what will eventually be a very complicated system?" asks program manager Christopher Serino, an assistant leader of the Airborne Radar Systems and Techniques Group. "From this feasibility testing, we showed that such a system could actually work, and identified the technology challenges. We knew those challenges would require innovative solutions, so that's where we focused our initial efforts."

WiSPR is based on multiple-element antenna arrays. Whether serving a radar or communications function, the arrays are phased, which means the phase between each antenna element is adjusted. This adjustment ensures all phases add together to steer the narrow beams in the desired direction. With this configuration of multiple elements phased up, the antenna becomes more directive in sending and receiving energy toward one location. (Such phased arrays are becoming ubiquitous in technologies like 5G smartphones, base stations, and satellites.)

To enable the tiny beams to continuously scan for objects, the team custom-built RFICs using state-of-the-art semiconductor technology and added digital capabilities to the chips. By controlling the behavior of these chips with custom firmware and software, the system can search for an object and, after the object is found, keep it in "track" while the search for additional objects continues — all without physically moving antennas or relying on an operator to tell the system what to do next.

"Phasing up elements in an array to get gain in a particular direction is standard practice," explains Deputy Program Manager David Conway, a senior staff member in the Integrated RF and Photonics Group. "What isn't standard is having this many elements with the RF at millimeter wavelengths still working together, still summing up their energy in transmit and receive, and capable of quickly scanning over very wide angles."

Line 'em up and cool 'em down

For the communications function, the team devised a novel beam alignment procedure.

"To be able to combine many antenna elements to have a radar reach out beyond typical MMW operating ranges — that's new," Serino says. "To be able to electronically scan the beams around as a radar with effectively zero latency between beams at these frequencies — that's new. Broadening some of those beams so you're not constantly reacquiring and repointing during communications — that's also new."

Another innovation key to WiSPR's development is a cooling arrangement that removes the large amount of heat dissipated in a small area behind the transmit elements, each having their own power amplifier.

Last year, the team demonstrated their prototype WiSPR system at the U.S. Army Aberdeen Proving Ground in Maryland, in collaboration with the U.S. Army Rapid Capabilities and Critical Technologies Office and the U.S. Army Test and Evaluation Command. WiSPR technology has since been transitioned to a vendor for production. By adopting WiSPR, Army units will be able to conduct their missions more effectively.

"We're anticipating that this system will be used in the not-too-distant future," Lyons says. "Our work has pushed the state of the art in MMW radars and communication systems for both military and commercial applications."

"This is exactly the kind of work Lincoln Laboratory is proud of: keeping an eye on the commercial sector and leveraging billions-of-dollars investments to build new technology, rather than starting from scratch," says Lincoln Laboratory assistant director Marc Viera.

This effort supported the U.S. Army Rapid Capabilities and Critical Technologies Office. The team consists of additional members from the laboratory's Airborne Radar Systems and Techniques, Integrated RF and Photonics, Mechanical Engineering, Advanced Capabilities and Systems, Homeland Protection Systems, and Transportation Safety and Resilience groups.

Novel AI model inspired by neural dynamics from the brain

Fri, 05/02/2025 - 3:30pm

Researchers from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) have developed a novel artificial intelligence model inspired by neural oscillations in the brain, with the goal of significantly advancing how machine learning algorithms handle long sequences of data.

AI often struggles with analyzing complex information that unfolds over long periods of time, such as climate trends, biological signals, or financial data. One new type of AI model, called "state-space models," has been designed specifically to understand these sequential patterns more effectively. However, existing state-space models often face challenges — they can become unstable or require a significant amount of computational resources when processing long data sequences.

To address these issues, CSAIL researchers T. Konstantin Rusch and Daniela Rus have developed what they call “linear oscillatory state-space models” (LinOSS), which leverage principles of forced harmonic oscillators — a concept deeply rooted in physics and observed in biological neural networks. This approach provides stable, expressive, and computationally efficient predictions without overly restrictive conditions on the model parameters.

"Our goal was to capture the stability and efficiency seen in biological neural systems and translate these principles into a machine learning framework," explains Rusch. "With LinOSS, we can now reliably learn long-range interactions, even in sequences spanning hundreds of thousands of data points or more."

The LinOSS model is unique in ensuring stable prediction by requiring far less restrictive design choices than previous methods. Moreover, the researchers rigorously proved the model’s universal approximation capability, meaning it can approximate any continuous, causal function relating input and output sequences.

Empirical testing demonstrated that LinOSS consistently outperformed existing state-of-the-art models across various demanding sequence classification and forecasting tasks. Notably, LinOSS outperformed the widely-used Mamba model by nearly two times in tasks involving sequences of extreme length.

Recognized for its significance, the research was selected for an oral presentation at ICLR 2025 — an honor awarded to only the top 1 percent of submissions. The MIT researchers anticipate that the LinOSS model could significantly impact any fields that would benefit from accurate and efficient long-horizon forecasting and classification, including health-care analytics, climate science, autonomous driving, and financial forecasting.

"This work exemplifies how mathematical rigor can lead to performance breakthroughs and broad applications," Rus says. "With LinOSS, we’re providing the scientific community with a powerful tool for understanding and predicting complex systems, bridging the gap between biological inspiration and computational innovation."

The team imagines that the emergence of a new paradigm like LinOSS will be of interest to machine learning practitioners to build upon. Looking ahead, the researchers plan to apply their model to an even wider range of different data modalities. Moreover, they suggest that LinOSS could provide valuable insights into neuroscience, potentially deepening our understanding of the brain itself.

Their work was supported by the Swiss National Science Foundation, the Schmidt AI2050 program, and the U.S. Department of the Air Force Artificial Intelligence Accelerator.

TeleAbsence: Poetic encounters with the past

Fri, 05/02/2025 - 3:15pm

In the dim light of the lab, friends, family, and strangers watched the image of a pianist playing for them, the pianist’s fingers projected onto the moving keys of a real grand piano that filled the space with music.

Watching the ghostly musicians, faces and bodies blurred at their edges, several listeners shared one strong but strange conviction: “feeling someone’s presence” while “also knowing that I am the only one in the room.”

“It’s tough to explain,” another listener said. “It felt like they were in the room with me, but at the same time, not.”

That presence of absence is at the heart of TeleAbsence, a project by the MIT Media Lab’s Tangible Media group that focuses on technologies that create illusory communication with the dead and with past selves.

But rather than a “Black Mirror”-type scenario of synthesizing literal loved ones, the project led by Hiroshi Ishii, the Jerome B. Wiesner Professor of Media Arts and Sciences, instead seeks what it calls “poetic encounters” that reach across time and memory.

The project recently published a positioning paper in PRESENCE: Virtual and Augmented Reality that presents the design principles behind TeleAbsence, and how it could help people cope with loss and plan for how they might be remembered.

The phantom pianists of the MirrorFugue project, created by Tangible Media graduate Xiao Xiao ’09, SM ’11, PhD ’16, are one of the best-known examples of the project. On April 30, Xiao, now director and principal investigator at the Institute for Future Technologies of Da Vinci Higher Education in Paris, shared results from the first experimental study of TeleAbsence through MirrorFugue at the 2025 CHI conference on Human Factors in Computing Systems in Yokohama, Japan.

When Ishii spoke about TeleAbsence at the XPANSE 2024 conference in Abu Dhabi, “about 20 people came up to me after, and all of them told me they had tears in their eyes … the talk reminded them about a wife or a father who passed away,” he says. “One thing is clear: They want to see them again and talk to them again, metaphorically.”

Messages in bottles

As the director of the Tangible Media group, Ishii has been a world leader in telepresence, using technologies to connect people over physical distance. But when his mother died in 1998, Ishii says the pain of the loss prompted him to think about how much we long to connect across the distance of time.

His mother wrote poetry, and one of his first experiments in TeleAbsence was the creation of a Twitterbot that would post snippets of her poetry. Others watching the account online were so moved that they began posting photos of flowers to the feed to honor the mother and son.

“That was a turning point for TeleAbsence, and I wanted to expand this concept,” Ishii says.

Illusory communication, like the posted poems, is one key design principle of TeleAbsence. Even though users know the “conversation” is one-way, the researchers write, it can be comforting and cathartic to have a tangible way to reach out across time.

Finding ways to make memories material is another important design principle. One of the projects created by Ishii and colleagues is a series of glass bottles, reminiscent of the soy sauce bottles Ishii’s mother used while cooking. Open one of the bottles, and the sounds of chopping, of sizzling onions, of a radio playing quietly in the background, of a maternal voice, reunite a son with his mother.

Ishii says sight and sound are the primary modalities of TeleAbsence technologies for now, because although the senses of touch, smell, and taste are known to be powerful memory triggers, “it is a very big challenge to record that kind of multimodal moment.”

At the same time, one of the other pillars of TeleAbsence is the presence of absence. These are the physical markers, or traces, of a person that serve to remind us both of the person and that the person is gone. One of the most powerful examples, the researchers write, is the permanent “shadow” of Hiroshima Japanese resident Mitsuno Ochi, her silhouette transferred to stone steps 260 meters from where the atomic bomb detonated in 1945.

“Abstraction is very important,” Ishii says. “We want something to recall a moment, not physically recreate it.”

With the bottles, for instance, people have asked Ishii and his colleagues whether it might be more evocative to fill them with a perfume or drink. “But our philosophy is to make a bottle completely empty,” he explains. “The most important thing to let people imagine, based on the memory.”

Other important design principles within TeleAbsence include traces of reflection — the ephemera of faint pen scratches and blotted ink on a preserved letter, for instance — and the concept of remote time. TeleAbsence should go beyond dredging up a memory of a loved one, the researchers insist, and should instead produce a sense of being transported to spend a moment in the past with them.

Time travelers

For Xiao, who has played the piano her whole life, MirrorFugue is a “deeply personal project” that allowed her to travel to a time in her childhood that was almost lost to her.

Her parents moved from China to the United States when she was a baby — but it took eight years for Xiao to follow. “The piano, in a sense, was almost like my first language,” she recalls. “And then when I moved to America, my brain overwrote bits of my childhood where my operating system used to be in Chinese, and now it’s very much in English. But throughout this whole time, music and the piano stayed constant.”

MirrorFugue’s “sense of kind-of being there and not being there, and the wish to connect with oneself from the past, comes from my own desire to connect with my own past self,” she adds.

The new MirrorFugue study puts some empirical data behind the concept of TeleAbsence, she says. Its 28 participants were fitted with sensors to measure changes in their heart rate and hand movements during the experience. They were extensively interviewed about their perceptions and emotions afterward. The recorded images came from pianists ranging in experience from children early in their lessons to professional pianists like the late Ryuichi Sakamoto.

The researchers found that emotional experiences described by the listeners were significantly influenced by whether the listeners knew the pianist, as well as whether the pianist was known by the listeners to be alive or dead.

Some participants placed their own hands alongside the ghosts to play impromptu duets. One daughter, who said she had not paid close attention to her father’s playing when he was alive, was newly impressed by his talent. One person felt empathy watching his past self struggle through a new piece of music. A young girl, mouth slightly open in concentration and fingers small on the keys, showed her mother a past daughter that wasn’t possible to see in old photos.

The longing for past people and past selves can be “a deep sadness that will never go away,” says Xiao. “You’ll always carry it with you, but it also makes you sensitive to certain aesthetic experiences that’s also beautiful.”

“Once you’ve had that experience, it really resonates,” she adds, “And I think that’s why TeleAbsence resonates with so many people.”

Uncanny valleys and curated memory

Acutely aware of the potential ethical dangers of their research, the TeleAbsence scientists have worked with grief researchers and psychologists to better understand the implications of building these bridges through time.

For instance, “one thing we learned is that it depends on how long ago a person passed away,” says Ishii. “Right after death, when it’s very difficult for many people, this representation matters. But you have to make important informed decisions about whether this drags out the grief too long.”

TeleAbsence could comfort the dying, he says, by “knowing there is a means by which they are going to live on for their descendants.” He encourages people to consider curating “high-quality, condensed information,” such as their social media posts, that could be used for this purpose.

“But of course many families do not have ideal relationships, so I can easily think of the case where a descendant might not have any interest” in interacting with their ancestors through TeleAbsence, Ishii notes.

TeleAbsence should never fully recreate or generate new content for a loved one, he insists, pointing to the rise of “ghost bot” startups, companies that collect data on a person to create an “artificial, generative AI-based avatar that speaks what they never spoke, or do gestures or facial expressions.”

A recent viral video of a mother in Korea “reunited” in virtual reality with an avatar of her dead daughter, Ishii says, made him “very depressed, because they’re doing grief as entertainment, consumption for an audience.”

Xiao thinks there might still be some role for generative AI in the TeleAbsence space. She is writing a research proposal for MirrorFugue that would include representations of past pianists. “I think right now we’re getting to the point with generative AI that we can generate hand movements and we can transcribe the MIDI from the audio so that we can conjure up Franz Listz or Mozart or somebody, a really historical figure.”

“Now of course, it gets a little bit tricky, and we have discussed this, the role of AI and how to avoid the uncanny valley, how to avoid deceiving people,” she says. “But from a researcher’s perspective, it actually excites me a lot, the possibility to be able to empirically test these things.”

The importance of emptiness

Along with Ishii’s mother, the PRESENCE paper was also dedicated “in loving memory” to Elise O’Hara, a beloved Media Lab administrative assistant who worked with Tangible Media until her unexpected death in 2023. Her presence — and her absence — are felt deeply every day, says Ishii.

He wonders if TeleAbsence could someday become a common word “to describe something that was there, but is now gone.”

“When there is a place on a bookshelf where a book should be,” he says, “my students say, ‘oh, that’s a teleabsence.’”

Like a sudden silence in the middle of a song, or the empty white space of a painting, emptiness can hold important meaning. It’s an idea that we should make more room for in our lives, Ishii says.

“Because now we’re so busy, so many notification messages from your smartphone, and we are all distracted, always,” he suggests. “So emptiness and impermanence, presence of absence, if those concepts can be accepted, then people can think a bit more poetically.”

Study of facial bacteria could lead to probiotics that promote healthy skin

Thu, 05/01/2025 - 11:00am

The composition of bacterial populations living on our faces plays a significant role in the development of acne and other skin conditions such as eczema. Two species of bacteria predominate in most people, but how they interact with each other, and how those interactions may contribute to disease, has been difficult to study.

MIT researchers have now revealed the dynamics of those interactions in more detail than previously possible, shedding light on when and how new bacterial strains emerge on the skin of the face. Their findings could help guide the development of new treatments for acne and other conditions, and may also help to optimize the timing of such treatments.

The researchers found that many new strains of Cutibacterium acnes, a species believed to contribute to the development of acne, are acquired during the early teenage years. But after that, the makeup of these populations becomes very stable and doesn’t change much even when exposed to new strains.

That suggests that this transitional stage could be the best window for introducing probiotic strains of C. acnes, says Tami Lieberman, an associate professor of civil and environmental engineering, a member of MIT’s Institute for Medical Engineering and Science, and the senior author of the study.

“We found that there are some surprising dynamics, and these dynamics provide insights for how to design probiotic therapy,” Lieberman says. “If we had a strain that we knew could prevent acne, these results would suggest we should make sure we apply them early during the transition to adulthood, to really get them to engraft.”

Jacob Baker PhD ’24, who is now the chief scientific officer at Taxa Technologies, is the lead author of the paper, which appears today in Cell Host and Microbe. Other authors include MIT graduate student Evan Qu, MIT postdoc Christopher Mancuso, Harvard University graduate student A. Delphine Tripp, and former MIT postdoc Arolyn Conwill PhD ’18.

Microbial dynamics

Although C. acnes has been implicated in the development of acne, it is still unclear exactly why acne develops in some people but not others — it may be that some strains are more likely to cause skin inflammation, or there may be differences in how the host immune system responds to the bacteria, Lieberman says. There are probiotic strains of C. acnes now available, which are thought to help prevent acne, but the benefits of these strains have not been proven.

Along with C. acnes, the other predominant bacterium found on the face is Staphylococcus epidermidis. Together, these two strains make up about 80 percent of the strains in the adult facial skin microbiome. Both of these species exist in different strains, or lineages, that vary by a small number of genetic mutations. However, until now, researchers had not been able to accurately measure this diversity or track how it changes over time.

Learning more about those dynamics could help researchers answer key questions that could help them develop new probiotic treatments for acne: How easy is it for new lineages to establish themselves on the skin, and what is the best time to introduce them?

To study these population shifts, the researchers had to measure how individual cells evolve over time. To do that, they began by obtaining microbiome samples from 30 children at a Boston-area school and from 27 of their parents. Studying members of the same family enabled the researchers to analyze the likelihood of different lineages being transferred between people in close contact.

For about half of the individuals, the researchers were able to take samples at multiple time points, and for the rest, only once. For each sample, they isolated individual cells and grew them into colonies, then sequenced their genomes.

This allowed the researchers to learn how many lineages were found on each person, how they changed over time, and how different cells from the same lineage were. From that information, the researchers could infer what had happened to those lineages in the recent past and how long they had been present on the individual.

Overall, the researchers identified a total of 89 C. acnes lineages and 78 S. epidermidis lineages, with up to 11 of each found in each person’s microbiome. Previous work had suggested that in each person’s facial skin microbiome, lineages of these two skin bacteria remain stable over long periods of time, but the MIT team found that these populations are actually more dynamic than previously thought.

“We wanted to know if these communities were truly stable, and if there could be times where they weren’t stable. In particular, if the transition to an adult skin like microbiome would have a higher rate of acquisition of new lineages,” Lieberman says.

During the early teens, an increase in hormone production results in increased oil on the skin, which is a good food source for bacteria. It has previously been shown that during this time, the density of bacteria on the skin of the face increases by about 10,000-fold. In this study, the researchers found that while the composition of C. acnes populations tended to remain very stable over time, the early teenage years present an opportunity for many more lineages of C. acnes to appear.

“For C. acnes, what we were able to show was that people do get strains throughout life, but very rarely,” Lieberman says. “We see the highest rate of influx when teenagers are transitioning to a more adult-like skin microbiome.”

The findings suggest that for topical probiotic treatments for acne, the best time to apply them is during the early teenage years, when there could be more opportunity for probiotic strains to become established.

Population turnover

Later in adulthood, there is a little bit of sharing of C. acnes strains between parents living in the same household, but the rate of turnover in any individual person’s microbiome is still very low, Lieberman says.

The researchers found that S. epidermidis has a much higher turnover rate than C. acnes — each S. epidermidis strain lives on the face for an average of less than two years. However, there was not very much overlap in the S. epidermidis lineages shared by members of the same household, suggesting that transfer of strains between people is not causing the high turnover rate.

“That suggests that something is preventing homogenization between people,” Lieberman says. “It could be host genetics or host behavior, or people using different topicals or different moisturizers, or it could be active restriction of new migrants from the bacteria that are already there at that moment.”

Now that they’ve shown that new C. acnes strains can be acquired during the early teenage years, the researchers hope to study whether the timing of this acquisition affects how the immune system responds to them. They also hope to learn more about how people maintain such different microbiome populations even when exposed to new lineages through close contact with family members.

“We want to understand why we each have unique strain communities despite the fact that there is this constant accessibility and high turnover, specifically for S. epidermidis,” Lieberman says. “What’s driving this constant turnover in S. epidermidis, and what are the implications of these new colonizations for acne during adolescence?”

The research was funded by the MIT Center for Microbiome Informatics and Therapeutics, a Smith Family Foundation Award for Excellence in Biomedical Research, and the National Institutes of Health.

Making AI models more trustworthy for high-stakes settings

Thu, 05/01/2025 - 12:00am

The ambiguity in medical imaging can present major challenges for clinicians who are trying to identify disease. For instance, in a chest X-ray, pleural effusion, an abnormal buildup of fluid in the lungs, can look very much like pulmonary infiltrates, which are accumulations of pus or blood.

An artificial intelligence model could assist the clinician in X-ray analysis by helping to identify subtle details and boosting the efficiency of the diagnosis process. But because so many possible conditions could be present in one image, the clinician would likely want to consider a set of possibilities, rather than only having one AI prediction to evaluate.

One promising way to produce a set of possibilities, called conformal classification, is convenient because it can be readily implemented on top of an existing machine-learning model. However, it can produce sets that are impractically large. 

MIT researchers have now developed a simple and effective improvement that can reduce the size of prediction sets by up to 30 percent while also making predictions more reliable.

Having a smaller prediction set may help a clinician zero in on the right diagnosis more efficiently, which could improve and streamline treatment for patients. This method could be useful across a range of classification tasks — say, for identifying the species of an animal in an image from a wildlife park — as it provides a smaller but more accurate set of options.

“With fewer classes to consider, the sets of predictions are naturally more informative in that you are choosing between fewer options. In a sense, you are not really sacrificing anything in terms of accuracy for something that is more informative,” says Divya Shanmugam PhD ’24, a postdoc at Cornell Tech who conducted this research while she was an MIT graduate student.

Shanmugam is joined on the paper by Helen Lu ’24; Swami Sankaranarayanan, a former MIT postdoc who is now a research scientist at Lilia Biosciences; and senior author John Guttag, the Dugald C. Jackson Professor of Computer Science and Electrical Engineering at MIT and a member of the MIT Computer Science and Artificial Intelligence Laboratory (CSAIL). The research will be presented at the Conference on Computer Vision and Pattern Recognition in June.

Prediction guarantees

AI assistants deployed for high-stakes tasks, like classifying diseases in medical images, are typically designed to produce a probability score along with each prediction so a user can gauge the model’s confidence. For instance, a model might predict that there is a 20 percent chance an image corresponds to a particular diagnosis, like pleurisy.

But it is difficult to trust a model’s predicted confidence because much prior research has shown that these probabilities can be inaccurate. With conformal classification, the model’s prediction is replaced by a set of the most probable diagnoses along with a guarantee that the correct diagnosis is somewhere in the set.

But the inherent uncertainty in AI predictions often causes the model to output sets that are far too large to be useful.

For instance, if a model is classifying an animal in an image as one of 10,000 potential species, it might output a set of 200 predictions so it can offer a strong guarantee.

“That is quite a few classes for someone to sift through to figure out what the right class is,” Shanmugam says.

The technique can also be unreliable because tiny changes to inputs, like slightly rotating an image, can yield entirely different sets of predictions.

To make conformal classification more useful, the researchers applied a technique developed to improve the accuracy of computer vision models called test-time augmentation (TTA).

TTA creates multiple augmentations of a single image in a dataset, perhaps by cropping the image, flipping it, zooming in, etc. Then it applies a computer vision model to each version of the same image and aggregates its predictions.

“In this way, you get multiple predictions from a single example. Aggregating predictions in this way improves predictions in terms of accuracy and robustness,” Shanmugam explains.

Maximizing accuracy

To apply TTA, the researchers hold out some labeled image data used for the conformal classification process. They learn to aggregate the augmentations on these held-out data, automatically augmenting the images in a way that maximizes the accuracy of the underlying model’s predictions.

Then they run conformal classification on the model’s new, TTA-transformed predictions. The conformal classifier outputs a smaller set of probable predictions for the same confidence guarantee.

“Combining test-time augmentation with conformal prediction is simple to implement, effective in practice, and requires no model retraining,” Shanmugam says.

Compared to prior work in conformal prediction across several standard image classification benchmarks, their TTA-augmented method reduced prediction set sizes across experiments, from 10 to 30 percent.

Importantly, the technique achieves this reduction in prediction set size while maintaining the probability guarantee.

The researchers also found that, even though they are sacrificing some labeled data that would normally be used for the conformal classification procedure, TTA boosts accuracy enough to outweigh the cost of losing those data.

“It raises interesting questions about how we used labeled data after model training. The allocation of labeled data between different post-training steps is an important direction for future work,” Shanmugam says.

In the future, the researchers want to validate the effectiveness of such an approach in the context of models that classify text instead of images. To further improve the work, the researchers are also considering ways to reduce the amount of computation required for TTA.

This research is funded, in part, by the Wistrom Corporation.

Studying work, life, and economics

Wed, 04/30/2025 - 4:50pm

For policymakers investigating the effective transition of an economy from agriculture to manufacturing and services, there are complex economic, institutional, and practical considerations. “Are certain regions trapped in an under-industrialization state?” asks Tishara Garg, an economics doctoral student at MIT. “If so, can government policy help them escape this trap and transition to an economy characterized by higher levels of industrialization and better-paying jobs?” 

Garg’s research focuses on trade, economic geography, and development. Her studies yielded the paper “Can Industrial Policy Overcome Coordination Failures: Theory and Evidence from Industrial Zones,” which investigates whether economic policy can shift an economy from an undesirable state to a desirable state. 

Garg’s work combines tools from industrial organization and numerical algebraic geometry. Her paper finds that regions in India with state-developed industrial zones are 38 percent more likely to shift from a low to high industrialization state over a 15-year period than those without such zones.  

The kinds of questions uncovered during her studies aren’t easily answered using standard technical and econometric tools, so she’s developing new ones. “One of my study’s main contributions is a methodological framework that draws on ideas from different areas,” she notes. “These tools not only help me study the question I want to answer, but are also general enough to help study a broader set of questions around multiple challenges.”

The new tools she’s developed, along with a willingness to engage with other disciplines, have helped her discover innovative ways to approach these challenges while learning to work with new ones, options she asserts are actively encouraged at an institution like MIT.

“I benefited from having an open mind and learning different things,” she says.

“I was introduced to academia late”

Garg’s journey from Kaithal, India, to MIT wasn’t especially smooth, as societal pressures exerted a powerful influence. “The traditional path for someone like me is to finish school, enter an arranged marriage, and start a family,” she says. “But I was good at school and wanted to do more.” 

Garg, who hails from a background with limited access to information on career development opportunities, took to math early. “I chose business in high school because I planned to become an accountant,” she recalls. “My uncle was an accountant.”

While pursuing the successful completion of a high school business track, she became interested in economics. “I didn’t know much about economics, but I came to enjoy it,” she says. Garg relishes the pursuit of deductive reasoning that begins with a set of assumptions and builds, step by step, toward a well-defined, clear conclusion. She especially enjoys grappling with the arguments she found in textbooks. She continued to study economics as an undergraduate at the University of Delhi, and later earned her master’s from the Indian Statistical Institute. Doctoral study wasn’t an option until she made it one.

“It took me some time to convince my parents,” she says. She spent a year at a hedge fund before applying to economics doctoral programs in the United States and choosing MIT. “I was introduced to academia late,” she notes. “But my heart was being drawn to the academic path.”

Answering ambitious and important questions

Garg, who hadn’t left India before her arrival in Cambridge, Massachusetts, found the transition challenging. “There were new cultural norms, a language barrier, different foods, and no preexisting social network,” she says. Garg relied on friends and MIT faculty for support when she arrived in 2019. 

“When Covid hit, the department looked out for me,” she says. Garg recalls regular check-ins from a faculty advisor and the kind of camaraderie that can grow from shared circumstances, like Covid-related sheltering protocols. A world that forced her to successfully navigate a new and unfamiliar reality helped reshape how she viewed herself. “Support from the community at MIT helped me grow in many ways,” she recalls, “I found my voice here.”

Once she began her studies, one of the major differences Garg found was the diversity of opinions in her field of inquiry. “At MIT, I could speak with students and faculty specializing in trade, development economics, industrial organization, macroeconomics, and more,” she says. “I had limited exposure to many of these subfields before coming to MIT.” 

She quickly found her footing, leaning heavily on both her past successes and the academic habits she developed during her studies in India. “I’m not a passive learner,” she says. “My style is active, critical, and engaged.”

Conducting her research exposes Garg to new ideas. She learned the value of exploring other disciplines’ approaches to problem-solving, which was encouraged and enabled at MIT. 

One of the classes she came to enjoy most was a course in industrial organization taught by Tobias Salz. “I had little familiarity with the material, and it was highly technical — but he taught it in such a clear and intuitive way that I found myself truly enjoying the class, even though it was held during the pandemic,” she recalls. That early experience laid the groundwork for future research. Salz went on to advise her dissertation, helping her engage with work she would build upon.

“Answering ambitious and important questions is what draws me to the work,” Garg says. “I enjoy learning, I enjoy the creative process of bringing different ideas together and MIT's environment has made it easy for me to pick up new things.”

Working with her advisors at MIT helped Garg formalize her research and appreciate the value of uncovering questions and developing approaches to answer them. Professor Abhijit Banerjee, an advisor and Nobel laureate, helped her understand the importance of appreciating different traditions while also staying true to how you think about the problem, she recalls. “That gave me the confidence to pursue the questions in ways that felt most compelling and personal to me,” she says, “even if they didn’t fit neatly into disciplinary boundaries.”

This encouragement, combined with the breadth of perspectives at MIT, pushed her to think creatively about research challenges and to look beyond traditional tools to discover solutions. “MIT’s faculty have helped me improve the way I think and refine my approach to this work,” she says.

Paying it forward

Garg, who will continue her research as a postdoc at Princeton University in the fall and begin her career as a professor at Stanford University in 2026, singles out her network of friends and advisors for special praise.

“From regular check-ins with my advisors to the relationships that help me find balance with my studies, the people at MIT have been invaluable,” she says. 

Garg is especially invested in mentorship opportunities available as a researcher and professor. “I benefited from the network of friends and mentors at MIT and I want to pay it forward — especially for women, and others from backgrounds like mine,” she says.

She cites the work of her advisors, David Atkin and Dave Donaldson — with whom she is also collaborating on research studying incidences of economic distortions — as both major influences on her development and a key reason she’s committed to mentoring others. “They’ve been with me every step of the way,” she says. 

Garg recommends keeping an open mind, above all. “Some of my students didn’t come from a math-heavy background and would restrict themselves or otherwise get discouraged from pursuing theoretical work,” she says. “But I always encouraged them to pursue their interests above all, even if it scared them.” 

The variety of ideas available in her area of inquiry still fascinates Garg, who’s excited about what’s next. “Don’t shy from big questions,” she says. “Explore the big idea.”

AI-enabled translations initiative empowers Ukrainian learners with new skills

Wed, 04/30/2025 - 4:35pm

With war continuing to disrupt education for millions of Ukrainian high school and college students, many are turning to online resources, including MIT OpenCourseWare, a part of MIT Open Learning offering educational materials from more than 2,500 MIT undergraduate and graduate courses.

For Ukrainian high school senior Sofiia Lipkevych and other students, MIT OpenCourseWare has provided valuable opportunities to take courses in key subject areas. However, while multiple Ukrainian students study English, many do not yet have sufficient command of the language to be able to fully understand and use the often very technical and complex OpenCourseWare content and materials.

“At my school, I saw firsthand how language barriers prevented many Ukrainian students from accessing world-class education,” says Lipkevych.

She was able to address this challenge as a participant in the Ukrainian Leadership and Technology Academy (ULTA), established by Ukrainian MIT students Dima Yanovsky and Andrii Zahorodnii. During summer 2024 at ULTA, Lipkevych worked on a browser extension that translated YouTube videos in real-time. Since MIT OpenCourseWare was a main source of learning materials for students participating in ULTA, she was inspired to translate OpenCourseWare lectures directly and to have this translation widely available on the OpenCourseWare website and YouTube channel. She reached out to Professor Elizabeth Wood, founding director of the MIT Ukraine Program, who connected her with MIT OpenCourseWare Director Curt Newton.

Although there had been some translations of MIT OpenCourseWare’s educational resources available beginning in 2004, these initial translations were conducted manually by several global partners, without the efficiencies of the latest artificial intelligence tools, and over time the programs couldn’t be sustained, and shut down.

“We were thrilled to have this contact with ULTA,” says Newton. “We’ve been missing having a vibrant translation community, and we are excited to have a ‘phase 2’ of translations emerge.”

The ULTA team selected courses to translate based on demand among Ukrainian students, focusing on foundational subjects that are prerequisites for advanced learning — particularly those for which high-quality, Ukrainian-language materials are scarce. Starting with caption translations on videos of lectures, the team has translated the following courses so far: 18.06 (Linear Algebra)2.003SC (Engineering Dynamics), 5.60 (Thermodynamics & Kinetics)6.006 (Introduction to Algorithms), and 6.0001 (Introduction to Computer Science and Programming in Python). They also worked directly with Andy Eskenazi, a PhD student in the MIT Department of Aeronautics and Astronautics, to translate 16.002 (How to CAD Almost Anything - Siemens NX Edition).

The ULTA team developed multiple tools to help break language barriers. For MIT OpenCourseWare’s PDF content available through the ULTA program, they created a specialized tool that uses optical character recognition to recognize LaTeX in documents — such as problem sets and other materials — and then used a few large language models to translate them, all while maintaining technical accuracy. The team built a glossary of technical terms used in the courses and their corresponding Ukrainian translations, to help make sure that the wording was correct and consistent. Each translation also undergoes human review to further ensure accuracy and high quality.

For video content, the team initially created a browser extension that can translate YouTube video captions in real-time. They ultimately collaborated with ElevenLabs, implementing their advanced AI dubbing editor that preserves the original speaker's tone, pace, and emotional delivery. The lectures are translated in the ElevenLabs dubbing editor, and then the audio is uploaded to the MIT OpenCourseWare YouTube channel.

The team is currently finalizing the translation of the audio for class 9.13 (The Human Brain), taught by MIT Professor Nancy Kanwisher, which Lipkevych says they selected for its interdisciplinary nature and appeal to a wide variety of learners.

This Ukrainian translation project highlights the transformative potential of the latest translation technologies, building upon a 2023 MIT OpenCourseWare experiment using the Google Aloud AI dubbing prototype on a few courses, including MIT Professor Patrick Winston’s How to Speak. The advanced capabilities of the dubbing editor used in this project are opening up possibilities for a much greater variety of language offerings throughout MIT OpenCourseWare materials.

“I expect that in a few years we’ll look back and see that this was the moment when things shifted for OpenCourseWare to be truly usable for the whole world,” says Newton.

Community-led language translations of MIT OpenCourseWare materials serve as a high-impact example of the power of OpenCourseWare’s Creative Commons licensing, which grants everyone the right to revise materials to suit their particular needs and redistribute those revisions to the world.

While there isn’t currently a way for users of the MIT OpenCourseWare platform to quickly identify which videos are available in which languages, MIT OpenCourseWare is working toward building this capability into its website, as well as expanding its number of offerings in different languages.

“This project represents more than just translation,” says Lipkevych. “We’re enabling thousands of Ukrainians to build skills that will be essential for the country’s eventual reconstruction. We’re also hoping this model of collaboration can be extended to other languages and institutions, creating a template for making high-quality education accessible worldwide.”

Pages