MIT Latest News

A new vaccine approach could help combat future coronavirus pandemics
A new experimental vaccine developed by researchers at MIT and Caltech could offer protection against emerging variants of SARS-CoV-2, as well as related coronaviruses, known as sarbecoviruses, that could spill over from animals to humans.
In addition to SARS-CoV-2, the virus that causes COVID-19, sarbecoviruses — a subgenus of coronaviruses — include the virus that led to the outbreak of the original SARS in the early 2000s. Sarbecoviruses that currently circulate in bats and other mammals may also hold the potential to spread to humans in the future.
By attaching up to eight different versions of sarbecovirus receptor-binding proteins (RBDs) to nanoparticles, the researchers created a vaccine that generates antibodies that recognize regions of RBDs that tend to remain unchanged across all strains of the viruses. That makes it much more difficult for viruses to evolve to escape vaccine-induced antibodies.
“This work is an example of how bringing together computation and immunological experiments can be fruitful,” says Arup K. Chakraborty, the John M. Deutch Institute Professor at MIT and a member of MIT’s Institute for Medical Engineering and Science and the Ragon Institute of MIT, MGH and Harvard University.
Chakraborty and Pamela Bjorkman, a professor of biology and biological engineering at Caltech, are the senior authors of the study, which appears today in Cell. The paper’s lead authors are Eric Wang PhD ’24, Caltech postdoc Alexander Cohen, and Caltech graduate student Luis Caldera.
Mosaic nanoparticles
The new study builds on a project begun in Bjorkman’s lab, in which she and Cohen created a “mosaic” 60-mer nanoparticle that presents eight different sarbecovirus RBD proteins. The RBD is the part of the viral spike protein that helps the virus get into host cells. It is also the region of the coronavirus spike protein that is usually targeted by antibodies against sarbecoviruses.
RBDs contain some regions that are variable and can easily mutate to escape antibodies. Most of the antibodies generated by mRNA COVID-19 vaccines target those variable regions because they are more easily accessible. That is one reason why mRNA vaccines need to be updated to keep up with the emergence of new strains.
If researchers could create a vaccine that stimulates production of antibodies that target RBD regions that can’t easily change and are shared across viral strains, it could offer broader protection against a variety of sarbecoviruses.
Such a vaccine would have to stimulate B cells that have receptors (which then become antibodies) that target those shared, or “conserved,” regions. When B cells circulating in the body encounter a vaccine or other antigen, their B cell receptors, each of which have two “arms,” are more effectively activated if two copies of the antigen are available for binding to each arm. The conserved regions tend to be less accessible to B cell receptors, so if a nanoparticle vaccine presents just one type of RBD, B cells with receptors that bind to the more accessible variable regions, are most likely to be activated.
To overcome this, the Caltech researchers designed a nanoparticle vaccine that includes 60 copies of RBDs from eight different related sarbecoviruses, which have different variable regions but similar conserved regions. Because eight different RBDs are displayed on each nanoparticle, it’s unlikely that two identical RBDs will end up next to each other. Therefore, when a B cell receptor encounters the nanoparticle immunogen, the B cell is more likely to become activated if its receptor can recognize the conserved regions of the RBD.
“The concept behind the vaccine is that by co-displaying all these different RBDs on the nanoparticle, you are selecting for B cells that recognize the conserved regions that are shared between them,” Cohen says. “As a result, you’re selecting for B cells that are more cross-reactive. Therefore, the antibody response would be more cross-reactive and you could potentially get broader protection.”
In studies conducted in animals, the researchers showed that this vaccine, known as mosaic-8, produced strong antibody responses against diverse strains of SARS-CoV-2 and other sarbecoviruses and protected from challenges by both SARS-CoV-2 and SARS-CoV (original SARS).
Broadly neutralizing antibodies
After these studies were published in 2021 and 2022, the Caltech researchers teamed up with Chakraborty’s lab at MIT to pursue computational strategies that could allow them to identify RBD combinations that would generate even better antibody responses against a wider variety of sarbecoviruses.
Led by Wang, the MIT researchers pursued two different strategies — first, a large-scale computational screen of many possible mutations to the RBD of SARS-CoV-2, and second, an analysis of naturally occurring RBD proteins from zoonotic sarbecoviruses.
For the first approach, the researchers began with the original strain of SARS-CoV-2 and generated sequences of about 800,000 RBD candidates by making substitutions in locations that are known to affect antibody binding to variable portions of the RBD. Then, they screened those candidates for their stability and solubility, to make sure they could withstand attachment to the nanoparticle and injection as a vaccine.
From the remaining candidates, the researchers chose 10 based on how different their variable regions were. They then used these to create mosaic nanoparticles coated with either two or five different RBD proteins (mosaic-2COM and mosaic-5COM).
In their second approach, instead of mutating the RBD sequences, the researchers chose seven naturally occurring RBD proteins, using computational techniques to select RBDs that were different from each other in regions that are variable, but retained their conserved regions. They used these to create another vaccine, mosaic-7COM.
Once the researchers produced the RBD-nanoparticles, they evaluated each one in mice. After each mouse received three doses of one of the vaccines, the researchers analyzed how well the resulting antibodies bound to and neutralized seven variants of SARS-CoV-2 and four other sarbecoviruses.
They also compared the mosaic nanoparticle vaccines to a nanoparticle with only one type of RBD displayed, and to the original mosaic-8 particle from their 2021, 2022, and 2024 studies. They found that mosaic-2COM and mosaic-5COM outperformed both of those vaccines, and mosaic-7COM showed the best responses of all. Mosaic-7COM elicited antibodies with binding to most of the viruses tested, and these antibodies were also able to prevent the viruses from entering cells.
The researchers saw similar results when they tested the new vaccines in mice that were previously vaccinated with a bivalent mRNA COVID-19 vaccine.
“We wanted to simulate the fact that people have already been infected and/or vaccinated against SARS-CoV-2,” Wang says. “In pre-vaccinated mice, mosaic-7COM is consistently giving the highest binding titers for both SARS-CoV-2 variants and other sarbecoviruses.”
Bjorkman’s lab has received funding from the Coalition for Epidemic Preparedness Innovations to do a clinical trial of the mosaic-8 RBD-nanoparticle. They also hope to move mosaic-7COM, which performed better in the current study, into clinical trials. The researchers plan to work on redesigning the vaccines so that they could be delivered as mRNA, which would make them easier to manufacture.
The research was funded by a National Science Foundation Graduate Research Fellowship, the National Institutes of Health, Wellcome Leap, the Bill and Melinda Gates Foundation, the Coalition for Epidemic Preparedness Innovations, and the Caltech Merkin Institute for Translational Research.
Toward video generative models of the molecular world
As the capabilities of generative AI models have grown, you've probably seen how they can transform simple text prompts into hyperrealistic images and even extended video clips.
More recently, generative AI has shown potential in helping chemists and biologists explore static molecules, like proteins and DNA. Models like AlphaFold can predict molecular structures to accelerate drug discovery, and the MIT-assisted “RFdiffusion,” for example, can help design new proteins. One challenge, though, is that molecules are constantly moving and jiggling, which is important to model when constructing new proteins and drugs. Simulating these motions on a computer using physics — a technique known as molecular dynamics — can be very expensive, requiring billions of time steps on supercomputers.
As a step toward simulating these behaviors more efficiently, MIT Computer Science and Artificial Intelligence Laboratory (CSAIL) and Department of Mathematics researchers have developed a generative model that learns from prior data. The team’s system, called MDGen, can take a frame of a 3D molecule and simulate what will happen next like a video, connect separate stills, and even fill in missing frames. By hitting the “play button” on molecules, the tool could potentially help chemists design new molecules and closely study how well their drug prototypes for cancer and other diseases would interact with the molecular structure it intends to impact.
Co-lead author Bowen Jing SM ’22 says that MDGen is an early proof of concept, but it suggests the beginning of an exciting new research direction. “Early on, generative AI models produced somewhat simple videos, like a person blinking or a dog wagging its tail,” says Jing, a PhD student at CSAIL. “Fast forward a few years, and now we have amazing models like Sora or Veo that can be useful in all sorts of interesting ways. We hope to instill a similar vision for the molecular world, where dynamics trajectories are the videos. For example, you can give the model the first and 10th frame, and it’ll animate what’s in between, or it can remove noise from a molecular video and guess what was hidden.”
The researchers say that MDGen represents a paradigm shift from previous comparable works with generative AI in a way that enables much broader use cases. Previous approaches were “autoregressive,” meaning they relied on the previous still frame to build the next, starting from the very first frame to create a video sequence. In contrast, MDGen generates the frames in parallel with diffusion. This means MDGen can be used to, for example, connect frames at the endpoints, or “upsample” a low frame-rate trajectory in addition to pressing play on the initial frame.
This work was presented in a paper shown at the Conference on Neural Information Processing Systems (NeurIPS) this past December. Last summer, it was awarded for its potential commercial impact at the International Conference on Machine Learning’s ML4LMS Workshop.
Some small steps forward for molecular dynamics
In experiments, Jing and his colleagues found that MDGen’s simulations were similar to running the physical simulations directly, while producing trajectories 10 to 100 times faster.
The team first tested their model’s ability to take in a 3D frame of a molecule and generate the next 100 nanoseconds. Their system pieced together successive 10-nanosecond blocks for these generations to reach that duration. The team found that MDGen was able to compete with the accuracy of a baseline model, while completing the video generation process in roughly a minute — a mere fraction of the three hours that it took the baseline model to simulate the same dynamic.
When given the first and last frame of a one-nanosecond sequence, MDGen also modeled the steps in between. The researchers’ system demonstrated a degree of realism in over 100,000 different predictions: It simulated more likely molecular trajectories than its baselines on clips shorter than 100 nanoseconds. In these tests, MDGen also indicated an ability to generalize on peptides it hadn’t seen before.
MDGen’s capabilities also include simulating frames within frames, “upsampling” the steps between each nanosecond to capture faster molecular phenomena more adequately. It can even “inpaint” structures of molecules, restoring information about them that was removed. These features could eventually be used by researchers to design proteins based on a specification of how different parts of the molecule should move.
Toying around with protein dynamics
Jing and co-lead author Hannes Stärk say that MDGen is an early sign of progress toward generating molecular dynamics more efficiently. Still, they lack the data to make these models immediately impactful in designing drugs or molecules that induce the movements chemists will want to see in a target structure.
The researchers aim to scale MDGen from modeling molecules to predicting how proteins will change over time. “Currently, we’re using toy systems,” says Stärk, also a PhD student at CSAIL. “To enhance MDGen’s predictive capabilities to model proteins, we’ll need to build on the current architecture and data available. We don’t have a YouTube-scale repository for those types of simulations yet, so we’re hoping to develop a separate machine-learning method that can speed up the data collection process for our model.”
For now, MDGen presents an encouraging path forward in modeling molecular changes invisible to the naked eye. Chemists could also use these simulations to delve deeper into the behavior of medicine prototypes for diseases like cancer or tuberculosis.
“Machine learning methods that learn from physical simulation represent a burgeoning new frontier in AI for science,” says Bonnie Berger, MIT Simons Professor of Mathematics, CSAIL principal investigator, and senior author on the paper. “MDGen is a versatile, multipurpose modeling framework that connects these two domains, and we’re very excited to share our early models in this direction.”
“Sampling realistic transition paths between molecular states is a major challenge,” says fellow senior author Tommi Jaakkola, who is the MIT Thomas Siebel Professor of electrical engineering and computer science and the Institute for Data, Systems, and Society, and a CSAIL principal investigator. “This early work shows how we might begin to address such challenges by shifting generative modeling to full simulation runs.”
Researchers across the field of bioinformatics have heralded this system for its ability to simulate molecular transformations. “MDGen models molecular dynamics simulations as a joint distribution of structural embeddings, capturing molecular movements between discrete time steps,” says Chalmers University of Technology associate professor Simon Olsson, who wasn’t involved in the research. “Leveraging a masked learning objective, MDGen enables innovative use cases such as transition path sampling, drawing analogies to inpainting trajectories connecting metastable phases.”
The researchers’ work on MDGen was supported, in part, by the National Institute of General Medical Sciences, the U.S. Department of Energy, the National Science Foundation, the Machine Learning for Pharmaceutical Discovery and Synthesis Consortium, the Abdul Latif Jameel Clinic for Machine Learning in Health, the Defense Threat Reduction Agency, and the Defense Advanced Research Projects Agency.
Physicists discover — and explain — unexpected magnetism in an atomically thin material
MIT physicists have created a new ultrathin, two-dimensional material with unusual magnetic properties that initially surprised the researchers before they went on to solve the complicated puzzle behind those properties’ emergence. As a result, the work introduces a new platform for studying how materials behave at the most fundamental level — the world of quantum physics.
Ultrathin materials made of a single layer of atoms have riveted scientists’ attention since the discovery of the first such material — graphene, composed of carbon — about 20 years ago. Among other advances since then, researchers have found that stacking individual sheets of the 2D materials, and sometimes twisting them at a slight angle to each other, can give them new properties, from superconductivity to magnetism. Enter the field of twistronics, which was pioneered at MIT by Pablo Jarillo-Herrero, the Cecil and Ida Green Professor of Physics at MIT.
In the current research, reported in the Jan. 7 issue of Nature Physics, the scientists, led by Jarillo-Herrero, worked with three layers of graphene. Each layer was twisted on top of the next at the same angle, creating a helical structure akin to the DNA helix or a hand of three cards that are fanned apart.
“Helicity is a fundamental concept in science, from basic physics to chemistry and molecular biology. With 2D materials, one can create special helical structures, with novel properties which we are just beginning to understand. This work represents a new twist in the field of twistronics, and the community is very excited to see what else we can discover using this helical materials platform!” says Jarillo-Herrero, who is also affiliated with MIT’s Materials Research Laboratory.
Do the twist
Twistronics can lead to new properties in ultrathin materials because arranging sheets of 2D materials in this way results in a unique pattern called a moiré lattice. And a moiré pattern, in turn, has an impact on the behavior of electrons.
“It changes the spectrum of energy levels available to the electrons and can provide the conditions for interesting phenomena to arise,” says Sergio C. de la Barrera, one of three co-first authors of the recent paper. De la Barrera, who conducted the work while a postdoc at MIT, is now an assistant professor at the University of Toronto.
In the current work, the helical structure created by the three graphene layers forms two moiré lattices. One is created by the first two overlapping sheets; the other is formed between the second and third sheets.
The two moiré patterns together form a third moiré, a supermoiré, or “moiré of a moiré,” says Li-Qiao Xia, a graduate student in MIT physics and another of the three co-first authors of the Nature Physics paper. “It’s like a moiré hierarchy.” While the first two moiré patterns are only nanometers, or billionths of a meter, in scale, the supermoiré appears at a scale of hundreds of nanometers superimposed over the other two. You can only see it if you zoom out to get a much wider view of the system.
A major surprise
The physicists expected to observe signatures of this moiré hierarchy. They got a huge surprise, however, when they applied and varied a magnetic field. The system responded with an experimental signature for magnetism, one that arises from the motion of electrons. In fact, this orbital magnetism persisted to -263 degrees Celsius — the highest temperature reported in carbon-based materials to date.
But that magnetism can only occur in a system that lacks a specific symmetry — one that the team’s new material should have had. “So the fact that we saw this was very puzzling. We didn’t really understand what was going on,” says Aviram Uri, an MIT Pappalardo postdoc in physics and the third co-first author of the new paper.
Other authors of the paper include MIT professor of physics Liang Fu; Aaron Sharpe of Sandia National Laboratories; Yves H. Kwan of Princeton University; Ziyan Zhu, David Goldhaber-Gordon, and Trithep Devakul of Stanford University; and Kenji Watanabe and Takashi Taniguchi of the National Institute for Materials Science in Japan.
What was happening?
It turns out that the new system did indeed break the symmetry that prohibits the orbital magnetism the team observed, but in a very unusual way. “What happens is that the atoms in this system aren’t very comfortable, so they move in a subtle orchestrated way that we call lattice relaxation,” says Xia. And the new structure formed by that relaxation does indeed break the symmetry locally, on the moiré length scale.
This opens the possibility for the orbital magnetism the team observed. However, if you zoom out to view the system on the supermoiré scale, the symmetry is restored. “The moiré hierarchy turns out to support interesting phenomena at different length scales,” says de la Barrera.
Concludes Uri: “It’s a lot of fun when you solve a riddle and it’s such an elegant solution. We’ve gained new insights into how electrons behave in these complex systems, insights that we couldn’t have had unless our experimental observations forced to think about these things.”
This work was supported by the Army Research Office, the National Science Foundation, the Gordon and Betty Moore Foundation, the Ross M. Brown Family Foundation, an MIT Pappalardo Fellowship, the VATAT Outstanding Postdoctoral Fellowship in Quantum Science and Technology, the JSPS KAKENHI, and a Stanford Science Fellowship.
New START.nano cohort is developing solutions in health, data storage, power, and sustainable energy
MIT.nano has announced seven new companies to join START.nano, a program aimed at speeding the transition of hard-tech innovation to market. The program supports new ventures through discounted use of MIT.nano’s facilities and access to the MIT innovation ecosystem.
The advancements pursued by the newly engages startups include wearables for health care, green alternatives to fossil fuel-based energy, novel battery technologies, enhancements in data systems, and interconnecting nanofabrication knowledge networks, among others.
“The transition of the grand idea that is imagined in the laboratory to something that a million people can use in their hands is a journey fraught with many challenges,” MIT.nano Director Vladimir Bulović said at the 2024 Nano Summit, where nine START.nano companies presented their work. The program provides resources to ease startups over the first two hurdles — finding stakeholders and building a well-developed prototype.
In addition to access to laboratory tools necessary to advance their technologies, START.nano companies receive advice from MIT.nano expert staff, are connected to MIT.nano Consortium companies, gain a broader exposure at MIT conferences and community events, and are eligible to join the MIT Startup Exchange.
“MIT.nano has allowed us to push our project to the frontiers of sensing by implementing advanced fabrication techniques using their machinery,” said Uroš Kuzmanović, CEO and founder of Biosens8. “START.nano has surrounded us with exciting peers, a strong support system, and a spotlight to present our work. By taking advantage of all that the program has to offer, BioSens8 is moving faster than we could anywhere else.”
Here are the seven new START.nano participants:
Analog Photonics is developing lidar and optical communications technology using silicon photonics.
Biosens8 is engineering novel devices to enable health ownership. Their research focuses on multiplexed wearables for hormones, neurotransmitters, organ health markers, and drug use that will give insight into the body's health state, opening the door to personalized medicine and proactive, data-driven health decisions.
Casimir, Inc. is working on power-generating nanotechnology that interacts with quantum fields to create a continuous source of power. The team compares their technology to a solar panel that works in the dark or a battery that never needs to be recharged.
Central Spiral focuses on lossless data compression. Their technology allows for the compression of any type of data, including those that are already compressed, reducing data storage and transmission costs, lowering carbon dioxide emissions, and enhancing efficiency.
FabuBlox connects stakeholders across the nanofabrication ecosystem and resolves issues of scattered, unorganized, and isolated fab knowledge. Their cloud-based platform combines a generative process design and simulation interface with GitHub-like repository building capabilities.
Metal Fuels is converting industrial waste aluminum to onsite energy and high-value aluminum/aluminum-oxide powders. Their approach combines existing mature technologies of molten metal purification and water atomization to develop a self-sustaining reactor that produces alumina of higher value than our input scrap aluminum feedstock, while also collecting the hydrogen off-gas.
PolyJoule, Inc. is an energy storage startup working on conductive polymer battery technology. The team’s goal is a grid battery of the future that is ultra-safe, sustainable, long living, and low-cost.
In addition to the seven startups that are actively using MIT.nano, nine other companies have been invited to join the latest START.nano cohort:
- Acorn Genetics
- American Boronite Corp.
- Copernic Catalysts
- Envoya Bio
- Helix Carbon
- Minerali
- Plaid Semiconductors
- Quantum Network Technologies
- Wober Tech
Launched in 2021, START.nano now comprises over 20 companies and eight graduates — ventures that have moved beyond the initial startup stages and some into commercialization.
Steven Strang, literary scholar and leader in writing and communication support at MIT, dies at 77
Steven Strang, a writer and literary scholar who founded MIT’s Writing and Communication Center in 1981 and directed it for 40 years, died with family at his side on Dec. 29, 2024. He was 77.
His vision for the center was ambitious. After an MIT working group identified gaps between the students’ technical knowledge and their ability to communicate it — particularly once in positions of leadership — Strang advocated an even broader approach rarely used at other universities. Rather than student-tutors working with peers, Strang hired instructors with doctorates, subject matter expertise, and teaching experience to help train all MIT community members for the current and future careers becoming increasingly reliant on persuasion and the need to communicate with varied audiences.
“He made an indelible mark on the MIT community,” wrote current director Elena Kallestinova in a message to WCC staff soon after Strang’s death. “He was deeply respected as a leader, educator, mentor, and colleague.”
Beginning his professional life as a journalist with the Bangor Daily News, Strang soon shifted to academia, receiving a PhD in English from Brown University and over the decades publishing countless pieces of fiction, poetry, and criticism, in addition to his pedagogical articles on writing and rhetoric.
But the Writing and Communication Center is his legacy. At his Jan. 11 memorial, longtime MIT lecturer and colleague Thalia Rubio called the WCC “Steve’s creation,” pointing out that it went on to serve many thousands of students and others. Another colleague, Bob Irwin, described in a note Strang’s commitment to making the WCC “a place that offered both friendliness and the highest professional standards of advice and consultation on all communication tasks and issues. Steve himself was conscientious, a respectful director, and a warm and reliable mentor to me and others. I think he was exemplary in his job.”
MIT recognized Strang’s major contributions with a Levitan Teaching Award, an Infinite Mile Award, and an Excellence Award. In nomination letters and testimonials, students and peers alike told of a “tireless commitment,” that “they might not have graduated, or been hired to the job they have today, or gained admittance to graduate school had it not been for the help of The Writing Center.”
Strang is also remembered for his work founding the MIT Writers Group, which he first offered as a creative writing workshop for Independent Activities Period in 2002. In yet another example of Strang recognizing and meeting a community need, about 70 people from across the Institute showed up that first year.
Strang is survived by a large extended family, including his wife Ayni and her two children, Elly and Marta, whom Strang adopted as his own. Donations in his memory can be made to The Rhode Island Society for the Prevention of Cruelty to Animals.
New general law governs fracture energy of networks across materials and length scales
Materials like car tires, human tissues, and spider webs are diverse in composition, but all contain networks of interconnected strands. A long-standing question about the durability of these materials asks: What is the energy required to fracture these diverse networks? A recently published paper by MIT researchers offers new insights.
“Our findings reveal a simple, general law that governs the fracture energy of networks across various materials and length scales,” says Xuanhe Zhao, the Uncas and Helen Whitaker Professor and professor of mechanical engineering and civil and environmental engineering at MIT. “This discovery has significant implications for the design of new materials, structures, and metamaterials, allowing for the creation of systems that are incredibly tough, soft, and stretchable.”
Despite an established understanding of the importance of failure resistance in design of such networks, no existing physical model effectively linked strand mechanics and connectivity to predict bulk fracture — until now. This new research reveals a universal scaling law that bridges length scales and makes it possible to predict the intrinsic fracture energy of diverse networks.
“This theory helps us predict how much energy it takes to break these networks by advancing a crack,” says graduate student Chase Hartquist, one of the paper’s lead authors. “It turns out that you can design tougher versions of these materials by making the strands longer, more stretchable, or resistant to higher forces before breaking.”
To validate their results, the team 3D-printed a giant, stretchable network, allowing them to demonstrate fracture properties in practice. They found that despite the differences in the networks, they all followed a simple and predictable rule. Beyond the changes to the strands themselves, a network can also be toughened by connecting the strands into larger loops.
“By adjusting these properties, car tires could last longer, tissues could better resist injury, and spider webs could become more durable,” says Hartquist.
Shu Wang, a postdoc in Zhao’s lab and fellow lead author of the paper, called the research findings “an extremely fulfilling moment ... it meant that the same rules could be applied to describe a wide variety of materials, making it easier to design the best material for a given situation.”
The researchers explain that this work represents progress in an exciting and emerging field called “architected materials,” where the structure within the material itself gives it unique properties. They say the discovery sheds light on how to make these materials even tougher, by focusing on designing the segments within the architecture stronger and more stretchable. The strategy is adaptable for materials across fields and can be applied to improve durability of soft robotic actuators, enhance the toughness of engineered tissues, or even create resilient lattices for aerospace technology.
Their open-access paper, “Scaling Law for Intrinsic Fracture Energy of Diverse Stretchable Networks,” is available now in Physical Review X, a leading journal in interdisciplinary physics.
“Forever grateful for MIT Open Learning for making knowledge accessible and fostering a network of curious minds”
Bia Adams, a London-based neuropsychologist, former professional ballet dancer, and MIT Open Learning learner, has built her career across decades of diverse, interconnected experiences and an emphasis on lifelong learning. She earned her bachelor’s degree in clinical and behavioral psychology, and then worked as a psychologist and therapist for several years before taking a sabbatical in her late 20s to study at the London Contemporary Dance School and The Royal Ballet — fulfilling a long-time dream.
“In hindsight, I think what drew me most to ballet was not so much the form itself,” says Adams, “but more of a subconscious desire to make sense of my body moving through space and time, my emotions and motivations — all within a discipline that is rigorous, meticulous, and routine-based. It’s an endeavor to make sense of the world and myself.”
After acquiring some dance-related injuries, Adams returned to psychology. She completed an online certificate program specializing in medical neuroscience via Duke University, focusing on how pathology arises out of the way the brain computes information and generates behavior.
In addition to her clinical practice, she has also worked at a data science and AI consultancy for neural network research.
In 2022, in search of new things to learn and apply to both her work and personal life, Adams discovered MIT OpenCourseWare within MIT Open Learning. She was drawn to class 8.04 (Quantum Physics I), which specifically focuses on quantum mechanics, as she was hoping to finally gain some understanding of complex topics that she had tried to teach herself in the past with limited success. She credits the course’s lectures, taught by Allan Adams (physicist and principal investigator of the MIT Future Ocean Lab), with finally making these challenging topics approachable.
“I still talk to my friends at length about exciting moments in these lectures,” says Adams. “After the first class, I was hooked.”
Adams’s journey through MIT Open Learning’s educational resources quickly led to a deeper interest in computational neuroscience. She learned how to use tools from mathematics and computer science to better understand the brain, nervous system, and behavior.
She says she gained many new insights from class 6.034 (Artificial Intelligence), particularly in watching the late Professor Patrick Winston’s lectures. She appreciated learning more about the cognitive psychology aspect of AI, including how pioneers in the field looked at how the brain processes information and aimed to build programs that could solve problems. She further enhanced her understanding of AI with the Minds and Machines course on MITx Online, part of Open Learning.
Adams is now in the process of completing Introduction to Computer Science and Programming Using Python, taught by John Guttag; Eric Grimson, former interim vice president for Open Learning; and Ana Bell.
“I am multilingual, and I think the way my brain processes code is similar to the way computers code,” says Adams. “I find learning to code similar to learning a foreign language: both exhilarating and intimidating. Learning the rules, deciphering the syntax, and building my own world through code is one of the most fascinating challenges of my life.”
Adams is also pursuing a master’s degree at Duke and the University College of London, focusing on the neurobiology of sleep and looking particularly at how the biochemistry of the brain can affect this critical function. As a complement to this research, she is currently exploring class 9.40 (Introduction to Neural Computation), taught by Michale Fee and Daniel Zysman, which introduces quantitative approaches to understanding brain and cognitive functions and neurons and covers foundational quantitative tools of data analysis in neuroscience.
In addition to the courses related more directly to her field, MIT Open Learning also provided Adams an opportunity to explore other academic areas. She delved into philosophy for the first time, taking Paradox and Infinity, taught by Professor Agustín Rayo, the Kenan Sahin Dean of the MIT School of Humanities, Arts, and Social Sciences, and Digital Learning Lab Fellow David Balcarras, which looks at the intersection of philosophy and mathematics. She also was able to explore in more depth immunology, which had always been of great interest to her, through Professor Adam Martin’s lectures on this topic in class 7.016 (Introductory Biology).
“I am forever grateful for MIT Open Learning,” says Adams, “for making knowledge accessible and fostering a network of curious minds, all striving to share, expand, and apply this knowledge for the greater good.”
For MIT-WHOI Joint Program student Faith Brooks, the sky’s the limit
Faith Brooks, a graduate student in the MIT-WHOI Joint Program, has had a clear dream since the age of 4: to become a pilot.
“At around 8 years old, my neighbor knew I wanted to fly and showed me pictures of her dad landing a jet on an aircraft carrier, and I was immediately captivated,” says Brooks. Further inspired by her grandfather’s experience in the U.S. Navy (USN), and owing to a lifelong fascination with aviation, she knew nothing would stand in her way.
Brooks explored several different paths to becoming a pilot, but she says one conversation with her longtime mentor, Capt. Matt Skone, USN (Ret.), changed the trajectory of her life.
“He asked if I had heard of the Naval Academy,” she recalls. “At the time, I hadn’t … I immediately knew that that was where I wanted to go, and everything else I learned about United States Naval Academy (USNA) reinforced that for me.”
In her “firstie” (senior) year at the USNA, Brooks was selected to go to Pensacola, Florida, and train to become a naval pilot as a student naval aviator, taking her one step closer to her dream. The USNA also helped guide her path to MIT. Her journey to joining the MIT-WHOI Joint Program began with the USNA’s professional knowledge curriculum, where she read about retired Capt. Wendy Lawrence SM ’88, a naval aviator and astronaut.
“Reading her bio prompted me to look into the program, and it sounded like the perfect program for me — where else could you get a better education in ocean engineering than MIT and Woods Hole Oceanographic Institution [WHOI]?”
In the MIT-WHOI Joint Program, Brooks is researching the impact of coastal pond breaching on preventing and mitigating harmful algal blooms. Her work focuses on the biannual mechanical breaching of Nantucket’s Sesachacha Pond to the ocean and the resultant impact on the pond’s water quality. This practice aims to improve water quality and mitigate harmful algal blooms (HABs), especially in summer.
Breaching in coastal ponds is a process that was initially used to enhance salinity for herring and shellfish habitats, but has since shifted to address water quality concerns. Traditionally, an excavator creates a breach in the pond, which naturally closes within one to five days, influenced by sediment transport and weather conditions. High winds and waves can accelerate sediment movement, limiting ocean water exchange and potentially increasing eutrophication, where excessive nutrients lead to dense plant growth and depletion of oxygen. In brackish water environments, harmful algal blooms are often driven by elevated nitrogen levels and higher temperatures, with higher nitrogen concentrating leading to more frequent and severe blooms as temperatures rise.
The Nantucket Natural Resources Department (NRD) has been collaborating with local homeowners to investigate the pond breaching process. Existing data are mainly anecdotal evidence and NRD’s monthly sampling since 2022, which has not shown the expected decrease in eutrophication. Brooks’ research will focus on data before, during, and after the breach at two pond sites to assess water changes to evaluate its effectiveness in improving water quality.
When Brooks isn’t knee-deep in the waters of the Sesachacha or training with her MIT Triathlon team, she takes additional opportunities to further her education. Last year, Brooks participated in the MIT-Portugal Marine Robotics Summer School in Faial, Azores, in Portugal, and immersed herself in a combination of a hands-on design projects and lectures on a variety of topics related to oceanography, engineering, and marine robotics.
“My favorite part of the program was how interdisciplinary it was. We had a combination of mechanical engineers, electrical engineers, computer scientists, marine biologists, and oceanographers, and we had teams that included each of these specialties,” she says. “Our project involved designing a lander equipped with an underwater camera connected to a surface buoy that would transmit the footage. Having worked in mostly just engineering teams previously, it was a great experience to work with a more diverse group and I gained a much better understanding of how to design instruments and systems in accordance with what the marine biologists need.”
Brooks also earned her Part 107 Small Unmanned Aircraft System (UAS) license to operate the lab’s drone with a multispectral camera for her upcoming fieldwork. When she graduates from the MIT-WHOI Joint Program next September, she’ll report to the Naval Aviation Schools Command in Pensacola, Florida, to begin flight training.
While she says she’ll miss Boston’s charm and history, as well as the Shining Sea Bikeway on crisp fall days in Woods Hole, Brooks is looking forward to putting her uniform back on, and starting her naval career and flight school. The time Brooks has spent at MIT will support her in these future endeavors. She advises others interested in a similar path to focus on research within their areas of interest.
“The biggest lesson that I’ve learned from both research theses is that any research project will change over time, and it’s often a good idea to take a step back and look at how your work fits into the larger picture,” she says. “I couldn’t recommend doing research more; it’s such a great opportunity to dig into something that you’re interested in, and is also very fulfilling.”
Toward sustainable decarbonization of aviation in Latin America
According to the International Energy Agency, aviation accounts for about 2 percent of global carbon dioxide emissions, and aviation emissions are expected to double by mid-century as demand for domestic and international air travel rises. To sharply reduce emissions in alignment with the Paris Agreement’s long-term goal to keep global warming below 1.5 degrees Celsius, the International Air Transport Association (IATA) has set a goal to achieve net-zero carbon emissions by 2050. Which raises the question: Are there technologically feasible and economically viable strategies to reach that goal within the next 25 years?
To begin to address that question, a team of researchers at the MIT Center for Sustainability Science and Strategy (CS3) and the MIT Laboratory for Aviation and the Environment has spent the past year analyzing aviation decarbonization options in Latin America, where air travel is expected to more than triple by 2050 and thereby double today’s aviation-related emissions in the region.
Chief among those options is the development and deployment of sustainable aviation fuel. Currently produced from low- and zero-carbon sources (feedstock) including municipal waste and non-food crops, and requiring practically no alteration of aircraft systems or refueling infrastructure, sustainable aviation fuel (SAF) has the potential to perform just as well as petroleum-based jet fuel with as low as 20 percent of its carbon footprint.
Focused on Brazil, Chile, Colombia, Ecuador, Mexico and Peru, the researchers assessed SAF feedstock availability, the costs of corresponding SAF pathways, and how SAF deployment would likely impact fuel use, prices, emissions, and aviation demand in each country. They also explored how efficiency improvements and market-based mechanisms could help the region to reach decarbonization targets. The team’s findings appear in a CS3 Special Report.
SAF emissions, costs, and sources
Under an ambitious emissions mitigation scenario designed to cap global warming at 1.5 C and raise the rate of SAF use in Latin America to 65 percent by 2050, the researchers projected aviation emissions to be reduced by about 60 percent in 2050 compared to a scenario in which existing climate policies are not strengthened. To achieve net-zero emissions by 2050, other measures would be required, such as improvements in operational and air traffic efficiencies, airplane fleet renewal, alternative forms of propulsion, and carbon offsets and removals.
As of 2024, jet fuel prices in Latin America are around $0.70 per liter. Based on the current availability of feedstocks, the researchers projected SAF costs within the six countries studied to range from $1.11 to $2.86 per liter. They cautioned that increased fuel prices could affect operating costs of the aviation sector and overall aviation demand unless strategies to manage price increases are implemented.
Under the 1.5 C scenario, the total cumulative capital investments required to build new SAF producing plants between 2025 and 2050 were estimated at $204 billion for the six countries (ranging from $5 billion in Ecuador to $84 billion in Brazil). The researchers identified sugarcane- and corn-based ethanol-to-jet fuel, palm oil- and soybean-based hydro-processed esters and fatty acids as the most promising feedstock sources in the near term for SAF production in Latin America.
“Our findings show that SAF offers a significant decarbonization pathway, which must be combined with an economy-wide emissions mitigation policy that uses market-based mechanisms to offset the remaining emissions,” says Sergey Paltsev, lead author of the report, MIT CS3 deputy director, and senior research scientist at the MIT Energy Initiative.
Recommendations
The researchers concluded the report with recommendations for national policymakers and aviation industry leaders in Latin America.
They stressed that government policy and regulatory mechanisms will be needed to create sufficient conditions to attract SAF investments in the region and make SAF commercially viable as the aviation industry decarbonizes operations. Without appropriate policy frameworks, SAF requirements will affect the cost of air travel. For fuel producers, stable, long-term-oriented policies and regulations will be needed to create robust supply chains, build demand for establishing economies of scale, and develop innovative pathways for producing SAF.
Finally, the research team recommended a region-wide collaboration in designing SAF policies. A unified decarbonization strategy among all countries in the region will help ensure competitiveness, economies of scale, and achievement of long-term carbon emissions-reduction goals.
“Regional feedstock availability and costs make Latin America a potential major player in SAF production,” says Angelo Gurgel, a principal research scientist at MIT CS3 and co-author of the study. “SAF requirements, combined with government support mechanisms, will ensure sustainable decarbonization while enhancing the region’s connectivity and the ability of disadvantaged communities to access air transport.”
Financial support for this study was provided by LATAM Airlines and Airbus.
Bryan Reimer named to FAA Rulemaking Committee
Bryan Reimer, a research scientist at the MIT Center for Transportation and Logistics (CTL), and the founder and co-leader of the Advanced Vehicle Technology Consortium and the Human Factors Evaluator for Automotive Demand Consortium in the MIT AgeLab, has been appointed to the Task Force on Human Factors in Aviation Safety Aviation Rulemaking Committee (HF Task Force ARC). The HF Task Force ARC will provide recommendations to the U.S. Federal Aviation Administration (FAA) on the most significant human factors and the relative contribution of these factors to aviation safety risk.
Reimer, who has worked at MIT since 2003, joins a committee whose operational or academic expertise includes air carrier operations, air traffic control, pilot experience, aeronautical information, aircraft maintenance and mechanics psychology, human-machine integration, and general aviation operations. Their recommendations to the FAA will help ensure safety for passengers, aircraft crews, and cargo for years to come. His appointment follows a year of serving on the Transforming Transportation Advisory Committee (TTAC) for the U.S. Department of Transportation, where he has taken on the role of vice chair on the Artificial Intelligence subcommittee. The TTAC recently released a report to the Secretary of Transportation in response to its charter.
As a mobility and technology futurist working at the intersection of technology, human behavior, and public policy, Reimer brings his expertise in human-machine integration, transportation safety, and AI to the committee. The committee, chartered by congressional mandate through the bipartisan FAA Reauthorization Act of 2024, specifically calls for a portion of the committee to have expertise on human factors but whose experience and training are not primarily in aviation, which Reimer will provide.
MIT CTL creates supply chain innovation and drives it into practice through the three pillars of research, outreach, and education, working with businesses, government, and nongovernmental organizations. As a longtime advocate of collaboration across public and private sectors to ensure consumers’ safety in transportation, Reimer’s particular expertise will help the FAA more broadly consider the human element of aviation safety. Yossi Sheffi, director of MIT CTL, says, “Aviation plays a critical role in the rapid and reliable transportation of goods across vast distances, making it essential for delivering time-sensitive products globally. We must understand the current human factors involved in this process to help ensure smooth operation of this indispensable service amid potential disruptions.”
Reimer recently discussed his research on an episode of The Ojo-Yoshida Report with Phil Koopman, a professor of electrical and computer engineering.
HF Task Force ARC members will serve a two-year term. The first ARC plenary meeting was held Jan. 15-16 in Washington.
The multifaceted challenge of powering AI
Artificial intelligence has become vital in business and financial dealings, medical care, technology development, research, and much more. Without realizing it, consumers rely on AI when they stream a video, do online banking, or perform an online search. Behind these capabilities are more than 10,000 data centers globally, each one a huge warehouse containing thousands of computer servers and other infrastructure for storing, managing, and processing data. There are now over 5,000 data centers in the United States, and new ones are being built every day — in the U.S. and worldwide. Often dozens are clustered together right near where people live, attracted by policies that provide tax breaks and other incentives, and by what looks like abundant electricity.
And data centers do consume huge amounts of electricity. U.S. data centers consumed more than 4 percent of the country’s total electricity in 2023, and by 2030 that fraction could rise to 9 percent, according to the Electric Power Research Institute. A single large data center can consume as much electricity as 50,000 homes.
The sudden need for so many data centers presents a massive challenge to the technology and energy industries, government policymakers, and everyday consumers. Research scientists and faculty members at the MIT Energy Initiative (MITEI) are exploring multiple facets of this problem — from sourcing power to grid improvement to analytical tools that increase efficiency, and more. Data centers have quickly become the energy issue of our day.
Unexpected demand brings unexpected solutions
Several companies that use data centers to provide cloud computing and data management services are announcing some surprising steps to deliver all that electricity. Proposals include building their own small nuclear plants near their data centers and even restarting one of the undamaged nuclear reactors at Three Mile Island, which has been shuttered since 2019. (A different reactor at that plant partially melted down in 1979, causing the nation’s worst nuclear power accident.) Already the need to power AI is causing delays in the planned shutdown of some coal-fired power plants and raising prices for residential consumers. Meeting the needs of data centers is not only stressing power grids, but also setting back the transition to clean energy needed to stop climate change.
There are many aspects to the data center problem from a power perspective. Here are some that MIT researchers are focusing on, and why they’re important.
An unprecedented surge in the demand for electricity
“In the past, computing was not a significant user of electricity,” says William H. Green, director of MITEI and the Hoyt C. Hottel Professor in the MIT Department of Chemical Engineering. “Electricity was used for running industrial processes and powering household devices such as air conditioners and lights, and more recently for powering heat pumps and charging electric cars. But now all of a sudden, electricity used for computing in general, and by data centers in particular, is becoming a gigantic new demand that no one anticipated.”
Why the lack of foresight? Usually, demand for electric power increases by roughly half-a-percent per year, and utilities bring in new power generators and make other investments as needed to meet the expected new demand. But the data centers now coming online are creating unprecedented leaps in demand that operators didn’t see coming. In addition, the new demand is constant. It’s critical that a data center provides its services all day, every day. There can be no interruptions in processing large datasets, accessing stored data, and running the cooling equipment needed to keep all the packed-together computers churning away without overheating.
Moreover, even if enough electricity is generated, getting it to where it’s needed may be a problem, explains Deepjyoti Deka, a MITEI research scientist. “A grid is a network-wide operation, and the grid operator may have sufficient generation at another location or even elsewhere in the country, but the wires may not have sufficient capacity to carry the electricity to where it’s wanted.” So transmission capacity must be expanded — and, says Deka, that’s a slow process.
Then there’s the “interconnection queue.” Sometimes, adding either a new user (a “load”) or a new generator to an existing grid can cause instabilities or other problems for everyone else already on the grid. In that situation, bringing a new data center online may be delayed. Enough delays can result in new loads or generators having to stand in line and wait for their turn. Right now, much of the interconnection queue is already filled up with new solar and wind projects. The delay is now about five years. Meeting the demand from newly installed data centers while ensuring that the quality of service elsewhere is not hampered is a problem that needs to be addressed.
Finding clean electricity sources
To further complicate the challenge, many companies — including so-called “hyperscalers” such as Google, Microsoft, and Amazon — have made public commitments to having net-zero carbon emissions within the next 10 years. Many have been making strides toward achieving their clean-energy goals by buying “power purchase agreements.” They sign a contract to buy electricity from, say, a solar or wind facility, sometimes providing funding for the facility to be built. But that approach to accessing clean energy has its limits when faced with the extreme electricity demand of a data center.
Meanwhile, soaring power consumption is delaying coal plant closures in many states. There are simply not enough sources of renewable energy to serve both the hyperscalers and the existing users, including individual consumers. As a result, conventional plants fired by fossil fuels such as coal are needed more than ever.
As the hyperscalers look for sources of clean energy for their data centers, one option could be to build their own wind and solar installations. But such facilities would generate electricity only intermittently. Given the need for uninterrupted power, the data center would have to maintain energy storage units, which are expensive. They could instead rely on natural gas or diesel generators for backup power — but those devices would need to be coupled with equipment to capture the carbon emissions, plus a nearby site for permanently disposing of the captured carbon.
Because of such complications, several of the hyperscalers are turning to nuclear power. As Green notes, “Nuclear energy is well matched to the demand of data centers, because nuclear plants can generate lots of power reliably, without interruption.”
In a much-publicized move in September, Microsoft signed a deal to buy power for 20 years after Constellation Energy reopens one of the undamaged reactors at its now-shuttered nuclear plant at Three Mile Island, the site of the much-publicized nuclear accident in 1979. If approved by regulators, Constellation will bring that reactor online by 2028, with Microsoft buying all of the power it produces. Amazon also reached a deal to purchase power produced by another nuclear plant threatened with closure due to financial troubles. And in early December, Meta released a request for proposals to identify nuclear energy developers to help the company meet their AI needs and their sustainability goals.
Other nuclear news focuses on small modular nuclear reactors (SMRs), factory-built, modular power plants that could be installed near data centers, potentially without the cost overruns and delays often experienced in building large plants. Google recently ordered a fleet of SMRs to generate the power needed by its data centers. The first one will be completed by 2030 and the remainder by 2035.
Some hyperscalers are betting on new technologies. For example, Google is pursuing next-generation geothermal projects, and Microsoft has signed a contract to purchase electricity from a startup’s fusion power plant beginning in 2028 — even though the fusion technology hasn’t yet been demonstrated.
Reducing electricity demand
Other approaches to providing sufficient clean electricity focus on making the data center and the operations it houses more energy efficient so as to perform the same computing tasks using less power. Using faster computer chips and optimizing algorithms that use less energy are already helping to reduce the load, and also the heat generated.
Another idea being tried involves shifting computing tasks to times and places where carbon-free energy is available on the grid. Deka explains: “If a task doesn’t have to be completed immediately, but rather by a certain deadline, can it be delayed or moved to a data center elsewhere in the U.S. or overseas where electricity is more abundant, cheaper, and/or cleaner? This approach is known as ‘carbon-aware computing.’” We’re not yet sure whether every task can be moved or delayed easily, says Deka. “If you think of a generative AI-based task, can it easily be separated into small tasks that can be taken to different parts of the country, solved using clean energy, and then be brought back together? What is the cost of doing this kind of division of tasks?”
That approach is, of course, limited by the problem of the interconnection queue. It’s difficult to access clean energy in another region or state. But efforts are under way to ease the regulatory framework to make sure that critical interconnections can be developed more quickly and easily.
What about the neighbors?
A major concern running through all the options for powering data centers is the impact on residential energy consumers. When a data center comes into a neighborhood, there are not only aesthetic concerns but also more practical worries. Will the local electricity service become less reliable? Where will the new transmission lines be located? And who will pay for the new generators, upgrades to existing equipment, and so on? When new manufacturing facilities or industrial plants go into a neighborhood, the downsides are generally offset by the availability of new jobs. Not so with a data center, which may require just a couple dozen employees.
There are standard rules about how maintenance and upgrade costs are shared and allocated. But the situation is totally changed by the presence of a new data center. As a result, utilities now need to rethink their traditional rate structures so as not to place an undue burden on residents to pay for the infrastructure changes needed to host data centers.
MIT’s contributions
At MIT, researchers are thinking about and exploring a range of options for tackling the problem of providing clean power to data centers. For example, they are investigating architectural designs that will use natural ventilation to facilitate cooling, equipment layouts that will permit better airflow and power distribution, and highly energy-efficient air conditioning systems based on novel materials. They are creating new analytical tools for evaluating the impact of data center deployments on the U.S. power system and for finding the most efficient ways to provide the facilities with clean energy. Other work looks at how to match the output of small nuclear reactors to the needs of a data center, and how to speed up the construction of such reactors.
MIT teams also focus on determining the best sources of backup power and long-duration storage, and on developing decision support systems for locating proposed new data centers, taking into account the availability of electric power and water and also regulatory considerations, and even the potential for using what can be significant waste heat, for example, for heating nearby buildings. Technology development projects include designing faster, more efficient computer chips and more energy-efficient computing algorithms.
In addition to providing leadership and funding for many research projects, MITEI is acting as a convenor, bringing together companies and stakeholders to address this issue. At MITEI’s 2024 Annual Research Conference, a panel of representatives from two hyperscalers and two companies that design and construct data centers together discussed their challenges, possible solutions, and where MIT research could be most beneficial.
As data centers continue to be built, and computing continues to create an unprecedented increase in demand for electricity, Green says, scientists and engineers are in a race to provide the ideas, innovations, and technologies that can meet this need, and at the same time continue to advance the transition to a decarbonized energy system.
For clean ammonia, MIT engineers propose going underground
Ammonia is the most widely produced chemical in the world today, used primarily as a source for nitrogen fertilizer. Its production is also a major source of greenhouse gas emissions — the highest in the whole chemical industry.
Now, a team of researchers at MIT has developed an innovative way of making ammonia without the usual fossil-fuel-powered chemical plants that require high heat and pressure. Instead, they have found a way to use the Earth itself as a geochemical reactor, producing ammonia underground. The processes uses Earth’s naturally occurring heat and pressure, provided free of charge and free of emissions, as well as the reactivity of minerals already present in the ground.
The trick the team devised is to inject water underground, into an area of iron-rich subsurface rock. The water carries with it a source of nitrogen and particles of a metal catalyst, allowing the water to react with the iron to generate clean hydrogen, which in turn reacts with the nitrogen to make ammonia. A second well is then used to pump that ammonia up to the surface.
The process, which has been demonstrated in the lab but not yet in a natural setting, is described today in the journal Joule. The paper’s co-authors are MIT professors of materials science and engineering Iwnetim Abate and Ju Li, graduate student Yifan Gao, and five others at MIT.
“When I first produced ammonia from rock in the lab, I was so excited,” Gao recalls. “I realized this represented an entirely new and never-reported approach to ammonia synthesis.’”
The standard method for making ammonia is called the Haber-Bosch process, which was developed in Germany in the early 20th century to replace natural sources of nitrogen fertilizer such as mined deposits of bat guano, which were becoming depleted. But the Haber-Bosch process is very energy intensive: It requires temperatures of 400 degrees Celsius and pressures of 200 atmospheres, and this means it needs huge installations in order to be efficient. Some areas of the world, such as sub-Saharan Africa and Southeast Asia, have few or no such plants in operation. As a result, the shortage or extremely high cost of fertilizer in these regions has limited their agricultural production.
The Haber-Bosch process “is good. It works,” Abate says. “Without it, we wouldn’t have been able to feed 2 out of the total 8 billion people in the world right now, he says, referring to the portion of the world’s population whose food is grown with ammonia-based fertilizers. But because of the emissions and energy demands, a better process is needed, he says.
Burning fuel to generate heat is responsible for about 20 percent of the greenhouse gases emitted from plants using the Haber-Bosch process. Making hydrogen accounts for the remaining 80 percent. But ammonia, the molecule NH3, is made up only of nitrogen and hydrogen. There’s no carbon in the formula, so where do the carbon emissions come from? The standard way of producing the needed hydrogen is by processing methane gas with steam, breaking down the gas into pure hydrogen, which gets used, and carbon dioxide gas that gets released into the air.
Other processes exist for making low- or no-emissions hydrogen, such as by using solar or wind-generated electricity to split water into oxygen and hydrogen, but that process can be expensive. That’s why Abate and his team worked on developing a system to produce what they call geological hydrogen. Some places in the world, including some in Africa, have been found to naturally generate hydrogen underground through chemical reactions between water and iron-rich rocks. These pockets of naturally occurring hydrogen can be mined, just like natural methane reservoirs, but the extent and locations of such deposits are still relatively unexplored.
Abate realized this process could be created or enhanced by pumping water, laced with copper and nickel catalyst particles to speed up the process, into the ground in places where such iron-rich rocks were already present. “We can use the Earth as a factory to produce clean flows of hydrogen,” he says.
He recalls thinking about the problem of the emissions from hydrogen production for ammonia: “The ‘aha!’ moment for me was thinking, how about we link this process of geological hydrogen production with the process of making Haber-Bosch ammonia?”
That would solve the biggest problem of the underground hydrogen production process, which is how to capture and store the gas once it’s produced. Hydrogen is a very tiny molecule — the smallest of them all — and hard to contain. But by implementing the entire Haber-Bosch process underground, the only material that would need to be sent to the surface would be the ammonia itself, which is easy to capture, store, and transport.
The only extra ingredient needed to complete the process was the addition of a source of nitrogen, such as nitrate or nitrogen gas, into the water-catalyst mixture being injected into the ground. Then, as the hydrogen gets released from water molecules after interacting with the iron-rich rocks, it can immediately bond with the nitrogen atoms also carried in the water, with the deep underground environment providing the high temperatures and pressures required by the Haber-Bosch process. A second well near the injection well then pumps the ammonia out and into tanks on the surface.
“We call this geological ammonia,” Abate says, “because we are using subsurface temperature, pressure, chemistry, and geologically existing rocks to produce ammonia directly.”
Whereas transporting hydrogen requires expensive equipment to cool and liquefy it, and virtually no pipelines exist for its transport (except near oil refinery sites), transporting ammonia is easier and cheaper. It’s about one-sixth the cost of transporting hydrogen, and there are already more than 5,000 miles of ammonia pipelines and 10,000 terminals in place in the U.S. alone. What’s more, Abate explains, ammonia, unlike hydrogen, already has a substantial commercial market in place, with production volume projected to grow by two to three times by 2050, as it is used not only for fertilizer but also as feedstock for a wide variety of chemical processes.
For example, ammonia can be burned directly in gas turbines, engines, and industrial furnaces, providing a carbon-free alternative to fossil fuels. It is being explored for maritime shipping and aviation as an alternative fuel, and as a possible space propellant.
Another upside to geological ammonia is that untreated wastewater, including agricultural runoff, which tends to be rich in nitrogen already, could serve as the water source and be treated in the process. “We can tackle the problem of treating wastewater, while also making something of value out of this waste,” Abate says.
Gao adds that this process “involves no direct carbon emissions, presenting a potential pathway to reduce global CO2 emissions by up to 1 percent.” To arrive at this point, he says, the team “overcame numerous challenges and learned from many failed attempts. For example, we tested a wide range of conditions and catalysts before identifying the most effective one.”
The project was seed-funded under a flagship project of MIT’s Climate Grand Challenges program, the Center for the Electrification and Decarbonization of Industry. Professor Yet-Ming Chiang, co-director of the center, says “I don’t think there’s been any previous example of deliberately using the Earth as a chemical reactor. That’s one of the key novel points of this approach.” Chiang emphasizes that even though it is a geological process, it happens very fast, not on geological timescales. “The reaction is fundamentally over in a matter of hours,” he says. “The reaction is so fast that this answers one of the key questions: Do you have to wait for geological times? And the answer is absolutely no.”
Professor Elsa Olivetti, a mission director of the newly established Climate Project at MIT, says, “The creative thinking by this team is invaluable to MIT’s ability to have impact at scale. Coupling these exciting results with, for example, advanced understanding of the geology surrounding hydrogen accumulations represent the whole-of-Institute efforts the Climate Project aims to support.”
“This is a significant breakthrough for the future of sustainable development,” says Geoffrey Ellis, a geologist at the U.S. Geological Survey, who was not associated with this work. He adds, “While there is clearly more work that needs to be done to validate this at the pilot stage and to get this to the commercial scale, the concept that has been demonstrated is truly transformative. The approach of engineering a system to optimize the natural process of nitrate reduction by Fe2+ is ingenious and will likely lead to further innovations along these lines.”
The initial work on the process has been done in the laboratory, so the next step will be to prove the process using a real underground site. “We think that kind of experiment can be done within the next one to two years,” Abate says. This could open doors to using a similar approach for other chemical production processes, he adds.
The team has applied for a patent and aims to work towards bringing the process to market.
“Moving forward,” Gao says, “our focus will be on optimizing the process conditions and scaling up tests, with the goal of enabling practical applications for geological ammonia in the near future.”
The research team also included Ming Lei, Bachu Sravan Kumar, Hugh Smith, Seok Hee Han, and Lokesh Sangabattula, all at MIT. Additional funding was provided by the National Science Foundation and was carried out, in part, through the use of MIT.nano facilities.
Modeling complex behavior with a simple organism
The roundworm C. elegans is a simple animal whose nervous system has exactly 302 neurons. Each of the connections between those neurons has been comprehensively mapped, allowing researchers to study how they work together to generate the animal’s different behaviors.
Steven Flavell, an MIT associate professor of brain and cognitive sciences and investigator with The Picower Institute for Learning and Memory at MIT and the Howard Hughes Medical Institute, uses the worm as a model to study motivated behaviors such as feeding and navigation, in hopes of shedding light on the fundamental mechanisms that may also determine how similar behaviors are controlled in other animals.
In recent studies, Flavell’s lab has uncovered neural mechanisms underlying adaptive changes in the worms’ feeding behavior, and his lab has also mapped how the activity of each neuron in the animal’s nervous system affects the worms’ different behaviors.
Such studies could help researchers gain insight into how brain activity generates behavior in humans. “It is our aim to identify molecular and neural circuit mechanisms that may generalize across organisms,” he says, noting that many fundamental biological discoveries, including those related to programmed cell death, microRNA, and RNA interference, were first made in C. elegans.
“Our lab has mostly studied motivated state-dependent behaviors, like feeding and navigation. The machinery that’s being used to control these states in C. elegans — for example, neuromodulators — are actually the same as in humans. These pathways are evolutionarily ancient,” he says.
Drawn to the lab
Born in London to an English father and a Dutch mother, Flavell came to the United States in 1982 at the age of 2, when his father became chief scientific officer at Biogen. The family lived in Sudbury, Massachusetts, and his mother worked as a computer programmer and math teacher. His father later became a professor of immunology at Yale University.
Though Flavell grew up in a science family, he thought about majoring in English when he arrived at Oberlin College. A musician as well, Flavell took jazz guitar classes at Oberlin’s conservatory, and he also plays the piano and the saxophone. However, taking classes in psychology and physiology led him to discover that the field that most captivated him was neuroscience.
“I was immediately sold on neuroscience. It combined the rigor of the biological sciences with deep questions from psychology,” he says.
While in college, Flavell worked on a summer research project related to Alzheimer’s disease, in a lab at Case Western Reserve University. He then continued the project, which involved analyzing post-mortem Alzheimer’s tissue, during his senior year at Oberlin.
“My earliest research revolved around mechanisms of disease. While my research interests have evolved since then, my earliest research experiences were the ones that really got me hooked on working at the bench: running experiments, looking at brand new results, and trying to understand what they mean,” he says.
By the end of college, Flavell was a self-described lab rat: “I just love being in the lab.” He applied to graduate school and ended up going to Harvard Medical School for a PhD in neuroscience. Working with Michael Greenberg, Flavell studied how sensory experience and resulting neural activity shapes brain development. In particular, he focused on a family of gene regulators called MEF2, which play important roles in neuronal development and synaptic plasticity.
All of that work was done using mouse models, but Flavell transitioned to studying C. elegans during a postdoctoral fellowship working with Cori Bargmann at Rockefeller University. He was interested in studying how neural circuits control behavior, which seemed to be more feasible in simpler animal models.
“Studying how neurons across the brain govern behavior felt like it would be nearly intractable in a large brain — to understand all the nuts and bolts of how neurons interact with each other and ultimately generate behavior seemed daunting,” he says. “But I quickly became excited about studying this in C. elegans because at the time it was still the only animal with a full blueprint of its brain: a map of every brain cell and how they are all wired up together.”
That wiring diagram includes about 7,000 synapses in the entire nervous system. By comparison, a single human neuron may form more than 10,000 synapses. “Relative to those larger systems, the C. elegans nervous system is mind-bogglingly simple,” Flavell says.
Despite their much simpler organization, roundworms can execute complex behaviors such as feeding, locomotion, and egg-laying. They even sleep, form memories, and find suitable mating partners. The neuromodulators and cellular machinery that give rise to those behaviors are similar to those found in humans and other mammals.
“C. elegans has a fairly well-defined, smallish set of behaviors, which makes it attractive for research. You can really measure almost everything that the animal is doing and study it,” Flavell says.
How behavior arises
Early in his career, Flavell’s work on C. elegans revealed the neural mechanisms that underlie the animal’s stable behavioral states. When worms are foraging for food, they alternate between stably exploring the environment and pausing to feed. “The transition rates between those states really depend on all these cues in the environment. How good is the food environment? How hungry are they? Are there smells indicating a better nearby food source? The animal integrates all of those things and then adjusts their foraging strategy,” Flavell says.
These stable behavioral states are controlled by neuromodulators like serotonin. By studying serotonergic regulation of the worm’s behavioral states, Flavell’s lab has been able to uncover how this important system is organized. In a recent study, Flavell and his colleagues published an “atlas” of the C. elegans serotonin system. They identified every neuron that produces serotonin, every neuron that has serotonin receptors, and how brain activity and behavior change across the animal as serotonin is released.
“Our studies of how the serotonin system works to control behavior have already revealed basic aspects of serotonin signaling that we think ought to generalize all the way up to mammals,” Flavell says. “By studying the way that the brain implements these long-lasting states, we can tap into these basic features of neuronal function. With the resolution that you can get studying specific C. elegans neurons and the way that they implement behavior, we can uncover fundamental features of the way that neurons act.”
In parallel, Flavell’s lab has also been mapping out how neurons across the C. elegans brain control different aspects of behavior. In a 2023 study, Flavell’s lab mapped how changes in brain-wide activity relate to behavior. His lab uses special microscopes that can move along with the worms as they explore, allowing them to simultaneously track every behavior and measure the activity of every neuron in the brain. Using these data, the researchers created computational models that can accurately capture the relationship between brain activity and behavior.
This type of research requires expertise in many areas, Flavell says. When looking for faculty jobs, he hoped to find a place where he could collaborate with researchers working in different fields of neuroscience, as well as scientists and engineers from other departments.
“Being at MIT has allowed my lab to be much more multidisciplinary than it could have been elsewhere,” he says. “My lab members have had undergrad degrees in physics, math, computer science, biology, neuroscience, and we use tools from all of those disciplines. We engineer microscopes, we build computational models, we come up with molecular tricks to perturb neurons in the C. elegans nervous system. And I think being able to deploy all those kinds of tools leads to exciting research outcomes.”
Explained: Generative AI’s environmental impact
In a two-part series, MIT News explores the environmental implications of generative AI. In this article, we look at why this technology is so resource-intensive. A second piece will investigate what experts are doing to reduce genAI’s carbon footprint and other impacts.
The excitement surrounding potential benefits of generative AI, from improving worker productivity to advancing scientific research, is hard to ignore. While the explosive growth of this new technology has enabled rapid deployment of powerful models in many industries, the environmental consequences of this generative AI “gold rush” remain difficult to pin down, let alone mitigate.
The computational power required to train generative AI models that often have billions of parameters, such as OpenAI’s GPT-4, can demand a staggering amount of electricity, which leads to increased carbon dioxide emissions and pressures on the electric grid.
Furthermore, deploying these models in real-world applications, enabling millions to use generative AI in their daily lives, and then fine-tuning the models to improve their performance draws large amounts of energy long after a model has been developed.
Beyond electricity demands, a great deal of water is needed to cool the hardware used for training, deploying, and fine-tuning generative AI models, which can strain municipal water supplies and disrupt local ecosystems. The increasing number of generative AI applications has also spurred demand for high-performance computing hardware, adding indirect environmental impacts from its manufacture and transport.
“When we think about the environmental impact of generative AI, it is not just the electricity you consume when you plug the computer in. There are much broader consequences that go out to a system level and persist based on actions that we take,” says Elsa A. Olivetti, professor in the Department of Materials Science and Engineering and the lead of the Decarbonization Mission of MIT’s new Climate Project.
Olivetti is senior author of a 2024 paper, “The Climate and Sustainability Implications of Generative AI,” co-authored by MIT colleagues in response to an Institute-wide call for papers that explore the transformative potential of generative AI, in both positive and negative directions for society.
Demanding data centers
The electricity demands of data centers are one major factor contributing to the environmental impacts of generative AI, since data centers are used to train and run the deep learning models behind popular tools like ChatGPT and DALL-E.
A data center is a temperature-controlled building that houses computing infrastructure, such as servers, data storage drives, and network equipment. For instance, Amazon has more than 100 data centers worldwide, each of which has about 50,000 servers that the company uses to support cloud computing services.
While data centers have been around since the 1940s (the first was built at the University of Pennsylvania in 1945 to support the first general-purpose digital computer, the ENIAC), the rise of generative AI has dramatically increased the pace of data center construction.
“What is different about generative AI is the power density it requires. Fundamentally, it is just computing, but a generative AI training cluster might consume seven or eight times more energy than a typical computing workload,” says Noman Bashir, lead author of the impact paper, who is a Computing and Climate Impact Fellow at MIT Climate and Sustainability Consortium (MCSC) and a postdoc in the Computer Science and Artificial Intelligence Laboratory (CSAIL).
Scientists have estimated that the power requirements of data centers in North America increased from 2,688 megawatts at the end of 2022 to 5,341 megawatts at the end of 2023, partly driven by the demands of generative AI. Globally, the electricity consumption of data centers rose to 460 terawatts in 2022. This would have made data centers the 11th largest electricity consumer in the world, between the nations of Saudi Arabia (371 terawatts) and France (463 terawatts), according to the Organization for Economic Co-operation and Development.
By 2026, the electricity consumption of data centers is expected to approach 1,050 terawatts (which would bump data centers up to fifth place on the global list, between Japan and Russia).
While not all data center computation involves generative AI, the technology has been a major driver of increasing energy demands.
“The demand for new data centers cannot be met in a sustainable way. The pace at which companies are building new data centers means the bulk of the electricity to power them must come from fossil fuel-based power plants,” says Bashir.
The power needed to train and deploy a model like OpenAI’s GPT-3 is difficult to ascertain. In a 2021 research paper, scientists from Google and the University of California at Berkeley estimated the training process alone consumed 1,287 megawatt hours of electricity (enough to power about 120 average U.S. homes for a year), generating about 552 tons of carbon dioxide.
While all machine-learning models must be trained, one issue unique to generative AI is the rapid fluctuations in energy use that occur over different phases of the training process, Bashir explains.
Power grid operators must have a way to absorb those fluctuations to protect the grid, and they usually employ diesel-based generators for that task.
Increasing impacts from inference
Once a generative AI model is trained, the energy demands don’t disappear.
Each time a model is used, perhaps by an individual asking ChatGPT to summarize an email, the computing hardware that performs those operations consumes energy. Researchers have estimated that a ChatGPT query consumes about five times more electricity than a simple web search.
“But an everyday user doesn’t think too much about that,” says Bashir. “The ease-of-use of generative AI interfaces and the lack of information about the environmental impacts of my actions means that, as a user, I don’t have much incentive to cut back on my use of generative AI.”
With traditional AI, the energy usage is split fairly evenly between data processing, model training, and inference, which is the process of using a trained model to make predictions on new data. However, Bashir expects the electricity demands of generative AI inference to eventually dominate since these models are becoming ubiquitous in so many applications, and the electricity needed for inference will increase as future versions of the models become larger and more complex.
Plus, generative AI models have an especially short shelf-life, driven by rising demand for new AI applications. Companies release new models every few weeks, so the energy used to train prior versions goes to waste, Bashir adds. New models often consume more energy for training, since they usually have more parameters than their predecessors.
While electricity demands of data centers may be getting the most attention in research literature, the amount of water consumed by these facilities has environmental impacts, as well.
Chilled water is used to cool a data center by absorbing heat from computing equipment. It has been estimated that, for each kilowatt hour of energy a data center consumes, it would need two liters of water for cooling, says Bashir.
“Just because this is called ‘cloud computing’ doesn’t mean the hardware lives in the cloud. Data centers are present in our physical world, and because of their water usage they have direct and indirect implications for biodiversity,” he says.
The computing hardware inside data centers brings its own, less direct environmental impacts.
While it is difficult to estimate how much power is needed to manufacture a GPU, a type of powerful processor that can handle intensive generative AI workloads, it would be more than what is needed to produce a simpler CPU because the fabrication process is more complex. A GPU’s carbon footprint is compounded by the emissions related to material and product transport.
There are also environmental implications of obtaining the raw materials used to fabricate GPUs, which can involve dirty mining procedures and the use of toxic chemicals for processing.
Market research firm TechInsights estimates that the three major producers (NVIDIA, AMD, and Intel) shipped 3.85 million GPUs to data centers in 2023, up from about 2.67 million in 2022. That number is expected to have increased by an even greater percentage in 2024.
The industry is on an unsustainable path, but there are ways to encourage responsible development of generative AI that supports environmental objectives, Bashir says.
He, Olivetti, and their MIT colleagues argue that this will require a comprehensive consideration of all the environmental and societal costs of generative AI, as well as a detailed assessment of the value in its perceived benefits.
“We need a more contextual way of systematically and comprehensively understanding the implications of new developments in this space. Due to the speed at which there have been improvements, we haven’t had a chance to catch up with our abilities to measure and understand the tradeoffs,” Olivetti says.
MIT Global SCALE Network named No. 1 supply chain and logistics master’s program for 2024-25
The MIT Global Supply Chain and Logistics Excellence (SCALE) Network has once again been ranked as the world’s top master’s program for supply chain and logistics management by Eduniversal’s 2024/2025 Best Masters Rankings. This recognition marks the eighth consecutive No. 1 ranking since 2016, reaffirming MIT’s unparalleled leadership in supply chain education, research, and practice.
Eduniversal evaluates more than 20,000 postgraduate programs globally each year, considering academic reputation, graduate employability, and student satisfaction.
The MIT SCALE Network’s sustained top ranking reflects its commitment to fostering international diversity; delivering hands-on, project-based learning; and success in developing a generation of supply chain leaders ready to tackle global supply chain challenges.
A growing global network with local impact
This year’s ranking coincides with the MIT SCALE Network’s expansion of its global footprint, highlighted by the recent announcement of the UK SCALE Center at Loughborough University. The center, which will welcome its inaugural cohort in fall 2025, underscores MIT’s commitment to advancing supply chain innovation and creating transformative opportunities for students and researchers.
The UK SCALE Center joins the network’s global community of centers in the United States, China, Spain, Colombia, and Luxembourg. Together, these centers deliver world-class education and practical solutions that address critical supply chain challenges across industries, empowering a global alumni base of more than 1,900 leaders representing over 50 different countries.
"The launch of the UK SCALE Center represents a fantastic opportunity for Loughborough University to showcase our cutting-edge research and data-driven, forward-thinking approach to supporting the U.K. supply chain industry,” says Jan Godsell, dean of Loughborough Business School. “Through projects like the InterAct Network and our implementation of the Made Smarter Innovation 'Leading Digital Transformation' program, we’re offering businesses and industry professionals the essential training and leading insights into the future of the supply chain ecosystem, which I’m excited to build on with the creation of this new MSc in supply chain management."
Other MIT SCALE centers also emphasized the network’s transformative impact:
“The MIT SCALE Network provides NISCI students with the tools, expertise, and global connections to lead in today’s rapidly evolving supply chain environment,” says Jay Guo, director of the Ningbo China Institute for Supply Chain Innovation.
Susana Val, director of Zaragoza Logistics Center (ZLC), highlights the program’s reach and influence: “For the last 21 years, ZLC has educated over 5,000 logistics professionals from more than 90 nationalities. We are proud of this recognition and look forward to continuing our alliance with the MIT SCALE Network, upholding the rigor and quality that define our teaching.”
From Luxembourg, Benny Mantin, director of the Luxembourg Center for Logistics and Supply Chain Management (LCL), adds, “Our students greatly appreciate the LCL’s SCALE Network membership as it provides them with superb experience and ample opportunities to network and expand their scope.”
The global presence and collaborative approach of the MIT SCALE Network help define its mission: to deliver education and research that drive transformative impact in every corner of the world.
A legacy of leadership
This latest recognition from Eduniversal underscores the MIT SCALE Network’s leadership in supply chain education. For over two decades, its master’s programs have shaped graduates who tackle pressing challenges across industries and geographies.
"This recognition reflects the dedication of our faculty, researchers, and global partners to delivering excellence in supply chain education," says Yossi Sheffi, director of the MIT Center for Transportation and Logistics. “The MIT SCALE Network’s alumni are proof of the program’s impact, applying their skills to tackle challenges across every industry and continent.”
Maria Jesus Saenz, executive director of the MIT SCM Master’s Program, emphasizes the strength of the global alumni network: “The MIT SCALE Network doesn’t just prepare graduates — it connects them to a global community of supply chain leaders. This powerful ecosystem fosters collaboration and innovation that transcends borders, enabling our graduates to tackle the world’s most pressing supply chain challenges.”
Founded in 2003, the MIT SCALE Network connects world-class research centers across multiple continents, offering top-ranked master’s and executive education programs that combine academic rigor with real-world application. Graduates are among the most sought-after professionals in the global supply chain field.
Making the art world more accessible
In the world of high-priced art, galleries usually act as gatekeepers. Their selective curation process is a key reason galleries in major cities often feature work from the same batch of artists. The system limits opportunities for emerging artists and leaves great art undiscovered.
NALA was founded by Benjamin Gulak ’22 to disrupt the gallery model. The company’s digital platform, which was started as part of an MIT class project, allows artists to list their art and uses machine learning and data science to offer personalized recommendations to art lovers.
By providing a much larger pool of artwork to buyers, the company is dismantling the exclusive barriers put up by traditional galleries and efficiently connecting creators with collectors.
“There’s so much talent out there that has never had the opportunity to be seen outside of the artists’ local market,” Gulak says. “We’re opening the art world to all artists, creating a true meritocracy.”
NALA takes no commission from artists, instead charging buyers an 11.5 percent commission on top of the artist’s listed price. Today more than 20,000 art lovers are using NALA's platform, and the company has registered more than 8,500 artists.
“My goal is for NALA to become the dominant place where art is discovered, bought, and sold online,” Gulak says. “The gallery model has existed for such a long period of time that they are the tastemakers in the art world. However, most buyers never realize how restrictive the industry has been.”
From founder to student to founder again
Growing up in Canada, Gulak worked hard to get into MIT, participating in science fairs and robotic competitions throughout high school. When he was 16, he created an electric, one-wheeled motorcycle that got him on the popular television show “Shark Tank” and was later named one of the top inventions of the year by Popular Science.
Gulak was accepted into MIT in 2009 but withdrew from his undergrad program shortly after entering to launch a business around the media exposure and capital from “Shark Tank.” Following a whirlwind decade in which he raised more than $12 million and sold thousands of units globally, Gulak decided to return to MIT to complete his degree, switching his major from mechanical engineering to one combining computer science, economics, and data science.
“I spent 10 years of my life building my business, and realized to get the company where I wanted it to be, it would take another decade, and that wasn’t what I wanted to be doing,” Gulak says. “I missed learning, and I missed the academic side of my life. I basically begged MIT to take me back, and it was the best decision I ever made.”
During the ups and downs of running his company, Gulak took up painting to de-stress. Art had always been a part of Gulak’s life, and he had even done a fine arts study abroad program in Italy during high school. Determined to try selling his art, he collaborated with some prominent art galleries in London, Miami, and St. Moritz. Eventually he began connecting artists he’d met on travels from emerging markets like Cuba, Egypt, and Brazil to the gallery owners he knew.
“The results were incredible because these artists were used to selling their work to tourists for $50, and suddenly they’re hanging work in a fancy gallery in London and getting 5,000 pounds,” Gulak says. “It was the same artist, same talent, but different buyers.”
At the time, Gulak was in his third year at MIT and wondering what he’d do after graduation. He thought he wanted to start a new business, but every industry he looked at was dominated by tech giants. Every industry, that is, except the art world.
“The art industry is archaic,” Gulak says. “Galleries have monopolies over small groups of artists, and they have absolute control over the prices. The buyers are told what the value is, and almost everywhere you look in the industry, there’s inefficiencies.”
At MIT, Gulak was studying the recommender engines that are used to populate social media feeds and personalize show and music suggestions, and he envisioned something similar for the visual arts.
“I thought, why, when I go on the big art platforms, do I see horrible combinations of artwork even though I’ve had accounts on these platforms for years?” Gulak says. “I’d get new emails every week titled ‘New art for your collection,’ and the platform had no idea about my taste or budget.”
For a class project at MIT, Gulak built a system that tried to predict the types of art that would do well in a gallery. By his final year at MIT, he had realized that working directly with artists would be a more promising approach.
“Online platforms typically take a 30 percent fee, and galleries can take an additional 50 percent fee, so the artist ends up with a small percentage of each online sale, but the buyer also has to pay a luxury import duty on the full price,” Gulak explains. “That means there’s a massive amount of fat in the middle, and that’s where our direct-to-artist business model comes in.”
Today NALA, which stands for Networked Artistic Learning Algorithm, onboards artists by having them upload artwork and fill out a questionnaire about their style. They can begin uploading work immediately and choose their listing price.
The company began by using AI to match art with its most likely buyer. Gulak notes that not all art will sell — “if you’re making rock paintings there may not be a big market” — and artists may price their work higher than buyers are willing to pay, but the algorithm works to put art in front of the most likely buyer based on style preferences and budget. NALA also handles sales and shipments, providing artists with 100 percent of their list price from every sale.
“By not taking commissions, we’re very pro artists,” Gulak says. “We also allow all artists to participate, which is unique in this space. NALA is built by artists for artists.”
Last year, NALA also started allowing buyers to take a photo of something they like and see similar artwork from its database.
“In museums, people will take a photo of masterpieces they’ll never be able to afford, and now they can find living artists producing the same style that they could actually put in their home,” Gulak says. “It makes art more accessible.”
Championing artists
Ten years ago, Ben Gulak was visiting Egypt when he discovered an impressive mural on the street. Gulak found the local artist, Ahmed Nofal, on Instagram and bought some work. Later, he brought Nofal to Dubai to participate in World Art Dubai. The artist’s work was so well-received he ended up creating murals for the Royal British Museum in London and Red Bull. Most recently, Nofal and Gulak collaborated together during Art Basel 2024 doing a mural at the Museum of Graffiti in Miami.
Gulak has worked personally with many of the artists on his platform. For more than a decade he’s travelled to Cuba buying art and delivering art supplies to friends. He’s also worked with artists as they work to secure immigration visas.
“Many people claim they want to help the art world, but in reality, they often fall back on the same outdated business models,” says Gulak. “Art isn’t just my passion — it’s a way of life for me. I’ve been on every side of the art world: as a painter selling my work through galleries, as a collector with my office brimming with art, and as a collaborator working alongside incredible talents like Raheem Saladeen Johnson. When artists visit, we create together, sharing ideas and brainstorming. These experiences, combined with my background as both an artist and a computer scientist, give me a unique perspective. I’m trying to use technology to provide artists with unparalleled access to the global market and shake things up.”
Karl Berggren named faculty head of electrical engineering in EECS
Karl K. Berggren, the Joseph F. and Nancy P. Keithley Professor of Electrical Engineering at MIT, has been named the new faculty head of electrical engineering in the Department of Electrical Engineering and Computer Science (EECS), effective Jan. 15.
“Karl’s exceptional interdisciplinary research combining electrical engineering, physics, and materials science, coupled with his experience working with industry and government organizations, makes him an ideal fit to head electrical engineering. I’m confident electrical engineering will continue to grow under his leadership,” says Anantha Chandrakasan, chief innovation and strategy officer, dean of engineering, and Vannevar Bush Professor of Electrical Engineering and Computer Science.
“Karl has made an incredible impact as a researcher and educator over his two decades in EECS. Students and faculty colleagues praise his thoughtful approach to teaching, and the care with which he oversaw the teaching labs in his prior role as undergraduate lab officer for the department. He will undoubtedly be an excellent leader, bringing his passion for education and collaborative spirit to this new role,” adds Daniel Huttenlocher, dean of the MIT Schwarzman College of Computing and the Henry Ellis Warren Professor of Electrical Engineering and Computer Science.
Berggren joins the leadership of EECS, which jointly reports to the MIT Schwarzman College of Computing and the School of Engineering. The largest academic department at MIT, EECS was reorganized in 2019 as part of the formation of the college into three overlapping sub-units in electrical engineering, computer science, and artificial intelligence and decision-making. The restructuring has enabled each of the three sub-units to concentrate on faculty recruitment, mentoring, promotion, academic programs, and community building in coordination with the others.
A member of the EECS faculty since 2003, Berggren has taught a range of subjects in the department, including Digital Communications, Circuits and Electronics, Fundamentals of Programming, Applied Quantum and Statistical Physics, Introduction to EECS via Interconnected Embedded Systems, Introduction to Quantum Systems Engineering, and Introduction to Nanofabrication. Before joining EECS, Berggren worked as a staff member at MIT Lincoln Laboratory for seven years. Berggren also maintains an active consulting practice and has experience working with industrial and government organizations.
Berggren’s current research focuses on superconductive circuits, electronic devices, single-photon detectors for quantum applications, and electron-optical systems. He heads the Quantum Nanostructures and Nanofabrication Group, which develops nanofabrication technology at the few-nanometer length scale. The group uses these technologies to push the envelope of what is possible with photonic and electrical devices, focusing on superconductive and free-electron devices.
Berggren has received numerous prestigious awards and honors throughout his career. Most recently, he was named an MIT MacVicar Fellow in 2024. Berggren is also a fellow of the AAAS, IEEE, and the Kavli Foundation, and a recipient of the 2015 Paul T. Forman Team Engineering Award from the Optical Society of America (now Optica). In 2016, he received a Bose Fellowship and was also a recipient of the EECS department’s Frank Quick Innovation Fellowship and the Burgess (’52) & Elizabeth Jamieson Award for Excellence in Teaching.
Berggren succeeds Joel Voldman, who has served as the inaugural electrical engineering faculty head since January 2020.
“Joel has been in leadership roles since 2018, when he was named associate department head of EECS. I am deeply grateful to him for his invaluable contributions to EECS since that time,” says Asu Ozdaglar, MathWorks Professor and head of EECS, who also serves as the deputy dean of the MIT Schwarzman College of Computing. “I look forward to working with Karl now and continuing along the amazing path we embarked on in 2019.”
MIT student encourages all learners to indulge their curiosity with MIT Open Learning's MITx
Shreya Mogulothu is naturally curious. As a high school student in New Jersey, she was interested in mathematics and theoretical computer science (TCS). So, when her curiosity compelled her to learn more, she turned to MIT Open Learning’s online resources and completed the Paradox and Infinity course on MITx Online.
“Coming from a math and TCS background, the idea of pushing against the limits of assumptions was really interesting,” says Mogulothu, now a junior at MIT. “I mean, who wouldn’t want to learn more about infinity?”
The class, taught by Agustín Rayo, professor of philosophy and the current dean of the School of Humanities, Arts, and Social Sciences, and David Balcarras, a former instructor in philosophy and fellow in the Digital Learning Lab at Open Learning, explores the intersection of math and philosophy and guides learners through thinking about paradoxes and open-ended problems, as well as the boundaries of theorizing and the limits of standard mathematical tools.
“We talked about taking regular assumptions about numbers and objects and pushing them to extremes,” Mogulothu says. “For example, what contradictions arise when you talk about an infinite set of things, like the infinite hats paradox?”
The infinite hats paradox, also known as Bacon’s Puzzle, involves an infinite line of people, each wearing one of two colors of hats. The puzzle posits that each individual can see only the hat of the person in front of them and must guess the color of their own hat. The puzzle challenges students to identify if there is a strategy that can ensure the least number of incorrect answers and to consider how strategy may change if there is a finite number of people. Mogulothu was thrilled that a class like this was available to her even though she wasn’t yet affiliated with MIT.
“My MITx experience was one of the reasons I came to MIT,” she says. “I really liked the course, and I was happy it was shared with people like me, who didn’t even go to the school. I thought that a place that encouraged even people outside of campus to learn like that would be a pretty good place to study.”
Looking back at the course, Balcarras says, “Shreya may have been the most impressive student in our online community of approximately 3,900 learners and 100 verified learners. I cannot single out another student whose performance rivaled hers.”
Because of her excellent performance, Mogulothu was invited to submit her work to the 2021 MITx Philosophy Awards. She won. In fact, Balcarras remembers, both papers she wrote for the course would have won. They demonstrated, he says, “an unusually high degree of precision, formal acumen, and philosophical subtlety for a high school student.”
Completing the course and winning the award was rewarding, Mogulothu says. It motivated her to keep exploring new things as a high school student, and then as a new student enrolled at MIT.
She came to college thinking she would declare a major in math or computer science. But when she looked at the courses she was most interested in, she realized she should pursue a physics major.
She has enjoyed the courses in her major, especially class STS.042J/8.225J (Einstein, Oppenheimer, Feynman: Physics in the 20th Century), taught by David Kaiser, the Germeshausen Professor of the History of Science and professor of physics. She took the course on campus, but it is also available on Open Learning’s MIT OpenCourseWare. As a student, she continues to use MIT Open Learning resources to check out courses and review syllabi as she plans her coursework.
In summer 2024, Mogulothu did research on gravitational wave detection at PIER, the partnership between research center DESY and the University of Hamburg, in Hamburg, Germany. She wants to pursue a PhD in physics to keep researching, expanding her mind, and indulging the curiosity that led her to MITx in the first place. She encourages all learners to feel comfortable and confident trying something entirely new.
“I went into the Paradox and Infinity course thinking, ‘yeah, math is cool, computer science is cool,’” she says. “But, actually taking the course and learning about things you don’t even expect to exist is really powerful. It increases your curiosity and is super rewarding to stick with something and realize how much you can learn and grow.”
More than an academic advisor
Advisors are meant to guide students academically, supporting their research and career objectives. For MIT graduate students, the Committed to Caring program recognizes those who go above and beyond.
Professors Iain Stewart and Roberto Fernandez are two of the 2023-25 Committed to Caring cohort, supporting their students through self-doubt, developing a welcoming environment, and serving as a friend.
Iain Stewart: Supportive, equitable, and inclusive
Iain Stewart is the Otto and Jane Morningstar Professor of Science and former director of the Center for Theoretical Physics (CTP). His research interests center around nuclear and particle physics, where he develops and applies effective field theories to understand interactions between elementary particles and particularly strong interactions described by quantum chromodynamics.
Stewart shows faith in his students’ abilities even when they doubt themselves. According to his nominators, the field of physics, like many areas of intellectual pursuit, can attract a wide range of personalities, including those who are highly confident as well as those who may grapple with self-doubt. He explains concepts in a down-to-earth manner and does not make his students feel less than they are.
For his students, Stewart’s research group comes as a refreshing change. Stewart emphasizes that graduate school is for learning, and that one is not expected to know everything from the onset.
Stewart shows a great level of empathy and emotional support for his students. For example, one of the nominators recounted a story about preparing for their oral qualification exam. The student had temporarily suspended research, and another faculty member made a disparaging comment about the student’s grasp of their research. The student approached Stewart in distress.
"As your advisor,” Stewart reassured them, “I can tell you confidently that you know your research and you are doing well, and it’s totally OK to put it off for a while to prepare for the qual."
Stewart’s words gave the student a sense of relief and validation, reminding them that progress is a journey, not a race, and that taking time to prepare thoughtfully was both wise and necessary.
Always emphasizing positivity in his feedback, Stewart reminds advisees of their achievements and progress, helping them develop a more optimistic mindset. Stewart’s mentorship style recognizes individual student needs, a trait that his students find uncommon. His research group flourishes due to this approach, and a large number of his graduate and postdoc students have achieved great success.
During his six years as director, Stewart has made significant contributions to the CTP. He has improved the culture and demographics due to strong and inclusive leadership. In particular, a noteworthy number of women have joined the CTP.
In his own research group, a large number of international and female students have found a place, which is uncommon for groups in theoretical physics. Currently, three out of seven group members are female in a field where fewer than 10 percent are women.
Stewart’s nominators believe that given the number of women he has mentored in his career, he is single-handedly contributing to improving the diversity in his field. His nominators say he supports diverse backgrounds, and financially supports and encourages participation for marginalized groups.
Roberto Fernandez: Professor and friend
Roberto Fernandez is the William F. Pounds Professor of Organization Studies at the MIT Sloan School of Management as well as the co-director of the Economic Sociology PhD Program. His research focuses on organizations, social networks, and race and gender stratification. He has extensive experience doing field research in organizations, and he currently focuses on the organizational processes surrounding the hiring of new talent.
Fernandez describes himself as a “full-service professor.” He tries to attend to differing needs and circumstances of students and the situations they find themselves in, offering advice and consolation.
Fernandez is very understanding of his students, and is happy to speak to them about academic and personal problems alike. He acknowledges that each student comes from a different background with individual experience, and Fernandez attempts to accommodate each one in an ideal manner.
He advises in a way that respects a student’s personal life, but still expects a reasonable amount of produced work that motivates the student, allows for them to excel, and keeps them to a high standard.
Fernandez says, “It is just my sense of duty to pay forward how my mentors treated me. I feel like I would dishonor their work if I were not to pass it on.”
A nominator shared that Fernandez serves as both a professor and a friend. He has gone out of his way to check in and chat with them. They said that Fernandez is the only professor who has taken the time to truly get to know their story, and Fernandez speaks to students like an equal.
The nominator noted that many people at MIT enjoy a high level of privilege. Despite the differences in their circumstances, however, the nominator feels comfortable talking to Fernandez.
Happily, the professor continued to touch base with the nominator long after their class had finished, and he is the one person who really made them feel like MIT was their home. This experience stood out as unique for the nominator, and played a large role in their experience.
In addition to providing genuine connections, Fernandez advises incoming graduate students about the need for a mindset shift. Graduate school is not like undergrad. Being an excellent student is necessary, but it is not sufficient to succeed in a PhD program. Excellent undergraduate students are consumers of knowledge; on the other hand, excellent graduate students are producers of knowledge.
The nominator enthused, “[Fernandez] really went above and beyond, and this means a lot.”
MIT philosopher Sally Haslanger honored with Quinn Prize
MIT philosopher Sally Haslanger has been named the 2024 recipient of the prestigious Philip L. Quinn Prize from the American Philosophical Association (APA).
The award recognizes Haslanger’s lifelong contributions to philosophy and philosophers. Haslanger, the Ford Professor of Philosophy and Women’s and Gender Studies, says she is deeply honored by the recognition.
“So many philosophers I deeply respect have come before me as awardees, including Judith Jarvis Thomson, my former colleague and lifelong inspiration,” Haslanger says. “Judy and I both were deeply engaged in doing metaphysics with an eye toward the moral/political domain. Both of us were committed feminists in a time when it was not professionally easy. Both of us believed in the power of institutions, such as the APA and the American Association of University Professors (AAUP), to sustain a flourishing intellectual community. Both of us have demanded that institutions we are part of abide by their values.”
Haslanger joined the MIT faculty in 1998.
Her research features explorations of the social construction of categories like gender, race, and the family; social explanation and social structure; and topics in feminist epistemology. She has also published in metaphysics and critical race theory. Broadly speaking, her work links issues of social justice with contemporary work in epistemology, metaphysics, philosophy of language, and philosophy of mind.
Her book, “Resisting Reality: Social Construction and Social Critique” (Oxford University Press, 2012), was awarded the Joseph B. Gittler prize for outstanding work in the philosophy of social science. She also co-authored “What is Race: Four Philosophical Views” (Oxford University Press, 2019). Her current book, “Doing Justice to the Social” (under contract with Oxford University Press), develops an account of social practices and structures, emphasizing their materiality, the role of ideology, and potential grounds for critique. She continues to document and ameliorate the underrepresentation of women and other minorities in philosophy.
Haslanger, a former president of the Eastern Division of the APA, singles out the collaborative nature of the field while also celebrating her peers’ recognition, noting her work is “inspired, nourished, and scaffolded by others.”
“Judy was a notable inspiration (and a clear example of how hard such work can be), but there are so many others who have been on this journey with me and kept me going, including feminist colleagues across the country and abroad, graduate students, staff members, and allies from many different disciplines and professions,” Haslanger says.
Awarded annually since 2007, the Quinn Prize honors the memory of Philip L. Quinn, a noted philosopher from the University of Notre Dame who served as president of the APA Central Division for many years. The prize carries a $2,500 award and an engraved plaque.
Kieran Setiya, the Peter de Florez Professor of Philosophy and head of the Department of Linguistics and Philosophy, says Haslanger has played a “transformative role in philosophy.”
“Sally’s influence on the field has been vast. Bridging a deep divide, she has brought critical social theory into conversation with analytic philosophy, arguing for an account of social structures and practices that does justice to their materiality,” Setiya says. “This work earned her a Guggenheim Fellowship as well as membership in the American Academy of Arts and Sciences, along with invitations to give lectures named after canonical philosophers past and present: Wittgenstein, Benjamin, Hempel, Kant, Spinoza, and others.”
Setiya noted Haslanger’s substantial contributions to the field, including her role in founding the Philosophy in an Inclusive Key Summer Institute (PIKSI) in Boston, which for 10 years has brought diverse undergraduates to MIT to show them that graduate study in philosophy is a meaningful option for them and to mentor them as they apply to graduate school.
“As Sally’s colleague, I am in awe not just of her extraordinary philosophical and professional achievements, but of her integrity and the seemingly limitless energy she invests in her students, in the Philosophy Section, in MIT, in the profession, and in fighting for social justice in the world from which academia is inextricable,” Setiya adds.