Feed aggregator
Victor K. McElheny, founding director of MIT’s Knight Science Journalism Program, dies at 89
Victor K. McElheny, the celebrated journalist and author who founded MIT’s Knight Science Journalism Program more than 40 years ago and served for 15 years as its director, died on July 14 in Lexington, Massachusetts, after a brief illness. He was 89.
Born in Boston and raised in Poughkeepsie, New York, McElheny’s storied journalism career spanned seven decades, during which he wrote for several of the nation’s leading newspapers and magazines, penned three critically acclaimed books, and produced groundbreaking coverage of national stories ranging from the Apollo moon landing to the sequencing of the human genome. He is remembered as a steadfast champion of science journalism who eloquently made the case for the profession’s importance in society and worked tirelessly to help the field — and its practitioners — thrive.
“Victor was a pioneering science journalist, at publications that included The Charlotte Observer, Science, and The New York Times, and an author of note, especially for his biographies of scientific luminaries from Edwin Land to James Watson,” says Deborah Blum, who now heads the MIT program McElheny founded. “Yet, he still found time in 1983 to create the Knight Science Journalism Program, to fight for it, find funding for it, and to build it into what it is today.”
A 1957 graduate of Harvard University, McElheny worked as a reporter for the school’s venerable newspaper, The Harvard Crimson, before eventually taking a job as a science reporter at The Charlotte Observer in North Carolina. In the decades that followed, he served as the European editor at Science magazine, science editor of the Boston Globe, and the technology specialist at The New York Times, among other prominent posts. McElheny’s 1970s reporting on emerging techniques in molecular biology earned the journalist a reputation as a leading reporter on the developing field of genetics — and helped lay the groundwork for his critically acclaimed 2003 biography, “Watson and DNA: Making a Scientific Revolution.” McElheny also authored a biography of Edwin Land, co-founder of the Polaroid Corp., and a well-received book about the groundbreaking effort to map the human genome.
The impact of McElheny’s own stalwart career is rivaled only by his indelible impact on the careers of legions of science journalists who have come behind him.
In 1983, after a stint as director of the Banbury Center at Cold Spring Harbor Laboratory, McElheny — along with then-MIT president Paul Gray and then-director of MIT’s Science, Technology, and Society Program, Carl Kaysen — helped launch a first-of-its-kind science journalism fellowship program, funded with support from the Alfred P. Sloan and Andrew W. Mellon foundations. “The notion took hold that it would be good for MIT to have a fellowship program for science journalists, on the model of the Nieman Fellowship at Harvard,” McElheny recalled in a 2013 MIT News story. (McElheny, himself, had been part of the Nieman’s 1962-63 fellowship class.) The goal, as he explained it, was to allow journalists to connect with researchers “to make acquaintances who will provide them not only with story tips, but with judgment.”
In 1987, McElheny secured a multimillion-dollar grant from the Knight Foundation, creating an endowment that continues to support the fellowship to this day. McElheny led the program — originally known as the Vannevar Bush Science Journalism Fellowship Program and later renamed the Knight Science Journalism Program — for 15 years before stepping down to make way for his successor, preeminent journalist and editor Boyce Rensberger.
“What motivated the man professionally was a deep desire that the public understand and appreciate science and technology,” Rensberger recalls of his predecessor. “And he knew the only way that could happen to people out of school was through science journalists and other science writers creating knowledgeable content for mass media.”
Over the Knight Science Journalism Program’s 42-year history, it has supported and helped advance the careers of more than 400 leading science journalists from around the world. Following his retirement, McElheny remained actively involved with the program, frequently visiting to drop in on seminars or share an inspiring word with incoming classes of fellows.
In 2018, McElheny and his wife, Ruth, teamed with Blum, who joined the program as director in 2015, to establish the Victor K. McElheny Award for local and regional science journalism. The award, which received early support from the Rita Allen Foundation, is now funded by a generous endowment created by the McElhenys. Now entering its seventh year, it has quickly built a reputation as a prestigious national competition honoring some of the country’s best local science journalism.
“Victor was a transformational figure for MIT,” says Agustín Rayo, dean of MIT’s School of Humanities, Arts, and Social Sciences, which houses the Knight Science Journalism Program. “He never ceased to impress me. He had an extraordinary understanding of the ways in which science and technology shape society, of the ways in which society has shaped MIT, and of the ways in which MIT can shape the world.”
“Victor touched so many lives in his long and storied career,” says Usha Lee McFarling, a former Knight Science Journalism Fellow who was recently named to succeed Blum as the program’s director. Even in recent weeks and months, she says, “Victor was bubbling over with ideas on how to keep the fellowship program he founded more than 40 years ago powerful and relevant.”
McElheny’s death was preceded by that of his wife, Ruth — also an accomplished science communicator — who died in April. He is survived by his brothers, Kenneth McElheny and Steven McElheny, and Steven’s wife Karen Sexton; his sister, Robin McElheny, and her husband Alex Griswold; his six nephews and nieces, Josiah and Tobias McElheny, Raphael Griswold, and Hanna, Molly, and Rosa McElheny; and Ruth’s nephew, Dennis Sullivan, and niece, Deirdre Sullivan.
Alumni of the Knight Science Journalism Program describe Victor McElheny’s passing as a huge loss for the entire field of science journalism — a loss of a visionary who generously shared both his remarkable knowledge of the history of the field and his inspiring vision of the possibilities for the future.
“Whether we’re talking about the stars, the Earth, the oceans, the atmosphere, or other planets, our level of understanding is increasing all the time,” McElheny mused to science writer Brittany Flaherty in a 2019 profile. “There’s always more — a lot more — for science journalists to do.”
For those who wish to honor McElheny’s memory, his family invites memorial gifts to the Victor K. McElheny Award Fund.
School of Architecture and Planning recognizes faculty with academic promotions in 2025
Seven faculty in the MIT School of Architecture and Planning (SA+P) have been honored for their contributions through promotions, effective July 1. Three faculty promotions are in the Department of Architecture; three are in the Department of Urban Studies and Planning; and one is in the Program in Media Arts and Sciences.
“Whether architects, urbanists, computer scientists, or nanotechnologists, they represent our school at its best, in its breadth of inquiry and mission to improve the relationship between human beings and their environments,” says SA+P Dean Hashim Sarkis.
Department of Architecture
Marcelo Coelho has been promoted to associate professor of the practice. Coelho is the director of the Design Intelligence Lab, which explores the intersection of human and machine intelligence across design, AI, and fabrication. His work ranges from light-based installations to physical computing. Recognition for his work includes two Prix Ars Electronica awards and Fast Company’s Innovation by Design Award. Coelho’s experimental approach redefines creative processes, transforming how we imagine and interact with intelligent systems. Coelho teaches courses that bring together industrial design, user experience, and artificial intelligence.
Holly Samuelson has been promoted to associate professor without tenure. Samuelson has co-authored over 40 peer-reviewed papers, winning a Best Paper award from the journal Energy and Building. As a recognized expert in architectural technology, she has been featured in media outlets such as The Washington Post, The Boston Globe, the BBC, and The Wall Street Journal.
Rafi Segal has been promoted to full professor. An award-winning designer, Segal works across architectural and urban scales, with projects ranging from Villa 003 in the ORDOS 100 series to the Kitgum Peace Museum in Uganda, the Ashdod Museum of Art in Israel, and the winning design proposal for the National Library of Israel in Jerusalem. His current work includes planning a new communal neighborhood for an Israeli kibbutz and curating the first exhibition on Alfred Neumann’s 1960s architecture.
Department of Urban Studies and Planning (DUSP)
Carlo Ratti has been reappointed as professor of the practice. Ratti is the director of the Senseable City Lab and a founding partner of the international design office Carlo Ratti Associati. He has co-authored over 500 publications and holds several patents. His work has been exhibited globally, including at the Venice Biennale, the Museum of Modern Art in New York City, and the Design Museum in Barcelona. Two of his projects, the Digital Water Pavilion and the Copenhagen Wheel, were named among TIME Magazine’s “Best Inventions of the Year.” He is the curator of the 2025 Venice Biennale’s 19th International Architecture Exhibition.
Albert Saiz has been promoted to full professor. Saiz serves as the director of MIT’s Urban Economics Lab, which conducts research on real estate economics, urban economics, housing markets, local public finance, zoning regulations, global real estate, and demographic trends affecting urban and real estate development worldwide. He also contributes to the broader research community as a visiting scholar at the Federal Reserve Bank of Philadelphia, a research fellow at the Institute for the Analysis of Labor, and editor for the Journal of Housing Economics.
Delia Wendel has been promoted to associate professor without tenure. Wendel’s research engages three main areas: forms of community repair after conflict and disaster, African urbanism, and spatial politics. Her interdisciplinary work draws together urban studies, critical peace studies, architectural history, cultural geography, and anthropology. At MIT DUSP, she leads the Planning for Peace critical collective and oversees the Mellon Foundation and the MIT Center for Art, Science and Technology-funded research and exhibition project, Memory Atlas for Repair. She also serves as the managing editor of Projections, the department’s annual peer-reviewed journal on critical issues in urban studies and planning.
Program in Media Arts and Sciences
Deblina Sarkar has been promoted to associate professor without tenure. As the director of the Nano-Cybernetic Biotrek Lab at the MIT Media Lab, she merges nanoelectronics, physics, and biology to create groundbreaking technologies, from ultra-thin quantum transistors to the first antenna that operates inside living cells. Her interdisciplinary work has earned her major honors, including the National Institutes of Health Director’s New Innovator Award and the IEEE Early Career Award in Nanotechnology.
“Encryption Backdoors and the Fourth Amendment”
Law journal article that looks at the Dual_EC_PRNG backdoor from a US constitutional perspective:
Abstract: The National Security Agency (NSA) reportedly paid and pressured technology companies to trick their customers into using vulnerable encryption products. This Article examines whether any of three theories removed the Fourth Amendment’s requirement that this be reasonable. The first is that a challenge to the encryption backdoor might fail for want of a search or seizure. The Article rejects this both because the Amendment reaches some vulnerabilities apart from the searches and seizures they enable and because the creation of this vulnerability was itself a search or seizure. The second is that the role of the technology companies might have brought this backdoor within the private-search doctrine. The Article criticizes the doctrine particularly its origins in Burdeau v. McDowelland argues that if it ever should apply, it should not here. The last is that the customers might have waived their Fourth Amendment rights under the third-party doctrine. The Article rejects this both because the customers were not on notice of the backdoor and because historical understandings of the Amendment would not have tolerated it. The Article concludes that none of these theories removed the Amendment’s reasonableness requirement...
Trump is escalating his attacks on wind, solar
House Republicans take aim at DC climate lawsuit
Maryland elevates environmental justice at state agencies
Advocacy groups form coalition to oppose science cuts
Maryland regulators, EPA spar over offshore wind permit
EU budget plan would deal ‘devastating blow’ to nature
Heavy rains in South Korea leave 17 dead, 11 others missing
Iran’s central government to shut down due to heat wave
Researchers worry for bees on a warming Earth
Heatwaves disrupt prey behaviour
Nature Climate Change, Published online: 22 July 2025; doi:10.1038/s41558-025-02393-z
Sublethal impacts of heat on reproductive outcomes are beginning to be considered as important drivers of population persistence under climate change. Now, research shows that the impact of transient heat on antipredator behaviours may be an underappreciated source of variation that could have far-reaching implications for survival.EFF to Court: Protect Our Health Data from DHS
The federal government is trying to use Medicaid data to identify and deport immigrants. So EFF and our friends at EPIC and the Protect Democracy Project have filed an amicus brief asking a judge to block this dangerous violation of federal data privacy laws.
Last month, the AP reported that the U.S. Department of Health and Human Services (HHS) had disclosed to the U.S. Department of Homeland Security (DHS) a vast trove of sensitive data obtained from states about people who obtain government-assisted health care. Medicaid is a federal program that funds health insurance for low-income people; it is partially funded and primarily managed by states. Some states, using their own funds, allow enrollment by non-citizens. HHS reportedly disclosed to DHS the Medicaid enrollee data from several of these states, including enrollee names, addresses, immigration status, and claims for health coverage.
In response, California and 19 other states sued HHS and DHS. The states allege, among other things, that these federal agencies violated (1) the data disclosure limits in the Social Security Act, the Privacy Act, and HIPAA, and (2) the notice-and-comment requirements for rulemaking under the Administrative Procedure Act (APA).
Our amicus brief argues that (1) disclosure of sensitive Medicaid data causes a severe privacy harm to the enrolled individuals, (2) the APA empowers federal courts to block unlawful disclosure of personal data between federal agencies, and (3) the broader public is harmed by these agencies’ lack of transparency about these radical changes in data governance.
A new agency agreement, recently reported by the AP, allows Immigration and Customs Enforcement (ICE) to access the personal data of Medicaid enrollees held by HHS’ Centers for Medicare and Medicaid Services (CMS). The agreement states: “ICE will use the CMS data to allow ICE to receive identity and location information on aliens identified by ICE.”
In the 1970s, in the wake of the Watergate and COINTELPRO scandals, Congress wisely enacted numerous laws to protect our data privacy from government misuse. This includes strict legal limits on disclosure of personal data within an agency, or from one agency to another. EFF sued over DOGE agents grabbing personal data from the U.S. Office of Personnel Management, and filed an amicus brief in a suit challenging ICE grabbing taxpayer data. We’ve also reported on the U.S. Department of Agriculture’s grab of food stamp data and DHS’s potential grab of postal data. And we’ve written about the dangers of consolidating all government information.
We have data protection rules for good reason, and these latest data grabs are exactly why.
You can read our new amicus brief here.
A new way to edit or generate images
AI image generation — which relies on neural networks to create new images from a variety of inputs, including text prompts — is projected to become a billion-dollar industry by the end of this decade. Even with today’s technology, if you wanted to make a fanciful picture of, say, a friend planting a flag on Mars or heedlessly flying into a black hole, it could take less than a second. However, before they can perform tasks like that, image generators are commonly trained on massive datasets containing millions of images that are often paired with associated text. Training these generative models can be an arduous chore that takes weeks or months, consuming vast computational resources in the process.
But what if it were possible to generate images through AI methods without using a generator at all? That real possibility, along with other intriguing ideas, was described in a research paper presented at the International Conference on Machine Learning (ICML 2025), which was held in Vancouver, British Columbia, earlier this summer. The paper, describing novel techniques for manipulating and generating images, was written by Lukas Lao Beyer, a graduate student researcher in MIT’s Laboratory for Information and Decision Systems (LIDS); Tianhong Li, a postdoc at MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL); Xinlei Chen of Facebook AI Research; Sertac Karaman, an MIT professor of aeronautics and astronautics and the director of LIDS; and Kaiming He, an MIT associate professor of electrical engineering and computer science.
This group effort had its origins in a class project for a graduate seminar on deep generative models that Lao Beyer took last fall. In conversations during the semester, it became apparent to both Lao Beyer and He, who taught the seminar, that this research had real potential, which went far beyond the confines of a typical homework assignment. Other collaborators were soon brought into the endeavor.
The starting point for Lao Beyer’s inquiry was a June 2024 paper, written by researchers from the Technical University of Munich and the Chinese company ByteDance, which introduced a new way of representing visual information called a one-dimensional tokenizer. With this device, which is also a kind of neural network, a 256x256-pixel image can be translated into a sequence of just 32 numbers, called tokens. “I wanted to understand how such a high level of compression could be achieved, and what the tokens themselves actually represented,” says Lao Beyer.
The previous generation of tokenizers would typically break up the same image into an array of 16x16 tokens — with each token encapsulating information, in highly condensed form, that corresponds to a specific portion of the original image. The new 1D tokenizers can encode an image more efficiently, using far fewer tokens overall, and these tokens are able to capture information about the entire image, not just a single quadrant. Each of these tokens, moreover, is a 12-digit number consisting of 1s and 0s, allowing for 212 (or about 4,000) possibilities altogether. “It’s like a vocabulary of 4,000 words that makes up an abstract, hidden language spoken by the computer,” He explains. “It’s not like a human language, but we can still try to find out what it means.”
That’s exactly what Lao Beyer had initially set out to explore — work that provided the seed for the ICML 2025 paper. The approach he took was pretty straightforward. If you want to find out what a particular token does, Lao Beyer says, “you can just take it out, swap in some random value, and see if there is a recognizable change in the output.” Replacing one token, he found, changes the image quality, turning a low-resolution image into a high-resolution image or vice versa. Another token affected the blurriness in the background, while another still influenced the brightness. He also found a token that’s related to the “pose,” meaning that, in the image of a robin, for instance, the bird’s head might shift from right to left.
“This was a never-before-seen result, as no one had observed visually identifiable changes from manipulating tokens,” Lao Beyer says. The finding raised the possibility of a new approach to editing images. And the MIT group has shown, in fact, how this process can be streamlined and automated, so that tokens don’t have to be modified by hand, one at a time.
He and his colleagues achieved an even more consequential result involving image generation. A system capable of generating images normally requires a tokenizer, which compresses and encodes visual data, along with a generator that can combine and arrange these compact representations in order to create novel images. The MIT researchers found a way to create images without using a generator at all. Their new approach makes use of a 1D tokenizer and a so-called detokenizer (also known as a decoder), which can reconstruct an image from a string of tokens. However, with guidance provided by an off-the-shelf neural network called CLIP — which cannot generate images on its own, but can measure how well a given image matches a certain text prompt — the team was able to convert an image of a red panda, for example, into a tiger. In addition, they could create images of a tiger, or any other desired form, starting completely from scratch — from a situation in which all the tokens are initially assigned random values (and then iteratively tweaked so that the reconstructed image increasingly matches the desired text prompt).
The group demonstrated that with this same setup — relying on a tokenizer and detokenizer, but no generator — they could also do “inpainting,” which means filling in parts of images that had somehow been blotted out. Avoiding the use of a generator for certain tasks could lead to a significant reduction in computational costs because generators, as mentioned, normally require extensive training.
What might seem odd about this team’s contributions, He explains, “is that we didn’t invent anything new. We didn’t invent a 1D tokenizer, and we didn’t invent the CLIP model, either. But we did discover that new capabilities can arise when you put all these pieces together.”
“This work redefines the role of tokenizers,” comments Saining Xie, a computer scientist at New York University. “It shows that image tokenizers — tools usually used just to compress images — can actually do a lot more. The fact that a simple (but highly compressed) 1D tokenizer can handle tasks like inpainting or text-guided editing, without needing to train a full-blown generative model, is pretty surprising.”
Zhuang Liu of Princeton University agrees, saying that the work of the MIT group “shows that we can generate and manipulate the images in a way that is much easier than we previously thought. Basically, it demonstrates that image generation can be a byproduct of a very effective image compressor, potentially reducing the cost of generating images several-fold.”
There could be many applications outside the field of computer vision, Karaman suggests. “For instance, we could consider tokenizing the actions of robots or self-driving cars in the same way, which may rapidly broaden the impact of this work.”
Lao Beyer is thinking along similar lines, noting that the extreme amount of compression afforded by 1D tokenizers allows you to do “some amazing things,” which could be applied to other fields. For example, in the area of self-driving cars, which is one of his research interests, the tokens could represent, instead of images, the different routes that a vehicle might take.
Xie is also intrigued by the applications that may come from these innovative ideas. “There are some really cool use cases this could unlock,” he says.
MIT Learn offers “a whole new front door to the Institute”
In 2001, MIT became the first higher education institution to provide educational resources for free to anyone in the world. Fast forward 24 years: The Institute has now launched a dynamic AI-enabled website for its non-degree learning opportunities, making it easier for learners around the world to discover the courses and resources available on MIT’s various learning platforms.
MIT Learn enables learners to access more than 12,700 educational resources — including introductory and advanced courses, courseware, videos, podcasts, and more — from departments across the Institute. MIT Learn is designed to seamlessly connect the existing Institute’s learning platforms in one place.
“With MIT Learn, we’re opening access to MIT’s digital learning opportunities for millions around the world,” says Dimitris Bertsimas, vice provost for open learning. “MIT Learn elevates learning with personalized recommendations powered by AI, guiding each learner toward deeper understanding. It is a stepping stone toward a broader vision of making these opportunities even more accessible to global learners through one unified learning platform.”
The goal for MIT Learn is twofold: to allow learners to find what they want to fulfill their curiosity, and to enable learners to develop a long-term relationship with MIT as a source of educational experiences.
“By fostering long-term connections between learners and MIT, we not only provide a pathway to continued learning, but also advance MIT’s mission to disseminate knowledge globally,” says Ferdi Alimadhi, chief technology officer for MIT Open Learning and the lead of the MIT Learn project. “With this initial launch of MIT Learn, we’re introducing AI-powered features that leverage emerging technologies to help learners discover the right content, engage with it more deeply, and stay supported as they shape their own educational journeys.”
With its sophisticated search, browse, and discovery capability, MIT Learn allows learners to explore topics without having to understand MIT’s organizational structure or know the names of departments and programs. An AI-powered recommendation feature called “Ask Tim” complements the site’s traditional search and browsing tools, helping learners quickly find courses and resources aligned with their personal and professional goals. Learners can also prompt “Ask Tim” for a summary of a course’s structure, topics, and expectations, leading to more-informed decisions before enrolling.
In select offerings, such as Molecular Biology: DNA Replication and Repair, Genetics: The Fundamentals, and Cell Biology: Transport and Signaling, learners can interact with an AI assistant by asking questions about a lecture, requesting flashcards of key concepts, and obtaining instant summaries. These select offerings also feature an AI tutor to support learners as they work through problem sets, guiding them toward the next step without giving away the answers. These features, Alimadhi says, are being introduced in a limited set of courses and modules to allow the MIT Open Learning team to gather insights and improve the learning experience before expanding more broadly.
“MIT Learn is a whole new front door to the Institute,” says Christopher Capozzola, senior associate dean for open learning, who worked with faculty across the Institute on the project. “Just as the Kendall Square renovations transformed the way that people interact with our physical campus, MIT Learn transforms how people engage with what we offer digitally.”
Learners who choose to create an account on MIT Learn receive personalized course recommendations and can create and curate lists of educational resources, follow their specific areas of interest, and receive notifications when new MIT content is available. They can also personalize their learning experience based on their specific interests and choose the format that is best suited to them.
"From anywhere and for anyone, MIT Learn makes lifelong learning more accessible and personalized, building on the Institute’s decades of global leadership in open learning,” says MIT Provost Anantha Chandrakasan.
MIT Learn was designed to account for a learner’s evolving needs throughout their learning journey. It highlights supplemental study materials for middle schoolers, high schoolers, and college students, upskilling opportunities for early-career professionals, reskilling programs for those considering a career shift, and resources for educators.
“MIT has an amazing collection of learning opportunities, covering a wide range of formats,” says Eric Grimson, chancellor for academic advancement, who oversaw the initial development of MIT Learn during his time as interim vice president for open learning. “The sheer size of that collection can be daunting, so creating a platform that brings all of those offerings together, in an easily searchable framework, greatly enhances our ability to serve learners.”
According to Peter Hirst, senior associate dean for executive education at MIT Sloan School of Management, one of the Institute's incredible strengths is its sheer volume and diversity of expertise, research, and learning opportunities. But it can be challenging to discover and follow all those opportunities — even for people who are immersed in the on-campus experience. MIT Learn, he says, is a solution to this problem.
“MIT Learn gathers all the knowledge and learning resources offered across all of MIT into a learner-friendly, curatable repository that enables anyone and everyone, whatever their interests or learning needs, to explore and engage in the wide range of learning resources and public certificate programs that MIT has to offer and that can help them achieve their goals,” Hirst says.
MIT Learn was spearheaded by MIT Open Learning, which aims to transform teaching and learning on and off the Institute’s campus. MIT Learn was developed with the direction of former provost Cynthia Barnhart, and in cooperation with Sloan Executive Education and Professional Education. During the design phase, OpenCourseWare Faculty Advisory Committee Chair Michael Short and MITx Faculty Advisory Committee Chair Caspar Hare contributed key insights, along with other numerous faculty involved with Open Learning’s product offerings, including OpenCourseWare, MITx, and MicroMasters programs. MIT Learn is also informed by the insights of the Ad Hoc Committee on MITx and MITx Online.
“For over 20 years, MIT staff and faculty have been creating a wealth of online resources, from lecture videos to practice problems, and from single online courses to entire credential-earning programs,” says Sara Fisher Ellison, a member of the Ad Hoc Committee on MITx and MITx Online and the faculty lead for the online MITx MicroMasters Program in Data, Economics, and Design of Policy. “Making these resources findable, searchable, and broadly available is a natural extension of MIT’s core educational mission. MIT Learn is a big, important step in that direction. We are excited for the world to see what we have to offer.”
Looking ahead, MIT Learn will also feature selected content from the MIT Press. As MIT Learn continues to grow, Open Learning is exploring collaborations with departments across the Institute with the goal of offering the fullest possible range of educational materials from MIT to learners around the world.
“MIT Learn is the latest step in a long tradition of the Institute providing innovative ways for learners to access knowledge,” Barnhart says. “This AI-enabled platform delivers on the Institute’s commitment to help people launch into learning journeys that can unlock life-changing opportunities.”
Dating Apps Need to Learn How Consent Works
Staying safe whilst dating online should not be the responsibility of users—dating apps should be prioritizing our privacy by default, and laws should require companies to prioritize user privacy over their profit. But dating apps are taking shortcuts in safeguarding the privacy and security of users in favour of developing and deploying AI tools on their platforms, sometimes by using your most personal information to train their AI tools.
Grindr has big plans for its gay wingman bot, Bumble launched AI Icebreakers, Tinder introduced AI tools to choose profile pictures for users, OKCupid teamed up with AI photo editing platform Photoroom to erase your ex from profile photos, and Hinge recently launched an AI tool to help users write prompts.
The list goes on, and the privacy harms are significant. Dating apps have built platforms that encourage people to be exceptionally open with sensitive and potentially dangerous personal information. But at the same time, the companies behind the platforms collect vast amounts of intimate details about their customers—everything from sexual preferences to precise location—who are often just searching for compatibility and connection. This data falling into the wrong hands can—and has—come with unacceptable consequences, especially for members of the LGBTQ+ community.
This is why corporations should provide opt-in consent for AI training data obtained through channels like private messages, and employ minimization practices for all other data. Dating app users deserve the right to privacy, and should have a reasonable expectation that the contents of conversations—from text messages to private pictures—are not going to be shared or used for any purpose that opt-in consent has not been provided for. This includes the use of personal data for building AI tools, such as chatbots and picture selection tools.
AI IcebreakersBack in December 2023, Bumble introduced AI Icebreakers to the ‘Bumble for Friends’ section of the app to help users start conversations by providing them with AI-generated messages. Powered by OpenAI’s ChatGPT, the feature was deployed in the app without ever asking for their consent. Instead, the company presented users with a pop-up upon entering the app which repeatedly nudged people to click ‘Okay’ or face the same pop-up every time the app is reopened until individuals finally relent and tap ‘Okay.’
Obtaining user data without explicit opt-in consent is bad enough. But Bumble has taken this even further by sharing personal user data from its platform with OpenAI to feed into the company’s AI systems. By doing this, Bumble has forced its AI feature on millions of users in Europe—without their consent but with their personal data.
In response, European nonprofit noyb recently filed a complaint with the Austrian data protection authority on Bumble’s violation of its transparency obligations under Article 5(1)(a) GDPR. In its report, noyb flagged concerns around Bumble’s data sharing with OpenAI, which allowed the company to generate an opening message based on information users shared on the app.
In its complaint, noyb specifically alleges that Bumble:
- Failed to provide information about the processing of personal data for its AI Icebreaker feature
- Confused users with a “fake” consent banner
- Lacks a legal basis under Article 6(1) GDPR as it never sought user consent and cannot legally claim to base its processing on legitimate interest
- Can only process sensitive data—such as data involving sexual orientation—with explicit consent per Article 9 GDPR
- Failed to adequately respond to the complainant’s access request, regulated through Article 15 GDPR.
Grindr recently launched its AI wingman. The feature operates like a chatbot and currently keeps track of favorite matches and suggests date locations. In the coming years, Grindr plans for the chatbot to send messages to other AI agents on behalf of users, and make restaurant reservations—all without human intervention. This might sound great: online dating without the time investment? A win for some! But privacy concerns remain.
The chatbot is being built in collaboration with a third party company called Ex-human, which raises concerns about data sharing. Grindr has communicated that its users’ personal data will remain on its own infrastructure, which Ex-Human does not have access to, and that users will be “notified” when AI tools are available on the app. The company also said that it will ask users for permission to use their chat history for AI training. But AI data poses privacy risks that do not seem fully accounted for, particularly in places where it’s not safe to be outwardly gay.
In building this ‘gay chatbot,’ Grindr’s CEO said one of its biggest limitations was preserving user privacy. It’s good that they are cognizant of these harms, particularly because the company has a terrible track record of protecting user privacy, and the company was also recently sued for allegedly revealing the HIV status of users. Further, direct messages on Grindr are stored on the company’s servers, where you have to trust they will be secured, respected, and not used to train AI models without your consent. Given Grindr’s poor record of not respecting user consent and autonomy on the platform, users need additional protections and guardrails for their personal data and privacy than currently being provided—especially for AI tools that are being built by third parties.
AI Picture SelectionIn the past year, Tinder and Bumble have both introduced AI tools to help users choose better pictures for their profiles. Tinder’s AI-powered feature, Photo Selector, requires users to upload a selfie, after which its facial recognition technology can identify the person in their camera roll images. The Photo Selector then chooses a “curated selection of photos” direct from users’ devices based on Tinder’s “learnings” about good profile images. Users are not informed about the parameters behind choosing photos, nor is there a separate privacy policy introduced to guardrail privacy issues relating to the potential collection of biometric data, and collection, storage, and sale of camera roll images.
The Way Forward: Opt-In Consent for AI Tools and Consumer Privacy LegislationPutting users in control of their own data is fundamental to protecting individual and collective privacy. We all deserve the right to control how our data is used and by whom. And when it comes to data like profile photos and private messages, all companies should require opt-in consent before processing those messages for AI. Finding love should not involve such a privacy impinging tradeoff.
At EFF, we’ve also long advocated for the introduction of comprehensive consumer privacy legislation to limit the collection of our personal data at its source and prevent retained data being sold or given away, breached by hackers, disclosed to law enforcement, or used to manipulate a user’s choices through online behavioral advertising. This would help protect users on dating apps as reducing the amount of data collected prevents the subsequent use in ways like building AI tools and training AI models.
The privacy options at our disposal may seem inadequate to meet the difficult moments ahead of us, especially for vulnerable communities, but these steps are essential to protecting users on dating apps. We urge companies to put people over profit and protect privacy on their platforms.
When Your Power Meter Becomes a Tool of Mass Surveillance
Simply using extra electricity to power some Christmas lights or a big fish tank shouldn’t bring the police to your door. In fact, in California, the law explicitly protects the privacy of power customers, prohibiting public utilities from disclosing precise “smart” meter data in most cases.
Despite this, Sacramento’s power company and law enforcement agencies have been running an illegal mass surveillance scheme for years, using our power meters as home-mounted spies. The Electronic Frontier Foundation (EFF) is seeking to end Sacramento’s dragnet surveillance of energy customers and have asked for a court order to stop this practice for good.
For a decade, the Sacramento Municipal Utilities District (SMUD) has been searching through all of its customers’ energy data, and passed on more than 33,000 tips about supposedly “high” usage households to police. Ostensibly looking for homes that were growing illegal amounts of cannabis, SMUD analysts have admitted that such “high” power usage could come from houses using air conditioning or heat pumps or just being large. And the threshold of so-called “suspicion” has steadily dropped, from 7,000 kWh per month in 2014 to just 2,800 kWh a month in 2023. One SMUD analyst admitted that they themselves “used 3500 [kWh] last month.”
This scheme has targeted Asian customers. SMUD analysts deemed one home suspicious because it was “4k [kWh], Asian,” and another suspicious because “multiple Asians have reported there.” Sacramento police sent accusatory letters in English and Chinese, but no other language, to residents who used above-average amounts of electricity.
In 2022, EFF and the law firm Vallejo, Antolin, Agarwal, Kanter LLP sued SMUD and the City of Sacramento, representing the Asian American Liberation Network and two Sacramento County residents. One is an immigrant from Vietnam. Sheriff’s deputies showed up unannounced at his home, falsely accused him of growing cannabis based on an erroneous SMUD tip, demanded entry for a search, and threatened him with arrest when he refused. He has never grown cannabis; rather, he consumes more than average electricity due to a spinal injury.
Last week, we filed our main brief explaining how this surveillance program violates the law and why it must be stopped. California’s state constitution bars unreasonable searches. This type of dragnet surveillance — suspicionless searches of entire zip codes worth of customer energy data — is inherently unreasonable. Additionally, a state statute generally prohibits public utilities from sharing such data. As we write in our brief, the Sacramento’s mass surveillance scheme does not qualify for one of the narrow exceptions to this rule.
Mass surveillance violates the privacy of many individuals, as police without individualized suspicion seek (possibly non-existent) evidence of some kind of offense by some unknown person. As we’ve seen time and time again, innocent people inevitably get caught in the dragnet. For decades, EFF has been exposing and fighting these kinds of dangerous schemes. We remain committed to protecting digital privacy, whether it’s being threatened by national governments – or your local power company.
Related Cases: Asian American Liberation Network v. SMUD, et al.The unique, mathematical shortcuts language models use to predict dynamic scenarios
Let’s say you’re reading a story, or playing a game of chess. You may not have noticed, but each step of the way, your mind kept track of how the situation (or “state of the world”) was changing. You can imagine this as a sort of sequence of events list, which we use to update our prediction of what will happen next.
Language models like ChatGPT also track changes inside their own “mind” when finishing off a block of code or anticipating what you’ll write next. They typically make educated guesses using transformers — internal architectures that help the models understand sequential data — but the systems are sometimes incorrect because of flawed thinking patterns. Identifying and tweaking these underlying mechanisms helps language models become more reliable prognosticators, especially with more dynamic tasks like forecasting weather and financial markets.
But do these AI systems process developing situations like we do? A new paper from researchers in MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) and Department of Electrical Engineering and Computer Science shows that the models instead use clever mathematical shortcuts between each progressive step in a sequence, eventually making reasonable predictions. The team made this observation by going under the hood of language models, evaluating how closely they could keep track of objects that change position rapidly. Their findings show that engineers can control when language models use particular workarounds as a way to improve the systems’ predictive capabilities.
Shell games
The researchers analyzed the inner workings of these models using a clever experiment reminiscent of a classic concentration game. Ever had to guess the final location of an object after it’s placed under a cup and shuffled with identical containers? The team used a similar test, where the model guessed the final arrangement of particular digits (also called a permutation). The models were given a starting sequence, such as “42135,” and instructions about when and where to move each digit, like moving the “4” to the third position and onward, without knowing the final result.
In these experiments, transformer-based models gradually learned to predict the correct final arrangements. Instead of shuffling the digits based on the instructions they were given, though, the systems aggregated information between successive states (or individual steps within the sequence) and calculated the final permutation.
One go-to pattern the team observed, called the “Associative Algorithm,” essentially organizes nearby steps into groups and then calculates a final guess. You can think of this process as being structured like a tree, where the initial numerical arrangement is the “root.” As you move up the tree, adjacent steps are grouped into different branches and multiplied together. At the top of the tree is the final combination of numbers, computed by multiplying each resulting sequence on the branches together.
The other way language models guessed the final permutation was through a crafty mechanism called the “Parity-Associative Algorithm,” which essentially whittles down options before grouping them. It determines whether the final arrangement is the result of an even or odd number of rearrangements of individual digits. Then, the mechanism groups adjacent sequences from different steps before multiplying them, just like the Associative Algorithm.
“These behaviors tell us that transformers perform simulation by associative scan. Instead of following state changes step-by-step, the models organize them into hierarchies,” says MIT PhD student and CSAIL affiliate Belinda Li SM ’23, a lead author on the paper. “How do we encourage transformers to learn better state tracking? Instead of imposing that these systems form inferences about data in a human-like, sequential way, perhaps we should cater to the approaches they naturally use when tracking state changes.”
“One avenue of research has been to expand test-time computing along the depth dimension, rather than the token dimension — by increasing the number of transformer layers rather than the number of chain-of-thought tokens during test-time reasoning,” adds Li. “Our work suggests that this approach would allow transformers to build deeper reasoning trees.”
Through the looking glass
Li and her co-authors observed how the Associative and Parity-Associative algorithms worked using tools that allowed them to peer inside the “mind” of language models.
They first used a method called “probing,” which shows what information flows through an AI system. Imagine you could look into a model’s brain to see its thoughts at a specific moment — in a similar way, the technique maps out the system’s mid-experiment predictions about the final arrangement of digits.
A tool called “activation patching” was then used to show where the language model processes changes to a situation. It involves meddling with some of the system’s “ideas,” injecting incorrect information into certain parts of the network while keeping other parts constant, and seeing how the system will adjust its predictions.
These tools revealed when the algorithms would make errors and when the systems “figured out” how to correctly guess the final permutations. They observed that the Associative Algorithm learned faster than the Parity-Associative Algorithm, while also performing better on longer sequences. Li attributes the latter’s difficulties with more elaborate instructions to an over-reliance on heuristics (or rules that allow us to compute a reasonable solution fast) to predict permutations.
“We’ve found that when language models use a heuristic early on in training, they’ll start to build these tricks into their mechanisms,” says Li. “However, those models tend to generalize worse than ones that don’t rely on heuristics. We found that certain pre-training objectives can deter or encourage these patterns, so in the future, we may look to design techniques that discourage models from picking up bad habits.”
The researchers note that their experiments were done on small-scale language models fine-tuned on synthetic data, but found the model size had little effect on the results. This suggests that fine-tuning larger language models, like GPT 4.1, would likely yield similar results. The team plans to examine their hypotheses more closely by testing language models of different sizes that haven’t been fine-tuned, evaluating their performance on dynamic real-world tasks such as tracking code and following how stories evolve.
Harvard University postdoc Keyon Vafa, who was not involved in the paper, says that the researchers’ findings could create opportunities to advance language models. “Many uses of large language models rely on tracking state: anything from providing recipes to writing code to keeping track of details in a conversation,” he says. “This paper makes significant progress in understanding how language models perform these tasks. This progress provides us with interesting insights into what language models are doing and offers promising new strategies for improving them.”
Li wrote the paper with MIT undergraduate student Zifan “Carl” Guo and senior author Jacob Andreas, who is an MIT associate professor of electrical engineering and computer science and CSAIL principal investigator. Their research was supported, in part, by Open Philanthropy, the MIT Quest for Intelligence, the National Science Foundation, the Clare Boothe Luce Program for Women in STEM, and a Sloan Research Fellowship.
The researchers presented their research at the International Conference on Machine Learning (ICML) this week.
Another Supply Chain Vulnerability
ProPublica is reporting:
Microsoft is using engineers in China to help maintain the Defense Department’s computer systems—with minimal supervision by U.S. personnel—leaving some of the nation’s most sensitive data vulnerable to hacking from its leading cyber adversary, a ProPublica investigation has found.
The arrangement, which was critical to Microsoft winning the federal government’s cloud computing business a decade ago, relies on U.S. citizens with security clearances to oversee the work and serve as a barrier against espionage and sabotage...