MIT CSAIL Principal Research Scientist Una-May O’Reilly discusses how she develops agents that reveal AI models’ security weaknesses before hackers do.
Starting with a single frame in a simulation, a new system uses generative AI to emulate the dynamics of molecules, connecting static molecular structures and developing blurry pictures into videos.
Biodiversity researchers tested vision systems on how well they could retrieve relevant nature images. More advanced models performed well on simple queries but struggled with more research-specific prompts.
When you ask a robot to place blocks in matching bowls, for example, MIT’s “PRoC3S” method helps an LLM create a viable action plan by testing each step in a simulation. This strategy could eventually aid in-home robots to complete more ambiguous chore requests.
Daniela Rus, Director of CSAIL and MIT EECS Professor, recently received the 2025 IEEE Edison Medal. The award recognizes her leadership and pioneering work in modern robotics.
Inspired by the mechanics of the human vocal tract, MIT CSAIL researchers’ AI model can produce and understand vocal imitations of everyday sounds. Their method could help build new sonic interfaces for entertainment and education.
A new design tool uses UV and RGB lights to change the color and textures of everyday objects. The system could enable surfaces to display dynamic patterns, such as health data and fashion designs.
“MouthIO” is an in-mouth device that users can digitally design and 3D print with integrated sensors and actuators to capture health data and interact with a computer or phone.
A new method can train a neural network to sort corrupted data while anticipating next steps. It can make flexible plans for robots, generate high-quality video, and help AI agents navigate digital environments.
“Co-LLM” algorithm helps a general-purpose AI model collaborate with an expert large language model by combining the best parts of both answers, leading to more factual responses.
“ScribblePrompt” is an interactive AI framework that can efficiently highlight anatomical structures across different medical scans, assisting medical workers to delineate regions of interest and abnormalities.
A new algorithm solves complicated partial differential equations by breaking them down into simpler problems, potentially guiding computer graphics and geometry processing.
An AI team coordinator aligns agents’ beliefs about how to achieve a task, intervening when necessary to potentially help with tasks in search and rescue, hospitals, and video games.
In controlled experiments, MIT CSAIL researchers discover simulations of reality developing deep within LLMs, indicating an understanding of language beyond simple mimicry.
New algorithm helps robots practice skills like sweeping and placing objects, potentially helping them improve at important tasks in houses, hospitals, and factories.
Neural network controllers provide complex robots with stability guarantees, paving the way for the safer deployment of autonomous vehicles and industrial machines.
LLMs trained primarily on text can generate complex visual concepts through code with self-correction. Researchers used these illustrations to train an image-free computer vision system to recognize real photos.
MIT CSAIL’s frugal deep-learning model infers the hidden physical properties of objects, then adapts to find the most stable grasps for robots in unstructured environments like homes and fulfillment centers.
“Alchemist” system adjusts the material attributes of specific objects within images to potentially modify video game models to fit different environments, fine-tune VFX, and diversify robotic training.
Three neurosymbolic methods help language models find better abstractions within natural language, then use those representations to execute complex tasks.
A CSAIL study highlights why it is so challenging to program a quantum computer to run a quantum algorithm, and offers a conceptual model for a more user-friendly quantum computer.
Adaptive smart glove from MIT CSAIL researchers can send tactile feedback to teach users new skills, guide robots with more precise manipulation, and help train surgeons and pilots.
The ambient light sensors responsible for smart devices’ brightness adjustments can capture images of touch interactions like swiping and tapping for hackers.
A multimodal system uses models trained on language, vision, and action data to help robots develop and execute plans for household, construction, and manufacturing tasks.
MIT CSAIL researchers established new connections between combinatorial and continuous optimization, which can find global solutions for complex motion planning puzzles.
By blending 2D images with foundation models to build 3D feature fields, a new MIT method helps robots understand and manipulate nearby objects with open-ended language prompts.
“Lightning” system connects photons to the electronic components of computers using a novel abstraction, creating the first photonic computing prototype to serve real-time machine-learning inference requests.
Developed by MIT researchers, BrightMarkers are invisible fluorescent tags embedded in physical objects to enhance motion tracking, virtual reality, and object detection.
FlexBoard is a flexible breadboard that enables rapid prototyping of objects with interactive sensors, actuators, and displays on curved and deformable surfaces.
The Association for Computing Machinery (ACM) recently awarded Yael Tauman Kalai, MIT Department of Electrical Engineering and Computer Science (EECS) adjunct professor, CSAIL member, and Senior Principal Researcher at Microsoft Research with the 2022 ACM Prize in Computing for her cryptography research.
Passionate about creating educational opportunities in India, PhD student Siddhartha Jayanti recently explored multiprocessor speed limits, in a paper written in the Indian language Telugu.