We aim to develop the science of autonomy toward a future with robots and AI systems integrated into everyday life, supporting people with cognitive and physical tasks.
Our projects are centered around the problems of navigation and mapping for autonomous mobile robots operating in underwater and terrestrial environments.
Our researchers create state-of-the-art systems to better recognize objects, people, scenes, behaviors and more, with applications in health-care, gaming, tagging systems and more.
Our mission is to work with policy makers and cybersecurity technologists to increase the trustworthiness and effectiveness of interconnected digital systems.
Our goal is to create an online risk-aware planner for vehicle maneuvers that can make driving safer and less stressful through a “parallel” autonomous system that assists the driver by watching for risky situations, and by helping the driver take proactive, compensating actions before they become crises.
Our goal is to understand what drivers can perceive when they are not attending to the road ahead and the consequences of their perceptions for safety.
Our goal is to allow planners to exploit the forces and contacts between objects so as to carry out complex manipulation tasks in the presence of uncertainty.
Our goal is to design algorithms that enable robots to operate in human environments by simultaneously reasoning about high-level task actions as well as low-level motions.
In this project, we aim to develop a framework that can ensure and certify the safety of an autonomous vehicle. By leveraging research from the area of formal verification, this framework aims to assess the safety, i.e., free of collisions, of a broad class of autonomous car controllers/planners for a given traffic model.
Our goal is to create a theoretical framework and effective machine learning algorithms for robust, reliable control of autonomous vehicles. Key threads include developing metrics of confidence; and designing deep learning algorithms for parallel autonomy.
We are developing a general framework that enforces privacy transparently enabling different kinds of machine learning to be developed that are automatically privacy preserving.
Developed at MIT’s Computer Science and Artificial Intelligence Laboratory, a team of robots can self-assemble to form different structures with applications in inspection, disaster response, and manufacturing
Honda Research Institute USA seeks to develop intelligent systems that use curiosity to understand people’s needs and empower human capability through cross-disciplinary research that aims to advance breakthroughs in artificial cognition.
Google AI’s Jeff Dean has a seemingly straightforward objective: he wants to use a collection of trainable mathematical units organized in layers to solve complicated tasks that will ultimately benefit many parts of society.
For all the progress made in self-driving technologies, there still aren’t many places where they can actually drive. Companies like Google only test their fleets in major cities where they’ve spent countless hours meticulously labeling the exact 3-D positions of lanes, curbs, off-ramps, and stop signs.
Artificial intelligence (AI) in the form of “neural networks” are increasingly used in technologies like self-driving cars to be able to see and recognize objects. Such systems could even help with tasks like identifying explosives in airport security lines.
Light lets us see the things that surround us, but what if we could also use it to see things hidden around corners? It sounds like science fiction, but that’s the idea behind a new algorithm out of MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) — and its discovery has implications for everything from emergency response to self-driving cars.
In recent years, a host of Hollywood blockbusters — including “The Fast and the Furious 7,” “Jurassic World,” and “The Wolf of Wall Street” — have included aerial tracking shots provided by drone helicopters outfitted with cameras. Those shots required separate operators for the drones and the cameras, and careful planning to avoid collisions. But a team of researchers from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) and ETH Zurich hope to make drone cinematography more accessible, simple, and reliable.
We aim to develop the science of autonomy toward a future with robots and AI systems integrated into everyday life, supporting people with cognitive and physical tasks.
Our mission is to work with policy makers and cybersecurity technologists to increase the trustworthiness and effectiveness of interconnected digital systems.
Our projects are centered around the problems of navigation and mapping for autonomous mobile robots operating in underwater and terrestrial environments.
Our researchers create state-of-the-art systems to better recognize objects, people, scenes, behaviors and more, with applications in health-care, gaming, tagging systems and more.
Alloy is a language for describing structures and a tool for exploring them. It has been used in a wide range of applications from finding holes in security mechanisms to designing telephone switching networks. Now one of the main challenges is to make the system more usable and understandable for the user.
Self-driving cars are likely to be safer, on average, than human-driven cars. But they may fail in new and catastrophic ways that a human driver could prevent. This project is designing a new architecture for a highly dependable self-driving car.
Our goal is to create an online risk-aware planner for vehicle maneuvers that can make driving safer and less stressful through a “parallel” autonomous system that assists the driver by watching for risky situations, and by helping the driver take proactive, compensating actions before they become crises.
Our goal is to understand what drivers can perceive when they are not attending to the road ahead and the consequences of their perceptions for safety.
We are developing a general framework that enforces privacy transparently enabling different kinds of machine learning to be developed that are automatically privacy preserving.
Our goal is to allow planners to exploit the forces and contacts between objects so as to carry out complex manipulation tasks in the presence of uncertainty.
Our goal is to create a theoretical framework and effective machine learning algorithms for robust, reliable control of autonomous vehicles. Key threads include developing metrics of confidence; and designing deep learning algorithms for parallel autonomy.
In this project, we aim to develop a framework that can ensure and certify the safety of an autonomous vehicle. By leveraging research from the area of formal verification, this framework aims to assess the safety, i.e., free of collisions, of a broad class of autonomous car controllers/planners for a given traffic model.
Developed at MIT’s Computer Science and Artificial Intelligence Laboratory, a team of robots can self-assemble to form different structures with applications in inspection, disaster response, and manufacturing
Honda Research Institute USA seeks to develop intelligent systems that use curiosity to understand people’s needs and empower human capability through cross-disciplinary research that aims to advance breakthroughs in artificial cognition.
Google AI’s Jeff Dean has a seemingly straightforward objective: he wants to use a collection of trainable mathematical units organized in layers to solve complicated tasks that will ultimately benefit many parts of society.
For all the progress made in self-driving technologies, there still aren’t many places where they can actually drive. Companies like Google only test their fleets in major cities where they’ve spent countless hours meticulously labeling the exact 3-D positions of lanes, curbs, off-ramps, and stop signs.
Artificial intelligence (AI) in the form of “neural networks” are increasingly used in technologies like self-driving cars to be able to see and recognize objects. Such systems could even help with tasks like identifying explosives in airport security lines.
Light lets us see the things that surround us, but what if we could also use it to see things hidden around corners? It sounds like science fiction, but that’s the idea behind a new algorithm out of MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) — and its discovery has implications for everything from emergency response to self-driving cars.
In recent years, a host of Hollywood blockbusters — including “The Fast and the Furious 7,” “Jurassic World,” and “The Wolf of Wall Street” — have included aerial tracking shots provided by drone helicopters outfitted with cameras. Those shots required separate operators for the drones and the cameras, and careful planning to avoid collisions. But a team of researchers from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) and ETH Zurich hope to make drone cinematography more accessible, simple, and reliable.