72 results on '"Teller, Seth"'
Search Results
2. Bounded-Error Interactive Ray Tracing
- Author
-
Bala, Kavita, Dorsey, Julie, Teller, Seth, Bala, Kavita, Dorsey, Julie, and Teller, Seth
- Published
- 2023
3. Immediate-Mode Ray-Casting
- Author
-
Alex, John, Teller, Seth, Alex, John, and Teller, Seth
- Abstract
We propose a simple modification to the classical polygon rasterization pipeline that enables exact, efficient raycasting of bounded implicit surfaces without the use of a global spatial data structure bounding hierarchy. Our algorithm requires two descr
- Published
- 2023
4. Matching and Pose Refinement with Camera Pose Estimates
- Author
-
Coorg, Satyan, Teller, Seth, Coorg, Satyan, and Teller, Seth
- Abstract
This paper describes novel algorithms that use absolute camera pose information to identify correspondence among point features in hundreds or thousands of images. Our incidence counting algorithm is a geometric approach to matching: it makes features by extruding them into an absolute 3-D coordinate system, then searching 3-D space for regions into which many features project.
- Published
- 2023
5. Automatic Extraction of Textured Vertical Facades from Pose Imagery
- Author
-
Coorg, Satvan, Teller, Seth, Coorg, Satvan, and Teller, Seth
- Abstract
Extracting 3-dimensional structure from real-world imagery and rendering it from unrestricted viewpoints is an important problem in computer vision, and increasingly, computer graphics. Despite many years of research, a system that automatically recovers
- Published
- 2023
6. Anchor-free Distributed Localization in Sensor Networks
- Author
-
Priyantha, Nissanka B., Balakrishnan, Hari, Demaine, Erik, Teller, Seth, Priyantha, Nissanka B., Balakrishnan, Hari, Demaine, Erik, and Teller, Seth
- Abstract
Many sensor network applications require that each node's sensor stream be annotated with its physical location in some common coordinate system. Manual measurement and configuration methods for obtaining location don't scale and are error-prone, and equipping sensors with GPS is often expensive and does not work in indoor and urban deployments. Sensor networks can therefore benefit from a self-configuring method where nodes cooperate with each other, estimate local distances to their neighbors, and converge to a consistent coordinate assignment. This paper describes a fully decentralized algorithm called AFL (Anchor-Free Localization) where nodes start from a random initial coordinate assignment and converge to a consistent solution using only local node interactions. The key idea in AFL is fold-freedom, where nodes first configure into a topology that resembles a scaled and unfolded version of the true configuration, and then run a force-based relaxation procedure. We show using extensive simulations under a variety of network sizes, node densities, and distance estimation errors that our algorithm is superior to previously proposed methods that incrementally compute the coordinate of nodes in the network, in terms of its ability to computer correct coordinates under a wider variety of conditions and its robuestness to measurement errors.
- Published
- 2023
7. Frustum Casting for Progressive, Interactive Rendering
- Author
-
Teller, Seth, Alex, John, Teller, Seth, and Alex, John
- Abstract
Efficient visible surface determination algorithms have long been a fundamental goal of computer graphics. We discuss the well-known ray casting problem: given a geometric scene description, a synthetic camera, and a viewport which discretizes the camer
- Published
- 2023
8. Automatic Recovery of Camera Positions in Urban Scenes
- Author
-
Antone, Matthew E., Teller, Seth, Antone, Matthew E., and Teller, Seth
- Abstract
Accurate camera calibration is crucial to the reconstruction of three-dimensional geometry and the recovery of photometric scene properties. Calibration involves the determination of intrinsic parameters (e.g. focal length, principal point, and radial lens distortion) and extrinsic parameters (orientation and position). In urban scenes and other environments containing sufficient geometric structure, it is possible to decouple extrinsic calibration into rotational and translational components that can be treated separately, simplifying the registration problem. Here we present such a decoupled formulation and describe methods for automatically recovering the positions of a large set of cameras given intrinsic calibration, relative rotations, and approximate positions. Our algorithm first estimates the directions of translation (up to an unknown scale factor) between adjacent camera pairs using point features but without requiring explicit correspondence between them. This technique combines the robustness and simplicity of a Hough transform with the accuracy of Monte Carlo expectation maximization. We then find a set of distances between the pairs that produces globally-consistent camera positions. Novel uncertainty formulations and match plausibility criteria improve reliability and accuracy. We assess our system's performance using both synthetic data and a large set of real panoramic imagery. The system produces camera positions accurate to within 5 centimeters in image networks extending over hundreds of meters.
- Published
- 2023
9. Conservative Radiance Interpolants for Ray Tracing
- Author
-
Teller, Seth, Bala, Kavita, Dorsey, Julie, Teller, Seth, Bala, Kavita, and Dorsey, Julie
- Abstract
Classical ray-tracing algorithms compute radiance returning to the eye along one or more sample rays through each pixel of an image. The output of a ray-tracing algorithm, although potentially photorealistic, is a two-dimensional quality an image array of radiance values and is not directly useful from any viewpoint other than the one for which it was computed.
- Published
- 2023
10. Polygonal Approximation of Voronoi Diagrams of Set of Triangles in Three Dimensions
- Author
-
Teichmann, Marek, Teller, Seth, Teichmann, Marek, and Teller, Seth
- Published
- 2023
11. A Spacially and Temporally Coherent Object Space Visibility Algorithm
- Author
-
Coorg, Satyan, Teller, Seth, Coorg, Satyan, and Teller, Seth
- Abstract
Efficiently identifying polygons that are visible from a changing synthetic viewpoint is an important problem in computer graphics. In many complex geometric models, most parts of the model are invisible from the instantaneous viewpoint. Despite this, hidden-surface algorithms like the z-buffer or BSP tree often expend significant computation processing invisible portions of the model.
- Published
- 2023
12. Acquisition of a Large Pose-Mosaic Dataset
- Author
-
Coorg, Satyan, Master, Neel, Teller, Seth, Coorg, Satyan, Master, Neel, and Teller, Seth
- Abstract
We describe the generation of a large pose-mosaic dataset: a collection of several thousand digital images, grouped by spatial position into spherical mosaics, each annotated with estimates of the acquiring camera's 6 DOF pose (3 DOF position and 3 DOF orientation) in an absolute coordinate system. The pose-mosaic dataset was generated by acquiring images, grouped by spatial position into nodes (essentially, spherical mosaics). A prototype mechanical pan-tilt head was manually deployed to acquire the data. Manual surverying provided initial position estimates for each node. A back-projecting scheme provided initial rotational estimates. Relative rotations within each node, along with internal camera parameters, were refined automatically by an optimization-correlation scheme. Relative translations and rotations among nodes were refined according to point correspondences, generated automatically and by a human operator. The resulting pose-imagery is self-consistent under a variety of evaluation metrics. Pose-mosaics are useful "first-class" data objects, for example in automatic reconstruction of textured 3D CAD models which represent urban exteriors.
- Published
- 2023
13. Scalable, Controlled Imagery Capture in Urban Environments
- Author
-
Teller, Seth and Teller, Seth
- Abstract
We describe the design considerations underlying a system for scalable, automated capture of precisely controlled imagery in urban scenes. The system operates for architectural scenes in which, from every camera position, some two vanishing points are visible. It has been used to capture thousands of controlled images in outdoor environments spanning hundreds of meters. The proposed system architecture forms the foundation for a future, fully robotic outdoor mapping capability for urban areas, analogous to existing, satellite-based robotic mapping systems which acquire images and models of natural terrain. Four key ideas distinguish our approach from other methods. First, our sensor acquires georeferencing metadata with every image, enabling related images to be efficiently identified and registered. Second, the sensor acquires omni-directional images; we show strong experimental evidence that such images are fundamentally more powerful observations than conventional (narrow-FOV) images. Third, the system uses a probabilistic, projective error formulation to account for uncertainty. By treating measurement error in an appropriate depth-free framework, and by deferring decisions about camera calibration and scene structure until many noisy observations can be fused, the system achieves superior robustness and accuracy. Fourth, the system's computational requirements scale linearly in the input size, the area of the acquisition region, and the size of the output model. This is in contrast to most previous methods, which either assume constant-size inputs or exhibit quadratic running time (or worse) asymptotically. These attributes enable the system to operate in a regime of scale and physical extent which is unachievable by any other method, whether manual or automated. Consequently, it can acquire the most complex calibrated terrestrial image sets in existence, while operating faster thanany existing manual or algorithmic method.
- Published
- 2023
14. Computational Geometry (Dagstuhl Seminar 9707)
- Author
-
Rolf Klein and Raimund Seidel and Seth Teller, Klein, Rolf, Seidel, Raimund, Teller, Seth, Rolf Klein and Raimund Seidel and Seth Teller, Klein, Rolf, Seidel, Raimund, and Teller, Seth
- Published
- 2021
- Full Text
- View/download PDF
15. Inferring Maps and Behaviors from Natural Language Instructions
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Walter, Matthew Robert, Howard, Thomas M., Hemachandra, Sachithra Madhawa, Teller, Seth, Roy, Nicholas, Duvallet, Felix, Oh, Jean, Stentz, Anthony, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Walter, Matthew Robert, Howard, Thomas M., Hemachandra, Sachithra Madhawa, Teller, Seth, Roy, Nicholas, Duvallet, Felix, Oh, Jean, and Stentz, Anthony
- Abstract
Natural language provides a flexible, intuitive way for people to command robots, which is becoming increasingly important as robots transition to working alongside people in our homes and workplaces. To follow instructions in unknown environments, robots will be expected to reason about parts of the environments that were described in the instruction, but that the robot has no direct knowledge about. However, most existing approaches to natural language understanding require that the robot’s environment be known a priori. This paper proposes a probabilistic framework that enables robots to follow commands given in natural language, without any prior knowledge of the environment. The novelty lies in exploiting environment information implicit in the instruction, thereby treating language as a type of sensor that is used to formulate a prior distribution over the unknown parts of the environment. The algorithm then uses this learned distribution to infer a sequence of actions that are most consistent with the command, updating our belief as we gather Keywords Natural Language; Mobile Robot; Parse Tree; World Model; Behavior Inference
- Published
- 2018
16. Bridging text spotting and SLAM with junction features
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Leonard, John J, Wang, Hsueh-Cheng, Paull, Liam, Rosenholtz, Ruth Ellen, Finn, Chelsea, Kaess, Michael, Teller, Seth, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Leonard, John J, Wang, Hsueh-Cheng, Paull, Liam, Rosenholtz, Ruth Ellen, Finn, Chelsea, Kaess, Michael, and Teller, Seth
- Abstract
Navigating in a previously unknown environment and recognizing naturally occurring text in a scene are two important autonomous capabilities that are typically treated as distinct. However, these two tasks are potentially complementary, (i) scene and pose priors can benefit text spotting, and (ii) the ability to identify and associate text features can benefit navigation accuracy through loop closures. Previous approaches to autonomous text spotting typically require significant training data and are too slow for real-time implementation. In this work, we propose a novel high-level feature descriptor, the “junction”, which is particularly well-suited to text representation and is also fast to compute. We show that we are able to improve SLAM through text spotting on datasets collected with a Google Tango, illustrating how location priors enable improved loop closure with text features., Andrea Bocelli Foundation, East Japan Railway Company, United States. Office of Naval Research (N00014-10-1-0936, N00014-11-1-0688, N00014-13-1-0588), National Science Foundation (U.S.) (IIS-1318392)
- Published
- 2017
17. Generalized Grounding Graphs: A Probabilistic Framework for Understanding Grounded Commands
- Author
-
Kollar, Thomas, Tellex, Stefanie, Walter, Matthew, Huang, Albert, Bachrach, Abraham, Hemachandra, Sachi, Brunskill, Emma, Banerjee, Ashis, Roy, Deb, Teller, Seth, Roy, Nicholas, Kollar, Thomas, Tellex, Stefanie, Walter, Matthew, Huang, Albert, Bachrach, Abraham, Hemachandra, Sachi, Brunskill, Emma, Banerjee, Ashis, Roy, Deb, Teller, Seth, and Roy, Nicholas
- Abstract
Many task domains require robots to interpret and act upon natural language commands which are given by people and which refer to the robot's physical surroundings. Such interpretation is known variously as the symbol grounding problem, grounded semantics and grounded language acquisition. This problem is challenging because people employ diverse vocabulary and grammar, and because robots have substantial uncertainty about the nature and contents of their surroundings, making it difficult to associate the constitutive language elements (principally noun phrases and spatial relations) of the command text to elements of those surroundings. Symbolic models capture linguistic structure but have not scaled successfully to handle the diverse language produced by untrained users. Existing statistical approaches can better handle diversity, but have not to date modeled complex linguistic structure, limiting achievable accuracy. Recent hybrid approaches have addressed limitations in scaling and complexity, but have not effectively associated linguistic and perceptual features. Our framework, called Generalized Grounding Graphs (G^3), addresses these issues by defining a probabilistic graphical model dynamically according to the linguistic parse structure of a natural language command. This approach scales effectively, handles linguistic diversity, and enables the system to associate parts of a command with the specific objects, places, and events in the external world to which they refer. We show that robots can learn word meanings and use those learned meanings to robustly follow natural language commands produced by untrained users. We demonstrate our approach for both mobility commands and mobile manipulation commands involving a variety of semi-autonomous robotic platforms, including a wheelchair, a micro-air vehicle, a forklift, and the Willow Garage PR2., Comment: Submitted to the Journal of Artificial Intelligence Research
- Published
- 2017
18. A Situationally Aware Voice-commandable Robotic Forklift Working Alongside People in Unstructured Outdoor Environments
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Davis, Randall, Walter, Matthew R., Antone, Matthew, Chuangsuwanich, Ekapol, Correa, Andrew, Fletcher, Luke, Frazzoli, Emilio, Friedman, Yuli, Glass, James R., How, Jonathan P., Jeon, Jeong hwan, Karaman, Sertac, Luders, Brandon, Roy, Nicholas, Tellex, Stefanie, Teller, Seth, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Davis, Randall, Walter, Matthew R., Antone, Matthew, Chuangsuwanich, Ekapol, Correa, Andrew, Fletcher, Luke, Frazzoli, Emilio, Friedman, Yuli, Glass, James R., How, Jonathan P., Jeon, Jeong hwan, Karaman, Sertac, Luders, Brandon, Roy, Nicholas, Tellex, Stefanie, and Teller, Seth
- Abstract
One long-standing challenge in robotics is the realization of mobile autonomous robots able to operate safely in human workplaces, and be accepted by the human occupants. We describe the development of a multiton robotic forklift intended to operate alongside people and vehicles, handling palletized materials within existing, active outdoor storage facilities. The system has four novel characteristics. The first is a multimodal interface that allows users to efficiently convey task-level commands to the robot using a combination of pen-based gestures and natural language speech. These tasks include the manipulation, transport, and placement of palletized cargo within dynamic, human-occupied warehouses. The second is the robot's ability to learn the visual identity of an object from a single user-provided example and use the learned model to reliably and persistently detect objects despite significant spatial and temporal excursions. The third is a reliance on local sensing that allows the robot to handle variable palletized cargo and navigate within dynamic, minimally prepared environments without a global positioning system. The fourth concerns the robot's operation in close proximity to people, including its human supervisor, pedestrians who may cross or block its path, moving vehicles, and forklift operators who may climb inside the robot and operate it manually. This is made possible by interaction mechanisms that facilitate safe, effective operation around people. This paper provides a comprehensive description of the system's architecture and implementation, indicating how real-world operational requirements motivated key design choices. We offer qualitative and quantitative analyses of the robot operating in real settings and discuss the lessons learned from our effort.
- Published
- 2016
19. A Situationally Aware Voice-commandable Robotic Forklift Working Alongside People in Unstructured Outdoor Environments
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Davis, Randall, Walter, Matthew R., Antone, Matthew, Chuangsuwanich, Ekapol, Correa, Andrew, Fletcher, Luke, Frazzoli, Emilio, Friedman, Yuli, Glass, James R., How, Jonathan P., Jeon, Jeong hwan, Karaman, Sertac, Luders, Brandon, Roy, Nicholas, Tellex, Stefanie, Teller, Seth, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Davis, Randall, Walter, Matthew R., Antone, Matthew, Chuangsuwanich, Ekapol, Correa, Andrew, Fletcher, Luke, Frazzoli, Emilio, Friedman, Yuli, Glass, James R., How, Jonathan P., Jeon, Jeong hwan, Karaman, Sertac, Luders, Brandon, Roy, Nicholas, Tellex, Stefanie, and Teller, Seth
- Abstract
One long-standing challenge in robotics is the realization of mobile autonomous robots able to operate safely in human workplaces, and be accepted by the human occupants. We describe the development of a multiton robotic forklift intended to operate alongside people and vehicles, handling palletized materials within existing, active outdoor storage facilities. The system has four novel characteristics. The first is a multimodal interface that allows users to efficiently convey task-level commands to the robot using a combination of pen-based gestures and natural language speech. These tasks include the manipulation, transport, and placement of palletized cargo within dynamic, human-occupied warehouses. The second is the robot's ability to learn the visual identity of an object from a single user-provided example and use the learned model to reliably and persistently detect objects despite significant spatial and temporal excursions. The third is a reliance on local sensing that allows the robot to handle variable palletized cargo and navigate within dynamic, minimally prepared environments without a global positioning system. The fourth concerns the robot's operation in close proximity to people, including its human supervisor, pedestrians who may cross or block its path, moving vehicles, and forklift operators who may climb inside the robot and operate it manually. This is made possible by interaction mechanisms that facilitate safe, effective operation around people. This paper provides a comprehensive description of the system's architecture and implementation, indicating how real-world operational requirements motivated key design choices. We offer qualitative and quantitative analyses of the robot operating in real settings and discuss the lessons learned from our effort.
- Published
- 2016
20. Drift-free humanoid state estimation fusing kinematic, inertial and LIDAR sensing
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Fallon, Maurice, Antone, Matthew, Roy, Nicholas, Teller, Seth, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Fallon, Maurice, Antone, Matthew, Roy, Nicholas, and Teller, Seth
- Abstract
This paper describes an algorithm for the probabilistic fusion of sensor data from a variety of modalities (inertial, kinematic and LIDAR) to produce a single consistent position estimate for a walking humanoid. Of specific interest is our approach for continuous LIDAR-based localization which maintains reliable drift-free alignment to a prior map using a Gaussian Particle Filter. This module can be bootstrapped by constructing the map on-the-fly and performs robustly in a variety of challenging field situations. We also discuss a two-tier estimation hierarchy which preserves registration to this map and other objects in the robot’s vicinity while also contributing to direct low-level control of a Boston Dynamics Atlas robot. Extensive experimental demonstrations illustrate how the approach can enable the humanoid to walk over uneven terrain without stopping (for tens of minutes), which would otherwise not be possible. We characterize the performance of the estimator for each sensor modality and discuss the computational requirements., United States. Air Force Research Laboratory (Award FA8750-12-1-0321)
- Published
- 2016
21. Automatic Calibration of Multiple Coplanar Sensors
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Brookshire, Jonathan David, Teller, Seth, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Brookshire, Jonathan David, and Teller, Seth
- Abstract
This paper describes an algorithm for recovering the rigid 3-DOF transformation (offset and rotation) between pairs of sensors mounted rigidly in a common plane on a mobile robot. The algorithm requires only a set of sensor observations made as the robot moves along a suitable path. Our method does not require synchronized sensors; nor does it require complete metrical reconstruction of the environment or the sensor path. We show that incremental pose measurements alone are sufficient to recover sensor calibration through nonlinear least squares estimation. We use the Fisher Information Matrix to compute a Cramer-Rao lower bound (CRLB) for the resulting calibration. Applying the algorithm in practice requires a non-degenerate motion path, a principled procedure for estimating per-sensopose displacements and their covariances, a way to temporally resample asynchronous sensor data, and a way to assess the quality of the recovered calibration. We give constructive methods for each step. We demonstrate and validate the end-to-end calibration procedure for both simulated and real LIDAR and inertial data, achieving CRLBs, and corresponding calibrations, accurate to millimeters and milliradians. Source code is available from http://rvsn.csail.mit.edu/calibration.
- Published
- 2015
22. An Architecture for Online Affordance-based Perception and Whole-body Planning
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Department of Mechanical Engineering, Massachusetts Institute of Technology. Laboratory for Manufacturing and Productivity, Fallon, Maurice Francis, Kuindersma, Scott, Karumanchi, Sisir B., Antone, Matthew, Schneider, Toby, Dai, Hongkai, Perez D'Arpino, Claudia, Deits, Robin Lloyd Henderson, DiCicco, Matt, Fourie, Dehann, Koolen, Frans Anton, Marion, James Patrick, Posa, Michael Antonio, Valenzuela, Andres Klee, Yu, Kuan-Ting, Shah, Julie A., Iagnemma, Karl, Tedrake, Russell Louis, Teller, Seth, Shah, Julie A, Tedrake, Russell L, Fallon, Maurice, Schneider, Toby Edwin, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Department of Mechanical Engineering, Massachusetts Institute of Technology. Laboratory for Manufacturing and Productivity, Fallon, Maurice Francis, Kuindersma, Scott, Karumanchi, Sisir B., Antone, Matthew, Schneider, Toby, Dai, Hongkai, Perez D'Arpino, Claudia, Deits, Robin Lloyd Henderson, DiCicco, Matt, Fourie, Dehann, Koolen, Frans Anton, Marion, James Patrick, Posa, Michael Antonio, Valenzuela, Andres Klee, Yu, Kuan-Ting, Shah, Julie A., Iagnemma, Karl, Tedrake, Russell Louis, Teller, Seth, Shah, Julie A, Tedrake, Russell L, Fallon, Maurice, and Schneider, Toby Edwin
- Abstract
The DARPA Robotics Challenge Trials held in December 2013 provided a landmark demonstration of dexterous mobile robots executing a variety of tasks aided by a remote human operator using only data from the robot's sensor suite transmitted over a constrained, field-realistic communications link. We describe the design considerations, architecture, implementation, and performance of the software that Team MIT developed to command and control an Atlas humanoid robot. Our design emphasized human interaction with an efficient motion planner, where operators expressed desired robot actions in terms of affordances fit using perception and manipulated in a custom user interface. We highlight several important lessons we learned while developing our system on a highly compressed schedule., United States. Defense Advanced Research Projects Agency (United States. Air Force Research Laboratory Award FA8750-12-1-0321), United States. Office of Naval Research (Award N00014-12-1-0071)
- Published
- 2015
23. Learning Articulated Motions From Visual Demonstration
- Author
-
Pillai, Sudeep, Walter, Matthew R., Teller, Seth, Pillai, Sudeep, Walter, Matthew R., and Teller, Seth
- Abstract
Many functional elements of human homes and workplaces consist of rigid components which are connected through one or more sliding or rotating linkages. Examples include doors and drawers of cabinets and appliances; laptops; and swivel office chairs. A robotic mobile manipulator would benefit from the ability to acquire kinematic models of such objects from observation. This paper describes a method by which a robot can acquire an object model by capturing depth imagery of the object as a human moves it through its range of motion. We envision that in future, a machine newly introduced to an environment could be shown by its human user the articulated objects particular to that environment, inferring from these "visual demonstrations" enough information to actuate each object independently of the user. Our method employs sparse (markerless) feature tracking, motion segmentation, component pose estimation, and articulation learning; it does not require prior object models. Using the method, a robot can observe an object being exercised, infer a kinematic model incorporating rigid, prismatic and revolute joints, then use the model to predict the object's motion from a novel vantage point. We evaluate the method's performance, and compare it to that of a previously published technique, for a variety of household objects., Comment: Published in Robotics: Science and Systems X, Berkeley, CA. ISBN: 978-0-9923747-0-9
- Published
- 2015
24. Drift-Free Humanoid State Estimation fusing Kinematic, Inertial and LIDAR Sensing
- Author
-
MASSACHUSETTS INST OF TECH CAMBRIDGE COMPUTER SCIENCE AND ARTIFICIAL INTELLIGENCE LAB, Fallon, Maurice F, Antone, Matthew, Roy, Nicholas, Teller, Seth, MASSACHUSETTS INST OF TECH CAMBRIDGE COMPUTER SCIENCE AND ARTIFICIAL INTELLIGENCE LAB, Fallon, Maurice F, Antone, Matthew, Roy, Nicholas, and Teller, Seth
- Abstract
This paper describes an algorithm for the probabilistic fusion of sensor data from a variety of modalities (inertial, kinematic and LIDAR) to produce a single consistent position estimate for a walking humanoid. Of specific interest is our approach for continuous LIDAR-based localization which maintains reliable drift-free alignment to a prior map using a Gaussian Particle Filter. This module can be bootstrapped by constructing the map on-the-fly and performs robustly in a variety of challenging field situations. We also discuss a two-tier estimation hierarchy which preserves registration to this map and other objects in the robot's vicinity while also contributing to direct low-level control of a Boston Dynamics Atlas robot. Extensive experimental demonstrations illustrate how the approach can enable the humanoid to walk over uneven terrain without stopping (for tens of minutes), which would otherwise not be possible. We characterize the performance of the estimator for each sensor modality and discuss the computational requirements., Under review.
- Published
- 2014
25. An Architecture for Online Affordance-based Perception and Whole-body Planning
- Author
-
Seth Teller, Robotics, Vision & Sensor Networks, Fallon, Maurice, Kuindersma, Scott, Karumanchi, Sisir, Antone, Matthew, Schneider, Toby, Dai, Hongkai, Perez D'Arpino, Claudia, Deits, Robin, DiCicco, Matt, Fourie, Dehann, Koolen, Twan, Marion, Pat, Posa, Michael, Valenzuela, Andres, Yu, Kuan-Ting, Shah, Julie, Iagnemma, Karl, Tedrake, Russ, Teller, Seth, Seth Teller, Robotics, Vision & Sensor Networks, Fallon, Maurice, Kuindersma, Scott, Karumanchi, Sisir, Antone, Matthew, Schneider, Toby, Dai, Hongkai, Perez D'Arpino, Claudia, Deits, Robin, DiCicco, Matt, Fourie, Dehann, Koolen, Twan, Marion, Pat, Posa, Michael, Valenzuela, Andres, Yu, Kuan-Ting, Shah, Julie, Iagnemma, Karl, Tedrake, Russ, and Teller, Seth
- Abstract
The DARPA Robotics Challenge Trials held in December 2013 provided a landmark demonstration of dexterous mobile robots executing a variety of tasks aided by a remote human operator using only data from the robot's sensor suite transmitted over a constrained, field-realistic communications link. We describe the design considerations, architecture, implementation and performance of the software that Team MIT developed to command and control an Atlas humanoid robot. Our design emphasized human interaction with an efficient motion planner, where operators expressed desired robot actions in terms of affordances fit using perception and manipulated in a custom user interface. We highlight several important lessons we learned while developing our system on a highly compressed schedule.
- Published
- 2014
26. An Architecture for Online Affordance-based Perception and Whole-body Planning
- Author
-
Seth Teller, Robotics, Vision & Sensor Networks, Fallon, Maurice, Kuindersma, Scott, Karumanchi, Sisir, Antone, Matthew, Schneider, Toby, Dai, Hongkai, Perez D'Arpino, Claudia, Deits, Robin, DiCicco, Matt, Fourie, Dehann, Koolen, Twan, Marion, Pat, Posa, Michael, Valenzuela, Andres, Yu, Kuan-Ting, Shah, Julie, Iagnemma, Karl, Tedrake, Russ, Teller, Seth, Seth Teller, Robotics, Vision & Sensor Networks, Fallon, Maurice, Kuindersma, Scott, Karumanchi, Sisir, Antone, Matthew, Schneider, Toby, Dai, Hongkai, Perez D'Arpino, Claudia, Deits, Robin, DiCicco, Matt, Fourie, Dehann, Koolen, Twan, Marion, Pat, Posa, Michael, Valenzuela, Andres, Yu, Kuan-Ting, Shah, Julie, Iagnemma, Karl, Tedrake, Russ, and Teller, Seth
- Abstract
The DARPA Robotics Challenge Trials held in December 2013 provided a landmark demonstration of dexterous mobile robots executing a variety of tasks aided by a remote human operator using only data from the robot's sensor suite transmitted over a constrained, field-realistic communications link. We describe the design considerations, architecture, implementation and performance of the software that Team MIT developed to command and control an Atlas humanoid robot. Our design emphasized human interaction with an efficient motion planner, where operators expressed desired robot actions in terms of affordances fit using perception and manipulated in a custom user interface. We highlight several important lessons we learned while developing our system on a highly compressed schedule.
- Published
- 2014
27. Learning Semantic Maps from Natural Language Descriptions
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Walter, Matthew R., Hemachandra, Sachithra Madhaw, Homberg, Bianca S., Tellex, Stefanie, Teller, Seth, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Walter, Matthew R., Hemachandra, Sachithra Madhaw, Homberg, Bianca S., Tellex, Stefanie, and Teller, Seth
- Abstract
This paper proposes an algorithm that enables robots to efficiently learn human-centric models of their environment from natural language descriptions. Typical semantic mapping approaches augment metric maps with higher-level properties of the robot’s surroundings (e.g., place type, object locations), but do not use this information to improve the metric map. The novelty of our algorithm lies in fusing high-level knowledge, conveyed by speech, with metric information from the robot’s low-level sensor streams. Our method jointly estimates a hybrid metric, topological, and semantic representation of the environment. This semantic graph provides a common framework in which we integrate concepts from natural language descriptions (e.g., labels and spatial relations) with metric observations from low-level sensors. Our algorithm efficiently maintains a factored distribution over semantic graphs based upon the stream of natural language and low-level sensor information. We evaluate the algorithm’s performance and demonstrate that the incorporation of information from natural language increases the metric, topological and semantic accuracy of the recovered environment model.
- Published
- 2014
28. Online pose classification and walking speed estimation using handheld devices
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Park, Jun-geun, Patel, Ami, Curtis, Dorothy, Teller, Seth, Ledlie, Jonathan, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Park, Jun-geun, Patel, Ami, Curtis, Dorothy, Teller, Seth, and Ledlie, Jonathan
- Abstract
We describe and evaluate two methods for device pose classification and walking speed estimation that generalize well to new users, compared to previous work. These machine learning based methods are designed for the general case of a person holding a mobile device in an unknown location and require only a single low-cost, low-power sensor: a triaxial accelerometer. We evaluate our methods in straight-path indoor walking experiments as well as in natural indoor walking settings. Experiments with 14 human participants to test user generalization show that our pose classifier correctly selects among four device poses with 94% accuracy compared to 82% for previous work, and our walking speed estimates are within 12-15% (straight/indoor walk) of ground truth compared to 17-22% for previous work. Implementation on a mobile phone demonstrates that both methods can run efficiently online.
- Published
- 2014
29. Motion Compatibility for Indoor Localization
- Author
-
Seth Teller, Robotics, Vision & Sensor Networks, Park, Jun-geun, Teller, Seth, Seth Teller, Robotics, Vision & Sensor Networks, Park, Jun-geun, and Teller, Seth
- Abstract
Indoor localization -- a device's ability to determine its location within an extended indoor environment -- is a fundamental enabling capability for mobile context-aware applications. Many proposed applications assume localization information from GPS, or from WiFi access points. However, GPS fails indoors and in urban canyons, and current WiFi-based methods require an expensive, and manually intensive, mapping, calibration, and configuration process performed by skilled technicians to bring the system online for end users. We describe a method that estimates indoor location with respect to a prior map consisting of a set of 2D floorplans linked through horizontal and vertical adjacencies. Our main contribution is the notion of "path compatibility," in which the sequential output of a classifier of inertial data producing low-level motion estimates (standing still, walking straight, going upstairs, turning left etc.) is examined for agreement with the prior map. Path compatibility is encoded in an HMM-based matching model, from which the method recovers the user s location trajectory from the low-level motion estimates. To recognize user motions, we present a motion labeling algorithm, extracting fine-grained user motions from sensor data of handheld mobile devices. We propose "feature templates," which allows the motion classifier to learn the optimal window size for a specific combination of a motion and a sensor feature function. We show that, using only proprioceptive data of the quality typically available on a modern smartphone, our motion labeling algorithm classifies user motions with 94.5% accuracy, and our trajectory matching algorithm can recover the user's location to within 5 meters on average after one minute of movements from an unknown starting location. Prior information, such as a known starting floor, further decreases the time required to obtain precise location estimate.
- Published
- 2014
30. Motion Compatibility for Indoor Localization
- Author
-
Seth Teller, Robotics, Vision & Sensor Networks, Park, Jun-geun, Teller, Seth, Seth Teller, Robotics, Vision & Sensor Networks, Park, Jun-geun, and Teller, Seth
- Abstract
Indoor localization -- a device's ability to determine its location within an extended indoor environment -- is a fundamental enabling capability for mobile context-aware applications. Many proposed applications assume localization information from GPS, or from WiFi access points. However, GPS fails indoors and in urban canyons, and current WiFi-based methods require an expensive, and manually intensive, mapping, calibration, and configuration process performed by skilled technicians to bring the system online for end users. We describe a method that estimates indoor location with respect to a prior map consisting of a set of 2D floorplans linked through horizontal and vertical adjacencies. Our main contribution is the notion of "path compatibility," in which the sequential output of a classifier of inertial data producing low-level motion estimates (standing still, walking straight, going upstairs, turning left etc.) is examined for agreement with the prior map. Path compatibility is encoded in an HMM-based matching model, from which the method recovers the user s location trajectory from the low-level motion estimates. To recognize user motions, we present a motion labeling algorithm, extracting fine-grained user motions from sensor data of handheld mobile devices. We propose "feature templates," which allows the motion classifier to learn the optimal window size for a specific combination of a motion and a sensor feature function. We show that, using only proprioceptive data of the quality typically available on a modern smartphone, our motion labeling algorithm classifies user motions with 94.5% accuracy, and our trajectory matching algorithm can recover the user's location to within 5 meters on average after one minute of movements from an unknown starting location. Prior information, such as a known starting floor, further decreases the time required to obtain precise location estimate.
- Published
- 2014
31. Sensor fusion for flexible human-portable building-scale mapping
- Author
-
Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Fallon, Maurice Francis, Johannsson, Hordur, Brookshire, Jonathan David, Teller, Seth, Leonard, John Joseph, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Fallon, Maurice Francis, Johannsson, Hordur, Brookshire, Jonathan David, Teller, Seth, and Leonard, John Joseph
- Abstract
This paper describes a system enabling rapid multi-floor indoor map building using a body-worn sensor system fusing information from RGB-D cameras, LIDAR, inertial, and barometric sensors. Our work is motivated by rapid response missions by emergency personnel, in which the capability for one or more people to rapidly map a complex indoor environment is essential for public safety. Human-portable mapping raises a number of challenges not encountered in typical robotic mapping applications including complex 6-DOF motion and the traversal of challenging trajectories including stairs or elevators. Our system achieves robust performance in these situations by exploiting state-of-the-art techniques for robust pose graph optimization and loop closure detection. It achieves real-time performance in indoor environments of moderate scale. Experimental results are demonstrated for human-portable mapping of several floors of a university building, demonstrating the system's ability to handle motion up and down stairs and to organize initially disconnected sets of submaps in a complex environment., Lincoln Laboratory, United States. Air Force (Contract FA8721-05-C-0002), United States. Office of Naval Research (Grant N00014-10-1-0936), United States. Office of Naval Research (Grant N00014-11-1-0688), United States. Office of Naval Research (Grant N00014-12-10020)
- Published
- 2013
32. Understanding natural language commands for robotic navigation and mobile manipulation
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Tellex, Stefanie A., Kollar, Thomas Fleming, Dickerson, Steven R., Walter, Matthew R., Banerjee, Ashis, Teller, Seth, Roy, Nicholas, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Tellex, Stefanie A., Kollar, Thomas Fleming, Dickerson, Steven R., Walter, Matthew R., Banerjee, Ashis, Teller, Seth, and Roy, Nicholas
- Abstract
This paper describes a new model for understanding natural language commands given to autonomous systems that perform navigation and mobile manipulation in semi-structured environments. Previous approaches have used models with fixed structure to infer the likelihood of a sequence of actions given the environment and the command. In contrast, our framework, called Generalized Grounding Graphs, dynamically instantiates a probabilistic graphical model for a particular natural language command according to the command's hierarchical and compositional semantic structure. Our system performs inference in the model to successfully find and execute plans corresponding to natural language commands such as "Put the tire pallet on the truck." The model is trained using a corpus of commands collected using crowdsourcing. We pair each command with robot actions and use the corpus to learn the parameters of the model. We evaluate the robot's performance by inferring plans from natural language commands, executing each plan in a realistic robot simulator, and asking users to evaluate the system's performance. We demonstrate that our system can successfully follow many natural language commands from the corpus.
- Published
- 2012
33. Multiple Relative Pose Graphs for Robust Cooperative Mapping
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Department of Mechanical Engineering, Kim, Been, Kaess, Michael, Fletcher, Luke Sebastian, Leonard, John Joseph, Bachrach, Abraham Galton, Roy, Nicholas, Teller, Seth, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Department of Mechanical Engineering, Kim, Been, Kaess, Michael, Fletcher, Luke Sebastian, Leonard, John Joseph, Bachrach, Abraham Galton, Roy, Nicholas, and Teller, Seth
- Abstract
This paper describes a new algorithm for cooperative and persistent simultaneous localization and mapping (SLAM) using multiple robots. Recent pose graph representations have proven very successful for single robot mapping and localization. Among these methods, incremental smoothing and mapping (iSAM) gives an exact incremental solution to the SLAM problem by solving a full nonlinear optimization problem in real-time. In this paper, we present a novel extension to iSAM to facilitate online multi-robot mapping based on multiple pose graphs. Our main contribution is a relative formulation of the relationship between multiple pose graphs that avoids the initialization problem and leads to an efficient solution when compared to a completely global formulation. The relative pose graphs are optimized together to provide a globally consistent multi-robot solution. Efficient access to covariances at any time for relative parameters is provided through iSAM, facilitating data association and loop closing. The performance of the technique is illustrated on various data sets including a publicly available multi-robot data set. Further evaluation is performed in a collaborative helicopter and ground robot experiment., United States. Office of Naval Research (Grant N00014-05-1-0244), United States. Office of Naval Research (Grant N00014-06-1-0043), United States. Office of Naval Research (Grant N00014-07-1-0749), Massachusetts Institute of Technology. Center for Technology, Policy, and Industrial Development. Ford-MIT Alliance
- Published
- 2012
34. Asymptotically-optimal path planning for manipulation using incremental sampling-based algorithms
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Laboratory for Information and Decision Systems, Perez, Alejandro Tomas, Karaman, Sertac, Shkolnik, Alexander C., Frazzoli, Emilio, Teller, Seth, Walter, Matthew R., Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Laboratory for Information and Decision Systems, Perez, Alejandro Tomas, Karaman, Sertac, Shkolnik, Alexander C., Frazzoli, Emilio, Teller, Seth, and Walter, Matthew R.
- Abstract
A desirable property of path planning for robotic manipulation is the ability to identify solutions in a sufficiently short amount of time to be usable. This is particularly challenging for the manipulation problem due to the need to plan over high-dimensional configuration spaces and to perform computationally expensive collision checking procedures. Consequently, existing planners take steps to achieve desired solution times at the cost of low quality solutions. This paper presents a planning algorithm that overcomes these difficulties by augmenting the asymptotically-optimal RRT* with a sparse sampling procedure. With the addition of a collision checking procedure that leverages memoization, this approach has the benefit that it quickly identifies low-cost feasible trajectories and takes advantage of subsequent computation time to refine the solution towards an optimal one. We evaluate the algorithm through a series of Monte Carlo simulations of seven, twelve, and fourteen degree of freedom manipulation planning problems in a realistic simulation environment. The results indicate that the proposed approach provides significant improvements in the quality of both the initial solution and the final path, while incurring almost no computational overhead compared to the RRT algorithm. We conclude with a demonstration of our algorithm for single-arm and dual-arm planning on Willow Garage's PR2 robot.
- Published
- 2012
35. Probabilistic lane estimation for autonomous driving using basis curves
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Huang, Albert S., Teller, Seth, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Huang, Albert S., and Teller, Seth
- Abstract
Lane estimation for autonomous driving can be formulated as a curve estimation problem, where local sensor data provides partial and noisy observations of spatial curves forming lane boundaries. The number of lanes to estimate are initially unknown and many observations may be outliers or false detections (due e.g. to shadows or non-boundary road paint). The challenges lie in detecting lanes when and where they exist, and updating lane estimates as new observations are made. This paper describes an efficient probabilistic lane estimation algorithm based on a novel curve representation. The key advance is a principled mechanism to describe many similar curves as variations of a single basis curve. Locally observed road paint and curb features are then fused to detect and estimate all nearby travel lanes. The system handles roads with complex multi-lane geometries and makes no assumptions about the position and orientation of the vehicle with respect to the roadway. We evaluate our algorithm using a ground truth dataset containing manually-labeled, fine-grained lane geometries for vehicle travel in two large and diverse datasets that include more than 300,000 images and 44 km of roadway. The results illustrate the capabilities of our algorithm for robust lane estimation in the face of challenging conditions and unknown roadways., United States. Defense Advanced Research Projects Agency (Urban Challenge, ARPA Order No. W369/00, Program Code DIRO, issued by DARPA/CMO under Contract No. HR0011-06-C-0149)
- Published
- 2012
36. One-shot visual appearance learning for mobile manipulation
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Walter, Matthew R., Teller, Seth, Friedman, Yuli, Antone, Matthew, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Walter, Matthew R., Teller, Seth, Friedman, Yuli, and Antone, Matthew
- Abstract
We describe a vision-based algorithm that enables a robot to robustly detect specific objects in a scene following an initial segmentation hint from a human user. The novelty lies in the ability to ‘reacquire’ objects over extended spatial and temporal excursions within challenging environments based upon a single training example. The primary difficulty lies in achieving an effective reacquisition capability that is robust to the effects of local clutter, lighting variation, and object relocation. We overcome these challenges through an adaptive detection algorithm that automatically generates multiple-view appearance models for each object online. As the robot navigates within the environment and the object is detected from different viewpoints, the one-shot learner opportunistically and automatically incorporates additional observations into each model. In order to overcome the effects of ‘drift’ common to adaptive learners, the algorithm imposes simple requirements on the geometric consistency of candidate observations. Motivating our reacquisition strategy is our work developing a mobile manipulator that interprets and autonomously performs commands conveyed by a human user. The ability to detect specific objects and reconstitute the user’s segmentation hints enables the robot to be situationally aware. This situational awareness enables rich command and control mechanisms and affords natural interaction. We demonstrate one such capability that allows the human to give the robot a ‘guided tour’ of named objects within an outdoor environment and, hours later, to direct the robot to manipulate those objects by name using spoken instructions. We implemented our appearance-based detection strategy on our robotic manipulator as it operated over multiple days in different outdoor environments. We evaluate the algorithm’s performance under challenging conditions that include scene clutter, lighting and viewpoint variation, object ambiguity, and object relocation. The, United States. Air Force (Contract FA8721-05-C-0002)
- Published
- 2012
37. Multimodal interaction with an autonomous forklift
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Davis, Randall, Correa, Andrew Thomas, Walter, Matthew R., Fletcher, Luke Sebastian, Glass, James R., Teller, Seth, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Davis, Randall, Correa, Andrew Thomas, Walter, Matthew R., Fletcher, Luke Sebastian, Glass, James R., and Teller, Seth
- Abstract
We describe a multimodal framework for interacting with an autonomous robotic forklift. A key element enabling effective interaction is a wireless, handheld tablet with which a human supervisor can command the forklift using speech and sketch. Most current sketch interfaces treat the canvas as a blank slate. In contrast, our interface uses live and synthesized camera images from the forklift as a canvas, and augments them with object and obstacle information from the world. This connection enables users to ¿draw on the world,¿ enabling a simpler set of sketched gestures. Our interface supports commands that include summoning the forklift and directing it to lift, transport, and place loads of palletized cargo. We describe an exploratory evaluation of the system designed to identify areas for detailed study. Our framework incorporates external signaling to interact with humans near the vehicle. The robot uses audible and visual annunciation to convey its current state and intended actions. The system also provides seamless autonomy handoff: any human can take control of the robot by entering its cabin, at which point the forklift can be operated manually until the human exits., United States. Army. Logistics Innovation Agency, United States. Army Combined Arms Support Command, United States. Dept. of the Air Force (Air Force Contract FA8721-05-C-0002)
- Published
- 2012
38. A High-Rate, Heterogeneous Data Set from the Darpa Urban Challenge
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Department of Mechanical Engineering, Huang, Albert S., Fletcher, Luke Sebastian, Moore, David C., Teller, Seth, Leonard, John Joseph, Olson, Edwin B., Antone, Matthew, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Department of Mechanical Engineering, Huang, Albert S., Fletcher, Luke Sebastian, Moore, David C., Teller, Seth, Leonard, John Joseph, Olson, Edwin B., and Antone, Matthew
- Abstract
This paper describes a data set collected by MIT’s autonomous vehicle Talos during the 2007 DARPA Urban Challenge. Data from a high-precision navigation system, five cameras, 12 SICK planar laser range scanners, and a Velodyne high-density laser range scanner were synchronized and logged to disk for 90 km of travel. In addition to documenting a number of large loop closures useful for developing mapping and localization algorithms, this data set also records the first robotic traffic jam and two autonomous vehicle collisions. It is our hope that this data set will be useful to the autonomous vehicle community, especially those developing robotic perception capabilities., United States. Defense Advanced Research Projects Agency (Urban Challenge, ARPA Order No. W369/00, Program Code DIRO, issued by DARPA/CMO under Contract No. HR0011-06-C-0149)
- Published
- 2012
39. Approaching the Symbol Grounding Problem with Probabilistic Graphical Models
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Tellex, Stefanie A., Kollar, Thomas Fleming, Dickerson, Steven R., Walter, Matthew R., Banerjee, Ashis, Teller, Seth, Roy, Nicholas, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Tellex, Stefanie A., Kollar, Thomas Fleming, Dickerson, Steven R., Walter, Matthew R., Banerjee, Ashis, Teller, Seth, and Roy, Nicholas
- Abstract
In order for robots to engage in dialog with human teammates, they must have the ability to map between words in the language and aspects of the external world. A solution to this symbol grounding problem (Harnad, 1990) would enable a robot to interpret commands such as “Drive over to receiving and pick up the tire pallet.” In this article we describe several of our results that use probabilistic inference to address the symbol grounding problem. Our specific approach is to develop models that factor according to the linguistic structure of a command. We first describe an early result, a generative model that factors according to the sequential structure of language, and then discuss our new framework, generalized grounding graphs (G3). The G3 framework dynamically instantiates a probabilistic graphical model for a natural language input, enabling a mapping between words in language and concrete objects, places, paths and events in the external world. We report on corpus-based experiments where the robot is able to learn and use word meanings in three real-world tasks: indoor navigation, spatial language video retrieval, and mobile manipulation., U.S. Army Research Laboratory. Collaborative Technology Alliance Program (Cooperative Agreement W911NF-10-2-0016), United States. Office of Naval Research (MURI N00014-07-1-0749)
- Published
- 2012
40. Ground Robot Navigation using Uncalibrated Cameras
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, Huang, Albert S., Koch, Olivier A., Walter, Matthew R., Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, Huang, Albert S., Koch, Olivier A., and Walter, Matthew R.
- Abstract
Precise calibration of camera intrinsic and extrinsic parameters, while often useful, is difficult to obtain during field operation and presents scaling issues for multi-robot systems. We demonstrate a vision-based approach to navigation that does not depend on traditional camera calibration, and present an algorithm for guiding a robot through a previously traversed environment using a set of uncalibrated cameras mounted on the robot. On the first excursion through an environment, the system builds a topological representation of the robot's exploration path, encoded as a place graph. On subsequent navigation missions, the method localizes the robot within the graph and provides robust guidance to a specified destination. We combine this method with reactive collision avoidance to obtain a system able to navigate the robot safely and reliably through the environment. We validate our approach with ground-truth experiments and demonstrate the method on a small ground rover navigating through several dynamic environments., Charles Stark Draper Laboratory
- Published
- 2012
41. Following and Interpreting Narrated Guided Tours
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, Hemachandra, Sachithra Madhaw, Kollar, Thomas Fleming, Roy, Nicholas, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, Hemachandra, Sachithra Madhaw, Kollar, Thomas Fleming, and Roy, Nicholas
- Abstract
URL to abstract on conference site, We describe a robotic tour-taking capability enabling a robot to acquire local knowledge of a human-occupied environment. A tour-taking robot autonomously follows a human guide through an environment, interpreting the guide’s spoken utterances and the shared spatiotemporal context in order to acquire a spatially segmented and semantically labeled metrical-topological representation of the environment. The described tour-taking capability enables scalable deployment of mobile robots into human-occupied environments, and natural human-robot interaction for commanded mobility. Our primary contributions are an efficient, socially acceptable autonomous tour-following behavior and a tour interpretation algorithm that partitions a map into spaces labeled according to the guide’s utterances. The tour-taking behavior is demonstrated in a multi-floor office building and evaluated by assessing the comfort of the tour guides, and by comparing the robot’s map partitions to those produced by humans.
- Published
- 2011
42. A Voice-Commandable Robotic Forklift Working Alongside Humans in Minimally-Prepared Outdoor Environments
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, Walter, Matthew R., Correa, Andrew Thomas, Davis, Randall, Fletcher, Luke Sebastian, Frazzoli, Emilio, Glass, Jim, How, Jonathan P., Huang, Albert S., Jeon, Jeong hwan, Karaman, Sertac, Luders, Brandon Douglas, Roy, Nicholas, Antone, Matthew, Sainath, Tara, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, Walter, Matthew R., Correa, Andrew Thomas, Davis, Randall, Fletcher, Luke Sebastian, Frazzoli, Emilio, Glass, Jim, How, Jonathan P., Huang, Albert S., Jeon, Jeong hwan, Karaman, Sertac, Luders, Brandon Douglas, Roy, Nicholas, Antone, Matthew, and Sainath, Tara
- Abstract
One long-standing challenge in robotics is the realization of mobile autonomous robots able to operate safely in existing human workplaces in a way that their presence is accepted by the human occupants. We describe the development of a multi-ton robotic forklift intended to operate alongside human personnel, handling palletized materials within existing, busy, semi-structured outdoor storage facilities. The system has three principal novel characteristics. The first is a multimodal tablet that enables human supervisors to use speech and pen-based gestures to assign tasks to the forklift, including manipulation, transport, and placement of palletized cargo. Second, the robot operates in minimally-prepared, semi-structured environments, in which the forklift handles variable palletized cargo using only local sensing (and no reliance on GPS), and transports it while interacting with other moving vehicles. Third, the robot operates in close proximity to people, including its human supervisor, other pedestrians who may cross or block its path, and forklift operators who may climb inside the robot and operate it manually. This is made possible by novel interaction mechanisms that facilitate safe, effective operation around people. We describe the architecture and implementation of the system, indicating how real-world operational requirements motivated the development of the key subsystems, and provide qualitative and quantitative descriptions of the robot operating in real settings., United States. Dept. of the Air Force (FA8721-05-C-0002)
- Published
- 2011
43. Spoken command of large mobile robots in outdoor environments
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, Chuangsuwanich, Ekapol, Cyphers, David Scott, Glass, James R., Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, Chuangsuwanich, Ekapol, Cyphers, David Scott, and Glass, James R.
- Abstract
We describe a speech system for commanding robots in human-occupied outdoor military supply depots. To operate in such environments, the robots must be as easy to interact with as are humans, i.e. they must reliably understand ordinary spoken instructions, such as orders to move supplies, as well as commands and warnings, spoken or shouted from distances of tens of meters. These design goals preclude close-talking microphones and “push-to-talk” buttons that are typically used to isolate commands from the sounds of vehicles, machinery and non-relevant speech. We used multiple microphones to provide omnidirectional coverage. A novel voice activity detector was developed to detect speech and select the appropriate microphone to listen to. Finally, we developed a recognizer model that could successfully recognize commands when heard amidst other speech within a noisy environment. When evaluated on speech data in the field, this system performed significantly better than a more computationally intensive baseline system, reducing the effective false alarm rate by a factor of 40, while maintaining the same level of precision.
- Published
- 2011
44. Implications of Device Diversity for Organic Localization
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, Park, Jun-geun, Curtis, Dorothy, Ledlie, Jonathan, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, Park, Jun-geun, Curtis, Dorothy, and Ledlie, Jonathan
- Abstract
paper listed on conference site, Many indoor localization methods are based on the association of 802.11 wireless RF signals from wireless access points (WAPs) with location labels. An “organic” RF positioning system relies on regular users, not dedicated surveyors, to build the map of RF fingerprints to location labels. However, signal variation due to device heterogeneity may degrade localization performance. We analyze the diversity of those signal characteristics pertinent to indoor localization — signal strength and AP detection — as measured by a variety of 802.11 devices. We first analyze signal strength diversity, and show that pairwise linear transformation alone does not solve the problem. We propose kernel estimation with a wide kernel width to reduce the difference in probability estimates. We also investigate diversity in access point detection. We demonstrate that localization performance may degrade significantly when AP detection rate is used as a feature for localization, and correlate the loss of performance to a device dissimilarity measure captured by Kullback-Leibler divergence. Based on this analysis, we show that using only signal strength, without incorporating negative evidence, achieves good localization performance when devices are heterogeneous., Nokia Research Center
- Published
- 2011
45. Appearance-based object reacquisition for mobile manipulation
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, Walter, Matthew R., Friedman, Yuli, Antone, Matthew, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, Walter, Matthew R., Friedman, Yuli, and Antone, Matthew
- Abstract
This paper describes an algorithm enabling a human supervisor to convey task-level information to a robot by using stylus gestures to circle one or more objects within the field of view of a robot-mounted camera. These gestures serve to segment the unknown objects from the environment. Our method's main novelty lies in its use of appearance-based object “reacquisition” to reconstitute the supervisory gestures (and corresponding segmentation hints), even for robot viewpoints spatially and/or temporally distant from the viewpoint underlying the original gesture. Reacquisition is particularly challenging within relatively dynamic and unstructured environments. The technical challenge is to realize a reacquisition capability robust enough to appearance variation to be useful in practice. Whenever the supervisor indicates an object, our system builds a feature-based appearance model of the object. When the object is detected from subsequent viewpoints, the system automatically and opportunistically incorporates additional observations, revising the appearance model and reconstituting the rough contours of the original circling gesture around that object. Our aim is to exploit reacquisition in order to both decrease the user burden of task specification and increase the effective autonomy of the robot. We demonstrate and analyze the approach on a robotic forklift designed to approach, manipulate, transport and place palletized cargo within an outdoor warehouse. We show that the method enables gesture reuse over long timescales and robot excursions (tens of minutes and hundreds of meters)., United States. Dept. of the Air Force (Air Force Contract FA8721-05-C-0002)
- Published
- 2011
46. Probabilistic Lane Estimation using Basis Curves
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, Huang, Albert S., Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Teller, Seth, and Huang, Albert S.
- Abstract
Lane estimation for autonomous driving can be formulated as a curve estimation problem, where local sensor data provides partial and noisy observations of spatial curves. The number of curves to estimate may be initially unknown and many of the observations may be outliers or false detections (due e.g. to to tree shadows or lens flare). The challenges lie in detecting lanes when and where they exist, and updating lane estimates as new observations are made. This paper describes an efficient probabilistic lane estimation algorithm based on a novel curve representation. The key advance is a principled mechanism to describe many similar curves as variations of a single basis curve. Locally observed road paint and curb features are then fused to detect and estimate all nearby travel lanes. The system handles roads with complex geometries and makes no assumptions about the position and orientation of the vehicle with respect to the roadway. We evaluate our algorithm with a ground truth dataset containing manually-labeled, fine-grained lane geometries for vehicle travel in two large and diverse datasets that include more than 300,000 images and 44km of roadway. The results illustrate the capabilities of our algorithm for robust lane estimation in the face of challenging conditions and unknown roadways.
- Published
- 2011
47. Closed-loop pallet manipulation in unstructured environments
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Laboratory for Information and Decision Systems, Teller, Seth, Walter, Matthew R., Karaman, Sertac, Frazzoli, Emilio, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Laboratory for Information and Decision Systems, Teller, Seth, Walter, Matthew R., Karaman, Sertac, and Frazzoli, Emilio
- Abstract
This paper addresses the problem of autonomous manipulation of a priori unknown palletized cargo with a robotic lift truck (forklift). Specifically, we describe coupled perception and control algorithms that enable the vehicle to engage and place loaded pallets relative to locations on the ground or truck beds. Having little prior knowledge of the objects with which the vehicle is to interact, we present an estimation framework that utilizes a series of classifiers to infer the objects' structure and pose from individual LIDAR scans. The classifiers share a low-level shape estimation algorithm that uses linear programming to robustly segment input data into sets of weak candidate features. We present and analyze the performance of the segmentation method, and subsequently describe its role in our estimation algorithm. We then evaluate the performance of a motion controller that, given an estimate of a pallet's pose, is employed to safely engage each pallet. We conclude with a validation of our algorithms for a set of real-world pallet and truck interactions., United States. Dept. of the Air Force (Air Force Contract FA8721-05-C-00)
- Published
- 2011
48. Growing an organic indoor location system
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Teller, Seth, Park, Jun-geun, Charrow, Ben, Curtis, Dorothy, Battat, Jonathan, Minkov, Einat, Hicks, Jamey, Ledlie, Jonathan, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Teller, Seth, Park, Jun-geun, Charrow, Ben, Curtis, Dorothy, Battat, Jonathan, Minkov, Einat, Hicks, Jamey, and Ledlie, Jonathan
- Abstract
Most current methods for 802.11-based indoor localization depend on surveys conducted by experts or skilled technicians. Some recent systems have incorporated surveying by users. Structuring localization systems "organically," however, introduces its own set of challenges: conveying uncertainty, determining when user input is actually required, and discounting erroneous and stale data. Through deployment of an organic location system in our nine-story building, which contains nearly 1,400 distinct spaces, we evaluate new algorithms for addressing these challenges. We describe the use of Voronoi regions for conveying uncertainty and reasoning about gaps in coverage, and a clustering method for identifying potentially erroneous user data. Our algorithms facilitate rapid coverage while maintaining positioning accuracy comparable to that achievable with survey-driven indoor deployments., Nokia Research Center
- Published
- 2011
49. Anytime Motion Planning using the RRT*
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Laboratory for Information and Decision Systems, Teller, Seth, Karaman, Sertac, Walter, Matthew R., Frazzoli, Emilio, Perez, Alejandro, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Aeronautics and Astronautics, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology. Laboratory for Information and Decision Systems, Teller, Seth, Karaman, Sertac, Walter, Matthew R., Frazzoli, Emilio, and Perez, Alejandro
- Abstract
The Rapidly-exploring Random Tree (RRT) algorithm, based on incremental sampling, efficiently computes motion plans. Although the RRT algorithm quickly produces candidate feasible solutions, it tends to converge to a solution that is far from optimal. Practical applications favor “anytime” algorithms that quickly identify an initial feasible plan, then, given more computation time available during plan execution, improve the plan toward an optimal solution. This paper describes an anytime algorithm based on the RRT* which (like the RRT) finds an initial feasible solution quickly, but (unlike the RRT) almost surely converges to an optimal solution. We present two key extensions to the RRT*, committed trajectories and branch-and-bound tree adaptation, that together enable the algorithm to make more efficient use of computation time online, resulting in an anytime algorithm for real-time implementation. We evaluate the method using a series of Monte Carlo runs in a high-fidelity simulation environment, and compare the operation of the RRT and RRT* methods. We also demonstrate experimental results for an outdoor wheeled robotic vehicle., United States. Army. Logistics Innovation Agency, United States. Army Combined Arms Support Command, United States. Dept. of the Air Force (Air Force Contract FA8721-05-C-0002)
- Published
- 2011
50. Collaborative future event recommendation
- Author
-
Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Jaakkola, Tommi S., Teller, Seth, Minkov, Einat, Charrow, Ben, Ledlie, Jonathan, Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science, Jaakkola, Tommi S., Teller, Seth, Minkov, Einat, Charrow, Ben, and Ledlie, Jonathan
- Abstract
We demonstrate a method for collaborative ranking of future events. Previous work on recommender systems typically relies on feedback on a particular item, such as a movie, and generalizes this to other items or other people. In contrast, we examine a setting where no feedback exists on the particular item. Because direct feedback does not exist for events that have not taken place, we recommend them based on individuals' preferences for past events, combined collaboratively with other peoples' likes and dislikes. We examine the topic of unseen item recommendation through a user study of academic (scientific) talk recommendation, where we aim to correctly estimate a ranking function for each user, predicting which talks would be of most interest to them. Then by decomposing user parameters into shared and individual dimensions, we induce a similarity metric between users based on the degree to which they share these dimensions. We show that the collaborative ranking predictions of future events are more effective than pure content-based recommendation. Finally, to further reduce the need for explicit user feedback, we suggest an active learning approach for eliciting feedback and a method for incorporating available implicit user cues., Nokia Research Center
- Published
- 2011
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.