Spatiotemporal Occupancy Prediction for Autonomous Driving


Book Description

Advancements in robotics, computer vision, machine learning and hardware have contributed to impressive developments of autonomous vehicles. However, there still exist challenges that must be tackled in order for the autonomous vehicles to be safely and seamlessly integrated into human environments. This is particularly the case in dense and cluttered urban settings. Autonomous vehicles must be able to understand and anticipate how their surroundings will evolve in both time and space. This capability will allow the autonomous vehicles to proactively plan safe trajectories and avoid other traffic agents. A common prediction approach is an agent-centric method (e.g., pedestrian or vehicle trajectory prediction). These methods require detection and tracking of all agents in the environment since trajectory prediction is performed on each agent. An alternative approach is a map-based (e.g., occupancy grid map) prediction method where the entire environment is discretized into grid cells and the collective occupancy probabilities for each grid cell are predicted. Hence, object detection and tracking capability is generally not needed. This makes a map-based occupancy prediction approach more robust to partial object occlusions and is capable of handling any arbitrary number of agents in the environments. However, a common problem with occupancy grid map prediction is the vanishing of objects from the predictions, especially at longer time horizons. In this thesis, we consider the problem of spatiotemporal environment prediction in urban environments. We merge tools from robotics, computer vision and deep learning to develop spatiotemporal occupancy prediction frameworks that leverage environment information. In our first research work, we developed an occupancy prediction methodology that leverages environment dynamic information, in terms of static-dynamic parts of the environment. Our model learns to predict the spatiotemporal evolution of the static and dynamic parts of the environment input separately and outputs the final occupancy grid map predictions of the entire environment. In our second research work, we further developed the prediction framework to be modular, by adding a learning-based static-dynamic segmentation module upstream of the occupancy prediction module. The addition addressed previous limitations that require the static and dynamic parts of the environment to be known in advance. Lastly, we developed an environment prediction framework that leverages environment semantic information. Our proposed model consists of two sub-modules, which are future semantic segmentation prediction and occupancy prediction. We proposed to represent environment semantics in the form of semantic gird maps that are similar to the occupancy grid representation. This allows a direct flow of semantic information to the occupancy prediction sub-module. Experiments validated on the real-world driving dataset show that our methods outperform other state-of-the-art models and reduce the issue of vanishing object in the predictions at longer time horizons.




Uncertainty-aware Spatiotemporal Perception for Autonomous Vehicles


Book Description

Autonomous vehicles are set to revolutionize transportation in terms of safety and efficiency. However, autonomous systems still have challenges operating in complex human environments, such as an autonomous vehicle in a cluttered, dynamic urban setting. A key obstacle to deploying autonomous systems on the road is understanding, anticipating, and making inferences about human behaviors. Autonomous perception builds a general understanding of the environment for a robot. This includes making inferences about human behaviors in both space and time. Humans are difficult to model due to their vastly diverse behaviors and rapidly evolving objectives. Moreover, in cluttered settings, there are computational and visibility limitations. However, humans also possess desirable capabilities, such as their ability to generalize beyond their observed environment. Although learning-based systems have had success in recent years in modeling and imitating human behavior, efficiently capturing the data and model uncertainty for these systems remains an open problem. This thesis proposes algorithmic advances to uncertainty-aware autonomous perception systems in human environments. We make system-level contributions to spatiotemporal robot perception that reasons about human behavior, and foundational advancements in uncertainty-aware machine learning models for trajectory prediction. These contributions enable robotic systems to make uncertainty- and socially-aware spatiotemporal inferences about human behavior. Traditional robot perception is object-centric and modular, consisting of object detection, tracking, and trajectory prediction stages. These systems can fail prior to the prediction stage due to partial occlusions in the environment. We thus propose an alternative end-to-end paradigm for spatiotemporal environment prediction from a map-centric occupancy grid representation. Occupancy grids are robust to partial occlusions, can handle an arbitrary number of human agents in the scene, and do not require a priori information regarding the environment. We investigate the performance of computer vision techniques in this context and develop new mechanisms tailored to the task of spatiotemporal environment prediction. Spatially, robots also need to reason about fully occluded agents in their environment, which may occur due to sensor limitations or other agents on the road obstructing the field of view. Humans excel at extrapolating from their experiences by making inferences from observed social behaviors. We draw inspiration from human intuition to fill in portions of the robot's map that are not observable by traditional sensors. We infer occupancy in these occluded regions by learning a multimodal mapping from observed human driver behaviors to the environment ahead of them, thus treating people as sensors. Our system handles multiple observed agents to maximally inform the occupancy map around the robot. In order to safely integrate human behavior modeling into the robot autonomy stack, the perception system must efficiently account for uncertainty. Human behavior is often modeled using discrete latent spaces in learning-based models to capture the multimodality in the distribution. For example, in a trajectory prediction task, there may be multiple valid future predictions given a past trajectory. To accurately model this latent distribution, the latent space needs to be sufficiently large, leading to tractability concerns for downstream tasks, such as path planning. We address this issue by proposing a sparsification algorithm for discrete latent sample spaces that can be applied post hoc without sacrificing model performance. Our approach successfully balances multimodality and sparsity to achieve efficient data uncertainty estimation. Aside from modeling data uncertainty, learning-based autonomous systems must be aware of their model uncertainty or what they do not know. Flagging out-of-distribution or unknown scenarios encountered in the real world could be helpful to downstream autonomy stack components and to engineers for further system development. Although the machine learning community has been prolific in model uncertainty estimation for small benchmark problems, relatively little work has been done on estimating this uncertainty in complex, learning-based robotic systems. We propose efficiently learning the model uncertainty over an interpretable, low-dimensional latent space in the context of a trajectory prediction task. The algorithms presented in this thesis were validated on real-world autonomous driving data and baselined against state-of-the-art techniques. We show that drawing inspiration from human-level reasoning while modeling the associated uncertainty can inform environment understanding for autonomous perception systems. The contributions made in this thesis are a step towards uncertainty- and socially-aware autonomous systems that can function seamlessly in human environments.




Belief State Planning for Autonomous Driving: Planning with Interaction, Uncertain Prediction and Uncertain Perception


Book Description

This work presents a behavior planning algorithm for automated driving in urban environments with an uncertain and dynamic nature. The algorithm allows to consider the prediction uncertainty (e.g. different intentions), perception uncertainty (e.g. occlusions) as well as the uncertain interactive behavior of the other agents explicitly. Simulating the most likely future scenarios allows to find an optimal policy online that enables non-conservative planning under uncertainty.




Computational Science – ICCS 2021


Book Description

The six-volume set LNCS 12742, 12743, 12744, 12745, 12746, and 12747 constitutes the proceedings of the 21st International Conference on Computational Science, ICCS 2021, held in Krakow, Poland, in June 2021.* The total of 260 full papers and 57 short papers presented in this book set were carefully reviewed and selected from 635 submissions. 48 full and 14 short papers were accepted to the main track from 156 submissions; 212 full and 43 short papers were accepted to the workshops/ thematic tracks from 479 submissions. The papers were organized in topical sections named: Part I: ICCS Main Track Part II: Advances in High-Performance Computational Earth Sciences: Applications and Frameworks; Applications of Computational Methods in Artificial Intelligence and Machine Learning; Artificial Intelligence and High-Performance Computing for Advanced Simulations; Biomedical and Bioinformatics Challenges for Computer Science Part III: Classifier Learning from Difficult Data; Computational Analysis of Complex Social Systems; Computational Collective Intelligence; Computational Health Part IV: Computational Methods for Emerging Problems in (dis-)Information Analysis; Computational Methods in Smart Agriculture; Computational Optimization, Modelling and Simulation; Computational Science in IoT and Smart Systems Part V: Computer Graphics, Image Processing and Artificial Intelligence; Data-Driven Computational Sciences; Machine Learning and Data Assimilation for Dynamical Systems; MeshFree Methods and Radial Basis Functions in Computational Sciences; Multiscale Modelling and Simulation Part VI: Quantum Computing Workshop; Simulations of Flow and Transport: Modeling, Algorithms and Computation; Smart Systems: Bringing Together Computer Vision, Sensor Networks and Machine Learning; Software Engineering for Computational Science; Solving Problems with Uncertainty; Teaching Computational Science; Uncertainty Quantification for Computational Models *The conference was held virtually. Chapter “Deep Learning Driven Self-adaptive hp Finite Element Method” is available open access under a Creative Commons Attribution 4.0 International License via link.springer.com.
















Transportation and Power Grid in Smart Cities


Book Description

With the increasing worldwide trend in population migration into urban centers, we are beginning to see the emergence of the kinds of mega-cities which were once the stuff of science fiction. It is clear to most urban planners and developers that accommodating the needs of the tens of millions of inhabitants of those megalopolises in an orderly and uninterrupted manner will require the seamless integration of and real-time monitoring and response services for public utilities and transportation systems. Part speculative look into the future of the world’s urban centers, part technical blueprint, this visionary book helps lay the groundwork for the communication networks and services on which tomorrow’s “smart cities” will run. Written by a uniquely well-qualified author team, this book provides detailed insights into the technical requirements for the wireless sensor and actuator networks required to make smart cities a reality.




Machine Learning for Transportation Research and Applications


Book Description

Transportation is a combination of systems that presents a variety of challenges often too intricate to be addressed by conventional parametric methods. Increasing data availability and recent advancements in machine learning provide new methods to tackle challenging transportation problems. This textbookis designed for college or graduate-level students in transportation or closely related fields to study and understand fundamentals in machine learning. Readers will learn how to develop and apply various types of machine learning models to transportation-related problems. Example applications include traffic sensing, data-quality control, traffic prediction, transportation asset management, traffic-system control and operations, and traffic-safety analysis. Introduces fundamental machine learning theories and methodologies Presents state-of-the-art machine learning methodologies and their incorporation into transportationdomain knowledge Includes case studies or examples in each chapter that illustrate the application of methodologies andtechniques for solving transportation problems Provides practice questions following each chapter to enhance understanding and learning Includes class projects to practice coding and the use of the methods