Automated Driving Toolbox™ provides a co-simulation framework that models driving algorithms in Simulink® and visualizes their performance in a 3D environment. This 3D simulation environment uses the Unreal Engine® from Epic Games®.
Simulink blocks related to the 3D simulation environment can be found in the Automated Driving Toolbox > Simulation 3D block library. These blocks provide the ability to:
Configure prebuilt scenes in the 3D simulation environment.
Place and move vehicles within these scenes.
Set up camera, radar, and lidar sensors on the vehicles.
Simulate sensor outputs based on the environment around the vehicle.
Obtain ground truth data for semantic segmentation and depth information.
This simulation tool is commonly used to supplement real data when developing, testing, and verifying the performance of automated driving algorithms. In conjunction with a vehicle model, you can use these blocks to perform realistic closed-loop simulations that encompass the entire automated driving stack, from perception to control.
For more details on the simulation environment, see How 3D Simulation for Automated Driving Works.
To access the Automated Driving Toolbox > Simulation 3D library, at the MATLAB® command prompt, enter:
To configure a model to co-simulate with the 3D simulation environment, add a Simulation 3D Scene Configuration block to the model. Using this block, you can choose from a set of prebuilt 3D scenes where you can test and visualize your driving algorithms. The following image is from the Virtual Mcity scene.
The toolbox includes these scenes.
Straight road segment
|Curved Road||Curved, looped road|
Empty parking lot
|Double Lane Change|
Straight road with barrels and traffic signs that are set up for executing a double lane change maneuver
Flat, black pavement surface with no road objects
|US City Block|
City block with multiple intersections
Highway with cones, barriers, an animal, traffic lights, and traffic signs
|Large Parking Lot|
Parking lot with parked cars, cones, curbs, and traffic signs
City environment that represents the University of Michigan proving grounds (see Mcity Test Facility); includes cones, barriers, an animal, traffic lights, and traffic signs
To define a virtual vehicle in a scene, add a Simulation 3D Vehicle with Ground Following block to your model. Using this block, you can control the movement of the vehicle by supplying the X, Y, and yaw values that define its position and orientation at each time step. The vehicle automatically moves along the ground.
You can also specify the color and type of vehicle. The toolbox includes these vehicle types:
You can define virtual sensors and attach them at various positions on the vehicles. The toolbox includes these sensor modeling and configuration blocks.
|Simulation 3D Camera||Camera model with lens. Includes parameters for image size, focal length, distortion, and skew.|
|Simulation 3D Fisheye Camera||Fisheye camera that can be described using the Scaramuzza camera model. Includes parameters for distortion center, image size, and mapping coefficients.|
|Simulation 3D Lidar||Scanning lidar sensor model. Includes parameters for detection range, resolution, and fields of view.|
|Simulation 3D Probabilistic Radar||Probabilistic radar model that returns a list of detections. Includes parameters for radar accuracy, radar bias, detection probability, and detection reporting. It does not simulate radar at an electromagnetic wave propagation level.|
|Simulation 3D Probabilistic Radar Configuration||Configures radar signatures for all actors detected by the Simulation 3D Probabilistic Radar blocks in a model.|
For more details on choosing a sensor, see Choose a Sensor for 3D Simulation.
Automated Driving Toolbox 3D simulation blocks provide the tools for testing and visualizing path planning, vehicle control, and perception algorithms.
You can use the 3D simulation environment to visualize the motion of a vehicle in
a prebuilt scene. This environment provides you with a way to analyze the
performance of path planning and vehicle control algorithms. After designing these
algorithms in Simulink, you can use the
drivingsim3d library to visualize
vehicle motion in one of the prebuilt scenes.
For an example of path planning and vehicle control algorithm visualization, see Visualize Automated Parking Valet Using 3D Simulation.
Automated Driving Toolbox provides several blocks for detailed camera, radar, and lidar sensor modeling. By mounting these sensors on vehicles within the virtual environment, you can generate synthetic sensor data or sensor detections to test the performance of your sensor models against perception algorithms.
You can also output and visualize ground truth data to validate depth estimation algorithms and train semantic segmentation networks. For an example, see Visualize Depth and Semantic Segmentation Data in 3D Environment.
After you design and test a perception system within the 3D simulation environment, you can then use it to drive a control system that actually steers a vehicle. In this case, rather than manually set up a trajectory, the vehicle uses the perception system to drive itself. By combining perception and control into a closed-loop system in the 3D simulation environment, you can develop and test more complex algorithms, such as lane keeping assist and adaptive cruise control.
For an example that discusses closed-loop simulation in the 3D environment, see Design of Lane Marker Detector in 3D Simulation Environment.