Automated Driving System Toolbox 소개

Size: px
Start display at page:

Download "Automated Driving System Toolbox 소개"

Transcription

1 1

2 Automated Driving System Toolbox 소개 이제훈차장 2017 The MathWorks, Inc. 2

3 Common Questions from Automated Driving Engineers vehicle How can I Visualize Sensor data? How can I design and verify Perception algorithms? How can I design and verify Sensor fusion? 3

4 Common Questions from Automated Driving Engineers vehicle How can I Visualize Sensor data? How can I design and verify Perception algorithms? How can I design and verify Sensor fusion? 4

5 Automated Driving Sensor data Camera Radar Object Detection Lidar Sensor fusion & Tracking IMU IMU: Inertial Measurement Unit 5

6 Automated Driving Sensor data Camera (640 x 480 x 3) Radar Detector SensorID = 2; Timestamp = ; Vision Detector NumDetections = 23; SensorID = 2541; Detections(1) Timestamp = ; TrackID: Lidar ( x 3) NumDetections = 2546; TrackStatus: Detections(1) Position: [ ] TrackID: Lane Detector Velocity: [ ] Classification: Left 5 Amplitude: Position: IsValid: [ ] Detections(2) Velocity: 255 Confidence: [ ] TrackID: Size: 255 BoundaryType: [ ] TrackStatus: Detections(2) Offset: Position: [ ] TrackID: Vision based 253 HeadingAngle: Inertial Measurement Velocity: [-8.02 Unit ] Classification: 253 Curvature: lane 253 detector Timestamp: Amplitude: Position: [ ] Right Velocity: Detections(3) Velocity: [ ] TrackID: IsValid: YawRate: TrackStatus:

7 7

8 Visualize Sensor data in vehicle coordinates ISO 8855 vehicle axis coordinate system Positive x is forward Positive y is left y x %% Plot in vehicle coordinates ax2 = axes(... 'Position',[ ]); bep = birdseyeplot(... 'Parent',ax2,... 'Xlimits',[0 45],... 'Ylimits',[-10 10]); legend('off'); Plot in vehicle coordinates with birdseyeplot 8

9 Visualize Sensor data - expected coverage area %% Create coverage area plotter covplot = coverageareaplotter(bep,... 'FaceColor','blue',... 'EdgeColor','blue'); %% Update coverage area plotter plotcoveragearea(covplot,... [sensorparams(1).x... % Position x sensorparams(1).y],... % Position y sensorparams(1).range,... sensorparams(1).yawangle,... sensorparams(1).fov(1)) % Field of view Plot sensor coverage area with coverageareaplotter 9

10 Visualize Sensor data - detected objects (vehicle coordinates) %% Create detection plotter detplot = detectionplotter(bep,... 'MarkerEdgeColor','blue',... 'Marker','^'); %% Update detection plotter n = round(currenttime/0.05); numdets = vision(n).numobjects; pos = zeros(numdets,3); vel = zeros(numdets,3); labels = repmat({''},numdets,1); for k = 1:numDets pos(k,:) = vision(n).object(k).position; vel(k,:) = vision(n).object(k).velocity; labels{k} = num2str(... vision(n).object(k).classification); end plotdetection(detplot,pos,vel,labels); Plot vision detections with detectionplotter detectionplotter can be used to visualize vision detector, radar detector, and lidar point cloud 10

11 Visualize Sensor data - detected objects (image coordinates) %% Bounding box positions in image coordinates imboxes = zeros(numdets,4); for k = 1:numDets if vision(n).object(k).classification == 5 vehposlr = vision(n).object(k).position(1:2)'; imposlr = vehicletoimage(sensor, vehposlr); boxheight = 1.4 * 1333 / vehposlr(1); boxwidth = 1.8 * 1333 / vehposlr(1); imboxes(k,:)=[imposlr(1) - boxwidth/2,... imposlr(2) - boxheight,... boxwidth, boxheight]; end end %% Draw bounding boxes on image frame frame = insertobjectannotation(frame,... 'Rectangle', imboxes, labels,... 'Color','yellow','LineWidth',2); im.cdata = frame; Draw boxes with insertobjectannotation Transform vehicle to image coordinates with vehicletoimage 11

12 Learn more about visualizing vehicle data by exploring examples in the Automated Driving System Toolbox Plot object detectors in vehicle coordinates Vision & radar detector Lane detectors Detector coverage areas Transform between vehicle and image coordinates Plot lidar point cloud 12

13 Common Questions from Automated Driving Engineers vehicle How can I Visualize Sensor data? How can I design and verify Perception algorithms? How can I design and verify Sensor fusion? 13

14 Automated Driving Perception Algorithms Pedestrian Detection Vehicle Detection Object Detection: Locate and classify object in image 14

15 MATLAB Tools to Train Detectors Images Label Ground Truth Ground Truth Train detector Object detector imageds = imagedatastore(dir) Easily manage large sets of images - Single line of code to access images - Operates on disk, database, big-data file system 15

16 MATLAB Tools to Train Detectors Images Label Ground Truth Ground Truth Train detector Object detector Automate Labeling of Ground Truth 16

17 MATLAB Tools to Train Detectors Images Label Ground Truth Ground Truth Train detector Object detector Design object detectors with the Computer Vision System Toolbox Machine Learning Deep Learning Aggregate Channel Feature Cascade R-CNN (Regions with Convolutional Neural Networks) Fast R-CNN Faster R-CNN trainacfobjectdetector traincascadeobjectdetector trainrcnnobjectdetector trainfastrcnnobjectdetector trainfasterrcnnobjectdetector 17

18 Designing Perception Algorithms Computer Vision Algorithms for Automated Driving Vehicle Detection Deep learning and ACF based (pre-trained) Pedestrian Detection ACF and HOG/SVM based (pre-trained) 18

19 Designing Perception Algorithms Additional Computer Vision Algorithms for Automated Driving Vehicle detection with distance estimation using mono-camera Lane Detection and Classification RANSAC-based lane boundary fitting Lane boundary visualization 19

20 Designing Perception Algorithms LiDAR Processing Algorithms 20

21 Example of Vision System Detection How can I verify this detection is correct? 21

22 Ground truth labeling to Train Detectors Images Label Ground Truth Ground Truth Train detector Object detector Ground truth labeling to Evaluate Detectors Images Label Ground Truth Ground Truth Evaluate detections Detections 22

23 23

24 Learn more about verifying perception algorithms by exploring examples in the Automated Driving System Toolbox Train object detector using deep learning and machine learning techniques Label detections with Ground Truth Labeler App Extend connectivity of Ground Truth Labeler App 24

25 Common Questions from Automated Driving Engineers vehicle How can I Visualize Sensor data? How can I design and verify Perception algorithms? How can I design and verify Sensor fusion? 25

26 Automated Driving Sensor fusion with radar and vision Can we fuse detections to better track the vehicle? 26

27 27

28 Sensor fusion framework Sensor fusion Framework Object Detections Track Manager Tracking Filter Tracks Time Measurement Measurement Noise Assigns detections to tracks Creates new tracks Updates existing tracks Removes old tracks Predicts and updates state of track Supports linear, extended, and unscented Kalman filters Time State State Covariance Track ID Age Is Confirmed Is Coasted 28

29 Sensor fusion - Data Association Radar Object Vision Object Radar Vision V 1 V 2 R 1 R 2 R m Pairs of visions and associated radars Assignments V 1 + R 2 V 2 + R 1 V n + R m Fusion f(v 1 ) + f(r 2 ) f(v 2 ) + f(r 1 ) f(v n ) + f(r m ) Fused Object List V n costmatrix [assignments, unassignedvisions, unassignedradars] =... assigndetectionstotracks(costmatrix, param.costofnonassignment); 29

30 Sensor fusion - Kalman Filter Initial state & covariance ˆx 0 P 0 Output of updated state xˆ k P k Previous state & covariance ˆ 1 x k P k 1 k k 1 Current becomes previous Time Update ( Predict ) (1) Predict state based on physical model and previous state xˆ k Axˆ k 1 Buk wk (2) Predict error covariance matrix P k AP k 1 A T Q Measurement Update ( Correct ) (1) Compute Kalman gain K k P k H T (HP k H T R) 1 (2) Update estimate state with measurement?x k?x k K k (z k H?x k ) (3) Update the error covariance matrix P k (I K k H) P k K : Kalman gain R : Sensor noise covariance matrix (measurement error) Measurement z k v H u w P Q E[ ww k ek : Control variable matrix : Process (state) noise Hx E[ e x k k k ek ] xˆ T minimize P k T v : Measurement noise ] k k : Process (state) covariance matrix (estimation error) : Process noise covariance matrix A : State matrix relates the state at the previous, k-1 to the state at the current, k : Output matrix relates the state to the measurement From sensor spec or experiment 30

31 Sensor fusion - Kalman Filter Initial state & covariance ˆx 0 P 0 Previous state & covariance ˆ 1 x k P k 1 k k 1 Current becomes previous Time Update ( Predict ) [z_pred,x_pred,p_pred] = predict(obj) z_pred : prediction of measurement x_pred : prediction of state P_pred : state estimation error covariance at the next time step Predicted state x_pred Output of updated state x_corr P_corr Measurement Update ( Correct ) Current Measurement Linear KF (trackingkf) [z_corr,x_corr,p_corr] = correct(obj,z) Extended KF (trackingekf) z_corr : correction of measurement x_corr : correction of state P_corr : state estimation error covariance zunscented KF (trackingukf) Constant velocity initcvkf initcvekf initcvukf Constant acceleration initcakf initcaekf initcaukf Constant turn Not applicable initctekf initctukf 31

32 Synthesize Driving Scenario for Sensor fusion Simulated data for worst-case scenarios Environment Models (Roads and actors) OEM specific test scenarios Sensor Models Fail Operation test scenarios Scenarios identified from real world test drive data 32

33 Synthesize Driving Scenario for Sensor fusion %% Create a new scenario s = drivingscenario('sampletime', 0.05); %% Create road road(s, [ 0 0;... % Centers [x,y] (m) 45 0],... 5); % Width (m) road(s, [35 20; ],... 5); Specify road centers and width as part of a drivingscenario %% Plot scenario p1 = uipanel('position',[ ]); a1 = axes('parent',p1); plot(s,'parent',a1,... 'Centerline','on','Waypoints','on') a1.xlim = [0 45]; a1.ylim = [-6 20]; 33

34 Synthesize Driving Scenario for Sensor fusion %% Add ego vehicle egocar = vehicle(s); waypoints = [ ;... % [x y] (m) ; ; ; ; ]; speed = 13.89; % (m/s) = 50 km/hr path(egocar, waypoints, speed); %% Play scenario while advance(s) pause(s.sampletime); end Specify ego vehicle path using waypoints and speeds 34

35 Synthesize Driving Scenario for Sensor fusion %% Add child pedestrian actor child = actor(s,'length',0.24,... 'Width',0.45,... 'Height',1.7,... 'Position',[40-5 0],... 'Yaw',180); path(child,... [30 15; 40 15],... % Waypoints (m) 1.39); % Speed (m/s) = 5 km/hr Specify pedestrian actor size and path %% Add Target vehicle targetvehicle = vehicle(s); Specify target vehicle with varying speed path(targetvehicle,... [44 1; -4 1],... % Waypoints (m) [5 ; 14]); % Speeds (m/s) 35

36 Synthesize Driving Scenario for Sensor fusion Measurement rate Mounting position on car Most common params, e.g. coverage More params (resolution, bias, etc.) 36

37 Euro NCAP TEST PROTOCOL AEB VRU systems Car-to-VRU Nearside Child (CVNC) Vehicle Travels towards a VRU 5 km/h VRU A child pedestrian crossing the road km/h Scenario Pedestrian`s path running from behind and obstruction vehicle strikes the pedestrian at 50% of the vehicle's width when no braking action is applied. 37

38 Learn more about sensor fusion by exploring examples in the Automated Driving System Toolbox Design multi-object tracker based on logged vehicle data Generate C/C++ code from algorithm which includes a multi-object tracker Synthesize driving scenario to test multi-object tracker 38

39 Common Questions from Automated Driving Engineers vehicle How can I Visualize Sensor data? How can I design and verify Perception algorithms? How can I design and verify Sensor fusion? 39

40 % Thank you 40

Introduction to Automated Driving System Toolbox: Design and Verify Perception Systems

Introduction to Automated Driving System Toolbox: Design and Verify Perception Systems Introduction to Automated Driving System Toolbox: Design and Verify Perception Systems Mark Corless Industry Marketing Automated Driving Segment Manager 2017 The MathWorks, Inc. 1 Some common questions

More information

Case study: Vision and radar-based sensor fusion

Case study: Vision and radar-based sensor fusion Automated Driving System Toolbox Case study: Vision and radar-based sensor fusion Seo-Wook Park Principal Application Engineer The MathWorks, Inc. 2017 The MathWorks, Inc. 1 Automated Driving System Toolbox

More information

다중센서기반자율시스템의모델설계및개발 이제훈차장 The MathWorks, Inc. 2

다중센서기반자율시스템의모델설계및개발 이제훈차장 The MathWorks, Inc. 2 1 다중센서기반자율시스템의모델설계및개발 이제훈차장 2017 The MathWorks, Inc. 2 What we will see today 3 Functional Segmentation of Autonomous System Aircraft/ Platform Sense Perceive Plan & Decide Control Connect/ Communicate

More information

Automated Driving Development

Automated Driving Development Automated Driving Development with MATLAB and Simulink MANOHAR REDDY M 2015 The MathWorks, Inc. 1 Using Model-Based Design to develop high quality and reliable Active Safety & Automated Driving Systems

More information

Sensor Fusion Using Synthetic Radar and Vision Data

Sensor Fusion Using Synthetic Radar and Vision Data Sensor Fusion Using Synthetic Radar and Vision Data Sensor fusion and control algorithms for automated driving systems require rigorous testing. Vehicle-based testing is not only time consuming to set

More information

Turning an Automated System into an Autonomous system using Model-Based Design Autonomous Tech Conference 2018

Turning an Automated System into an Autonomous system using Model-Based Design Autonomous Tech Conference 2018 Turning an Automated System into an Autonomous system using Model-Based Design Autonomous Tech Conference 2018 Asaf Moses Systematics Ltd., Technical Product Manager aviasafm@systematics.co.il 1 Autonomous

More information

Forward Collision Warning Using Sensor Fusion

Forward Collision Warning Using Sensor Fusion Forward Collision Warning Using Sensor Fusion Automated driving systems use vision, radar, ultrasound, and combinations of sensor technologies to automate dynamic driving tasks. These tasks include steering,

More information

Developing Algorithms for Robotics and Autonomous Systems

Developing Algorithms for Robotics and Autonomous Systems Developing Algorithms for Robotics and Autonomous Systems Jorik Caljouw 2015 The MathWorks, Inc. 1 Key Takeaway of this Talk Success in developing an autonomous robotics system requires: 1. Multi-domain

More information

Visual Perception Using Monocular Camera

Visual Perception Using Monocular Camera Visual Perception Using Monocular Camera Automated Driving System Toolbox provides a suite of computer vision algorithms that use data from cameras to detect and track objects of interest such as lane

More information

W4. Perception & Situation Awareness & Decision making

W4. Perception & Situation Awareness & Decision making W4. Perception & Situation Awareness & Decision making Robot Perception for Dynamic environments: Outline & DP-Grids concept Dynamic Probabilistic Grids Bayesian Occupancy Filter concept Dynamic Probabilistic

More information

Creating Affordable and Reliable Autonomous Vehicle Systems

Creating Affordable and Reliable Autonomous Vehicle Systems Creating Affordable and Reliable Autonomous Vehicle Systems Shaoshan Liu shaoshan.liu@perceptin.io Autonomous Driving Localization Most crucial task of autonomous driving Solutions: GNSS but withvariations,

More information

Fusion Framework for Moving-Object Classification. Omar Chavez, Trung-Dung Vu (UJF) Trung-Dung Vu (UJF) Olivier Aycard (UJF) Fabio Tango (CRF)

Fusion Framework for Moving-Object Classification. Omar Chavez, Trung-Dung Vu (UJF) Trung-Dung Vu (UJF) Olivier Aycard (UJF) Fabio Tango (CRF) Fusion Framework for Moving-Object Classification Omar Chavez, Trung-Dung Vu (UJF) Trung-Dung Vu (UJF) Olivier Aycard (UJF) Fabio Tango (CRF) Introduction Advance Driver Assistant Systems (ADAS) help drivers

More information

Designing a Pick and Place Robotics Application Using MATLAB and Simulink

Designing a Pick and Place Robotics Application Using MATLAB and Simulink Designing a Pick and Place Robotics Application Using MATLAB and Simulink Carlos Santacruz-Rosero, PhD Sr Application Engineer Robotics Pulkit Kapur Sr Industry Marketing Manager Robotics 2017 The MathWorks,

More information

S7348: Deep Learning in Ford's Autonomous Vehicles. Bryan Goodman Argo AI 9 May 2017

S7348: Deep Learning in Ford's Autonomous Vehicles. Bryan Goodman Argo AI 9 May 2017 S7348: Deep Learning in Ford's Autonomous Vehicles Bryan Goodman Argo AI 9 May 2017 1 Ford s 12 Year History in Autonomous Driving Today: examples from Stereo image processing Object detection Using RNN

More information

Error Simulation and Multi-Sensor Data Fusion

Error Simulation and Multi-Sensor Data Fusion Error Simulation and Multi-Sensor Data Fusion AERO4701 Space Engineering 3 Week 6 Last Week Looked at the problem of attitude determination for satellites Examined several common methods such as inertial

More information

Vehicle Tracking with Heading Estimation using a Mono Camera System

Vehicle Tracking with Heading Estimation using a Mono Camera System Master of Science Thesis in Electrical Engineering Department of Electrical Engineering, Linköping University, 218 Vehicle Tracking with Heading Estimation using a Mono Camera System Fredrik Nilsson Master

More information

Sensory Augmentation for Increased Awareness of Driving Environment

Sensory Augmentation for Increased Awareness of Driving Environment Sensory Augmentation for Increased Awareness of Driving Environment Pranay Agrawal John M. Dolan Dec. 12, 2014 Technologies for Safe and Efficient Transportation (T-SET) UTC The Robotics Institute Carnegie

More information

2 OVERVIEW OF RELATED WORK

2 OVERVIEW OF RELATED WORK Utsushi SAKAI Jun OGATA This paper presents a pedestrian detection system based on the fusion of sensors for LIDAR and convolutional neural network based image classification. By using LIDAR our method

More information

Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices

Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices Technical University of Cluj-Napoca Image Processing and Pattern Recognition Research Center www.cv.utcluj.ro Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving

More information

Korea Autonomous Vehicle Contest 2013

Korea Autonomous Vehicle Contest 2013 Korea Autonomous Vehicle Contest 2013 Smart Mobility Team (RTOS Lab. & Dyros Lab.) Graduate School of Convergence Science and Technology Seoul National University Page 1 Contents 1. Contest Information

More information

Simulation: A Must for Autonomous Driving

Simulation: A Must for Autonomous Driving Simulation: A Must for Autonomous Driving NVIDIA GTC 2018 (SILICON VALLEY) / Talk ID: S8859 Rohit Ramanna Business Development Manager Smart Virtual Prototyping, ESI North America Rodolphe Tchalekian EMEA

More information

Team Aware Perception System using Stereo Vision and Radar

Team Aware Perception System using Stereo Vision and Radar Team Aware Perception System using Stereo Vision and Radar System Development Review 03/ 08 / 2017 Amit Agarwal Harry Golash Yihao Qian Menghan Zhang Zihao (Theo) Zhang Project Description Develop a standalone

More information

OPEN SIMULATION INTERFACE. INTRODUCTION AND OVERVIEW.

OPEN SIMULATION INTERFACE. INTRODUCTION AND OVERVIEW. OPEN SIMULATION INTERFACE. INTRODUCTION AND OVERVIEW. DIFFERENTIATION OF SIMULATION DATA INTERFACES. Odometry / Dynamics Interface Open Interface (OSI) Map Interface Dynamics Model Map Model Vehicle Dynamics

More information

A System for Real-time Detection and Tracking of Vehicles from a Single Car-mounted Camera

A System for Real-time Detection and Tracking of Vehicles from a Single Car-mounted Camera A System for Real-time Detection and Tracking of Vehicles from a Single Car-mounted Camera Claudio Caraffi, Tomas Vojir, Jiri Trefny, Jan Sochman, Jiri Matas Toyota Motor Europe Center for Machine Perception,

More information

Towards Autonomous Vehicle. What is an autonomous vehicle? Vehicle driving on its own with zero mistakes How? Using sensors

Towards Autonomous Vehicle. What is an autonomous vehicle? Vehicle driving on its own with zero mistakes How? Using sensors 7 May 2017 Disclaimer Towards Autonomous Vehicle What is an autonomous vehicle? Vehicle driving on its own with zero mistakes How? Using sensors Why Vision Sensors? Humans use both eyes as main sense

More information

Realtime Object Detection and Segmentation for HD Mapping

Realtime Object Detection and Segmentation for HD Mapping Realtime Object Detection and Segmentation for HD Mapping William Raveane Lead AI Engineer Bahram Yoosefizonooz Technical Director NavInfo Europe Advanced Research Lab Presented at GTC Europe 2018 AI in

More information

DEEP NEURAL NETWORKS CHANGING THE AUTONOMOUS VEHICLE LANDSCAPE. Dennis Lui August 2017

DEEP NEURAL NETWORKS CHANGING THE AUTONOMOUS VEHICLE LANDSCAPE. Dennis Lui August 2017 DEEP NEURAL NETWORKS CHANGING THE AUTONOMOUS VEHICLE LANDSCAPE Dennis Lui August 2017 THE RISE OF GPU COMPUTING APPLICATIONS 10 7 10 6 GPU-Computing perf 1.5X per year 1000X by 2025 ALGORITHMS 10 5 1.1X

More information

Towards Fully-automated Driving. tue-mps.org. Challenges and Potential Solutions. Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA

Towards Fully-automated Driving. tue-mps.org. Challenges and Potential Solutions. Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA Towards Fully-automated Driving Challenges and Potential Solutions Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA Mobile Perception Systems 6 PhDs, 1 postdoc, 1 project manager, 2 software engineers

More information

Training models for road scene understanding with automated ground truth Dan Levi

Training models for road scene understanding with automated ground truth Dan Levi Training models for road scene understanding with automated ground truth Dan Levi With: Noa Garnett, Ethan Fetaya, Shai Silberstein, Rafi Cohen, Shaul Oron, Uri Verner, Ariel Ayash, Kobi Horn, Vlad Golder,

More information

Object Fusion for an Advanced Emergency Braking System (AEBS) Jonny Andersson

Object Fusion for an Advanced Emergency Braking System (AEBS) Jonny Andersson Object Fusion for an Advanced Emergency Braking System (AEBS) Agenda 1. Rear- end collisions & EU legislation 2. How the AEB system works 3. Object fusion methods 4. Simulink implementation 5. Sensor visualisation

More information

Sensor Fusion: Potential, Challenges and Applications. Presented by KVH Industries and Geodetics, Inc. December 2016

Sensor Fusion: Potential, Challenges and Applications. Presented by KVH Industries and Geodetics, Inc. December 2016 Sensor Fusion: Potential, Challenges and Applications Presented by KVH Industries and Geodetics, Inc. December 2016 1 KVH Industries Overview Innovative technology company 600 employees worldwide Focused

More information

Vision based autonomous driving - A survey of recent methods. -Tejus Gupta

Vision based autonomous driving - A survey of recent methods. -Tejus Gupta Vision based autonomous driving - A survey of recent methods -Tejus Gupta Presently, there are three major paradigms for vision based autonomous driving: Directly map input image to driving action using

More information

On Board 6D Visual Sensors for Intersection Driving Assistance Systems

On Board 6D Visual Sensors for Intersection Driving Assistance Systems On Board 6D Visual Sensors for Intersection Driving Assistance Systems S. Nedevschi, T. Marita, R. Danescu, F. Oniga, S. Bota, I. Haller, C. Pantilie, M. Drulea, C. Golban Sergiu.Nedevschi@cs.utcluj.ro

More information

Evaluation of a laser-based reference system for ADAS

Evaluation of a laser-based reference system for ADAS 23 rd ITS World Congress, Melbourne, Australia, 10 14 October 2016 Paper number ITS- EU-TP0045 Evaluation of a laser-based reference system for ADAS N. Steinhardt 1*, S. Kaufmann 2, S. Rebhan 1, U. Lages

More information

2015 The MathWorks, Inc. 1

2015 The MathWorks, Inc. 1 2015 The MathWorks, Inc. 1 개발에서구현까지 MATLAB 환경에서의딥러닝 김종남 Application Engineer 2015 The MathWorks, Inc. 2 3 Why MATLAB for Deep Learning? MATLAB is Productive MATLAB is Fast MATLAB Integrates with Open Source

More information

A real-time Road Boundary Detection Algorithm Based on Driverless Cars Xuekui ZHU. , Meijuan GAO2, b, Shangnian LI3, c

A real-time Road Boundary Detection Algorithm Based on Driverless Cars Xuekui ZHU. , Meijuan GAO2, b, Shangnian LI3, c 4th National Conference on Electrical, Electronics and Computer Engineering (NCEECE 2015) A real-time Road Boundary Detection Algorithm Based on Driverless Cars Xuekui ZHU 1, a, Meijuan GAO2, b, Shangnian

More information

Pedestrian Detection Using Multi-layer LIDAR

Pedestrian Detection Using Multi-layer LIDAR 1 st International Conference on Transportation Infrastructure and Materials (ICTIM 2016) ISBN: 978-1-60595-367-0 Pedestrian Detection Using Multi-layer LIDAR Mingfang Zhang 1, Yuping Lu 2 and Tong Liu

More information

Multi camera tracking. Jan Baan. Content. VBM Multicamera VBM A270 test site Helmond

Multi camera tracking. Jan Baan. Content. VBM Multicamera VBM A270 test site Helmond Jan Baan Content VBM Multicamera VBM A270 test site Helmond 2 1 3 Introductie/Doel video vehicle detection image to world tracking video track filtering VBM Database Vehicle data trajectory (x,y,t) dimensions

More information

Pedestrian Detection with Radar and Computer Vision

Pedestrian Detection with Radar and Computer Vision Pedestrian Detection with Radar and Computer Vision camera radar sensor Stefan Milch, Marc Behrens, Darmstadt, September 25 25 / 26, 2001 Pedestrian accidents and protection systems Impact zone: 10% opposite

More information

Detection and Classification of Vehicles

Detection and Classification of Vehicles Detection and Classification of Vehicles Gupte et al. 2002 Zeeshan Mohammad ECG 782 Dr. Brendan Morris. Introduction Previously, magnetic loop detectors were used to count vehicles passing over them. Advantages

More information

Live Metric 3D Reconstruction on Mobile Phones ICCV 2013

Live Metric 3D Reconstruction on Mobile Phones ICCV 2013 Live Metric 3D Reconstruction on Mobile Phones ICCV 2013 Main Contents 1. Target & Related Work 2. Main Features of This System 3. System Overview & Workflow 4. Detail of This System 5. Experiments 6.

More information

ASTRIUM Space Transportation

ASTRIUM Space Transportation SIMU-LANDER Hazard avoidance & advanced GNC for interplanetary descent and soft-landing S. Reynaud, E. Ferreira, S. Trinh, T. Jean-marius 3rd International Workshop on Astrodynamics Tools and Techniques

More information

A Longitudinal Control Algorithm for Smart Cruise Control with Virtual Parameters

A Longitudinal Control Algorithm for Smart Cruise Control with Virtual Parameters ISSN (e): 2250 3005 Volume, 06 Issue, 12 December 2016 International Journal of Computational Engineering Research (IJCER) A Longitudinal Control Algorithm for Smart Cruise Control with Virtual Parameters

More information

Cloud-based Large Scale Video Analysis

Cloud-based Large Scale Video Analysis Cloud-based Large Scale Video Analysis Marcos Nieto Principal Researcher Vicomtech-IK4 Joachim Kreikemeier Manager V-Drive Valeo Schalter und Sensoren GmbH INDEX 1. Cloud-LSVA project 2. ADAS validation

More information

Extended target tracking using PHD filters

Extended target tracking using PHD filters Ulm University 2014 01 29 1(35) With applications to video data and laser range data Division of Automatic Control Department of Electrical Engineering Linöping university Linöping, Sweden Presentation

More information

Multi-View 3D Object Detection Network for Autonomous Driving

Multi-View 3D Object Detection Network for Autonomous Driving Multi-View 3D Object Detection Network for Autonomous Driving Xiaozhi Chen, Huimin Ma, Ji Wan, Bo Li, Tian Xia CVPR 2017 (Spotlight) Presented By: Jason Ku Overview Motivation Dataset Network Architecture

More information

Advanced Driver Assistance: Modular Image Sensor Concept

Advanced Driver Assistance: Modular Image Sensor Concept Vision Advanced Driver Assistance: Modular Image Sensor Concept Supplying value. Integrated Passive and Active Safety Systems Active Safety Passive Safety Scope Reduction of accident probability Get ready

More information

open robinos specification

open robinos specification open robinos specification architecture, mechanisms, interfaces and testing Version: V 1.0.1 Date: June 29 th, 2016 open robinos specification Version: 1.0.0 Page: 2/42 Table of contents 1 OBJECTIVE...

More information

E80. Experimental Engineering. Lecture 9 Inertial Measurement

E80. Experimental Engineering. Lecture 9 Inertial Measurement Lecture 9 Inertial Measurement http://www.volker-doormann.org/physics.htm Feb. 19, 2013 Christopher M. Clark Where is the rocket? Outline Sensors People Accelerometers Gyroscopes Representations State

More information

RECURRENT NEURAL NETWORKS

RECURRENT NEURAL NETWORKS RECURRENT NEURAL NETWORKS Methods Traditional Deep-Learning based Non-machine Learning Machine-Learning based method Supervised SVM MLP CNN RNN (LSTM) Localizati on GPS, SLAM Self Driving Perception Pedestrian

More information

navigation Isaac Skog

navigation Isaac Skog Foot-mounted zerovelocity aided inertial navigation Isaac Skog skog@kth.se Course Outline 1. Foot-mounted inertial navigation a. Basic idea b. Pros and cons 2. Inertial navigation a. The inertial sensors

More information

Tracking driver actions and guiding phone usage for safer driving. Hongyu Li Jan 25, 2018

Tracking driver actions and guiding phone usage for safer driving. Hongyu Li Jan 25, 2018 Tracking driver actions and guiding phone usage for safer driving Hongyu Li Jan 25, 2018 1 Smart devices risks and opportunities Phone in use 14% Other distractions 86% Distraction-Affected Fatalities

More information

Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains

Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains PhD student: Jeff DELAUNE ONERA Director: Guy LE BESNERAIS ONERA Advisors: Jean-Loup FARGES Clément BOURDARIAS

More information

Egomotion Estimation by Point-Cloud Back-Mapping

Egomotion Estimation by Point-Cloud Back-Mapping Egomotion Estimation by Point-Cloud Back-Mapping Haokun Geng, Radu Nicolescu, and Reinhard Klette Department of Computer Science, University of Auckland, New Zealand hgen001@aucklanduni.ac.nz Abstract.

More information

Real Time Multi-Sensor Data Acquisition and Processing for a Road Mapping System

Real Time Multi-Sensor Data Acquisition and Processing for a Road Mapping System Real Time Multi-Sensor Data Acquisition and Processing for a Road Mapping System by Xiang Luo A thesis submitted for the degree of Master of Engineering (Research) Faculty of Engineering and Information

More information

Object Detection. Part1. Presenter: Dae-Yong

Object Detection. Part1. Presenter: Dae-Yong Object Part1 Presenter: Dae-Yong Contents 1. What is an Object? 2. Traditional Object Detector 3. Deep Learning-based Object Detector What is an Object? Subset of Object Recognition What is an Object?

More information

컴퓨터비전의최신기술 : Deep Learning, 3D Vision and Embedded Vision

컴퓨터비전의최신기술 : Deep Learning, 3D Vision and Embedded Vision 1 컴퓨터비전의최신기술 : Deep Learning, 3D Vision and Embedded Vision 김종남 Application Engineer 2017 The MathWorks, Inc. 2 Three Main Topics New capabilities for computer vision system design: Deep Learning 3-D Vision

More information

Demystifying Deep Learning

Demystifying Deep Learning Demystifying Deep Learning Let the computers do the hard work Jérémy Huard 2015 The MathWorks, Inc. 1 2 Why MATLAB for Deep Learning? MATLAB is Productive MATLAB is Fast MATLAB Integrates with Open Source

More information

(1) and s k ωk. p k vk q

(1) and s k ωk. p k vk q Sensing and Perception: Localization and positioning Isaac Sog Project Assignment: GNSS aided INS In this project assignment you will wor with a type of navigation system referred to as a global navigation

More information

Camera and Inertial Sensor Fusion

Camera and Inertial Sensor Fusion January 6, 2018 For First Robotics 2018 Camera and Inertial Sensor Fusion David Zhang david.chao.zhang@gmail.com Version 4.1 1 My Background Ph.D. of Physics - Penn State Univ. Research scientist at SRI

More information

Computer Vision with MATLAB MATLAB Expo 2012 Steve Kuznicki

Computer Vision with MATLAB MATLAB Expo 2012 Steve Kuznicki Computer Vision with MATLAB MATLAB Expo 2012 Steve Kuznicki 2011 The MathWorks, Inc. 1 Today s Topics Introduction Computer Vision Feature-based registration Automatic image registration Object recognition/rotation

More information

Critical Design Review

Critical Design Review Critical Design Review Perception System Using Stereo Vision and Radar Team A Amit Agarwal, Harry Golash, Yihao Qian, Menghan Zhang, Zihao (Theo) Zhang Sponsored by: Delphi Automotive December 15, 2016

More information

MULTI-MODAL MAPPING. Robotics Day, 31 Mar Frank Mascarich, Shehryar Khattak, Tung Dang

MULTI-MODAL MAPPING. Robotics Day, 31 Mar Frank Mascarich, Shehryar Khattak, Tung Dang MULTI-MODAL MAPPING Robotics Day, 31 Mar 2017 Frank Mascarich, Shehryar Khattak, Tung Dang Application-Specific Sensors Cameras TOF Cameras PERCEPTION LiDAR IMU Localization Mapping Autonomy Robotic Perception

More information

Satellite and Inertial Navigation and Positioning System

Satellite and Inertial Navigation and Positioning System Satellite and Inertial Navigation and Positioning System Project Proposal By: Luke Pfister Dan Monroe Project Advisors: Dr. In Soo Ahn Dr. Yufeng Lu EE 451 Senior Capstone Project December 10, 2009 PROJECT

More information

Pedestrian Path Prediction with Recursive Bayesian Filters: A Comparative Study

Pedestrian Path Prediction with Recursive Bayesian Filters: A Comparative Study Pedestrian Path Prediction with Recursive Bayesian Filters: A Comparative Study N. Schneider 1,2 and D. M. Gavrila 1,2 1 Environment Perception, Daimler R&D, Ulm, Germany 2 Intelligent Systems Laboratory,

More information

Odometry-based Online Extrinsic Sensor Calibration

Odometry-based Online Extrinsic Sensor Calibration 213 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) November 3-7, 213. Toyo, Japan Odometry-based Online Extrinsic Sensor Calibration Sebastian Schneider, Thorsten Luettel and

More information

Vehicle Localization. Hannah Rae Kerner 21 April 2015

Vehicle Localization. Hannah Rae Kerner 21 April 2015 Vehicle Localization Hannah Rae Kerner 21 April 2015 Spotted in Mtn View: Google Car Why precision localization? in order for a robot to follow a road, it needs to know where the road is to stay in a particular

More information

Stochastic Road Shape Estimation, B. Southall & C. Taylor. Review by: Christopher Rasmussen

Stochastic Road Shape Estimation, B. Southall & C. Taylor. Review by: Christopher Rasmussen Stochastic Road Shape Estimation, B. Southall & C. Taylor Review by: Christopher Rasmussen September 26, 2002 Announcements Readings for next Tuesday: Chapter 14-14.4, 22-22.5 in Forsyth & Ponce Main Contributions

More information

Automotive 3D Object Detection Without Target Domain Annotations

Automotive 3D Object Detection Without Target Domain Annotations Master of Science Thesis in Electrical Engineering Department of Electrical Engineering, Linköping University, 2018 Automotive 3D Object Detection Without Target Domain Annotations Erik Linder-Norén and

More information

Vision-based Frontal Vehicle Detection and Tracking

Vision-based Frontal Vehicle Detection and Tracking Vision-based Frontal and Tracking King Hann LIM, Kah Phooi SENG, Li-Minn ANG and Siew Wen CHIN School of Electrical and Electronic Engineering The University of Nottingham Malaysia campus, Jalan Broga,

More information

Demystifying Deep Learning

Demystifying Deep Learning Demystifying Deep Learning Mandar Gujrathi Mandar.Gujrathi@mathworks.com.au 2015 The MathWorks, Inc. 1 2 Deep Learning Applications Voice assistants (speech to text) Teaching character to beat video game

More information

Training models for road scene understanding with automated ground truth Dan Levi

Training models for road scene understanding with automated ground truth Dan Levi Training models for road scene understanding with automated ground truth Dan Levi With: Noa Garnett, Ethan Fetaya, Shai Silberstein, Rafi Cohen, Shaul Oron, Uri Verner, Ariel Ayash, Kobi Horn, Vlad Golder

More information

FUSION Multitarget-Multisensor Bias Estimation. T. Kirubarajan July Estimation, Tracking and Fusion Laboratory (ETFLab)

FUSION Multitarget-Multisensor Bias Estimation. T. Kirubarajan July Estimation, Tracking and Fusion Laboratory (ETFLab) Multitarget-Multisensor Bias Estimation Estimation, Tracking and Fusion Laboratory (ETFLab) E. Taghavi, R. Tharmarasa, T. Kirubarajan T. Kirubarajan July 2017 Motivation (1) Fusion of different sensors

More information

Synscapes A photorealistic syntehtic dataset for street scene parsing Jonas Unger Department of Science and Technology Linköpings Universitet.

Synscapes A photorealistic syntehtic dataset for street scene parsing Jonas Unger Department of Science and Technology Linköpings Universitet. Synscapes A photorealistic syntehtic dataset for street scene parsing Jonas Unger Department of Science and Technology Linköpings Universitet 7D Labs VINNOVA https://7dlabs.com Photo-realistic image synthesis

More information

Detecting the Unexpected: The Path to Road Obstacles Prevention in Autonomous Driving

Detecting the Unexpected: The Path to Road Obstacles Prevention in Autonomous Driving Detecting the Unexpected: The Path to Road Obstacles Prevention in Autonomous Driving Shmoolik Mangan, PhD Algorithms Development Manager, VAYAVISION AutonomousTech TLV Israel 2018 VAYAVISION s approach

More information

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS

Mobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS Mobile Robotics Mathematics, Models, and Methods Alonzo Kelly Carnegie Mellon University HI Cambridge UNIVERSITY PRESS Contents Preface page xiii 1 Introduction 1 1.1 Applications of Mobile Robots 2 1.2

More information

Technical Paper of HITCSC Team for The International Aerial Robotics Competition

Technical Paper of HITCSC Team for The International Aerial Robotics Competition Technical Paper of HITCSC Team for The International Aerial Robotics Competition Qingtao Yu, Yuanhong Li, Hang Yu, Haodi Yao, Rui Xing, Qi Zhao, Xiaowei Xing Harbin Institute of Tecnology ABSTRACT This

More information

Distributed Vision-Aided Cooperative Navigation Based on Three-View Geometry

Distributed Vision-Aided Cooperative Navigation Based on Three-View Geometry Distributed Vision-Aided Cooperative Navigation Based on hree-view Geometry Vadim Indelman, Pini Gurfil Distributed Space Systems Lab, Aerospace Engineering, echnion Ehud Rivlin Computer Science, echnion

More information

Radar Detection Improvement by Integration of Multi- Object Tracking

Radar Detection Improvement by Integration of Multi- Object Tracking Radar Detection Improvement by Integration of Multi- Object Tracing Lingmin Meng Research and Technology Center Robert Bosch Corp. Pittsburgh, PA, U.S.A. lingmin.meng@rtc.bosch.com Wolfgang Grimm Research

More information

CSE 586 Final Programming Project Spring 2011 Due date: Tuesday, May 3

CSE 586 Final Programming Project Spring 2011 Due date: Tuesday, May 3 CSE 586 Final Programming Project Spring 2011 Due date: Tuesday, May 3 What I have in mind for our last programming project is to do something with either graphical models or random sampling. A few ideas

More information

Data Association for SLAM

Data Association for SLAM CALIFORNIA INSTITUTE OF TECHNOLOGY ME/CS 132a, Winter 2011 Lab #2 Due: Mar 10th, 2011 Part I Data Association for SLAM 1 Introduction For this part, you will experiment with a simulation of an EKF SLAM

More information

open robinos specification

open robinos specification architecture, mechanisms, erfaces, and testing 1.1.0, Released Legal disclaimer All brand names, trademarks and registered trademarks are property of their rightful owners and are used only for description.

More information

Relating Local Vision Measurements to Global Navigation Satellite Systems Using Waypoint Based Maps

Relating Local Vision Measurements to Global Navigation Satellite Systems Using Waypoint Based Maps Relating Local Vision Measurements to Global Navigation Satellite Systems Using Waypoint Based Maps John W. Allen Samuel Gin College of Engineering GPS and Vehicle Dynamics Lab Auburn University Auburn,

More information

TomTom Innovation. Hans Aerts VP Software Development Business Unit Automotive November 2015

TomTom Innovation. Hans Aerts VP Software Development Business Unit Automotive November 2015 TomTom Innovation Hans Aerts VP Software Development Business Unit Automotive November 2015 Empower Movement Simplify complex technology From A to BE Innovative solutions Maps Consumer Connect people and

More information

Pedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016

Pedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016 edestrian Detection Using Correlated Lidar and Image Data EECS442 Final roject Fall 2016 Samuel Rohrer University of Michigan rohrer@umich.edu Ian Lin University of Michigan tiannis@umich.edu Abstract

More information

Probabilistic Sensor Models for Virtual Validation Use Cases and Benefits

Probabilistic Sensor Models for Virtual Validation Use Cases and Benefits Probabilistic Sensor Models for Virtual Validation Use Cases and Benefits Dr. Robin Schubert Co-Founder & CEO BASELABS GmbH 2 BASELABS enables data fusion results. 3 Who we are What we do We partner Data

More information

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech

Dealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech Dealing with Scale Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why care about size? The IMU as scale provider: The

More information

6D-Vision: Fusion of Stereo and Motion for Robust Environment Perception

6D-Vision: Fusion of Stereo and Motion for Robust Environment Perception 6D-Vision: Fusion of Stereo and Motion for Robust Environment Perception Uwe Franke, Clemens Rabe, Hernán Badino, and Stefan Gehrig DaimlerChrysler AG, 70546 Stuttgart, Germany {uwe.franke,clemens.rabe,hernan.badino,stefan.gehrig}@daimlerchrysler.com

More information

Deep learning in MATLAB From Concept to CUDA Code

Deep learning in MATLAB From Concept to CUDA Code Deep learning in MATLAB From Concept to CUDA Code Roy Fahn Applications Engineer Systematics royf@systematics.co.il 03-7660111 Ram Kokku Principal Engineer MathWorks ram.kokku@mathworks.com 2017 The MathWorks,

More information

Mobile Mapping and Navigation. Brad Kohlmeyer NAVTEQ Research

Mobile Mapping and Navigation. Brad Kohlmeyer NAVTEQ Research Mobile Mapping and Navigation Brad Kohlmeyer NAVTEQ Research Mobile Mapping & Navigation Markets Automotive Enterprise Internet & Wireless Mobile Devices 2 Local Knowledge & Presence Used to Create Most

More information

ENY-C2005 Geoinformation in Environmental Modeling Lecture 4b: Laser scanning

ENY-C2005 Geoinformation in Environmental Modeling Lecture 4b: Laser scanning 1 ENY-C2005 Geoinformation in Environmental Modeling Lecture 4b: Laser scanning Petri Rönnholm Aalto University 2 Learning objectives To recognize applications of laser scanning To understand principles

More information

Following Dirt Roads at Night-Time: Sensors and Features for Lane Recognition and Tracking

Following Dirt Roads at Night-Time: Sensors and Features for Lane Recognition and Tracking Following Dirt Roads at Night-Time: Sensors and Features for Lane Recognition and Tracking Sebastian F. X. Bayerl, Thorsten Luettel and Hans-Joachim Wuensche Abstract The robust perception of roads is

More information

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor COSC160: Detection and Classification Jeremy Bolton, PhD Assistant Teaching Professor Outline I. Problem I. Strategies II. Features for training III. Using spatial information? IV. Reducing dimensionality

More information

Indoor navigation using smartphones. Chris Hide IESSG, University of Nottingham, UK

Indoor navigation using smartphones. Chris Hide IESSG, University of Nottingham, UK Indoor navigation using smartphones Chris Hide IESSG, University of Nottingham, UK Overview Smartphones Available sensors Current positioning methods Positioning research at IESSG 1. Wi-Fi fingerprinting

More information

Attack Resilient State Estimation for Vehicular Systems

Attack Resilient State Estimation for Vehicular Systems December 15 th 2013. T-SET Final Report Attack Resilient State Estimation for Vehicular Systems Nicola Bezzo (nicbezzo@seas.upenn.edu) Prof. Insup Lee (lee@cis.upenn.edu) PRECISE Center University of Pennsylvania

More information

Vision Based Tangent Point Detection Algorithm, Evaluation and Validation

Vision Based Tangent Point Detection Algorithm, Evaluation and Validation 15-6 MVA2009 IAPR Conference on Machine Vision Applications, May 20-22, 2009, Yokohama, JAPAN Vision Based Tangent Point Detection Algorithm, Evaluation and Validation Romain Gallen romain.gallen@lcpc.fr

More information

openpass / OSI World Contents Overview OSI objects and OSI world implementation Date Reinhard Biegel, in-tech

openpass / OSI World Contents Overview OSI objects and OSI world implementation Date Reinhard Biegel, in-tech openpass / OSI World Overview OSI objects and OSI world implementation Date 05.04.2018 Author Reinhard Biegel, in-tech Contents 1 Introduction... 2 2 World requirements... 2 3 Architectural overview...

More information

arxiv: v1 [cs.cv] 17 Apr 2019

arxiv: v1 [cs.cv] 17 Apr 2019 2D Car Detection in Radar Data with PointNets Andreas Danzer, Thomas Griebel, Martin Bach, and Klaus Dietmayer arxiv:1904.08414v1 [cs.cv] 17 Apr 2019 Abstract For many automated driving functions, a highly

More information

VISION FOR AUTOMOTIVE DRIVING

VISION FOR AUTOMOTIVE DRIVING VISION FOR AUTOMOTIVE DRIVING French Japanese Workshop on Deep Learning & AI, Paris, October 25th, 2017 Quoc Cuong PHAM, PhD Vision and Content Engineering Lab AI & MACHINE LEARNING FOR ADAS AND SELF-DRIVING

More information

Sensor Integration and Image Georeferencing for Airborne 3D Mapping Applications

Sensor Integration and Image Georeferencing for Airborne 3D Mapping Applications Sensor Integration and Image Georeferencing for Airborne 3D Mapping Applications By Sameh Nassar and Naser El-Sheimy University of Calgary, Canada Contents Background INS/GPS Integration & Direct Georeferencing

More information