Automated Driving System Toolbox 소개
|
|
- Aldous Nichols
- 6 years ago
- Views:
Transcription
1 1
2 Automated Driving System Toolbox 소개 이제훈차장 2017 The MathWorks, Inc. 2
3 Common Questions from Automated Driving Engineers vehicle How can I Visualize Sensor data? How can I design and verify Perception algorithms? How can I design and verify Sensor fusion? 3
4 Common Questions from Automated Driving Engineers vehicle How can I Visualize Sensor data? How can I design and verify Perception algorithms? How can I design and verify Sensor fusion? 4
5 Automated Driving Sensor data Camera Radar Object Detection Lidar Sensor fusion & Tracking IMU IMU: Inertial Measurement Unit 5
6 Automated Driving Sensor data Camera (640 x 480 x 3) Radar Detector SensorID = 2; Timestamp = ; Vision Detector NumDetections = 23; SensorID = 2541; Detections(1) Timestamp = ; TrackID: Lidar ( x 3) NumDetections = 2546; TrackStatus: Detections(1) Position: [ ] TrackID: Lane Detector Velocity: [ ] Classification: Left 5 Amplitude: Position: IsValid: [ ] Detections(2) Velocity: 255 Confidence: [ ] TrackID: Size: 255 BoundaryType: [ ] TrackStatus: Detections(2) Offset: Position: [ ] TrackID: Vision based 253 HeadingAngle: Inertial Measurement Velocity: [-8.02 Unit ] Classification: 253 Curvature: lane 253 detector Timestamp: Amplitude: Position: [ ] Right Velocity: Detections(3) Velocity: [ ] TrackID: IsValid: YawRate: TrackStatus:
7 7
8 Visualize Sensor data in vehicle coordinates ISO 8855 vehicle axis coordinate system Positive x is forward Positive y is left y x %% Plot in vehicle coordinates ax2 = axes(... 'Position',[ ]); bep = birdseyeplot(... 'Parent',ax2,... 'Xlimits',[0 45],... 'Ylimits',[-10 10]); legend('off'); Plot in vehicle coordinates with birdseyeplot 8
9 Visualize Sensor data - expected coverage area %% Create coverage area plotter covplot = coverageareaplotter(bep,... 'FaceColor','blue',... 'EdgeColor','blue'); %% Update coverage area plotter plotcoveragearea(covplot,... [sensorparams(1).x... % Position x sensorparams(1).y],... % Position y sensorparams(1).range,... sensorparams(1).yawangle,... sensorparams(1).fov(1)) % Field of view Plot sensor coverage area with coverageareaplotter 9
10 Visualize Sensor data - detected objects (vehicle coordinates) %% Create detection plotter detplot = detectionplotter(bep,... 'MarkerEdgeColor','blue',... 'Marker','^'); %% Update detection plotter n = round(currenttime/0.05); numdets = vision(n).numobjects; pos = zeros(numdets,3); vel = zeros(numdets,3); labels = repmat({''},numdets,1); for k = 1:numDets pos(k,:) = vision(n).object(k).position; vel(k,:) = vision(n).object(k).velocity; labels{k} = num2str(... vision(n).object(k).classification); end plotdetection(detplot,pos,vel,labels); Plot vision detections with detectionplotter detectionplotter can be used to visualize vision detector, radar detector, and lidar point cloud 10
11 Visualize Sensor data - detected objects (image coordinates) %% Bounding box positions in image coordinates imboxes = zeros(numdets,4); for k = 1:numDets if vision(n).object(k).classification == 5 vehposlr = vision(n).object(k).position(1:2)'; imposlr = vehicletoimage(sensor, vehposlr); boxheight = 1.4 * 1333 / vehposlr(1); boxwidth = 1.8 * 1333 / vehposlr(1); imboxes(k,:)=[imposlr(1) - boxwidth/2,... imposlr(2) - boxheight,... boxwidth, boxheight]; end end %% Draw bounding boxes on image frame frame = insertobjectannotation(frame,... 'Rectangle', imboxes, labels,... 'Color','yellow','LineWidth',2); im.cdata = frame; Draw boxes with insertobjectannotation Transform vehicle to image coordinates with vehicletoimage 11
12 Learn more about visualizing vehicle data by exploring examples in the Automated Driving System Toolbox Plot object detectors in vehicle coordinates Vision & radar detector Lane detectors Detector coverage areas Transform between vehicle and image coordinates Plot lidar point cloud 12
13 Common Questions from Automated Driving Engineers vehicle How can I Visualize Sensor data? How can I design and verify Perception algorithms? How can I design and verify Sensor fusion? 13
14 Automated Driving Perception Algorithms Pedestrian Detection Vehicle Detection Object Detection: Locate and classify object in image 14
15 MATLAB Tools to Train Detectors Images Label Ground Truth Ground Truth Train detector Object detector imageds = imagedatastore(dir) Easily manage large sets of images - Single line of code to access images - Operates on disk, database, big-data file system 15
16 MATLAB Tools to Train Detectors Images Label Ground Truth Ground Truth Train detector Object detector Automate Labeling of Ground Truth 16
17 MATLAB Tools to Train Detectors Images Label Ground Truth Ground Truth Train detector Object detector Design object detectors with the Computer Vision System Toolbox Machine Learning Deep Learning Aggregate Channel Feature Cascade R-CNN (Regions with Convolutional Neural Networks) Fast R-CNN Faster R-CNN trainacfobjectdetector traincascadeobjectdetector trainrcnnobjectdetector trainfastrcnnobjectdetector trainfasterrcnnobjectdetector 17
18 Designing Perception Algorithms Computer Vision Algorithms for Automated Driving Vehicle Detection Deep learning and ACF based (pre-trained) Pedestrian Detection ACF and HOG/SVM based (pre-trained) 18
19 Designing Perception Algorithms Additional Computer Vision Algorithms for Automated Driving Vehicle detection with distance estimation using mono-camera Lane Detection and Classification RANSAC-based lane boundary fitting Lane boundary visualization 19
20 Designing Perception Algorithms LiDAR Processing Algorithms 20
21 Example of Vision System Detection How can I verify this detection is correct? 21
22 Ground truth labeling to Train Detectors Images Label Ground Truth Ground Truth Train detector Object detector Ground truth labeling to Evaluate Detectors Images Label Ground Truth Ground Truth Evaluate detections Detections 22
23 23
24 Learn more about verifying perception algorithms by exploring examples in the Automated Driving System Toolbox Train object detector using deep learning and machine learning techniques Label detections with Ground Truth Labeler App Extend connectivity of Ground Truth Labeler App 24
25 Common Questions from Automated Driving Engineers vehicle How can I Visualize Sensor data? How can I design and verify Perception algorithms? How can I design and verify Sensor fusion? 25
26 Automated Driving Sensor fusion with radar and vision Can we fuse detections to better track the vehicle? 26
27 27
28 Sensor fusion framework Sensor fusion Framework Object Detections Track Manager Tracking Filter Tracks Time Measurement Measurement Noise Assigns detections to tracks Creates new tracks Updates existing tracks Removes old tracks Predicts and updates state of track Supports linear, extended, and unscented Kalman filters Time State State Covariance Track ID Age Is Confirmed Is Coasted 28
29 Sensor fusion - Data Association Radar Object Vision Object Radar Vision V 1 V 2 R 1 R 2 R m Pairs of visions and associated radars Assignments V 1 + R 2 V 2 + R 1 V n + R m Fusion f(v 1 ) + f(r 2 ) f(v 2 ) + f(r 1 ) f(v n ) + f(r m ) Fused Object List V n costmatrix [assignments, unassignedvisions, unassignedradars] =... assigndetectionstotracks(costmatrix, param.costofnonassignment); 29
30 Sensor fusion - Kalman Filter Initial state & covariance ˆx 0 P 0 Output of updated state xˆ k P k Previous state & covariance ˆ 1 x k P k 1 k k 1 Current becomes previous Time Update ( Predict ) (1) Predict state based on physical model and previous state xˆ k Axˆ k 1 Buk wk (2) Predict error covariance matrix P k AP k 1 A T Q Measurement Update ( Correct ) (1) Compute Kalman gain K k P k H T (HP k H T R) 1 (2) Update estimate state with measurement?x k?x k K k (z k H?x k ) (3) Update the error covariance matrix P k (I K k H) P k K : Kalman gain R : Sensor noise covariance matrix (measurement error) Measurement z k v H u w P Q E[ ww k ek : Control variable matrix : Process (state) noise Hx E[ e x k k k ek ] xˆ T minimize P k T v : Measurement noise ] k k : Process (state) covariance matrix (estimation error) : Process noise covariance matrix A : State matrix relates the state at the previous, k-1 to the state at the current, k : Output matrix relates the state to the measurement From sensor spec or experiment 30
31 Sensor fusion - Kalman Filter Initial state & covariance ˆx 0 P 0 Previous state & covariance ˆ 1 x k P k 1 k k 1 Current becomes previous Time Update ( Predict ) [z_pred,x_pred,p_pred] = predict(obj) z_pred : prediction of measurement x_pred : prediction of state P_pred : state estimation error covariance at the next time step Predicted state x_pred Output of updated state x_corr P_corr Measurement Update ( Correct ) Current Measurement Linear KF (trackingkf) [z_corr,x_corr,p_corr] = correct(obj,z) Extended KF (trackingekf) z_corr : correction of measurement x_corr : correction of state P_corr : state estimation error covariance zunscented KF (trackingukf) Constant velocity initcvkf initcvekf initcvukf Constant acceleration initcakf initcaekf initcaukf Constant turn Not applicable initctekf initctukf 31
32 Synthesize Driving Scenario for Sensor fusion Simulated data for worst-case scenarios Environment Models (Roads and actors) OEM specific test scenarios Sensor Models Fail Operation test scenarios Scenarios identified from real world test drive data 32
33 Synthesize Driving Scenario for Sensor fusion %% Create a new scenario s = drivingscenario('sampletime', 0.05); %% Create road road(s, [ 0 0;... % Centers [x,y] (m) 45 0],... 5); % Width (m) road(s, [35 20; ],... 5); Specify road centers and width as part of a drivingscenario %% Plot scenario p1 = uipanel('position',[ ]); a1 = axes('parent',p1); plot(s,'parent',a1,... 'Centerline','on','Waypoints','on') a1.xlim = [0 45]; a1.ylim = [-6 20]; 33
34 Synthesize Driving Scenario for Sensor fusion %% Add ego vehicle egocar = vehicle(s); waypoints = [ ;... % [x y] (m) ; ; ; ; ]; speed = 13.89; % (m/s) = 50 km/hr path(egocar, waypoints, speed); %% Play scenario while advance(s) pause(s.sampletime); end Specify ego vehicle path using waypoints and speeds 34
35 Synthesize Driving Scenario for Sensor fusion %% Add child pedestrian actor child = actor(s,'length',0.24,... 'Width',0.45,... 'Height',1.7,... 'Position',[40-5 0],... 'Yaw',180); path(child,... [30 15; 40 15],... % Waypoints (m) 1.39); % Speed (m/s) = 5 km/hr Specify pedestrian actor size and path %% Add Target vehicle targetvehicle = vehicle(s); Specify target vehicle with varying speed path(targetvehicle,... [44 1; -4 1],... % Waypoints (m) [5 ; 14]); % Speeds (m/s) 35
36 Synthesize Driving Scenario for Sensor fusion Measurement rate Mounting position on car Most common params, e.g. coverage More params (resolution, bias, etc.) 36
37 Euro NCAP TEST PROTOCOL AEB VRU systems Car-to-VRU Nearside Child (CVNC) Vehicle Travels towards a VRU 5 km/h VRU A child pedestrian crossing the road km/h Scenario Pedestrian`s path running from behind and obstruction vehicle strikes the pedestrian at 50% of the vehicle's width when no braking action is applied. 37
38 Learn more about sensor fusion by exploring examples in the Automated Driving System Toolbox Design multi-object tracker based on logged vehicle data Generate C/C++ code from algorithm which includes a multi-object tracker Synthesize driving scenario to test multi-object tracker 38
39 Common Questions from Automated Driving Engineers vehicle How can I Visualize Sensor data? How can I design and verify Perception algorithms? How can I design and verify Sensor fusion? 39
40 % Thank you 40
Introduction to Automated Driving System Toolbox: Design and Verify Perception Systems
Introduction to Automated Driving System Toolbox: Design and Verify Perception Systems Mark Corless Industry Marketing Automated Driving Segment Manager 2017 The MathWorks, Inc. 1 Some common questions
More informationCase study: Vision and radar-based sensor fusion
Automated Driving System Toolbox Case study: Vision and radar-based sensor fusion Seo-Wook Park Principal Application Engineer The MathWorks, Inc. 2017 The MathWorks, Inc. 1 Automated Driving System Toolbox
More information다중센서기반자율시스템의모델설계및개발 이제훈차장 The MathWorks, Inc. 2
1 다중센서기반자율시스템의모델설계및개발 이제훈차장 2017 The MathWorks, Inc. 2 What we will see today 3 Functional Segmentation of Autonomous System Aircraft/ Platform Sense Perceive Plan & Decide Control Connect/ Communicate
More informationAutomated Driving Development
Automated Driving Development with MATLAB and Simulink MANOHAR REDDY M 2015 The MathWorks, Inc. 1 Using Model-Based Design to develop high quality and reliable Active Safety & Automated Driving Systems
More informationSensor Fusion Using Synthetic Radar and Vision Data
Sensor Fusion Using Synthetic Radar and Vision Data Sensor fusion and control algorithms for automated driving systems require rigorous testing. Vehicle-based testing is not only time consuming to set
More informationTurning an Automated System into an Autonomous system using Model-Based Design Autonomous Tech Conference 2018
Turning an Automated System into an Autonomous system using Model-Based Design Autonomous Tech Conference 2018 Asaf Moses Systematics Ltd., Technical Product Manager aviasafm@systematics.co.il 1 Autonomous
More informationForward Collision Warning Using Sensor Fusion
Forward Collision Warning Using Sensor Fusion Automated driving systems use vision, radar, ultrasound, and combinations of sensor technologies to automate dynamic driving tasks. These tasks include steering,
More informationDeveloping Algorithms for Robotics and Autonomous Systems
Developing Algorithms for Robotics and Autonomous Systems Jorik Caljouw 2015 The MathWorks, Inc. 1 Key Takeaway of this Talk Success in developing an autonomous robotics system requires: 1. Multi-domain
More informationVisual Perception Using Monocular Camera
Visual Perception Using Monocular Camera Automated Driving System Toolbox provides a suite of computer vision algorithms that use data from cameras to detect and track objects of interest such as lane
More informationW4. Perception & Situation Awareness & Decision making
W4. Perception & Situation Awareness & Decision making Robot Perception for Dynamic environments: Outline & DP-Grids concept Dynamic Probabilistic Grids Bayesian Occupancy Filter concept Dynamic Probabilistic
More informationCreating Affordable and Reliable Autonomous Vehicle Systems
Creating Affordable and Reliable Autonomous Vehicle Systems Shaoshan Liu shaoshan.liu@perceptin.io Autonomous Driving Localization Most crucial task of autonomous driving Solutions: GNSS but withvariations,
More informationFusion Framework for Moving-Object Classification. Omar Chavez, Trung-Dung Vu (UJF) Trung-Dung Vu (UJF) Olivier Aycard (UJF) Fabio Tango (CRF)
Fusion Framework for Moving-Object Classification Omar Chavez, Trung-Dung Vu (UJF) Trung-Dung Vu (UJF) Olivier Aycard (UJF) Fabio Tango (CRF) Introduction Advance Driver Assistant Systems (ADAS) help drivers
More informationDesigning a Pick and Place Robotics Application Using MATLAB and Simulink
Designing a Pick and Place Robotics Application Using MATLAB and Simulink Carlos Santacruz-Rosero, PhD Sr Application Engineer Robotics Pulkit Kapur Sr Industry Marketing Manager Robotics 2017 The MathWorks,
More informationS7348: Deep Learning in Ford's Autonomous Vehicles. Bryan Goodman Argo AI 9 May 2017
S7348: Deep Learning in Ford's Autonomous Vehicles Bryan Goodman Argo AI 9 May 2017 1 Ford s 12 Year History in Autonomous Driving Today: examples from Stereo image processing Object detection Using RNN
More informationError Simulation and Multi-Sensor Data Fusion
Error Simulation and Multi-Sensor Data Fusion AERO4701 Space Engineering 3 Week 6 Last Week Looked at the problem of attitude determination for satellites Examined several common methods such as inertial
More informationVehicle Tracking with Heading Estimation using a Mono Camera System
Master of Science Thesis in Electrical Engineering Department of Electrical Engineering, Linköping University, 218 Vehicle Tracking with Heading Estimation using a Mono Camera System Fredrik Nilsson Master
More informationSensory Augmentation for Increased Awareness of Driving Environment
Sensory Augmentation for Increased Awareness of Driving Environment Pranay Agrawal John M. Dolan Dec. 12, 2014 Technologies for Safe and Efficient Transportation (T-SET) UTC The Robotics Institute Carnegie
More information2 OVERVIEW OF RELATED WORK
Utsushi SAKAI Jun OGATA This paper presents a pedestrian detection system based on the fusion of sensors for LIDAR and convolutional neural network based image classification. By using LIDAR our method
More informationMachine learning based automatic extrinsic calibration of an onboard monocular camera for driving assistance applications on smart mobile devices
Technical University of Cluj-Napoca Image Processing and Pattern Recognition Research Center www.cv.utcluj.ro Machine learning based automatic extrinsic calibration of an onboard monocular camera for driving
More informationKorea Autonomous Vehicle Contest 2013
Korea Autonomous Vehicle Contest 2013 Smart Mobility Team (RTOS Lab. & Dyros Lab.) Graduate School of Convergence Science and Technology Seoul National University Page 1 Contents 1. Contest Information
More informationSimulation: A Must for Autonomous Driving
Simulation: A Must for Autonomous Driving NVIDIA GTC 2018 (SILICON VALLEY) / Talk ID: S8859 Rohit Ramanna Business Development Manager Smart Virtual Prototyping, ESI North America Rodolphe Tchalekian EMEA
More informationTeam Aware Perception System using Stereo Vision and Radar
Team Aware Perception System using Stereo Vision and Radar System Development Review 03/ 08 / 2017 Amit Agarwal Harry Golash Yihao Qian Menghan Zhang Zihao (Theo) Zhang Project Description Develop a standalone
More informationOPEN SIMULATION INTERFACE. INTRODUCTION AND OVERVIEW.
OPEN SIMULATION INTERFACE. INTRODUCTION AND OVERVIEW. DIFFERENTIATION OF SIMULATION DATA INTERFACES. Odometry / Dynamics Interface Open Interface (OSI) Map Interface Dynamics Model Map Model Vehicle Dynamics
More informationA System for Real-time Detection and Tracking of Vehicles from a Single Car-mounted Camera
A System for Real-time Detection and Tracking of Vehicles from a Single Car-mounted Camera Claudio Caraffi, Tomas Vojir, Jiri Trefny, Jan Sochman, Jiri Matas Toyota Motor Europe Center for Machine Perception,
More informationTowards Autonomous Vehicle. What is an autonomous vehicle? Vehicle driving on its own with zero mistakes How? Using sensors
7 May 2017 Disclaimer Towards Autonomous Vehicle What is an autonomous vehicle? Vehicle driving on its own with zero mistakes How? Using sensors Why Vision Sensors? Humans use both eyes as main sense
More informationRealtime Object Detection and Segmentation for HD Mapping
Realtime Object Detection and Segmentation for HD Mapping William Raveane Lead AI Engineer Bahram Yoosefizonooz Technical Director NavInfo Europe Advanced Research Lab Presented at GTC Europe 2018 AI in
More informationDEEP NEURAL NETWORKS CHANGING THE AUTONOMOUS VEHICLE LANDSCAPE. Dennis Lui August 2017
DEEP NEURAL NETWORKS CHANGING THE AUTONOMOUS VEHICLE LANDSCAPE Dennis Lui August 2017 THE RISE OF GPU COMPUTING APPLICATIONS 10 7 10 6 GPU-Computing perf 1.5X per year 1000X by 2025 ALGORITHMS 10 5 1.1X
More informationTowards Fully-automated Driving. tue-mps.org. Challenges and Potential Solutions. Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA
Towards Fully-automated Driving Challenges and Potential Solutions Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA Mobile Perception Systems 6 PhDs, 1 postdoc, 1 project manager, 2 software engineers
More informationTraining models for road scene understanding with automated ground truth Dan Levi
Training models for road scene understanding with automated ground truth Dan Levi With: Noa Garnett, Ethan Fetaya, Shai Silberstein, Rafi Cohen, Shaul Oron, Uri Verner, Ariel Ayash, Kobi Horn, Vlad Golder,
More informationObject Fusion for an Advanced Emergency Braking System (AEBS) Jonny Andersson
Object Fusion for an Advanced Emergency Braking System (AEBS) Agenda 1. Rear- end collisions & EU legislation 2. How the AEB system works 3. Object fusion methods 4. Simulink implementation 5. Sensor visualisation
More informationSensor Fusion: Potential, Challenges and Applications. Presented by KVH Industries and Geodetics, Inc. December 2016
Sensor Fusion: Potential, Challenges and Applications Presented by KVH Industries and Geodetics, Inc. December 2016 1 KVH Industries Overview Innovative technology company 600 employees worldwide Focused
More informationVision based autonomous driving - A survey of recent methods. -Tejus Gupta
Vision based autonomous driving - A survey of recent methods -Tejus Gupta Presently, there are three major paradigms for vision based autonomous driving: Directly map input image to driving action using
More informationOn Board 6D Visual Sensors for Intersection Driving Assistance Systems
On Board 6D Visual Sensors for Intersection Driving Assistance Systems S. Nedevschi, T. Marita, R. Danescu, F. Oniga, S. Bota, I. Haller, C. Pantilie, M. Drulea, C. Golban Sergiu.Nedevschi@cs.utcluj.ro
More informationEvaluation of a laser-based reference system for ADAS
23 rd ITS World Congress, Melbourne, Australia, 10 14 October 2016 Paper number ITS- EU-TP0045 Evaluation of a laser-based reference system for ADAS N. Steinhardt 1*, S. Kaufmann 2, S. Rebhan 1, U. Lages
More information2015 The MathWorks, Inc. 1
2015 The MathWorks, Inc. 1 개발에서구현까지 MATLAB 환경에서의딥러닝 김종남 Application Engineer 2015 The MathWorks, Inc. 2 3 Why MATLAB for Deep Learning? MATLAB is Productive MATLAB is Fast MATLAB Integrates with Open Source
More informationA real-time Road Boundary Detection Algorithm Based on Driverless Cars Xuekui ZHU. , Meijuan GAO2, b, Shangnian LI3, c
4th National Conference on Electrical, Electronics and Computer Engineering (NCEECE 2015) A real-time Road Boundary Detection Algorithm Based on Driverless Cars Xuekui ZHU 1, a, Meijuan GAO2, b, Shangnian
More informationPedestrian Detection Using Multi-layer LIDAR
1 st International Conference on Transportation Infrastructure and Materials (ICTIM 2016) ISBN: 978-1-60595-367-0 Pedestrian Detection Using Multi-layer LIDAR Mingfang Zhang 1, Yuping Lu 2 and Tong Liu
More informationMulti camera tracking. Jan Baan. Content. VBM Multicamera VBM A270 test site Helmond
Jan Baan Content VBM Multicamera VBM A270 test site Helmond 2 1 3 Introductie/Doel video vehicle detection image to world tracking video track filtering VBM Database Vehicle data trajectory (x,y,t) dimensions
More informationPedestrian Detection with Radar and Computer Vision
Pedestrian Detection with Radar and Computer Vision camera radar sensor Stefan Milch, Marc Behrens, Darmstadt, September 25 25 / 26, 2001 Pedestrian accidents and protection systems Impact zone: 10% opposite
More informationDetection and Classification of Vehicles
Detection and Classification of Vehicles Gupte et al. 2002 Zeeshan Mohammad ECG 782 Dr. Brendan Morris. Introduction Previously, magnetic loop detectors were used to count vehicles passing over them. Advantages
More informationLive Metric 3D Reconstruction on Mobile Phones ICCV 2013
Live Metric 3D Reconstruction on Mobile Phones ICCV 2013 Main Contents 1. Target & Related Work 2. Main Features of This System 3. System Overview & Workflow 4. Detail of This System 5. Experiments 6.
More informationASTRIUM Space Transportation
SIMU-LANDER Hazard avoidance & advanced GNC for interplanetary descent and soft-landing S. Reynaud, E. Ferreira, S. Trinh, T. Jean-marius 3rd International Workshop on Astrodynamics Tools and Techniques
More informationA Longitudinal Control Algorithm for Smart Cruise Control with Virtual Parameters
ISSN (e): 2250 3005 Volume, 06 Issue, 12 December 2016 International Journal of Computational Engineering Research (IJCER) A Longitudinal Control Algorithm for Smart Cruise Control with Virtual Parameters
More informationCloud-based Large Scale Video Analysis
Cloud-based Large Scale Video Analysis Marcos Nieto Principal Researcher Vicomtech-IK4 Joachim Kreikemeier Manager V-Drive Valeo Schalter und Sensoren GmbH INDEX 1. Cloud-LSVA project 2. ADAS validation
More informationExtended target tracking using PHD filters
Ulm University 2014 01 29 1(35) With applications to video data and laser range data Division of Automatic Control Department of Electrical Engineering Linöping university Linöping, Sweden Presentation
More informationMulti-View 3D Object Detection Network for Autonomous Driving
Multi-View 3D Object Detection Network for Autonomous Driving Xiaozhi Chen, Huimin Ma, Ji Wan, Bo Li, Tian Xia CVPR 2017 (Spotlight) Presented By: Jason Ku Overview Motivation Dataset Network Architecture
More informationAdvanced Driver Assistance: Modular Image Sensor Concept
Vision Advanced Driver Assistance: Modular Image Sensor Concept Supplying value. Integrated Passive and Active Safety Systems Active Safety Passive Safety Scope Reduction of accident probability Get ready
More informationopen robinos specification
open robinos specification architecture, mechanisms, interfaces and testing Version: V 1.0.1 Date: June 29 th, 2016 open robinos specification Version: 1.0.0 Page: 2/42 Table of contents 1 OBJECTIVE...
More informationE80. Experimental Engineering. Lecture 9 Inertial Measurement
Lecture 9 Inertial Measurement http://www.volker-doormann.org/physics.htm Feb. 19, 2013 Christopher M. Clark Where is the rocket? Outline Sensors People Accelerometers Gyroscopes Representations State
More informationRECURRENT NEURAL NETWORKS
RECURRENT NEURAL NETWORKS Methods Traditional Deep-Learning based Non-machine Learning Machine-Learning based method Supervised SVM MLP CNN RNN (LSTM) Localizati on GPS, SLAM Self Driving Perception Pedestrian
More informationnavigation Isaac Skog
Foot-mounted zerovelocity aided inertial navigation Isaac Skog skog@kth.se Course Outline 1. Foot-mounted inertial navigation a. Basic idea b. Pros and cons 2. Inertial navigation a. The inertial sensors
More informationTracking driver actions and guiding phone usage for safer driving. Hongyu Li Jan 25, 2018
Tracking driver actions and guiding phone usage for safer driving Hongyu Li Jan 25, 2018 1 Smart devices risks and opportunities Phone in use 14% Other distractions 86% Distraction-Affected Fatalities
More informationTightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains
Tightly-Integrated Visual and Inertial Navigation for Pinpoint Landing on Rugged Terrains PhD student: Jeff DELAUNE ONERA Director: Guy LE BESNERAIS ONERA Advisors: Jean-Loup FARGES Clément BOURDARIAS
More informationEgomotion Estimation by Point-Cloud Back-Mapping
Egomotion Estimation by Point-Cloud Back-Mapping Haokun Geng, Radu Nicolescu, and Reinhard Klette Department of Computer Science, University of Auckland, New Zealand hgen001@aucklanduni.ac.nz Abstract.
More informationReal Time Multi-Sensor Data Acquisition and Processing for a Road Mapping System
Real Time Multi-Sensor Data Acquisition and Processing for a Road Mapping System by Xiang Luo A thesis submitted for the degree of Master of Engineering (Research) Faculty of Engineering and Information
More informationObject Detection. Part1. Presenter: Dae-Yong
Object Part1 Presenter: Dae-Yong Contents 1. What is an Object? 2. Traditional Object Detector 3. Deep Learning-based Object Detector What is an Object? Subset of Object Recognition What is an Object?
More information컴퓨터비전의최신기술 : Deep Learning, 3D Vision and Embedded Vision
1 컴퓨터비전의최신기술 : Deep Learning, 3D Vision and Embedded Vision 김종남 Application Engineer 2017 The MathWorks, Inc. 2 Three Main Topics New capabilities for computer vision system design: Deep Learning 3-D Vision
More informationDemystifying Deep Learning
Demystifying Deep Learning Let the computers do the hard work Jérémy Huard 2015 The MathWorks, Inc. 1 2 Why MATLAB for Deep Learning? MATLAB is Productive MATLAB is Fast MATLAB Integrates with Open Source
More information(1) and s k ωk. p k vk q
Sensing and Perception: Localization and positioning Isaac Sog Project Assignment: GNSS aided INS In this project assignment you will wor with a type of navigation system referred to as a global navigation
More informationCamera and Inertial Sensor Fusion
January 6, 2018 For First Robotics 2018 Camera and Inertial Sensor Fusion David Zhang david.chao.zhang@gmail.com Version 4.1 1 My Background Ph.D. of Physics - Penn State Univ. Research scientist at SRI
More informationComputer Vision with MATLAB MATLAB Expo 2012 Steve Kuznicki
Computer Vision with MATLAB MATLAB Expo 2012 Steve Kuznicki 2011 The MathWorks, Inc. 1 Today s Topics Introduction Computer Vision Feature-based registration Automatic image registration Object recognition/rotation
More informationCritical Design Review
Critical Design Review Perception System Using Stereo Vision and Radar Team A Amit Agarwal, Harry Golash, Yihao Qian, Menghan Zhang, Zihao (Theo) Zhang Sponsored by: Delphi Automotive December 15, 2016
More informationMULTI-MODAL MAPPING. Robotics Day, 31 Mar Frank Mascarich, Shehryar Khattak, Tung Dang
MULTI-MODAL MAPPING Robotics Day, 31 Mar 2017 Frank Mascarich, Shehryar Khattak, Tung Dang Application-Specific Sensors Cameras TOF Cameras PERCEPTION LiDAR IMU Localization Mapping Autonomy Robotic Perception
More informationSatellite and Inertial Navigation and Positioning System
Satellite and Inertial Navigation and Positioning System Project Proposal By: Luke Pfister Dan Monroe Project Advisors: Dr. In Soo Ahn Dr. Yufeng Lu EE 451 Senior Capstone Project December 10, 2009 PROJECT
More informationPedestrian Path Prediction with Recursive Bayesian Filters: A Comparative Study
Pedestrian Path Prediction with Recursive Bayesian Filters: A Comparative Study N. Schneider 1,2 and D. M. Gavrila 1,2 1 Environment Perception, Daimler R&D, Ulm, Germany 2 Intelligent Systems Laboratory,
More informationOdometry-based Online Extrinsic Sensor Calibration
213 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) November 3-7, 213. Toyo, Japan Odometry-based Online Extrinsic Sensor Calibration Sebastian Schneider, Thorsten Luettel and
More informationVehicle Localization. Hannah Rae Kerner 21 April 2015
Vehicle Localization Hannah Rae Kerner 21 April 2015 Spotted in Mtn View: Google Car Why precision localization? in order for a robot to follow a road, it needs to know where the road is to stay in a particular
More informationStochastic Road Shape Estimation, B. Southall & C. Taylor. Review by: Christopher Rasmussen
Stochastic Road Shape Estimation, B. Southall & C. Taylor Review by: Christopher Rasmussen September 26, 2002 Announcements Readings for next Tuesday: Chapter 14-14.4, 22-22.5 in Forsyth & Ponce Main Contributions
More informationAutomotive 3D Object Detection Without Target Domain Annotations
Master of Science Thesis in Electrical Engineering Department of Electrical Engineering, Linköping University, 2018 Automotive 3D Object Detection Without Target Domain Annotations Erik Linder-Norén and
More informationVision-based Frontal Vehicle Detection and Tracking
Vision-based Frontal and Tracking King Hann LIM, Kah Phooi SENG, Li-Minn ANG and Siew Wen CHIN School of Electrical and Electronic Engineering The University of Nottingham Malaysia campus, Jalan Broga,
More informationDemystifying Deep Learning
Demystifying Deep Learning Mandar Gujrathi Mandar.Gujrathi@mathworks.com.au 2015 The MathWorks, Inc. 1 2 Deep Learning Applications Voice assistants (speech to text) Teaching character to beat video game
More informationTraining models for road scene understanding with automated ground truth Dan Levi
Training models for road scene understanding with automated ground truth Dan Levi With: Noa Garnett, Ethan Fetaya, Shai Silberstein, Rafi Cohen, Shaul Oron, Uri Verner, Ariel Ayash, Kobi Horn, Vlad Golder
More informationFUSION Multitarget-Multisensor Bias Estimation. T. Kirubarajan July Estimation, Tracking and Fusion Laboratory (ETFLab)
Multitarget-Multisensor Bias Estimation Estimation, Tracking and Fusion Laboratory (ETFLab) E. Taghavi, R. Tharmarasa, T. Kirubarajan T. Kirubarajan July 2017 Motivation (1) Fusion of different sensors
More informationSynscapes A photorealistic syntehtic dataset for street scene parsing Jonas Unger Department of Science and Technology Linköpings Universitet.
Synscapes A photorealistic syntehtic dataset for street scene parsing Jonas Unger Department of Science and Technology Linköpings Universitet 7D Labs VINNOVA https://7dlabs.com Photo-realistic image synthesis
More informationDetecting the Unexpected: The Path to Road Obstacles Prevention in Autonomous Driving
Detecting the Unexpected: The Path to Road Obstacles Prevention in Autonomous Driving Shmoolik Mangan, PhD Algorithms Development Manager, VAYAVISION AutonomousTech TLV Israel 2018 VAYAVISION s approach
More informationMobile Robotics. Mathematics, Models, and Methods. HI Cambridge. Alonzo Kelly. Carnegie Mellon University UNIVERSITY PRESS
Mobile Robotics Mathematics, Models, and Methods Alonzo Kelly Carnegie Mellon University HI Cambridge UNIVERSITY PRESS Contents Preface page xiii 1 Introduction 1 1.1 Applications of Mobile Robots 2 1.2
More informationTechnical Paper of HITCSC Team for The International Aerial Robotics Competition
Technical Paper of HITCSC Team for The International Aerial Robotics Competition Qingtao Yu, Yuanhong Li, Hang Yu, Haodi Yao, Rui Xing, Qi Zhao, Xiaowei Xing Harbin Institute of Tecnology ABSTRACT This
More informationDistributed Vision-Aided Cooperative Navigation Based on Three-View Geometry
Distributed Vision-Aided Cooperative Navigation Based on hree-view Geometry Vadim Indelman, Pini Gurfil Distributed Space Systems Lab, Aerospace Engineering, echnion Ehud Rivlin Computer Science, echnion
More informationRadar Detection Improvement by Integration of Multi- Object Tracking
Radar Detection Improvement by Integration of Multi- Object Tracing Lingmin Meng Research and Technology Center Robert Bosch Corp. Pittsburgh, PA, U.S.A. lingmin.meng@rtc.bosch.com Wolfgang Grimm Research
More informationCSE 586 Final Programming Project Spring 2011 Due date: Tuesday, May 3
CSE 586 Final Programming Project Spring 2011 Due date: Tuesday, May 3 What I have in mind for our last programming project is to do something with either graphical models or random sampling. A few ideas
More informationData Association for SLAM
CALIFORNIA INSTITUTE OF TECHNOLOGY ME/CS 132a, Winter 2011 Lab #2 Due: Mar 10th, 2011 Part I Data Association for SLAM 1 Introduction For this part, you will experiment with a simulation of an EKF SLAM
More informationopen robinos specification
architecture, mechanisms, erfaces, and testing 1.1.0, Released Legal disclaimer All brand names, trademarks and registered trademarks are property of their rightful owners and are used only for description.
More informationRelating Local Vision Measurements to Global Navigation Satellite Systems Using Waypoint Based Maps
Relating Local Vision Measurements to Global Navigation Satellite Systems Using Waypoint Based Maps John W. Allen Samuel Gin College of Engineering GPS and Vehicle Dynamics Lab Auburn University Auburn,
More informationTomTom Innovation. Hans Aerts VP Software Development Business Unit Automotive November 2015
TomTom Innovation Hans Aerts VP Software Development Business Unit Automotive November 2015 Empower Movement Simplify complex technology From A to BE Innovative solutions Maps Consumer Connect people and
More informationPedestrian Detection Using Correlated Lidar and Image Data EECS442 Final Project Fall 2016
edestrian Detection Using Correlated Lidar and Image Data EECS442 Final roject Fall 2016 Samuel Rohrer University of Michigan rohrer@umich.edu Ian Lin University of Michigan tiannis@umich.edu Abstract
More informationProbabilistic Sensor Models for Virtual Validation Use Cases and Benefits
Probabilistic Sensor Models for Virtual Validation Use Cases and Benefits Dr. Robin Schubert Co-Founder & CEO BASELABS GmbH 2 BASELABS enables data fusion results. 3 Who we are What we do We partner Data
More informationDealing with Scale. Stephan Weiss Computer Vision Group NASA-JPL / CalTech
Dealing with Scale Stephan Weiss Computer Vision Group NASA-JPL / CalTech Stephan.Weiss@ieee.org (c) 2013. Government sponsorship acknowledged. Outline Why care about size? The IMU as scale provider: The
More information6D-Vision: Fusion of Stereo and Motion for Robust Environment Perception
6D-Vision: Fusion of Stereo and Motion for Robust Environment Perception Uwe Franke, Clemens Rabe, Hernán Badino, and Stefan Gehrig DaimlerChrysler AG, 70546 Stuttgart, Germany {uwe.franke,clemens.rabe,hernan.badino,stefan.gehrig}@daimlerchrysler.com
More informationDeep learning in MATLAB From Concept to CUDA Code
Deep learning in MATLAB From Concept to CUDA Code Roy Fahn Applications Engineer Systematics royf@systematics.co.il 03-7660111 Ram Kokku Principal Engineer MathWorks ram.kokku@mathworks.com 2017 The MathWorks,
More informationMobile Mapping and Navigation. Brad Kohlmeyer NAVTEQ Research
Mobile Mapping and Navigation Brad Kohlmeyer NAVTEQ Research Mobile Mapping & Navigation Markets Automotive Enterprise Internet & Wireless Mobile Devices 2 Local Knowledge & Presence Used to Create Most
More informationENY-C2005 Geoinformation in Environmental Modeling Lecture 4b: Laser scanning
1 ENY-C2005 Geoinformation in Environmental Modeling Lecture 4b: Laser scanning Petri Rönnholm Aalto University 2 Learning objectives To recognize applications of laser scanning To understand principles
More informationFollowing Dirt Roads at Night-Time: Sensors and Features for Lane Recognition and Tracking
Following Dirt Roads at Night-Time: Sensors and Features for Lane Recognition and Tracking Sebastian F. X. Bayerl, Thorsten Luettel and Hans-Joachim Wuensche Abstract The robust perception of roads is
More informationCOSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor
COSC160: Detection and Classification Jeremy Bolton, PhD Assistant Teaching Professor Outline I. Problem I. Strategies II. Features for training III. Using spatial information? IV. Reducing dimensionality
More informationIndoor navigation using smartphones. Chris Hide IESSG, University of Nottingham, UK
Indoor navigation using smartphones Chris Hide IESSG, University of Nottingham, UK Overview Smartphones Available sensors Current positioning methods Positioning research at IESSG 1. Wi-Fi fingerprinting
More informationAttack Resilient State Estimation for Vehicular Systems
December 15 th 2013. T-SET Final Report Attack Resilient State Estimation for Vehicular Systems Nicola Bezzo (nicbezzo@seas.upenn.edu) Prof. Insup Lee (lee@cis.upenn.edu) PRECISE Center University of Pennsylvania
More informationVision Based Tangent Point Detection Algorithm, Evaluation and Validation
15-6 MVA2009 IAPR Conference on Machine Vision Applications, May 20-22, 2009, Yokohama, JAPAN Vision Based Tangent Point Detection Algorithm, Evaluation and Validation Romain Gallen romain.gallen@lcpc.fr
More informationopenpass / OSI World Contents Overview OSI objects and OSI world implementation Date Reinhard Biegel, in-tech
openpass / OSI World Overview OSI objects and OSI world implementation Date 05.04.2018 Author Reinhard Biegel, in-tech Contents 1 Introduction... 2 2 World requirements... 2 3 Architectural overview...
More informationarxiv: v1 [cs.cv] 17 Apr 2019
2D Car Detection in Radar Data with PointNets Andreas Danzer, Thomas Griebel, Martin Bach, and Klaus Dietmayer arxiv:1904.08414v1 [cs.cv] 17 Apr 2019 Abstract For many automated driving functions, a highly
More informationVISION FOR AUTOMOTIVE DRIVING
VISION FOR AUTOMOTIVE DRIVING French Japanese Workshop on Deep Learning & AI, Paris, October 25th, 2017 Quoc Cuong PHAM, PhD Vision and Content Engineering Lab AI & MACHINE LEARNING FOR ADAS AND SELF-DRIVING
More informationSensor Integration and Image Georeferencing for Airborne 3D Mapping Applications
Sensor Integration and Image Georeferencing for Airborne 3D Mapping Applications By Sameh Nassar and Naser El-Sheimy University of Calgary, Canada Contents Background INS/GPS Integration & Direct Georeferencing
More information