sensor fusion of Udacity

2023-10-22 06:49
文章标签 sensor fusion udacity

本文主要是介绍sensor fusion of Udacity,希望对大家解决编程问题提供一定的参考价值,需要的开发者们随着小编来一起学习吧!

文章目录

        • Radar Principles
          • 01 why use Radar
          • 02 the radar sensor
        • Camera
          • 06 Combining Camera and Lidar

Radar Principles
01 why use Radar

在这里插入图片描述
在这里插入图片描述
在这里插入图片描述
在这里插入图片描述
同时Radar能够在多雨多雪的环境下使用

02 the radar sensor

在这里插入图片描述

Camera
06 Combining Camera and Lidar

这块投影的公式见知乎willian
我这里结合透视投影做一个介绍:
这个转换的前提是camera的坐标系是:天东北
(1)小孔成像,三维坐标转公制坐标(m)
x ′ = f x z x{'}=f \frac xz x=fzx
y ′ = f y z y{'}=f \frac yz y=fzy

(2) 公制坐标转换成像素坐标(图像中心转到左上角):
( x , y , z ) T − > f ∗ k ∗ x z + C x , f ∗ l ∗ y z + C y (x,y,z)^{T}->f *k*\frac xz + C_{x}, f *l*\frac yz + C_{y} (x,y,z)T>fkzx+Cx,flzy+Cy
其中k,l为参数:pixel/m, 即多少pixel每米

结合透视投影的例子:
选择图像像素为:800*600

k = 800 / ( t a n ( 4 0 ο ) ∗ f ∗ 2 ) p i x e l / m k=800/(tan(40^{\omicron})*f*2) pixel/m k=800/(tan(40ο)f2)pixel/m
40是arithum_fov/2

l = 600 / ( t a n ( 2 0 ο ) ∗ f ∗ 2 ) p i x e l / m l=600/(tan(20^{\omicron})*f*2) pixel/m l=600/(tan(20ο)f2)pixel/m
20是垂直_fov/2
而如果我们做转换是是需要求
f ∗ k 与 f ∗ l f*k与f*l fkfl
这样f就约掉了

因此在本例子中需要确认的是:
Lidar的FOV
Lidar水平与垂直的像素点尺寸组合
(图打通lidar像素水平与垂直比例为4:3因此确定像素组合:800600、400300 or 1200*900这个尺寸都是去测试,看效果如何(python的调试工程)
4:3怎么来的就是点数之比:
水平:65/0.16=403
垂直:40/0.13=300

多传感器融合的论文推荐:

Additional Resources on Sensor Fusion and Object Detection & Tracking

Nice work reaching the end of the sensor fusion content! While you still have the project left to do here, we’re also providing some additional resources and recent research on the topic that you can come back to if you have time later on.

Reading research papers is a great way to get exposure to the latest and greatest in the field, as well as expand your learning. However, just like the project ahead, it’s often best to learn by doing - if you find a paper that really excites you, try to implement it (or even something better) yourself!
Optional Reading

All of these are completely optional reading - you could spend days reading through the entirety of these! We suggest moving onto the project first so you have Kalman Filters fresh on your mind, before coming back to check these out.

We’ve categorized these papers to hopefully help you narrow down which ones might be of interest, as well as highlighted a couple key reads by category by including their Abstract section, which summarizes the paper. We’ve also included some additional papers you might consider as well if you want to delve even deeper.
Tracking Multiple Objects and Sensor Fusion

The below papers and resources concern tracking multiple objects, using Kalman Filters as well as other techniques!

No Blind Spots: Full-Surround Multi-Object Tracking for Autonomous Vehicles using Cameras & LiDARs by A. Rangesh and M. Trivedi

Abstract: Online multi-object tracking (MOT) is extremely important for high-level spatial reasoning and path planning for autonomous and highly-automated vehicles. In this paper, we present a modular framework for tracking multiple objects (vehicles), capable of accepting object proposals from different
sensor modalities (vision and range) and a variable number of sensors, to produce continuous object tracks. […]
We demonstrate that our framework is well-suited to track objects through entire maneuvers around the ego-vehicle, some of which take more than a few minutes to complete. We also leverage the modularity of our approach by comparing the effects of including/excluding different sensors, changing the total number of sensors, and the quality of object proposals on the final tracking result.

Multiple Sensor Fusion and Classification for Moving Object Detection and Tracking by R.O. Chavez-Garcia and O. Aycard

Abstract: […] We believe that by including the objects classification from multiple sensors detections as a key component of the object’s representation and the perception process, we can improve the perceived model of the environment. First, we define a composite object representation to
include class information in the core object’s description. Second, we propose a complete perception fusion architecture based on the Evidential framework to solve the Detection and Tracking of Moving Objects (DATMO) problem by integrating the composite representation and uncertainty management. Finally, we integrate our fusion approach in a real-time application inside a vehicle demonstrator from the
interactIVe IP European project which includes three main sensors: radar, lidar and camera. […]

Stereo cameras

The below papers cover various methods of using stereo camera set-ups for object detection and tracking.

Robust 3-D Motion Tracking from Stereo Images: A Model-less Method by Y.K. Yu, et. al.

Abstract: Traditional vision-based 3-D motion estimation algorithms require given or calculated 3-D models while the motion is being tracked. We propose a high-speed extended Kalman filter-based approach that recovers camera position and orientation from stereo image sequences without prior knowledge as well as the procedure for the reconstruction of 3-D structures. […] The proposed method has been applied to recover
the motion from stereo image sequences taken by a robot and a hand-held stereo rig. The results are accurate compared to the ground truths. It is shown in the experiment that our algorithm is not susceptible to outlying point features with the application of a validation gate.

Vehicle Tracking and Motion Estimation Based on Stereo Vision Sequences by A. Barth
(long read)

Abstract: In this dissertation, a novel approach for estimating trajectories of road vehicles such as
cars, vans, or motorbikes, based on stereo image sequences is presented. Moving objects
are detected and reliably tracked in real-time from within a moving car. […]
The focus of this contribution is on oncoming traffic, while most existing work in the literature
addresses tracking the lead vehicle. The overall approach is generic and scalable
to a variety of traffic scenes including inner city, country road, and highway scenarios. […]
The key idea is to derive these parameters from a set of tracked 3D points on the
object’s surface, which are registered to a time-consistent object coordinate system, by
means of an extended Kalman filter. Combining the rigid 3D point cloud model with
the dynamic model of a vehicle is one main contribution of this thesis. […]
The experimental results show the proposed system is able to accurately estimate the
object pose and motion parameters in a variety of challenging situations, including night
scenes, quick turn maneuvers, and partial occlusions.

Deep Learning-based approaches

The below papers include various deep learning-based approaches to 3D object detection and tracking.

Fast and Furious: Real Time End-to-End 3D Detection, Tracking and Motion Forecasting with a Single Convolutional Net by W. Luo, et. al.

Abstract: In this paper we propose a novel deep neural network that is able to jointly reason about 3D detection, tracking and motion forecasting given data captured by a 3D sensor. By jointly reasoning about these tasks, our holistic approach is more robust to occlusion as well as sparse data at range. Our approach performs 3D convolutions across space and time over a bird’s eye view representation of the 3D world, which is very efficient in terms of both memory and computation. Our experiments on a new very large scale dataset captured in several north american cities, show that we can outperform the state-of-the-art by a
large margin. Importantly, by sharing computation we can perform all tasks in as little as 30 ms.

VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection by Y. Zhou and O. Tuzel

Abstract: Accurate detection of objects in 3D point clouds is a central problem in many applications, such as autonomous navigation, housekeeping robots, and augmented/virtual reality. To interface a highly sparse LiDAR point cloud with a region proposal network (RPN), most existing efforts have focused on hand-crafted feature representations, for example, a bird's eye view projection. In this work, we remove the need of manual feature engineering for 3D point clouds and propose VoxelNet, a generic 3D detection network that unifies feature extraction and bounding box prediction into a single stage, end-to-end trainable deep network. […] Experiments on the KITTI car detection benchmark show that VoxelNet outperforms the state-of-the-art LiDAR based 3D detection methods by a large margin. Furthermore, our network learns an effective discriminative representation of objects with various geometries, leading to encouraging results in 3D detection of pedestrians and cyclists, based on only LiDAR.

Other papers on Tracking Multiple Objects and Sensor Fusion

The below papers and resources concern tracking multiple objects, using Kalman Filters as well as other techniques! We have not included the abstracts here for brevity, but you should check those out first to see which of these you want to take a look at.

Multiple Object Tracking using Kalman Filter and Optical Flow by S. Shantaiya, et. al.Kalman Filter Based Multiple Objects Detection-Tracking Algorithm Robust to Occlusion by J-M Jeong, et. al.Tracking Multiple Moving Objects Using Unscented Kalman Filtering Techniques by X. Chen, et. al.LIDAR-based 3D Object Perception by M. Himmelsbach, et. alFast multiple objects detection and tracking fusing color camera and 3D LIDAR for intelligent vehicles by S. Hwang, et. al.3D-LIDAR Multi Object Tracking for Autonomous Driving by A.S. Rachman (long read)

这篇关于sensor fusion of Udacity的文章就介绍到这儿,希望我们推荐的文章对编程师们有所帮助!



http://www.chinasem.cn/article/259788

相关文章

VMware Fusion Pro 13 Mac版虚拟机 安装Win11系统教程

Mac分享吧 文章目录 Win11安装完成,软件打开效果一、VMware安装Windows11虚拟机1️⃣:准备镜像2️⃣:创建虚拟机3️⃣:虚拟机设置4️⃣:安装虚拟机5️⃣:解决连不上网问题 安装完成!!! Win11安装完成,软件打开效果 一、VMware安装Windows11虚拟机 首先确保自己的mac开启了网络共享。不然虚拟机连不上👀的 1️⃣:准备镜像

VMware Fusion Pro 13 for Mac虚拟机软件

Mac分享吧 文章目录 效果一、下载软件二、开始安装安装完成!!! 效果 一、下载软件 下载软件 地址:www.macfxb.cn 二、开始安装 安装完成!!!

Structure-Aware Feature Fusion for Unsupervised Domain Adaptation

Structure-Aware Feature Fusion for Unsupervised Domain Adaptation 摘要引言相关工作UDA中的分布匹配方法 Model 摘要 无监督域适应(Unsupervised Domain Adaptation,UDA)旨在从有标注的源域中学习并迁移通用特征到无任何标注的目标域。现有的方法仅对高层次的表示进行对齐,却没有利

8 自研rgbd相机基于rk3566之sensor图像解析与AWB算法原理

自研rgbd相机基于rk3566之sensor图像解析与AWB开发 1、sensor-raw图解析介绍2、sensor-raw图解析程序详解常规raw10-to-raw16数据解析sc2310-raw10-to-raw16图像解析gc2053/gc2093 raw图像解析 3、sensor-awb及常用算法程序bayer bggr to rgb 图像解析简单白平衡算法灰度世界法白平衡算法完美

VMware fusion下ubuntu(16)虚拟机编译安卓源码笔记

前两天刚刚编译成功了android源码,历尽九九八十一难。因为过程艰苦,不想就这么遗忘,所以写出来与大家共享,希望为后来者提供借鉴。 我开发的环境是mac os10.12.3,vmware fusion8.1.1,ubuntu16.04.2。 环境搭建: 关于vmware fusion的安装我就不想说什么了,因为实在太简单了。但是在安装ubuntu的时候一定要注意,创建好虚拟机以后首先要设置

android 传感器 Sensor

/SensorTest/res/layout/activity_main.xml <?xml version="1.0" encoding="utf-8"?><LinearLayout xmlns:android="http://schemas.android.com/apk/res/android"android:layout_width="fill_parent"android:lay

Mac不连网怎么办,Mac上安装的VMware Fusion不连网怎么办(续篇)

昨天下午改了下以太网的IP然后点击应用(网络肯定是不能用的),再改回原来的IP地址后神奇般的好了,但好景不长 今天一早开机后,重新连接以太网网络一切正常可用,打开VMware Fusion也可以连网 解决方案: Mac端以太网连接,配置IPv4为手动,IP地址,子网掩码,路由器,DNS服务器设置根据所在网络自行配置即可,Wi-Fi就开着不要管了 VMware Fusion端,打开网络和共享

Mac不连网怎么办,Mac上安装的VMware Fusion不连网怎么办

这两天我的Mac mini抽风似得断网了,测试网线完好木有问题,试过重启、拔网线重连、修改以太网的配置均未成功,临时解决方案使用wifi上网 1、步骤: 打开Mac的偏好设置==>选择网络==>以太网(停用服务)==>选择Wi-Fi==>选择网络名称并输入密码==>打开Wi-Fi 酱紫Mac就可以临时用了,根本解决方案不晓得,有待挖掘 接下来就是VMware Fusion的问题了 之前网

image sensor 关于像素的误解

CMOS Image Sensor一些概念误区: 误解1: 130W Pixel 像素感光单元,每个单元能感应R,G,B值; 正解: 130W个像素感光单元,每个感光单元只能感应单一颜色(RGB中某种色彩),不能同时感应R,G,B三个值; (PS: Foveon X3感光元件能同时感应RGB三个值) 误解2: 感光单元的R/G/B值只为单一像素试用 正解: 每个像素感光单元只

论文阅读--Efficient Hybrid Zoom using Camera Fusion on Mobile Phones

这是谷歌影像团队 2023 年发表在 Siggraph Asia 上的一篇文章,主要介绍的是利用多摄融合的思路进行变焦。 单反相机因为卓越的硬件性能,可以非常方便的实现光学变焦。不过目前的智能手机,受制于物理空间的限制,还不能做到像单反一样的光学变焦。目前主流的智能手机,都是采用多摄的设计,一般来说一个主摄搭配一个长焦,为了实现主摄与长焦之间的变焦,目前都是采用数字变焦的方式,数字变焦相比于光学