• Session No.36 Research on the Recognition Technology Required for Automated Driving (OS)
  • May 22Room G30313:40-15:45
  • Chair: Takayoshi Yamashita (Chubu University)
Contents
In this session, we will discuss about research results and challenges related to recognition technology required for autonomous driving systems, which have been developed in recent years.
Committee
Autonomous Driving Technology Committee
Organizer
Naoki Suganuma (Kanazawa University), Takayoshi Yamashita (Chubu University), Junichi Meguro (Meijo University)
For presentations that will not be available video streaming after congress, a “✕” is displayed in the “Video” column, so please check.
No. Video Title・Author (Affiliation)
1

Enhancing Accuracy of Multi-Camera BEV Perception using Stereo Disparity

Shuntaro Tsuchiya・Yui Tanaka・Takeru Ninomiya・Hideaki Kido (Hitachi)・Kota Irie・Yoshitaka Okuyama (Hitachi Astemo)

Development of BEV (Bird Eye View) model that integrates multi-camera images in a bird's-eye view space is in progress. However, image-based recognition is still challenging due to low ranging accuracy. This report shows that using the relative distance obtained by stereo matching techniques improves the accuracy of 3D recognition by BEV model.

2

Robust 3D Object Detection in Rain with LiDAR

Keigo Hariya・Keisuke Yoneda・Yukiya Fukuda・Naoki Suganuma (Kanazawa University)

Detecting surrounding traffic participants is essential for safe autonomous driving. In 3D object detection with LiDAR, false-positive objects caused by rain splatting noise lead autonomous vehicles to unsafe path planning. This research proposes robust 3D object detection method against false-positive objects to rain splatting noise.

3

Implementing Localization Using Deep Learning with LiDAR Point Clouds

Kengo Kawahara・Keisuke Yoneda・Ryo Yanase・Amane Kinoshita・Naoki Suganuma (Kanazawa University)

To achieve safe autonomous driving, self-localization is essential. This study proposes a matching method that retains the features of LiDAR point clouds in a Pillar structure and converts them into pseudo-images, using deep learning to estimate the vehicle's position as a likelihood distribution. The proposed method aims to achieve more robust localization compared to conventional point cloud matching methods.

4

Simulation of Infrastructure LiDAR Using CARLA and Pedestrian Detection

Riku Nikaido・Keigo Hariya・Keisuke Yoneda・Naoki Suganuma (Kanazawa University)

LiDAR is widely used in autonomous driving perception and as an infrastructure sensor in urban environments. In this paper, we simulate a stationary LiDAR system specialized for pedestrian detection using CARLA. As a detection method, we leverage multi-frame 3D point clouds for object detection. Furthermore, the study aims a robust detection model capable of accurately identifying pedestrians in sparse point cloud data.

5

Development of ultrasonic transmission simulation to predict the effect of sonar false detection

Motoyasu Ukai・Takaaki Nakamura (Aisin)

High-precision sensing technology is essential to realize autonomous driving and autonomous parking. AEB (Automatic Emergency Braking System) detects obstacles using ultrasonic sonar, but false detections occur due to the influence of surrounding vehicle parts. In this study, we report on our efforts to develop CAE technology that can predict sensor signals when installed in a vehicle and judge whether they are good or bad.

Back to Top