BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//C2SMART Home - ECPv6.15.20//NONSGML v1.0//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
X-ORIGINAL-URL:https://c2smart.engineering.nyu.edu
X-WR-CALDESC:Events for C2SMART Home
REFRESH-INTERVAL;VALUE=DURATION:PT1H
X-Robots-Tag:noindex
X-PUBLISHED-TTL:PT1H
BEGIN:VTIMEZONE
TZID:America/New_York
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20210314T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20211107T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20220313T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20221106T060000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20230312T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20231105T060000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20220519T120000
DTEND;TZID=America/New_York:20220519T130000
DTSTAMP:20260506T010645
CREATED:20220502T194244Z
LAST-MODIFIED:20220503T152103Z
UID:76546-1652961600-1652965200@c2smart.engineering.nyu.edu
SUMMARY:Webinar: Collaborative and Adversarial 3D Perception for Autonomous Driving
DESCRIPTION:Robust and reliable perception systems serve as the “eyes” of autonomous vehicles. LiDAR is a widely applied perception sensor in autonomous vehicles for capturing 3D geometry information of the environment. However\, LiDAR-based perception faces many challenges such as data sparsity\, occlusions\, and motion distortion. In this talk\, I will show how we design novel 3D deep learning algorithms from two aspects\, collaborative and adversarial\, in order to improve the robustness of LiDAR-based 3D perception. For effective and efficient collaborative perception\, we propose DiscoNet. It uses a dynamic directed graph with matrix-valued edge weight for an ego-vehicle to adaptively retrieve the most important complementary information from its neighboring vehicles\, which could improve its own perception performance and robustness. Besides collaborative perception\, we also study the adversarial robustness of LiDAR-based perception\, and reveal an often-overlooked vulnerability that lies in the LiDAR motion correction process. We show that spoofing of a vehicle’s trajectory estimation with small adversarial perturbations can jeopardize LiDAR perception. We hope our collaborative and adversarial 3D perception research can help improve the robustness and safety of autonomous driving systems. \n\n\n\n\n\n\n\n \nSpeaker Biography: Yiming Li is a Ph.D. candidate in AI4CE Lab at New York University (NYU) with the Dean’s PhD Fellowship. His research interest primarily lies in robot vision and learning\, with its applications in cyber-physical systems\, autonomous driving\, and human-robot interaction. More specifically\, he is interested in collaborative and adversarial perception\, egocentric vision\, multi-modal perception\, and embodied AI. His works have been published in top-tier conferences including NeurIPS\, CVPR\, ICCV\, ICRA\, and IROS. During his first Ph.D. year\, he visited MARS Lab in Institute for Interdisciplinary Information Sciences (IIIS) at Tsinghua University\, MediaBrain Group in School of Electronic Information and Electrical Engineering at Shanghai Jiao Tong University (SJTU)\, and Institute for AI Industry Research (AIR) at Tsinghua University. He obtained a bachelor degree in mechatronics\, manufacture\, and automation from Tongji University at Shanghai with honors.
URL:https://c2smart.engineering.nyu.edu/event/webinar-collaborative-and-adversarial-3d-perception-for-autonomous-driving/
ORGANIZER;CN="C2SMART":MAILTO:c2smart@nyu.edu
LOCATION:
END:VEVENT
END:VCALENDAR