N22A-T020 TITLE: Lidar-like 3D Imaging System for Accurate Scene Understanding
OUSD (R&E) MODERNIZATION PRIORITY: Artificial Intelligence (AI)/Machine Learning (ML);Autonomy
TECHNOLOGY AREA(S): Information Systems;Sensors
OBJECTIVE: Develop inexpensive Lidar-like 3D imaging sensors that have high depth and lateral resolution, have a large field-of view for reliable object detection, respond in real time, and work at medium to long ranges in indoor and outdoor environments.
DESCRIPTION: 3D scene understanding (i.e., 3D scene segmentation and parsing, depth estimation, object detection and recognition) are essential components in a vision system. 3D sensors similar to Microsoft Kinect are inexpensive and high resolution but have limited range outdoors, thus not suited for many robotics applications. Lidars have long range and high depth accuracy, but are very expensive; for example, those used in self-driving cars are typically several times more expensive than other car components. Another drawback of current Lidars is their small "vertical" field-of-view, which results in limited vertical resolution and accuracy in object detection because Lidars (even the more expensive ones) have at most 64 scan lines, which could fail to detect small objects even at medium range distances.
The goal of this STTR topic is to develop inexpensive, high-resolution, high-accuracy 3D imaging sensors for wide use on a variety of large and small ground and aerial robotic platforms that can work in dynamic environments under different conditions. ONR expects recent promising advances along a number of directions including machine learning-based algorithms for improved depth estimation with stereo cameras [Refs 2, 5], active illumination technologies [Ref 1], and optimal time-of-flight coding [Ref 3], etc., open new approaches to building hybrid systems that combine optical cameras and laser ranging for developing such 3D imaging sensors. Combining these advances (ML-based stereo imaging, utilizing active illumination for 3D imaging, and novel time-of-flight coding for improved range estimation) requires innovative approaches.
PHASE I: Design the system architecture including sensors and computing hardware, and processing and inference algorithms for building inexpensive, high-resolution, accurate, 3D imaging sensors. Since these sensors are intended for use on various UGVs and UAVs deployed in dynamic and cluttered environments, the design should consider tradeoff estimates among size, weight, and power (SWAP), as well as resolution, detection accuracy, operating range, frame rate, and cost. Develop a breadboard version to demonstrate the feasibility of the design. Develop a Phase II plan.
PHASE II: Perform experiments in a variety of situations and refine the system. Goals for Phase II are: (a) the system should have a field-of-view and resolution comparable to optical cameras; (b) Demonstrate the systemís capability for human detection. Normal vision can detect humans up to a distance of about 300m in daylight. At nighttime, typical headlights illuminate the road up to a distance of about 60m [Ref 4]. The minimum detection range should be the aforementioned distances in daylight and nighttime. (c) Develop a compact prototype imaging system that is small, lightweight, and low power, suitable for portability by personnel and small autonomous platforms (UxVs).
PHASE III DUAL USE APPLICATIONS: Perform additional experiments in a variety of situations and further refine the system for transition and commercialization. Ensure that the real-time imaging system is operable in real-world dynamic environments, thus extending the imaging to handle real-time acquisition, that is, at least 30 fps. This technology could be used in the commercial sector for self-driving cars, and in surveillance and navigation on any land or air vehicle.
KEYWORDS: Lidar-like 3D imaging sensor; hybrid imaging; high-resolution sensor with large field of vision; FOV; outdoor imaging; indoor imaging
** TOPIC NOTICE **
The Navy Topic above is an "unofficial" copy from the overall DoD 22.A STTR BAA. Please see the official DoD Topic website at rt.cto.mil/rtl-small-business-resources/sbir-sttr/ for any updates.
The DoD issued its 22.A STTR BAA pre-release on December 1, 2021, which opens to receive proposals on January 12, 2022, and closes February 10, 2022 (12:00pm est).
Direct Contact with Topic Authors: During the pre-release period (December 1, 2021 thru January 11, 2022) proposing firms have an opportunity to directly contact the Technical Point of Contact (TPOC) to ask technical questions about the specific BAA topic. Once DoD begins accepting proposals on January 12, 2022 no further direct contact between proposers and topic authors is allowed unless the Topic Author is responding to a question submitted during the Pre-release period.
SITIS Q&A System: After the pre-release period, proposers may submit written questions through SITIS (SBIR/STTR Interactive Topic Information System) at www.dodsbirsttr.mil/topics-app/, login and follow instructions. In SITIS, the questioner and respondent remain anonymous but all questions and answers are posted for general viewing.
Topics Search Engine: Visit the DoD Topic Search Tool at www.dodsbirsttr.mil/topics-app/ to find topics by keyword across all DoD Components participating in this BAA.