Seeing around corners with Prof. Andreas Velten
We caught up with Prof. Andreas Velten, Department of Biostatistics and Medical Informatics, Department of Electrical and Computer Engineering to share his views and thoughts on developing new imaging methods to see around corners, the industry-wide challenges with imaging systems yet to match the performance of the human eye and the importance of dialog between basic research and short term applications.
Prof Velten will be presenting “Robust Inexpensive Frequency Domain LiDAR using Hamiltonian Coding” at AutoSens in May 2019.
You worked as a Postdoctoral Associate at the MIT Media Lab, what did you work on there?
I developed ultra-fast imaging systems to capture videos the propagation of light. I captured videos of laser pulses moving through soda bottles or illuminating little still life scenes. We used the time of flight information captured in the videos to develop methods to see around corners. We illuminate a relay surface in a scene with our light pulses. After hitting the relay surface the light travels into the scene and reflects off objects. We capture video of the light that comes back to the relay wall and use it to reconstruct images of the scene as it can be seen from the relay wall. In my own group we are using this method to image scenes like office cubicles through a window, or the inside of caves from the air.
You’re the co-founder and CTO of OnLume, what does the company do?
OnLume develops cameras to better visualize tumors, nerves, and other anatomy during surgery.
What is your main research focus within the Computational Optics group?
Our group develops new imaging methods that can achieve things that normal vision systems can’t, by combining new methods of light capture with computational algorithms to create images from the captured data. A large focus is on fast imaging systems that can measure the time of flight of light through the scene. We use this information to see through fog, around corners, and create high resolution 3D images.
Watch latest videos from AutoSens on YouTube
► Deep learning processing technologies for embedded systems
► Multiclass road object detection for advanced driver assistance using deep neural networks
► Improving and implementing traditional computer vision algorithms…
What do you see as the biggest challenges for imaging for automotive?
Imaging systems can’t yet match the performance of the human eye. Especially with respect to dynamic range and efficiency. Trying to write an algorithm that performs like a human driver, but with inferior data, is challenging. Imaging methods like LiDAR can provide algorithms with better data to not only match, but exceed the capability of human vision. Finding ways of providing these technologies in a cost effective way and using them to the largest benefit to the driver (which could be a human or an algorithm) provides a fascinating challenge.
Your presentation covers Hamiltonian Coding. Can you explain what that is and how it applies to automotive?
Hamiltonian Coding is a way to improve the performance of “frequency domain” LiDAR systems. These systems illuminate the target with a modulated light source (i.e. one that is blinking on and off in a particular pattern). The pattern seen by the camera is shifted in time from the illumination and by comparing the two patterns the camera can determine the distance to the target. This is an inexpensive way to perform 3D LiDAR imaging and is used in existing devices like the Microsoft Kinect. They are fast, robust, and don’t have moving parts like other LiDAR imaging systems. Our research analyzes what the best patterns or “codes” are to send out into the scene. By choosing our optimal patterns we can improve the performance of these systems by an order of magnitude and make them more robust to changes in ambient light with minimal changes in hardware. We hope that these changes will make frequency domain methods competitive with the pulsed LiDAR systems often used in automotive applications.
What are you looking forward to about presenting at AutoSens?
Presenting at a venue focusing on such an important emerging application area for advanced vision is of great importance to me. I believe the dialog between basic research and short term applications is very important to make sure our research remains relevant. So I am hoping for lots of questions and interesting conversations.
You can hear Prof. Andreas Velten, Department of Biostatistics and Medical Informatics, Department of Electrical and Computer Engineering at AutoSens in Detroit this Spring. Tickets are available here >>
Read latest news and updates from AutoSens
- Regulatory landscape for autonomous driving in the US ~ 16 April 19
- AutoSens takes aim at the biggest challenges facing autonomous driving by encouraging collaboration ~ 16 April 19
- Simulation is the key to public confidence in AVs ~ 2 April 19