I Was Selling Something on Craigslist and I Found a Buyer

We decided that we could both meet at Atlantic Terminal Mall, this Saturday at 2pm. You’re supposed to pick a public place — and give some sort of physical description of yourself so they can easily…

Smartphone

独家优惠奖金 100% 高达 1 BTC + 180 免费旋转




The Camera IS The Lidar

With the tradeoffs between both sensing modalities in mind, we set out to bring the best aspects of lidars and cameras together in a single device from the very beginning. Today we’re releasing a firmware upgrade and update to our open source driver that deliver on that goal. Our OS-1 lidar now outputs fixed resolution depth images, signal images, and ambient images in real time, all without a camera. The data layers are perfectly spatially correlated, with zero temporal mismatch or shutter effects, and have 16 bits per pixel and linear photo response. Check it out:

The OS-1’s optical system has a larger aperture than most DSLRs, and the photon counting ASIC we developed has extreme low light sensitivity so we’re able to collect ambient imagery even in low light conditions. The OS-1 captures both the signal and ambient data in the near infrared, so the data closely resembles visible light images of the same scenes, which gives the data a natural look and a high chance that algorithms developed for cameras translate well to the data. In the future we’ll work to remove fixed pattern noise from these ambient images, but we wanted to let customers get their hands on the data in the meantime!

Customers who have been given early access to the update have been blown away, and we encourage anyone curious about the OS-1 to watch our unedited videos online, or download our raw data and play it back yourself with the visualizer.

We’ve seen multiple lidar companies market a lidar/camera fusion solution by co-mounting a separate camera with a lidar, performing a shoddy extrinsic calibration, and putting out a press release for what ultimately is a useless product. We didn’t do that, and to prove it we want to share some examples of how powerful the OS-1 sensor data can be, which brings us back to deep learning.

Because the sensor outputs fixed resolution image frames with depth, signal, and ambient data at each pixel, we’re able to feed these images directly into deep learning algorithms that were originally developed for cameras. We encode the depth, intensity, and ambient information in a vector much like a network for color images would encode the red, green, and blue channels at the input layer. The networks we’ve trained have generalized extremely well to the new lidar data types.

As one example, we trained a per pixel semantic classifier to identify driveable road, vehicles, pedestrians, and bicyclists on a series of depth and intensity frames from around San Francisco. We’re able to run the resultant network on an NVIDIA GTX 1060 in real time and achieved encouraging results, especially considering this is the first implementation we’ve attempted. Take a look:

Because all data is provided per pixel, we’re able to seamlessly translate the 2D masks into the 3D frame for additional real time processing like bounding box estimation and tracking.

On careful inspection, it’s clear that the network is picking up different key points in each image. Anyone who has worked on lidar and visual odometry will grasp the value of the redundancy embodied in this result. Lidar odometry struggles in geometrically uniform environments like tunnels and highways, while visual odometry struggles with textureless and poorly lit environments. The OS-1’s camera/lidar fusion provides a multi-modal solution to this long standing problem.

It’s results like these that make us confident that well fused lidar and camera data are much more than the sum of their parts, and we expect further convergence between lidars and cameras in the future.

Look out for more exciting developments from Ouster in the future!

Sunny drive with ambient data:

Add a comment

Related posts:

Container Super Meetup 2018

Every year in the run up to Container Camp UK we organise a community collider meetup which brings together all of the local user groups focussing on containers and related technologies. This year we…

Benefits of MBBS in Kazakhstan.

One of the major problems that students in India face is of choosing a college for a prestigious degree like Engineering or medicals. The entrance exam for these courses — NEET and JEE are high…

Putting on the squeeze

The purpose of the impeachment inquiry and public hearings is to render the president impotent. To exert enormous pressure on him and this keep him in check, constitutionally.