Broadcast News
26/07/2017
Measuring Distance With A Single Photo
Most cameras just record colour but now the 3D shapes of objects, captured through only a single lens, can be accurately estimated using new software developed by UCL computer scientists.
The method, published at CVPR 2017, gives state-of-the-art results and works with existing photos, allowing any camera to map the depth for every pixel it captures.
The technology has a wide variety of applications, from augmented reality in computer games and apps, to robot interaction, and self-driving cars. Historical images and videos can also be analysed by the software, which is useful for reconstruction of incidents or to automatically convert 2D films into immersive 3D.
Dr Gabriel Brostow, UCL Computer Science supervising researcher, said: "Inferring object-range from a simple image by using real-time software has a whole host of potential uses. Depth mapping is critical for self-driving cars to avoid collisions, for example. Currently, car manufacturers use a combination of laser-scanners and/or radar sensors, which have limitations.
"They all use cameras too, but the individual cameras couldn't provide meaningful depth information. So far, we've optimised the software for images of residential areas, and it gives unparalleled depth mapping, even when objects are on the move."
The new software was developed using machine learning methods and has been trained and tested in outdoor and urban environments.
It successfully estimates depths for thin structures such as street signs and poles, as well as people and cars, and quickly predicts a dense depth map for each 512 x 256 pixel image, running at over 25 frames per second.
Currently, depth mapping systems rely on bulky binocular stereo rigs or a single camera paired with a laser or light-pattern projector that don't work well outdoors because objects move too fast and sunlight dwarfs the projected patterns.
There are other machine-learning based systems also seeking to get depth from single photographs, but those are trained in different ways, with some needing elusive high-quality depth information. The new technology doesn't need real-life depth datasets, and outperforms all the other systems. Once trained, it runs in the field by processing one normal single-lens photo after another.
First author and UCL PhD student, Clément Godard (UCL Computer Science), added: "Understanding the shape of a scene from a single image is a fundamental problem. We aren't the only ones working on it, but we have got the highest quality outdoors results, and are looking to develop it further to work with 360 degree cameras. A 360 degree depth map would be fantastically useful – it could drive wearable tech to assist disabled people with navigation, or to map real-life locations for virtual reality gaming, for example."
Co-author, Dr Oisin Mac Aodha, previously at UCL and now at Caltech, said: "At the moment, our software requires a desktop computer to process individual images, but we plan on miniaturising it, so it can be run on hand-held devices such as phones and tablets, making it more accessible to app developers. We've also only optimised it for outdoor use, so our next step is to train it on indoor environments."
The team has patented the technology for commercial use through UCL Business, but has made the code free for academic use. Funding for the research was kindly provided by the Engineering and Physical Sciences Research Council.
Go to the projects homepage, here.
The method, published at CVPR 2017, gives state-of-the-art results and works with existing photos, allowing any camera to map the depth for every pixel it captures.
The technology has a wide variety of applications, from augmented reality in computer games and apps, to robot interaction, and self-driving cars. Historical images and videos can also be analysed by the software, which is useful for reconstruction of incidents or to automatically convert 2D films into immersive 3D.
Dr Gabriel Brostow, UCL Computer Science supervising researcher, said: "Inferring object-range from a simple image by using real-time software has a whole host of potential uses. Depth mapping is critical for self-driving cars to avoid collisions, for example. Currently, car manufacturers use a combination of laser-scanners and/or radar sensors, which have limitations.
"They all use cameras too, but the individual cameras couldn't provide meaningful depth information. So far, we've optimised the software for images of residential areas, and it gives unparalleled depth mapping, even when objects are on the move."
The new software was developed using machine learning methods and has been trained and tested in outdoor and urban environments.
It successfully estimates depths for thin structures such as street signs and poles, as well as people and cars, and quickly predicts a dense depth map for each 512 x 256 pixel image, running at over 25 frames per second.
Currently, depth mapping systems rely on bulky binocular stereo rigs or a single camera paired with a laser or light-pattern projector that don't work well outdoors because objects move too fast and sunlight dwarfs the projected patterns.
There are other machine-learning based systems also seeking to get depth from single photographs, but those are trained in different ways, with some needing elusive high-quality depth information. The new technology doesn't need real-life depth datasets, and outperforms all the other systems. Once trained, it runs in the field by processing one normal single-lens photo after another.
First author and UCL PhD student, Clément Godard (UCL Computer Science), added: "Understanding the shape of a scene from a single image is a fundamental problem. We aren't the only ones working on it, but we have got the highest quality outdoors results, and are looking to develop it further to work with 360 degree cameras. A 360 degree depth map would be fantastically useful – it could drive wearable tech to assist disabled people with navigation, or to map real-life locations for virtual reality gaming, for example."
Co-author, Dr Oisin Mac Aodha, previously at UCL and now at Caltech, said: "At the moment, our software requires a desktop computer to process individual images, but we plan on miniaturising it, so it can be run on hand-held devices such as phones and tablets, making it more accessible to app developers. We've also only optimised it for outdoor use, so our next step is to train it on indoor environments."
The team has patented the technology for commercial use through UCL Business, but has made the code free for academic use. Funding for the research was kindly provided by the Engineering and Physical Sciences Research Council.
Go to the projects homepage, here.
Top Related Stories
Click here for the latest broadcast news stories.
10/07/2024
Vizrt Launches PTZ3 PLUS And PTZ3 UHD PLUS Cameras
Vizrt is bringing the BroadcastAV market sharply into focus with the launch of the PTZ3 PLUS and PTZ3 UHD PLUS cameras. Bridging the gap between ProAV
Vizrt Launches PTZ3 PLUS And PTZ3 UHD PLUS Cameras
Vizrt is bringing the BroadcastAV market sharply into focus with the launch of the PTZ3 PLUS and PTZ3 UHD PLUS cameras. Bridging the gap between ProAV
17/10/2024
V1 Productions Deploys Marshall Cameras
When Chopper Cops, the all-new docuseries from V1 Productions and SEE IT NOW Studios, took flight on CBS/Paramount+, producers relied on a selection o
V1 Productions Deploys Marshall Cameras
When Chopper Cops, the all-new docuseries from V1 Productions and SEE IT NOW Studios, took flight on CBS/Paramount+, producers relied on a selection o
26/09/2024
Marshall To Display Cutting-Edge Cameras At NAB New York
Marshall Electronics is to highlight a range of its cutting-edge cameras at NAB New York 2024 (Booth 1352). Marshall offers a wide selection of PTZ an
Marshall To Display Cutting-Edge Cameras At NAB New York
Marshall Electronics is to highlight a range of its cutting-edge cameras at NAB New York 2024 (Booth 1352). Marshall offers a wide selection of PTZ an
13/06/2024
Ikegami Demonstrates Broadcast Production Cameras
Following a well attended April 2024 NAB Show in Las Vegas, Ikegami demonstrated examples from its wide range of broadcast production cameras, control
Ikegami Demonstrates Broadcast Production Cameras
Following a well attended April 2024 NAB Show in Las Vegas, Ikegami demonstrated examples from its wide range of broadcast production cameras, control
10/06/2024
Canon Europe To Offer Firmware For 4K Remote Cameras
Canon Europe is to offer firmware for 4K remote cameras and remote camera controllers intended to improve the functionality and performance of functio
Canon Europe To Offer Firmware For 4K Remote Cameras
Canon Europe is to offer firmware for 4K remote cameras and remote camera controllers intended to improve the functionality and performance of functio
16/05/2024
Ikegami To Unveil Range Of Production Cameras At Broadcast Asia 2024
Ikegami is to introduce recent additions to its wide range of broadcast production cameras, monitors and related equipment on stand 6D3-1 (Hall 6) at
Ikegami To Unveil Range Of Production Cameras At Broadcast Asia 2024
Ikegami is to introduce recent additions to its wide range of broadcast production cameras, monitors and related equipment on stand 6D3-1 (Hall 6) at
10/05/2024
S&L Integrated Systems Relies On Marshall Cameras
S&L Integrated Systems, a leader in audio-visual technology solutions, has recently transformed courtroom experiences throughout Georgia and the surro
S&L Integrated Systems Relies On Marshall Cameras
S&L Integrated Systems, a leader in audio-visual technology solutions, has recently transformed courtroom experiences throughout Georgia and the surro
11/11/2024
SBS Medianet Introduces Ikegami UHK-X700 4K Cameras
Korean cable television network SBS Medianet has expanded its production base with the purchase of Ikegami UHK-X700 cameras for live coverage of a maj
SBS Medianet Introduces Ikegami UHK-X700 4K Cameras
Korean cable television network SBS Medianet has expanded its production base with the purchase of Ikegami UHK-X700 cameras for live coverage of a maj
28/02/2024
FOR-A Implements Complete System For MAP
FOR-A has implemented a complete system for Media and Art Production (MAP). The new facility is now online in Dubai Studio City. MAP is a leading prod
FOR-A Implements Complete System For MAP
FOR-A has implemented a complete system for Media and Art Production (MAP). The new facility is now online in Dubai Studio City. MAP is a leading prod
07/04/2022
MAP Installs Lawo AoIP Infrastructure
Morocco's Radio-TV Broadcaster MAP has installed Lawo AoIP Infrastructure. According to a Moroccan proverb, the earth is a peacock, and Morocco is its
MAP Installs Lawo AoIP Infrastructure
Morocco's Radio-TV Broadcaster MAP has installed Lawo AoIP Infrastructure. According to a Moroccan proverb, the earth is a peacock, and Morocco is its
08/02/2022
FOR-A Delivers Complete Flyaway Production Kit To MAP
FOR-A has delivered a complete flyaway production kit to Media and Art Production (MAP), based in Fujairah, United Arab Emirates. The kit was designed
FOR-A Delivers Complete Flyaway Production Kit To MAP
FOR-A has delivered a complete flyaway production kit to Media and Art Production (MAP), based in Fujairah, United Arab Emirates. The kit was designed
01/11/2024
Hitachi Kokusai Appoints East Coast Sales Manager
Hitachi Kokusai has added proven industry talent to its North American sales team with the appointment of Bo Reames as East Coast Sales Manager. Repor
Hitachi Kokusai Appoints East Coast Sales Manager
Hitachi Kokusai has added proven industry talent to its North American sales team with the appointment of Bo Reames as East Coast Sales Manager. Repor
16/06/2004
Allegorithmic announces MaP|TIME Video Filter Designer for Adobe After Effects and Discreet Combustion
Allegorithmic, a young award-winning company specialising in complex media modelling technology, has announced MaP|TIME, a brand new plug-in for Adobe
Allegorithmic announces MaP|TIME Video Filter Designer for Adobe After Effects and Discreet Combustion
Allegorithmic, a young award-winning company specialising in complex media modelling technology, has announced MaP|TIME, a brand new plug-in for Adobe
27/10/2014
zLense Launches Depth-Mapping Camera Solution
zLense has launched the world's first depth-mapping camera solution that captures 3D data and scenery in real-time and adds a 3D layer, which is optim
zLense Launches Depth-Mapping Camera Solution
zLense has launched the world's first depth-mapping camera solution that captures 3D data and scenery in real-time and adds a 3D layer, which is optim
26/09/2014
SMPTE Conference To Feature In-Depth Sessions
The Society of Motion Picture and Television Engineers (SMPTE), has announced that the SMPTE 2014 Annual Technical Conference and Exhibition (SMPTE 20
SMPTE Conference To Feature In-Depth Sessions
The Society of Motion Picture and Television Engineers (SMPTE), has announced that the SMPTE 2014 Annual Technical Conference and Exhibition (SMPTE 20