Advances in 2D/3D Video Analysis: using context & 3D

4/8/2015
Title
Advances in 2D/3D Video Analysis:
using context & 3D depth information
Prof.dr.ir. Peter H.N. de With (TU/e VCA)
With contributions from:
(In cooperation with ViNotion)
Dennis vd Wouw, Willem Sanberg.
Egor Bondarev, Lingni Ma (TU/e VCA)
Slide - 1
EE- SPS / Video Coding & Architectures
• Faculty Electrical Engineering
• Dept. Signal Processing Systems
• Video is a rich area (computing, memory,
speed, bandwidth)
• Healthcare, AV Multimedia, PC systems,
Security, Geo-referencing, etc.
• Video Coding and Architectures
• Research group of about 25-30 researchers
and 3 postdoc staff members (approx. 2/3
PhDs)
• Covers Professional imaging
• Also implementations
• real-time architectures, fast coding
systems, special embedded applications
Slide 2
SPS-VCA Group Overview
Slide - 2
1
4/8/2015
SPS-VCA - (1) /
Video Content Analysis and Applications
Ship & context
detection, Watervisie
R’dam
Video Proc.
Fast & realtime SW
Traffic sign detection &
3D positioning,
CycloMedia Techn.
Content Analysis
Fall incidents & Medical
training for hospitals &
Kempenh. Inst.
Bank robbery
application analysis in
EU ITEA Cantata
Object detection, ViNotion
Scalable Video Coding &
Analysis, VDG Security
Content Anal. & RT prototypes,
spin-off
SPS-VCA Group Overview
Slide - 3
Slide 3
8-4-2015
Contents
1. Use of Context Information: improved ship
detection
2. Robust obstacle detection with 3D depth
3. Fast stixel proc.: depth-based object detection
4. 3D reconstruction of spaces/environments
Slide - 4
2
4/8/2015
Outline
1. Use of Context Information for
improved ship detection
Slide - 5
Ship Detection / Motivation and Goals
• Why detecting ships using video camera?
– Video-based ship detection is promising in port surveillance
Low cost
Easy to manage
Small, non-metal ships
Valuable supplement to radar system
• Challenges
Water, harbor infrastructure,
vegetation, etc.
– Dynamic and complex background
– Large variations of ship appearances in size, shape and color
• Our Goals
– Develop robust ship detection based on water region detection
– Find entire ship for various applications: size estimation, tracking,
etc.
6
Slide - 6
3
4/8/2015
Context: Water Detection(1/2)
False positives
can be reduced
• Why Context
– Ships only travel inside the water region
– Beneficial if water region is a-priori extracted and
provided as contextual information
• Context Extraction: Water Detection
Segmentation + Region-based Classification
Slide - 7
Context: Water Detection(2/2)
contextual information
for ship detection
Original surveillance
images
Results of segmentation
Classification of water
regions
8
Slide - 8
4
4/8/2015
Improved Ship Detection (1/2)
• Improvements:
• Detection of entire body of the ship, not only the cabin
• Merge the redundantly detected cabins for one ship
• Temporal information is added for improving the
detection coherence
• Refinement of the bounding box indicating the ship
9
Slide - 9
10
Slide - 10
Improved Ship Detection (2/2)
Flow Chart of Offline Demo
Ⅱ
5
4/8/2015
Demonstration of Results
• Comparison between 2 Demos
Video of Online Demo
Video of Offline Demo
11
Slide - 11
Conclusions on Ship Detection
• Conclusions
• Robust automatic ship detection for camera-based port
surveillance
• Water detection provides contextual information to facilitate ship
detection
• Temporal information improves the detection
• Enables multiple ship detection for both moving and static ships
• Future work
• Dealing with occlusion problems add robustness where needed
• Develop and exploit a tracking algorithm for combined
detection-tracking strategies to employ intelligent scenarios in
usage
12
Slide - 12
6
4/8/2015
Outline
2. Depth reconstruction
with stereo imaging
Robust collision
avoidance with 3D
Slide - 13
Next step in improved reliability in object detection
Possible extension for higher quality
• Apply local 3D reconstruction
• Overlapping cameras
• Advanced camera calibration
• Homography allows 3D reconstruction
• Enables distance calibration
• More accurate detection, higher robustness road sampling
Slide - 14
7
4/8/2015
Obstacle detection in 3D / distinguish from background
●
Alternative to change detection method
●
Two cameras to estimate depth
–
Similar to humans using two eyes to perceive depth
–
Measure displacement between L/R images → estimate depth
–
Obstacles have different depth than background
Left
Right
Displacement
Tram track
detection
Obstacle detection
in 3D
Object
recognition
Collision
prediction
Warning
Slide - 15
Facilitate object recognition with large variations
●
Automatic detection of most important traffic participants
–
Detect objects with large appearance variation in single image
–
Classes: pedestrians, cyclists, cars
–
Object class informative for risk
Pedestrian
Car
Cyclist
Tram track
detection
Obstacle detection
in 3D
Object
recognition
Collision
prediction
Warning
Slide - 16
8
4/8/2015
Stereo data acquisition / Experimental Setup – (1)
●
Mount camera on tram
●
Depth information desired (obstacle detection in 3D)
–
Two cameras
4/8/2015
Collision avoidance - Phase 1
Slide 17 of 33Slide - 17
Stereo data acquisition / Experimental Setup – (2)
●
●
Custom fabricated stereo-rig
–
Variable baseline
–
High-end 20 MPixel cameras
–
Accurate GPS positioning
–
Flexible: emulate simple system
Computer system for processing
–
Hexa core Intel i7-3960x
–
Dual high-end GPU
–
Large storage
4/8/2015
Collision avoidance - Phase 1
Slide 18 of 33Slide - 18
9
4/8/2015
3D Obstacle detection pipeline
Camera
interface
Depth
reconstruction
Track/ lane
detection
Ground surface
modelling
Obstacle
detection
Exploit depth to find obstacles
See if any collide with the trajectory
Segmentation
& semantics
Collision
prediction
Warn
driver
Slide - 19
Experimental Results – obstacle detection in 3D
●
●
Accurate localization of obstacles
Detect any object
(car, bike, tram, pedestrian, ….)
●
No learning and GPS required
●
Conclusion:
–
Obstacle detection is
straightforward with depth sensing
–
DEMO video
Slide 20 of 33Slide - 20
10
4/8/2015
Conclusions – Benefits for Object recognition
●
Enhance functionality: Can be used to classify an object
(e.g. car, pedestrian, cyclist)
●
Robustness against variations in light and object appearance
●
Conclusion:
–
Can be used to increase system functionality
–
Assign importance based on object class
(e.g. pedestrian more fragile than car, or security risk)
Slide - 21
Outline
3.
Fast stixel processing for
depth-based object detection
Slide - 22
11
4/8/2015
Stixel principle: processing concept
•
Column-wise probabilistic analysis
•
Models the world in rectangular column patches
•
•
•
Assumed geometry: linear parallellax with
superimposed fronto-parallel obstacles
Segment and label into ground or obstacle
•
Ground disparity:
•
Obstacle disparity:
≅
≅
Efficient and optimal solution
using Dynamic Programming
•
Cost function: log-likelihood
Baseline system:
D. Pfeiffer, “The Stixel World”, Ph.D. dissertation, Humboldt-Universitat Berlin, 2011
Slide - 23
Our Stixel proc. system: extension with color
Fuse Color data into core of the algorithm (assume independence)
∗
•
arg max
∈
•
•
~ ∏&'(
$)*
,
•
#$ , %$
Learn
| ,
$
~∏
3, - ,
#$ , %$
,.
12
0
)( ∏,),/
$
$
+, - ,
3, - ,
online and self-supervised
•
Starting with a uniform color-distribution for ground and obstacles,
update that model for each evaluated frame
•
Highly adaptive to new environments
Our system:
W.P. Sanberg, G. Dubbelman, P.H.N. de With,
“Extending the Stixel World with Online Self-supervised Color Modeling for Road-versus obstacle Segmentation”,
IEEE Conf. on Intelligent Transportation Systems (ITSC), Oct. 2014 [IN PRESS]
Slide - 24
12
4/8/2015
Experiments – Our Stixel proc. system
• Experiments with:
•
Color space
•
Number of histogram bins
•
Range of learning window
•
Sample selection
• Results:
•
Indexed RGB (adaptive!)
•
Single 3-seconds-old frame gives enough color information
(low computational constraints!)
•
Increase F-score ‘drivable distance’ from 0.86 to 0.97
Our system:
W.P. Sanberg, G. Dubbelman, P.H.N. de With,
“Extending the Stixel World with Online Self-supervised Color Modeling for Road-versus obstacle Segmentation”,
IEEE Conf. on Intelligent Transportation Systems (ITSC), Oct. 2014 [IN PRESS]
Slide - 25
Future outlook – Our Stixel system
Slide - 26
13
4/8/2015
Future outlook – Our Stixel system
Slide - 27
Outline
4. Multi-modal 3D
reconstruction
processing
Algorithms for full-3D
Segmentation of objects
Slide - 28
14
4/8/2015
Sensing Platforms
Set of 2D/3D Sensing Platforms
• Few laser-scanner platforms (LIDARs)
• RGB-D mapping system (Kinect)
• Stereo Cameras, Panoramic 360 camera
• Autonomous robot
Slide - 29
Multi-modal 3D Reconstruction processing – (1)
Full 3D Reconstruction
Model of environment
Slide - 30
15
4/8/2015
2D/3D Data Fusion Pipeline
Pointclouds
from laser-scan
Point- cloud filtering &
Mesh generation
Generated 3D meshed model
Mapping the textures
on meshes
Enhancement &
Alignment of meshes
3D model with mapped texture
Total volumetric 3D model
Image samples from 360-camera
Slide - 31
Experim. Results: Detailed Reconstruction & Texturing
Slide - 32
16
4/8/2015
Technologies / Significant processing…
Simultaneous Localization and Mapping (SLAM):
For both mono-camera and RGB-D sensors
Real-time
Loop closure and bundle adjustment
Point-cloud filtering and registration
3D Segmentation, decimation and semantic classification
Kinfu Large Scale (Kinect pipeline)
Multi-view texturing of point clouds and meshes
Data fusion:
LIDAR data with stereo/kinect data
Point-cloud data with multi-view textures
Video data into 3D model
PAGE 33
Slide - 33
Outline contributions to 3D processing
Real-time photo-realistic RGB-D SLAM
Efficient mesh reconstruction
16m x 8m office reconstructed in real time with RGB-D camera
Slide - 34
17
4/8/2015
RGB-D SLAM / introduction
Challenges
noval sensing interface, first system in 2011
tracking and mapping at 30fps
consistency vs. noisy measurement
data representation, 9 million points per second
Slide - 35
RGB-D SLAM / Overview of proposed algorithms
Schematic overview
Slide - 36
18
4/8/2015
Mesh Reconstruction / introduction
Triangulation
from discrete points to continuous surface
Applications
3D modeling
rendering / visualization
3D printing
robot reasoning
Slide - 37
Mesh Reconstruction / introduction
Challenges
large-scale point cloud
preserve geometry
redundancy (planar regions)
5.6 million points
11.2 million triangles
Slide - 38
19
4/8/2015
Mesh Reconstruction / Algorithm overview
Hybrid triangulation
Required
- Plane detection in 3D space
- Non-plane segments
Slide - 39
Mesh Reconstruction / Fast triangulation algorithm
Quadtree-based planar triangulation
in
input plane
extract boundary
construct quadtree
Slide - 40
20
4/8/2015
Mesh Reconstruction / result
90% planar points are removed
90% planar triangles are reduced
Slide - 41
Mesh Reconstruction / Demonstration result
DEMO MESHING
DEMO RENDERING
Slide - 42
21
4/8/2015
Conclusion on near future of video content analysis
• Exploiting context information for improved ship detect.
• Exploit sensor sensitivity and full resolution
• Better camera calibration, exploit further sensing modalities
• Extended obstacle detection with local 3D information
• Enhanced knowledge about object, broader viewing angle
• Better situational awareness, better signaling to operator
• Fast stixel processing object detection in depth signals
• 3D space sensing with e.g. Lidar & advanced Im. Proc.
• 3D modeling and reconstruction
• 3D object detection and classification
Slide - 43
7.2 Conclusions – (2)
Thanks for your attention
• Successful participation in October 2012 NATO Trial
• Performing in the top of the participants
• Only one showing a real-time system with live reporting
Slide - 44
22