slide1 n.
Skip this Video
Download Presentation
Martin Kozák Miroslav Hlaváč

Loading in 2 Seconds...

play fullscreen
1 / 14

Martin Kozák Miroslav Hlaváč - PowerPoint PPT Presentation

  • Uploaded on

Fish position determination in 3D space by stereo vision. Martin Kozák Miroslav Hlaváč. 27 . 07. 2011. Project goals. Design low budget system to determinate 3D position of fish in water environment in real time Explore the capabilities of two cameras system

I am the owner, or an agent authorized to act on behalf of the owner, of the copyrighted work described.
Download Presentation

PowerPoint Slideshow about 'Martin Kozák Miroslav Hlaváč' - yul

Download Now An Image/Link below is provided (as is) to download presentation

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.

- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -
Presentation Transcript

Fish position determination

in 3D space by stereo vision

Martin Kozák

Miroslav Hlaváč

27. 07. 2011

project goals
Project goals
  • Design low budget system to determinate 3D position of fish in water environment in real time
  • Explore the capabilities of two cameras system
  • Explore the capabilities of the Kinect dept sensor
  • Testing of both systems in different conditions
  • Compare results from cameras and Kinect
  • Designed system will be used to track differences in fish motion
used equipment and software
Used equipment and software
  • Aquarium (60x30x30cm)–similar one is planned to be used in real application of this project
  • Two Microsoft LifeCam Studio webcams
  • Calibration object (chessboard)
  • Kinect for Xbox 360
  • Rubber testing object
  • Matlab
two cameras system
Two cameras system
  • The system of two cameras is emulating human eyes
  • We need to do calibration of cameras to determine the system parameters
  • These parameters are then used to compute 3D coordinates from two different views of scene (epipolar geometry)
epipolar geometry
Epipolar geometry
  • We can determine position of point from one image, but to determine depth we need the information from the second camera
  • Selection of one point in left image and finding corresponding point on epipolar line in right image
  • Computing 3D coordinates from those two points
cameras calibration
Cameras calibration
  • Two sets of parameters for cameras
    • Extrinsic (rotation and translation between cameras)
    • Intrinsic (focal length, skew and pixel distortion for each camera)
  • Gaming device for Xbox 360
  • Projecting IR light pattern on the scene through special grid
  • Computing depth information from the projected grid distortion
cameras results 1
Cameras results 1
  • Manual corresponding points selection
  • Selecting the white point onrubber testing object manually and computing 3D trajectory
  • 3D coordinates accuracy is ± 0.5 mm
camera results 2
Camera results 2
  • We developed online tracking system – 7fps
  • Automatic corresponding point selection
  • Image thresholding
  • Binary image opening to eliminate small distortions
  • By computing mean position of white pixels we will get corresponding points in both images
kinect accuracy
Kinect accuracy
  • Real and Kinect distance dependence on water depth
  • Depth independent
  • Kinect accuracy in x-axis in water
  • x-axis accuracy is ±3.5pixels

measured distance [cm]

real distance [cm]

object size [pixel]

shift from the center of view [cm]

kinect results
Kinect results
  • We developed online tracking system – 30 fps
  • Maximum measurable depth in clear water is 40 cm
  • Maximum measurable depth in dirty water is 20 cm
  • Depth of fish is obtained by depth thresholding
  • Minimal measurable distance 80cm
kinect vs cameras
Kinect vs. cameras



Precision (+)

Environment independence (+)

Image segmentation(-)

Localization of (-)corresponding points

Calibration for each new (-) system position

Requires more processing (-) power

  • No need for calibration (+)
  • Depth map is direct output(+)
  • No color and outer light (+) dependence
  • Maximal water depth (-) limitation
  • IR reflecting material cause (-) errors in depth map
  • Lower accuracy in water(-)
  • Minimal distance 80cm(-)
  • Both systems are usable for online 3D fish position determination in water
  • We would recommend using Kinect in environment where accuracy is not the main concern the water is shallow and clean and where we need more mobility
  • Cameras offer higher accuracy and environment independence but they require more processing power (corresponding points detection) and initial calibration

We would like to thank Ing. PetrCísař, Ph.D. for leading us through this project and for his advices.