Patents
Literature
Hiro is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Hiro

175 results about "Time-of-flight camera" patented technology

A time-of-flight camera (ToF camera) is a range imaging camera system that employs time-of-flight techniques to resolve distance between the camera and the subject for each point of the image, by measuring the round trip time of an artificial light signal provided by a laser or an LED. Laser-based time-of-flight cameras are part of a broader class of scannerless LIDAR, in which the entire scene is captured with each laser pulse, as opposed to point-by-point with a laser beam such as in scanning LIDAR systems.

Target orientation estimation using depth sensing

A system for estimating orientation of a target based on real-time video data uses depth data included in the video to determine the estimated orientation. The system includes a time-of-flight camera capable of depth sensing within a depth window. The camera outputs hybrid image data (color and depth). Segmentation is performed to determine the location of the target within the image. Tracking is used to follow the target location from frame to frame. During a training mode, a target-specific training image set is collected with a corresponding orientation associated with each frame. During an estimation mode, a classifier compares new images with the stored training set to determine an estimated orientation. A motion estimation approach uses an accumulated rotation / translation parameter calculation based on optical flow and depth constrains. The parameters are reset to a reference value each time the image corresponds to a dominant orientation.
Owner:HONDA MOTOR CO LTD +1

Method and apparatus for fusing distance data from a distance sensing camera with an image

It is inter alia disclosed a method comprising: projecting a distance value and position associated with each of a plurality of pixels of an array of pixel sensors in an image sensor of a time of flight camera system onto a three dimensional world coordinate space as a plurality of depth sample points; and merging pixels from a two dimensional camera image with the plurality of depth sample points of the three dimensional world coordinate space to produce a fused two dimensional depth mapped image.
Owner:NOKIA TECHNOLOGLES OY

3D time-of-flight camera and method

The present invention relates to a 3D time-of-flight camera and a corresponding method for acquiring information about a scene, in particular for acquiring depth images of a scene, information about phase shifts between a reference signal and incident radiation of a scene or environmental information about the scene. To increase the frame rate, the proposed camera comprises a radiation source (12) that generates and emits electromagnetic radiation (13) for illuminating said scene (2), a radiation detector (14) that detects electromagnetic radiation (15a, 15b) reflected from said scene (2), said radiation detector (14) comprising one or more pixels (16), in particular an array of pixels, wherein said one or more pixels individually detect electromagnetic radiation reflected from said scene, wherein a pixel comprises two or more detection units (161-1, 161-2, ...) each detecting samples of a sample set of two or more samples and an evaluation unit (18) that evaluates said sample sets of said two or more detection units and generates scene-related information from said sample sets, wherein said evaluation unit comprises a rectification unit (20) that rectifies a subset of samples of said sample sets by use of a predetermined rectification operator defining a correlation between samples detected by two different detection units of a particular pixel, and an information value calculator (22) that determines an information value of said scene-related information from said subset of rectified samples and the remaining samples of the sample sets.
Owner:SONY CORP

Depth calculation imaging method based on flight time TOF camera

ActiveCN102663712AAvoid changeImplementing the super-resolution processImage enhancementImage analysisComputer visionHigh resolution
The invention belongs to the field of computer vision. In order to achieve the balance between general quantization error and overload error to ensure that the quantization output noise-signal ratio is optimum, the method adopts the technical scheme of a depth calculation imaging method based on a flight time TOF camera, and the method comprises the following steps: firstly, obtaining respective internal parameters including focal distances and optical centers and external parameters including rotation and translation of the TOF camera and a color camera after camera calibration, and obtaining a plurality of depth scatters on a high resolution diagram; secondly, building an autoregression model item of an energy function; thirdly, building an basic data item and a final solve equation of the energy function, building a data item of the energy function through an initial depth scatter diagram, and combining the data item and an autoregression item with a factor Lambada into a whole body to be served as a final solve equation through a lagrange equation; and fourthly, performing solving on an optimized equation through a linear function optimization method. The method is mainly applied to digital image processing.
Owner:TIANJIN UNIV

Methods and Apparatus for Coded Time-of-Flight Camera

In illustrative implementations, a time-of-flight camera robustly measures scene depths, despite multipath interference. The camera emits amplitude modulated light. An FPGA sends at least two electrical signals, the first being to control modulation of radiant power of a light source and the second being a reference signal to control modulation of pixel gain in a light sensor. These signals are identical, except for time delays. These signals comprise binary codes that are m-sequences or other broadband codes. The correlation waveform is not sinusoidal. During measurements, only one fundamental modulation frequency is used. One or more computer processors solve a linear system by deconvolution, in order to recover an environmental function. Sparse deconvolution is used if the scene has only a few objects at a finite depth. Another algorithm, such as Wiener deconvolution, is used is the scene has global illumination or a scattering media.
Owner:MASSACHUSETTS INST OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products