41 |
Method and apparatus for encoding object information of a video object
plane |
US984030 |
1997-12-02 |
US6125142A |
2000-09-26 |
Seok-Won Han |
A method for encoding a video signal including a video object plane(VOP) and object information thereof, the VOP having an object therein, divides the VOP into a plurality of blocks and detecting boundary blocks and object blocks, wherein each boundary block includes therein background pixels and object pixels and each object block includes therein object pixels only, the background and the object pixels representing pixels residing outside and inside the object, respectively. Then the object information is converted into a set of object symbols and each object symbol is added to a background pixel included in a boundary block to thereby produce processing blocks, the processing blocks including the object blocks and one or more processed boundary blocks having therein the object symbols and the remaining unprocessed boundary blocks. The processing blocks are encoded to thereby generate an encoded video signal. |
42 |
Motion estimation and compensation of video object planes for interlaced digital video |
US10028007 |
2001-12-20 |
USRE38564E1 |
2004-08-10 |
Robert O. Eifrig; Xuemin Chen; Ajay Luthra |
A motion estimation and compensation technique is provided for interlaced digital video such as video object planes (VOPs). Predictor motion vectors for use in differentially encoding a current field coded macroblock are obtained using the median of motion vectors of surrounding blocks or macroblocks. When a surrounding macroblock is itself interlaced coded, an average motion vector for that macroblock is used, with fractional pixel values being mapped to the half-pixel. When the current block is not interlaced coded but a surrounding block is, the field motion vectors may be used individually or averaged. In a repetitive padding technique for an interlaced coded VOP, the even and odd lines of the VOP and surrounding block are grouped. Within each field, exterior pixels are padded by setting them to the value of the nearest boundary pixel, or to an average of two boundary pixels. The lines are then reordered to provide a single padded reference VOP image. |
43 |
Prediction and coding of bi-directionally predicted video object planes
for interlaced digital video |
US944118 |
1997-10-06 |
US5991447A |
1999-11-23 |
Robert O. Eifrig; Xuemin Chen; Ajay Luthra |
A system for coding of digital video images such as bi-directionally predicted video object planes (B-VOPs), in particular, where the B-VOP and/or a reference image used to code the B-VOP is interlaced coded. For a B-VOP macroblock which is co-sited with a field predicted macroblock of a future anchor picture, direct mode prediction is made by calculating four field motion vectors, then generating the prediction macroblock. The four field motion vectors and their reference fields are determined from (1) an offset term of the current macroblock's coding vector, (2) the two future anchor picture field motion vectors, (3) the reference field used by the two field motion vectors of the co-sited future anchor macroblock, and (4) the temporal spacing, in field periods, between the current B-VOP fields and the anchor fields. Additionally, a coding mode decision process for the current MB selects a forward, backward, or average field coding mode according to a minimum sum of absolute differences (SAD) error which is obtained over the top and bottom fields of the current MB. |
44 |
Bidirectionally predicted pictures or video object planes for efficient and flexible video coding |
US12030928 |
2008-02-14 |
US07545863B1 |
2009-06-09 |
Barin Geoffry Haskell; Atul Puri; Robert Lewis Schmidt |
A method is provided for decoding a bit stream representing an image that has been encoded The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficients and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes. |
45 |
Bidirectionally predicted pictures or video object planes for efficient and flexible video coding |
US11212262 |
2005-08-27 |
US07463685B1 |
2008-12-09 |
Barin Geoffry Haskell; Atul Puri; Robert Lewis Schmidt |
A method is provided for decoding a bit stream representing an image that has been encoded The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficents and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes. |
46 |
Bidirectionally predicted pictures or video object planes for efficient and flexible video coding |
US10728658 |
2003-12-06 |
US07356081B1 |
2008-04-08 |
Barin Geoffry Haskell; Atul Puri; Robert Louis Schmidt |
A method is provided for decoding a bit stream representing an image that has been encoded. The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficients and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes. |
47 |
Bidirectionally predicted pictures or video object planes for efficient and flexible video coding |
US09988786 |
2001-11-20 |
US20020075956A1 |
2002-06-20 |
Barin
Geoffry
Haskell; Atul
Puri; Robert
Louis
Schmidt |
A method is provided for decoding a bit stream representing an image that has been encoded The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficents and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes. |
48 |
Motion estimation and compensation of video object planes for interlaced
digital video |
US301141 |
1999-04-28 |
US6026195A |
2000-02-15 |
Robert O. Eifrig; Xuemin Chen; Ajay Luthra |
A motion estimation and compensation technique is provided for interlaced digital video such as video object planes (VOPs). Predictor motion vectors for use in differentially encoding a current field coded macroblock are obtained using the median of motion vectors of surrounding blocks or macroblocks. When a surrounding macroblock is itself interlaced coded, an average motion vector for that macroblock is used, with fractional pixel values being mapped to the half-pixel. When the current block is not interlaced coded but a surrounding block is, the field motion vectors may be used individually or averaged. In a repetitive padding technique for an interlaced coded VOP, the even and odd lines of the VOP and surrounding block are grouped. Within each field, exterior pixels are padded by setting them to the value of the nearest boundary pixel, or to an average of two boundary pixels. The lines are then reordered to provide a single padded reference VOP image. |
49 |
Bidirectional predicted pictures or video object planes for efficient and flexible video coding |
US11924657 |
2007-10-26 |
US07656953B1 |
2010-02-02 |
Barin Geoffrey Haskell; Atul Puri; Robert Louis Schmidt |
A method is provided for decoding a bit stream representing an image that has been encoded The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficients and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes. |
50 |
Prediction and coding of bi-directionally predicted video object planes for interlaced digital video |
US09995997 |
2001-11-19 |
USRE38563E1 |
2004-08-10 |
Robert O. Eifrig; Xuemin Chen; Ajay Luthra |
A system for coding of digital video images such as bi-directionally predicted video object planes (B-VOPs), in particular, where the B-VOP and/or a reference image used to code the B-VOP is interlaced coded. For a B-VOP macroblock which is co-sited with a field predicted macroblock of a future anchor picture, direct mode prediction is made by calculating four field motion vectors, then generating the prediction macroblock. The four field motion vectors and their reference fields are determined from (1) an offset term of the current macroblock's coding vector, (2) the two future anchor picture field motion vectors, (3) the reference field used by the two field motion vectors of the co-sited future anchor macroblock, and (4) the temporal spacing, in field periods, between the current B-VOP fields and the anchor fields. Additionally, a coding mode decision process for the current MB selects a forward, backward, or average field coding mode according to a minimum sum of absolute differences (SAD) error which is obtained over the top and bottom fields of the current MB. |
51 |
Bidirectionally predicted pictures or video object planes for efficient and flexible video coding |
US09988786 |
2001-11-20 |
US06704360B2 |
2004-03-09 |
Barin Geoffry Haskell; Atul Puri; Robert Louis Schmidt |
A method is provided for decoding a bit stream representing an image that has been encoded The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficents and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes. |
52 |
Bidirectionally predicted pictures or video object planes for efficient and flexible video coding |
US08827142 |
1997-03-27 |
US06404813B1 |
2002-06-11 |
Barin Geoffry Haskell; Atul Puri; Robert Louis Schmidt |
A method is provided for decoding a bit stream representing an image that has been encoded. The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficents and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes. |
53 |
Bidirectional predicted pictures or video object planes for efficient and flexible coding |
US12436366 |
2009-05-06 |
US08208547B2 |
2012-06-26 |
Barin Geoffry Haskell; Atul Puri; Robert Louis Schmidt |
A method is provided for decoding a bit stream representing an image that has been encoded The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficients and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes. |
54 |
Method and apparatus for reducing the bit rate in a video object planes
sequence coder |
US613963 |
1996-03-08 |
US5778100A |
1998-07-07 |
Homer H. Chen; Barin G. Haskell; Joern Ostermann |
A method and apparatus for reducing the output bit rate in a video object planes sequence encoder. An approximation of the shape of an object on a current frame is generated and used, along with a predicted image of the object from a subsequent frame and an estimate of the displacement of the object, to predict an image of the object in a current frame. A stripe is formed around the predicted image and the picture elements contained in the stripe are assigned texture values. The predicted image and stripe are subtracted from the current frame to form an error signal which is masked by the approximate shape of the object and provided to an entropy coder for transmission at an output bit rate. |
55 |
PROCEDE ET DISPOSITIFS DE REALITE AUGMENTEE UTILISANT UN SUIVI AUTOMATIQUE, EN TEMPS REEL, D'OBJETS GEOMETRIQUES PLANAIRES TEXTURES, SANS MARQUEUR, DANS UN FLUX VIDEO |
EP08761786.6 |
2008-01-18 |
EP2132710A2 |
2009-12-16 |
LEFEVRE, Valentin; LIVET, Nicolas |
The invention relates to a method and to devices for the real-time tracking of one or more substantially planar geometrical objects of a real scene in at least two images of a video stream for an augmented-reality application. After receiving a first image of the video stream (300), the first image including the object to be tracked, the position and orientation of the object in the first image are determined from a plurality of previously determined image blocks (320), each image block of said plurality of image blocks being associated with an exposure of the object to be tracked. The first image and the position and the orientation of the object to be tracked in the first image define a key image. After receiving a second image from the video stream, the position and orientation of the object to be tracked in the second image are evaluated from the key image (300). The second image and the corresponding position and orientation of the object to be tracked can be stored as a key image. If the position and the orientation of the object to be tracked cannot be found again in the second image from the key image, the position and the orientation of this object in the second image are determined from the plurality of image blocks and the related exposures (320). |
56 |
Method and apparatus for generating chrominance shape information of a
video object plane in a video signal |
US703594 |
1996-08-27 |
US5822460A |
1998-10-13 |
Jong-Il Kim |
An apparatus generates chrominance shape information based on luminance shape information represented by binary values, to thereby describe an object in a video object plane effectively. The apparatus divides the luminance shape information into a multiplicity of sample blocks, a sample block including 2.times.2 pixels, and determines, for each sample block, a chrominance value based on all logic values in a sample block. The chrominance shape information is produced in a form of a matrix based on the chrominance values for all sample blocks. |
57 |
AUGMENTED REALITY METHOD AND DEVICES USING A REAL TIME AUTOMATIC TRACKING OF MARKER-FREE TEXTURED PLANAR GEOMETRICAL OBJECTS IN A VIDEO STREAM |
US13606978 |
2012-09-07 |
US20120328158A1 |
2012-12-27 |
Valentin Lefevre; Nicolas Livet |
Methods and devices for the real-time tracking of an object in a video stream for an augmented-reality application are disclosed herein. |
58 |
Augmented reality method and devices using a real time automatic tracking of marker-free textured planar geometrical objects in a video stream |
US12524101 |
2008-01-18 |
US08315432B2 |
2012-11-20 |
Valentin Lefevre; Nicolas Livet |
The invention relates to a method and to devices for the real-time tracking of one or more substantially planar geometrical objects of a real scene in at least two images of a video stream for an augmented-reality application. After receiving a first image of the video stream (300), the first image including the object to be tracked, the position and orientation of the object in the first image are determined from a plurality of previously determined image blocks (320), each image block of said plurality of image blocks being associated with an exposure of the object to be tracked. The first image and the position and the orientation of the object to be tracked in the first image define a key image. After receiving a second image from the video stream, the position and orientation of the object to be tracked in the second image are evaluated from the key image (300). The second image and the corresponding position and orientation of the object to be tracked can be stored as a key image. If the position and the orientation of the object to be tracked cannot be found again in the second image from the key image, the position and the orientation of this object in the second image are determined from the plurality of image blocks and the related exposures (320). |
59 |
Augmented reality method and devices using a real time automatic tracking of marker-free textured planar geometrical objects in a video stream |
US13606978 |
2012-09-07 |
US08374395B2 |
2013-02-12 |
Valentin Lefevre; Nicolas Livet |
Methods and devices for the real-time tracking of an object in a video stream for an augmented-reality application are disclosed herein. |
60 |
Augmented reality method and devices using a real time automatic tracking of marker-free textured planar geometrical objects in a video stream |
US13606912 |
2012-09-07 |
US08374394B2 |
2013-02-12 |
Valentin Lefevre; Nicolas Livet |
Methods and devices for the real-time tracking of one or more objects of a real scene in a video stream for an augmented-reality application are disclosed herein. |