序号 专利名 申请号 申请日 公开(公告)号 公开(公告)日 发明人
41 Method and apparatus for encoding object information of a video object plane US984030 1997-12-02 US6125142A 2000-09-26 Seok-Won Han
A method for encoding a video signal including a video object plane(VOP) and object information thereof, the VOP having an object therein, divides the VOP into a plurality of blocks and detecting boundary blocks and object blocks, wherein each boundary block includes therein background pixels and object pixels and each object block includes therein object pixels only, the background and the object pixels representing pixels residing outside and inside the object, respectively. Then the object information is converted into a set of object symbols and each object symbol is added to a background pixel included in a boundary block to thereby produce processing blocks, the processing blocks including the object blocks and one or more processed boundary blocks having therein the object symbols and the remaining unprocessed boundary blocks. The processing blocks are encoded to thereby generate an encoded video signal.
42 Motion estimation and compensation of video object planes for interlaced digital video US10028007 2001-12-20 USRE38564E1 2004-08-10 Robert O. Eifrig; Xuemin Chen; Ajay Luthra
A motion estimation and compensation technique is provided for interlaced digital video such as video object planes (VOPs). Predictor motion vectors for use in differentially encoding a current field coded macroblock are obtained using the median of motion vectors of surrounding blocks or macroblocks. When a surrounding macroblock is itself interlaced coded, an average motion vector for that macroblock is used, with fractional pixel values being mapped to the half-pixel. When the current block is not interlaced coded but a surrounding block is, the field motion vectors may be used individually or averaged. In a repetitive padding technique for an interlaced coded VOP, the even and odd lines of the VOP and surrounding block are grouped. Within each field, exterior pixels are padded by setting them to the value of the nearest boundary pixel, or to an average of two boundary pixels. The lines are then reordered to provide a single padded reference VOP image.
43 Prediction and coding of bi-directionally predicted video object planes for interlaced digital video US944118 1997-10-06 US5991447A 1999-11-23 Robert O. Eifrig; Xuemin Chen; Ajay Luthra
A system for coding of digital video images such as bi-directionally predicted video object planes (B-VOPs), in particular, where the B-VOP and/or a reference image used to code the B-VOP is interlaced coded. For a B-VOP macroblock which is co-sited with a field predicted macroblock of a future anchor picture, direct mode prediction is made by calculating four field motion vectors, then generating the prediction macroblock. The four field motion vectors and their reference fields are determined from (1) an offset term of the current macroblock's coding vector, (2) the two future anchor picture field motion vectors, (3) the reference field used by the two field motion vectors of the co-sited future anchor macroblock, and (4) the temporal spacing, in field periods, between the current B-VOP fields and the anchor fields. Additionally, a coding mode decision process for the current MB selects a forward, backward, or average field coding mode according to a minimum sum of absolute differences (SAD) error which is obtained over the top and bottom fields of the current MB.
44 Bidirectionally predicted pictures or video object planes for efficient and flexible video coding US12030928 2008-02-14 US07545863B1 2009-06-09 Barin Geoffry Haskell; Atul Puri; Robert Lewis Schmidt
A method is provided for decoding a bit stream representing an image that has been encoded The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficients and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes.
45 Bidirectionally predicted pictures or video object planes for efficient and flexible video coding US11212262 2005-08-27 US07463685B1 2008-12-09 Barin Geoffry Haskell; Atul Puri; Robert Lewis Schmidt
A method is provided for decoding a bit stream representing an image that has been encoded The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficents and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes.
46 Bidirectionally predicted pictures or video object planes for efficient and flexible video coding US10728658 2003-12-06 US07356081B1 2008-04-08 Barin Geoffry Haskell; Atul Puri; Robert Louis Schmidt
A method is provided for decoding a bit stream representing an image that has been encoded. The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficients and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes.
47 Bidirectionally predicted pictures or video object planes for efficient and flexible video coding US09988786 2001-11-20 US20020075956A1 2002-06-20 Barin Geoffry Haskell; Atul Puri; Robert Louis Schmidt
A method is provided for decoding a bit stream representing an image that has been encoded The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficents and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes.
48 Motion estimation and compensation of video object planes for interlaced digital video US301141 1999-04-28 US6026195A 2000-02-15 Robert O. Eifrig; Xuemin Chen; Ajay Luthra
A motion estimation and compensation technique is provided for interlaced digital video such as video object planes (VOPs). Predictor motion vectors for use in differentially encoding a current field coded macroblock are obtained using the median of motion vectors of surrounding blocks or macroblocks. When a surrounding macroblock is itself interlaced coded, an average motion vector for that macroblock is used, with fractional pixel values being mapped to the half-pixel. When the current block is not interlaced coded but a surrounding block is, the field motion vectors may be used individually or averaged. In a repetitive padding technique for an interlaced coded VOP, the even and odd lines of the VOP and surrounding block are grouped. Within each field, exterior pixels are padded by setting them to the value of the nearest boundary pixel, or to an average of two boundary pixels. The lines are then reordered to provide a single padded reference VOP image.
49 Bidirectional predicted pictures or video object planes for efficient and flexible video coding US11924657 2007-10-26 US07656953B1 2010-02-02 Barin Geoffrey Haskell; Atul Puri; Robert Louis Schmidt
A method is provided for decoding a bit stream representing an image that has been encoded The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficients and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes.
50 Prediction and coding of bi-directionally predicted video object planes for interlaced digital video US09995997 2001-11-19 USRE38563E1 2004-08-10 Robert O. Eifrig; Xuemin Chen; Ajay Luthra
A system for coding of digital video images such as bi-directionally predicted video object planes (B-VOPs), in particular, where the B-VOP and/or a reference image used to code the B-VOP is interlaced coded. For a B-VOP macroblock which is co-sited with a field predicted macroblock of a future anchor picture, direct mode prediction is made by calculating four field motion vectors, then generating the prediction macroblock. The four field motion vectors and their reference fields are determined from (1) an offset term of the current macroblock's coding vector, (2) the two future anchor picture field motion vectors, (3) the reference field used by the two field motion vectors of the co-sited future anchor macroblock, and (4) the temporal spacing, in field periods, between the current B-VOP fields and the anchor fields. Additionally, a coding mode decision process for the current MB selects a forward, backward, or average field coding mode according to a minimum sum of absolute differences (SAD) error which is obtained over the top and bottom fields of the current MB.
51 Bidirectionally predicted pictures or video object planes for efficient and flexible video coding US09988786 2001-11-20 US06704360B2 2004-03-09 Barin Geoffry Haskell; Atul Puri; Robert Louis Schmidt
A method is provided for decoding a bit stream representing an image that has been encoded The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficents and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes.
52 Bidirectionally predicted pictures or video object planes for efficient and flexible video coding US08827142 1997-03-27 US06404813B1 2002-06-11 Barin Geoffry Haskell; Atul Puri; Robert Louis Schmidt
A method is provided for decoding a bit stream representing an image that has been encoded. The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficents and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes.
53 Bidirectional predicted pictures or video object planes for efficient and flexible coding US12436366 2009-05-06 US08208547B2 2012-06-26 Barin Geoffry Haskell; Atul Puri; Robert Louis Schmidt
A method is provided for decoding a bit stream representing an image that has been encoded The method includes the steps of: performing an entropy decoding of the bit stream to form a plurality of transform coefficients and a plurality of motion vectors; performing an inverse transformation on the plurality of transform coefficients to form a plurality of error blocks; determining a plurality of predicted blocks based on bidirectional motion estimation that employs the motion vectors, wherein the bidirectional motion estimation includes a direct prediction mode and a second prediction mode; and, adding the plurality of error blocks to the plurality of predicted blocks to form the image. The second prediction mode may include forward, backward, and interpolated prediction modes.
54 Method and apparatus for reducing the bit rate in a video object planes sequence coder US613963 1996-03-08 US5778100A 1998-07-07 Homer H. Chen; Barin G. Haskell; Joern Ostermann
A method and apparatus for reducing the output bit rate in a video object planes sequence encoder. An approximation of the shape of an object on a current frame is generated and used, along with a predicted image of the object from a subsequent frame and an estimate of the displacement of the object, to predict an image of the object in a current frame. A stripe is formed around the predicted image and the picture elements contained in the stripe are assigned texture values. The predicted image and stripe are subtracted from the current frame to form an error signal which is masked by the approximate shape of the object and provided to an entropy coder for transmission at an output bit rate.
55 PROCEDE ET DISPOSITIFS DE REALITE AUGMENTEE UTILISANT UN SUIVI AUTOMATIQUE, EN TEMPS REEL, D'OBJETS GEOMETRIQUES PLANAIRES TEXTURES, SANS MARQUEUR, DANS UN FLUX VIDEO EP08761786.6 2008-01-18 EP2132710A2 2009-12-16 LEFEVRE, Valentin; LIVET, Nicolas
The invention relates to a method and to devices for the real-time tracking of one or more substantially planar geometrical objects of a real scene in at least two images of a video stream for an augmented-reality application. After receiving a first image of the video stream (300), the first image including the object to be tracked, the position and orientation of the object in the first image are determined from a plurality of previously determined image blocks (320), each image block of said plurality of image blocks being associated with an exposure of the object to be tracked. The first image and the position and the orientation of the object to be tracked in the first image define a key image. After receiving a second image from the video stream, the position and orientation of the object to be tracked in the second image are evaluated from the key image (300). The second image and the corresponding position and orientation of the object to be tracked can be stored as a key image. If the position and the orientation of the object to be tracked cannot be found again in the second image from the key image, the position and the orientation of this object in the second image are determined from the plurality of image blocks and the related exposures (320).
56 Method and apparatus for generating chrominance shape information of a video object plane in a video signal US703594 1996-08-27 US5822460A 1998-10-13 Jong-Il Kim
An apparatus generates chrominance shape information based on luminance shape information represented by binary values, to thereby describe an object in a video object plane effectively. The apparatus divides the luminance shape information into a multiplicity of sample blocks, a sample block including 2.times.2 pixels, and determines, for each sample block, a chrominance value based on all logic values in a sample block. The chrominance shape information is produced in a form of a matrix based on the chrominance values for all sample blocks.
57 AUGMENTED REALITY METHOD AND DEVICES USING A REAL TIME AUTOMATIC TRACKING OF MARKER-FREE TEXTURED PLANAR GEOMETRICAL OBJECTS IN A VIDEO STREAM US13606978 2012-09-07 US20120328158A1 2012-12-27 Valentin Lefevre; Nicolas Livet
Methods and devices for the real-time tracking of an object in a video stream for an augmented-reality application are disclosed herein.
58 Augmented reality method and devices using a real time automatic tracking of marker-free textured planar geometrical objects in a video stream US12524101 2008-01-18 US08315432B2 2012-11-20 Valentin Lefevre; Nicolas Livet
The invention relates to a method and to devices for the real-time tracking of one or more substantially planar geometrical objects of a real scene in at least two images of a video stream for an augmented-reality application. After receiving a first image of the video stream (300), the first image including the object to be tracked, the position and orientation of the object in the first image are determined from a plurality of previously determined image blocks (320), each image block of said plurality of image blocks being associated with an exposure of the object to be tracked. The first image and the position and the orientation of the object to be tracked in the first image define a key image. After receiving a second image from the video stream, the position and orientation of the object to be tracked in the second image are evaluated from the key image (300). The second image and the corresponding position and orientation of the object to be tracked can be stored as a key image. If the position and the orientation of the object to be tracked cannot be found again in the second image from the key image, the position and the orientation of this object in the second image are determined from the plurality of image blocks and the related exposures (320).
59 Augmented reality method and devices using a real time automatic tracking of marker-free textured planar geometrical objects in a video stream US13606978 2012-09-07 US08374395B2 2013-02-12 Valentin Lefevre; Nicolas Livet
Methods and devices for the real-time tracking of an object in a video stream for an augmented-reality application are disclosed herein.
60 Augmented reality method and devices using a real time automatic tracking of marker-free textured planar geometrical objects in a video stream US13606912 2012-09-07 US08374394B2 2013-02-12 Valentin Lefevre; Nicolas Livet
Methods and devices for the real-time tracking of one or more objects of a real scene in a video stream for an augmented-reality application are disclosed herein.
QQ群二维码
意见反馈