Inter-frame predicted image synthesizing method

Pulse or digital communications – Bandwidth reduction or expansion – Television or motion video signal

Reexamination Certificate

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

Reexamination Certificate

active

06526095

ABSTRACT:

BACKGROUND OF THE INVENTION
The present invention relates to a method and apparatus for coding and decoding an image. More particularly the present invention relates to a method and apparatus of coding and decoding an image by the use of a motion compensation method wherein an image is divided, into patches and each patch is deformed by affine or bilinear transformation.
In the case of high-efficiency coding of a dynamic image by a coder, it is known that motion compensation is very effective for data compression due to similarity between frames temporally close to each other. The processing of an image by motion compensation is performed according to equation 1, shown below. In this case, it is assumed that the predicted image of a frame (current frame) to be coded is P(x,y) and a reference image (decoded image of a frame which is temporally close to P and whose coding is already completed) is R(x,y). Moreover, it is assumed that x and y are integers and a pixel is present at a point whose coordinate values are integers for P and R. In this case, the relation between P and R is also shown by the equation 1 shown below.
P
(
x,y
)=
R
(
f
i
(
x,y
),
g
i
(
x,y
)),(
x,y
)&egr;
P
i
,0≦
i<N
  Equation 1
In this case, Pi denotes a pixel included in the i-th patch of an image by assuming that the image is divided into N small regions (patches). Moreover, transformation functions fi(x,y) and gi(x,y) show the spatial correspondence between the image of a current frame and a reference image. In this case, the motion vector of the pixel (x,y) in Pi can be shown by (fi(x,y)−x, gi(x,y)−y) by using the coordinates of a pixel in a predicted image as a starting point and the coordinates of a correspondent point in the reference image as an ending point.
In the case of H.261, of Motion Picture Experts Group (MPEG)1, and MPEG2 which are the international standards for the video coding method, a method referred to as block matching is used in which fi(x,y)−x and gi(x,y)−y are constants unrelated to x or y. However, to achieve a data compression ratio higher than these standard coding methods, it is required to use a higher-level motion compensation method. As the above new motion compensation method, a motion compensation method has recently been proposed which allows that fi(x,y)−x and gi(x,y)−y are not constants but pixels in the same patch have different motion vectors. As transformation functions of these methods, the following examples have been disclosed.
“Basic study of motion compensation according to triangle patch”, by Nakaya et al., Technical report of IEICE, IE90-106, Hei 2-03 discloses an example of affine transform as follows:
fi
(
x,y
)=
a
i0
x+a
i1
y+a
i2
gi
(
x,y
)=
a
i3
x+a
i4
y+a
i5
  Equation 2
“Motion compensation for video compression control grid interpolation”, by G. J. Sullivan et al., Proc. ICASSP '91, M9.1, pp. 2713-2716, 1991-05 discloses an example of bilinear transform as follows:
fi
(
x,y
)=
b
i0
xy+b
i1
x+b
i2
y+b
i3
gi
(
x,y
)=
b
i4
xy+b
i5
x+b
i6
y+b
i7
  Equation 3
In the above equations, aij and bij denote motion parameters estimated for each patch. When the value of a transformation function is not an integer, coordinate values are not integers. Therefore, it is necessary to obtain the luminance value of a point where no pixel is actually present in a reference image. In this case, bilinear interpolation using four peripheral pixels is frequently performed. When describing the above interpolation method in the form of the equation, R(x+&xgr;,y+&eegr;) is shown below by assuming 0≦&xgr;, &eegr;<1.
R
(
x+&xgr;,y+&eegr;
)=(1−&eegr;)((1−&xgr;)
R
(
x,y
)+&xgr;
R
(
x+
1,
y
))+&eegr;(1−&xgr;)
R
(
x,y+
1)+&xgr;
R
(
x+
1
,y+
1))   Equation 4
Hereinafter, the motion compensation method using the transformation functions in the above equations 2 and 3 is referred to as warping prediction.
To transmit motion information, a video coder must transmit information capable of specifying the motion parameter of a transformation function to the receiving side in some manner. For example, the transformation function uses affine transform and the shape of a patch is assumed as a triangle. In this case, even if six motion parameters are directly transmitted to or the motion vectors of three apexes of the patch are transmitted to the receiving side, it must be possible for the receiving side to reproduce the six motion parameters ai1 to ai6.
FIGS. 1
a-d
illustrate an example of the motion compensation method of transmitting motion vectors of apexes (grid points) of a triangle patch.
FIGS. 1
a-d
illustrate the processing for synthesizing a predicted image of an original image
102
of a current frame by using a reference image
101
. First, the current frame is divided into a plurality of polygonal patches and formed into a patch-divided image
108
as illustrated in
FIG. 1
d
. An apex of a patch is referred to as a grid point and each grid point is shared by a plurality of patches. For example, a patch
109
is formed with grid points
110
,
111
, and
112
and these grid points also serve as apexes of another patch. Thus, after an image is divided into a plurality of patches, motion estimation is performed.
In the case of the example shown in
FIGS. 1
a-b
, motion estimation is performed between the predicted image and the reference image for each grid point. As a result, each patch is deformed by a reference image
103
after motion estimation as illustrated in
FIG. 1
c
. For example, the patch
109
corresponds to a deformed patch
104
. This is because it is estimated that grid points
105
,
106
, and
107
respectively correspond to the grid points
110
,
111
, and
112
. In the case of this example, when assuming the coordinates of the grid points
110
,
111
, and
112
as (I,J), (I+r,J), and (I,J+s) (where I, J, r, and s are integers) respectively and the motion vectors of the points as (U
1
,V
1
), (U
2
,V
2
), and (U
3
,V
3
) respectively, the motion vector (ua(x,y),va(x,y)) at the point (x,y) in the patch can be shown by the following equation 5 in accordance with the relation of the equation 2.
u
a

(
x
,
y
)
=


U
1
-
U
0
r

(
x
-
I
)
+
U
2
-
U
0
s

(
y
-
J
)
+
U
0
v
a

(
x
,
y
)
=


V
1
-
V
0
r

(
x
-
I
)
+
V
2
-
V
0
s

(
y
-
J
)
+
V
0
Equation 5
By using the relation, it is possible to obtain the motion vector for each pixel and synthesize a predicted image.
Moreover, when deforming a quadrangular patch by using bilinear transform, the motion vector ub(x,y),vb(x,y) at a point (x,y) in the patch can be shown by the following equation in accordance with the relation of the equation 3 by assuming the coordinates of grid points as (I,J), (I+r,J), (I,J+s), and (I+r,J+s) and the motion vectors of the points as (U
1
,V
1
), (U
2
,V
2
), (U
3
,V
3
), and (U
4
,V
4
).
ub

(
x
,
y
)
=


J
+
s
-
y
s

(
I
+
r
-
x
r

U
0
+
x
-
I
r

U
1
)
+
y
-
J
s

(
I
+
r
-
x
r

U
2
+
x
-
I
r

U
3
)
=


U
0
-
U
1
-
U
2
+
U
3
rs

(
x
-
I
)

(
y
-
J
)
+
-
U
0
+
U
1
r

(
x
-
I
)
+


-
U
0
+
U
2
s

(
y
-
J
)
+
U
0
vb

(
x
,
y
)
=


V
0
-
V
1
-
V
2
+
V
3
rs

(
x
-
I
)

(
y
-
J
)
+
-
V
0
+
V
1
r

(
x
-
I
)
+


-
V
0
+
V
2
s

(
y
-
J
)
+
V
0
Equation 6
By introducing the above-described warping prediction, it is possible to accurately approximate the motion vectors of an image sequence and realize a high data compression ratio. However, the throughput for coding and decoding decreases compared to a conventional method. Particularly, the divisions performed in the equations 5 and 6 are large factors for complicating the processing. Therefore, the warping prediction using affine transform and bilinear transform has a problem that the throughput

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Inter-frame predicted image synthesizing method does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Inter-frame predicted image synthesizing method, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Inter-frame predicted image synthesizing method will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-3146705

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.