US7659944B2 - Digital video signal processing apparatus and method for field-based adaptive spatio-temporal Y/C separation - Google Patents

Digital video signal processing apparatus and method for field-based adaptive spatio-temporal Y/C separation Download PDF

Info

Publication number
US7659944B2
US7659944B2 US11/331,360 US33136006A US7659944B2 US 7659944 B2 US7659944 B2 US 7659944B2 US 33136006 A US33136006 A US 33136006A US 7659944 B2 US7659944 B2 US 7659944B2
Authority
US
United States
Prior art keywords
data
field
coefficient
mask
video signal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US11/331,360
Other versions
US20060181648A1 (en
Inventor
Sung-Cheol Park
Hyung-Jun Lim
Jae-Hong Park
Kyoung-Mook Lim
Heo-Jin Byeon
Dong-Suk Shin
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Assigned to SAMSUNG ELECTRONICS CO., LTD. reassignment SAMSUNG ELECTRONICS CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LIM, HYUNG-JUN, LIM, KYOUNG-MOOK, PARK, JAE-HONG, PARK, SUNG-CHEOL, SHIN, DONG-SUK, BYEON, HEO-JIN
Publication of US20060181648A1 publication Critical patent/US20060181648A1/en
Application granted granted Critical
Publication of US7659944B2 publication Critical patent/US7659944B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/77Circuits for processing the brightness signal and the chrominance signal relative to each other, e.g. adjusting the phase of the brightness signal relative to the colour signal, correcting differential gain or differential phase
    • H04N9/78Circuits for processing the brightness signal and the chrominance signal relative to each other, e.g. adjusting the phase of the brightness signal relative to the colour signal, correcting differential gain or differential phase for separating the brightness signal or the chrominance signal from the colour television signal, e.g. using comb filter
    • EFIXED CONSTRUCTIONS
    • E02HYDRAULIC ENGINEERING; FOUNDATIONS; SOIL SHIFTING
    • E02DFOUNDATIONS; EXCAVATIONS; EMBANKMENTS; UNDERGROUND OR UNDERWATER STRUCTURES
    • E02D5/00Bulkheads, piles, or other structural elements specially adapted to foundation engineering
    • E02D5/74Means for anchoring structural elements or bulkheads
    • E02D5/80Ground anchors
    • E02D5/808Ground anchors anchored by using exclusively a bonding material
    • EFIXED CONSTRUCTIONS
    • E02HYDRAULIC ENGINEERING; FOUNDATIONS; SOIL SHIFTING
    • E02DFOUNDATIONS; EXCAVATIONS; EMBANKMENTS; UNDERGROUND OR UNDERWATER STRUCTURES
    • E02D2600/00Miscellaneous
    • E02D2600/20Miscellaneous comprising details of connection between elements
    • EFIXED CONSTRUCTIONS
    • E02HYDRAULIC ENGINEERING; FOUNDATIONS; SOIL SHIFTING
    • E02DFOUNDATIONS; EXCAVATIONS; EMBANKMENTS; UNDERGROUND OR UNDERWATER STRUCTURES
    • E02D2600/00Miscellaneous
    • E02D2600/30Miscellaneous comprising anchoring details
    • EFIXED CONSTRUCTIONS
    • E02HYDRAULIC ENGINEERING; FOUNDATIONS; SOIL SHIFTING
    • E02DFOUNDATIONS; EXCAVATIONS; EMBANKMENTS; UNDERGROUND OR UNDERWATER STRUCTURES
    • E02D2600/00Miscellaneous
    • E02D2600/40Miscellaneous comprising stabilising elements

Definitions

  • the present invention relates to a digital video signal processing apparatus. More particularly, the present invention relates to a digital video signal processing apparatus and method for field-based adaptive spatio-temporal, i.e., 3-dimensional (3D), Y/C separation in NTSC/PAL (National Television System Committee/Phase Alternation by Line) systems.
  • 3-dimensional (3D) 3-dimensional
  • NTSC/PAL National Television System Committee/Phase Alternation by Line
  • the display of an NTSC/PAL broadcasting system includes a device for processing a CVBS (Composite Video Blanking Sync) signal, which is a composite of a Y (luminance) signal and a C (chrominance) signal.
  • the C signal is quadrature-amplitude-modulated with a subcarrier frequency fsc.
  • characteristics of the C signal are determined by its frequency and phase.
  • a digital video signal processing apparatus at a receiving stage separates Y and C signals with reference to the characteristic of the C signal and displays an image based on the separated signals.
  • FIG. 1 illustrates a block diagram of a conventional video signal processing apparatus 100 .
  • the video signal processing apparatus 100 includes a comb filter 110 , a one-dimensional bandpass filter (1D-BPF) 120 , a weight determination unit 130 , a combiner 140 and a subtracter 150 .
  • the comb filter 110 one-dimensionally band-pass-filters an input video signal in the vertical direction.
  • the 1D-BPF 120 one-dimensionally band-pass-filters the input video signal in the horizontal direction.
  • the weight determination unit 130 determines weights of the output of the comb filter 110 and the output of the 1D-BPF 120 with reference to vertical/horizontal correlation and the phase of a C signal.
  • the combiner 140 combines the output signals of the comb filter 110 and 1D-BPF 120 using the weights to generate a C signal.
  • the subtracter 150 subtracts the C signal from the input CVBS signal to generate a Y signal.
  • FIG. 2 illustrates a block diagram of another conventional video signal processing apparatus 200 .
  • the video signal processing apparatus 200 includes a 2D-BPF 210 , a subtracter 220 and a post-processor 230 .
  • the 2D-BPF 210 performs a two-dimensional convolution to extract a modulated C signal.
  • the extracted C signal and a Y signal generated by the subtracter 220 are processed by the post-processor 230 .
  • the post-processor 230 compensates the Y/C signal to generate a compensated Y/C signal.
  • Y/C separation when the edge of a detected image has high vertical correlation, Y/C separation is performed based on comb filtering and when the edge of a detected image has high horizontal correlation, Y/C separation is performed based on 1D bandpass filtering.
  • system performance largely depends on a threshold used for edge detection. That is, Y/C separation can be executed incorrectly or unstably when the filtering method is wrongly selected due to inaccurate edge detection.
  • Conventional techniques that do not select one of the filtering methods but combine the results of the filtering operations can solve this problem to some extent. However, these techniques are based on horizontal or vertical one-dimensional filtering, and thus, artifacts may remain in the generated signal due to the inconstant edge directions.
  • cross-luma which occurs when a C component exists in the separated Y signal, resulting in dotted artifacts
  • cross-color which occurs when a Y component exists in the separated C signal, resulting in a rainbow pattern artifact
  • spatio-temporal filtering is used by conventional video signal processing apparatuses.
  • the correlations of pixel data of the previous field and pixel data of the next field with the current pixel or the correlations of pixel data of the previous frame and pixel data of the next frame with the current pixel is considered.
  • the spatio-temporal filtering method requires a memory for storing the pixel data of the previous and next fields or frames.
  • the spatio-temporal filter is more expensive than the spatial filter, the spatio-temporal filtering technique is frequently used when images with high picture quality are required.
  • conventional Y/C separation techniques which discontinuously select the spatial filter and the spatio-temporal filter in response to inter-frame/intra-frame correlation of the CVBS signal, cause artifacts such as cross-luma and cross-color when there is an error in the correlation measurement result.
  • the conventional techniques have limitations in two-dimensional spatial Y/C separation.
  • conventional Y/C separation techniques using intra-frame correlation have an advantage of using pixel data from the same spatial location.
  • these techniques use data before and after one frame temporally distant by more than one field, resulting in deterioration of the spectrum characteristics of filters.
  • the present invention is therefore directed to a digital video signal processing apparatus and method, which substantially overcome one or more of the problems due to the limitations and disadvantages of the related art.
  • At least one of the above and other features and advantages of the present invention may be realized by providing a video signal processing method including generating 2D weight coefficients and 3D weight coefficients using data of a plurality of fields of an input video signal and generating a C signal of the input video signal by carrying out convolution of a horizontal coefficient mask, a vertical coefficient mask and a time coefficient mask, which are composed based on the weight coefficients, and corresponding data windows of the plurality of field data, wherein generating the C signal includes local vertical comb filtering, horizontal band pass filtering and 2D band pass filtering according to spatio-temporal local characteristics of the input video signal.
  • Local vertical comb filtering may be performed when local characteristics indicate high vertical correlation.
  • Local horizontal band pass filtering may be performed when local characteristics indicate high horizontal correlation.
  • 2D or 3D band pass filtering may be carried out when local characteristics indicate high or low correlation in multiple directions using at least three field data having an interval of one field.
  • the 2D weight coefficients may include a first coefficient proportional to vertical and upward correlations with the current pixel, a second coefficient proportional to vertical and downward correlations with the current pixel, a third coefficient proportional to horizontal and leftward correlations with the current pixel, and a fourth coefficient proportional to horizontal and rightward correlations with the current pixel.
  • the generated C signal may be subtracted from the input video signal and outputting as a Y signal.
  • Generating the C signal may include performing convolution of a corresponding data window of a first field data of the input video signal and a first filter mask, performing convolution of a corresponding data window of a second field data of the input video signal and a second filter mask, performing convolution of a corresponding data window of a third field data of the input video signal and a third filter mask, and combining the convolution results to output the combined result as the C signal, wherein the first, second and third filter masks are generated from 3D convolution of the horizontal coefficient mask, the vertical coefficient mask and the time coefficient mask.
  • the method may be applied to a PAL system or an NTSC system.
  • the first field data, the second field data and the third field data may have an interval of one field.
  • the horizontal coefficient mask, the vertical coefficient mask and the time coefficient mask may each be one-dimensional.
  • the 3D weight coefficients may include a first coefficient in proportion to a correlation between the currently processed pixel data and data of the field before the current field, and a second coefficient in proportion to a correlation between the currently processed pixel data and data of the field after the current field.
  • Generating the C signal may further include performing convolution of a corresponding data window of a fourth field data of the input video signal and a fourth filter mask, performing convolution of a corresponding data window of a fifth field data of the input video signal and a fifth filter mask, combining the five convolution results to be output the combined result as the generated C signal, wherein the time coefficient mask includes first and second coefficient masks, and the first to fifth filter masks are generated from 3D convolution of the horizontal coefficient mask, the vertical coefficient mask, the first coefficient mask and the second coefficient mask.
  • the 3D weight coefficients used for the first coefficient mask may include a first coefficient in proportion to the correlation in a first direction between the currently processed pixel data and data of the field before the current field, and a second coefficient in proportion to the correlation in the second direction between the currently processed pixel data and data of the field after the current field.
  • the 3D weight coefficients used for the second coefficient mask may include a third coefficient in proportion to the correlation in a third direction between the currently processed pixel data and data of the field before the current field, and a fourth coefficient in proportion to the correlation in a fourth direction between the currently processed pixel data and data of the field after the current field.
  • a video signal processing apparatus including a weight determination unit to generate 2D weight coefficients and 3D weight coefficients using data of a plurality of fields of an input video signal and a filter to convolute a horizontal coefficient mask, a vertical coefficient mask and a time coefficient mask, the masks respectively composed of the weight coefficients, with corresponding data windows of the data of the plurality of fields to generate a C signal of the input video signal, wherein the filter adaptively performs local vertical comb filtering, horizontal band pass filtering and 2D band pass filtering in response to spatio-temporal local characteristics of the input video signal.
  • FIG. 1 illustrates a block diagram of a conventional video signal processing apparatus
  • FIG. 2 illustrates a block diagram of another conventional video signal processing apparatus
  • FIG. 3 illustrates a block diagram of a video signal processing apparatus according to an embodiment of the present invention
  • FIG. 4 illustrates chrominance signal phases of an NTSC system
  • FIG. 5 illustrates chrominance signal phases of a PAL system
  • FIG. 6 illustrates spatio-temporal chrominance signal phases of the NTSC system
  • FIG. 7 illustrates spatio-temporal chrominance signal phases of the PAL system
  • FIG. 8 illustrates a diagram for explaining a horizontal filter coefficient of the NTSC system
  • FIG. 9 illustrates a diagram for explaining a vertical filter coefficient of the NTSC system
  • FIG. 10 is a diagram for explaining a first temporal filter coefficient of the NTSC system
  • FIG. 11 illustrates a diagram for explaining a second temporal filter coefficient of the NTSC system
  • FIG. 12 is a diagram for explaining a vertical/horizontal filter coefficient of the NTSC system
  • FIG. 13 illustrates a diagram for explaining a vertical/temporal filter coefficient of the NTSC system
  • FIG. 14 illustrates a diagram for explaining a horizontal filter coefficient of the PAL system
  • FIG. 15 illustrates a diagram for explaining a vertical filter coefficient of the PAL system
  • FIG. 16 illustrates a diagram for explaining a temporal filter coefficient of the PAL system
  • FIG. 17 illustrates a vertical variation in pixels
  • FIG. 18 illustrates a horizontal variation in pixels
  • FIG. 19 illustrates a block diagram of the 3D BPF of FIG. 3 ;
  • FIG. 20 illustrates spectrum characteristics representing Y and C components with respect to horizontal and vertical frequencies of a video signal
  • FIG. 21 illustrates spectrum characteristics representing Y and C components with respect to temporal and vertical frequencies of a video signal
  • FIG. 22 illustrates a block diagram of the 3D BPF of FIG. 3 .
  • Korean Patent Application No. 10-2005-0003178 filed on Jan. 13, 2005, in the Korean Intellectual Property Office, and entitled “Digital Video Signal Processing Apparatus and Method for Adaptive and Temporal and Spatial Y/C Separation Based on Field Period,” is incorporated by reference herein in its entirety.
  • FIG. 3 illustrates a block diagram of a video signal processing apparatus 300 according to an embodiment of the present invention.
  • the video signal processing apparatus 300 may include a memory 310 , a weight determination unit 320 , a 3D BPF 330 , and a subtracter 340 .
  • the video signal processing apparatus 300 can be used for both NTSC systems and PAL systems.
  • the video signal processing apparatus 300 receives a digital CVBS signal as an input video signal and separates the input video signal into a Y signal and a C signal.
  • the input video signal can be a digital signal obtained by sampling the active video region of an analog CVBS signal at a predetermined frequency e.g., 4f sc , where f sc is a subcarrier frequency.
  • CVBS(t) Y+U *sin2 ⁇ f SC t+V *cos2 ⁇ f SC t [Equation 1]
  • pixel signals sampled at 4f sc have chrominance signal phases as shown in FIG. 4 . That is, the pixel signals are repeated in the form of Y+U, Y+V, Y ⁇ U, Y ⁇ V, . . . in each horizontal line.
  • FIG. 4 shows only the phases of chrominance signal components. In the NTSC system, the chrominance signal phases are shifted by 180° in the vertical direction in neighboring horizontal scan lines, as shown in FIG. 4 .
  • the input video signal CVBS(t) of the NTSC system may be represented spatio-temporally, as shown in FIG. 6 .
  • the phase of the C component of the currently processed central pixel at (i, j, t) has a particular phase, e.g., +U.
  • Pixel signals spatio-temporally disposed in a first diagonal direction from the central pixel, i.e., a pixel at (i ⁇ 1, j, t+1) of the next field and a pixel at (i, j, t ⁇ 1) of the previous field, have the same phase as the central pixel.
  • Pixel signals spatio-temporally disposed in a second diagonal direction from the central pixel i.e., a pixel at (i, j, t+1) of the next field and a pixel at (i ⁇ 1, j, t ⁇ 1) of the previous field, have a phase opposite the phase, e.g., ⁇ U, of the C component of the central pixel at (i, j, t).
  • a phase opposite the phase e.g., ⁇ U, of the C component of the central pixel at (i, j, t).
  • one frame is composed of two fields, i.e., an odd field and an even field
  • CVBS(t) Y+U *sin2 ⁇ f SC t ⁇ V *cos2 ⁇ f SC t [Equation 2]
  • pixel signals in the PAL system have chrominance signal phases as shown in FIG. 5 . That is, the pixel signals are repeated in the form of Y+U, Y+V, Y ⁇ U, Y ⁇ V, . . . or Y+U, Y ⁇ V, Y ⁇ U, Y+V, . . . . in each horizontal line, and the chrominance signal phases are shifted by 180° in the vertical direction every other horizontal scan line.
  • the input video signal CVBS(t) of the PAL system may be represented spatio-temporally, as shown in FIG. 7 .
  • the phase of the C signal component of the currently processed central pixel at (i, j, t) has a particular phase, e.g., +U.
  • Pixel signals spatio-temporally disposed in a diagonal direction from the current pixel, i.e., a pixel at (i ⁇ 1, j, t+1) of the next field and a pixel at (i, j, t ⁇ 1) of the previous field, have a phase opposite the phase of the current pixel signal (i, j, t) of the current field.
  • a pixel at (i+1, j, t+2) located one frame (two fields) after the central pixel and a pixel at (i, j, t ⁇ 2) located one frame (two fields) before the central pixel also have the phase opposite the phase of the current pixel at (i, j, t) of the current field.
  • Pixels disposed vertically from the central pixel i.e., pixels at (i ⁇ 2, j, t) and (i+2, j, t) of the same field, also have the phase opposite the phase of the current pixel at (i, j, t) of the current field.
  • the Y and C signals (U and V signals) separated by the video signal processing apparatus 300 are converted to a format required for an external circuit and then stored or transmitted to a display device.
  • the Y and C signals can be interpolated with three color signals, e.g., red, green and blue color signals, to be displayed on a liquid crystal display (LCD).
  • LCD liquid crystal display
  • the video signal processing apparatus 300 adaptively operates in response to spatio-temporal local characteristics of the input video signal CVBS.
  • the memory 310 stores digital data corresponding to a plurality of fields of the input video signal CVBS.
  • the memory 310 includes a plurality of field memories 311 , 312 , 313 and 314 which each store data corresponding to one field. While FIG. 3 shows four field memories 311 , 312 , 313 and 314 , more memories can be added to the digital video signal processing apparatus 300 if required for filtering.
  • the data corresponding to the plurality of fields, stored in the memory 310 , and currently input video data are output to the weight determination unit 320 .
  • the weight determination unit 320 generates 2D weight coefficients Wu, Wd, Wl and Wr, which will be used for filtering in the 3D BPF 330 , using the plurality of field data. Furthermore, the weight determination unit 320 generates 3D weight coefficients Wb 1 , Wb 2 , Wf 1 and Wf 2 /Wb and Wf in response to a network parameter control (NPC) signal.
  • NPC network parameter control
  • 3D weight coefficients Wb 1 , Wb 2 , Wf 1 and Wf 2 may be generated for the NTSC system having the chrominance signal phases shown in FIG. 4 .
  • the control signal NPC has a logic low level
  • 3D weight coefficients Wb and Wf may be generated for the PAL system having the chrominance signal phases shown in FIG. 5 .
  • While the 2D weight coefficients Wu, Wd, Wl and Wr and the 3D weight coefficients Wb 1 , Wb 2 , Wf 1 and Wf 2 /Wb and Wf may be generated for both the NTSC and PAL systems, data used to generate the coefficients for the NTSC system may be different from data used to generate the coefficients for the PAL system.
  • the 3D BPF 330 generates the C signal of the input video signal using the 2D weight coefficients Wu, Wd, Wl and Wr and the 3D weight coefficients Wb 1 , Wb 2 , Wf 1 and Wf 2 /Wb and Wf.
  • the 3D BPF 330 uses corresponding data windows of three fields or corresponding data windows of five fields, which will be convoluted with filter masks in response to the control signal NPC. For example, when the control signal NPC has a logic high level, the 3D BPF 330 may be operated for the NTSC system and use five data windows BBFID (before before field), BEFID, (before field), PRFID (present field), AFFID (after field) and AAFID (after after field). When the control signal NPC has a logic low level, the 3D BPF 330 may be operated for the PAL system and use three data windows BEFID, PRFID and AFFID.
  • the subtracter 340 subtracts the C signal (U or V signal), output from the 3D BPF 330 , from the video signal CVBS of the currently processed pixel to obtain the Y signal. For example, when the U signal output as the C signal is subtracted from the video signal of the current pixel Y+U, the Y signal is obtained.
  • the 3D BPF 330 continuously and adaptively performs local vertical comb filtering, local horizontal band pass filtering and 2D/3D band pass filtering, in response to spatio-temporal local characteristics of the input video signal. That is, the 3D BPF 330 performs filtering using filter masks for convolution of field data items having an interval of one field. The 3D BPF 330 performs local vertical comb filtering if the local characteristics of the central pixel to be filtered show high vertical correlation, and executes local horizontal band pass filtering if the local characteristics show high horizontal correlation. Furthermore, the 3D BPF 330 performs 2D/3D band pass filtering if the local characteristics show high or low correlation in multiple directions.
  • the 3D BPF 330 performs filtering using the filter masks, that is, a horizontal coefficient mask, a vertical coefficient mask and a time coefficient mask, which are composed based on the 2D weight coefficients Wu, Wd, Wl and Wr and the 3D weight coefficients Wb 1 , Wb 2 , Wf 1 and Wf 2 /Wb and Wf.
  • the time coefficient mask is composed of first and second time coefficient masks, and the 2D weight coefficients Wu, Wd, Wl and Wr and the 3D weight coefficients Wb 1 , Wb 2 , Wf 1 and Wf 2 are used for filtering.
  • the horizontal coefficient mask is h h (i, j, t)
  • the vertical coefficient mask is h v (i, j, t)
  • the first time coefficient mask is h t1 (i, j, t)
  • the second time coefficient mask is h t2 (i, j, t)
  • the C signal C(i, j, t) output from the 3D BPF 330 is represented as follows.
  • the 3D BPF 330 outputs the result of 3-dimensional convolution of the horizontal coefficient mask h h (i, j, t), the vertical coefficient mask h v (i, j, t), the first time coefficient mask h t1 (i, j, t), the second time coefficient mask h t2 (i, j, t) and the corresponding data window CVBS(i, j, t) of the plurality of field data items stored in the memory 310 as the C signal C(i, j, t) of the input video signal.
  • the Y signal Y(i, j, t) output from the subtracter 340 is represented as the result of subtracting the C signal C(i, j, t) from the CVBS data CVBS(i, j, t) of the currently processed central pixel, which is represented as follows.
  • Y ( i,j,t ) CVBS ( i,j,t ) ⁇ C ( i,j,t ) [Equation 4]
  • the horizontal coefficient mask h h (i, j, t), the vertical coefficient mask h v (i, j, t), the first time coefficient mask h t1 (i, j, t) and the second time coefficient mask h t2 (i, j, t) have a 1-dimensional form as represented by Equation 5.
  • coefficients Nh, Nv, Nt 1 and Nt 2 can be set to values when normalization is carried out such that the sum of the absolute values of Nh, Nv, Nt 1 and Nt 2 and the absolute values of the 2D weight coefficients and 3D weight coefficients are equal to 1.
  • Equation 6 3D convolution of the four 1D coefficient masks h h , h v , h t1 and h t2 can be represented by Equation 6.
  • Pixel data items corresponding to the horizontal/vertical coefficient mask h hv (i, j, t) and vertical/time coefficient mask h t1t2 (i, j, t) of the NTSC system are shown in FIGS. 12 and 13 .
  • vertical pixel data items along the time axis T represent data items of different fields.
  • pixel data 49 belongs to first field data
  • 43 and 48 belong to second field data
  • 41 , 46 and 47 belong to third field data
  • 42 and 45 belong to fourth field data
  • 44 belongs to fifth field data.
  • the currently processed central pixel 41 belongs to the third field data.
  • coefficient masks used for convolution are determined according to Equation 6.
  • Equation 3 The result of 3D convolution of Equation 3 is identical to the result of 3D convolution of the spatial domain coefficient mask h hv and the time domain coefficient mask h t1t2 in Equation 6. Since h t1t2 extends over five fields along the time axis, coefficient masks for the respective fields of Equation 3 are defined in the five fields.
  • data items which are 2D-convoluted with the first through fifth coefficient masks h 1 , h 2 , h 3 , h 4 and h 5 , are CVBS data windows BBFID, BEFID, PRFID, AFFID and AAFID of five fields having an interval of one field. That is, the currently processed field data window PRFID and h 3 (i, j, t) is convoluted.
  • the C signal C(i, j, t) output from the 3D BPF 330 is represented as follows.
  • the 3D BPF 330 outputs the result of 3-dimensional convolution of the horizontal coefficient mask h h (i, j, t), the vertical coefficient mask h v (i, j, t), the time coefficient mask h t (i, j, t) and the corresponding data window CVBS(i, j, t) of the plurality of field data items stored in the memory 310 as the C signal C(i, j, t) of the input video signal.
  • the Y signal Y(i, j, t) output from the subtracter 340 is represented by Equation 4.
  • the horizontal coefficient mask h h (i, j, t), the vertical coefficient mask h v (i, j, t) and the time coefficient mask h t (i, j, t) have a 1-dimensional form as represented by Equation 8.
  • coefficients Nh, Nv and Nt can be set to values when normalization is carried out such that the sum of the absolute values of Nh, Nv and Nt and the absolute values of the 2D weight coefficients and 3D weight coefficients is equal to 1.
  • 3D spatial masks obtained from the three 1D coefficient masks h h , h v and h t of Equation 7 are respectively composed of h 1 (i, j, t), h 2 (i, j, t) and h 3 (i, j, t) represented by Equation 9.
  • the data CVBS(i, j, t) used for the convolution of the 3D BPF 330 includes three data windows corresponding to the filter masks h 1 , h 2 and h 3 . That is, CVBS data of a 3 ⁇ 5 matrix corresponding to each of the filer masks of Equation 9 is used for the convolution of Equation 3.
  • h 2 (i, j, t) is the coefficient mask for the currently processed field
  • h 1 (i, j, t) is the coefficient mask for the previous field
  • h 3 (i, j, t) is the coefficient mask for the next field. That is, data items BEFID, PRFID and AFFID of three fields having an interval of one field are used for the 3D convolution of Equation 3.
  • Wu, Wd, Wl and Wr are weights applied to the pixels ( ⁇ U of FIG. 4 ) having the opposite phase to the central pixel (i, j) in the currently processed field, and they have values of ⁇ 0.5 to 0 in response to the spatial local characteristics of an image in each direction. Accordingly, diagonal weights WuWl, WdWl, WuWr and WdWr respectively have values of 0 to 0.25 in response to the local characteristics of the image.
  • the 3D weight coefficients Wb 1 , Wb 2 , Wf 1 and Wf 2 /Wb and Wf are used for reflecting a motion variation on the time axis.
  • the 3D weight coefficients Wb 1 , Wb 2 , Wf 1 and Wf 2 /Wb and Wf have values of ⁇ 0.5 to 0 for pixels having the opposite phase to the central pixel (i, j) and values of 0 to 0.5 for pixels having the same phase as the central pixel (i, j).
  • the 3D weight coefficients Wb 1 , Wb 2 , Wf 1 and Wf 2 /Wb and Wf can use any value that represents correlation on the time axis.
  • Nh is applied to the currently processed central pixel (i, j, t)
  • Wl is applied to a pixel (i, j ⁇ 2, t) located two pixels to the left of the central pixel (i, j, t)
  • Wr is applied to a pixel (i, j+2, t) located two pixels to the right of the central pixel (i, j, t).
  • FIG. 9 NTSC system and FIG.
  • Nv is applied to the currently processed central pixel (i, j, t) and Wu is applied to a pixel (i ⁇ 1, j, t) one horizontal scan line above the central pixel (i, j, t) in the NTSC system and to a pixel (i ⁇ 2, j, t) two horizontal scan lines above the central pixel (i, j, t) in the PAL system.
  • Wd is applied to a pixel (i+1, j, t) one horizontal scan line under the central pixel (i, j, t) in the NTSC system and to a pixel (i+2, j, t) two horizontal scan lines under the central pixel (i, j, t) in the PAL system.
  • Nt 1 and Nt 2 are applied to the central pixel (i, j, t) of the currently processed field.
  • Wb 1 and Wb 2 are respectively applied to a pixel (i, j, t ⁇ 1) and a pixel (i ⁇ 1, j, t ⁇ 1) in data of the field before the current field.
  • the pixel (i, j, t ⁇ 1) and a pixel (i ⁇ 1, j, t ⁇ 1) are located in positions corresponding to the position of the central pixel (i, j, t) of the current field and respectively have the same and the opposite phase to the currently processed central pixel (i, j, t).
  • Wf 1 and Wf 2 are respectively applied to a pixel (i ⁇ 1, j, t+1) and a pixel (i, j, t+1) in data of the field after the current field.
  • the pixel (i ⁇ 1, j, t+1) and the pixel (i, j, t+1) are located in positions corresponding to the position of the central pixel (i, j, t) of the current field and respectively have the same and the opposite phase to the currently processed central pixel (i, j, t).
  • Nt is applied to the central pixel (i, j, t) of the currently processed field
  • Wb is applied to a pixel (i, j, t ⁇ 1) in data of the field before the current field, which is located in the position corresponding to the position of the central pixel (i, j, t) of the current field
  • Wf is applied to a pixel (i ⁇ 1, j, t+1) in data of the field after the current field, which is located in the position corresponding to the position of the central pixel (i, j, t) of the current field.
  • the 2D weight coefficients Wu, Wd, Wl and Wr and 3D weight coefficients Wb 1 , Wb 2 , Wf 1 and Wf 2 used in Equation 5 are determined such that they satisfy Equation 10.
  • the 2D weight coefficients Wu, Wd, Wl and Wr and 3D weight coefficients Wb and Wf used in Equation 8 are determined such that they satisfy Equation 11.
  • VARv denotes a vertical variation in the spatial domain
  • VARh represents a horizontal variation in the spatial domain
  • VARt, VARt 1 and VARt 2 represent temporal variations in the time domain.
  • VARu denotes an upward variation in the spatial domain
  • VARd is a downward variation in the spatial domain
  • VARl denotes a leftward variation in the spatial domain
  • VARr is a rightward variation in the spatial domain
  • VARb, VARb 1 and VARb 2 denote previous variations in the time domain
  • VARf, VARf 1 , VARf 2 are subsequent variations in the time domain.
  • Equation 12 To represent Equations 10 and 11 in more detail, values of Equation 12 are proposed.
  • Difh dl+dr+dh
  • dr is the absolute value of the difference between the pixel data (i, j) and pixel data (i, j+4)
  • dh is the absolute value of the difference between the pixel data (i, j ⁇ 2) and pixel data (i, j+2).
  • du is the absolute value of the difference between the pixel data (i, j) and pixel data (i ⁇ 2, j)
  • dd is the absolute value of the difference between the pixel data (i, j) and pixel data (i+2, j)
  • dv is the absolute value of the difference between pixel data (i ⁇ 1, j) and pixel data (i+1, j) in the NTSC system.
  • du is the absolute value of the difference between the pixel data (i, j) and pixel data (i ⁇ 4, j)
  • dd is the absolute value of the difference between the pixel data (i, j) and pixel data (i+4, j)
  • dv is the absolute value of the difference between pixel data (i ⁇ 2, j) and pixel data (i+2, j).
  • Dift denotes a variation on the time axis
  • Diff and Difb represent variations to subsequent and previous field directions.
  • Dift, Diff and Difb are calculated by combining pixels having the same phase as the current pixel on the time axis.
  • the weights Wu, Wd, Wl, Wr, Wf and Wb can be determined such that they represent horizontal/vertical/temporal/leftward/rightward/upward/downward variations using various combinations of pixel data items having the same phase.
  • Wu may be proportional to the vertical and upward correlation for the central pixel (i, j).
  • Wd may be proportional to the vertical and downward correlation for the central pixel (i, j).
  • Wl may be proportional to the horizontal and leftward correlation for the central pixel (i, j).
  • Wr may be proportional to the horizontal and rightward correlation for the central pixel (i, j).
  • Wb may be proportional to the correlation of the currently processed central pixel and data of the field before the current pixel, or the difference between pixels in data of the field/frame before the current field, which have the same phase as the currently processed central pixel, as shown in FIG. 16 .
  • Wf may be proportional to the correlation of the currently processed central pixel and data of the field after the current pixel, or the difference between pixels in data of the field/frame after the current field, which have the same phase as the currently processed central pixel,
  • Wb 1 may be proportional to the correlation of the currently processed central pixel and data of the field before the current pixel, or the difference between the central pixel and a pixel in data of the field/frame before the current field, which have the same phase as the currently processed central pixel (i, j, t) in a first direction, as shown in FIG. 10 .
  • Wf 1 may be proportional to the correlation of the currently processed central pixel and data of the field after the current pixel, or the difference between the central pixel and a pixel in data of the field/frame after the current field, which have the same phase as the currently processed central pixel (i, j, t) in a second direction.
  • Wb 2 may be proportional to the correlation of the currently processed central pixel and data of the field before the current pixel, or the difference between the central pixel and a pixel in data of the field/frame before the current field, which have the opposite phase to the currently processed central pixel (i, j, t) in a third direction, as shown in FIG. 11 .
  • Wf 2 may be proportional to the correlation of the currently processed central pixel and data of the field after the current pixel, or the difference between the central pixel and a pixel in data of the field/frame after the current field, which have the opposite phase to the currently processed central pixel (i, j, t) in a fourth direction, as shown in FIG. 11 .
  • the 3D BPF 330 performs filtering through convolution of Equation 3 using the 2D weight coefficients Wu, Wd, Wl and Wr and the third weight coefficients Wb 1 , Wb 2 , Wf 1 and Wf 2 /Wb and Wf, which are determined as above.
  • FIG. 19 is a block diagram of the 3D BPF 330 of FIG. 3 .
  • the 3D BPF 330 includes a first filter 331 , a second filter 332 , a third filter 333 , a fourth filter 334 , a fifth filter 335 , and a combiner 336 .
  • the 3D BPF 330 may perform NTSC or PAL system operation in response to the control signal NPC.
  • the 3D BPF 330 may perform NTSC system operation using the corresponding data windows BBFID, BEFID, PRFID, AFFID and MFID of five fields.
  • the 3D BPF 330 may perform PAL system operation using the corresponding data windows BEFID, PRFID and AAFID of three fields.
  • the first filter 331 , the second filter 332 , the third filter 333 , the fourth filter 334 , the fifth filter 335 and the combiner 336 may all be operated.
  • the first filter 331 performs convolution of the corresponding data window of the first field data BBFID of the input video signal and the first filer mask h 1 (i, j, t) determined by Equation 6.
  • the data window of the first field data BBFID exists in field data that is two fields before the third field data PRFID currently processed.
  • the second filter 332 performs convolution of the corresponding data window of the second field data BEFID of the input video signal and the second filer mask h 2 (i, j, t) determined by Equation 6.
  • the data window of the second field data BEFID exists in field data that is one field before the third field data PRFID currently processed.
  • the third filter 333 performs convolution of the corresponding data window of the third field data PRFID of the input video signal and the third filer mask h 3 (i, j, t) determined by Equation 6.
  • the data window of the third field data PRFID exists in the currently processed field data.
  • the fourth filter 334 performs convolution of the corresponding data window of the fourth field data AFFID of the input video signal and the fourth filer mask h 4 (i, j, t) determined by Equation 6.
  • the data window of the fourth field data AFFID exists in field data that is one field after the third field data PRFID currently processed.
  • the fifth filter 335 performs convolution of the corresponding data window of the fifth field data AAFID of the input video signal and the fifth filer mask h 5 (i, j, t) determined by Equation 6.
  • the data window of the fourth field data AAFID exists in field data that is two fields after the third field data PRFID currently processed.
  • the second filter 332 In the PAL system, the second filter 332 , the third filter 333 , the fourth filter 334 and the combiner 336 may be operated.
  • the second filter 332 performs convolution of the filter mask h 1 (i, j, t) determined by Equation 9 and the corresponding data window of the field data BEFID of the input video signal. That is, the second filter 332 uses data items in vertical and horizontal directions, which have the opposite C component phase to the central pixel, and data items in diagonal directions, which have the same C component phase as the central pixel, among digital data of three continuous horizontal scan lines of the field data BEFID that is one field before the current field, as represented by Equation 9.
  • the third filter 333 performs convolution of the filter mask h 2 (i, j, t) determined by Equation 9 and the corresponding data window of the field data PRFID of the input video signal.
  • the third filter 333 uses data items in vertical and horizontal directions, which have the opposite C component phase to the central pixel, and data items in diagonal directions, which have the same C component phase as the central pixel, among digital data of three continuous horizontal scan lines of the current field data PRFID, as represented by Equation 9.
  • the fourth filter 334 performs convolution of the filter mask h 3 (i, j, t) determined by Equation 9 and the corresponding data window of the field data AFFID of the input video signal.
  • the fourth filter 334 uses data items in vertical and horizontal directions, which have the opposite C component phase to the central pixel, and data items in diagonal directions, which have the same C component phase as the central pixel, among digital data of three continuous horizontal scan lines of the field data AFFID that is one field after the current field, as represented by Equation 9.
  • the combiner 336 appropriately combines the convolution results of the first to fifth filters 331 to 335 , i.e., performs addition or subtraction on the convolution results, to output the result as the C signal.
  • the combiner 336 appropriately combines the convolution results of the second, third and fourth filter 332 , 333 and 334 , i.e., performs addition or subtraction on the convolution results, to output the result as the C signal.
  • the 3D BPF 330 continuously and adaptively performs comb filtering, 1D band pass filtering and 2D/3D band pass filtering in response to spatio-temporal local characteristics of the input video signal.
  • the C component When correlation is low in all directions, the C component is separated in the sharpest form, and thus cross-color artifacts are minimized and the high frequency component of the Y signal is extracted in a distinct manner.
  • the high frequency component 16 of the C signal, included in the Y component 15 is extracted distinctly like a desired sharp form 17 , to thereby remove artifacts.
  • the 3D BPF 330 may serve as any of the three types of filters, e.g., a local vertical comb filter, horizontal band pass filter, and/or a 2D/3D band pass filter, as illustrated schematically in FIG. 22 .
  • the adaptive 3D BPF 330 performs Y/C separation according to local comb filtering/1D band pass filtering when the edge direction is fixed vertically/horizontally based on fields in response to spatio-temporal local characteristics of an image using spatio-temporal filters.
  • the 3D BPF 330 performs 2D/3D band pass filtering in all directions.
  • the 3D BPF continuously performs comb filtering, 1D band pass filtering and 2D band pass filtering in response to the spatio-temporal local characteristics of the image.
  • the video signal processing apparatus continuously and adaptively performs local comb filtering/1D band pass filtering or 2D band pass filtering in all directions in response to spatio-temporal local characteristics of an image in the NTSC/PAL system.
  • the video signal processing apparatus of the present invention may detect the edge of the image accurately more than a conventional apparatus that discretely selects filtering methods, and produce better and more stable spectrum characteristics than a frame-based Y/C separation filter. Accordingly, artifacts such as cross-luma and cross-color may be reduced or removed when the video signal processing apparatus of the present invention is applied to a display system, resulting in improved image quality.

Abstract

A digital video signal processing apparatus and method for field-based adaptive Y/C separation includes an adaptive 3D BPF performing Y/C separation according to local comb filtering/1D band pass filtering when the edge direction is fixed vertically/horizontally in response to spatio-temporal local characteristics of an image using spatio-temporal filters. When the edge direction is not fixed horizontally or vertically, the 3D BPF performs 2D/3D band pass filtering in all directions. The 3D BPF adaptively and continuously carries out comb filtering, 1D band pass filtering and 2D/3D band pass filtering in response to the spatio-temporal local characteristics of the image.

Description

BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention relates to a digital video signal processing apparatus. More particularly, the present invention relates to a digital video signal processing apparatus and method for field-based adaptive spatio-temporal, i.e., 3-dimensional (3D), Y/C separation in NTSC/PAL (National Television System Committee/Phase Alternation by Line) systems.
2. Description of the Related Art
The display of an NTSC/PAL broadcasting system includes a device for processing a CVBS (Composite Video Blanking Sync) signal, which is a composite of a Y (luminance) signal and a C (chrominance) signal. The C signal is quadrature-amplitude-modulated with a subcarrier frequency fsc. Thus, characteristics of the C signal are determined by its frequency and phase. A digital video signal processing apparatus at a receiving stage separates Y and C signals with reference to the characteristic of the C signal and displays an image based on the separated signals.
FIG. 1 illustrates a block diagram of a conventional video signal processing apparatus 100. Referring to FIG. 1, the video signal processing apparatus 100 includes a comb filter 110, a one-dimensional bandpass filter (1D-BPF) 120, a weight determination unit 130, a combiner 140 and a subtracter 150. The comb filter 110 one-dimensionally band-pass-filters an input video signal in the vertical direction. The 1D-BPF 120 one-dimensionally band-pass-filters the input video signal in the horizontal direction. The weight determination unit 130 determines weights of the output of the comb filter 110 and the output of the 1D-BPF 120 with reference to vertical/horizontal correlation and the phase of a C signal. The combiner 140 combines the output signals of the comb filter 110 and 1D-BPF 120 using the weights to generate a C signal. The subtracter 150 subtracts the C signal from the input CVBS signal to generate a Y signal.
FIG. 2 illustrates a block diagram of another conventional video signal processing apparatus 200. Referring to FIG. 2, the video signal processing apparatus 200 includes a 2D-BPF 210, a subtracter 220 and a post-processor 230. The 2D-BPF 210 performs a two-dimensional convolution to extract a modulated C signal. The extracted C signal and a Y signal generated by the subtracter 220 are processed by the post-processor 230. When the 2D-BPF 210 carries out Y/C separation incorrectly, the post-processor 230 compensates the Y/C signal to generate a compensated Y/C signal.
In conventional Y/C separation techniques, when the edge of a detected image has high vertical correlation, Y/C separation is performed based on comb filtering and when the edge of a detected image has high horizontal correlation, Y/C separation is performed based on 1D bandpass filtering. As described above, when selecting one of these filtering methods according to a conventional Y/C separation technique, system performance largely depends on a threshold used for edge detection. That is, Y/C separation can be executed incorrectly or unstably when the filtering method is wrongly selected due to inaccurate edge detection. Conventional techniques that do not select one of the filtering methods but combine the results of the filtering operations can solve this problem to some extent. However, these techniques are based on horizontal or vertical one-dimensional filtering, and thus, artifacts may remain in the generated signal due to the inconstant edge directions.
In other words, when the direction in which the edge of an image extends is not uniform, cross-luma, which occurs when a C component exists in the separated Y signal, resulting in dotted artifacts, and cross-color, which occurs when a Y component exists in the separated C signal, resulting in a rainbow pattern artifact, can appear on a displayed image when Y/C separation is not properly performed by comb filtering or 1D bandpass filtering selected discretely.
To improve spatial filtering, spatio-temporal filtering is used by conventional video signal processing apparatuses. In this case, when processing a current pixel, the correlations of pixel data of the previous field and pixel data of the next field with the current pixel or the correlations of pixel data of the previous frame and pixel data of the next frame with the current pixel is considered. The spatio-temporal filtering method requires a memory for storing the pixel data of the previous and next fields or frames. Although the spatio-temporal filter is more expensive than the spatial filter, the spatio-temporal filtering technique is frequently used when images with high picture quality are required.
However, conventional Y/C separation techniques, which discontinuously select the spatial filter and the spatio-temporal filter in response to inter-frame/intra-frame correlation of the CVBS signal, cause artifacts such as cross-luma and cross-color when there is an error in the correlation measurement result. In addition, the conventional techniques have limitations in two-dimensional spatial Y/C separation. Furthermore, conventional Y/C separation techniques using intra-frame correlation have an advantage of using pixel data from the same spatial location. However, these techniques use data before and after one frame temporally distant by more than one field, resulting in deterioration of the spectrum characteristics of filters.
SUMMARY OF THE INVENTION
The present invention is therefore directed to a digital video signal processing apparatus and method, which substantially overcome one or more of the problems due to the limitations and disadvantages of the related art.
It is a feature of an embodiment of the present invention to provide a digital video signal processing apparatus used for both NTSC and PAL systems.
It is a feature of an embodiment of the present invention to adaptively and continuously perform field/frame-based Y/C separation in multiple directions according to temporal and spatial characteristics of an input video signal.
It is yet another feature of an embodiment of the present invention to use a spatio-temporal filter having spectral characteristics suitable for Y/C separation of the input video signal.
At least one of the above and other features and advantages of the present invention may be realized by providing a video signal processing method including generating 2D weight coefficients and 3D weight coefficients using data of a plurality of fields of an input video signal and generating a C signal of the input video signal by carrying out convolution of a horizontal coefficient mask, a vertical coefficient mask and a time coefficient mask, which are composed based on the weight coefficients, and corresponding data windows of the plurality of field data, wherein generating the C signal includes local vertical comb filtering, horizontal band pass filtering and 2D band pass filtering according to spatio-temporal local characteristics of the input video signal.
Local vertical comb filtering may be performed when local characteristics indicate high vertical correlation. Local horizontal band pass filtering may be performed when local characteristics indicate high horizontal correlation. 2D or 3D band pass filtering may be carried out when local characteristics indicate high or low correlation in multiple directions using at least three field data having an interval of one field.
The 2D weight coefficients may include a first coefficient proportional to vertical and upward correlations with the current pixel, a second coefficient proportional to vertical and downward correlations with the current pixel, a third coefficient proportional to horizontal and leftward correlations with the current pixel, and a fourth coefficient proportional to horizontal and rightward correlations with the current pixel.
The generated C signal may be subtracted from the input video signal and outputting as a Y signal.
Generating the C signal may include performing convolution of a corresponding data window of a first field data of the input video signal and a first filter mask, performing convolution of a corresponding data window of a second field data of the input video signal and a second filter mask, performing convolution of a corresponding data window of a third field data of the input video signal and a third filter mask, and combining the convolution results to output the combined result as the C signal, wherein the first, second and third filter masks are generated from 3D convolution of the horizontal coefficient mask, the vertical coefficient mask and the time coefficient mask.
The method may be applied to a PAL system or an NTSC system.
The first field data, the second field data and the third field data may have an interval of one field. The horizontal coefficient mask, the vertical coefficient mask and the time coefficient mask may each be one-dimensional. The 3D weight coefficients may include a first coefficient in proportion to a correlation between the currently processed pixel data and data of the field before the current field, and a second coefficient in proportion to a correlation between the currently processed pixel data and data of the field after the current field.
Generating the C signal may further include performing convolution of a corresponding data window of a fourth field data of the input video signal and a fourth filter mask, performing convolution of a corresponding data window of a fifth field data of the input video signal and a fifth filter mask, combining the five convolution results to be output the combined result as the generated C signal, wherein the time coefficient mask includes first and second coefficient masks, and the first to fifth filter masks are generated from 3D convolution of the horizontal coefficient mask, the vertical coefficient mask, the first coefficient mask and the second coefficient mask.
The 3D weight coefficients used for the first coefficient mask may include a first coefficient in proportion to the correlation in a first direction between the currently processed pixel data and data of the field before the current field, and a second coefficient in proportion to the correlation in the second direction between the currently processed pixel data and data of the field after the current field. The 3D weight coefficients used for the second coefficient mask may include a third coefficient in proportion to the correlation in a third direction between the currently processed pixel data and data of the field before the current field, and a fourth coefficient in proportion to the correlation in a fourth direction between the currently processed pixel data and data of the field after the current field.
At least one of the above and other features and advantages of the present invention may be realized by providing a video signal processing apparatus, including a weight determination unit to generate 2D weight coefficients and 3D weight coefficients using data of a plurality of fields of an input video signal and a filter to convolute a horizontal coefficient mask, a vertical coefficient mask and a time coefficient mask, the masks respectively composed of the weight coefficients, with corresponding data windows of the data of the plurality of fields to generate a C signal of the input video signal, wherein the filter adaptively performs local vertical comb filtering, horizontal band pass filtering and 2D band pass filtering in response to spatio-temporal local characteristics of the input video signal.
BRIEF DESCRIPTION OF THE DRAWINGS
The above and other features and advantages of the present invention will become more apparent to those of ordinary skill in the art by describing in detail exemplary embodiments thereof with reference to the attached drawings in which:
FIG. 1 illustrates a block diagram of a conventional video signal processing apparatus;
FIG. 2 illustrates a block diagram of another conventional video signal processing apparatus;
FIG. 3 illustrates a block diagram of a video signal processing apparatus according to an embodiment of the present invention;
FIG. 4 illustrates chrominance signal phases of an NTSC system;
FIG. 5 illustrates chrominance signal phases of a PAL system;
FIG. 6 illustrates spatio-temporal chrominance signal phases of the NTSC system;
FIG. 7 illustrates spatio-temporal chrominance signal phases of the PAL system;
FIG. 8 illustrates a diagram for explaining a horizontal filter coefficient of the NTSC system;
FIG. 9 illustrates a diagram for explaining a vertical filter coefficient of the NTSC system;
FIG. 10 is a diagram for explaining a first temporal filter coefficient of the NTSC system;
FIG. 11 illustrates a diagram for explaining a second temporal filter coefficient of the NTSC system;
FIG. 12 is a diagram for explaining a vertical/horizontal filter coefficient of the NTSC system;
FIG. 13 illustrates a diagram for explaining a vertical/temporal filter coefficient of the NTSC system;
FIG. 14 illustrates a diagram for explaining a horizontal filter coefficient of the PAL system;
FIG. 15 illustrates a diagram for explaining a vertical filter coefficient of the PAL system;
FIG. 16 illustrates a diagram for explaining a temporal filter coefficient of the PAL system;
FIG. 17 illustrates a vertical variation in pixels;
FIG. 18 illustrates a horizontal variation in pixels;
FIG. 19 illustrates a block diagram of the 3D BPF of FIG. 3;
FIG. 20 illustrates spectrum characteristics representing Y and C components with respect to horizontal and vertical frequencies of a video signal;
FIG. 21 illustrates spectrum characteristics representing Y and C components with respect to temporal and vertical frequencies of a video signal; and
FIG. 22 illustrates a block diagram of the 3D BPF of FIG. 3.
DETAILED DESCRIPTION OF THE INVENTION
Korean Patent Application No. 10-2005-0003178, filed on Jan. 13, 2005, in the Korean Intellectual Property Office, and entitled “Digital Video Signal Processing Apparatus and Method for Adaptive and Temporal and Spatial Y/C Separation Based on Field Period,” is incorporated by reference herein in its entirety.
The present invention will now be described more fully with reference to the accompanying drawings, in which exemplary embodiments of the invention are shown. The invention may, however, be embodied in many different forms and should not be construed as being limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the concept of the invention to those skilled in the art. Throughout the drawings, like reference numerals refer to like elements.
FIG. 3 illustrates a block diagram of a video signal processing apparatus 300 according to an embodiment of the present invention. Referring to FIG. 3, the video signal processing apparatus 300 may include a memory 310, a weight determination unit 320, a 3D BPF 330, and a subtracter 340. The video signal processing apparatus 300 can be used for both NTSC systems and PAL systems. The video signal processing apparatus 300 receives a digital CVBS signal as an input video signal and separates the input video signal into a Y signal and a C signal. The input video signal can be a digital signal obtained by sampling the active video region of an analog CVBS signal at a predetermined frequency e.g., 4fsc, where fsc is a subcarrier frequency.
An input video signal CVBS(t) of the NTSC system can be represented as follows.
CVBS(t)=Y+U*sin2πf SC t+V*cos2πf SC t  [Equation 1]
where U and V are C components, fsc is a subcarrier frequency, and t is time. Thus, in the NTSC system, pixel signals sampled at 4fsc have chrominance signal phases as shown in FIG. 4. That is, the pixel signals are repeated in the form of Y+U, Y+V, Y−U, Y−V, . . . in each horizontal line. FIG. 4 shows only the phases of chrominance signal components. In the NTSC system, the chrominance signal phases are shifted by 180° in the vertical direction in neighboring horizontal scan lines, as shown in FIG. 4.
The input video signal CVBS(t) of the NTSC system may be represented spatio-temporally, as shown in FIG. 6. Referring to FIG. 6, the phase of the C component of the currently processed central pixel at (i, j, t) has a particular phase, e.g., +U. Pixel signals spatio-temporally disposed in a first diagonal direction from the central pixel, i.e., a pixel at (i−1, j, t+1) of the next field and a pixel at (i, j, t−1) of the previous field, have the same phase as the central pixel. Pixel signals spatio-temporally disposed in a second diagonal direction from the central pixel, i.e., a pixel at (i, j, t+1) of the next field and a pixel at (i−1, j, t−1) of the previous field, have a phase opposite the phase, e.g., −U, of the C component of the central pixel at (i, j, t). In the NTSC/PAL system, one frame is composed of two fields, i.e., an odd field and an even field
An input video signal CVBS(t) of a PAL system can be represented as follows.
CVBS(t)=Y+U*sin2πf SC t±V*cos2πf SC t  [Equation 2]
where U and V are C components, fsc is a subcarrier frequency, and t is time. In Equation 2, the sign of the V component alternates between positive and negative every field. Thus, pixel signals in the PAL system have chrominance signal phases as shown in FIG. 5. That is, the pixel signals are repeated in the form of Y+U, Y+V, Y−U, Y−V, . . . or Y+U, Y−V, Y−U, Y+V, . . . . in each horizontal line, and the chrominance signal phases are shifted by 180° in the vertical direction every other horizontal scan line.
The input video signal CVBS(t) of the PAL system may be represented spatio-temporally, as shown in FIG. 7. Referring to FIG. 7, the phase of the C signal component of the currently processed central pixel at (i, j, t) has a particular phase, e.g., +U. Pixel signals spatio-temporally disposed in a diagonal direction from the current pixel, i.e., a pixel at (i−1, j, t+1) of the next field and a pixel at (i, j, t−1) of the previous field, have a phase opposite the phase of the current pixel signal (i, j, t) of the current field. A pixel at (i+1, j, t+2) located one frame (two fields) after the central pixel and a pixel at (i, j, t−2) located one frame (two fields) before the central pixel also have the phase opposite the phase of the current pixel at (i, j, t) of the current field. Pixels disposed vertically from the central pixel, i.e., pixels at (i−2, j, t) and (i+2, j, t) of the same field, also have the phase opposite the phase of the current pixel at (i, j, t) of the current field.
The Y and C signals (U and V signals) separated by the video signal processing apparatus 300 are converted to a format required for an external circuit and then stored or transmitted to a display device. For example, the Y and C signals can be interpolated with three color signals, e.g., red, green and blue color signals, to be displayed on a liquid crystal display (LCD).
Rather than discretely carrying out spatio-temporal comb filtering or 1D band pass filtering, the video signal processing apparatus 300 according to an embodiment of the present invention adaptively operates in response to spatio-temporal local characteristics of the input video signal CVBS. For this, the memory 310 stores digital data corresponding to a plurality of fields of the input video signal CVBS. The memory 310 includes a plurality of field memories 311, 312, 313 and 314 which each store data corresponding to one field. While FIG. 3 shows four field memories 311, 312, 313 and 314, more memories can be added to the digital video signal processing apparatus 300 if required for filtering.
The data corresponding to the plurality of fields, stored in the memory 310, and currently input video data are output to the weight determination unit 320. The weight determination unit 320 generates 2D weight coefficients Wu, Wd, Wl and Wr, which will be used for filtering in the 3D BPF 330, using the plurality of field data. Furthermore, the weight determination unit 320 generates 3D weight coefficients Wb1, Wb2, Wf1 and Wf2/Wb and Wf in response to a network parameter control (NPC) signal. For example, when the NPC signal has a logic high level, 3D weight coefficients Wb1, Wb2, Wf1 and Wf2 may be generated for the NTSC system having the chrominance signal phases shown in FIG. 4. When the control signal NPC has a logic low level, 3D weight coefficients Wb and Wf may be generated for the PAL system having the chrominance signal phases shown in FIG. 5. While the 2D weight coefficients Wu, Wd, Wl and Wr and the 3D weight coefficients Wb1, Wb2, Wf1 and Wf2/Wb and Wf may be generated for both the NTSC and PAL systems, data used to generate the coefficients for the NTSC system may be different from data used to generate the coefficients for the PAL system.
The 3D BPF 330 generates the C signal of the input video signal using the 2D weight coefficients Wu, Wd, Wl and Wr and the 3D weight coefficients Wb1, Wb2, Wf1 and Wf2/Wb and Wf. The 3D BPF 330 uses corresponding data windows of three fields or corresponding data windows of five fields, which will be convoluted with filter masks in response to the control signal NPC. For example, when the control signal NPC has a logic high level, the 3D BPF 330 may be operated for the NTSC system and use five data windows BBFID (before before field), BEFID, (before field), PRFID (present field), AFFID (after field) and AAFID (after after field). When the control signal NPC has a logic low level, the 3D BPF 330 may be operated for the PAL system and use three data windows BEFID, PRFID and AFFID.
The subtracter 340 subtracts the C signal (U or V signal), output from the 3D BPF 330, from the video signal CVBS of the currently processed pixel to obtain the Y signal. For example, when the U signal output as the C signal is subtracted from the video signal of the current pixel Y+U, the Y signal is obtained.
In the NTSC/PAL system, the 3D BPF 330 continuously and adaptively performs local vertical comb filtering, local horizontal band pass filtering and 2D/3D band pass filtering, in response to spatio-temporal local characteristics of the input video signal. That is, the 3D BPF 330 performs filtering using filter masks for convolution of field data items having an interval of one field. The 3D BPF 330 performs local vertical comb filtering if the local characteristics of the central pixel to be filtered show high vertical correlation, and executes local horizontal band pass filtering if the local characteristics show high horizontal correlation. Furthermore, the 3D BPF 330 performs 2D/3D band pass filtering if the local characteristics show high or low correlation in multiple directions.
The 3D BPF 330 performs filtering using the filter masks, that is, a horizontal coefficient mask, a vertical coefficient mask and a time coefficient mask, which are composed based on the 2D weight coefficients Wu, Wd, Wl and Wr and the 3D weight coefficients Wb1, Wb2, Wf1 and Wf2/Wb and Wf.
In the NTSC system, the time coefficient mask is composed of first and second time coefficient masks, and the 2D weight coefficients Wu, Wd, Wl and Wr and the 3D weight coefficients Wb1, Wb2, Wf1 and Wf2 are used for filtering. If the horizontal coefficient mask is hh(i, j, t), the vertical coefficient mask is hv(i, j, t), the first time coefficient mask is ht1(i, j, t) and the second time coefficient mask is ht2(i, j, t), then the C signal C(i, j, t) output from the 3D BPF 330 is represented as follows.
C(i,j,t)=h(i,j,t)***CVBS(i,j,t) h(i,j,t)=h h(i,j,t)***h v(i,j,t)***h t1(i,j,t)***h t2(i,j,t)***CVBS(i,j,t)   [Equation 3]
That is, the 3D BPF 330 outputs the result of 3-dimensional convolution of the horizontal coefficient mask hh(i, j, t), the vertical coefficient mask hv(i, j, t), the first time coefficient mask ht1(i, j, t), the second time coefficient mask ht2(i, j, t) and the corresponding data window CVBS(i, j, t) of the plurality of field data items stored in the memory 310 as the C signal C(i, j, t) of the input video signal. The Y signal Y(i, j, t) output from the subtracter 340 is represented as the result of subtracting the C signal C(i, j, t) from the CVBS data CVBS(i, j, t) of the currently processed central pixel, which is represented as follows.
Y(i,j,t)=CVBS(i,j,t)−C(i,j,t)  [Equation 4]
The horizontal coefficient mask hh(i, j, t), the vertical coefficient mask hv(i, j, t), the first time coefficient mask ht1(i, j, t) and the second time coefficient mask ht2(i, j, t) have a 1-dimensional form as represented by Equation 5.
h h = [ Wl Nh Wr ] h t 1 = [ Wb 1 Nt 1 Wf 1 ] h t 2 = [ Wb 2 Nt 2 Wf 2 ] h v = [ Wu Nv Wd ] [ Equation 5 ]
In Equation 5, coefficients Nh, Nv, Nt1 and Nt2 can be set to values when normalization is carried out such that the sum of the absolute values of Nh, Nv, Nt1 and Nt2 and the absolute values of the 2D weight coefficients and 3D weight coefficients are equal to 1.
3D convolution of the four 1D coefficient masks hh, hv, ht1 and ht2 can be represented by Equation 6.
h hv ( i , j , t ) = h h ( i , j , t ) ** * h v ( i , j , t ) = Wl · Wu 0 Wu 0 Wr · Wu Wl 0 Ns 0 Wr Wl · Wd 0 Wd 0 Wr · Wd h t 1 t 2 ( i , j , t ) = h t 1 ( i , j , t ) ** * h t 2 ( i , j , t ) = 0 0 Wb 1 · Wf 2 0 0 0 Wb 1 0 Wf 2 0 Wb 1 · Wb 2 0 Nt 0 Wf 1 · Wf 2 0 Wb 2 0 Wf 1 0 0 0 Wb 2 · Wf 1 0 0 [ Equation 6 ]
That is, C(i, j, t) of Equation 3 is arranged as C(i, j, t)=hhv(i, j, t)*** ht1t2(i, j, t)***CVBS(i, j, t).
Pixel data items corresponding to the horizontal/vertical coefficient mask hhv(i, j, t) and vertical/time coefficient mask ht1t2(i, j, t) of the NTSC system are shown in FIGS. 12 and 13. Data items corresponding to elements composed of nonzero coefficients in the vertical/horizontal coefficient mask hhv(i, j, t) in Equation 6, i.e., data items (32, 33, 34 and 35 of FIG. 12) in vertical and horizontal directions, which have the opposite C component phases to the central pixel (31 of FIG. 12), and pixel data items (36, 37, 38 and 39 of FIG. 12) in diagonal directions, which have the same C component phases as the central pixel, among digital data items of three continuous horizontal scan lines, are used for the aforementioned convolution. Furthermore, data items corresponding to elements composed of nonzero coefficients in the vertical/time coefficient mask ht1t2(i, j, t) in Equation 6, i.e., the central pixel (41 of FIG. 13), data items (44, 45, 46, 47, 48 and 49 of FIG. 13) having the opposite C component phases to the central pixel, and pixel data items (42 and 43 of FIG. 13) having the same C component phases as the central pixel, among digital data items of three continuous horizontal scan lines arranged over five fields, are used for the aforementioned convolution.
In FIG. 13, vertical pixel data items along the time axis T represent data items of different fields. For example, pixel data 49 belongs to first field data, 43 and 48 belong to second field data, 41, 46 and 47 belong to third field data, 42 and 45 belong to fourth field data, and 44 belongs to fifth field data. Here, the currently processed central pixel 41 belongs to the third field data. As described above, coefficient masks used for convolution are determined according to Equation 6.
The result of 3D convolution of Equation 3 is identical to the result of 3D convolution of the spatial domain coefficient mask hhv and the time domain coefficient mask ht1t2 in Equation 6. Since ht1t2 extends over five fields along the time axis, coefficient masks for the respective fields of Equation 3 are defined in the five fields. When the coefficient masks convoluted in these fields are respectively referred to as h1(i, j, t), h2(i, j, t), h3(i, j, t), h4(i, j, t) and h5(i, j, t), data items, which are 2D-convoluted with the first through fifth coefficient masks h1, h2, h3, h4 and h5, are CVBS data windows BBFID, BEFID, PRFID, AFFID and AAFID of five fields having an interval of one field. That is, the currently processed field data window PRFID and h3(i, j, t) is convoluted. In addition, previous field data windows BBFID and BEFID and next field data windows AFFID and AAFID are convoluted with the filter masks h1(i, j, t), h2(i, j, t), h3(i, j, t) and h5(i, j, t), respectively. In Equation 6, Ns=Nh*Nv.
In the PAL system, there is a single time coefficient mask and the 2D weight coefficients Wu, Wd, Wl and Wr and the 3D weight coefficients Wb and Wf are used. If the horizontal coefficient mask is hh(i, j, t), the vertical coefficient mask is hv(i, j, t), and the time coefficient mask is ht(i, j, t), then the C signal C(i, j, t) output from the 3D BPF 330 is represented as follows.
C(i,j,t)=h(i,j,t)***CVBS(i,j,t) h(i,j,t)=h h(i,j,t)***h v(i,j,t)***h t(i,j,t)  [Equation 7]
That is, the 3D BPF 330 outputs the result of 3-dimensional convolution of the horizontal coefficient mask hh(i, j, t), the vertical coefficient mask hv(i, j, t), the time coefficient mask ht(i, j, t) and the corresponding data window CVBS(i, j, t) of the plurality of field data items stored in the memory 310 as the C signal C(i, j, t) of the input video signal. The Y signal Y(i, j, t) output from the subtracter 340 is represented by Equation 4.
The horizontal coefficient mask hh(i, j, t), the vertical coefficient mask hv(i, j, t) and the time coefficient mask ht(i, j, t) have a 1-dimensional form as represented by Equation 8.
h h ( i , j , t ) = [ Wl Nh Wr ] h t ( i , j , t ) = [ Wb Nt Wf ] h v ( i , j , t ) = [ Wu Nv Wd ] [ Equation 8 ]
In Equation 8, coefficients Nh, Nv and Nt can be set to values when normalization is carried out such that the sum of the absolute values of Nh, Nv and Nt and the absolute values of the 2D weight coefficients and 3D weight coefficients is equal to 1.
3D spatial masks obtained from the three 1D coefficient masks hh, hv and ht of Equation 7 are respectively composed of h1(i, j, t), h2(i, j, t) and h3(i, j, t) represented by Equation 9.
h 1 ( i , j , t ) = [ Wb · Wl · Wu 0 Wb · Wu 0 Wb · Wr · Wu Wb · Wl 0 Ns · Wb 0 Wb · Wr Wb · Wl · Wd 0 Wb · Wd 0 Wb · Wr · Wd ] h 2 ( i , j , t ) = [ Nt · Wl · Wu 0 Nt · Wu 0 Nt · Wr · Wu Nt · Wl 0 Ns · Nt 0 Nt · Wr Nt · Wl · Wd 0 Nt · Wd 0 Nt · Wr · Wd ] h 3 ( i , j , t ) = [ Wf · Wl · Wu 0 Wf · Wu 0 Wf · Wr · Wu Wf · Wl 0 Ns · Wf 0 Wf · Wr Wf · Wl · Wd 0 Wf · Wd 0 Wf · Wr · Wd ] [ Equation 9 ]
Accordingly, the data CVBS(i, j, t) used for the convolution of the 3D BPF 330, represented by Equation 7, includes three data windows corresponding to the filter masks h1, h2 and h3. That is, CVBS data of a 3×5 matrix corresponding to each of the filer masks of Equation 9 is used for the convolution of Equation 3. In Equation 9, h2(i, j, t) is the coefficient mask for the currently processed field, h1(i, j, t) is the coefficient mask for the previous field, and h3(i, j, t) is the coefficient mask for the next field. That is, data items BEFID, PRFID and AFFID of three fields having an interval of one field are used for the 3D convolution of Equation 3.
In Equations 5 and 8, Wu, Wd, Wl and Wr are weights applied to the pixels (−U of FIG. 4) having the opposite phase to the central pixel (i, j) in the currently processed field, and they have values of −0.5 to 0 in response to the spatial local characteristics of an image in each direction. Accordingly, diagonal weights WuWl, WdWl, WuWr and WdWr respectively have values of 0 to 0.25 in response to the local characteristics of the image. In addition, the 3D weight coefficients Wb1, Wb2, Wf1 and Wf2/Wb and Wf are used for reflecting a motion variation on the time axis. The 3D weight coefficients Wb1, Wb2, Wf1 and Wf2/Wb and Wf have values of −0.5 to 0 for pixels having the opposite phase to the central pixel (i, j) and values of 0 to 0.5 for pixels having the same phase as the central pixel (i, j). The 3D weight coefficients Wb1, Wb2, Wf1 and Wf2/Wb and Wf can use any value that represents correlation on the time axis.
Specifically, referring to FIG. 8 (NTSC system) and FIG. 14 (PAL sys tem), Nh is applied to the currently processed central pixel (i, j, t), Wl is applied to a pixel (i, j−2, t) located two pixels to the left of the central pixel (i, j, t), and Wr is applied to a pixel (i, j+2, t) located two pixels to the right of the central pixel (i, j, t). Referring to FIG. 9 (NTSC system) and FIG. 15 (PAL system), Nv is applied to the currently processed central pixel (i, j, t) and Wu is applied to a pixel (i−1, j, t) one horizontal scan line above the central pixel (i, j, t) in the NTSC system and to a pixel (i−2, j, t) two horizontal scan lines above the central pixel (i, j, t) in the PAL system. Wd is applied to a pixel (i+1, j, t) one horizontal scan line under the central pixel (i, j, t) in the NTSC system and to a pixel (i+2, j, t) two horizontal scan lines under the central pixel (i, j, t) in the PAL system. Referring to FIGS. 10 and 11, in the NTSC system, Nt1 and Nt2 are applied to the central pixel (i, j, t) of the currently processed field. Wb1 and Wb2 are respectively applied to a pixel (i, j, t−1) and a pixel (i−1, j, t−1) in data of the field before the current field. The pixel (i, j, t−1) and a pixel (i−1, j, t−1) are located in positions corresponding to the position of the central pixel (i, j, t) of the current field and respectively have the same and the opposite phase to the currently processed central pixel (i, j, t). In addition, Wf1 and Wf2 are respectively applied to a pixel (i−1, j, t+1) and a pixel (i, j, t+1) in data of the field after the current field. Here, the pixel (i−1, j, t+1) and the pixel (i, j, t+1) are located in positions corresponding to the position of the central pixel (i, j, t) of the current field and respectively have the same and the opposite phase to the currently processed central pixel (i, j, t). Referring to FIG. 16, in the PAL system, Nt is applied to the central pixel (i, j, t) of the currently processed field, and Wb is applied to a pixel (i, j, t−1) in data of the field before the current field, which is located in the position corresponding to the position of the central pixel (i, j, t) of the current field. In addition, Wf is applied to a pixel (i−1, j, t+1) in data of the field after the current field, which is located in the position corresponding to the position of the central pixel (i, j, t) of the current field.
The 2D weight coefficients Wu, Wd, Wl and Wr and 3D weight coefficients Wb1, Wb2, Wf1 and Wf2 used in Equation 5 are determined such that they satisfy Equation 10. The 2D weight coefficients Wu, Wd, Wl and Wr and 3D weight coefficients Wb and Wf used in Equation 8 are determined such that they satisfy Equation 11.
Wu ( VARh & VARt ) ( or 1 VARv ) and Wu 1 VARu ( or VARd ) Wd ( VARh & VARt ) ( or 1 VARv ) and Wd 1 VARd ( or VARu ) Wl ( VARv & VARt ) ( or 1 VARh ) and Wl 1 VARl ( or VARr ) Wr ( VARv & VARt ) ( or 1 VARh ) and Wr 1 VARr ( or VARl ) Wb 1 ( VARv & VARh ) ( or 1 VARt 1 ) and Wb 1 1 VARb 1 ( or VARf 1 ) Wf 1 ( VARv & VARh ) ( or 1 VARt 1 ) and Wf 1 1 VARf 1 ( or VARb 1 ) Wb 2 ( VARv & VARh ) ( or 1 VARt 2 ) and Wb 2 1 VARb 2 ( or VARf 2 ) Wf 2 ( VARv & VARh ) ( or 1 VARt 2 ) and Wf 2 1 VARf 2 ( or VARb 2 ) [ Equation 10 ] Wu ( VARh & VARt ) ( or 1 VARv ) and Wu 1 VARu ( or VARd ) Wd ( VARh & VARt ) ( or 1 VARv ) and Wd 1 VARd ( or VARu ) Wl ( VARv & VARt ) ( or 1 VARh ) and Wl 1 VARl ( or VARr ) Wr ( VARv & VARt ) ( or 1 VARh ) and Wr 1 VARr ( or VARl ) Wb ( VARv & VARh ) ( or 1 VARt ) and Wb 1 VARb ( or VARf ) Wf ( VARv & VARh ) ( or 1 VARt ) and Wf 1 VARf ( or VARb ) [ Equation 11 ]
In Equations 10 and 11, VARv denotes a vertical variation in the spatial domain, VARh represents a horizontal variation in the spatial domain, and VARt, VARt1 and VARt2 represent temporal variations in the time domain. VARu denotes an upward variation in the spatial domain, and VARd is a downward variation in the spatial domain; VARl denotes a leftward variation in the spatial domain, and VARr is a rightward variation in the spatial domain; VARb, VARb1 and VARb2 denote previous variations in the time domain, and VARf, VARf1, VARf2 are subsequent variations in the time domain.
To represent Equations 10 and 11 in more detail, values of Equation 12 are proposed.
Wu = - 0.5 × Difh + Dift Difh + Difv + Dift × Difd Difu + Difd Wd = - 0.5 × Difh + Dift Difh + Difv + Dift × Difu Difu + Difd Wl = - 0.5 × Difv + Dift Difh + Difv + Dift × Difr Difr + Difl Wr = - 0.5 × Difv + Dift Difh + Difv + Dift × Difl Difr + Difl Wf = - 0.5 × Difv + Difh Difh + Difv + Dift × Difb Diff + Difb Wb = - 0.5 × Difv + Difh Difh + Difv + Dift × Diff Diff + Difb [ Equation 12 ]
In Equation 12, Difv is the absolute value of a vertical difference of the input video signal, Difh is the absolute value of a horizontal difference, Difu is the absolute value of an upward difference, and Difd is the absolute value of a downward difference. In addition, Difl is the absolute value of a leftward difference and Difr is the absolute value of a rightward difference. FIGS. 17 and 18 are diagrams for explaining vertical and horizontal variations in pixels of the NTSC system. Referring to FIG. 17, Difv=du+dd+dv, Difu=Du and Difd=dd for the currently processed central pixel (i, j). Referring to FIG. 18, Difh=dl+dr+dh, Difl=dl and Difr=dr. Here, dl is the absolute value of the difference between the pixel data (i, j) and pixel data (i, j−4), dr is the absolute value of the difference between the pixel data (i, j) and pixel data (i, j+4), and dh is the absolute value of the difference between the pixel data (i, j−2) and pixel data (i, j+2). Furthermore, du is the absolute value of the difference between the pixel data (i, j) and pixel data (i−2, j), dd is the absolute value of the difference between the pixel data (i, j) and pixel data (i+2, j), and dv is the absolute value of the difference between pixel data (i−1, j) and pixel data (i+1, j) in the NTSC system.
In the PAL system, du is the absolute value of the difference between the pixel data (i, j) and pixel data (i−4, j), dd is the absolute value of the difference between the pixel data (i, j) and pixel data (i+4, j), and dv is the absolute value of the difference between pixel data (i−2, j) and pixel data (i+2, j). Furthermore, Dift denotes a variation on the time axis, and Diff and Difb represent variations to subsequent and previous field directions. Here, Dift, Diff and Difb are calculated by combining pixels having the same phase as the current pixel on the time axis. In addition, the weights Wu, Wd, Wl, Wr, Wf and Wb can be determined such that they represent horizontal/vertical/temporal/leftward/rightward/upward/downward variations using various combinations of pixel data items having the same phase.
Consequently, Wu may be proportional to the vertical and upward correlation for the central pixel (i, j). Wd may be proportional to the vertical and downward correlation for the central pixel (i, j). Wl may be proportional to the horizontal and leftward correlation for the central pixel (i, j). Wr may be proportional to the horizontal and rightward correlation for the central pixel (i, j). Wb may be proportional to the correlation of the currently processed central pixel and data of the field before the current pixel, or the difference between pixels in data of the field/frame before the current field, which have the same phase as the currently processed central pixel, as shown in FIG. 16. Furthermore, Wf may be proportional to the correlation of the currently processed central pixel and data of the field after the current pixel, or the difference between pixels in data of the field/frame after the current field, which have the same phase as the currently processed central pixel,
In the NTSC system, Wb1 may be proportional to the correlation of the currently processed central pixel and data of the field before the current pixel, or the difference between the central pixel and a pixel in data of the field/frame before the current field, which have the same phase as the currently processed central pixel (i, j, t) in a first direction, as shown in FIG. 10. Wf1 may be proportional to the correlation of the currently processed central pixel and data of the field after the current pixel, or the difference between the central pixel and a pixel in data of the field/frame after the current field, which have the same phase as the currently processed central pixel (i, j, t) in a second direction. Furthermore, Wb2 may be proportional to the correlation of the currently processed central pixel and data of the field before the current pixel, or the difference between the central pixel and a pixel in data of the field/frame before the current field, which have the opposite phase to the currently processed central pixel (i, j, t) in a third direction, as shown in FIG. 11. Wf2 may be proportional to the correlation of the currently processed central pixel and data of the field after the current pixel, or the difference between the central pixel and a pixel in data of the field/frame after the current field, which have the opposite phase to the currently processed central pixel (i, j, t) in a fourth direction, as shown in FIG. 11.
The 3D BPF 330 performs filtering through convolution of Equation 3 using the 2D weight coefficients Wu, Wd, Wl and Wr and the third weight coefficients Wb1, Wb2, Wf1 and Wf2/Wb and Wf, which are determined as above.
FIG. 19 is a block diagram of the 3D BPF 330 of FIG. 3. Referring to FIG. 19, the 3D BPF 330 includes a first filter 331, a second filter 332, a third filter 333, a fourth filter 334, a fifth filter 335, and a combiner 336. The 3D BPF 330 may perform NTSC or PAL system operation in response to the control signal NPC.
When the control signal NPC has a logic high level, for example, the 3D BPF 330 may perform NTSC system operation using the corresponding data windows BBFID, BEFID, PRFID, AFFID and MFID of five fields. When the control signal NPC has a logic low level, the 3D BPF 330 may perform PAL system operation using the corresponding data windows BEFID, PRFID and AAFID of three fields.
In the NTSC system, the first filter 331, the second filter 332, the third filter 333, the fourth filter 334, the fifth filter 335 and the combiner 336 may all be operated. The first filter 331 performs convolution of the corresponding data window of the first field data BBFID of the input video signal and the first filer mask h1(i, j, t) determined by Equation 6. As described above with reference to FIG. 13, the data window of the first field data BBFID exists in field data that is two fields before the third field data PRFID currently processed.
The second filter 332 performs convolution of the corresponding data window of the second field data BEFID of the input video signal and the second filer mask h2(i, j, t) determined by Equation 6. The data window of the second field data BEFID exists in field data that is one field before the third field data PRFID currently processed.
The third filter 333 performs convolution of the corresponding data window of the third field data PRFID of the input video signal and the third filer mask h3(i, j, t) determined by Equation 6. The data window of the third field data PRFID exists in the currently processed field data.
The fourth filter 334 performs convolution of the corresponding data window of the fourth field data AFFID of the input video signal and the fourth filer mask h4(i, j, t) determined by Equation 6. The data window of the fourth field data AFFID exists in field data that is one field after the third field data PRFID currently processed.
The fifth filter 335 performs convolution of the corresponding data window of the fifth field data AAFID of the input video signal and the fifth filer mask h5(i, j, t) determined by Equation 6. The data window of the fourth field data AAFID exists in field data that is two fields after the third field data PRFID currently processed.
In the PAL system, the second filter 332, the third filter 333, the fourth filter 334 and the combiner 336 may be operated. Here, the second filter 332 performs convolution of the filter mask h1(i, j, t) determined by Equation 9 and the corresponding data window of the field data BEFID of the input video signal. That is, the second filter 332 uses data items in vertical and horizontal directions, which have the opposite C component phase to the central pixel, and data items in diagonal directions, which have the same C component phase as the central pixel, among digital data of three continuous horizontal scan lines of the field data BEFID that is one field before the current field, as represented by Equation 9.
The third filter 333 performs convolution of the filter mask h2(i, j, t) determined by Equation 9 and the corresponding data window of the field data PRFID of the input video signal. Here, the third filter 333 uses data items in vertical and horizontal directions, which have the opposite C component phase to the central pixel, and data items in diagonal directions, which have the same C component phase as the central pixel, among digital data of three continuous horizontal scan lines of the current field data PRFID, as represented by Equation 9.
The fourth filter 334 performs convolution of the filter mask h3(i, j, t) determined by Equation 9 and the corresponding data window of the field data AFFID of the input video signal. Here, the fourth filter 334 uses data items in vertical and horizontal directions, which have the opposite C component phase to the central pixel, and data items in diagonal directions, which have the same C component phase as the central pixel, among digital data of three continuous horizontal scan lines of the field data AFFID that is one field after the current field, as represented by Equation 9.
In the NTSC system, the combiner 336 appropriately combines the convolution results of the first to fifth filters 331 to 335, i.e., performs addition or subtraction on the convolution results, to output the result as the C signal. In the PAL system, the combiner 336 appropriately combines the convolution results of the second, third and fourth filter 332, 333 and 334, i.e., performs addition or subtraction on the convolution results, to output the result as the C signal.
Through the aforementioned operation, the 3D BPF 330 continuously and adaptively performs comb filtering, 1D band pass filtering and 2D/3D band pass filtering in response to spatio-temporal local characteristics of the input video signal.
For example, when correlation is high or low in all directions of the spatio-temporal domain, all the 2D weight coefficients Wu, Wd, Wl and Wr and the 3D weight coefficients Wb1, Wb2, Wf1 and Wf2/Wb and Wf of Equations 5 and 8 affect the filtering operation of the 3D BPF 330. Accordingly, in spectrum characteristics with respect to horizontal and vertical frequencies H and V, shown in FIG. 20, the high frequency component 12 of the C signal, included in the Y component 11, is extracted distinctly like a desired sharp form 13, to thereby remove artifacts. Specifically, an averaging effect among surrounding pixels stabilizes the signal in response to the weight coefficients of Equations 5 and 8 when correlation is high in all directions. When correlation is low in all directions, the C component is separated in the sharpest form, and thus cross-color artifacts are minimized and the high frequency component of the Y signal is extracted in a distinct manner. Similarly, in spectrum characteristics with respect to time and vertical frequencies T and V, shown in FIG. 21, the high frequency component 16 of the C signal, included in the Y component 15, is extracted distinctly like a desired sharp form 17, to thereby remove artifacts.
On the other hand, if correlation is spatio-temporally high in a specific direction, plane coefficients in the corresponding direction become large but plane coefficients in the other directions become small. Thus, comb filtering or 1D band pass filtering is carried out on pixels surrounding the central pixel to separate Y and C signals, thereby minimizing artifacts.
While the weight coefficients will determine the filtering performed, the 3D BPF 330 may serve as any of the three types of filters, e.g., a local vertical comb filter, horizontal band pass filter, and/or a 2D/3D band pass filter, as illustrated schematically in FIG. 22.
As described above, in the video signal processing apparatus 300 according to the present invention, the adaptive 3D BPF 330 performs Y/C separation according to local comb filtering/1D band pass filtering when the edge direction is fixed vertically/horizontally based on fields in response to spatio-temporal local characteristics of an image using spatio-temporal filters. When the edge direction is not fixed horizontally/vertically, the 3D BPF 330 performs 2D/3D band pass filtering in all directions. The 3D BPF continuously performs comb filtering, 1D band pass filtering and 2D band pass filtering in response to the spatio-temporal local characteristics of the image.
The video signal processing apparatus according to the present invention continuously and adaptively performs local comb filtering/1D band pass filtering or 2D band pass filtering in all directions in response to spatio-temporal local characteristics of an image in the NTSC/PAL system. Thus, the video signal processing apparatus of the present invention may detect the edge of the image accurately more than a conventional apparatus that discretely selects filtering methods, and produce better and more stable spectrum characteristics than a frame-based Y/C separation filter. Accordingly, artifacts such as cross-luma and cross-color may be reduced or removed when the video signal processing apparatus of the present invention is applied to a display system, resulting in improved image quality.
Exemplary embodiments of the present invention have been disclosed herein, and although specific terms are employed, they are used and are to be interpreted in a generic and descriptive sense only and not for purpose of limitation. For example, while embodiments of the present invention has been described relative to a hardware implementation, the processing of the present invention may be implemented in software, e.g., by an article of manufacture having a machine-accessible medium including data that, when accessed by a machine, cause the machine to separate the Y and C components. Accordingly, it will be understood by those of ordinary skill in the art that various changes in form and details may be made without departing from the spirit and scope of the present invention as set forth in the following claims.

Claims (24)

1. A video signal processing method, comprising:
generating 2D weight coefficients and 3D weight coefficients using data of a plurality of fields of an input video signal; and
generating a C signal of the input video signal by carrying out convolution of a horizontal coefficient mask, a vertical coefficient mask and a time coefficient mask, which are composed based on the weight coefficients, and corresponding data windows of the plurality of field data,
wherein generating the C signal includes local vertical comb filtering, horizontal band pass filtering and 2D band pass filtering according to spatio-temporal local characteristics of the input video signal.
2. The method as claimed in claim 1, wherein local vertical comb filtering is performed when local characteristics indicate high vertical correlation, local horizontal band pass filtering is performed when local characteristics indicate high horizontal correlation, and 2D or 3D band pass filtering is carried out when local characteristics indicate high or low correlation in multiple directions using at least three field data having an interval of one field.
3. The method as claimed in claim 1, wherein the 2D weight coefficients include a first coefficient proportional to vertical and upward correlations with the current pixel, a second coefficient proportional to vertical and downward correlations with the current pixel, a third coefficient proportional to horizontal and leftward correlations with the current pixel, and a fourth coefficient proportional to horizontal and rightward correlations with the current pixel.
4. The method as claimed in claim 1, further comprising:
subtracting the generated C signal from the input video signal; and
outputting the subtraction result as a Y signal.
5. The method as claimed in claim 1, wherein generating the C signal comprises:
performing convolution of a corresponding data window of a first field data of the input video signal and a first filter mask;
performing convolution of a corresponding data window of a second field data of the input video signal and a second filter mask;
performing convolution of a corresponding data window of a third field data of the input video signal and a third filter mask; and
combining the convolution results to output the combined result as the C signal,
wherein the first, second and third filter masks are generated from 3D convolution of the horizontal coefficient mask, the vertical coefficient mask and the time coefficient mask.
6. The method as claimed in claim 5, wherein the video signal method is applied to a PAL system.
7. The method as claimed in claim 5, wherein the first field data, the second field data and the third field data have an interval of one field.
8. The method as claimed in claim 5, wherein the horizontal coefficient mask, the vertical coefficient mask and the time coefficient mask each are one-dimensional.
9. The method as claimed in claim 5, wherein the 3D weight coefficients include a first coefficient in proportion to a correlation between the currently processed pixel data and data of the field before the current field, and a second coefficient in proportion to a correlation between the currently processed pixel data and data of the field after the current field.
10. The method as claimed in claim 5, wherein generating the C signal further comprises:
performing convolution of a corresponding data window of a fourth field data of the input video signal and a fourth filter mask;
performing convolution of a corresponding data window of a fifth field data of the input video signal and a fifth filter mask;
combining the five convolution results to be output the combined result as the generated C signal,
wherein the time coefficient mask includes first and second coefficient masks, and the first to fifth filter masks are generated from 3D convolution of the horizontal coefficient mask, the vertical coefficient mask, the first coefficient mask and the second coefficient mask.
11. The method as claimed in claim 10, wherein the video signal processing method is applied to an NTSC system.
12. The method as claimed in claim 10, wherein:
the 3D weight coefficients used for the first coefficient mask include;
a first coefficient in proportion to the correlation in a first direction between the currently processed pixel data and data of the field before the current field, and
a second coefficient in proportion to the correlation in the second direction between the currently processed pixel data and data of the field after the current field; and
the 3D weight coefficients used for the second coefficient mask include;
a third coefficient in proportion to the correlation in a third direction between the currently processed pixel data and data of the field before the current field, and
a fourth coefficient in proportion to the correlation in a fourth direction between the currently processed pixel data and data of the field after the current field.
13. A video signal processing apparatus, comprising:
a weight determination unit to generate 2D weight coefficients and 3D weight coefficients using data of a plurality of fields of an input video signal; and
a filter to convolute a horizontal coefficient mask, a vertical coefficient mask and a time coefficient mask, the masks respectively composed of the weight coefficients, with corresponding data windows of the data of the plurality of fields to generate a C signal of the input video signal,
wherein the filter adaptively performs local vertical comb filtering, horizontal band pass filtering and 2D band pass filtering in response to spatio-temporal local characteristics of the input video signal.
14. The apparatus as claimed in claim 13, wherein the filter performs:
local vertical comb filtering when the local characteristics of the input video signal indicate high vertical correlation;
local horizontal band pass filtering when the local characteristics of the input video signal indicate high horizontal correlation; and
2D or 3D band pass filtering when the local characteristics of the input video signal indicate high or low correlation in multiple directions using at least three field data items having an interval of one field.
15. The apparatus as claimed in claim 13, wherein the 2D weight coefficients include:
a first coefficient proportional to vertical and upward correlations with a current pixel;
a second coefficient proportional to vertical and downward correlations with the current pixel;
a third coefficient proportional to horizontal and leftward correlations with the current pixel; and
a fourth coefficient proportional to horizontal and rightward correlations with the current pixel.
16. The apparatus as claimed in claim 13, further comprising a subtracter to subtract the C signal output from the filter from the input video signal to obtain the subtraction result as a Y signal.
17. The apparatus as claimed in claim 13, wherein the filter comprises:
a first filter to convolute a first data window of a first field data of the input video signal and a first filter mask;
a second filter to convolute a second data window of a second field data of the input video signal and a second filter mask;
a third filter to convolute a third data window of a third field data of the input video signal and a third filter mask; and
a combiner to combine the convolution results to first, second and thrid filters to generate the C signal,
wherein the first, second and third filter masks are generated from 3D convoluting the horizontal coefficient mask, the vertical coefficient mask and the time coefficient mask.
18. The apparatus as claimed in claim 17, wherein the video signal apparatus is applied to a PAL system.
19. The apparatus as claimed in claim 17, wherein the first field data, the second field data and the third field data have an interval of one field.
20. The apparatus as claimed in claim 17, wherein the horizontal coefficient mask, the vertical coefficient mask and the time coefficient mask each are one-dimensional.
21. The apparatus as claimed in claim 17, wherein the 3D weight coefficients include:
a first coefficient proportional to a correlation between a current pixel data and data of a field before a current field; and
a second coefficient proportional to a correlation between the current pixel data and data of a field after the current field.
22. The apparatus as claimed in claim 17, wherein the filter further comprises:
a fourth filter to convolute of a fourth data window of a fourth field data of the input video signal and a fourth filter mask; and
a fifth filter to convolute a fifth data window of a fifth field data of the input video signal and a fifth filter mask,
wherein the combiner combines the convolution results of the first to fifth filters to generate the C signal, the time coefficient mask includes first and second coefficient masks, and the first to fifth filter masks are generated from 3D convolution of the horizontal coefficient mask, the vertical coefficient mask, the first coefficient mask and the second coefficient mask.
23. The apparatus as claimed in claim 22, wherein the video signal processing apparatus is applied to an NTSC system.
24. The apparatus as claimed in claim 22, wherein:
the 3D weight coefficients used for the first coefficient mask include
a first coefficient proportional to a correlation in a first direction between the current pixel data and data of the field before the current field, and
a second coefficient proportional to a correlation in the second direction between the currently processed pixel data and data of the field after the current field; and
the 3D weight coefficients used for the second coefficient mask include
a third coefficient proportional to a correlation in a third direction between the currently processed pixel data and data of the field before the current field, and
a fourth coefficient proportional to the correlation in a fourth direction between the current pixel data and data of the field after the current field.
US11/331,360 2005-01-13 2006-01-13 Digital video signal processing apparatus and method for field-based adaptive spatio-temporal Y/C separation Expired - Fee Related US7659944B2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR1020050003178A KR100688519B1 (en) 2005-01-13 2005-01-13 Digital video signal processing apparatus and method for adaptive and temporal and spatial Y/C separation based on field period
KR10-2005-0003178 2005-01-13

Publications (2)

Publication Number Publication Date
US20060181648A1 US20060181648A1 (en) 2006-08-17
US7659944B2 true US7659944B2 (en) 2010-02-09

Family

ID=36815246

Family Applications (1)

Application Number Title Priority Date Filing Date
US11/331,360 Expired - Fee Related US7659944B2 (en) 2005-01-13 2006-01-13 Digital video signal processing apparatus and method for field-based adaptive spatio-temporal Y/C separation

Country Status (4)

Country Link
US (1) US7659944B2 (en)
KR (1) KR100688519B1 (en)
CN (1) CN1805553B (en)
TW (1) TWI318537B (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7944508B1 (en) * 2007-06-08 2011-05-17 Kolorific, Inc. Method and system for automatic detection and suppression of cross-luma and cross-color in a component video signal
KR20090013513A (en) * 2007-08-02 2009-02-05 삼성전자주식회사 Video processing method and video processing apparatus using the same
US8593572B2 (en) * 2008-01-30 2013-11-26 Csr Technology Inc. Video signal motion detection
TWI415479B (en) * 2010-01-27 2013-11-11 Sunplus Technology Co Ltd System and method for cvbs signal decoding and de-interlacing
FR2968878A1 (en) * 2010-12-14 2012-06-15 Thomson Licensing METHOD AND DEVICE FOR GENERATING IMAGES COMPRISING KINETIC FLOU
CN104038749B (en) * 2013-03-08 2016-02-03 矽统科技股份有限公司 Image compensating device
TWI602422B (en) * 2016-08-17 2017-10-11 Chipone Technology (Beijing)Co Ltd Image processing device and method thereof

Citations (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5473389A (en) * 1992-11-10 1995-12-05 Sony Corporation Y/C separator using 3-D, 2-D and 1-D filters
JPH0818997A (en) 1994-06-30 1996-01-19 Sony Corp Filter circuit
US5502509A (en) * 1993-06-21 1996-03-26 Mitsubishi Denki Kabushiki Kaisha Chrominance-luminance separation method and filter performing selective spatial filter based on detected spatial correlation
US5541669A (en) * 1990-03-29 1996-07-30 Mitsubishi Denki Kabushiki Denki Motion adaptive luminance signal and color signal separation filter
US5585861A (en) * 1991-03-14 1996-12-17 Mitsubishi Denki Kabushiki Kaisha Luminance and chrominance signals separating filter adaptive to movement of image
KR970068684A (en) 1996-03-25 1997-10-13 빈센트 비. 인그라시아 Y / C Separating Circuit
US5686972A (en) * 1994-06-18 1997-11-11 Samsung Electronics Co., Ltd. Circuit for adaptively separating luminance and color signals based on expended vertical and horizontal correlations and method therefor
US5909255A (en) * 1996-02-19 1999-06-01 Matsushita Electric Industrial Co., Ltd. Y/C separation apparatus
US5990978A (en) * 1996-03-23 1999-11-23 Lg Electronics Inc. Luminance/chrominance signal separating device
US6055024A (en) * 1998-09-02 2000-04-25 Rockwell Semiconductor Systems, Inc. Adaptive method for Y-C separation in video signals
JP2001204045A (en) 2000-01-24 2001-07-27 Mitsubishi Electric Corp Motion detector
US6288754B1 (en) * 1997-05-21 2001-09-11 Victor Company Of Japan, Ltd. YC separation circuit
US6300985B1 (en) * 1999-03-12 2001-10-09 Fortel Dtv Composite video decoder for providing a high quality serial digital output
US6504579B1 (en) 1997-08-05 2003-01-07 Micronas Intermettal Gmbh Adaptive filter for processing video signals
US6674488B1 (en) * 2000-02-08 2004-01-06 Oki Electric Industry Co., Ltd. Luminance and color difference signal separator for adaptively selecting color difference signals output from comb and line filters
US6774954B1 (en) * 2001-06-28 2004-08-10 Ndsp Corporation Apparatus and method for adaptive three dimensional television Y/C separation comb filter bank
US20040174464A1 (en) * 2003-03-05 2004-09-09 Macinnis Alexander G. System and method for three dimensional comb filtering
US6795126B1 (en) * 2001-12-14 2004-09-21 Ndsp Corporation Method and system for adaptive three-dimensional color television Y/C separation comb filter design
US20040201781A1 (en) * 2003-04-09 2004-10-14 Masaya Kobayashi Video signal processor and TV receiver using the same
US6809778B2 (en) * 2000-07-10 2004-10-26 Matsushita Electric Industrial Co., Ltd. Apparatus and method for YC separation and three-line correlation detection
US6914638B2 (en) * 2000-12-20 2005-07-05 Intel Corporation Three-dimensional enhancement processing for television broadcasting signals
US6956620B2 (en) * 2001-03-12 2005-10-18 Samsung Electronics Co., Ltd. Apparatus for separating a luminance signal and a chrominance signal from an NTSC composite video signal
US20060077302A1 (en) * 2002-12-16 2006-04-13 Koninklijke Philips Electronics N.V. Comb filter
US7046306B2 (en) * 2003-03-31 2006-05-16 Texas Instruments Incorporated Processing a video signal using motion estimation to separate luminance information from chrominance information in the video signal
US20060176406A1 (en) 2005-01-13 2006-08-10 Samsung Electronics Co., Ltd. Digital video signal processing apparatus and method for frame-based adaptive temporal and spatial Y/C separation
US7176984B2 (en) * 2003-03-21 2007-02-13 Mediatek Inc. Video signal separating method and related apparatus
US7196736B2 (en) * 2002-05-31 2007-03-27 Victor Company Of Japan, Limited Method of YC separation of YC separation filter
US7274408B2 (en) * 2003-12-22 2007-09-25 Sunplus Technology Co., Ltd. [Method of motion detection for 3D comb filter video decoder]
US7304688B1 (en) * 2003-05-20 2007-12-04 Pixelworks, Inc. Adaptive Y/C separator
US7324163B2 (en) * 2003-12-23 2008-01-29 Genesis Microchip Inc. Sub-carrier detection (SCD) for motion threshold logic
US7420624B2 (en) * 2004-08-31 2008-09-02 Sunplus Technology Co., Ltd. Three-dimensional video composite signal demodulating method and system capable of inhibiting pseudo-color noises
US7453525B2 (en) * 2005-05-24 2008-11-18 Texas Instruments Incorporated Motion detector for a video display system

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100246911B1 (en) * 1996-01-11 2000-03-15 윤종용 Auto compensation apparatus and method for delaying difference of between luminance and chrominance signal
DE19609193A1 (en) * 1996-03-09 1997-09-11 Thomson Brandt Gmbh Method and circuit arrangement for separating luminance and chrominance signals of a CVBS signal

Patent Citations (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5541669A (en) * 1990-03-29 1996-07-30 Mitsubishi Denki Kabushiki Denki Motion adaptive luminance signal and color signal separation filter
US5585861A (en) * 1991-03-14 1996-12-17 Mitsubishi Denki Kabushiki Kaisha Luminance and chrominance signals separating filter adaptive to movement of image
US5473389A (en) * 1992-11-10 1995-12-05 Sony Corporation Y/C separator using 3-D, 2-D and 1-D filters
US5502509A (en) * 1993-06-21 1996-03-26 Mitsubishi Denki Kabushiki Kaisha Chrominance-luminance separation method and filter performing selective spatial filter based on detected spatial correlation
US5686972A (en) * 1994-06-18 1997-11-11 Samsung Electronics Co., Ltd. Circuit for adaptively separating luminance and color signals based on expended vertical and horizontal correlations and method therefor
JPH0818997A (en) 1994-06-30 1996-01-19 Sony Corp Filter circuit
US5909255A (en) * 1996-02-19 1999-06-01 Matsushita Electric Industrial Co., Ltd. Y/C separation apparatus
US5990978A (en) * 1996-03-23 1999-11-23 Lg Electronics Inc. Luminance/chrominance signal separating device
KR970068684A (en) 1996-03-25 1997-10-13 빈센트 비. 인그라시아 Y / C Separating Circuit
US6288754B1 (en) * 1997-05-21 2001-09-11 Victor Company Of Japan, Ltd. YC separation circuit
US6504579B1 (en) 1997-08-05 2003-01-07 Micronas Intermettal Gmbh Adaptive filter for processing video signals
US6055024A (en) * 1998-09-02 2000-04-25 Rockwell Semiconductor Systems, Inc. Adaptive method for Y-C separation in video signals
US6300985B1 (en) * 1999-03-12 2001-10-09 Fortel Dtv Composite video decoder for providing a high quality serial digital output
JP2001204045A (en) 2000-01-24 2001-07-27 Mitsubishi Electric Corp Motion detector
US6693676B2 (en) 2000-01-24 2004-02-17 Mitsubishi Denki Kabushiki Kaisha Motion detecting apparatus for detecting motion information of picture within signal
US6674488B1 (en) * 2000-02-08 2004-01-06 Oki Electric Industry Co., Ltd. Luminance and color difference signal separator for adaptively selecting color difference signals output from comb and line filters
US6809778B2 (en) * 2000-07-10 2004-10-26 Matsushita Electric Industrial Co., Ltd. Apparatus and method for YC separation and three-line correlation detection
US6914638B2 (en) * 2000-12-20 2005-07-05 Intel Corporation Three-dimensional enhancement processing for television broadcasting signals
US6956620B2 (en) * 2001-03-12 2005-10-18 Samsung Electronics Co., Ltd. Apparatus for separating a luminance signal and a chrominance signal from an NTSC composite video signal
US6774954B1 (en) * 2001-06-28 2004-08-10 Ndsp Corporation Apparatus and method for adaptive three dimensional television Y/C separation comb filter bank
US6795126B1 (en) * 2001-12-14 2004-09-21 Ndsp Corporation Method and system for adaptive three-dimensional color television Y/C separation comb filter design
US7196736B2 (en) * 2002-05-31 2007-03-27 Victor Company Of Japan, Limited Method of YC separation of YC separation filter
US20060077302A1 (en) * 2002-12-16 2006-04-13 Koninklijke Philips Electronics N.V. Comb filter
US7227587B2 (en) * 2003-03-05 2007-06-05 Broadcom Corporation System and method for three dimensional comb filtering
US20040174464A1 (en) * 2003-03-05 2004-09-09 Macinnis Alexander G. System and method for three dimensional comb filtering
US20070153127A1 (en) * 2003-03-05 2007-07-05 Broadcom Corporation System and method for three dimensional comb filtering
US7176984B2 (en) * 2003-03-21 2007-02-13 Mediatek Inc. Video signal separating method and related apparatus
US7046306B2 (en) * 2003-03-31 2006-05-16 Texas Instruments Incorporated Processing a video signal using motion estimation to separate luminance information from chrominance information in the video signal
US20040201781A1 (en) * 2003-04-09 2004-10-14 Masaya Kobayashi Video signal processor and TV receiver using the same
US7133080B2 (en) * 2003-04-09 2006-11-07 Sony Corporation Video signal processor and TV receiver using the same
US7304688B1 (en) * 2003-05-20 2007-12-04 Pixelworks, Inc. Adaptive Y/C separator
US7274408B2 (en) * 2003-12-22 2007-09-25 Sunplus Technology Co., Ltd. [Method of motion detection for 3D comb filter video decoder]
US7324163B2 (en) * 2003-12-23 2008-01-29 Genesis Microchip Inc. Sub-carrier detection (SCD) for motion threshold logic
US7420624B2 (en) * 2004-08-31 2008-09-02 Sunplus Technology Co., Ltd. Three-dimensional video composite signal demodulating method and system capable of inhibiting pseudo-color noises
TW200633550A (en) 2005-01-13 2006-09-16 Samsung Electronics Co Ltd Digital video signal processing apparatus and method for frame-based adaptive temporal and spatial Y/C separation
US20060176406A1 (en) 2005-01-13 2006-08-10 Samsung Electronics Co., Ltd. Digital video signal processing apparatus and method for frame-based adaptive temporal and spatial Y/C separation
US7453525B2 (en) * 2005-05-24 2008-11-18 Texas Instruments Incorporated Motion detector for a video display system

Also Published As

Publication number Publication date
KR100688519B1 (en) 2007-03-02
TWI318537B (en) 2009-12-11
TW200701801A (en) 2007-01-01
KR20060082606A (en) 2006-07-19
CN1805553A (en) 2006-07-19
CN1805553B (en) 2010-09-22
US20060181648A1 (en) 2006-08-17

Similar Documents

Publication Publication Date Title
US6545719B1 (en) Apparatus and method for concealing interpolation artifacts in a video interlaced to progressive scan converter
US7659944B2 (en) Digital video signal processing apparatus and method for field-based adaptive spatio-temporal Y/C separation
US7477319B2 (en) Systems and methods for deinterlacing video signals
KR950002659B1 (en) Video signal interpolator
US7869650B2 (en) Digital video signal processing apparatus and method for compensating a chrominance signal
US7605870B2 (en) Digital video signal processing apparatus and method for frame-based adaptive temporal and spatial Y/C separation
EP1411719A2 (en) Method and system for converting interlaced video signal to progressive video signal using a color edge detection
US7663702B2 (en) Digital video signal processing apparatus and method for adaptive temporal and spatial Y/C separation in multiple directions
JP5173135B2 (en) Digital video signal processing apparatus and method for adaptive Y / C separation
JPH02272894A (en) Color image sigal encoding
JPH0779476B2 (en) Luminance signal Color signal separation circuit
US20060125965A1 (en) Method and device for separating a chrominance signal from a composite video baseband signal
US7430020B1 (en) Separation of composite video signals using dot crawl detector
JP3024465B2 (en) Inter-frame luminance signal / color signal separation device
KR940002198B1 (en) Motion adaptive luminance signal and chrominance signal separation filter
JPH0716255B2 (en) Motion adaptive luminance signal color signal separation device
JP2646680B2 (en) Interpolation signal formation circuit
JP2557511B2 (en) Motion detection circuit for television display screen
KR940002197B1 (en) Motion adaptive luminance signal and chrominance signal separation filter
JPH02202192A (en) Motion detecting circuit for television display screen
JPH03274888A (en) Motion adaptive type luminance signal/chrominance signal separating filter
JPH04186991A (en) Moving adaptive luminance signal chrominance signal separation filter
JPH06245225A (en) Motion adaptive luminance signal chrominance signal separator circuit
JPH03219791A (en) Movement detector
JPH08256356A (en) Y/c separator

Legal Events

Date Code Title Description
AS Assignment

Owner name: SAMSUNG ELECTRONICS CO., LTD.,KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PARK, SUNG-CHEOL;LIM, HYUNG-JUN;PARK, JAE-HONG;AND OTHERS;SIGNING DATES FROM 20060313 TO 20060314;REEL/FRAME:017845/0572

Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PARK, SUNG-CHEOL;LIM, HYUNG-JUN;PARK, JAE-HONG;AND OTHERS;REEL/FRAME:017845/0572;SIGNING DATES FROM 20060313 TO 20060314

FPAY Fee payment

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.)

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.)

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20180209