Repository logo
  • English
  • 中文
Log In
Have you forgotten your password?
  1. Home
  2. College of Electrical Engineering and Computer Science / 電機資訊學院
  3. Electronics Engineering / 電子工程學研究所
  4. Algorithm and Architecture Analysis of the Video Signal Conversion for 2D and 3D Video
 
  • Details

Algorithm and Architecture Analysis of the Video Signal Conversion for 2D and 3D Video

Date Issued
2007
Date
2007
Author(s)
Chang, Yu-Lin
DOI
en-US
URI
http://ntur.lib.ntu.edu.tw//handle/246246/57260
Abstract
Human are pursuing the reality of vision devices. The video devices improve from monochrome television to 3D-LCD today. The video signals also vary in all of these devices. In this dissertation, the video signal conversion for 2D and 3D video are discussed in two different parts: de-interlacing and 2D-to-3D conversion. The deinterlacing methods recover the lost data in temporal and spatial domain of a 2D video sequence. The 2D-to-3D conversion produces the whole dimensional data as the depth map of a 2D video, then it converts the depth map and 2D video into 3D video. The transition between interlaced scanned TV signals and progressive scanned TV signals hindered the quality improvement of the new display panels. Post-processing such as de-interlacing has become a great index for a TV decoder showing its performance. In Part I, three kinds of de-interlacing methods are described first: the intrafield de-interlacing, the motion adaptive de-interlacing, and the motion compensated deinterlacing. Second, for better de-interlaced image quality, we proposed an intra-field deinterlacing algorithm named “Extended Intelligent Edge-based Line Average” (EIELA). Its VLSI module implementation is also stated. Third, for near-perfect de-interlaced image quality, a de-interlacing algorithm using adaptive global and local motion estimation/- compensation is proposed. It consists of the global and local motion estimation/compensation, 4-field motion adaptation, the block-based directional edge interpolation, and the GMC/MC/MA block mode decision module. All defects such as jagged effects, blurring, line-crawling, and feathering are suppressed lower than the traditional methods. Moreover, the true motion information is extracted accurately by the 4-field motion estimation and global motion information. In Part II, we first make a detailed survey for different kinds of 3D video capturing methods. There are three kinds of 3D video capturing methods: the active sensor based methods, the passive sensor based methods, and the 2D-to-3D conversion. After analyzing the previous works, a real-time automatic depth fusion 2D-to-3D conversion system is proposed for the home multimedia platform. In Part III, we tried to convert the binocular, monocular, and pictorial depth cue to depth reconstruction algorithms. Five novel algorithms and hardware architecture are presented. The depth reconstruction algorithms can be classified into three categories: the motion parallax based depth reconstruction which utilizes the binocular depth cue, the image based depth reconstruction which uses the monocular depth cue, and the consciousness based depth reconstruction which map the perspective in pictorial depth cue to depth gradient. After the depth reconstruction, a priority depth fusion algorithm is proposed to integrate all the depth maps. Then a multiview depth image based rendering method is presented to provide multiview image rendering technique for the multiview 3D-LCD. One-dimensional cross search dense disparity estimation is proposed for the motion parallax based depth reconstruction. The fast algorithm utilizes the characteristics of the motion parallax and trinocular cameras. As the motion parallax is induced by camera motion, 1D cross search tends to find better and more smooth results for a true depth map. A symmetric trinocular property for trinocular camera stereo matching is also described. Then a 2D full search dense disparity estimation hardware architecture design is designed for the real-time operation of the motion parallax based depth reconstruction. The dense disparity estimation needs to calculate the disparity vectors of each depth pixel. With the features of resolution switching and high specification, the proposed hardware architecture uses a data assignment unit as a small buffer to achieve a IP-based design. The hardware can be switched to three different depth pixel resolution in real-time. Depth from Focus and short-term motion assisted color segmentation are proposed for the image based depth reconstruction. The DfF method adapts the “blurriness” characteristic while taking pictures with large aperture camera. After extracting the object from the blurring areas, the depth of the object is set to the focus distance of the taken picture. The second image-based depth reconstruction method is the depth map generation by short-term motion assisted color segmentation. It achieves a smooth depth map generation both in the spatial and temporal domain. But both methods would face the moving cameras problem and the tuning of various different type image sequences in the future. They should be combined with the depth from geometry perspective and other depth cues to produce more accurate depth map. For the consciousness based depth reconstruction, we have presented a fundamental detection algorithm based on the structural components analysis with robustness. It is suitable for images with distinct object edges. The proposed method for vanishing line and vanishing point detection provides direct analysis from image structure without complicated math calculation. The proposed method is feasible for a particular image sequence without prior temporal information, and guarantees that dominant vanishing lines are detected correctly with high probability and accuracy. The proposed block-based algorithm which still holds the regular block data flow feature is much faster, simpler and efficient. As for the 2D-to-3D conversion procedure, the proposed vanishing line and point detection gives great help for the overall scene knowledge, and the conversion proceeds more easily. After retrieving all the depth maps from different depth cues, we proposed a priority depth fusion method to integrate the three depth maps. It considers the priority of the depth maps in six aspects: the scene adaptability, the temporal consistency, perceptibility, correctness, fineness, and cover area. In order to obtain a comfortable depth map, the six aspects should be deliberated to decide the priority. We also proposed a per-pixel texture mapping depth image based rendering algorithm which can be accelerated by the GPU. The proposed algorithm converts points to vertices. Then an image plane represents the original frame and depth map is constructed. Through the GPU pipeline, the left and right image can be rendered out. Even for a free viewpoint application, as long as the GPU draws more than 49.6Mtriangles per second, the multiview DIBR still can run at real-time. And the proposed DIBR also speed up the previous design for 38 times. After having these algorithms, an automatic depth fusing 2D-to-3D conversion system is described. The proposed system generates the depth map of most of the commercial video with hardware acceleration. With the calculation of GPU, the depth map and the original 2D image are converted to stereo images for showing on the 3D display devices. Huge amount of the 2D contents such as DVD or TV programs are able to convert and show on the 3D display devices on the enduser side. In summary, this dissertation presents an intra-field de-interlacing hardware architecture, named extended intelligent edge based line average and an adaptive local/- global motion compensated de-interlacing method for the de-interlacing of 2D video. For 2D video signals to 3D video signals, an automatic depth fusing 2D-to-3D conversion system is proposed to utilize the human depth cues to convert 2D video to 3D video. There are five algorithms proposed in this 2D-to-3D conversion system using different depth cues: one dimensional cross search dense disparity estimation, depth map generation with short-term motion assisted color segmentation, block-based vanishing line/point detection, per-pixel multiview depth image based rendering, and priority depth fusion. There is also a hardware architecture of 2D full search dense disparity estimation implemented to combine with the whole 2D-to-3D conversion system. The proposed 2D-to-3D conversion not only produces acceptable depth map for 2D video but also renders multiview video from the depth information and the 2D video.
Subjects
去交錯
電視後處理
二維至三維影像轉換
深度圖生成
三維影像
De-interlacing
TV Post-processing
2D-to-3D conversion
depth map generation
3D video
Type
thesis
File(s)
Loading...
Thumbnail Image
Name

ntu-96-F91943052-1.pdf

Size

23.31 KB

Format

Adobe PDF

Checksum

(MD5):f094d55cc833a991a73a6887f50d1a8f

臺大位居世界頂尖大學之列,為永久珍藏及向國際展現本校豐碩的研究成果及學術能量,圖書館整合機構典藏(NTUR)與學術庫(AH)不同功能平台,成為臺大學術典藏NTU scholars。期能整合研究能量、促進交流合作、保存學術產出、推廣研究成果。

To permanently archive and promote researcher profiles and scholarly works, Library integrates the services of “NTU Repository” with “Academic Hub” to form NTU Scholars.

總館學科館員 (Main Library)
醫學圖書館學科館員 (Medical Library)
社會科學院辜振甫紀念圖書館學科館員 (Social Sciences Library)

開放取用是從使用者角度提升資訊取用性的社會運動,應用在學術研究上是透過將研究著作公開供使用者自由取閱,以促進學術傳播及因應期刊訂購費用逐年攀升。同時可加速研究發展、提升研究影響力,NTU Scholars即為本校的開放取用典藏(OA Archive)平台。(點選深入了解OA)

  • 請確認所上傳的全文是原創的內容,若該文件包含部分內容的版權非匯入者所有,或由第三方贊助與合作完成,請確認該版權所有者及第三方同意提供此授權。
    Please represent that the submission is your original work, and that you have the right to grant the rights to upload.
  • 若欲上傳已出版的全文電子檔,可使用Open policy finder網站查詢,以確認出版單位之版權政策。
    Please use Open policy finder to find a summary of permissions that are normally given as part of each publisher's copyright transfer agreement.
  • 網站簡介 (Quickstart Guide)
  • 使用手冊 (Instruction Manual)
  • 線上預約服務 (Booking Service)
  • 方案一:臺灣大學計算機中心帳號登入
    (With C&INC Email Account)
  • 方案二:ORCID帳號登入 (With ORCID)
  • 方案一:定期更新ORCID者,以ID匯入 (Search for identifier (ORCID))
  • 方案二:自行建檔 (Default mode Submission)
  • 方案三:學科館員協助匯入 (Email worklist to subject librarians)

Built with DSpace-CRIS software - Extension maintained and optimized by 4Science