http://chineseinput.net/에서 pinyin(병음)방식으로 중국어를 변환할 수 있습니다.
변환된 중국어를 복사하여 사용하시면 됩니다.
Relative SATD-based Minimum Risk Bayesian Framework for Fast Intra Decision of HEVC
( Daehyeok Gwon ),( Haechul Choi ) 한국인터넷정보학회 2019 KSII Transactions on Internet and Information Syst Vol.13 No.1
High Efficiency Video Coding (HEVC) enables significantly improved compression performance relative to existing standards. However, the advance also requires high computational complexity. To accelerate the intra prediction mode decision, a minimum risk Bayesian classification framework is introduced. The classifier selects a small number of candidate modes to be evaluated by a rate-distortion optimization process using the sum of absolute Hadamard transformed difference (SATD). Moreover, the proposed method provides a loss factor that is a good trade-off model between computational complexity and coding efficiency. Experimental results show that the proposed method achieves a 31.54% average reduction in the encoding run time with a negligible coding loss of 0.93% BD-rate relative to HEVC test model 16.6 for the Intra_Main common test condition.
V-PCC 를 위한 Occupancy 정보 기반의 Texture 영상 부호화 방법
권대혁(Daehyeok Gwon),최해철(Haechul Choi) 한국방송·미디어공학회 2021 한국방송공학회 학술발표대회 논문집 Vol.2021 No.6
포인트 클라우드는 특정 개체 혹은 장면을 다수의 3 차원 포인터를 사용하여 표현하는 데이터의 표현 방식 중 하나로 3D 데이터를 정밀하게 수집하고 표현할 수 있는 방법이다. 하지만 방대한 양의 데이터를 필요로 하기 때문에 효율적인 압축이 필수적이다. 이에 따라 국제 표준화 단체인 Moving Picture Experts Group 에서는 포인트 클라우드 데이터의 효율적인 압축 방법 중 하나로 Video based Point Cloud Compression(V-PCC)에 대한 표준을 제정하였다. V-PCC 는 포인트 클라우드 정보를 Occupancy, Geometry, Texture 와 같은 다수의 2D 영상으로 변환하고 각 2D 영상을 전통적인 2D 비디오 코덱을 활용하여 압축하는 방법이다. 본 논문에서는 V-PCC 에서 변환하는 Occupancy 의 정보를 활용하여 효율적으로 Texture 영상을 압축할 수 있은 방법을 소개한다. 또한 제안 방법이 V-PCC 에서 약 1%의 부호화 효율을 얻을 수 있음을 보인다.
타일의 독립적 복호를 위한 HEVC 부호화 방법 및 문제점 분석
권대혁(Daehyeok Gwon),백아람(Aram Beak),최해철(Haechul Choi) 한국방송·미디어공학회 2017 방송공학회논문지 Vol.22 No.6
Ultra-high definition videos, panorama contents, and ultra-wide viewing videos have a huge spatial resolution. However, a whole region of images is not always interesting to viewers due to limitations of system resources and display devices. To allow one or more interesting tiles to be decoded freely without decoding other tiles, this paper introduces a tile-based independent coding method. The propose method restraints motion vectors to be within a tile region shifting motion search area and modifying an initial motion vector. Experiments results show that any desired tile is capable to be decoded independently of other tiles.
Adaptive block tree structure for video coding
Baek, Aram,Gwon, Daehyeok,Son, Sohee,Lee, Jinho,Kang, Jung-Won,Kim, Hui Yong,Choi, Haechul Electronics and Telecommunications Research Instit 2021 ETRI Journal Vol.43 No.2
The Joint Video Exploration Team (JVET) has studied future video coding (FVC) technologies with a potential compression capacity that significantly exceeds that of the high-efficiency video coding (HEVC) standard. The joint exploration test model (JEM), a common platform for the exploration of FVC technologies in the JVET, employs quadtree plus binary tree block partitioning, which enhances the flexibility of coding unit partitioning. Despite significant improvement in coding efficiency for chrominance achieved by separating luminance and chrominance tree structures in I slices, this approach has intrinsic drawbacks that result in the redundancy of block partitioning data. In this paper, an adaptive tree structure correlating luminance and chrominance of single and dual trees is presented. Our proposed method resulted in an average reduction of -0.24% in the Y Bjontegaard Delta rate relative to the intracoding of JEM 6.0 common test conditions.
박대민(Daemin Park),권대혁(Daehyeok Gwon),최진혁(Jinhuyck Choi),이인재(Injae Lee),최해철(Haechul Choi) 한국방송·미디어공학회 2015 방송공학회논문지 Vol.20 No.1
HCI(Human Computer Interaction) enables the interaction between people and computers by using a human-familiar interface called as Modality. Recently, to provide an optimal interface according to various devices and service environment, an advanced HCI method using multiple modalities is intensively studied. However, the multimodal interface has difficulties that modalities have different data formats and are hard to be cooperated efficiently. To solve this problem, a multimodal communicator is introduced, which is based on EMMA(Extensible Multimodal Annotation Markup language) and MMI(Multimodal Interaction Framework) of W3C(World Wide Web Consortium) standards. This standard based framework consisting of modality component, interaction manager, and presentation component makes multiple modalities interoperable and provides a wide expansion capability for other modalities. Experimental results show that the multimodal communicator is facilitated by using multiple modalities of eye tracking and gesture recognition for a map browsing scenario.