http://chineseinput.net/에서 pinyin(병음)방식으로 중국어를 변환할 수 있습니다.
변환된 중국어를 복사하여 사용하시면 됩니다.
다중 카메라로 관심선수를 촬영한 동영상에서 베스트 뷰 추출방법
홍호탁,엄기문,낭종호 한국정보과학회 2017 정보과학회논문지 Vol.44 No.12
In recent years, the number of video cameras that are used to record and broadcast live sporting events has increased, and selecting the shots with the best view from multiple cameras has been an actively researched topic. Existing approaches have assumed that the background in video is fixed. However, this paper proposes a best view selection method for cases in which the background is not fixed. In our study, an athlete of interest was recorded in video during motion with multiple cameras. Then, each frame from all cameras is analyzed for establishing rules to select the best view. The frames were selected using our system and are compared with what human viewers have indicated as being the most desirable. For the evaluation, we asked each of 20 non-specialists to pick the best and worst views. The set of the best views that were selected the most coincided with 54.5% of the frame selection using our proposed method. On the other hand, the set of views most selected as worst through human selection coincided with 9% of best view shots selected using our method, demonstrating the efficacy of our proposed method. 최근 스포츠 중계에 동원되는 카메라 대수가 증가함에 따라 수많은 카메라 화면 중 순간적으로 최고의 화면을 고르는데 어려움이 있다. 지금까지 스포츠 경기를 촬영한 영상들에서 자동으로 최고의 화면을 선택하는 방법들이 연구되어 왔지만 배경이 고정된 영상들만을 고려해 배경이 움직이는 영상들을 고려하는 연구가 필요하다. 본 논문에서는 각 영상 별로 관심선수를 추적하여 획득한 영상 내 관심선수 영역을 대상으로 관심선수의 활동량, 얼굴 가시성, 다른 선수와의 겹침 정도, 이미지 블러 현상 정도를 매프레임 마다 정량적으로 나타내어 정량화된 값을 기반으로 최고의 화면을 선택한다. 이렇게 선택된 베스트뷰를 20명의 일반 사람들에게 베스트 뷰와 워스트 뷰를 선택하게 하여 사람들이 선택한 베스트 뷰, 워스트 뷰와 비교한 결과 베스트 뷰와 일치율이 54.5%로 낮았지만 반대로 워스트 뷰와 일치율이 9%로 확실히 사람들이 선호하지 않는 화면은 선택하지 않는 것을 알 수 있었다.
김희영,김동민,류기환,홍호탁 국제문화기술진흥원 2022 International Journal of Advanced Culture Technolo Vol.10 No.1
This study is to select an optimal object detection algorithm for designing a self-checkout counter to improve the inconvenience of payment systems for products without existing barcodes. To this end, a performance comparison analysis of YOLO v2, Tiny YOLO v2, and the latest YOLO v5 among deep learning-based object detection algorithms was performed to derive results. In this paper, performance comparison was conducted by forming learning data as an example of 'donut' in a bakery store, and the performance result of YOLO v5 was the highest at 96.9% of mAP. Therefore, YOLO v5 was selected as the artificial intelligence object detection algorithm to be applied in this paper. As a result of performance analysis, when the optimal threshold was set for each donut, the precision and reproduction rate of all donuts exceeded 0.85, and the majority of donuts showed excellent recognition performance of 0.90 or more. We expect that the results of this paper will be helpful as the fundamental data for the development of an automatic payment system using AI self-service technology that is highly usable in the non-face-to-face era.