
Please use this identifier to cite or link to this item:
http://thuvienso.dut.udn.vn/handle/DUT/5553
DC Field | Value | Language |
---|---|---|
dc.contributor.advisor | TS. Nguyễn, Quang Như Quỳnh | en_US |
dc.contributor.advisor | Prof. Wen-Nung Lie | en_US |
dc.contributor.author | Nguyễn, Hữu Thắng | en_US |
dc.date.accessioned | 2025-02-18T03:25:26Z | - |
dc.date.available | 2025-02-18T03:25:26Z | - |
dc.date.issued | 2019 | - |
dc.identifier.uri | http://thuvienso.dut.udn.vn/handle/DUT/5553 | - |
dc.description | 37 Tr. | en_US |
dc.description.abstract | This thesis is “3D Human Skeleton Extraction for Action Analysis”. We want to generate an accurate 3D pose estimation from a single view RGB image to understand more about people in images and videos. The 3D human pose estimation algorithm contained two steps. First, we put all of the 3D Human samples into K-means algorithm, and take the mean pose in each cluster as anchor pose ground truth in training part. Second, we trained first model called “3D anchor pose estimator”, which use 2D human pose as input ground truth, and the result of clustering as output ground truth. Anchor poses are some common human posture in our daily life. Final we train second model called “3D human pose estimator” and combine 2D human pose and 3D anchor pose to estimate final 3D human pose. In this work, we show a systematic design for how Fully Convolutional Network (FCNs) can be incorporated for the task of pose estimation. Our deep learning network contains two-stage: The first network (FCN1) estimate a 3D Anchor Pose from the 2D skeleton, then pass the results to the second stage (FCN2) to further regress/refine the 3D Anchor Pose to yield the final 3D human pose estimation. According to the experiments, our two-stage FCN network can generate a 3D human pose with an average MPJPE (Mean per Joint Position Error) of 62.99 mm when a 2D skeleton prediction is used as the input. The 2D skeleton predictions are produced by a pre-trained model called Stacked Hourglass. Stateof-the-art results are achieved on the H36M benchmark | en_US |
dc.language.iso | en | en_US |
dc.publisher | Trường Đại học Bách khoa - Đại học Đà Nẵng | en_US |
dc.subject | 3D Human Pose Estimation | en_US |
dc.subject | Fully Convolutional Network | en_US |
dc.subject | Deep Learning | en_US |
dc.title | 3D human skeleton extraction for action analysis | en_US |
dc.type | Đồ án | en_US |
dc.identifier.id | 2.DA.FA.19.010 | - |
item.cerifentitytype | Publications | - |
item.languageiso639-1 | en | - |
item.openairecristype | http://purl.org/coar/resource_type/c_18cf | - |
item.fulltext | Có toàn văn | - |
item.openairetype | Đồ án | - |
item.grantfulltext | restricted | - |
Appears in Collections: | DA.Điện tử - Viễn thông |
Files in This Item:
File | Description | Size | Format | Existing users please Login |
---|---|---|---|---|
2.DA.FA.19.010.Nguyen Huu Thang.pdf | Thuyết minh | 8.39 MB | Adobe PDF | ![]() |
CORE Recommender
Page view(s) 50
9
checked on May 9, 2025
Download(s) 50
5
checked on May 9, 2025
Google ScholarTM
Check
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.