{"created":"2023-07-25T10:21:27.777025+00:00","id":473,"links":{},"metadata":{"_buckets":{"deposit":"d76d352a-d8e9-43c6-ab3b-4d953d3d2128"},"_deposit":{"created_by":3,"id":"473","owners":[3],"pid":{"revision_id":0,"type":"depid","value":"473"},"status":"published"},"_oai":{"id":"oai:air.repo.nii.ac.jp:00000473","sets":["590:591:592:593"]},"author_link":["1336","1332","1334","1337","1333","1335"],"item_10002_alternative_title_34":{"attribute_name":"別タイトル","attribute_value_mlt":[{"subitem_alternative_title":"<Original Papers>Extraction of Motion Primitives and Automated Buyo-fu Assignment"}]},"item_10002_biblio_info_36":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicIssueDates":{"bibliographicIssueDate":"2002-10-31","bibliographicIssueDateType":"Issued"},"bibliographicPageEnd":"40","bibliographicPageStart":"33","bibliographicVolumeNumber":"23","bibliographic_titles":[{"bibliographic_title":"秋田大学工学資源学部 研究報告"}]}]},"item_10002_description_29":{"attribute_name":"内容記述(抄録)","attribute_value_mlt":[{"subitem_description":"We have proposed a human motion description method using Buyo-fu, which aims at establishing a new recording method of human motions and re-using a time series of three-dimensional human motion data(mocap data) obtained by using a motion capturing system. The mocap data can be partitioned into some basic motions. We assign a code to each basic motion, and call it Buyo-fu. We can efficiently describe the human motions by using Buyo-fu. At present, we have a problem that it takes much time and effort to make Buyo-fu, because only a person who has professional knowledge about the concerned motions can extract Buyo-fu, In this paper, we propose a method which can automatically extract Buyo-fu from the mocap data, In our proposed method, 1) a series of the human motions is partitioned into the basic motions based on the speed of the motions, 2) since we consider that similar basic motions appear several times in a series of motions, we do clustering the obtained basic motions, 3) we assign a label to each clustered basic motion, and call it a primitive motion, 4) we can express a series of motions by means of the concatenation of the primitive motions, 5) when we can find out the same pattern of the primitive motions in the human motions, we define this pattern to be a new primitive. Each primitive motion which can be obtained using our proposed method can become Buyo-fu. Performing the experiment with the typical form of \"tensho\" and \"sanchin\" of Karate-do, we show the proposed method is a good candidate for automatically extracting Buyo-fu from the mocap data.","subitem_description_type":"Other"}]},"item_10002_publisher_30":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"秋田大学工学資源学部"}]},"item_10002_source_id_27":{"attribute_name":"ISSN","attribute_value_mlt":[{"subitem_source_identifier":"13457241","subitem_source_identifier_type":"ISSN"}]},"item_10002_source_id_35":{"attribute_name":"NCID","attribute_value_mlt":[{"subitem_source_identifier":"AA11410906","subitem_source_identifier_type":"NCID"}]},"item_10002_text_25":{"attribute_name":"著者所属","attribute_value_mlt":[{"subitem_text_value":"秋田経済法科大学短期大学部"},{"subitem_text_value":"秋田大学工学資源学部"},{"subitem_text_value":"秋田大学工学資源学部"}]},"item_10002_version_type_37":{"attribute_name":"著者版フラグ","attribute_value_mlt":[{"subitem_version_resource":"http://purl.org/coar/version/c_970fb48d4fbd8a85","subitem_version_type":"VoR"}]},"item_creator":{"attribute_name":"著者","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"湯川, 崇"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"小原, 直子"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"玉本, 英夫"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"YUKAWA, Takashi","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"OBARA, Naoko","creatorNameLang":"en"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"TAMAMOTO, Hideo","creatorNameLang":"en"}],"nameIdentifiers":[{}]}]},"item_files":{"attribute_name":"ファイル情報","attribute_type":"file","attribute_value_mlt":[{"accessrole":"open_date","date":[{"dateType":"Available","dateValue":"2017-02-16"}],"displaytype":"detail","filename":"KJ00000048348.pdf","filesize":[{"value":"748.0 kB"}],"format":"application/pdf","licensetype":"license_note","mimetype":"application/pdf","url":{"label":"KJ00000048348.pdf","url":"https://air.repo.nii.ac.jp/record/473/files/KJ00000048348.pdf"},"version_id":"739e8bed-7ae3-467f-ab0b-1dda34e0f6d9"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"jpn"}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourcetype":"departmental bulletin paper","resourceuri":"http://purl.org/coar/resource_type/c_6501"}]},"item_title":"<論文>動作プリミティブ抽出と舞踊符割当ての自動化","item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"<論文>動作プリミティブ抽出と舞踊符割当ての自動化"}]},"item_type_id":"10002","owner":"3","path":["593"],"pubdate":{"attribute_name":"公開日","attribute_value":"2007-12-14"},"publish_date":"2007-12-14","publish_status":"0","recid":"473","relation_version_is_last":true,"title":["<論文>動作プリミティブ抽出と舞踊符割当ての自動化"],"weko_creator_id":"3","weko_shared_id":-1},"updated":"2023-07-25T12:08:13.489978+00:00"}