{"created":"2021-03-01T06:19:02.041435+00:00","id":2425,"links":{},"metadata":{"_buckets":{"deposit":"aaa6c235-f730-4769-a468-6aa55cc056f7"},"_deposit":{"id":"2425","owners":[],"pid":{"revision_id":0,"type":"depid","value":"2425"},"status":"published"},"_oai":{"id":"oai:repository.dl.itc.u-tokyo.ac.jp:00002425"},"item_7_alternative_title_1":{"attribute_name":"\u305d\u306e\u4ed6\u306e\u30bf\u30a4\u30c8\u30eb","attribute_value_mlt":[{"subitem_alternative_title":"\u63cf\u753b\u30ed\u30dc\u30c3\u30c8 : \u8996\u899a\u306b\u57fa\u3065\u304f\u9ad8\u30ec\u30d9\u30eb\u30d7\u30e9\u30f3\u30cb\u30f3\u30b0"}]},"item_7_biblio_info_7":{"attribute_name":"\u66f8\u8a8c\u60c5\u5831","attribute_value_mlt":[{"bibliographicIssueDates":{"bibliographicIssueDate":"2008-03","bibliographicIssueDateType":"Issued"},"bibliographic_titles":[{}]}]},"item_7_date_granted_25":{"attribute_name":"\u5b66\u4f4d\u6388\u4e0e\u5e74\u6708\u65e5","attribute_value_mlt":[{"subitem_dategranted":"2008-03-24"}]},"item_7_degree_grantor_23":{"attribute_name":"\u5b66\u4f4d\u6388\u4e0e\u6a5f\u95a2","attribute_value_mlt":[{"subitem_degreegrantor":[{"subitem_degreegrantor_name":"University of Tokyo (\u6771\u4eac\u5927\u5b66)"}]}]},"item_7_degree_name_20":{"attribute_name":"\u5b66\u4f4d\u540d","attribute_value_mlt":[{"subitem_degreename":"\u535a\u58eb(\u60c5\u5831\u7406\u5de5\u5b66)"}]},"item_7_description_5":{"attribute_name":"\u6284\u9332","attribute_value_mlt":[{"subitem_description":"Recently, many areas of research on humanoid robots have been studied, such as motion control, man-machine interfaces, artificial intelligence (AI), and so on. Among them many research projects have tried to create artist robots, with the common objective of exploring new sensing, artificial intelligence, and manipulation techniques. The research described in this paper explores new vision and manipulation techniques through painting tasks. The ultimate goal is to create a robot painter that has capabilities similar to those of human artists. Regarding vision, the key problems of 2D/3D object segmentation, color perception, orientation mapping, and geometric edge processing are directly addressed by our method. This research focuses, first, on an effective 2D segmentation scheme using local and global classifiers. Our proposed method can effectively deal with a foreground cut, multiple cuts, and cut before matting. Then it is shown how to exploit normal stereo cameras to roughly extract the object automatically, based on 3D background subtraction and other vision techniques, and how to use our 2D segmentation to extract the object area correctly. The robot must analyze color distribution of the object to select the best set of colors to use. Normally, clustering colors face the tendency to produce colors with low contrast. We solve this problem by incorporating two clustering methods: maximum distance clustering and K-means. Then, in order to draw brush strokes meaningfully, the robot senses the orientation of the object. To smooth the orientation of the whole object, we apply global orientation that exploits the radial basis function to generate a style similar to Van Gogh, for instance, for the entire object. Furthermore, some human artists usually use edges to enhance their paintings. Technically, many researchers use gradient information to represent edges of objects. However, this would be affected by the color information on the surface. Hence, we decided to use 3D geometric edges as an input. We then extract 2D edges from the 3D model. Finally, the 2D edges are processed into brush strokes. We show how to apply these methods to high-level manipulation using a robot platform that consists of two arms and multi-fingered hands. The robot also has a stereo vision system. Based on the derived information, the robot then performs a visual feedback drawing. First, it detects a brush and grabs it using cameras and force sensors. Second, it calculates the position of the brush tip using principal component analysis (PCA). Third, it then draws and compares the canvas with the picture produced by the stereo cameras. Finally, as the trajectories planned by the robot may not be realized on the real robot platform because of its physical limitations, this research presents a method to filter and optimize trajectories targeting offline and online applications. All physical attributes, namely angle, collision, velocity, and dynamic force, are considered as a set of constraints to be met and represented as B-spline coefficients, making the limits guaranteed. The proposed method will be shown to outperform the current methods in the sense of correctness and minimal user interaction, and it does so in a reasonable computation time.","subitem_description_type":"Abstract"}]},"item_7_dissertation_number_26":{"attribute_name":"\u5b66\u4f4d\u6388\u4e0e\u756a\u53f7","attribute_value_mlt":[{"subitem_dissertationnumber":"\u7532\u7b2c23947\u53f7"}]},"item_7_full_name_3":{"attribute_name":"\u8457\u8005\u5225\u540d","attribute_value_mlt":[{"nameIdentifiers":[{"nameIdentifier":"6686","nameIdentifierScheme":"WEKO"}],"names":[{"name":"\u30eb\u30c1\u30e3\u30cc\u30e9\u30c3\u30af, \u30df\u30c6\u30a3"}]}]},"item_7_identifier_registration":{"attribute_name":"ID\u767b\u9332","attribute_value_mlt":[{"subitem_identifier_reg_text":"10.15083/00002419","subitem_identifier_reg_type":"JaLC"}]},"item_7_select_21":{"attribute_name":"\u5b66\u4f4d","attribute_value_mlt":[{"subitem_select_item":"doctoral"}]},"item_7_subject_13":{"attribute_name":"\u65e5\u672c\u5341\u9032\u5206\u985e\u6cd5","attribute_value_mlt":[{"subitem_subject":"548","subitem_subject_scheme":"NDC"}]},"item_7_text_22":{"attribute_name":"\u5b66\u4f4d\u5206\u91ce","attribute_value_mlt":[{"subitem_text_value":"Information Science and Technology (\u60c5\u5831\u7406\u5de5\u5b66)"}]},"item_7_text_24":{"attribute_name":"\u7814\u7a76\u79d1\u30fb\u5c02\u653b","attribute_value_mlt":[{"subitem_text_value":"Department of Information and Communication Engineering, Graduate School of Information Science and Technology (\u60c5\u5831\u7406\u5de5\u5b66\u7cfb\u7814\u7a76\u79d1\u96fb\u5b50\u60c5\u5831\u5b66\u5c02\u653b)"}]},"item_7_text_27":{"attribute_name":"\u5b66\u4f4d\u8a18\u756a\u53f7","attribute_value_mlt":[{"subitem_text_value":"\u535a\u60c5\u7b2c192\u53f7"}]},"item_7_text_4":{"attribute_name":"\u8457\u8005\u6240\u5c5e","attribute_value_mlt":[{"subitem_text_value":"\u5927\u5b66\u9662\u60c5\u5831\u7406\u5de5\u5b66\u7cfb\u7814\u7a76\u79d1\u96fb\u5b50\u60c5\u5831\u5b66\u5c02\u653b"},{"subitem_text_value":"Graduate School of Information Science and Technology Department of Information and Communication Engineering The University of Tokyo"}]},"item_creator":{"attribute_name":"\u8457\u8005","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"Ruchanurucks, Miti"}],"nameIdentifiers":[{"nameIdentifier":"6685","nameIdentifierScheme":"WEKO"}]}]},"item_files":{"attribute_name":"\u30d5\u30a1\u30a4\u30eb\u60c5\u5831","attribute_type":"file","attribute_value_mlt":[{"accessrole":"open_date","date":[{"dateType":"Available","dateValue":"2017-05-31"}],"displaytype":"detail","filename":"48057414.pdf","filesize":[{"value":"5.8 MB"}],"format":"application/pdf","licensetype":"license_note","mimetype":"application/pdf","url":{"label":"48057414.pdf","url":"https://repository.dl.itc.u-tokyo.ac.jp/record/2425/files/48057414.pdf"},"version_id":"063bb619-4f24-4fb8-8b4e-a30bd8d3804c"}]},"item_keyword":{"attribute_name":"\u30ad\u30fc\u30ef\u30fc\u30c9","attribute_value_mlt":[{"subitem_subject":"Robotics","subitem_subject_scheme":"Other"},{"subitem_subject":"Segmentation","subitem_subject_scheme":"Other"},{"subitem_subject":"Color reduction","subitem_subject_scheme":"Other"},{"subitem_subject":"Clustering","subitem_subject_scheme":"Other"},{"subitem_subject":"Orientation","subitem_subject_scheme":"Other"},{"subitem_subject":"Manipulation","subitem_subject_scheme":"Other"},{"subitem_subject":"Trajectory planning","subitem_subject_scheme":"Other"},{"subitem_subject":"B-spline","subitem_subject_scheme":"Other"}]},"item_language":{"attribute_name":"\u8a00\u8a9e","attribute_value_mlt":[{"subitem_language":"eng"}]},"item_resource_type":{"attribute_name":"\u8cc7\u6e90\u30bf\u30a4\u30d7","attribute_value_mlt":[{"resourcetype":"thesis","resourceuri":"http://purl.org/coar/resource_type/c_46ec"}]},"item_title":"Robot Painter : High-Level Planning Based on Visual Perception","item_titles":{"attribute_name":"\u30bf\u30a4\u30c8\u30eb","attribute_value_mlt":[{"subitem_title":"Robot Painter : High-Level Planning Based on Visual Perception"}]},"item_type_id":"7","owner":"1","path":["9/233/280","34/105/330"],"pubdate":{"attribute_name":"\u516c\u958b\u65e5","attribute_value":"2012-03-01"},"publish_date":"2012-03-01","publish_status":"0","recid":"2425","relation_version_is_last":true,"title":["Robot Painter : High-Level Planning Based on Visual Perception"],"weko_creator_id":"1","weko_shared_id":null},"updated":"2021-03-01T19:59:17.656610+00:00"}