{"created":"2023-05-15T12:35:57.871929+00:00","id":5095,"links":{},"metadata":{"_buckets":{"deposit":"b2bd2887-9b7f-45a9-9aa5-805b4447b6d1"},"_deposit":{"created_by":3,"id":"5095","owners":[3],"pid":{"revision_id":0,"type":"depid","value":"5095"},"status":"published"},"_oai":{"id":"oai:nitech.repo.nii.ac.jp:00005095","sets":["31"]},"author_link":["8623","8653","17025","13168"],"item_10001_alternative_title_24":{"attribute_name":"その他(別言語等)のタイトル","attribute_value_mlt":[{"subitem_alternative_title":"キョウカ ガクシュウ ケッカ ノ サイコウチク ヘノ ガイネン ガクシュウ ノ テキヨウ"},{"subitem_alternative_title":"Using Concept Learning for Restructuring Control Policy in Reinforcement Learning"}]},"item_10001_biblio_info_28":{"attribute_name":"書誌情報","attribute_value_mlt":[{"bibliographicIssueDates":{"bibliographicIssueDate":"2002-11-01","bibliographicIssueDateType":"Issued"},"bibliographicIssueNumber":"2","bibliographicPageEnd":"144","bibliographicPageStart":"135","bibliographicVolumeNumber":"17","bibliographic_titles":[{"bibliographic_title":"人工知能学会論文誌 / 人工知能学会"}]}]},"item_10001_description_36":{"attribute_name":"内容記述","attribute_value_mlt":[{"subitem_description":"Conventional reinforcement learning has focused on learning in a stable environment. However, an agent may be given another environment which differs from the old environment. Thus, an autonomous agent needs a method to learn efficiently a new policy suited for the new environment. In this paper, we propose a method to adapt to a new environment for an agent which has a task to reach goals. When an agent is provided with a new environment, our method learns a new partial policy using the precondition of agent’s old policy. The precondition of a policy is a condition that says what must be satisfied in order to reach goals by using the policy. Similarly to learning the precondition of an action from the instances of action’s success or failure by using concept learning, our method learns the precondition of a policy from the instances of policy’s success or failure by using concept learning. We describe a method using inductive logic programming (ILP) as a concept learning method. Since ILP provides methods for learning relational knowledge that is not expressible in attribute-value learning, our method can use relational representation for the precondition. We applied our method to a blocks-world problem for evaluation. We have come to conclusion that our method is effective when the cost to carry out the task is high.","subitem_description_type":"Other"}]},"item_10001_description_38":{"attribute_name":"フォーマット","attribute_value_mlt":[{"subitem_description":"application/pdf","subitem_description_type":"Other"}]},"item_10001_full_name_27":{"attribute_name":"著者別名","attribute_value_mlt":[{"nameIdentifiers":[{},{}],"names":[{"name":"Inuzuka, Nobuhiro"}]},{"nameIdentifiers":[{}],"names":[{"name":"Seki, Hirohisa"}]},{"nameIdentifiers":[{}],"names":[{"name":"Ito, Hidenori"}]}]},"item_10001_publisher_29":{"attribute_name":"出版者","attribute_value_mlt":[{"subitem_publisher":"人工知能学会"}]},"item_10001_relation_34":{"attribute_name":"DOI","attribute_value_mlt":[{"subitem_relation_name":[{"subitem_relation_name_text":"10.1527/tjsai.17.135"}],"subitem_relation_type":"isIdenticalTo","subitem_relation_type_id":{"subitem_relation_type_id_text":"http://dx.doi.org/10.1527/tjsai.17.135","subitem_relation_type_select":"DOI"}}]},"item_10001_source_id_30":{"attribute_name":"ISSN","attribute_value_mlt":[{"subitem_source_identifier":"13460714","subitem_source_identifier_type":"ISSN"}]},"item_10001_source_id_32":{"attribute_name":"書誌レコードID(NCID)","attribute_value_mlt":[{"subitem_source_identifier":"AA11579226","subitem_source_identifier_type":"NCID"}]},"item_10001_version_type_33":{"attribute_name":"著者版フラグ","attribute_value_mlt":[{"subitem_version_resource":"http://purl.org/coar/version/c_970fb48d4fbd8a85","subitem_version_type":"VoR"}]},"item_creator":{"attribute_name":"著者","attribute_type":"creator","attribute_value_mlt":[{"creatorNames":[{"creatorName":"松井, 藤五郎"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"犬塚, 信博"},{"creatorName":"イヌヅカ, ノブヒロ","creatorNameLang":"ja-Kana"}],"nameIdentifiers":[{},{}]},{"creatorNames":[{"creatorName":"世木, 博久"},{"creatorName":"セキ, ヒロヒサ","creatorNameLang":"ja-Kana"}],"nameIdentifiers":[{}]},{"creatorNames":[{"creatorName":"伊藤, 英則"}],"nameIdentifiers":[{}]}]},"item_files":{"attribute_name":"ファイル情報","attribute_type":"file","attribute_value_mlt":[{"accessrole":"open_date","date":[{"dateType":"Available","dateValue":"2017-01-25"}],"displaytype":"detail","filename":"JSAI17_135.pdf","filesize":[{"value":"1.7 MB"}],"format":"application/pdf","license_note":"Copyright c 社団法人人工知能学会","licensetype":"license_note","mimetype":"application/pdf","url":{"label":"本文_fulltext","url":"https://nitech.repo.nii.ac.jp/record/5095/files/JSAI17_135.pdf"},"version_id":"ee920006-6398-41b6-b6ec-e7cf62659456"}]},"item_language":{"attribute_name":"言語","attribute_value_mlt":[{"subitem_language":"jpn"}]},"item_resource_type":{"attribute_name":"資源タイプ","attribute_value_mlt":[{"resourcetype":"journal article","resourceuri":"http://purl.org/coar/resource_type/c_6501"}]},"item_title":"強化学習結果の再構築への概念学習の適用","item_titles":{"attribute_name":"タイトル","attribute_value_mlt":[{"subitem_title":"強化学習結果の再構築への概念学習の適用"}]},"item_type_id":"10001","owner":"3","path":["31"],"pubdate":{"attribute_name":"公開日","attribute_value":"2013-06-25"},"publish_date":"2013-06-25","publish_status":"0","recid":"5095","relation_version_is_last":true,"title":["強化学習結果の再構築への概念学習の適用"],"weko_creator_id":"3","weko_shared_id":3},"updated":"2023-05-15T14:33:55.363987+00:00"}