{"id":3517,"date":"2023-07-03T05:59:00","date_gmt":"2023-07-02T20:59:00","guid":{"rendered":"https:\/\/devneko.jp\/wordpress\/?p=3517"},"modified":"2023-07-03T05:59:00","modified_gmt":"2023-07-02T20:59:00","slug":"%e3%83%9e%e3%83%ab%e3%83%81%e3%83%a2%e3%83%bc%e3%83%80%e3%83%ab%e3%81%aa%e5%a4%a7%e8%a6%8f%e6%a8%a1%e8%a8%80%e8%aa%9e%e3%83%a2%e3%83%87%e3%83%ab%e3%81%ae%e3%82%b5%e3%83%bc%e3%83%99%e3%82%a4","status":"publish","type":"post","link":"https:\/\/devneko.jp\/wordpress\/?p=3517","title":{"rendered":"\u30de\u30eb\u30c1\u30e2\u30fc\u30c0\u30eb\u306a\u5927\u898f\u6a21\u8a00\u8a9e\u30e2\u30c7\u30eb\u306e\u30b5\u30fc\u30d9\u30a4&#038;\u30c1\u30e5\u30fc\u30c8\u30ea\u30a2\u30eb"},"content":{"rendered":"\n<ul class=\"wp-block-list\">\n<li><strong>A Survey on Multimodal Large Language Models&nbsp;<\/strong>[56.8]<br>\u30de\u30eb\u30c1\u30e2\u30fc\u30c0\u30eb\u8a00\u8a9e\u30e2\u30c7\u30eb(MLLM)\u306f\u3001\u30de\u30eb\u30c1\u30e2\u30fc\u30c0\u30eb\u30bf\u30b9\u30af\u3092\u5b9f\u884c\u3059\u308b\u305f\u3081\u306b\u3001\u8133\u3068\u3057\u3066\u5f37\u529b\u306a\u5927\u898f\u6a21\u8a00\u8a9e\u30e2\u30c7\u30eb\u3092\u4f7f\u7528\u3059\u308b\u3002 MLLM\u306e\u9a5a\u304f\u3079\u304d\u5275\u767a\u7684\u80fd\u529b\u3001\u4f8b\u3048\u3070\u753b\u50cf\u306b\u57fa\u3065\u304f\u30b9\u30c8\u30fc\u30ea\u30fc\u306e\u4f5c\u6210\u3084OCR\u306e\u306a\u3044\u6570\u5b66\u63a8\u8ad6\u306f\u3001\u4f1d\u7d71\u7684\u306a\u624b\u6cd5\u3067\u306f\u307e\u308c\u3067\u3042\u308b\u3002<br><a href=\"http:\/\/arxiv.org\/abs\/2306.13549v1\">\u8ad6\u6587<\/a>&nbsp;&nbsp;<a href=\"https:\/\/fugumt.com\/fugumt\/paper_check\/2306.13549v1\">\u53c2\u8003\u8a33\uff08\u30e1\u30bf\u30c7\u30fc\u30bf\uff09<\/a>&nbsp; &nbsp;(Fri, 23 Jun 2023 15:21:52 GMT)<\/li>\n\n\n\n<li>\u30de\u30eb\u30c1\u30e2\u30fc\u30c0\u30eb\u306a\u5927\u898f\u6a21\u8a00\u8a9e\u30e2\u30c7\u30eb\u306e\u30b5\u30fc\u30d9\u30a4\u3002Multimodal Instruction Tuning (MIT)\u3001Multimodal In-Context Learning (M-ICL)\u3001 Multimodal Chain-of-Thought (M-CoT)\u3001LLM-Aided Visual Reasoning (LAVR)\u306e\u30ab\u30c6\u30b4\u30ea\u3067\u6574\u7406\u3002LLM\u3092\u4e2d\u5fc3\u306b\u69d8\u3005\u306a\u30c8\u30e9\u30a4\u304c\u3055\u308c\u3066\u3044\u308b\u3053\u3068\u304c\u5206\u304b\u308b\u3002<\/li>\n\n\n\n<li>\u30ea\u30dd\u30b8\u30c8\u30ea\u306f<a href=\"https:\/\/github.com\/BradyFU\/Awesome-Multimodal-Large-Language-Models\">GitHub &#8211; BradyFU\/Awesome-Multimodal-Large-Language-Models: :sparkles::sparkles:Latest Papers and Datasets on Multimodal Large Language Models, and Their Evaluation.<\/a>\u3067\u3001\u3053\u306e\u8ad6\u6587\u30ea\u30b9\u30c8\u3082\u76f8\u5f53\u4fa1\u5024\u304c\u9ad8\u3044\u3002<\/li>\n<\/ul>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Large Multimodal Models: Notes on CVPR 2023 Tutorial&nbsp;<\/strong>[29.8]<br>\u3053\u306e\u30c1\u30e5\u30fc\u30c8\u30ea\u30a2\u30eb\u30ce\u30fc\u30c8\u306f\u3001CVPR 2023 tutorial on recent Advances in Vision Foundation Models&#8217; \u306e\u4e00\u90e8\u3067\u3042\u308b\u3002 \u8996\u899a\u30fb\u8a00\u8a9e\u30e2\u30c7\u30ea\u30f3\u30b0\u306e\u305f\u3081\u306e\u6700\u8fd1\u306eGPT\u306e\u3088\u3046\u306a\u5927\u898f\u6a21\u30e2\u30c7\u30eb\u306b\u3064\u3044\u3066,\u307e\u305a\u305d\u306e\u80cc\u666f\u3092\u7d39\u4ecb\u3059\u308b\u3002 \u524d\u63d0\u6761\u4ef6\u3068\u3057\u3066,\u5927\u898f\u6a21\u8a00\u8a9e\u30e2\u30c7\u30eb\u306b\u304a\u3051\u308b\u30a4\u30f3\u30b9\u30c8\u30e9\u30af\u30b7\u30e7\u30f3\u30c1\u30e5\u30fc\u30cb\u30f3\u30b0\u306e\u57fa\u790e\u306b\u3064\u3044\u3066\u8ff0\u3079\u308b\u3002 \u6700\u5f8c\u306b\u3001\u30aa\u30fc\u30d7\u30f3\u30bd\u30fc\u30b9\u30ea\u30bd\u30fc\u30b9\u3092\u7528\u3044\u305f\u30de\u30eb\u30c1\u30e2\u30fc\u30c0\u30ebGPT-4\u306e\u3088\u3046\u306a\u30e2\u30c7\u30eb\u306e\u6700\u5c0f\u9650\u306e\u30d7\u30ed\u30c8\u30bf\u30a4\u30d7\u3092\u69cb\u7bc9\u3059\u308b\u65b9\u6cd5\u306b\u3064\u3044\u3066\u8aac\u660e\u3059\u308b\u3002<br><a href=\"http:\/\/arxiv.org\/abs\/2306.14895v1\">\u8ad6\u6587<\/a>&nbsp;&nbsp;<a href=\"https:\/\/fugumt.com\/fugumt\/paper_check\/2306.14895v1\">\u53c2\u8003\u8a33\uff08\u30e1\u30bf\u30c7\u30fc\u30bf\uff09<\/a>&nbsp; &nbsp;(Mon, 26 Jun 2023 17:59:31 GMT)<\/li>\n\n\n\n<li>CVPR\u306e\u30de\u30eb\u30c1\u30e2\u30fc\u30c0\u30eb\u30e2\u30c7\u30eb\u306e\u30c1\u30e5\u30fc\u30c8\u30ea\u30a2\u30eb<\/li>\n\n\n\n<li>\u30b9\u30e9\u30a4\u30c9\uff1ahttps:\/\/tinyurl.com\/5c2c2mtm\u3001\u52d5\u753b<a href=\"https:\/\/www.youtube.com\/watch?v=mkI7EPD1vp8\">[CVPR2023 Tutorial Talk] Large Multimodal Models: Towards Building and Surpassing Multimodal GPT-4 &#8211; YouTube<\/a>\u306a\u3069\u3068\u3066\u3082\u6709\u7528<\/li>\n<\/ul>\n","protected":false},"excerpt":{"rendered":"","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[2],"tags":[223,387,524],"class_list":["post-3517","post","type-post","status-publish","format-standard","hentry","category-arxiv","tag-llm","tag-survey","tag-524"],"_links":{"self":[{"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/posts\/3517","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=3517"}],"version-history":[{"count":0,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/posts\/3517\/revisions"}],"wp:attachment":[{"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=3517"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=3517"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=3517"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}