{"id":6261,"date":"2025-02-28T05:19:00","date_gmt":"2025-02-27T20:19:00","guid":{"rendered":"https:\/\/devneko.jp\/wordpress\/?p=6261"},"modified":"2025-02-28T05:19:00","modified_gmt":"2025-02-27T20:19:00","slug":"mme-cot-benchmarking-chain-of-thought-in-large-multimodal-models-for-reasoning-quality-robustness-and-efficiency","status":"publish","type":"post","link":"https:\/\/devneko.jp\/wordpress\/?p=6261","title":{"rendered":"MME-CoT: Benchmarking Chain-of-Thought in Large Multimodal Models for Reasoning Quality, Robustness, and Efficiency\u00a0"},"content":{"rendered":"\n<ul class=\"wp-block-list\">\n<li><strong>MME-CoT: Benchmarking Chain-of-Thought in Large Multimodal Models for Reasoning Quality, Robustness, and Efficiency\u00a0<\/strong>[63.2]<br>CoT (Chain-of-Thought) \u306f,Large Language Models (LLMs) \u306e\u63a8\u8ad6\u80fd\u529b\u3092\u5927\u5e45\u306b\u5411\u4e0a\u3055\u305b\u305f\u3002 \u6211\u3005\u306f,LMM\u306eCoT\u63a8\u8ad6\u6027\u80fd\u3092\u8a55\u4fa1\u3059\u308b\u7279\u5225\u30d9\u30f3\u30c1\u30de\u30fc\u30af\u3067\u3042\u308bMME-CoT\u3092\u7d39\u4ecb\u3059\u308b\u3002 \u6211\u3005\u306f\u6700\u5148\u7aef\u306eLMM\u306e\u8a73\u7d30\u306a\u5206\u6790\u3092\u884c\u3044\u3001\u3044\u304f\u3064\u304b\u306e\u91cd\u8981\u306a\u77e5\u898b\u3092\u660e\u3089\u304b\u306b\u3057\u305f\u3002<br><a href=\"http:\/\/arxiv.org\/abs\/2502.09621v1\">\u8ad6\u6587<\/a>\u00a0\u00a0<a href=\"https:\/\/fugumt.com\/fugumt\/paper_check\/2502.09621v1\">\u53c2\u8003\u8a33\uff08\u30e1\u30bf\u30c7\u30fc\u30bf\uff09<\/a>\u00a0 \u00a0(Thu, 13 Feb 2025 18:59:46 GMT)<\/li>\n\n\n\n<li>\u300cwe introduce MMECoT, a specialized benchmark evaluating the CoT reasoning performance of LMMs, spanning six domains: math, science, OCR, logic, space-time, and general scenes.\u300d\u3068\u3044\u3046\u30d9\u30f3\u30c1\u30de\u30fc\u30af<\/li>\n\n\n\n<li>\u30d7\u30ed\u30b8\u30a7\u30af\u30c8\u30b5\u30a4\u30c8\u306f<a href=\"https:\/\/mmecot.github.io\/\">MME-CoT: Benchmarking Chain-of-Thought in Large Multimodal Models for Reasoning Quality, Robustness, and Efficiency<\/a>\u3001Leaderboard\u30c8\u30c3\u30d7\u304cKimi k1.5\u3067GPT-4o\u3092\u8d85\u3048\u3066\u3044\u308b\u3068\u3044\u3046\u9a5a\u304d\u306e\u7d50\u679c\u3002<\/li>\n<\/ul>\n","protected":false},"excerpt":{"rendered":"","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[2],"tags":[59,517,524],"class_list":["post-6261","post","type-post","status-publish","format-standard","hentry","category-arxiv","tag-chain-of-thought","tag-517","tag-524"],"_links":{"self":[{"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/posts\/6261","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=6261"}],"version-history":[{"count":0,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/posts\/6261\/revisions"}],"wp:attachment":[{"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=6261"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=6261"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=6261"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}