{"id":884,"date":"2021-10-21T04:00:00","date_gmt":"2021-10-20T19:00:00","guid":{"rendered":"https:\/\/devneko.jp\/wordpress\/?p=884"},"modified":"2021-10-21T04:00:00","modified_gmt":"2021-10-20T19:00:00","slug":"ml-defense-models-competition","status":"publish","type":"post","link":"https:\/\/devneko.jp\/wordpress\/?p=884","title":{"rendered":"ML Defense Models Competition"},"content":{"rendered":"\n<ul class=\"wp-block-list\"><li><strong>Adversarial Attacks on ML Defense Models Competition\u00a0<\/strong>[82.4]<br>\u6e05\u83ef\u5927\u5b66\u306eTSAIL\u30b0\u30eb\u30fc\u30d7\u3068Alibaba Security\u30b0\u30eb\u30fc\u30d7\u304c\u3053\u306e\u7af6\u4e89\u3092\u7d44\u7e54\u3057\u305f\u3002 \u3053\u306e\u7af6\u4e89\u306e\u76ee\u7684\u306f\u3001\u6575\u306e\u5805\u7262\u6027\u3092\u8a55\u4fa1\u3059\u308b\u305f\u3081\u306b\u3001\u65b0\u3057\u3044\u653b\u6483\u30a2\u30eb\u30b4\u30ea\u30ba\u30e0\u3092\u52d5\u6a5f\u4ed8\u3051\u308b\u3053\u3068\u3067\u3042\u308b\u3002<br><a href=\"http:\/\/arxiv.org\/abs\/2110.08042v1\">\u8ad6\u6587<\/a>\u00a0\u00a0<a href=\"https:\/\/fugumt.com\/fugumt\/paper_check\/2110.08042v1\">\u53c2\u8003\u8a33\uff08\u30e1\u30bf\u30c7\u30fc\u30bf\uff09<\/a>\u00a0 \u00a0(Fri, 15 Oct 2021 12:12:41 GMT)<ul><li>\u6e05\u83ef\u5927\u5b66\u3068Alibaba\u30bb\u30ad\u30e5\u30ea\u30c6\u30a3\u30b0\u30eb\u30fc\u30d7\u306b\u3088\u308b<a href=\"https:\/\/aisecure-workshop.github.io\/amlcvpr2021\/\">CVPR 2021 workshop on adversarial machine learning<\/a>\u00a0\u306e\u5831\u544a\u3002\u5404\u30c1\u30fc\u30e0\u306e\u30a2\u30d7\u30ed\u30fc\u30c1\u306e\u6982\u8981\u304c\u53c2\u8003\u306b\u306a\u308b\u3002<\/li><li>\u30d7\u30ed\u30b8\u30a7\u30af\u30c8\u30b5\u30a4\u30c8\u306f<a href=\"https:\/\/ml.cs.tsinghua.edu.cn\/ares-bench\/\">https:\/\/ml.cs.tsinghua.edu.cn\/ares-bench\/<\/a><\/li><\/ul><\/li><\/ul>\n","protected":false},"excerpt":{"rendered":"<p>Adversarial Attacks on ML Defense Models Competition\u00a0[82.4]\u6e05\u83ef\u5927\u5b66\u306eTSAIL\u30b0\u30eb\u30fc\u30d7\u3068Alibaba Security\u30b0\u30eb\u30fc\u30d7\u304c\u3053\u306e\u7af6\u4e89\u3092\u7d44\u7e54\u3057\u305f\u3002 \u3053\u306e\u7af6\u4e89 &hellip; <a href=\"https:\/\/devneko.jp\/wordpress\/?p=884\" class=\"more-link\"><span class=\"screen-reader-text\">&#8220;ML Defense Models Competition&#8221; \u306e<\/span>\u7d9a\u304d\u3092\u8aad\u3080<\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[2],"tags":[11,517],"class_list":["post-884","post","type-post","status-publish","format-standard","hentry","category-arxiv","tag-adversarial-attack","tag-517"],"_links":{"self":[{"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/posts\/884","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=884"}],"version-history":[{"count":0,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=\/wp\/v2\/posts\/884\/revisions"}],"wp:attachment":[{"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=884"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=884"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/devneko.jp\/wordpress\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=884"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}