¿­·¢Óé·¢K8com



¿­·¢Óé·¢K8com¿Æ¼¼

¿­·¢Óé·¢K8com



ÓëÌÚѶ¡¢¹þ¹¤´ǫ́ͬ¾º¼¼£¬¿­·¢Óé·¢K8com»ñ×ÔÈ»ÓïÑÔ´¦ÀíÁìÓò¹ú¼Ê¶¥»áNAACL2021¹Ú¾ü

2021-06-08

2021Äê6ÔÂ6ÈÕ-11ÈÕ£¬×ÔÈ»ÓïÑÔ´¦Àí£¨NLP£©ÁìÓòµÄ¶¥¼¶»áÒéNAACLÔÚÏßÉϾٰì¡£¿­·¢Óé·¢K8com¿Æ¼¼DeepBlueAIÍŶӲμÓÁËMulti-Hop Inference Explanation Regeneration (TextGraphs-15) ¹²ÏíÈÎÎñ±ÈÈü£¬²¢»ñµÃÁ˵ÚÒ»£¬¸Ã·½°¸¶àÓÃÓÚ¿Æѧ֪ʶÎÊ´ðµÈÁìÓò¡£Í¬ÈüµÀ¾º¼¼µÄ»¹ÓÐÌÚѶ¡¢¹þ¶û±õ¹¤Òµ´óѧ×é³ÉµÄÍŶÓÒÔ¼°Ð¼ÓÆ¿Ƽ¼Éè¼Æ´óѧÍŶӵÈ¡£

undefined

ͼ1 ³É¼¨ÅÅÃû


NAACLÈ«³ÆΪ Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies£¬ÓëACL¡¢EMNLP²¢³ÆNLPÁìÓòµÄÈý´ó¶¥»á¡£

¹Ú¾ü·½°¸·ÖÏí

ÈÎÎñ½éÉÜ

¶àÌõÍÆÀí£¨Multi-Hop Inference£©ÈÎÎñÊǽáºÏ¶àÌõÐÅϢȥ½â¾öÍÆÀíÈÎÎñ£¬Èç¿ÉÒÔ´ÓÊéÖлòÕßÍøÂçÉÏÑ¡ÔñÓÐÓõľä×Ó£¬»òÕß¼¯ºÏһЩ֪ʶ¿âµÄ֪ʶȥ»Ø´ðËûÈËÌá³öµÄÎÊÌâ¡£ÈçÏÂͼËùʾ£¬ÈçÐè»Ø´ðµ±Ç°ÎÊÌ⣬Ҫ½áºÏͼÖÐËùʾÈýÖÖÐÅÏ¢²ÅÄÜÍê³ÉÍÆÀí£¬µÃµ½ÕýÈ·µÄ´ð°¸¡£¶ø½âÊÍÔÙÉú£¨Explanation Regeneration£©ÈÎÎñÊǶàÌõÍÆÀíÈÎÎñµÄ»ù´¡£¬ÆäÄ¿µÄÊǹ¹½¨¿ÆѧÎÊÌâµÄ½âÊÍ£¬Ã¿¸ö½âÊͶ¼±íʾΪһ¸ö¡°½âÊÍͼ¡±£¬Ò»×éÔ­×ÓÊÂʵ£¨Ã¿¸ö½âÊÍ°üº¬1-16¸ö£¬´Ó9000¸öÊÂʵµÄ֪ʶ¿âÖÐÌáÈ¡£©£¬ËüÃÇÒ»Æð¹¹³ÉÁ˶ԻشðºÍ½âÊÍÎÊÌâ½øÐÐÍÆÀí½âÎöµÄÏêϸ½âÊÍ¡£

undefined

ͼ2 ÈÎÎñʾÀý

¶ÔÓÚµ±Ç°ÈÎÎñ£¬¾Ù°ì·½½«Æ䶨ÒåΪһ¸öÅÅÐòÈÎÎñ£¬ÊäÈëµÄÊÇÎÊÌâ¼°Æä¶ÔÓ¦µÄÕýÈ·´ð°¸£¬ÒªÇóϵͳÄܹ»¶ÔÌṩµÄ°ë½á¹¹»¯ÖªÊ¶¿âÖеÄÔ­×ÓÊÂʵ½âÊͽøÐÐÅÅÐò£¬ÒÔ±ãÅÅÃû¿¿Ç°µÄÔ­×ÓÊÂʵ½âÊÍÄܹ»Îª´ð°¸Ìṩ¸ü¼ÓÏêϸºÍÈ·ÇеÄ˵Ã÷¡£


Êý ¾Ý

´Ë¹²ÏíÈÎÎñÖÐʹÓõÄÊý¾Ý°üº¬´Ó AI2 ÍÆÀíÌôÕ½ (ARC) Êý¾Ý¼¯ÖÐÌáÈ¡µÄ´óÔ¼ 5,100 µÀ¿Æѧ¿¼ÊÔÌ⣬ÒÔ¼°´Ó WorldTree V2.1[2] ½âÊÍÖÐÌáÈ¡µÄÕýÈ·´ð°¸µÄÊÂʵ½âÊÍÓïÁϿ⣬²¢ÔÚ´Ë»ù´¡ÉÏÔö¼ÓÁËר¼ÒÉú³ÉµÄÏà¹ØÐÔÆÀ¼¶¡£Ö§³ÖÕâЩÎÊÌâ¼°Æä½âÊ͵Ä֪ʶ¿â°üº¬´óÔ¼ 9,000 ¸öÊÂʵ£¬ÖªÊ¶¿â¼È¿ÉÒÔ×÷Ϊ´¿Îı¾¾ä×Ó£¨·Ç½á¹¹»¯£©Ò²¿ÉÒÔ×÷Ϊ°ë½á¹¹»¯±í¸ñʹÓá£

·½ °¸

¸ÃÈÎÎñΪһ¸öÅÅÐòÈÎÎñ£¬¾ßÌå±íÏÖΪ¸ø¶¨ÎÊÌâºÍ´ð°¸£¬½«ÖªÊ¶¿âÖеÄ9,000¸öÔ­×ÓÊÂʵ½âÊͽøÐÐÅÅÐò£¬ÆÀ¼Û·½Ê½ÎªNDCG¡£·½°¸Ö÷ÒªÓÉÕٻغÍÅÅÐòÁ½²¿·Ö×é³É£¬µÚÒ»²½ÏÈÕÙ»ØTop-K£¨k> 100£©¸ö½âÊÍ£¬µÚ¶þ²½¶ÔÕٻصÄTop-K¸ö½âÊͽøÐÐÅÅÐò¡£Õë¶ÔÕٻغÍÅÅÐòÈÎÎñ£¬Èç¹ûÖ±½Ó²ÉÓà Interaction-Based£¨½»»¥ÐÍ£¬¼´ÎÊÌâÎı¾ºÍÊÂʵ½âÊÍÔÚÄ£ÐÍÖдæÔÚÐÅÏ¢½»»¥£© ÀàÐ͵ÄÍøÂ磬¼ÆËãÁ¿½«¾Þ´ó£¬Òò´Ë½»»¥ÐÍÍøÂçÔÚµ±Ç°µÄÈÎÎñÖÐÎÞ·¨Ê¹Óã¬ÍŶÓ×îÖÕ²ÉÓÃÁËÏòÁ¿»¯¼ìË÷µÄ·½Ê½½øÐÐÅÅÐò¡£

ΪÁËÌáÈ¡¸üÉîµÄÓïÒåÐÅÏ¢Éú³É±È½ÏºÃµÄÌØÕ÷ÏòÁ¿£¬ÍŶÓûÓвÉÓÃTF-IDF¡¢BM25¡¢DSSM[3]µÈ³£ÓõÄÅÅÐòÄ£ÐÍ£¬¶øÊDzÉÓÃÁ˵±Ç°±È½ÏÁ÷ÐеÄԤѵÁ·Ä£ÐÍ×÷ΪÌØÕ÷ÌáÈ¡Æ÷£¬½áºÏTriplet loss[4]ѵÁ·ÁËÒ»¸öTriplet NetworkÀ´Íê³ÉÏòÁ¿»¯ÅÅÐò£¬ÆäÖÐÔÚÕٻز¿·ÖºÍÅÅÐò²¿·Ö¾ù²ÉÓÃTriplet Network¡£

undefined

ͼ3 Triplet loss

undefined


Ä£ ÐÍ

Õë¶Ôµ±Ç°ÈÎÎñ£¬ÈçÏÂͼËùʾ£¬Ãªµã£¨Anchor£©Ñù±¾ÎªÎÊÌâºÍ´ð°¸Á¬½ÓµÄÎı¾£¬ÕýÑù±¾£¨Positive£©ÎªÎÊÌâ¶ÔÓ¦µÄ½âÊÍÎı¾£¬¸ºÑù±¾£¨Negative£©ÎªÆäËûËæ»úÑ¡ÔñÓëÕýÑù±¾²»Í¬µÄ½âÊÍÎı¾£¬ÆäÖÐËûÃÇÈý¸öÊäÈë¹²ÏíÒ»Ì×ԤѵÁ·ÓïÑÔÄ£ÐÍ£¨Pre-trained language model £ºPLM£©²ÎÊý¡£ÑµÁ·Ê±½«ÉÏÊöÈý¸öÎı¾ÊäÈëµ½PLMÄ£ÐÍÖУ¬Ñ¡È¡PLMÄ£Ð͵ÄËùÓÐToken ÏòÁ¿Æ½¾ù×÷ΪÊä³ö£¬½«Èý¸öÊäÈëÏòÁ¿¾­¹ýTriplet Loss µÃµ½ËðʧֵÍê³ÉÄ£Ð͵ÄѵÁ·¡£

 
undefined

ͼ4 Ä£ÐÍͼ


¸º ²É Ñù

ΪÁ˸üºÃµØѵÁ·Ä£ÐÍ£¬ÍŶÓÔÚÕٻؽ׶βÉÓÃÁËÈýÖÖ²»Í¬µÄ¸º²ÉÑù·½Ê½£º

È«¾ÖËæ»ú¸º²ÉÑù£¬¼´ÔÚ9,000¸ö½âÊÍÎı¾ÖÐËæ»úÑ¡È¡Ò»¸ö²»ÊÇÕýÑù±¾µÄÑù±¾×÷Ϊ¸ºÑù±¾;

BatchÄÚ¸º²ÉÑù£¬¼´ÔÚµ±Ç°BatchÄÚÑ¡È¡ÆäËûÎÊÌâµÄ¶ÔÓ¦µÄ½âÊÍÕýÑù±¾£¬×÷Ϊµ±Ç°ÎÊÌâµÄ¸ºÑù±¾;

Ïà½üÑù±¾¸º²ÉÑù£¬ÔÚͬһ¸ö±íÖÐËæ»úÑ¡È¡Ò»¸öÑù±¾×÷Ϊ¸ºÑù±¾£¬ÒòΪͬһ¸ö±íÖеÄÑù±¾±È½ÏÏà½ü¡£


ÔÚÅÅÐò½×¶ÎͬÑù²ÉÈ¡ÁËÈýÖÖ²»Í¬µÄ¸º²ÉÑù·½Ê½£º


Top-K Ëæ»ú¸º²ÉÑù£¬¼´ÔÚÔÚÕٻصÄTop-K¸öÑù±¾ÖÐËæ»úÑ¡È¡Ò»¸ö¸ºÑù±¾£»

BatchÄÚ¸º²ÉÑù£¬ºÍÕٻؽ׶ÎÏàͬ£»

Top-N Ëæ»ú¸º²ÉÑù£¬ÎªÁËÇ¿»¯Ç°ÃæһЩÑù±¾µÄÅÅÐòЧ¹û£¬Ôö´óÁËÇ°ÃæN¸öÑù±¾µÄ²ÉÑù¸ÅÂÊ£¨NԶԶСÓÚK£©¡£


ʵ Ñé

ÍŶӲÉÓÃÁËÁ½ÖÖԤѵÁ·Ä£ÐÍRoBERTa[5] ºÍERNIE 2.0[6]£¬²¢½«Á½¸öÄ£Ð͵ÄÔ¤²â½á¹û½øÐÐÁËÈںϡ£ÔÚÕٻغÍÅÅÐò½×¶Î£¬²ÉÓÃÁËͬÑùµÄ²ÎÊý£¬Ö÷Òª²ÎÊýÈç²ÉÓÃÈýÖÖ¸º²ÉÑù·½Ê½£¬Ã¿ÖÖ¸º²ÉÑù·½Ê½Ñ¡È¡16¸öÑù±¾£¬×îÖÕµÄbatch sizeΪ48£¬epochΪ15¡£Í¬Ê±£¬Ê¹ÓÃÁËAdamÓÅ»¯Æ÷²¢²ÉÓÃÁËѧϰÂÊË¥¼õ²ßÂÔ£¬´Ó1e-5Ë¥¼õµ½0¡£

ÍŶӷֱðÆÀ²âÁËNDCG @100¡¢NDCG @500¡¢NDCG @1000¡¢NDCG @2000µÄ½á¹û£¬×îÖÕЧ¹ûÈçϱíËùʾ£¬ÆäÖÐBaselineΪTFIDFÄ£ÐÍ¡¢RecallΪÕٻؽ׶Ρ¢Re-rankerΪÕë¶ÔÕٻصĽá¹ûÖØÐÂÅÅÐòµÄ½á¹û¡£´Ó±íÖпÉÒÔ¿´³ö»ùÓÚԤѵÁ·Ä£Ð͵ķ½·¨¶Ô±ÈBaselineÓÐןܴóµÄÌáÉý£¬Í¬Ê±ÖØÅÅÒ²ÓÐ×ÅÏÔÖøµÄÌáÉý£¬Í¬Ê±´ÓÅÅÐаñÖпÉÒÔ¿´³öDeepBlueAIÍŶӵÄÄ£ÐͶԱÈËûÈËÒ²ÓÐ׎ϴóµÄÁìÏÈ¡£

undefined

Table 1 The final results compared with different models



²Î¿¼ÎÄÏ×

[1] Clark P, Cowhey I, Etzioni O, et al. Think you have solved question answering? try arc, the ai2 reasoning challenge[J]. arXiv preprint arXiv:1803.05457, 2018.

[2] Xie Z, Thiem S, Martin J, et al. Worldtree v2: A corpus of science-domain structured explanations and inference patterns supporting multi-hop inference[C]//Proceedings of The 12th Language Resources and Evaluation Conference. 2020: 5456-5473.

[3] Huang, Po-Sen, et al. "Learning deep structured semantic models for web search using clickthrough data." *Proceedings of the 22nd ACM international conference on Information & Knowledge Management*. 2013.

[4] Schroff, Florian, Dmitry Kalenichenko, and James Philbin. "Facenet: A unified embedding for face recognition and clustering." *Proceedings of the IEEE conference on computer vision and pattern recognition*. 2015.

[5] Liu Y, Ott M, Goyal N, et al. Roberta: A robustly optimized bert pretraining approach[J]. arXiv preprint arXiv:1907.11692, 2019.

[6] Sun Y, Wang S, Li Y, et al. Ernie 2.0: A continual pre-training framework for language understanding[C]//Proceedings of the AAAI Conference on Artificial Intelligence. 2020, 34(05): 8968-8975.

Document

¿­·¢Óé·¢K8com