中大機構典藏-NCU Institutional Repository-提供博碩士論文、考古題、期刊論文、研究計畫等下載:Item 987654321/90791
English  |  正體中文  |  简体中文  |  全文笔数/总笔数 : 78852/78852 (100%)
造访人次 : 35272744      在线人数 : 1005
RC Version 7.0 © Powered By DSPACE, MIT. Enhanced by NTU Library IR team.
搜寻范围 查询小技巧:
  • 您可在西文检索词汇前后加上"双引号",以获取较精准的检索结果
  • 若欲以作者姓名搜寻,建议至进阶搜寻限定作者字段,可获得较完整数据
  • 进阶搜寻


    jsp.display-item.identifier=請使用永久網址來引用或連結此文件: http://ir.lib.ncu.edu.tw/handle/987654321/90791


    题名: 利用與摘要相關的文章重點句結合對比學習改進摘要模型的事實一致性;Combining Key Sentences Related to the Abstract with Contrastive Learning to Improve Summarization Factual Inconsistency
    作者: 張景泰;Chang, Ching-Tai
    贡献者: 軟體工程研究所
    关键词: 抽象式摘要;預訓練模型;對比學習;事實一致性;Abstractive Summarization;Pre-trained Model;Factual Inconsistency;Hallucination;Contrastive Learning
    日期: 2023-02-02
    上传时间: 2023-05-09 18:03:15 (UTC+8)
    出版者: 國立中央大學
    摘要: 摘要中的事實不一致性代表摘要中的訊息無法從來源文章中獲得驗證,是抽象式摘要中棘手的問題,研究顯示模型產出的摘要有30\%擁有事實不一致的問題,使得抽象式摘要難以應用在生活中,近幾年研究者也開始重視這個問題。

    過去的方法傾向於提供額外的背景知識,將其融入於模型中,或者在模型解碼後對產出的結果進行檢查及更正。

    對比學習是近幾年新的模型訓練方法,它在影像領域取得了卓越的成績,其概念是利用正樣本、負樣本之間的對比性,使得模型學習出來的向量物以類聚,正樣本經過模型得到的向量彼此間的距離會較貼近,負樣本經過模型得到的向量彼此間的距離會較疏遠。如此模型在一定程度上擁有了區分不同事物的能力。

    在我們的研究中,首先對原始文章找出與摘要每一句最相關的句子,接著對編碼器使用了對比學習方法使得編碼過後的向量可以獲得與摘要更加相關的原始文章向量使得解碼器產出的摘要更符合事實一致。;Hallucination, also known as factual inconsistency, is when models generate summaries that contain incorrect information or information not mentioned in source text.

    It is a critical problem in abstractive summarization and makes summaries generated by models hard to use in practice.
    Previous works prefer to add additional information such as background knowledge into the model or use post-correct/rank method after decoding to improve this headache.

    Contrastive learning is a new model-training method and has achieved excellent results in the Image Processing field. The concept is to use the contrast between positive and negative samples to make vectors learned by the model cluster together. Given the anchor point, the distance between the anchor point and the positive samples will be closer, and the distance between the anchor point and the negative samples will be farther. This way, the model has the ability to distinguish positive examples from negative examples to a certain extent.

    We propose a new method to improve factual consistency by separating representation of the most relevant sentences and the least relevant sentences from the source document during the training phase through contrastive learning so that the model can learn how to generate summaries that are more relevant to the main points of the source documents.
    显示于类别:[軟體工程研究所 ] 博碩士論文

    文件中的档案:

    档案 描述 大小格式浏览次数
    index.html0KbHTML68检视/开启


    在NCUIR中所有的数据项都受到原著作权保护.

    社群 sharing

    ::: Copyright National Central University. | 國立中央大學圖書館版權所有 | 收藏本站 | 設為首頁 | 最佳瀏覽畫面: 1024*768 | 建站日期:8-24-2009 :::
    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library IR team Copyright ©   - 隱私權政策聲明