Multi-view Contrastive Learning for Medical Question Summarization

Publisher:
IEEE
Publication Type:
Conference Proceeding
Citation:
2024 27th International Conference on Computer Supported Cooperative Work in Design (CSCWD), 2024, 00, pp. 1826-1831
Issue Date:
2024-07-10
Full metadata record
Most Seq2Seq neural model based medical question summarization MQS systems have a severe mismatch between training and inference i e exposure bias However this problem remains unexplored in the MQS task To bridge this research gap and alleviate the problem of exposure bias we propose a novel re ranking training framework for MQS called Multi view Contrastive Learning MvCL MvCL simultaneously considers the similarity scores between medical questions and candidate summaries as well as the average similarity scores between candidate summaries and other candidates within the same group and utilizes contrastive learning to optimize the model s ranking ability Additionally we propose a new multilevel inference approach to adapt to this training strategy The approach first filters out candidate summaries that are dissimilar to the original medical question and then selects the summary with the highest average similarity to other candidate summaries from the remaining candidates as the final output We conducted extensive experiments and the results demonstrate that our proposed MvCL framework achieves state of the art results on the majority of evaluation metrics across four datasets 1
Please use this identifier to cite or link to this item: