Abstract
Automatic summarization technology uses some concise sentences to summarize the general idea of the article so that readers can understand the article's main content only by reading the abstract. Instead of simply selecting and rearranging sentences from the original text — just as the general extractive summarization does, the rewriting system model based on seq2seq applies extracted sentences as input to generate more consistent summaries with human language conventions. The BERT pre-trained model has also been applied which achieved good results. To increase coherence, we propose the Rewriting Summarization Based on Sentence Order Prediction (SOP) and Scoped-Context: after the extractive summarization, we use the SOP tasks of the ALBERT model to reorder the sentence sequence; In the abstractive summarization, we apply the group sub-tag-based attention mechanism problem to the seq2seq situation. To further reduce redundancy and irrelevance, each extracted sentence is taken as the input of the rewriter altogether, with its context within a specific scope. Our method can enhance model performance, achieve higher ROUGE scores, and maintain lower computational complexity.
© (2023) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Yixuan Han "Summarization rewriting based on SOP and scoped-context", Proc. SPIE 12721, Second International Symposium on Computer Applications and Information Systems (ISCAIS 2023), 1272115 (26 June 2023); https://doi.org/10.1117/12.2683420
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Data modeling

Education and training

Ablation

Neural networks

Image processing

Local area networks

Performance modeling

Back to Top