KEYWORDS: Data modeling, Education and training, Performance modeling, Systems modeling, Autoregressive models, Mathematical optimization, Machine learning, Quantization, Transformers, Deep learning
In order to solve the complex module dependencies of dialogue systems and improve the system's ability to understand deep knowledge in natural language and produce more coherent texts, this paper introduces an end-to-end dialogue system based on large language models. First, low-rank adaption is used to fine-tune sequence-to-sequence large language models, which reduces system complexity and model fine-tuning cost. Then, the training method of reinforcement learning from human feedback is adopted to make the generated responses more aligned with human expectations. Finally, in-context learning is used to adapt to specific tasks, improving model flexibility and adaptability. Experimental results show that the system performs well in both automatic evaluation and practical use and has strong application value.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.