Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining 2023
DOI: 10.1145/3580305.3599931
|View full text |Cite
|
Sign up to set email alerts
|

WebGLM: Towards An Efficient Web-Enhanced Question Answering System with Human Preferences

Abstract: We present WebGLM, a web-enhanced question-answering system based on the General Language Model (GLM). Its goal is to augment a pre-trained large language model (LLM) with web search and retrieval capabilities while being efficient for real-world deployments. To achieve this, we develop WebGLM with strategies for the LLM-augmented retriever, bootstrapped generator, and human preference-aware scorer. Specifically, we identify and address the limitations of WebGPT (OpenAI), through which WebGLM is enabled with a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2024
2024
2025
2025

Publication Types

Select...
3
2
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 17 publications
(4 citation statements)
references
References 23 publications
0
4
0
Order By: Relevance
“…The pre-trained base model is fine-tuned, and the resulting RedGPT model (Yang et al, 2023b) is further used for instruction generation in an iterative manner to obtain a massive amount of high-quality data. WebGLM-QA (Liu et al, 2023e) generates data in three stages: Prompt Formulation, Instruction Inducting, and Few-shot In-context Learning. Wizard evol instruct 196K (Xu et al, 2023b) and Wizard evol instruct 70K (Xu et al, 2023b) use the Evol-Instruct method, subjecting 175 seed instructions to four evolution stages to enhance the complexity of generated instructions.…”
Section: Model Constructed Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…The pre-trained base model is fine-tuned, and the resulting RedGPT model (Yang et al, 2023b) is further used for instruction generation in an iterative manner to obtain a massive amount of high-quality data. WebGLM-QA (Liu et al, 2023e) generates data in three stages: Prompt Formulation, Instruction Inducting, and Few-shot In-context Learning. Wizard evol instruct 196K (Xu et al, 2023b) and Wizard evol instruct 70K (Xu et al, 2023b) use the Evol-Instruct method, subjecting 175 seed instructions to four evolution stages to enhance the complexity of generated instructions.…”
Section: Model Constructed Datasetsmentioning
confidence: 99%
“…In the end, approximately 240K instructions are obtained. • WebGLM-QA (Liu et al, 2023e). The WebGLM-QA dataset is designed for training the WebGLM generation module and comprises approximately 43K high-quality samples.…”
Section: B12 Model Constructed Datasetsmentioning
confidence: 99%
“…bGLM (Liu et al 2023b) augments LLMs with web search and retrieval capabilities. One major limitation of these approaches is the retrieved text is question-related, thus cannot guarantee the correctness of the question-unrelated portions in the generations.…”
Section: Related Workmentioning
confidence: 99%
“…ML tasks performed well on the SQuAD dataset, released less than a year ago. A logistic regression (LR) model based on linguistic features by Liu et al (2023) in June 2016 achieved an F1-score of 51%, up from 20%. The author reaches 77.3% F1 using BiDAF encoding, a bidirectional LSTM, and multistage decoding ( Chen et al, 2017 ).…”
Section: Literature Surveymentioning
confidence: 99%