Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Baize: An Open-Source Chat Model with Parameter-Efficient Tuning on Self-Chat Data

About

Chat models, such as ChatGPT, have shown impressive capabilities and have been rapidly adopted across numerous domains. However, these models are only accessible through a restricted API, creating barriers for new research and progress in the field. We propose a pipeline that can automatically generate a high-quality multi-turn chat corpus by leveraging ChatGPT to engage in a conversation with itself. Subsequently, we employ parameter-efficient tuning to enhance LLaMA, an open-source large language model. The resulting model, named Baize, demonstrates good performance in multi-turn dialogues with guardrails that minimize potential risks. Furthermore, we propose a new technique called Self-Distill with Feedback, to further improve the performance of the Baize models with feedback from ChatGPT. The Baize models and data are released for research purposes only at https://github.com/project-baize/baize-chatbot. An online demo is also available at https://huggingface.co/spaces/project-baize/chat-with-baize.

Canwen Xu, Daya Guo, Nan Duan, Julian McAuley• 2023

Related benchmarks

TaskDatasetResultRank
Molecular property predictionQM9 (test)--
174
Reagent PredictionMol-Instructions
Exact Match0.00e+0
24
Forward reaction predictionMol-Instructions
Exact Match0.00e+0
24
RetrosynthesisMol-Instructions
Exact Match0.00e+0
24
Proactive next utterance predictionCrossWOZ (test)
LLM-Judge46.1
17
Proactive next utterance predictionLMSYS (test)
LLM-Judge55.09
17
Proactive next utterance predictionShareGPT (test)
LLM-Judge46.3
17
Proactive next utterance predictionWildChat (test)
LLM-Judge48.5
17
Molecule CaptioningMol-Instructions
ROUGE-L0.148
17
True-or-FalseMol-Instructions
Accuracy48
13
Showing 10 of 25 rows

Other info

Follow for update