This repo is an archive of the code and data used in the vicuna blog post.
This repo is deprecated and we recommend using our new question set and evaluation pipeline at fastchat.llm_judge.
We do not recommend using this repo because its questions are relatively easy and it does not address limitations of GPT-4 based evaluation such as position bias.
Our AI-enhanced evaluation pipeline is based on GPT-4. This section provides a high-level summary of the pipeline. For detailed instructions, please refer to the evaluation documentation.
-
Generate answers from different models: Use
qa_baseline_gpt35.py
for ChatGPT, or specify the model checkpoint and runget_model_answer.py
for Vicuna and other models. -
Generate reviews with GPT-4: Use GPT-4 to generate reviews automatically. This step can also be performed manually if the GPT-4 API is not available to you.
-
Generate visualization data: Run
generate_webpage_data_from_table.py
to generate data for a static website, which allows you to visualize the evaluation data. -
Visualize the data: Serve a static website under the
webpage
directory. You can usepython3 -m http.server
to serve the website locally.
We use a data format encoded with JSON Lines for evaluation. The format includes information on models, prompts, reviewers, questions, answers, and reviews.
You can customize the evaluation process or contribute to our project by accessing the relevant data.
For detailed instructions, please refer to the evaluation documentation.