Skip to content
/ ATLAS Public

A principled instruction benchmark on formulating effective queries and prompts for large language models (LLMs). Our paper: https://arxiv.org/abs/2312.16171

License

Notifications You must be signed in to change notification settings

VILA-Lab/ATLAS

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Owl

Data License Python 3.10+ Code style: black

ATLAS: An LLM Inquiry Principle Benchmark

This repository contains resources and research on formulating effective queries and prompts for large language models (LLMs). The primary contribution is the introduction of 26 guiding principles aimed at optimizing interactions with LLMs of various scales, such as LLaMA-1/2, GPT-3.5, and GPT-4.

Overview

Our work aims to simplify the underlying concepts of formulating questions for different scales of large language models. By examining their abilities and enhancing user comprehension, we focus on optimizing the design of instructions and prompts. Extensive experiments conducted on models like LLaMA-1/2 and GPT-3.5/4 have verified the effectiveness of the proposed principles.

Data Release

Our dataset, comprising 13k data points, supports the study of LLM prompting principles. The data is curated to facilitate understanding and application of the 26 principles. Our project includes two types of datasets, catering to different needs and research focuses:

  1. General Dataset (general_dataset.json): This comprehensive dataset combines all the examples from each of the 26 principles into a single file, offering a holistic view of our research and its diverse applications.

    • File: general_dataset.json
    • Structure:
      • Each entry contains an Instruction field describing the task.
      • The {model_name}_conversations field provides the model-generated response to the instruction.

    Example:

    {
       "id": "identity_0",
       "model": "gpt4",
       "conversations": [
         {
           "from": "human",
           "value": "Explain the process of photosynthesis."
         },
         {
           "from": "gpt",
           "value": "Photosynthesis is a biochemical process..."
         }
       ]
     },
    
  2. Individual Principle Datasets: We offer separate datasets for boosting and correctness for each of the 26 principles for a more focused study. These files allow researchers to explore and analyze data of specific principles in isolation.

Principled Instruction Finetuning

Our benchmark is compatible with Stanford Alpaca or FastChat. We will further provide the finetuned models.

Third-party Apps

  1. Prompt Enhancer and its GitHub is a web service and tool designed to enable users to select specific principles to enhance the prompts. Special thanks to Lim Hyo Jeong for the excellent contribution!

  2. Magic Prompts is an easy tool to generate high-quality prompts. Special thanks to Lyzr for the excellent contribution!

  3. Prompt-builder and its GitHub is a small Python tool and application that allows users to generate natural language prompts adhering to specific rules and guidelines defined in our paper. Special thanks to lypsoty112 for the excellent contribution!

Citation

@article{bsharat2023principled,
  title={Principled Instructions Are All You Need for Questioning LLaMA-1/2, GPT-3.5/4},
  author={Sondos Mahmoud Bsharat, Aidar Myrzakhan, Zhiqiang Shen},
  journal={arXiv preprint arXiv:2312.16171},
  year={2023},
}

Contributing

We welcome contributions and suggestions to improve our principles and expand the dataset.

Acknowledgements

Stanford Alpaca

About

A principled instruction benchmark on formulating effective queries and prompts for large language models (LLMs). Our paper: https://arxiv.org/abs/2312.16171

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages