Overview
When we have a large pool of unlabeled examples that could be used in a prompt, how should we decide which examples to manually label? Active prompting identifies effective examples for human annotation using:- Uncertainty Estimation: Measure uncertainty on each example.
- Selection: Choose the most uncertain examples for human labeling.
- Annotation: Humans label selected examples.
- Inference: Use newly labeled data to improve prompts.
Uncertainty Estimation (Disagreement)
Query the same example k times and measure disagreement: unique responses / total responses.Example
Selection & Annotation
Select the top-n most uncertain unlabeled examples for human annotation.Inference
Use newly annotated examples as few-shot context during inference.References
- Active Prompting with Chain-of-Thought for Large Language Models (https://arxiv.org/abs/2302.12246)
- The Prompt Report: A Systematic Survey of Prompting Techniques (https://arxiv.org/abs/2406.06608)