Batch Prompting: Efficient Inference with Large Language Model APIs

Batch Prompting: Efficient Inference with Large Language Model APIs