https://github.com/tatsu-lab/alpaca_eval/blob/9752edf6ffcea3293a788c9ce2ddc7fd4a30b287/src/alpaca_eval/decoders/openai.py#L190

openai_api_base: Optional[str] = os.getenv("OPENAI_API_BASE") if os.getenv("OPENAI_API_BASE") else openai.base_url,

check OpenAI server builder

look at

Evaluating a model

Making a new evaluator or specify an evaluator

Commands

command: alpaca_eval --model_outputs alpaca_eval/results/mistral-medium/model_outputs.json --annotators_config autoj-13b/configs.yaml --output_path /scratch/hbenoit/swiss-ai/alpaca_eval_results

lm_eval —model hf —model_args pretrained=/scratch/hbenoit/swiss-ai/downloads/llama-slimpajama6b-final/llama-medium/11000 —tasks hellaswag,arc_easy,arc_challenge —batch_size auto:4 —output_path results/llama2-7b-chat-hf.json —show_config