LLM swap test configurations for comparing extraction quality across models. ## How to use Edit `config.yaml` to point extractor at a different model: ```yaml extraction: model: ollama-remote/qwen3:8b # default: llama3.2:3b alternatives: - ollama-remote/llama3.2:3b - ollama-remote/qwen3:8b - ollama-remote/phi4:14b ``` Run comparison mode to see which model produces better extractions on the same input text. ## Test results Run comparisons and log results here: ``` ./llm-swaps/eval_results/-.json ```