✂️ Open Provence Context Pruner & Reranker
This demo showcases the OpenProvence project, which applies the Provence approach to remove question-irrelevant context while simultaneously computing reranker scores. Modern AI agents and context-engineering pipelines often recurse through large batches of search results. Irrelevant passages inflate the tokens we send to an LLM. By inserting a compact pre-processing model such as Provence, we pass along only the evidence that matters before the LLM generates an answer.
OpenProvence is trained on public datasets with fully open training and inference code, and the weights are released under the MIT License.
Demo Highlights 🚀
- ⚡️ 30M-parameter compact model handles both English and Japanese on CPU -- lightning-fast on GPU.
- ⚙️ Additional model families are available. Swap the model ID to try a different checkpoint.
- Model 🔗: Provide a local path or Hugging Face model ID that points to an OpenProvence checkpoint.
- threshold (recommend: 0.05-0.5): Range 0.00-1.00. Higher values prune more aggressively. Lower values favor recall. Adjust to suit your task.
- Title: Supplying a concise title can improve pruning quality when the context is long.
- Sentence segmentation and language detection run in-model -- no manual tuning needed.
- For implementation notes and evaluation results, see the Open Provence documentation.
0 1
Sample input
| Question (required) | Title (optional) | Context (required, long passages supported) |
|---|