A method to quantify reflection biases in embedding models by permuting document segments. It reveals systematic positional and language biases, ensuring all parts of a document are adequately represented in its embedding for search.
This framework helps evaluate how well AI models represent all parts of a document, especially long ones. It works by shuffling document sections to see if the model unfairly favors certain positions or languages, helping engineers build more balanced search systems.
Was this definition helpful?