Add human-centered trust & uncertainty metrics to Hugging Face Evaluate #728
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Summary
This PR adds a new evaluation metric suite,
human_ai_trust, to Hugging Face Evaluate.The goal of this metric is to support human-centered evaluation of AI systems by operationalizing:
Unlike traditional metrics focused solely on predictive performance, this metric surfaces how users interpret, trust, and act on model outputs under uncertainty.
What's Included
This PR introduces:
human_ai_trustMotivation
Accuracy alone is insufficient for evaluating AI systems in high-stakes or vulnerable contexts.
From a human–AI interaction perspective:
This metric suite provides theory-grounded, computational signals for evaluating these human-centered failure modes.
Design Philosophy
This metric is:
Related Work
This metric complements existing evaluation approaches:
Key differences:
Usage Example
Testing
All metrics are fully unit-tested with 100% code coverage, including:
Run locally:
Companion Dataset
A small, theory-grounded demo dataset is available at:
https://huggingface.co/datasets/Dyra1204/human_ai_trust_demo
It demonstrates:
Limitations
These metrics are intended as descriptive human-centered evaluation signals.
They do not:
Checklist
Feedback Welcome
I would very much welcome feedback on:
I'm happy to adjust the implementation to better align with Hugging Face Evaluate conventions.