Longterm Wiki

Alignment Evaluations

Evaluationactive

Testing whether AI systems are actually aligned, not just capable of appearing aligned.

Cluster: Evaluation
Parent Area: AI Evaluations

Tags

function:assurancescope:technique

Sub-Areas1

NameStatusOrgsPapers
Epistemic Virtue EvaluationsTesting AI systems for epistemic honesty, calibration, and intellectual humility.emerging00