Ahead of the highly anticipated AI Safety Summit, leading AI scientists from the US, the PRC, the UK and other countries agreed on the importance of global cooperation and jointly called for research and policies to prevent unacceptable risks from advanced AI.
How do we measure the efficacy of language model explainability methods? The authors of this paper present ALMANACS, a language model explainability benchmark that scores explainability methods on simulatability.
Subscribe to our mailing list
If you would like to receive our newsletters and updates, then please subscribe.