E. Presani
YOU?
Author Swipe
View article: Introducing v0.5 of the AI Safety Benchmark from MLCommons
Introducing v0.5 of the AI Safety Benchmark from MLCommons Open
This paper introduces v0.5 of the AI Safety Benchmark, which has been created by the MLCommons AI Safety Working Group. The AI Safety Benchmark has been designed to assess the safety risks of AI systems that use chat-tuned language models.…
View article: ROBBIE: Robust Bias Evaluation of Large Generative Language Models
ROBBIE: Robust Bias Evaluation of Large Generative Language Models Open
As generative large language models (LLMs) grow more performant and prevalent, we must develop comprehensive enough tools to measure and improve their fairness. Different prompt-based datasets can be used to measure social bias across mult…
View article: ROBBIE: Robust Bias Evaluation of Large Generative Language Models
ROBBIE: Robust Bias Evaluation of Large Generative Language Models Open
David Esiobu, Xiaoqing Tan, Saghar Hosseini, Megan Ung, Yuchen Zhang, Jude Fernandes, Jane Dwivedi-Yu, Eleonora Presani, Adina Williams, Eric Smith. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. 20…
View article: "I'm sorry to hear that": Finding New Biases in Language Models with a Holistic Descriptor Dataset
"I'm sorry to hear that": Finding New Biases in Language Models with a Holistic Descriptor Dataset Open
As language models grow in popularity, it becomes increasingly important to clearly measure all possible markers of demographic identity in order to avoid perpetuating existing societal harms. Many datasets for measuring bias currently exi…
View article: “I’m sorry to hear that”: Finding New Biases in Language Models with a Holistic Descriptor Dataset
“I’m sorry to hear that”: Finding New Biases in Language Models with a Holistic Descriptor Dataset Open
As language models grow in popularity, it becomes increasingly important to clearly measure all possible markers of demographic identity in order to avoid perpetuating existing societal harms. Many datasets for measuring bias currently exi…
View article: Bringing Citations and Usage Metrics Together to Make Data Count
Bringing Citations and Usage Metrics Together to Make Data Count Open
Over the last years, many organizations have been working on infrastructure to facilitate sharing and reuse of research data. This means that researchers now have ways of making their data available, but not necessarily incentives to do so…