Attention: You are using an outdated browser, device or you do not have the latest version of JavaScript downloaded and so this website may not work as expected. Please download the latest software or switch device to avoid further issues.

GenAI for Better Evaluation: EvalAssist

Intro text + Embed Line

Who may be interested:

Are you a leader or policymaker considering generative artificial intelligence-driven (GenAI) deployments, or a researcher or evaluator interested in the science and real-world application of GenAI evaluation? Read on to learn about the development and assessment of EvalAssist, an open-source GenAI tool to augment early career evaluators’ capacity to design rigorous and ethical program evaluations. 

About:

In 2025, Data Foundation Senior Fellow Lauren Damme, Ph.D. led a project to develop a simple custom GPT to support early career evaluators in developing and designing program evaluations. The goal: to create an open-source generative artificial intelligence (GenAI) augmentation tool and thought partner to help ensure evaluation designs were rigorous, building upon best practices from the social sciences; and ethical, accounting for the highest standards in human subjects research. 

However, the nascent field of GenAI evaluation lacks established paradigms to uncover the human and social impacts of GenAI use. With support from over 45 highly-experienced evaluators representing 15 countries and the Data Foundation’s Data Coalition members, the development of EvalAssist not only resulted in a support tool for evaluators, but also generated advancements in the way we approach assessments of GenAI deployments. Evaluators and researchers exploring GenAI applications may be particularly interested in the measurement rubric developed to understand human and social impacts of tool use.

Key Resource: 

Stay tuned to read more about the model development and evaluation rubric in our upcoming Technical Paper (forthcoming, December 2025)

Access EvalAssist:

Use EvalAssist to help you design program evaluations: Coming soon (December 2025)

Get Involved: 

Are you an evaluator with more than five years of experience who would like to help assess the model? Take a brief intake survey and we will be in touch!

Evaluation Expert Advisory Group*:

Juliette Berg

Jacqueline Berman

Mamoun Besaiso

Hannah Betesh

Kerry Bruce

Amanda Cash

Aubrey Comperatore

Lauren Decker-Woodrow

Danuta Dobosz

Melanie Dubuis

Kate Dunham

Clement Dupont

Meghan Ecker-Lyster

Gizelle Gopez

Gusimerjeet Gursimerjeet

Sumera Idris

Susan Jenkins

Natalie Joseph

Vicky Kaisidou

Sharon Lacuyer

Yuna Liang

Kris Lugo-Graulich

Baptiste Millet

Claudia Mir Cervantes

Nomsey Okosa

Carlos Javier Rodriguez Cuellar

Patricia Rogers

Radha Roy

Sutirtha Sahariah

Brandie Sasser

Deena Schwartz

Aylin Talgar Pietz

Ignacio Torrano

Jessica Weitzel

Brean Witmer

*Additional members requested not to be publicly acknowledged.



image

DATA FOUNDATION
1100 13TH STREET NORTHWEST
SUITE 800, WASHINGTON, DC
20005, UNITED STATES

INFO@DATAFOUNDATION.ORG

This website is powered by
ToucanTech