Attention: You are using an outdated browser, device or you do not have the latest version of JavaScript downloaded and so this website may not work as expected. Please download the latest software or switch device to avoid further issues.
Who may be interested:
Are you a leader or policymaker considering generative artificial intelligence-driven (GenAI) deployments, or a researcher or evaluator interested in the science and real-world application of GenAI evaluation? Read on to learn about the development and assessment of EvalAssist, an open-source GenAI tool to augment early career evaluators’ capacity to design rigorous and ethical program evaluations.
About:
In 2025, Data Foundation Senior Fellow Lauren Damme, Ph.D. led a project to develop a simple custom GPT to support early career evaluators in developing and designing program evaluations. The goal: to create an open-source generative artificial intelligence (GenAI) augmentation tool and thought partner to help ensure evaluation designs were rigorous, building upon best practices from the social sciences; and ethical, accounting for the highest standards in human subjects research.
However, the nascent field of GenAI evaluation lacks established paradigms to uncover the human and social impacts of GenAI use. With support from over 45 highly-experienced evaluators representing 15 countries and the Data Foundation’s Data Coalition members, the development of EvalAssist not only resulted in a support tool for evaluators, but also generated advancements in the way we approach assessments of GenAI deployments. Evaluators and researchers exploring GenAI applications may be particularly interested in the measurement rubric developed to understand human and social impacts of tool use.
Key Resource:
Stay tuned to read more about the model development and evaluation rubric in our upcoming Technical Paper (forthcoming, December 2025)
Access EvalAssist:
Use EvalAssist to help you design program evaluations: Coming soon (December 2025)
Get Involved:
Are you an evaluator with more than five years of experience who would like to help assess the model? Take a brief intake survey and we will be in touch!
Evaluation Expert Advisory Group*:
Juliette Berg
Jacqueline Berman
Mamoun Besaiso
Hannah Betesh
Kerry Bruce
Amanda Cash
Aubrey Comperatore
Lauren Decker-Woodrow
Danuta Dobosz
Melanie Dubuis
Kate Dunham
Clement Dupont
Meghan Ecker-Lyster
Gizelle Gopez
Gusimerjeet Gursimerjeet
Sumera Idris
Susan Jenkins
Natalie Joseph
Vicky Kaisidou
Sharon Lacuyer
Yuna Liang
Kris Lugo-Graulich
Baptiste Millet
Claudia Mir Cervantes
Nomsey Okosa
Carlos Javier Rodriguez Cuellar
Patricia Rogers
Radha Roy
Sutirtha Sahariah
Brandie Sasser
Deena Schwartz
Aylin Talgar Pietz
Ignacio Torrano
Jessica Weitzel
Brean Witmer
*Additional members requested not to be publicly acknowledged.