Research on machine learning (ML) algorithms, as well as on their
ethical impacts, has focused largely on mathematical or computational
questions. However, for algorithmic systems to be useful, reliable, and
safe for human users, ML research must also wrangle with how users’
psychology and social context affect how they interact with algorithms.
This talk will address how novel research on how people interact with ML
systems can benefit from decades-old ideas in social science. The first
part of the talk will address how well-worn ideas from psychology and
behavioral research methods can inform how ML researchers develop and
evaluate algorithmic systems. The second part of the talk will address
how foundational ideas from organizational and institutional theory can
help ML ethicists develop tools and interventions that have practical
utility in tomorrow’s real-world technology.
Bio:
Amy Winecoff is a DataX data scientist at CITP. Her primary interests
are in human-algorithm interactions and fairness in machine learning
systems. Winecoff received her Ph.D. in psychology and neuroscience from
Duke University. After graduate school, she was an assistant professor
at Bard College, where she taught neuroscience, abnormal psychology, and
research methods. After leaving academia, she conducted research and
developed machine learning models for government agencies such as DARPA
and the U.S. Air Force to explain and predict human behavior. As a
senior data scientist at True Fit and Chewy, she developed product
recommendation and search systems. She also conducted quantitative user
research to assess how users’ psychology informs their evaluation of
algorithmic predictions. Winecoff is passionate about diversity and
inclusion in the technology industry.