Machine learning (ML) has been described as a modern Oracle of Delphi—a way to quickly solve problems in different domains, whether auto-completing email messages or predicting the presence of malignant tumors. Computer scientists are theorizing and designing ML technology into our social and personal lives. ML has justifiably stirred tremendous excitement in research, industry, and the popular zeitgeist of artificial intelligence (AI).
However, the enthusiastic adoption of ML has also had negative consequences. ML is being used for unsavory and controversial purposes, such as generating "deep fake" videos and reproducing facial discrimination.8,41 On the research side, new research points to worrisome trends of chasing metrics over more principled approaches and questionable gains in deep learning's performance compared to linear models.10,17,30 As a researcher working in applied machine learning for mental health, I have seen the consequences of this firsthand, where principled approaches to methods are neglected13 and people become the objects of prediction in a pipeline.11 Computer science (CS) is at a stage where interest in ML has both brought energy to new questions as well as raised reasonable concerns.
In response, a new area called "human-centered machine learning" (HCML) promises to balance technological possibilities with human needs and values. Put simply, HCML couples technical innovations in ML with social values like fairness, equality, and justice. The focus for HCML is broad; it includes fair and transparent algorithm design,39 human-in-the-loop decision-making,48 design for human-AI collaborations,4 and exploring the social impacts of ML.
However, there are no unifying guidelines on what "human-centered" means, nor how HCML research and practice should be conducted. People have worked to articulate a nascent set of values for HCML,11,44,47 but the concept is not clear and definitions come from many sides. This moment provides an opportunity to refine HCML into a unifying and interdisciplinary force across CS rather than risk fracture with each sub-field of CS taking ownership of an independent vision of HCML.29
This article draws on the interdisciplinary history of human-centered thinking, HCI, AI, and science and technology studies (STS) to propose best practices for HCML. I also draw from lessons learned from my experiences as a scholar who transitioned from media studies to human-computer interaction and applied ML. By adapting these lessons learned and discussing computing's responsibility for society through a sociotechnical lens,43 this article provides a beginning set of HCML practices to help scholars conduct research that is rigorous without losing sight of values and pro-social goals. These practices are intended to be revisited through the whole HCML pipeline of problem brain-storming, development, and deployment. Finally, I conclude with pragmatic examples that extend these practices into disciplinary changes to ensure HCML's future success.
HCML has captured the interest of ML and other computing disciplines. This interest has been reflected in a series of workshops and activities on human-centered perspectives at major HCI venues, such as ACM SIGCHI in 2016,22 and two in 2019.26,28 There has been attention in ML too, as shown through the rapid ascent of the FAT ML workshop to its own independent ACM FAccT conference. The term "human-centric" has also appeared in workshops in NeurIPsa and in a special issue of ACM Transactions on Interactive Intelligent Systems.20 These workshops and special issues discuss common themes, such as prioritizing human needs and values like explainability, studying the human labor in ML engineering, and audits of deployment. Collectively, these represent opportunities for humans to interface with ML systems.
I have been working in HCML since 2014, researching how machine learning can detect mental illness expressed through social media data.11,12,14 This research agenda is complex, with technical challenges around data size and modeling capabilities; challenges of mapping psychological constructs to ambiguous social media data; and social issues of governance, privacy, and equity. At the start of my journey, there were no terms that explained my interest in both ML and social implications. HCML and its umbrella of concern provides a crucial home for my research that jointly considers technical innovation alongside social concerns.
Human-centered machine learning couples technical innovations in ML with social values like fairness, equality, and justice.
But having a term to define the field in my head has not made it easier to describe to others. People often ask me what "counts" as HCML and interrogate what it means. Does working with human-generated data count? If I do interviews or ask users what they want, does that count? What if I don't build an AI system, but ask people about its impacts? How about if I point out the racial disparity of AI systems but never build anything and don't have a CS degree? Attaining a definition is also complicated because, although HCI and ML are players in this area, other disciplines such as critical data studies and STS contribute to conversations about HCML. Was HCML a term that could only be applied to CS research, but not more than that?
In short, pinning down a definition for HCML is tricky because it must contend with these questions. In response, I found that many proposed definitions pull HCML to narrow conceptualizations of its scope and impact. Some researchers conceptualize the term around values and operationalize those as guideposts for HCML. The most popular of these values is "fairness" and the closely related concept of "justice." Other scholars focus on methodological plurality, often drawing from qualitative methods (such as interviews and user observation) in HCI, the social sciences, and the humanities. Yet another contingency conceptualizes HCML as applied research areas as human-centered domains, like ML for health. What "counts" as HCML, then, becomes bounded in one of these boxes: values, methods, or domains. But to me, these definitions always fell flat. In focusing on procedures or values, these definitions missed a key part of the whole concept—what it meant to be "human-centered."
I returned to prior work to examine how others have thought about the concept. Some may assume human-centered is a novel term to this technical moment, but in fact it has close precedents to "human-oriented" research in CS and other disciplines. The earliest mention of the concept comes from over 50 years ago in the 1970s when engineers coding computational systems used "human-oriented design" to distinguish themselves from work that just had a technical focus. In so doing, they highlighted the interplay of human values in engineering software systems that supported human labor.25
The phrase "human-centered" or "human-oriented" is also not exclusively the property of CS and engineering fields. The term has also appeared in information sciences and STS. In information sciences, scholars argued that being "human-centered … must take account of varied social units that structure work and information—organizations and teams, communities, and their distinctive social processes and practices."28 Perhaps the most straightforward definition was offered by Gasson:21
Human-centered systems production should concern itself with the joint questions of 'What can be produced?' and 'What should be produced?' The first is about what is technically feasible, the second about what is socially desirable.
What struck me in this definition was the understanding of the gap between methods and values of what can be produced and what should be produced, which is a definitely moral call to action. I was struck by the similarity of this definition of Gasson's with Mark Ackerman's "social-technical gap," as proposed in the early days of Computer Supported Cooperative Work (CSCW).1 At the time, CSCW was mostly focused on building technology to facilitate collaboration in workplaces. Ackerman pointed out that CSCW must contend with the gap between our social imaginations and the technical realities of what we could engineer. To do so, he argues, we must acknowledge that our systems are "rigid and brittle" about the nuanced social work in which people work. Ackerman suggests ways to manage this gap in the same ways prior work on human-centeredness has approached this problem—considering the ways to "bridge" the social-technical gap by re-introducing theory that recontextualizes CSCW. Bridges are eminently practical in that they provide a path to connect two distant things—in this case, methods and values.
The field of ML now has a similar challenge designing for and bridging social imaginations and values with ML methods that Ackerman describes for CSCW several decades ago. I think of this as a "human-centered" bridge between values and methods and domains.
Using this prior work as groundwork for discussing practices with HCML, I propose a more unifying definition statement for how we should think about HCML:
Human-centered machine learning is a set of practices for building, evaluating, deploying, and critiquing ML systems that balances technical innovation with an equivalent focus on human and social concerns.
HCML is a set of practices that guide better decision-making throughout the whole ML pipeline of problem brainstorming, development, and deployment. It attempts to bridge the difficult world of values with methodological and technical innovation. Our technical innovations are used by others, and the acknowledgment of this can assist in mitigating, avoiding, and even improving outcomes of ML systems. When HCML serves as a dual focus rather than a laundry list of methods or values, we can return to the initial intentions of human-centered thinking as it was introduced to CS in the 1970s—a way to approach problem-solving. Therefore, questions from my colleagues that interrogate the edges of the concept of HCML are answered with, "We can look at the details of how something was conceptualized, executed, and evaluated to know if it was HCML."
Why focus on practices? Practices are actionable compared to values because they offer ways to move forward and bridge values into real decisions. I'm not the first to propose that practices are important.33,42 As mentioned before, Ackerman raised similar questions in CSCW of how to overcome the social-technical gap,1 and Philip Agre did this in the mid-1990s with "critical technical practice."3 This article's emphasis on practices also answers Mittelstadt's call for practices that can map abstract ideas like values onto everyday practices.33 Practices are also embedded in the thought process with these decisions from their start because they directly affect what you do. In these ways, they provide opportunities for reflection for decision making.
Here, I outline practices that are key to conducting HCML. These practices draw from areas essential to bridging the social and technical realities of HCML research as well as the computational traditions from HCI and AI. The challenges of representing "messy" social contexts in computer code have been a central concern of both of these areas,24 and scholars have spoken to complementary challenges of translating social worlds to the needed abstraction of systems.36,43 Additionally, I look to ideas from critical perspectives within STS, and critical data studies HCI3,18,45 and synthesize practices from emerging work that is actively grappling with these problems.22,26,38 Many of these ideas have been inspired by my experiences working in this space without the HCML label to accompany it.
Ensure ML is the right solution and approach to take. Due to hype and industry interest, many assume that ML is an appropriate (and often the best) solution for a problem. Some engineers and researchers use ML as a silver bullet, a straightforward solution to complex problems that have resisted other approaches. However, there is an assumption embedded in this popular narrative of ML's success that suggests: because we can use ML, we should. Treating ML as a silver bullet is what Evgeny Mozorov coined as the term "technological solutionism" nearly a decade ago,34 and what Selbst et al. further clarifies as the "solutionism trap," in which CS narratives imply that technology can solve a problem.43 The first practice, then, is to reframe these traps into an actionable component of HCML problem formulation and conceptualization. For each application, we should ask ourselves: "Is ML the right approach to take in the first place?"
Avoiding technological solutionism is more nuanced than just weighing the pros and cons of a potential ML solution. This practice asks us to evaluate whether ML is a good solution and if we have correctly articulated the problem that ML says it solves in the first place. In my work, we have considered building ML systems that could prevent online content moderators from seeing graphic content in mental health postings. This statement presumes that moderators should not see graphic content and that ML could make a meaningful dent in how much graphic content a moderator sees. This solution does not consider there are other problems in their work—like unsafe working conditions and psychological trauma—that may not be solved by AI alone but in tandem with other changes, such as corporate structure and compensation. In many solutions, the people involved are not asked what an ML solution could look like for them.
What is one way to make progress toward this practice? Jess Holbrook recalls saying to engineers, "Don't expect Machine learning to figure out what problems to solve."b He calls on us to use needs-finding exercises to triangulate what peoples' needs are before engineering energy is spent on developing outcomes for them. ML engineers are not the experts of a target population, and by assuming we know better, we risk causing more harm than good. At a minimum, we should go ask our target populations what they want and need.
Furthermore, attenuating ourselves to this practice is more than just asking "what are our target population's needs?" because that question doesn't consider if we should use ML at all to address a problem. Nushi et al. describes this as "accountability" of our technical prowess to social outcomes.35 Considering this question early and often in the thought process may raise the uncomfortable truth that the ML solution should never happen in the first place.6 We have responsibilities to our systems and the sociotechnical impacts of their design and deployment—the process is not simply data-design-deploy-done.
At its most radical, HCML should consider the possibility that the best way to do this is by refusing ML development. Refusal gives power back to the communities that ML may take agency away from. This refusal might look like manifest-nosc of communities that rightfully resist technological intervention as thought of by computer scientists. Similarly, a truly human-centered approach allows communities the space to resist ML and instead advocate for the most just and feminist futures. By including this work inside the space of HCML, we allow for these important conversations to happen rather than relegate them to the sidelines of CS.
Acknowledge ML problem statements take positions. At the point where ML has been chosen as a solution, the next step is formalizing the problem statement that HCML is going to solve. Problem statements are necessary to operationalize abstract ideas into concrete items. A classic framing of ML problem statement is, "Given a series of P posts in X domain, what is the outcome of the binary prediction task t?" Problem statements help teams focus on outcomes, divide up work, and establish parameters for a successful project.
However, after deciding that something is a problem and breaking it down into discrete tasks, ML engineers make subtle decisions to accomplish those goals. These decisions will impact how that problem will be instantiated in the world and therefore how data is sampled and labeled, what methods are chosen, and how to evaluate the success of the solution.
I believe a problem statement for an HCML solution is a position statement that reflects peoples' perspectives and assumptions. The idea of positions comes from feminist scholars who recognized that peoples' identity characteristics and social facets, like their gender and race, influenced how people understood and perceived the world.15,37 To be clear, positionality is more than just acknowledging that men or women have different perspectives on the world—these facets have influenced what and how we make becomes technical reality.16 For example, in my research, sometimes I frame diagnosis of mental illness as a binary state of being, rather than representing relapse, recovery, or transitions between states of better or worse well-being. In this example, I make assumptions to make the problem statement easier to solve. These decisions have impacts on the downstream system that is built and people whose lives could be influenced by this system.
To illustrate this, I offer an example: ML systems that predict criminality from facial features of mugshots take a position on who may be engaging in criminal activity (people with police photographs) and then decide that facial features contain information about criminality. None of this is grounded in immutable truth about the world. Criminality is determined by the legal system under which someone operates, which differs across governments. There is enormous variance in who is arrested for potentially committing crimes, who may have a police photograph taken, and cultural as well as legal considerations of who is eventually found guilty of a crime. This is notwithstanding the highly controversial question of whether cultural traits of criminals can be physiologically detected in facial features. And, in fact, my own social and identity characteristics, like my ethnicity, race, and gender, have implications on how I perceive the criminal justice system and what facial features I see as relevant for prediction if any.19 These decisions are impacted by the social world we live in, namely the cultural and legal system, not by objective facts about how criminality is defined.
How might this perspective on positions be integrated into an HCML agenda? Positions say, "I see that my experiences of the world impact how I view this problem, and I'm going to document, address, and interrogate them as best I can." This can be in published position statements in papers and documentation about the perspectives the team brings to the task and how these positions informed decision-making. For example, a large body of my research focuses on high-risk mental illnesses, from eating disorders to opioid use disorder and suicidality. I have personal experiences with mental illness, though not with the ones that appear in my work. Driven by my experiences, my position has influenced how I report data and protect participant privacy, consider ethics board review and care of the research, and how I describe the outcomes of studies. The point of positions is not virtue signaling to get published, nor self-flagellation or emotional theater. We all have unique perspectives on how we have experienced the world—it is ok to acknowledge that it impacts how we do our work in HCML. This practice asks us to ask ourselves: Who are we as researchers, and what perspectives do we bring to problems that influence our choices?
Move beyond users and interaction as the definition of human. To start any HCML project, it is essential to clarify who matters in our research pursuits. Historically, CS has mapped "humans" to the notion of "use"—a single person's interaction with or relationship to a technology.5 This concept is useful because it constrains attention on who and what to focus on during design.
ML challenges who a user is and what constitutes interaction. Take the example of "users" of facial recognition technology in airport security. Traditionally, the user is mapped to the person whose face is recognized or the customs agent analyzing the results of this process. However, other individuals that engage with the system aren't captured in traditional interaction metaphors. In the facial recognition example, a person may be in the background and ambiently recognized without directly engaging with or "using" the technology. There are other stakeholders as well, including all the people whose photographs served as training data, with or without their consent, as well as those that refuse to use facial recognition technology.
To avoid these pitfalls and reveal the true diversity of interactions between machine learning and people, HCML must move beyond just an "interaction mode" of understanding users.5 We must consider multiple ways of engaging with technology—directed and interactive, as well as ambient and non-use—as important questions to engage with for understanding ML systems and broader implications for HCML. Clarifying who is involved can help manage who should be involved in the development life cycle of a system.
Several existing analytical lenses might facilitate this process for HCML. One lens that has been proposed is the "post-user" lens, which Baumer and Brubaker define as meaningfully engaging with the complex interactions that people can have with technology. Another approach is an ecosystemic model, or systems-style thinking, in which the considerations of users evolve to understanding stakeholders,23 each with needs and preferences. Eco-systems recognize distributed impacts of people impacted by technology by using their data, ambient use of systems, or decision to not use technological systems at all. Finally, there is inspiration to be taken from the area of needs assessment, asking who are our stakeholders and what are their needs?
A truly human-centered approach allows communities the space to resist ML and instead advocate for the most just and feminist futures.
Legitimize domain contributions/collaborations as field priorities. In addition to prioritizing the needs of stakeholders, we should also look to collaborative fields that have been in conversation with HCML. The true potential of a human-centered perspective considers human and social factors in technological development, and therefore we should consider other fields as equitable contributors to HCML.
Domains outside of ML possess the theory, methods, and insights to help HCML build more thoughtful systems. Current research in HCI-AI collaborations points to starting opportunities for value-sensitive approaches in HCML. Examples of new techniques to the area include "human-subject experiments, surveys, and ethnographic inquiries."47 Many of these traditions draw from HCI, meaning that this is an easy avenue for initial collaborations and research,4,48 and much work in HCML currently reflects this exchange between HCI and ML. Researchers should be rewarded for engaging in this work, such as producing fairness audits and interpretability tools.
However, interdisciplinarity for HCML raises challenging questions within CS about what constitutes computing research. CS has a longstanding history of using external domains as fertile grounds for advancing its insights. This history includes CS theory being advanced from economics and mechanism design, and neural and genetic algorithms being adapted from biological systems. Classic ML textbooks acknowledge that ML is indebted to an interdisciplinary history of fields like psychology, statistics, and organizational studies—other areas notwithstanding.32 Like the interdisciplinary histories of AI and HCI—in which building tools that try to resolve technical problems are only one step toward embracing a human-centered agenda—we must start investigating these perspectives in our work by rewarding them.
We need different epistemological approaches to ML that push the applications of ML to other domain areas. There should be more than just a collaboration inside CS with its sub-fields who, given common disciplinary backgrounds, may inadvertently slide into old thought patterns.2 We should also look to external domains for pieces to the puzzle of our work and invite contributions into our discipline and positions in our departments and research teams.40 These domains may include critical data studies, sociology, gender, and race studies, and organizational studies.
Design HCML anticipating and iterating on sociotechnical failure. ML historically prioritizes technical conceptions of performance through quantitative metrics, such as evaluating error rates and efficiency. These metrics are essential in understanding the success or failure of a given system.35
In addition to the evaluation of metrics for failures, it is becoming obvious that models produce other kinds of failures. These failures are not readily diagnosed with mathematical techniques like evaluating performance. Instead, these failures are embodied within the social and political world that the technical model inhabits. These sociotechnical failures articulate what has been made "marginal" in our designs, where our abstractions and representations of ML do not match reality.2 Consider the challenges of training language systems on African-American Vernacular English.7 This failure is a breakdown of a system to technically manage different speech patterns, but also a socio-technical failure of not prioritizing a dialect that is biased against an important ethnic community.
Rather than viewing failure as defeat, what if we were to view failure as a new lens of understanding for how should HCML work? The study of technical and sociotechnical failure is a crucial opportunity for HCML to inspect its sociotechnical impacts on the world, avoiding Selbst et al.'s notion of "traps" in abstraction.43 In the study of design, scholars have discussed the notion of sociotechnical failure through metaphors like failures, gaps, brittleness, and breakdowns.1,27 As a scholar of why technology breaks down (a closely related idea to failure), Steve Jackson argues that "worlds of maintenance and repair and the instances of breakdown … are not separate or alternative to innovation, but sites for some of its most interesting and consequential operations."27
An HCML-led study of breakdown and failure should encompass more than how ML operates at a mathematical level, and what it does in the world; it should study the contextual, embedded, and "messy" use of the technology.18 A solid study of failure, therefore, would attend to not only the WHAT of failure and its cause, but WHO is impacted by failure, in what ways are they impacted by, how they cope with, and operate around failure. Current exemplars of this work are studying the places of AI systems failure,35 and navigation points around assumptions of failure as built into high-stakes medical decision-making.9
Domains outside of ML possess the theory, methods, and insights to help HCML build more thoughtful systems.
To truly move the needle on HCML, we must go beyond expecting engineers and researchers to ask the intended users and stakeholders what failure may be. Technologists should commit to following their designs into the messiness of the world to see how people grapple with failures, whether they be unintended quirks or more insidious long-term inequalities.
Finally, an anticipatory focus on failure highlights places where technical innovation should never happen in the first place,6 returning to the question posed earlier. Machine learning cannot solve wicked problems in the world, so rather than shy away from this responsibility, we ought to embrace the messiness in our designs.
What should we do to see these practices adopted in HCML? Let us look to CS as a field to begin.
For example, to embed the practice of moving beyond HCI notions of the user, research could start by using stakeholder analysis and needs assessment to guide value-sensitive algorithm design.48 When deciding if an ML solution is appropriate, interdisciplinarity calls us to read external literature and identify collaborations. From there, a decision may be reached about whether the costs of the ML system are worth it. More humanistic positions can be identified; for example, if classifying mental illness has problematic outcomes on how we define or diagnose illness, ML may not be worth applying in these scenarios or we may need to adjust our training data.
However, these practices also influence our thinking about disciplines writ large. One recent example is NeurIPS, which requires that authors must disclose the negative implications of design. This effort follows the ACM Future of Computing Academy's recent blog post arguing for similar practices."d These considerations of potential implications directly advance the study of sociotechnical failure being normalized in the community. This move by NeurIPS echoes the call of Winograd and Flores to open "the space of possibilities for actions" of the engineer to engage with their creations.46 To see the space of possibility, I encourage readers to reflect on the practices mentioned above in their work throughout the pipeline of ML.
Similarly, academic institutions can support these practices through administrative efforts like cross-appointments and creating interdisciplinary funding opportunities. At higher levels, HCML might be driven through reorganization around concepts like "computing" rather than just "computer science," or incentivized through formal mechanisms like paper publication standards and review guidance like the example of NeurIPS provided earlier. Pragmatic work across teams that incorporates human-centered design and evaluation perspectives31 could also open collaborative opportunities in industry. Grant opportunities that center considerations of these principles alongside traditional calls would be another way to further entrench them in this area.
Practices are everyday yet instrumental to how work is conducted. Therefore, I consider HCML to be driven by practices more than just a checklist of methods or self-stated values. By applying these practices, we are better able to direct our decisions, which in turn inform meta-level considerations for the field of computing moving forward.
Figure. Watch the author discuss this work in the exclusive Communications video. https://cacm.acm.org/videos/human-centered-machine-learning
1. Ackerman, M. The intellectual challenge of CSCW: The gap between social requirements and technical feasibility. Human-Computer Interaction 15, 2 (2020), 179–203; https://doi.org/10.1207/S15327051HCI1523_5
2. Agre, A. Computation and Human Experience. Cambridge University Press, 1997.
3. Agre, P., Bowker, G., Gasser, L., Star, L., and Turner, B. Toward a critical technical practice: Lessons learned in trying to reform AI. Bridging the Great Divide: Social Science, Technical Systems, and Cooperative Work, 1997, 131–157; http://polaris.gseis.ucla.edu/pagre/
4. Amershi, S. et al. Guidelines for human-AI interaction. In Proceedings of the 2019 Conf. Human Factors in Computing Systems, 13; https://doi.org/10.1145/3290605.3300233
5. Baumer, E.P.S., and Brubaker, J.R. Post-userism, 2017, 6291–6303; https://doi.org/10.1145/3025453.3025740
6. Baumer, E.P.S. and Silberman, M.S. When the implication is not to design (technology), 2011, 2271; https://doi.org/10.1145/1978942.1979275
7. Blodgett, S.L., Barocas, S., Daumé, H. and Wallach, H. Language (Technology) is power: A critical survey of "bias" in NLP. ACL. 2020; http://arxiv.org/abs/2005.14050
8. Buolamwini, J. and Gebru, T. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Proceedings of 2018 Machine Learning Research; https://doi.org/10.2147/OTT.S126905
9. Cai, C.J. et al. Human-centered tools for coping with imperfect algorithms during medical decision-making. In Proceedings of the 2019 Conf. on Human Factors in Computing Systems, 1–14; https://doi.org/10.1145/3290605.3300234
10. Caruana, R. and Niculescu-Mizil, A. An empirical comparison of supervised learning algorithms. In Proceedings of the ACM 2006 Intern. Conf. Series 148, 161–168; https://doi.org/10.1145/1143844.1143865
11. Chancellor, S. Baumer, E.P.S., and De Choudhury, M. Who is the "human" in human-centered machine learning: The case of predicting mental health from social media. PACM-HCI 3, Nov. 2019.
12. Chancellor, S., Birnbaum, M.L., Caine, E.D., Silenzio, V.M.B. and De Choudhury, M. A taxonomy of ethical tensions in inferring mental health states from social media. In Proceedings of FAT* 2019, 79–88; https://doi.org/10.1145/3287560.3287587
13. Chancellor, S. and De Choudhury, M. 2020. Methods in predictive techniques for mental health status on social media: a critical review. npj Digital Medicine 3, 1 (2020), 43; https://doi.org/10.1038/s41746-020-0233-7
14. Chancellor, S., Lin, Z., Goodman, E.L., Zerwas, S. and De Choudhury, M. Quantifying and predicting mental illness severity in online pro-eating disorder communities. In Proceedings of 2016 CSCW, 1169–1182; https://doi.org/10.1145/2818048.2819973
15. Collins, P.H. Intersectionality as Critical Social Theory. Duke University Press, 2019.
16. Costanza-Chock, S. Design justice: Towards an intersectional feminist framework for design theory and practice. In Proceedings of the Design Research Society 2018; https://doi.org/10.21606/dma.2017.679
17. Dacrema, M.F., Cremonesi, P., and Jannach, D. Are we really making much progress? A worrying analysis of recent neural recommendation approaches. In Proceedings of the 13th ACM Conf. Recommender Systems, 2019, 101–109; https://doi.org/10.1145/3298689.3347058
18. Dourish. P. Where the Action Is: The Foundations of Embodied Interaction. MIT Press, 2001; https://doi.org/10.1162/leon.2003.36.5.412
19. Elliot, D.S. Lies, Damn Lies and Arrest Statistics, 1995.
20. Fiebrink, R. and Gillies, M. Introduction to the special issue on human-centered machine learning. TiiS, 2018; https://doi.org/10.1177/019263650408863901
21. Gasson, S. Human-centered vs. user-centered approaches to information system design. J. Info. Technology Theory and Application 5, 2 (2003), 29–46.
22. Gillies, M. et al. Human-centered machine learning introduction and background, 2016, 3558–3565; http://dx.doi.org/10.1145/2851581.2856492
23. Goodpaster, K.E. Business Ethics and Stakeholder Analysis. Cambridge University Press, 2016; http://www.jstor.org/stable/3857592
24. Grudin, J. AI and HCI: Two fields divided by a common focus. AI Magazine 30, 4 (2009); https://doi.org/10.1609/aimag.v30i4.2271
25. Hedberg, E., Mumfor, B. The design of computer systems: Man's vision of man as an integral part of the system design process. In Proceedings of the 1975 IFIP Conf. Human Choice and Computers, 31–59; https://doi.org/10.16309/j.cnki.issn.1007-1776.2003.03.004
26. Inkpen, K., Veale, M., Chancellor, S., De Choudhury, M. and P.S. Baumer, E.P.S. Where is the human? Bridging the gap between AI and HCI. In Proceedings of the 2019 Conf. Human Factors in Computing Systems, 1–9; https://doi.org/10.1145/3290607.3299002
27. Jackson, S.J. Rethinking repair. Media Technologies, 2014, 221–240; https://doi.org/10.7551/mitpress/9780262525374.003.0011
28. Kling, R. and Star, S.L. Human centered systems in the perspective of organizational and social informatics. ACM SIGCAS Computers and Society 28, 1 (1997), 22–29; https://doi.org/10.1145/277351.277356
29. Kuhn, T.S. Scientific paradigms. Sociology of Science. B. Barnes (ed.). Penguin Books, Middlesex, England, U.K., 1972.
30. Lipton, Z.C. and Steinhardt, J. Troubling trends in machine learning scholarship, 2018, 1–15; https://doi.org/10.1089/end.2009.0245
31. Maguire, M. Methods to support human-centred design. Intern. J. Human Computer Studies 55, 4 (2001); 587–634; https://doi.org/10.1006/ijhc.2001.0503
32. Mitchell, T.M. Machine Learning. McGraw-Hill Higher Education, 1997.
33. Mittelstadt, B. Principles alone cannot guarantee ethical AI. Nature Machine Intelligence 1, 11 (2019), 501–507; https://doi.org/10.1038/s42256-019-0114-4
34. Morozov. E. To Save Everything, Click Here: The Folly of Technological Solutionism. Public Affairs, 2013.
35. Nushi, B., Kamar, E., and Horvitz, E. Towards accountable AI: Hybrid human-machine analyses for characterizing system failure. Hcomp, 2018, 126–135; http://arxiv.org/abs/1809.07424
36. Passi, S. and Barocas, S. Problem formulation and fairness. arXiv, 2019, 39–48; https://doi.org/10.1145/3287560.3287567
37. Ramazanoglu, C. and Holland, J. From truth / reality to knowledge / power: Taking a feminist standpoint. Feminist Methodology, 2015, 60–80.
38. Ramos, G. et al. Emerging perspectives in human-centered machine learning. In Proceedings of the 2019 Conference on Human Factors in Computing Systems, 1–8; https://doi.org/10.1145/3290607.3299014
39. Ribeiro, M.T., Singh, S. and Guestrin, C. 'Why should I trust you?' In Proceedings of the 22nd ACM SIGKDD Intern. Conf. Knowledge Discovery and Data Mining 2016, 1135–1144; https://doi.org/10.1145/2939672.2939778
40. Ribes, D., Hoffman, A.S., Slota, S.C., and Bowker, G.C. The logic of domains. Social Studies of Science 49, 3 (2019), 281–309; https://doi.org/10.1177/0306312719849709
41. Scheuerman, M.K., Paul, J.M., and Brubaker, J.R. How computers see gender: An evaluation of gender classification in commercial facial analysis and image labeling services. In Proceedings of the 2019 ACM on Human-Computer Interaction 3, CSCW; https://doi.org/10.1145/3359246
42. Schiff, D., Rakova, B., Ayesh, A., Fanti, A. and Lennon, M. Principles to practices for responsible AI: Closing the gap, 2020; http://arxiv.org/abs/2006.04707
43. Selbst, A.D., boyd, d., Friedler, S.A., Venkatasubramanian, S. and Vertesi, J. Fairness and abstraction in sociotechnical systems. In Proceedings of 2019 FAT*, 59–68; https://doi.org/10.1145/3287560.3287598
44. Shneiderman, B. Human-centered artificial intelligence: Reliable, safe and tustworthy. Intern. J. Human-Computer Interaction 36, 6 (2020), 495–504; https://doi.org/10.1080/10447318.2020.1741118
45. Suchman, L. Human-Machine Reconfigurations: Plans and Situated Actions, 2nd Edition. Cambridge University Press, 2006.
46. Winograd, T. and Flores, F. On understanding computers and cognition: A new foundation for design. A response to the reviews. Artificial Intelligence 31, 2 (1987), 250–261; https://doi.org/10.1016/0004-3702(87)90026-9
47. Vaughan, J.W. and Wallach, H. A Human-Centered Agenda for Intelligible Machine Learning.
48. Zhu, H., Yu, B., Halfaker, A., and Terveen, L. Value-sensitive algorithm design: Method, case study, and lessons. In Proceedings of the ACM 2018 Human-Computer Interaction 2, CSCW; https://doi.org/10.1145/3274463
Copyright held by author/owner. Publication rights licensed to ACM.
Request permission to publish from firstname.lastname@example.org
The Digital Library is published by the Association for Computing Machinery. Copyright © 2023 ACM, Inc.
No entries found