Who and What Drives Algorithm Development: Ethnographic Study of AI Start-up Organizational Formation

Share Share Share Share Share
[s2If !is_user_logged_in()] [/s2If] [s2If is_user_logged_in()]

CONCLUSION: PUZZLING THROUGH EPISTEMIC EDGES IN ALGORITHM R&D

We have examined dimensions of problem formation in data science and the splits in certain forms of expertise and knowledge that could have practical implications for the shaping of an early stage machine learning start-up. When we thought about the procedures for defining the borders and scope of our case study we became increasingly aware of the shifting role boundaries, problems in-process, and people and their expertise being marginalized and yet continuing to push ahead with algorithm development.

A case study that was dynamic with multiple lines of flight took us back again and again to puzzling through the human messiness of data science knowledge and what got put outside or was made as other in machine learning knowledge today and in the foreseeable future. Machine learning knowledge did not travel alone without being enmeshed in people, organizations, bodies, data and patients, and caught up in certain denials and acknowledgments of human pain and suffering. In other words, we acknowledged the observable tensions and looked for ways to advocate for silences lurking underneath the obvious.

Methodologically, capturing algorithms required daily ethnographic note taking, mundane but consistent participation, and an interpretative stance similar to Geertz’s that allowed one to depict the speech, gestures and uncertainties that came from everyday organizational challenges. We suggest, such analysis drew from a knowledge in data science at least an appreciation of how data was not just wrangled, handled or acquired but was organizing. To work with computerized data (in our case healthcare data and medical image data) occasioned being organized by its constraints, possibilities and by strategic partnerships as we have seen with our health insurance company focused on fraud detection. Such consistent participation in the life of an early stage start-up allowed us to gain a larger sense of the textures of data and human patterns similar to the spirit of Levi-Strauss that took us out of the particular and into a generalizing conception through algorithm development.

As we have seen, machine learning could be considered universal viewed with cross-industry application irrespective of specific expertise that has shaped those industries. In our case, a universalizing dimension was applied to algorithms, a domain specific boundedness was applied to the content creators of those industries. Thus, one of the negative outcomes we encountered was epistemic splitting, instead of epistemic inclusiveness, parsing knowledge into plug-ins and universalizing machine learning concepts above other forms of knowledge that weakened organization and compromised quality and flow of ideas that ran through algorithm development. Splitting knowledge into bits also split people and eroded organizational bonds that could have forged professional curiosity across disciplines. Domain expertise suggested a depth of knowledge in a particular field, however it might have also suggested an ongoing attempt to be pollinated by and extended into other areas of expertise. It had a potential to be a truly T-shaped enterprise.

A Note on AI-Inflected Ethnography

The idea of an AI-Inflected ethnography emerged at this time. It came together for us as a way to bring qualitative analysis to silences and to the downstream impact of algorithm development at the moment of conception and problem formation.

We believe an AI-Inflected ethnography has a potential to focus on the silences, intentionalities, gaps, aspirations and conceptions that may be in one moment accepted and in another moment forgotten or rendered invisible. An AI-Inflected ethnography we suggest, is not about what has been built but what has not yet been built and the reasoning or emotions that go into and are fought over to arrive at the product or algorithmic problem worthy of engineering time and worthy of gaining sometimes expensive data resources.

When we think about AI-Inflected ethnography we should not conjure pictures of methodological reasoning. Perhaps such reasoning will come over time and over further case studies but here we are marking active algorithm development today and its implications on people in a start-up that struggled and often failed to inhabit different points of view. An “inflected” ethnography focused on AI development does not offer up suggestions for research tools, recording procedures or discussion on synchronous (real-time) or asynchronous (non-real-time) of research data capture and technique in the field. Instead we have chosen to examine dimensions of problem formation in data science and epistemic splits in certain forms of expertise that have had practical implications and anxieties within an early stage machine learning start-up.

This type of analysis we believe is best located in moments when we can tease out the collaborative opportunities and imagination in algorithm development in an everyday when diversity of thought may be up for grabs and when algorithm problem formation may be held open. This form of analysis is unstable and in some ways an outcome of role murkiness between data scientist and ethnographer. But such analytical instability comes as a benefit. It is at these formative moments when certain epistemic events can be made visible through gaps in idea generation.

One of the strengths of AI-Inflected ethnography is paying close attention not only to the content of knowledge, but also to the processes of knowledge development focused on shifting edges of algorithm R&D. These edges of organizational knowledge and practice are not simply learned, known and then applied ideas but are locally contested and puzzled through. Deciphering epistemic borders suggests that everyone in an early stage start-up is engaged on one level or another with processes of the formation, co-creation and development of algorithmic knowledge and outcome even when these outcomes are very uncertain and distant. The process of building algorithms paradoxically engages and silences organizational members. The epistemological glue that represents machine learning knowledge does not draw together all forms of knowledge as domain knowledge or as a binary of usefulness/non-usefulness. Instead this glue provides a kind of organizational coherence as much as it limits the flow of forms of knowledge and fashions types of algorithms and types of organizations. These inflection points in AI development are tensions and realities not to fend off and avoid but to identify and transform.

It is important to keep in mind that what characterizes an AI-Inflected ethnography is not consultative attendings but an organizational embeddedness of an ethnographer/researcher that helps capture gaps and slippages within an organizational environment operating as a kind of disjointed body pulling together its internal fortitude to take on and build algorithms that can be integrated into software and into people’s lives. The odds of early stage start-up success are daunting, the odds of their failure are well over 90%.

Could these odds be improved upon with a true T-shaped organization and a true diversity of perspective?

Where Do We Go From Here? – What we gained access to was a deeper image of an organizational group of people trying to genuinely integrate a patient’s ‘thank you’ letter into their thinking but unable to seize upon its message of gratitude and transform this message into actionable algorithm development. We gained access to the potential consequences of radiologist’s insight as consultative input and domain knowledge. We gained a sense that the generalizing and particularizing aspects of machine learning were open and unsettled. There were forms of knowledge and emotional life that algorithm development resisted. We believe there may have been other forms of knowledge and diversity of knowledge that has not yet been organizationally imagined to produce very different algorithms for very different outcomes. We do not yet know how these forms will appear in the future but we can as researchers prepare for their emergence by laying the groundwork for the conceptualizing of more malleable and inclusive algorithms.

One of the most disembodied technologies in development today requires embodied and embedded research and engagement. AI calls for us as researchers to not only step back into the social but to step into the daily grind of AI’s silences and epistemic threads that are constantly being shredded and mended to transform organizational coherence.

If we use algorithm development to screen out the pain and suffering of vulnerable people like patients, instead of finding ways to integrate their journeys, then we may also find ever-narrowing algorithm development populating or even taking over our everyday lives.

It is an extensive kind of labor to evaluate, question, chart, chronicle how a start-up and AI development together forge forms of thinking and acting across algorithm, people, product and clients. It seems to be a worthy endeavor. However, if this kind of research needs to get done who is going to do it? Who will listen and make use of it? What kind of organization will it produce? Most importantly, what kind of commitment will it take to produce it?

Rodney Sappington Ph.D. is a data scientist and ethnographer who leads machine learning strategy and product development. His research focuses on the intersections of knowledge, human-machine relations and behavior, and societal outcomes from algorithm development. Rodney is CEO and Founder of Acesio Inc. rodney@acesio.org

Laima Serksnyte Ph.D. has 15+ years of experience in advanced research, consulting and coaching in the areas of executive leadership and organizational psychodynamics. Her research focuses on psycho-social, behavioral and societal levels of analysis in areas of AI development, healthcare, education and organizational member advancement. Laima is Head of Behavioral and Organizational Research at Acesio Inc. laima@acesio.org

NOTES

Acknowledgements – Thank you to the reviewers and curators of EPIC, especially to Dawn Nafus for her continued insights and feedback. A big thank you to colleagues in applied deep learning and data science for careful consideration of the concepts in this paper.

1. A “unicorn” company is a start-up that reaches $1B in revenue.

2. “Moonshots” are typically defined as ambitious and aspirational projects and companies reaching to develop bold and future-oriented products and services.

3. As a cautionary note, our educational and psychological systems have to keep up with AI projected development otherwise fall behind and in disuse, creating an elite group or risk a population whose knowledge is woefully behind, not relevant, or worse considered a danger to AI development. Bladerunner and many other dystopian societal images come to mind.

4. Claims (insurer) transactional data can have multiple uses in machine learning. Claims data is standardized and medically coded and covers a wide area of the patient’s journey. For example, prescription and behavioral trends can be captured across pain medications (opioids) and cholesterol lowering drugs (statins) and this same data can also be used to track physician practice claims, types of claims and time points when outlier claims could indicate probability for fraud.

5. Spiculated margins of a lung nodule are uneven edges that can indicate a higher risk of cancer. “Most nodules that are not cancer have very smooth or rounded margins or look like several rounded nodules together (also called “lobulated”). See Lung Cancer Alliances explanation “Understanding Lung Nodules: A Guide for the Patient.” https://lungcanceralliance.org/wpcontent/uploads/2017/09/Understanding_Lung_Nodules_Brochure_dig.pdf

2018 Ethnographic Praxis in Industry Conference Proceedings, pp. 245–263, ISSN 1559-8918

REFERENCES CITED

Dean, Jeff
2018     Deep Learning to Build Intelligent Systems by Jeff Dean. AI NEXTCon Silicon Valley 2018, April 10-13, Santa Clara.
https://www.youtube.com/watch?v=4skC-kGCYRA

Dean, Jeff
2016     The head of Google’s Brain team is more worried about the lack of diversity in artificial intelligence than an AI apocalypse. Recode. August 13, 2016.
https://www.recode.net/2016/8/13/12467506/google-brain-jeff-dean-ama-reddit-artificial-intelligence-robot-takeover

Faggella, Daniel
2018     Valuing the Artificial Intelligence Market, Graphs and Predictions. Techemergence. Updated September 16, 2018
https://www.techemergence.com/valuing-the-artificial-intelligence-market-graphs-and-predictions/

Ferrary, Michel and Mark Granovetter
2009     The role of venture capital firms in Silicon Valley’s complex innovation network. In Economy and Society. 38:2, 326-359.

Geertz, Clifford
1973.     The interpretation of cultures. New York: Basic Books.

Garling, Caleb
2015     Andrew Ng: Why ‘Deep Learning’ Is a Mandate for Humans, Not Just Machines. Wired. May 5, 2015.
https://www.wired.com/brandlab/2015/05/andrew-ng-deep-learning-mandate-humans-not-just-machines

Glasner, Joanna
2018     The Unicorns Behind San Francisco’s Burdensome Startup Success. Crunchbase News. April 18, 2018
https://news.crunchbase.com/news/unicorns-behind-san-franciscos-burdensome-startup-success/

Koulopoulos, Thomas
2018     According to Peter Diamandis and Ray Kurzweil, These Are the Most Dangerous and Disruptive Ideas. Inc. January 19, 2018
https://www.inc.com/thomas-koulopoulos/according-to-peter-diamandis-ray-kurtzweil-this-is-how-you-build-future.html

Levi-Strauss, Claude
1953     An Appraisal of Anthropology Today. Eds., Sol Tax and others. Chicago, IL: University of Chicago Press.

Ng, Andrew
2018     AI is the new electricity. Startup Grind. March 6, 2018.
https://www.youtube.com/watch?v=fgbBtnCvcDI

Quoc V. Le, Marc Aurelio Ranzato, Andrew Ng, Jeff Dean et al
2012     Building High-level Features Using Large Scale Unsupervised Learning
https://static.googleusercontent.com/media/research.google.com/en//pubs/archive/38115.pdf

Ruiz, Rebecca R.
2017     U.S. Charges 412, Including Doctors, in $1.3 Billion Health Fraud. New York Times. July 13, 2017
https://www.nytimes.com/2017/07/13/us/politics/health-care-fraud.html

Strathern, Marilyn
2004     The Whole Person And Its Artifacts. Annual Review of Anthropology. 2004. 33: 1-19

[/s2If]

Leave a Reply