Don’t give OpenAI all the credit for GPT-3: You might have helped create the latest ‘astonishing’ advance in AI too

The much-celebrated GPT-3 that can answer questions, write poems, and more wouldn’t be possible without content written by millions of people around the world. Shouldn’t they get some credit?

Published

September 22, 2020

Original Substack post: https://dataleverage.substack.com/p/dont-give-openai-all-the-credit-for

This is a blog post I originally wrote the for the People, Space, and Algorithms blog in September 2020 (you can find the original here). I’ve copied it here because I think the perspective here is a useful data point for ongoing discussions about generative AI.

.

You may have heard about OpenAI’s “GPT-3” (Brown et al. 2020) — an “astonishing” machine learning system that can produce impressive poems, code and op-eds. However, it wasn’t just OpenAI that built this: it was also millions of people writing, posting, and voting on content. In fact, it’s possible that you played a role in creating GPT3!

Indeed, content and data you helped generate may have been used to build AI systems like GPT-3 in the past. It’s even more likely your data will be used to build AI systems in the future. Do you deserve some of the credit for AI’s success? What about the profits (for instance, through a “data dividend”)? What might you do if you aren’t happy about the AI systems your data has helped to create?

GPT-3, like many modern AI systems, needs to be “trained”: in order to produce text, GPT-3 needs to be fed examples of existing text. In the case of GPT-3, the training data is sourced from across the web. One source is called “WebText”, which consists of the text from webpages that were posted on Reddit and had a “karma” (“upvotes” from Reddit users minus “downvotes”) of 3 or more. Another was the English version of Wikipedia. Finally, the researchers also trained GPT-3 on a large number of published books and the “Common Crawl”, a collection of text from across the web that includes everything accessible to web-scraping bots.

If you’ve edited English Wikipedia, posted links to Reddit, voted on Reddit posts, or written text that appeared in a book or the Common Crawl, you almost certainly played a role in creating GPT-3! In a sense, you were an “employee” working on creating this widely-celebrated AI system. This is just one example of what some authors have called “data labor”, in reference that the fact that your labor (editing, reading, voting) helps create the data that fuels AI. Other examples of data labor include contributing to Wikipedia (Vincent and Hecht 2021a), writing restaurants reviews, and most other interactions with online platforms that can be recorded.

Wikipedia and Reddit contributions may have been especially important, as OpenAI’s researchers emphasized datasets that are “higher-quality”. Specifically, GPT-3 only looked some of the text from published books and the Common Crawl, but it looked at all of English Wikipedia and Reddit-filtered web links “2-3 times”. If you’ve edited Wikipedia, GPT-3 may have “read” your Wikipedia edit, not just once, but three times!

For massive systems like GPT-3, it likely isn’t helpful to try to make claims like, “you helped with 1% of the training and I helped with 2%”. Nonetheless, it is meaningful to claim that the success of GPT-3 (and similar projects) is attributable to collective effort by millions of people engaging in (frequently unwitting) “data labor”.

As tech and tech companies become more powerful, there may be opportunities to leverage your valuable data labor. For instance, a group of people could delete or withhold their data (perhaps exerting rights defined in  privacy regulation) to harm an AI system, an idea we’ve called a “data strike” (Vincent, Hecht, and Sen 2019). For cases where “data strikes” aren’t feasible, you could instead funnel your data labor (Vincent and Hecht 2021b) towards organizations that align with your personal values. Our data fuels AI, so the more powerful AI becomes, the more powerful our “data leverage” (Vincent et al. 2021) may be. In the long run, data leverage could be a path towards an AI paradigm that shares its benefits more broadly, perhaps through a “data dividend”.

If you want to learn more about how the data you generate contributes to AI – and emerging research about how you might leverage those contributions in your relationships with tech companies – below are a few good resources:

Some additional notes:

References

Brown, Tom B., Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, et al. 2020. “Language Models Are Few-Shot Learners.” NeurIPS 2020. https://papers.nips.cc/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html.
Vincent, Nicholas, and Brent Hecht. 2021a. “A Deeper Investigation of the Importance of Wikipedia Links to Search Engine Results.” Proceedings of the ACM on Human-Computer Interaction. https://doi.org/10.1145/3449078.
———. 2021b. “Can "Conscious Data Contribution" Help Users to Exert "Data Leverage" Against Technology Companies?” Proceedings of the ACM on Human-Computer Interaction. https://doi.org/10.1145/3449177.
Vincent, Nicholas, Brent Hecht, and Shilad Sen. 2019. “"Data Strikes": Evaluating the Effectiveness of a New Form of Collective Action Against Technology Companies.” In The World Wide Web Conference (WWW). https://doi.org/10.1145/3308558.3313742.
Vincent, Nicholas, Isaac Johnson, Patrick Sheehan, and Brent Hecht. 2019. “Measuring the Importance of User-Generated Content to Search Engines.” In Proceedings of the International AAAI Conference on Web and Social Media. https://doi.org/10.1609/icwsm.v13i01.3248.
Vincent, Nicholas, Hanlin Li, Nicole Tilly, Stevie Chancellor, and Brent Hecht. 2021. “Data Leverage: A Framework for Empowering the Public in Its Relationship with Technology Companies.” In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, 215–27. https://doi.org/10.1145/3442188.3445885.