Artwork

Inhoud geleverd door re:verb, Calvin Pollak, and Alex Helberg. Alle podcastinhoud, inclusief afleveringen, afbeeldingen en podcastbeschrijvingen, wordt rechtstreeks geüpload en geleverd door re:verb, Calvin Pollak, and Alex Helberg of hun podcastplatformpartner. Als u denkt dat iemand uw auteursrechtelijk beschermde werk zonder uw toestemming gebruikt, kunt u het hier beschreven proces https://nl.player.fm/legal volgen.
Player FM - Podcast-app
Ga offline met de app Player FM !

E82: The Rhetoric of AI Hype (w/ Dr. Emily M. Bender)

53:02
 
Delen
 

Manage episode 372547057 series 2460300
Inhoud geleverd door re:verb, Calvin Pollak, and Alex Helberg. Alle podcastinhoud, inclusief afleveringen, afbeeldingen en podcastbeschrijvingen, wordt rechtstreeks geüpload en geleverd door re:verb, Calvin Pollak, and Alex Helberg of hun podcastplatformpartner. Als u denkt dat iemand uw auteursrechtelijk beschermde werk zonder uw toestemming gebruikt, kunt u het hier beschreven proces https://nl.player.fm/legal volgen.

Are you a writing instructor or student who’s prepared to turn over all present and future communication practices to the magic of ChatGPT? Not so fast! On today’s show, we are joined by Dr. Emily M. Bender, Professor in the Department of Linguistics at the University of Washington and a pre-eminent academic critic of so-called “generative AI” technologies. Dr. Bender’s expertise involves not only how these technologies work computationally, but also how language is used in popular media to hype, normalize, and even obfuscate AI and its potential to affect our lives.

Dr. Bender’s most well-known scholarly work related to this topic is a co-authored conference paper from 2021 entitled, “On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?” In our conversation, Emily explains why she and her co-authors chose the “stochastic parrot” metaphor – how this helps us to understand large language models and other related technologies more accurately than many competing metaphors. We go on to discuss several actual high-stakes, significant issues related to these technologies, before Dr. Bender provides a helpful index of some the most troublesome ways they are talked about in the media: synthetic text “gotcha”s, infancy metaphors, linear models of progress, inevitability framings, and many other troublesome tropes.

We conclude with a close reading of a recent piece in the Chronicle of Higher Education about using synthetic text generators in writing classrooms: “Why I’m Excited About Chat GPT” by Jenny Young. Young’s article exemplifies many of the tropes Emily discussed earlier, as well as capturing lots of strange prevailing ideas about writing pedagogy, genre, and rhetoric in general. We hope that you enjoy this podcast tour through the world of AI hype media, and we ask that you please remain non-synthetic ‘til next time – no shade to parrots! 🦜

Check out the podcast Mystery AI Hype Theater 3000, hosted by Dr. Emily M. Bender and Alex Hanna:

Audio Podcast

Video Podcast

Works & Concepts Cited in this episode:

Baria, A. T., & Cross, K. (2021). The brain is a computer is a brain: neuroscience's internal debate and the social significance of the Computational Metaphor. arXiv preprint arXiv:2107.14042.

Bender, E. M., Gebru, T., McMillan-Major, A., & Shmitchell, S. (2021, March). On the dangers of stochastic parrots: Can language models be too big?🦜. In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency (pp. 610-623).

Bender, E. M., & Koller, A. (2020, July). Climbing towards NLU: On meaning, form, and understanding in the age of data. In Proceedings of the 58th annual meeting of the association for computational linguistics (pp. 5185-5198).

Birhane, A., Kalluri, P., Card, D., Agnew, W., Dotan, R., & Bao, M. (2022, June). The values encoded in machine learning research. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency (pp. 173-184).

Birhane, A. (2021). Algorithmic injustice: a relational ethics approach. Patterns, 2(2).

Doyle-Burke, D. & Smith, J. (Hosts).(2020, 8 Jul.). Episode 17: Science Fiction, Science Fact, and AI Consciousness with Beth Singler. [Audio Podcast Episode]. The Radical AI Podcast. Radical AI.

Dubal, V. (2023). On algorithmic wage discrimination. Washington Center for Equitable Growth Working Papers Series.

Fahnestock, J. (1986). Accommodating science: The rhetorical life of scientific facts. Written communication, 3(3), 275-296.

Greene, D., Hoffmann, A. L., & Stark, L. (2019). Better, nicer, clearer, fairer: A critical assessment of the movement for ethical artificial intelligence and machine learning. Proceedings of the 52nd Hawaii International Conference on System Sciences.

Li, H. (2022). Language models: past, present, and future. Communications of the ACM, 65(7), 56-63. [Provides overview and context of Claude Shannon and Andrey Markov’s work on language modeling].

McDermott, D. (1976). Artificial intelligence meets natural stupidity. Acm Sigart Bulletin, (57), 4-9. [on the concept of A.I. as a “wishful mnemonic”]

Mitchell, M. (2021). Why AI is harder than we think. arXiv preprint arXiv:2104.12871. [More on A.I. as “wishful mnemonic”]

Quintarelli, S. (2019, 4 Nov.). Let’s forget the term AI. Let’s call them Systematic Approaches to Learning Algorithms and Machine Inferences (SALAMI). Quinta’s weblog.

Strubell, E., Ganesh, A., & McCallum, A. (2019). Energy and policy considerations for deep learning in NLP. arXiv preprint arXiv:1906.02243.

Torres, É.P. (2023, 15 Jun.). The Acronym Behind Our Wildest AI Dreams and Nightmares. TruthDig. [Émile Torres on the background behind the TESCREAL ideologies].

An accessible transcript of this episode is available here

  continue reading

96 afleveringen

Artwork
iconDelen
 
Manage episode 372547057 series 2460300
Inhoud geleverd door re:verb, Calvin Pollak, and Alex Helberg. Alle podcastinhoud, inclusief afleveringen, afbeeldingen en podcastbeschrijvingen, wordt rechtstreeks geüpload en geleverd door re:verb, Calvin Pollak, and Alex Helberg of hun podcastplatformpartner. Als u denkt dat iemand uw auteursrechtelijk beschermde werk zonder uw toestemming gebruikt, kunt u het hier beschreven proces https://nl.player.fm/legal volgen.

Are you a writing instructor or student who’s prepared to turn over all present and future communication practices to the magic of ChatGPT? Not so fast! On today’s show, we are joined by Dr. Emily M. Bender, Professor in the Department of Linguistics at the University of Washington and a pre-eminent academic critic of so-called “generative AI” technologies. Dr. Bender’s expertise involves not only how these technologies work computationally, but also how language is used in popular media to hype, normalize, and even obfuscate AI and its potential to affect our lives.

Dr. Bender’s most well-known scholarly work related to this topic is a co-authored conference paper from 2021 entitled, “On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?” In our conversation, Emily explains why she and her co-authors chose the “stochastic parrot” metaphor – how this helps us to understand large language models and other related technologies more accurately than many competing metaphors. We go on to discuss several actual high-stakes, significant issues related to these technologies, before Dr. Bender provides a helpful index of some the most troublesome ways they are talked about in the media: synthetic text “gotcha”s, infancy metaphors, linear models of progress, inevitability framings, and many other troublesome tropes.

We conclude with a close reading of a recent piece in the Chronicle of Higher Education about using synthetic text generators in writing classrooms: “Why I’m Excited About Chat GPT” by Jenny Young. Young’s article exemplifies many of the tropes Emily discussed earlier, as well as capturing lots of strange prevailing ideas about writing pedagogy, genre, and rhetoric in general. We hope that you enjoy this podcast tour through the world of AI hype media, and we ask that you please remain non-synthetic ‘til next time – no shade to parrots! 🦜

Check out the podcast Mystery AI Hype Theater 3000, hosted by Dr. Emily M. Bender and Alex Hanna:

Audio Podcast

Video Podcast

Works & Concepts Cited in this episode:

Baria, A. T., & Cross, K. (2021). The brain is a computer is a brain: neuroscience's internal debate and the social significance of the Computational Metaphor. arXiv preprint arXiv:2107.14042.

Bender, E. M., Gebru, T., McMillan-Major, A., & Shmitchell, S. (2021, March). On the dangers of stochastic parrots: Can language models be too big?🦜. In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency (pp. 610-623).

Bender, E. M., & Koller, A. (2020, July). Climbing towards NLU: On meaning, form, and understanding in the age of data. In Proceedings of the 58th annual meeting of the association for computational linguistics (pp. 5185-5198).

Birhane, A., Kalluri, P., Card, D., Agnew, W., Dotan, R., & Bao, M. (2022, June). The values encoded in machine learning research. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency (pp. 173-184).

Birhane, A. (2021). Algorithmic injustice: a relational ethics approach. Patterns, 2(2).

Doyle-Burke, D. & Smith, J. (Hosts).(2020, 8 Jul.). Episode 17: Science Fiction, Science Fact, and AI Consciousness with Beth Singler. [Audio Podcast Episode]. The Radical AI Podcast. Radical AI.

Dubal, V. (2023). On algorithmic wage discrimination. Washington Center for Equitable Growth Working Papers Series.

Fahnestock, J. (1986). Accommodating science: The rhetorical life of scientific facts. Written communication, 3(3), 275-296.

Greene, D., Hoffmann, A. L., & Stark, L. (2019). Better, nicer, clearer, fairer: A critical assessment of the movement for ethical artificial intelligence and machine learning. Proceedings of the 52nd Hawaii International Conference on System Sciences.

Li, H. (2022). Language models: past, present, and future. Communications of the ACM, 65(7), 56-63. [Provides overview and context of Claude Shannon and Andrey Markov’s work on language modeling].

McDermott, D. (1976). Artificial intelligence meets natural stupidity. Acm Sigart Bulletin, (57), 4-9. [on the concept of A.I. as a “wishful mnemonic”]

Mitchell, M. (2021). Why AI is harder than we think. arXiv preprint arXiv:2104.12871. [More on A.I. as “wishful mnemonic”]

Quintarelli, S. (2019, 4 Nov.). Let’s forget the term AI. Let’s call them Systematic Approaches to Learning Algorithms and Machine Inferences (SALAMI). Quinta’s weblog.

Strubell, E., Ganesh, A., & McCallum, A. (2019). Energy and policy considerations for deep learning in NLP. arXiv preprint arXiv:1906.02243.

Torres, É.P. (2023, 15 Jun.). The Acronym Behind Our Wildest AI Dreams and Nightmares. TruthDig. [Émile Torres on the background behind the TESCREAL ideologies].

An accessible transcript of this episode is available here

  continue reading

96 afleveringen

Alle afleveringen

×
 
Loading …

Welkom op Player FM!

Player FM scant het web op podcasts van hoge kwaliteit waarvan u nu kunt genieten. Het is de beste podcast-app en werkt op Android, iPhone en internet. Aanmelden om abonnementen op verschillende apparaten te synchroniseren.

 

Korte handleiding