Hosted by Ben Byford, The Machine Ethics Podcast brings together interviews with academics, authors, business leaders, designers and engineers on the subject of autonomous algorithms, artificial intelligence, machine learning, and technology’s impact on society.
This episode we’re chatting with Alex Shvartsman about our AI future, human crafted storytelling, the generative AI use backlash, disclaimers for generated text, human vs AI authorship, practical or functional goals of LLMs, changing themes in science fiction, a diversity of international perspectives and more…
Listen to the episode here:
Alex Shvartsman resides in Brooklyn, New York, and is the author of Kakistocracy (2023), The Middling Affliction (2022), and Eridani’s Crown (2019) fantasy novels. Over 120 of his stories have appeared in Analog, Nature, Strange Horizons, etc. He won the WSFA Small Press Award for Short Fiction and was a three-time finalist for the Canopus Award for Excellence in Interstellar Fiction.
His translations from Russian have appeared in F&SF, Clarkesworld, Tor.com, Analog, Asimov’s, etc. Alex has edited over a dozen anthologies, including the long-running Unidentified Funny Objects series.
This podcast was created and is run by Ben Byford and collaborators. The podcast, and other content was first created to extend Ben’s growing interest in both the AI domain and in the associated ethics. Over the last few years the podcast has grown into a place of discussion and dissemination of important ideas, not only in AI but in tech ethics generally. As the interviews unfold on they often veer into current affairs, the future of work, environmental issues, and more. Though the core is still AI and AI Ethics, we release content that is broader and therefore hopefully more useful to the general public and practitioners.
The hope for the podcast is for it to promote debate concerning technology and society, and to foster the production of technology (and in particular, decision making algorithms) that promote human ideals.
Join in the conversation by getting in touch via email here or following us on Twitter and Instagram.