Interlude - The Simple Truth
First published
03/03/2015
Genres:
society
philosophy
Summary
Book I: Map and Territory - Part D: Mysterious Answers - Interlude - The Simple Truth
Duration
Parent Podcast
Rationality: From AI to Zombies
View PodcastSimilar Episodes
Methods of Rationality in the Time of Hype
Release Date: 10/13/2023
Authors: Thomas Krendl Gilbert and Nathan Lambert
Description: This week, Tom and Nate discuss some of the core and intriguing dynamics of AI. We discuss the history of the rationality movement and where Harry Potter fan fiction fits in, if AI will ever not feel hypey, the do's and don'ts of Sam Altman, and other topics.(Editor note: sorry for some small issues in Nate's audio. That will be fixed in the next episode)Some links that are references:* HP MOR (Harry Potter and the Methods of Rationality). * A tweet referencing Sam Altman's funny (?) profile change.* Nathan's recent post on Interconnects on the job market craziness.
Explicit: No
Alignment Newsletter #25 by Rohin Shah
Release Date: 11/17/2021
Description: Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Alignment Newsletter #25, published by Rohin Shah on the AI Alignment Forum. Highlights Towards a New Impact Measure (Alex Turner): This post introduces a new idea for an impact measure. It defines impact as change in our ability to achieve goals. So, to measure impact, we can simply measure how much easier or harder it is to achieve goals -- this gives us Attainable Utility Preservation (AUP). This will penalize actions that restrict our ability to reach particular outcomes (opportunity cost) as well as ones that enlarge them (instrumental convergence). Alex then attempts to formalize this. For every action, the impact of that action is the absolute difference between attainable utility after the action, and attainable utility if the agent takes no action. Here, attainable utility is calculated as the sum of expected Q-values (over m steps) of every computable utility function (weighted by 2^{-length of description}). For a plan, we sum up the penalties for each action in the plan. (This is not entirely precise, but you'll have to read the post for the math.) We can then choose one canonical action, calculate its impact, and allow the agent to have impact equivalent to at most N of these actions. He then shows some examples, both theoretical and empirical. The empirical ones are done on the suite of examples from AI safety gridworlds used to test relative reachability. Since the utility functions here are indicators for each possible state, AUP is penalizing changes in your ability to reach states. Since you can never increase the number of states you reach, you are penalizing decrease in ability to reach states, which is exactly what relative reachability does, so it's not surprising that it succeeds on the environments where relative reachability succeeded. It does have the additional feature of handling shutdowns, which relative reachability doesn't do. Since changes in probability of shutdown drastically change the attainable utility, any such changes will be heavily penalized. We can use this dynamic to our advantage, for example by committing to shut down the agent if we see it doing something we disapprove of. My opinion: This is quite a big improvement for impact measures -- it meets many desiderata that weren't satisfied simultaneously before. My main critique is that it's not clear to me that an AUP-agent would be able to do anything useful. For example, perhaps the action used to define the impact unit is well-understood and accepted, but any other action makes humans a little bit more likely to turn off the agent. Then the agent won't be able to take those actions. Generally, I think that it's hard to satisfy the conjunction of three desiderata -- objectivity (no dependence on values), safety (preventing any catastrophic plans) and non-trivialness (the AI is still able to do some useful things). There's a lot more discussion in the comments. Realism about rationality (Richard Ngo): In the same way that moral realism claims that there is one true morality (even though we may not know it yet), rationality realism is the claim that there is one "correct" algorithm for rationality or intelligence. This post argues that many disagreements can be traced back to differences on how much one identifies with the rationality realism mindset. For example, people who agree with rationality realism are more likely to think that there is a simple theoretical framework that captures intelligence, that there is an "ideal" decision theory, that certain types of moral reasoning are "correct", that having contradictory preferences or beliefs is really bad, etc. The author's skepticism about this mindset also makes them skeptical about agent foundations research. My opinion: This does feel like an important generator of many disagreements I've had. I'd split rationality real...
Explicit: No
Rationality and The English Language
Release Date: 03/04/2015
Description: Book II: How to Actually Change Your Mind - Part F: Politics and Rationality - Rationality and The English Language
Explicit: No
The Scales of Justice, the Notebook of Rationality
Release Date: 03/04/2015
Description: Book II: How to Actually Change Your Mind - Part F: Politics and Rationality - The Scales of Justice, the Notebook of Rationality
Explicit: No
Similar Podcasts
Rationality: From AI to Zombies - The Podcast
Release Date: 03/24/2021
Authors: Walter and James
Description: Podcast of "Rationality: From AI to Zombies" by Eliezer Yudkowsky (chapters)
Explicit: No
Rationality: From AI to Zombies - The Podcast
Release Date: 08/21/2020
Authors: Walter & James
Description: Rationality: From AI to ZombiesChapter by ChapterWritten by Eliezer Yudkowsky ; Read by Walter & James
Explicit: No
ZombiesHeroes.com: Movie & TV Show Reviews - Zombies Heroes
Release Date: 08/12/2020
Authors: Rohan Prashanth
Description: ZombiesHeroes.com: A Movie & TV Show Review Podcast - From Zombies To Super Heroes And Everything In-between - Zombies Heroes
Explicit: No
LIBRARY OF THE LIVING DEAD
Release Date: 03/22/2021
Authors: LIBRARY OF THE LIVING DEAD
Description: A podcast dedicated to zombies and only zombies.
Explicit: No
Rational Chidiya
Release Date: 01/05/2022
Authors: Pratibha Shrivastava
Description: A Hindi Show to bring rationality with positivity
Explicit: No
Rationality
Release Date: 09/18/2021
Authors: Kendra Hummel
Description: A podcast that talks about the voting age. Should the Voting age be raised or lowered? Many have different views of this argument and we need to state the facts of how we should settle this argument.
Explicit: No
Rationality
Release Date: 04/27/2021
Authors: Hector Kidwell
Description: Exchanging Views, Not Insults - Political and societal discourse is so broken; entrenched and closed minded. It's like people pick a position and stick to it regardless of changing facts. Join the three of us, who all have very distinct political outlooks (one right, one left, one centre) there can be productive and valuable exchange of views, where conversation is respectful, passionate and fascinating. Get in touch with Hector on Twitter: @hectorkidwell See acast.com/privacy for privacy and opt-out information.
Explicit: No
The Zombie iPocalypse
Release Date: 09/23/2020
Authors: Zombie iPocalypse
Description: A Podcast by Zombies, For Zombies, about the iPocalypse. Each episode we will review and discuss topics stemming from the horror genre and industry
Explicit: No
Zombies To Zoinks!
Release Date: 08/24/2020
Authors: Zombies to Zoinks - Daniel, Stephan, & Trask
Description: Is Shaggy Rogers the illegitimate son of Captain America? What would happen if we sorted every Disney villain into a Hogwarts house? Could Winnie the Pooh and gang survive a zombie apocalypse? We're Daniel, Stephan and Trask and THESE are the important questions boggling the minds of nerds the world over covering everything from Zombies to Zoinks! If you're stuck inside just like the rest of us or are looking for some sultry voices to soothe you on your long distance post-apocalyptic travel excursions, then you've found the right podcast. Have a question or topic suggestion? Email us at [email protected] out our Link Tree to find us on Facebook, Twitter, Instagram, YouTube and more!https://linktr.ee/zombiestozoinks
Explicit: No
One Great Leap Presented by AI Truth
Release Date: 08/12/2020
Authors: Noah Wittman
Description: Artificial Intelligence is changing the world and we are taking One Great Leap into the next frontier. But, the ways that AI is changing the world are not always accurately portrayed by popular media. Join us as we explore the truth about AI, why EVERYONE needs to care, and why it doesn't involve killer robot zombies.
Explicit: No