Why AI might not be an existential risk to humanity W42
This week, we’re looking at counterarguments to the basic case for why AI is an existential risk to humanity, looking at how strong AI might come very soon, and sharing interesting papers.
🎉 You can now subscribe to our newsletter and listen to these updates in your favorite podcasting app. Check out newsletter.apartresearch.com and podcast.apartresearch.com.
This week, we’re looking at counterarguments to the basic case for why AI is an existential risk to humanity, looking at how strong AI might come very soon, and sharing interesting papers.
- Apply for the SERI MATS programme: https://ais.pub/serimats
- Apply for a fellowship from FLI for your PhD or postdoc: https://ais.pub/fli
- Join Redwood’s technical staff: https://ais.pub/redwoodjob
- Join CHAI’s research staff: https://ais.pub/chaiintern
- Visit the Alignment Jam page: https://ais.pub/jam
- Sign up for our newsletters: https://ais.pub/news
- Visit our podcast: https://ais.pub/pod
Sources:
- X-risk counter arguments: https://www.alignmentforum.org/posts/LDRQ5Zfqwi8GjzPYG/counterarguments-to-the-basic-ai-x-risk-case
- Response 1: https://www.alignmentforum.org/posts/GQat3Nrd9CStHyGaq/response-to-katja-grace-s-ai-x-risk-counterarguments
- Conversation: https://www.alignmentforum.org/posts/iXuJLARFBZbaBGxW3/a-conversation-about-katja-s-counterarguments-to-ai-risk
- Comprehensive AI services: https://www.lesswrong.com/posts/x3fNwSe5aWZb5yXEG
- Strong general AI coming soon: https://www.lesswrong.com/posts/K4urTDkBbtNuLivJx/why-i-think-strong-general-ai-is-coming-soon
- Nate’s decision theory != nice things: https://www.alignmentforum.org/posts/rP66bz34crvDudzcJ/decision-theory-does-not-imply-that-we-get-to-have-nice
- Moral RL agent: https://openreview.net/pdf?id=CtS2Rs_aYk
- Non-experts’ perspective on interpretability (high-stakes, otherwise accuracy): https://www.nature.com/articles/s41467-022-33417-3
- Neel Nanda’s interpretability paper list: https://www.lesswrong.com/posts/SfPrNY45kQaBozwmu/an-extremely-opinionated-annotated-list-of-my-favourite
- John’s bid for prediction markets: https://www.lesswrong.com/posts/oxSX9XDQHLu5YLpaD/how-to-make-prediction-markets-useful-for-alignment-work
- Physics on LLMs: https://arxiv.org/pdf/2210.05359.pdf
- Deep learning theory: https://www.lesswrong.com/posts/bumgqvRjTadFFkoAd/science-of-deep-learning-a-technical-agenda
