This Week in Responsible AI: Oct 12, 2023
This Week in Responsible AI: Oct 12, 2023
New Research
-
Who Audits the Auditors? Recommendations from a field scan of the algorithmic auditing ecosystem
-
Rigorously Assessing Natural Language Explanations of Neurons
-
Sparse Autoencoders Find Highly Interpretable Features in Language Models
-
Goodtriever: Adaptive Toxicity Mitigation with Retrieval-augmented Models
Representation
Policy
-
CODE IS SPEECH, AND SPEECH IS FREE: An argument in favor of open-sourcing AI
-
Fight for the Future’s Lia Holland On A.I. Copyright, Human Art and More
Privacy
Generative AI
-
AI firms working on “constitutions” to keep AI from spewing toxic content
-
Stable Signature: A new method for watermarking images created by open source generative AI
-
(CW: descriptions of anti-semitism) The Folly of DALL-E: How 4chan is Abusing Bing’s New Image Model
Compiled by Leif Hancox-Li
Don't miss what's next. Subscribe to This Week in Responsible AI: