AI for Epistemics is about helping to leverage AI for better truthseeking mechanisms — at the level of individual users, the whole of society, or in transparent ways within the AI systems themselves.
Sponsored by Elicit & Manifund, with a $10k prize pool. Come hack alongside folks from Anthropic, Manifold, Metaculus, METR, OpenAI, OpenPhil, SF Compute & many others!
# Schedule
9:30 -- Doors open
10 -- Kickoff
12 -- Lunch
3 -- Break, boba run
6 -- Dinner (also, open to public!)
7 -- Demos on stage
8 -- Judging, feedback, awards
9 -- Continue to showcase projects & hang out
# What to hack on?
For the hackathon, we have three broad categories. These are things we’d be excited to see, and if you want to run with one of those, that seems great! But variations or completely different ideas might be even better!
More details in this outline, by Owen Cotton-Barratt: https://docs.google.com/document/d/18_h13T-Yx5dxN7DRP8WXgKk-twObsMgrOAXoM2GLJ1k/edit?tab=t.nip7ri77dom0
# Tools for Thought
AI systems could help people to think things through — to deepen their understanding and come to a more accurate sense of what they should be doing.
Example projects:
- Honest Friend — A system that will skewer a piece of writing in its review, going after its weakest points, but do so in an open, friendly, and constructive way
- Crux Finder — A system which helps two people who have a disagreement to quickly get to the bottom of the disagreement
# Tools for Awareness
There is a massive amount of information in the world, and that is presented to us as we browse the internet. AI could help us to make sense of this.
Example projects:
- Argument Parser — A system that takes a written text and maps out what its arguments actually consist of … what the implicit assumptions are, which pieces of evidence are offered in support of which conclusions, etc.
- Community Notes for Everything — Community Notes on Twitter/X seem like a great example of an epistemic intervention. An AI system could potentially simulate this process, and provide the information that would have been surfaced by community notes on any tweet … or more broadly on any sentence or paragraph on any website.
# Epistemic Evals
We only get what we can measure. Even if people want epistemically virtuous AI systems, that won’t happen unless we can assess what it means to be epistemically virtuous. Epistemic evals could be used as a research development tool; they could also be used to create public pressure to make systems better on the measured dimensions.
Example projects:
- Pedantry Metrics — It’s possible to use language to distort, mislead, or paper over things. It would be nice to train systems that would go out of their way to avoid doing that. Could we assess performance on this?
- Sycophancy Evals — Various research shows LLMs are often sycophantic in various ways. But could we automate assessment, so that there’s a go-to place on the internet which routinely scores publicly available systems?
# More ideas
See this post by Lukas Finnveden: https://www.lesswrong.com/posts/D2n5uduYGXuexkv7v/what-s-important-in-ai-for-epistemics-2
Loading...
Restricted Access
Verify your phone number to view event details and activity