🛡️

AI for Epistemics Hackathon

Linkpost for https://manifund.substack.com/p/ai-for-epistemics-hackathon

AI for Epistemics is about helping to leverage AI for better truthseeking mechanisms — at the level of individual users, the whole of society, or in transparent ways within the AI systems themselves. Manifund & Elicit recently hosted a hackathon to explore new projects in the space, with about 40 participants, 9 projects judged, and 3 winners splitting a $10k prize pool. Read on to see what we built!

image

Resources

  • See the project showcase: https://moxsf.com/ai4e-hacks
  • Watch the recordings: project demos, opening speeches
  • See the outline of project ideas: link
    • Thanks to Owen Cotton-Barratt, Raymond Douglas, and Ben Goldhaber for preparing this!
  • Lukas Finnveden on “What's important in ‘AI for epistemics’?”: link
  • Automation of Wisdom and Philosophy essay contest: link

Why this hackathon?

From the opening speeches; lightly edited.

Andreas StuhlmĂźller: Why I'm excited about AI for Epistemics

In short - AI for Epistemics is important and tractable.

Why is it important? If you think about the next few years, things could get pretty chaotic. As everyone rushes to integrate AI systems into every part of the economy, the world could change more rapidly than it does today. There's significant risk that people and organizations will make mistakes for relatively uninteresting reasons—simply because they didn't have enough time to think things through.

If we can make it easier for people to think clearly and carefully, that's really important. People will use AI tools to help them make decisions either way; eventually unassisted decision-making just won’t be competitive anymore. This is a lever: the more these tools actually help people make wise decisions, or help them figure out whether they're right or wrong about something, the better off we'll be.

AI for Epistemics is also tractable now in a way it wasn't before. We're just reaching the point where models are good enough and cheap enough to apply at scale. You can now realistically say, "Let's analyze all news articles," or "Let's review all scientific papers," or thoroughly check every sentence of a document, at a level of detail that wasn't feasible before.

Given good ideas for epistemic tools, the implementation cost has dropped dramatically. Building significant products in hackathons has become much easier. You can basically copy and paste your project description into Cursor, type "please continue" five times, and you'll have a working demo.

The key challenge we'll need to think about today is: how can we tell if we're actually making things better? What evidence can we see that would lead us to believe a tool genuinely improves people's thinking, rather than just being a fun UI with knobs to play with?

I'm really excited about this hackathon. This is the event I've been most excited about for quite a while. I'm very grateful to Austin for creating this space for us.

Austin Chen: Why a hackathon?

Andreas first talked to me a couple months ago, saying we want to do more for the AI for Epistemics field. We were thinking about some ideas: “oh, maybe we should do a grants program, or a fellowship program, or something like that”.

But I have a special place in my heart for hackathons specifically. So I really sold him hard: we're gonna do a hackathon. We can do all that other stuff too later, but: first things first. (Andreas, wryly: “I was very hard to sell.”)

I like hackathons for a lot of reasons:

  • Hackathons are a sandbox. They're a place where you can play with an idea a little bit. You don't have to worry about whether this thing will be great down the line, or even live past the end of the day. So it gives you a chance to be a bit more creative, try riskier things.
  • It's a blank canvas. You don't have to worry about what your current users will think, or “will this make money?”. You can just… do stuff.
  • Hackathons are a forcing function. There's the demos in eight hours. We're all gonna get up there and present and talk about what we did. So you have to sit there and actually build your idea. You can't just keep spinning your wheels, thinking forever.
  • And it's a chance to meet people. It's a filtering function to find people who care a lot about this particular niche. Right now, AI for Epistemics is a tiny field. All the people who care about it are maybe in this room right now (plus a few others who are remote). But hopefully, it will grow down the line. And this is your chance to meet each other, talk to each other, share your ideas, build stuff out.

Those are some of the reasons I'm excited about hackathons. I'm glad that Andreas and the Elicit team are happy to host this with us today.

… (projects omitted)

What went well

  • Hacks were pretty cool! Especially given that they all represented ~8 hours of work
    • Many are minimal versions of products I really want to exist, and play with more
    • Almost all of them felt like a worthwhile showcase, exploration of something interesting, and relevant to this field
  • Lots of great people came for this! Very hard to think of more central folks for AI for Epistemics:
  • our lovely faces, once more

    image
  • From left to right: Rafe Kennedy, Oli Habryka, Evan Hadfield, Kirill Chesnov, Owain Evans, Charlie George, Panda Smith, Gustavo Lacerda, Andreas StuhlmĂźller, Austin Chen, David Nachbach (virtual), Lukas Finnveden, Tamera Lanham, Noa Nabeshima, Campbell Hutcheson, Keri Warr, Xyra Sinclair, Tilman Bayer, Raymon Arnold, Chris Lakin, Ozzie Gooen
    1. Not pictured participants and viewers: William Saunders, Ben Goldhaber, Deger Turan, Vishal Maini, Ross Rheingans-Yoo, Ethan Alley, Dan Selsam, Stephen Grugett, David Chee, Saul Munn, Gavriel Kleinwaks and many others…

    2. Some of the goal of the hackathon, as with any event, is just to bring people together and have them stay in contact
  • Good conversations at the beginning while participants were ideating, and throughout. Both on AI for Epistemics and other topics.
  • Overall event felt smooth and cohesive, especially for being pulled together on not that much organizer time
    • Pretty happy with the continuing artifacts that we produced out of this hackathon (the showcase page, the video recordings, this writeup)
    • Somewhat more effortful to do all this compared to typical hackathons Austin has run, but hopefully worthwhile when trying to incubate a new field
  • Mox seemed to be a good venue for this event! This was just our third week of operating, but I think our venue supported the hackathon well.
    • One participant remarked:
Something I like about your office is that it seems to naturally create the Cal Newport Deep Work architecture, where the further in you go the more deepworky it is

What could have gone better

  • Fewer submitted hacks than we’d hoped for
    • Had ~40 people around but only ~10 submissions. Ideally more like 15-20?
    • Maybe we should have promoted this event harder, or cast a wider net?
      • There’s a tradeoff on average participant quality vs number of submissions.
      • But maybe projects are hit-based, so having the best projects matters more than having a high average quality
    • Maybe try to get higher commitment from folks, if we run this again
  • Hoping to have discovered more people from outside our current networks, who are excited for AI for Epistemics
    • 2 of the 3 prizes went to teams from Elicit
      • (Which says something about how great the Elicit team is, in case anyone out there is thinking about finding a new job…)
  • Unclear path to deployment for these projects, or continuing impact
    • Admittedly, this is a standard problem with a hackathon form factor, especially when the hackathon isn’t housed within an org/for specific product features
  • Not sure we made great use of the ideas doc & categories that Owen/Raymond/Ben compiled?
    • But hopefully, their work will set a stage of “this is what AI for Epistemics is about”
    • Perhaps having such specified categories was too confusing for participants
      • One participant asked, “do I have to do something in these categories?” (A: no, but it’s bad that this wasn’t clear)
  • As judges: hard to give great judgements and feedback in short amount of time, by just looking at demos and asking questions
    • The format of “demos in front of an audience” do bias towards presentation ability and flashiness, over usability of a core product
    • Might change up the structure for next time
      • More time for judges and audience to play with hackathon demos?
      • Open up voting to the public, so it’s more democratized?

Final notes

Overall, we’re very happy with how this hackathon turned out. Building a new field from scratch is difficult, high-dimensional problem, and this is just one step along the way; but I think we made meaningful progress, with ideas we brainstormed, hacks we demoed, and people we gathered.

After the end of the hackathon, a few of the judges and participants continued to discuss: “What’s next for AI for Epistemics? How does one build a nascent field? Is ‘AI for Epistemics’ even a good name?” We’ll try to share more on this in the coming days; until then, if AI for Epistemics excites you, leave a comment or reach out to us!