Steering AIs

Sat May 10 2025

I’ve been spending some time thinking about ideas that Vitalik shared about having “AI as the engine, humans as the steering wheel” and participating in some of the experiments they are doing around them. As a Kaggle enthusiast and data person working around the blockchain space, experiments like Deepfunding align perfectly with my interests and skillset!

What is the main idea?

The gist is simple. Leverage ML models (I don’t really like calling these “AI”) as a powerful “engine” to execute a vast number of decisions, while humans provide high-quality, concise guidance (training and test data) as the “steering wheel”.

Models scale more efficiently than humans. Adding human oversight on values and objectives to the models make them useful. Instead of relying on a single, complex AI model, what if we create a simple, open, and stable mechanisms (like a Kaggle competition of sorts) where diverse models compete as players to best align with the human data and goals? Humans set the “objective function” and/or verify model outputs, while models compete to optimize for it like, again, in a Kaggle competition.

Building Useful Models

So far, I’m having fun participating in the ML competitions hosted by Pond. Right now, there is a diverse set of competitions that are open to anyone. Download or gather some data, train a model, submit to the leaderboard and see how it performs! Reward prizes are generous and for now, there are not that many participants.

While the experience is engaging and is working well (especially given that these are new processes and infrastructure), I’ve been thinking about how we could make it better for every party involved (hosts and participants).

For that, let me share a very high level view of the challenges I’ve noticed so far.

  1. Transparency and Verifiability (credible neutrality)
    • Once the competition ends, how do we know that the leaderboard is what it is without trusting the organizers?
    • How do we check that there were no bugs in the scoring script?
  2. Mechanism Design and Evaluation
    • Is perhaps the Nth model better at a subset of the data than the top 3 models? Shouldn’t that be rewarded?
    • How can prizes get fairly distributed in a way that doesn’t incentivize gaming the system? Right now, the incentive is to create many accounts in the competition platform and send lots of submissions just in case there is a great one.
  3. Data and Metrics Diversity
    • How can collectively improve the competition designs to allow more diverse data and metrics to avoid Goodhart’s law?
  4. Trust and Incentives
    • Is there a way to run things without a central trusted party (insider advantage, selling datasets, …)? Pond is great but, wouldn’t it be better if these competitions were viable without a central platform?
    • How to incentivize reporting bad behavior (whitehat style) or raising issues with the competition that will make the competition better?

Some Ideas

I want to mention again that this space is very novel and still evolving and learning! It is great to have the opportunity to participate and help shape the future of programs like Deepfunding!

Alright, here are a few things I thought would make the experience better.

Conclusion

The space is moving fast and it’s exciting to be part of it! Participating in these competitions is a great way to learn about ML and explore interesting problems. Looking forward to see how this evolves and how can I use my skills to make it better!

← Back to home!