Decentralized Team Collaboration in AI Agent Development

AI Agents Hackathon — LabLab × MindsDB

196 teams, 77 MVPs, blind judging — remote and hybrid teams competing with onsite groups in San Francisco.

Teams coding at the AI Agents Hackathon at the MindsDB venue

Case study

Overview

The AI Agents Hackathon (LabLab and MindsDB) stress-tested autonomous agents and whether remote-first and hybrid teams could match or beat onsite execution.

Case study

Objectives

  • Ship functional AI agents for real workflows

  • Measure decentralized collaboration

  • Use cloud tooling and frameworks for distributed work

  • Validate production-style MVPs on a short clock

Case study

Format — September 2024

  • Hybrid: onsite in San Francisco and global remote

  • 4,000+ registered; 196 teams; 926 remote / 188 onsite

  • 77 final MVPs; 20+ expert judges

Case study

Top three (blind evaluation)

1st — Aquinas (fully remote)

Social media engagement agent — distributed time zones, async throughput, strong API usage.

2nd — DEV AI AGENT (onsite)

App-building agent on Llama 3.1 — fast iteration and low-latency local work.

3rd — TEMO (hybrid)

Emotional-support AI for children with autism — Composio, Upstage Solar Pro, coordination across zones.

Case study

Partners and stack

  • Upstage — credits, mentors, workshops

  • Composio — 100+ tools, GitHub automation

  • Together AI — credits, GPU

  • AI/ML API — broad model access

  • Meta (Llama 3.1) — judges, mentors, Llama Impact Grants

MetricValue
Teams196
MVPs delivered77
Remote teams in top 32
Onsite teams in top 31
JudgingBlind to team location

Case study

Takeaways

  • Remote-first delivery worked for advanced agent systems

  • Hybrid needs strong coordination but can win

  • Judges could not infer location from output quality

The event set a precedent for remote innovation — AI agents weren’t just built; they were built better, faster, and together.