Sandbox Cold-Start Benchmark: Vercel vs Daytona

I’ve been building an AI agent that needs to run arbitrary code — execute scripts, call tools, inspect outputs. For this I need an execution sandbox: an isolated Linux environment I can spin up on demand, run a command in, and tear down. The faster it starts, the more responsive the agent feels.

AI & Automation 2025 Yearly Wrap-Up

This post is an adaptation of an internal wrap-up I shared at Playtomic to celebrate our AI journey in 2025. I’ve kept the public-facing industry insights and general learnings while removing internal-only details.

Claude Code vs Cursor Experiment at Playtomic

At Playtomic, we believe in sharing our engineering journey with the broader tech community. Recently, we conducted a comprehensive experiment to compare Claude Code against our existing standard, Cursor, to evaluate how they fit into our diverse development workflows. This follows our previous experiment with AI code review tools, where we evaluated tools like BugBot, CodeRabbit, and GitHub Copilot. After an intensive six-week trial during September and October 2025, we gathered some valuable insights that we hope will help other teams navigating similar decisions.

Experimenting with AI Code Review Tools at Playtomic

At Playtomic, we are constantly exploring how AI can improve our development lifecycle. After experimenting with different IDE tools, we decided to turn our attention to the Pull Request review process.

Switching to AI Engineering

After 15 years focused on mobile development, I am excited to announce that I am switching to an AI engineering role, becoming the Principal AI Engineer at Playtomic.