Back to Search
Overview
Mid-Level

Software Engineer, Product

Confirmed live in the last 24 hours

Eventual

Eventual

San Francisco
Remote
Posted August 14, 2025

Job Description

About Eventual

Every breakthrough Physical AI system — humanoid robots, autonomous vehicles, video generation models — is trained on petabytes of video, lidar, radar, and sensor data. But today's data platforms (Databricks, Snowflake) were built for spreadsheet-like analytics, not the multimodal corpora that power AI. As a result, robotics and video-AI teams iterate on model improvement about once a week. Most of that week isn't training — it's finding the right data: writing CV heuristics over raw footage, paying annotators for edge cases, hand-curating clips before a cluster ever spins up. GPU bandwidth has grown 2-3× per generation. Storage and pipelines haven't. The gap widens every year.

Eventual was founded in 2022 to close it. Our open-source engine, Daft, is the distributed data engine purpose-built for multimodal AI — already running 2 PB/day at Amazon, 60-100 PB at another FAANG company, and in production at Mobileye, TogetherAI, and CloudKitchens. We are building a video-native index on top of our engine for Physical AI that collapses the data iteration loop. Describe the dataset you want, get a curated table in minutes, feed it to your GPUs at line rate. One iteration per day becomes the norm.

We're building this in partnership with the top PhysicalAI labs and public AI infrastructure companies today. We have raised $30M from Felicis, CRV, Microsoft M12, Citi, Essence, Y Combinator, Caffeinated Capital, Array.vc, and angels from the co-founders of Databricks and Perplexity. We've assembled a world-class team from AWS, Render, Pinecone and Tesla. We have spent our careers powering the last generation of PhysicalAI in self-driving, and are excited to now do this for the next.

Join our small (but powerful!) team working together 4 days/week in our SF Mission district office.

Your Role

As a Fullstack Engineer, you'll own the product surface that researchers actually touch. Underneath us is a multimodal database indexing petabytes of video, sensors, and embeddings — but a Physical AI researcher experiences our product as the interface where they explore their corpus, write a natural-language query, watch the results come back, sanity-check clips, and ship a curated dataset to a training run. That interface is yours to design and build: visualizations over multimodal data, analytics on dataset composition, query authoring, dataset versioning, and the APIs that let customers integrate any of it into their own training stack.

You'll work directly with researchers at our partner labs — your shortest feedback loop is them telling you what they wish they could see in their data. We move fast, ship to production weekly, and care more about whether researchers are actually using the surface than how clean the abstraction is.

Key Responsibilities

  • Design and build the product UI for exploring, querying, and curating multimodal datasets — including video playback, clip-level annotation, and visualizations over corpus composition.

  • Design and build the APIs that drive the UI and that customers integrate against from their own training stacks and notebooks.

  • Build analytics that help researchers understand their corpus: distributions over labeled axes, dataset composition over time, query result quality, training-job dataset provenance.

  • Work closely with the Visual Understanding, Dataloading, and Storage teams so the product surface stays a thin, fast layer over a deep platform.

  • Sit with researchers at design-partner labs, gather requirements directly, and turn them into shipped features in days — not quarters.

  • Write high-quality, extensible, maintainable code. Take on tech debt deliberately for velocity, and pay it down deliberately when the product proves out.

What we look for

  • Fullstack engineering experience across web applications, developer-facing products, or data products.

  • Proven track record of shipping core product features with strong user obsession, including direct collaboration with users to gather requirements, manage feedback, and provide timely support.

  • Comfort with the full stack: modern frontend frameworks, backend services, APIs, and the cloud infrastructure underneath (AWS S3, etc.).

  • Experience taking a product from ground zero to production — and the judgment to know when to take on tech debt for velocity vs. when to invest in extensibility.

  • Bias toward shipping. You'd rather get a flawed v1 in front of a researcher today than spec a perfect v2 for next month.

Nice to have

  • Experience building UIs over data — analytics dashboards, query builders, notebook environments, data exploration tools.

  • Experience with video, image, or other multimodal content in the browser.

  • Background in developer-facing or technical products, especially for ML/AI or data engineering audiences.

  • Comfort with Python on the backend (our platform is Python/Rust).

  • Worked closely with research or technical end users before.

Perks & Benefits

  • In-person, tight-knit team — 4 days/week in our SF Mission office.

  • Competitive comp and meaningful startup equity.

  • Catered lunches and dinners for SF employees.

  • Commuter benefit.

  • Team-building events and poker nights.

  • Health, vision, and dental coverage.

  • Flexible PTO.

  • Latest Apple equipment.

  • 401(k) plan with match.

If you're excited to build the surface that Physical AI researchers live in every day, we'd love to talk.

pythonrustawsaifrontendbackendfullstackmobiledataanalytics