The Build - Ai dev and product show.

EP 11 - Open ai OSS via open coder, Langchain open SWE, local inference with ollama turbo, virtual audiences content testing

We recorded August 7th, right before ChatGPT launched.

We dove into GPT open source, OpenCode, Ollama Turbo, and deep agent setups.

I wanted to see LangChain’s open suite and test agent environments.

OpenCode stood out for its flexibility — multiple model providers, easy local setup, works with Ollama Turbo for $20/month.

LM Studio runs similarly.

I’m considering a high-spec NVIDIA rig and DGX Spark for local inference.

GPT-OSS is cheap, fast, and excellent for coding and tool-calling, but weaker on general knowledge.

Running it locally means more setup work but more control.

Hybrid local-plus-cloud routing feels inevitable.

We demoed OpenAgent Platform — fast, multi-provider agents without writing code.

Then explored LangChain SWE — an open-source, multi-threaded coding agent with planner/programmer loops, GitHub integration, Daytona sandboxes, and detailed token-cost tracking.

We looked at Vercel’s v0 API for quick generative UI, and the potential to run it privately for internal teams.

I closed with Google’s upcoming AI-mode ads and Societies.io — a virtual audience simulation tool for testing and optimizing content before publishing.

Chapters

00:00 Introduction to ChatGPT Launch and Demos

01:40 Exploring Open Code and LangChain

04:37 Local Inference and Olamma Integration

07:25 Cloud Acceleration with Turbo Service

10:11 Open Source Model Benchmarks and Feedback