
EP 11 - Open ai OSS via open coder, Langchain open SWE, local inference with ollama turbo, virtual audiences content testing
We recorded August 7th, right before ChatGPT launched.
We dove into GPT open source, OpenCode, Ollama Turbo, and deep agent setups.
I wanted to see LangChain’s open suite and test agent environments.
OpenCode stood out for its flexibility — multiple model providers, easy local setup, works with Ollama Turbo for $20/month.
LM Studio runs similarly.
I’m considering a high-spec NVIDIA rig and DGX Spark for local inference.
GPT-OSS is cheap, fast, and excellent for coding and tool-calling, but weaker on general knowledge.
Running it locally means more setup work but more control.
Hybrid local-plus-cloud routing feels inevitable.
We demoed OpenAgent Platform — fast, multi-provider agents without writing code.
Then explored LangChain SWE — an open-source, multi-threaded coding agent with planner/programmer loops, GitHub integration, Daytona sandboxes, and detailed token-cost tracking.
We looked at Vercel’s v0 API for quick generative UI, and the potential to run it privately for internal teams.
I closed with Google’s upcoming AI-mode ads and Societies.io — a virtual audience simulation tool for testing and optimizing content before publishing.
Chapters
00:00 Introduction to ChatGPT Launch and Demos
01:40 Exploring Open Code and LangChain
04:37 Local Inference and Olamma Integration
07:25 Cloud Acceleration with Turbo Service
10:11 Open Source Model Benchmarks and Feedback
信息
- 节目
- 频率一周一更
- 发布时间2025年8月8日 UTC 20:44
- 长度1 小时 39 分钟
- 季1
- 单集11
- 分级儿童适宜