
EP 11 - Open ai OSS via open coder, Langchain open SWE, local inference with ollama turbo, virtual audiences content testing
We recorded August 7th, right before ChatGPT launched.
We dove into GPT open source, OpenCode, Ollama Turbo, and deep agent setups.
I wanted to see LangChain’s open suite and test agent environments.
OpenCode stood out for its flexibility — multiple model providers, easy local setup, works with Ollama Turbo for $20/month.
LM Studio runs similarly.
I’m considering a high-spec NVIDIA rig and DGX Spark for local inference.
GPT-OSS is cheap, fast, and excellent for coding and tool-calling, but weaker on general knowledge.
Running it locally means more setup work but more control.
Hybrid local-plus-cloud routing feels inevitable.
We demoed OpenAgent Platform — fast, multi-provider agents without writing code.
Then explored LangChain SWE — an open-source, multi-threaded coding agent with planner/programmer loops, GitHub integration, Daytona sandboxes, and detailed token-cost tracking.
We looked at Vercel’s v0 API for quick generative UI, and the potential to run it privately for internal teams.
I closed with Google’s upcoming AI-mode ads and Societies.io — a virtual audience simulation tool for testing and optimizing content before publishing.
Chapters
00:00 Introduction to ChatGPT Launch and Demos
01:40 Exploring Open Code and LangChain
04:37 Local Inference and Olamma Integration
07:25 Cloud Acceleration with Turbo Service
10:11 Open Source Model Benchmarks and Feedback
ข้อมูล
- รายการ
- ความถี่อัปเดตทุกสัปดาห์
- ออกอากาศวันที่8 สิงหาคม 2568 เวลา 20:44 UTC
- ความยาว1 ชม. 39 นาที
- ซีซัน1
- ตอน11
- การจัดระดับเหมาะสม