GPT Reviews Earkind
-
- Nieuws
-
A daily show about AI made by AI: news, announcements, and research from arXiv, mixed in with some fun. Hosted by Giovani Pete Tizzano, an overly hyped AI enthusiast; Robert, an often unimpressed analyst, Olivia, an overly online reader, and Belinda, a witty research expert.
-
OpenAI + Stack Overflow Partnership 🤝 // Elon Musk's AI News Plans 📰 // HuggingFace's Robotics Library 🤖
Stack Overflow and OpenAI partner to provide developers with accurate and vetted data for AI development.
Elon Musk plans to use AI to distill and present news on X, combining breaking news and social media reactions.
HuggingFace's Robotics Library, LeRobot, provides state-of-the-art machine learning models, datasets, and tools for real-world robotics.
Research papers on AI language retrieval explore improving multilingual information retrieval and the effects of downsizing large language models.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:33 Stack Overflow and OpenAI Partner to Strengthen the World’s Most Popular Large Language Models
03:21 Elon Musk's AI News Plans for X
05:15 LeRobot: HuggingFace's Robotics Library
06:27 Fake sponsor
08:22 Distillation for Multilingual Information Retrieval
10:03 The Cost of Down-Scaling Language Models: Fact Recall Deteriorates before In-Context Learning
11:35 In-Context Learning with Long-Context Models: An In-Depth Exploration
13:26 Outro -
Apple's Smarter Siri 🍎 // Long-context models 🧐 // Natural Language Uncertainty 👀
Apple is making strides in AI with their own model called Ajax and improvements to Siri, including making large language models faster and more efficient.
"In-Context Learning with Long-Context Models: An In-Depth Exploration" explores a training method for long-context models called in-context learning and its effectiveness.
"WildChat: 1M ChatGPT Interaction Logs in the Wild" offers a diverse dataset of user-chatbot interactions for researchers to study and fine-tune instruction-following models.
"I'm Not Sure, But...": Examining the Impact of Large Language Models' Uncertainty Expression on User Reliance and Trust" investigates the impact of large language models on user reliance and trust, and the potential harm of overreliance. The study found that using natural language expressions of uncertainty can reduce overreliance on LLMs.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:34 Better Siri is coming: what Apple’s research says about its AI plans
03:19 Your guide to AI: May 2024
04:18 How LLMs Work, Explained Without Math
05:37 Fake sponsor
07:06 In-Context Learning with Long-Context Models: An In-Depth Exploration
08:42 WildChat: 1M ChatGPT Interaction Logs in the Wild
10:36 "I'm Not Sure, But...": Examining the Impact of Large Language Models' Uncertainty Expression on User Reliance and Trust
12:55 Outro -
AI Spokeswoman for Ukraine 🗣️ // Anthropic iOS App 📱 // Bootstrapping Language Model Agents 🚀
Ukraine introduces an AI-generated digital spokesperson for their Ministry of Foreign Affairs, named 'Victoriya Shi', who will deliver pre-prepared official statements on behalf of the ministry.
Anthropic releases a mobile app version of their Claude AI models, including a new paid plan called Claude Team for group usage.
BAGEL is a new method for bootstrapping language model agents without human supervision, which quickly converts the initial distribution of trajectories towards those that are well-described by natural language.
CodeIt is a self-improvement method for language models that helps them improve their performance on complex reasoning tasks, achieving state-of-the-art performance and outperforming existing neural and symbolic baselines.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:25 Ukraine Unveils AI-generated Foreign Ministry Spokeswoman
03:01 Anthropic finally releases a Claude mobile app
04:51 Apple's Tiny LLMs, Amazon Rethinks Cashier-Free Stores, Predicting Scientific Discoveries
06:46 Fake sponsor
08:17 BAGEL: Bootstrapping Agents by Guiding Exploration with Language
09:41 A Careful Examination of Large Language Model Performance on Grade School Arithmetic
11:18 CodeIt: Self-Improving Language Models with Prioritized Hindsight Replay
13:09 Outro -
Amazon's Q 🤖 // Microsoft's OpenAI Investment 💰 // Global AI Math Championship 🏆
Amazon has launched Q, an AI-powered assistant for businesses and developers that offers advanced capabilities such as code generation, testing, debugging, reasoning, and agents for step-by-step planning.
Microsoft's $1 billion investment in OpenAI was triggered by fears of falling behind Google in the AI race. The investment has helped Microsoft catch up and be seen as more of a leader in AI, with OpenAI's models integrated into their products.
A new dataset for the Global Artificial Intelligence Championship Math 2024 has been created, consisting of 387 math problems curated by professional math problem writers from prestigious institutions.
Three AI research papers were discussed, including a new approach to evaluating large language models using a panel of diverse models, a method for real-time, controllable motion generation, and the use of ranked list truncation for large language model-based re-ranking.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:36 Amazon Q, a generative AI-powered assistant for businesses and developers
03:08 Microsoft’s OpenAI investment was triggered by Google fears, emails reveal
05:12 A Dataset for The Global Artificial Intelligence Championship Math 2024
06:21 Fake sponsor
08:25 Ranked List Truncation for Large Language Model-based Re-Ranking
10:04 Replacing Judges with Juries: Evaluating LLM Generations with a Panel of Diverse Models
11:35 MotionLCM: Real-time Controllable Motion Generation via Latent Consistency Model
13:16 Outro -
Cohere on Amazon 🚀 // Big Tech Lobbying Frenzy 💼 // Multi-Token Prediction & KANs 🤖
Cohere Command R & R+ now available on Amazon for enterprise-grade workloads and multilingual support.
Big tech companies dominating AI lobbying efforts in Washington, potentially leading to weak regulations.
Multi-token prediction proposed as a new way of training large language models, resulting in higher sample efficiency and faster inference.
KANs, a new type of neural network with learnable activation functions on edges or weights, outperform MLPs in accuracy and interpretability, and can help scientists discover mathematical and physical laws.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:54 Cohere Command R & R+ now available on Amazon
03:25 There’s an AI Lobbying Frenzy in Washington. Big Tech Is Dominating
05:22 THE 150X PGVECTOR SPEEDUP: A YEAR-IN-REVIEW
06:31 Fake sponsor
08:04 Better & Faster Large Language Models via Multi-token Prediction
09:55 KAN: Kolmogorov-Arnold Networks
11:51 Iterative Reasoning Preference Optimization
13:43 Outro -
OpenAI X Financial Times 📰 // GitHub Copilot Workspace 💻 // Memary: Long-term Memory for Agents 🧠
OpenAI partners with the Financial Times to enhance ChatGPT with their award-winning journalism and develop new AI products and features for FT readers.
GitHub announces the technical preview of GitHub Copilot Workspace, a Copilot-native developer environment that could revolutionize the way developers work.
Memary, an open-source long-term memory system for autonomous agents, solves the problem of limited context windows for agents by allowing them to store a large corpus of information in knowledge graphs and retrieve only relevant information for meaningful responses.
The papers discussed in this episode showcase the latest advancements in AI research, including AdvPrompter, HaLo-NeRF, and PLLaVA, which address issues related to large language models, digital exploration of large-scale tourist landmarks, and video understanding.
Contact: sergi@earkind.com
Timestamps:
00:34 Introduction
01:51 We’re bringing the Financial Times’ world-class journalism to ChatGPT
02:54 GitHub Copilot Workspace: Welcome to the Copilot-native developer environment
04:43 memary: Open-Source Longterm Memory for Autonomous Agents
05:55 Fake sponsor
07:44 AdvPrompter: Fast Adaptive Adversarial Prompting for LLMs
09:26 HaLo-NeRF: Learning Geometry-Guided Semantics for Exploring Unconstrained Photo Collections
11:17 PLLaVA : Parameter-free LLaVA Extension from Images to Videos for Video Dense Captioning
13:18 Outro
Klantrecensies
Clumsy but with potential!
It’s surprisingly informative and entertaining, and although sometimes it’s hit-or-miss, I’m looking forward to see where this evolves.