
09x02: Moving Beyond Text for Agentic AI Applications with ApertureData
Our online interactions include audio, video, and sensor data, but most AI applications are still focused on text. This episode of Utilizing Tech considers how we can integrate multimodal data with agentic applications with Vishakha Gupta, founder and CEO of ApertureData, Frederic Van Haren of HighFens, and Stephen Foskett of Tech Field Day. After decades of developing AI models to process spoken word, images, video, and other multimodal data, the ascendance of large language models has largely focused on text. This is changing, as AI applications are increasingly leveraging multimodal data, including text, audio, video, and sensors. Many agentic applications still pass data as structured or unstructured text, but it is possible to use multimedia data as well, for example passing a clip of a video from agent to agent if the system has true multimodal understanding. Enterprise applications are moving beyond text to include voice and video, data in PDFs like charts and diagrams, medical sensors and images, and more.
Guest:
Vishakha Gupta, CEO and Founder, ApertureData
Hosts:
Stephen Foskett, President of the Tech Field Day Business Unit and Organizer of the Tech Field Day Event Series
Frederic Van Haren, Founder and CTO of HighFens, Inc.
Guy Currier, Chief Analyst at Visible Impact, The Futurum Group.
For more episodes of Utilizing Tech, head to the dedicated website and follow the show on X/Twitter, on Bluesky, and on Mastodon.
信息
- 节目
- 频率一周一更
- 发布时间2025年10月6日 UTC 13:00
- 长度38 分钟
- 季9
- 单集2
- 分级儿童适宜