EDGE AI POD

EDGE AI FOUNDATION

Discover the cutting-edge world of energy-efficient machine learning, edge AI, hardware accelerators, software algorithms, and real-world use cases with this podcast feed from all things in the world's largest EDGE AI community.  These are shows like EDGE AI Talks, EDGE AI Blueprints as well as EDGE AI FOUNDATION event talks on a range of research, product and business topics. Join us to stay informed and inspired!

  1. 2D AGO

    Powering Intelligence: Anaflash's Revolutionary AI Microcontroller with Embedded Flash Memory

    Memory bottlenecks, not computational limitations, are the true barrier holding back Edge AI. This revelation lies at the heart of Anaflash's revolutionary approach to intelligent edge computing – a breakthrough AI microcontroller with embedded flash memory that transforms how we think about power efficiency and cost in smart devices. The team has engineered a solution that addresses the two fundamental challenges facing Edge AI adoption: power efficiency and cost. Their microcontroller features zero-standby, power-weight memory with 4-bit per-cell embedded flash technology seamlessly integrated with computation resources. Unlike traditional non-volatile memory options that demand extra processing steps and offer limited storage density, this technology requires no additional masks and scales efficiently. At the core of this innovation is the Near Memory Computing Unit (NMCU), which establishes a tight coupling with flash memory through a wide I/O interface on a single chip. This architecture eliminates the need to fetch data from external memory after booting or waking from deep sleep – a game-changing feature for battery-powered devices. The NMCU's sophisticated three-part design enhances parallel computations while minimizing CPU intervention: control logic manages weight addresses and buffer flow, 16 processing elements share weights through high-bandwidth connections, and a quantization block efficiently converts computational results. Fabricated using Samsung Foundry's 28nm standard logic process in a compact 4 by 4.5 mm² die, the microcontroller delivers impressive results. Testing with MNIST and Deep Auto Encoder models demonstrates accuracy levels virtually identical to software baselines – over 95% and 0.878 AUC respectively. The overstress-free waterline driver circuit extends flash cell margins, further enhancing reliability and performance. Ready to transform your Edge AI applications with technology that combines unprecedented efficiency, performance, and cost-effectiveness? Experience the future of intelligent edge computing with Anaflash's embedded flash microcontroller – where memory and computation unite to power the next generation of smart devices. Send us a text Support the show Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    15 min
  2. NOV 4

    Enhancing Field Oriented Control of Electric Drives with tiny Neural Network

    Ever wondered how the electric vehicles of tomorrow will squeeze every last drop of efficiency from their batteries? The answer lies at the fascinating intersection of artificial intelligence and motor control. The electrification revolution in automotive technology demands increasingly sophisticated control systems for permanent magnet synchronous motors - the beating heart of electric vehicle propulsion. These systems operate at mind-boggling speeds, with control loops closing every 50 microseconds (that's 20,000 times per second!), and future systems pushing toward 10 microseconds. Traditional PID controllers, while effective under steady conditions, struggle with rapid transitions, creating energy-wasting overshoots that drain precious battery life. Our groundbreaking research presents a neural network approach that drastically reduces these inefficiencies. By generating time-varying compensation factors, our AI solution cuts maximum overshoots by up to 70% in challenging test scenarios. The methodology combines MatWorks' development tools with ST's microcontroller technology in a deployable package requiring just 1,700 parameters - orders of magnitude smaller than typical deep learning models. While we've made significant progress, challenges remain. Current deployment achieves 70-microsecond inference times on automotive-grade microcontrollers, still shy of our ultimate 10-microsecond target. Hardware acceleration represents the next frontier, along with exploring higher-level models and improved training methodologies. This research opens exciting possibilities for squeezing maximum efficiency from electric vehicles, turning previously wasted energy into extended range and performance. Curious about the technical details? Our complete paper is available on arXiv - scan the QR code to dive deeper into the future of smart motor control. Send us a text Support the show Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    17 min
  3. OCT 28

    Transforming Human-Computer Interaction with OpenVINO

    The gap between science fiction and reality is closing rapidly. Remember when talking to computers was just a fantasy in movies? Raymond Lo's presentation on building chatbots with OpenVINO reveals how Intel is transforming ordinary PCs into extraordinary AI companions. Imagine generating a photorealistic teddy bear image in just eight seconds on your laptop's integrated GPU. Or having a natural conversation with a locally-running chatbot that doesn't need cloud connectivity. These scenarios aren't futuristic dreams – they're happening right now thanks to breakthroughs in optimizing AI models for consumer hardware. The key breakthrough isn't just raw computational power but intelligent optimization. When Raymond's team first attempted to run large language models locally, they didn't face computational bottlenecks – they hit memory walls. Models simply wouldn't fit in available RAM. Through sophisticated compression techniques like quantization, they've reduced memory requirements by 75% while maintaining remarkable accuracy. The Neural Network Compression Framework (NNCF) now allows developers to experiment with different compression techniques to find the perfect balance between size and performance. What makes this particularly exciting is the deep integration with Windows and other platforms. Microsoft's AI Foundry now incorporates OpenVINO technology, meaning when you purchase a new PC, it comes ready to deliver optimized AI experiences out of the box. This represents a fundamental shift in how we think about computing – from tools we command with keyboards and mice to companions we converse with naturally. For developers, OpenVINO offers a treasure trove of resources – hundreds of notebooks with examples ranging from computer vision to generative AI. This dramatically accelerates development cycles, turning what used to take months into weeks. As Raymond revealed, even complex demos can be created in just two weeks using these tools. Ready to transform your PC into an AI powerhouse? Explore OpenVINO today and join the revolution in human-computer interaction. Your next conversation partner might be sitting on your desk already. Send us a text Support the show Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    43 min
  4. OCT 21

    Energy Efficient and high throughput inference using compressed tsetlin machine

    Logic beats arithmetic in the machine learning revolution happening at Newcastle University. From a forgotten Soviet mathematician's work in the 1960s to modern embedded systems, Settle Machine represents a paradigm shift in how we approach artificial intelligence. Unlike traditional neural networks that rely on complex mathematical operations, Settle Machine harnesses Boolean logic - simple yes/no questions similar to how humans naturally think. This "white box" approach creates interpretable models using only AND gates, OR gates, and NOT gates without any multiplication operations. The result? Machine learning that's not only understandable but dramatically more efficient. The technical magic happens through a process called Booleanization, converting input data into binary questions that feed learning automata. These finite state machines work in parallel, creating logical patterns that combine to make decisions. What's remarkable is the natural sparsity of the resulting models - for complex tasks like image recognition, more than 99% of potential features are automatically excluded. By further optimizing this sparsity and removing "weak includes," Newcastle's team has achieved astonishing efficiency improvements. The numbers don't lie: 10x faster inference time than Binarized Neural Networks, dramatically lower memory footprint, and energy efficiency improvements around 20x on embedded platforms. Their latest microchip implementation consumes just 8 nanojoules per frame for MNIST character recognition - likely the lowest energy consumption ever published for this benchmark. For edge computing and IoT applications where power constraints are critical, this breakthrough opens new possibilities. Beyond efficiency, Settle Machine addresses the growing demand for explainable AI. As regulations tighten around automated decision-making, the clear logical propositions generated by this approach provide transparency that black-box neural networks simply can't match. Ready to explore this revolutionary approach? Visit settlemachine.org or search for the unified GitHub repository to get started with open-source implementations. Send us a text Support the show Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    20 min
  5. OCT 14

    Applying GenAI to Mice Monitoring

    The AI revolution isn't just for tech giants with unlimited computing resources. Small and medium enterprises represent a crucial frontier for edge generative AI adoption, but they face unique challenges when implementing these technologies. This fascinating exploration takes us into an unexpected application: smart laboratory mouse cages enhanced with generative AI. Laboratory mice represent valuable assets in pharmaceutical research, with their welfare being a top priority. While fixed-function AI already monitors basic conditions like water and food availability through camera systems, the next evolution requires predicting animal behavior and intentions. By analyzing just 16 frames of VGA-resolution video, this edge-based system can predict a mouse's next actions, potentially protecting animals from harm when human intervention isn't immediately possible due to clean-room protocols. The technical journey demonstrates how generative AI can be scaled appropriately for edge devices. Starting with a 240-million parameter model (far smaller than headline-grabbing LLMs), the team optimized to 170 million parameters while actually improving accuracy. Running on a Raspberry Pi 5 without hardware acceleration, the system achieves inference times under 300 milliseconds – and could potentially reach real-time performance (30ms) with specialized hardware. The pipeline combines three generative neural networks: a video-to-my model, an OPT transformer, and a text-to-speech component for natural interaction. This case study provides valuable insights for anyone looking to implement edge generative AI in resource-constrained environments. While currently limited to monitoring single mice, the approach demonstrates that meaningful AI applications don't require supercomputers or billion-parameter models – opening doors for businesses of all sizes to harness generative AI's potential. Send us a text Support the show Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    17 min
  6. OCT 7

    Simple Cost Effective Vision AI Solutions at the edge

    Sony's revolutionary IMX500 stands at the forefront of a quiet revolution in edge computing and smart city technology. This isn't just another image sensor—it's the first to integrate AI processing directly on the chip, transforming how visual data becomes actionable intelligence while preserving privacy and minimizing infrastructure requirements. The power of this innovation lies in its elegant simplicity. Rather than sending complete images to cloud servers or external GPUs for processing, the IMX500 performs AI inference locally and transmits only the resulting metadata. This approach slashes bandwidth requirements to mere kilobytes, dramatically reduces power consumption, and—perhaps most critically—protects individual privacy by ensuring that identifiable images never leave the device. For urban environments where surveillance concerns often clash with safety imperatives, this represents a breakthrough compromise. Real-world deployments already demonstrate the technology's transformative potential. In Lakewood, Colorado, where a one-mile stretch of road had become notorious for traffic fatalities, Sony's solution achieved 100% performance in identifying dangerous situations—outperforming three competing technologies while costing less. Through partnership with ITRON, these sensors can be seamlessly deployed using existing streetlight infrastructure, creating mesh networks of intelligent sensors without requiring expensive new installation work or dedicated power sources. This practical approach to deployment makes citywide implementation financially viable even for budget-constrained municipalities. The implications extend far beyond traffic monitoring. From retail analytics to manufacturing quality control, the same core technology can be applied wherever visual intelligence provides value. By bringing AI to the edge in a form factor that addresses privacy, power, and practical deployment challenges, Sony has created a foundation for the next generation of smart infrastructure. Explore how this technology could transform your environment—whether an urban center, commercial space, or industrial facility—by leveraging the power of visual intelligence without the traditional limitations. Send us a text Support the show Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    19 min
  7. SEP 30

    Low Code No Code Platform for Developing AI algorithms

    Revolutionizing edge computing just got easier. This eye-opening exploration of ST Microelectronics' ST-IoT Craft platform reveals how everyday developers can now harness the power of artificial general intelligence without writing a single line of code. The modern IoT landscape presents a paradox: billions of devices generate zettabytes of valuable data, yet transforming that raw information into intelligent systems remains frustratingly complex. ST's innovative low-code/no-code platform elegantly solves this problem by distributing intelligence across three key components: smart sensors with embedded AI algorithms, intelligent gateways that filter data transmission, and cloud services that handle model training and adaptation. At the heart of this revolution is truly remarkable in-sensor AI technology. Imagine sensors that don't just collect data but actually think – detecting whether a laptop is on a desk or in a bag, whether an industrial asset is stationary or being handled, or whether a person is walking or running. These decisions happen directly on the sensor itself, dramatically reducing power consumption and network traffic while enabling real-time responses. The platform offers 31 different features including mean, variance, energy in bands, peak-to-peak values, and zero crossing that can be automatically selected and applied to your data. What makes ST-IoT Craft truly accessible is its browser-based interface with six pre-built examples spanning industrial and consumer applications. Users can visualize sensor data in real-time, train models with a single button click, and deploy finished solutions directly to hardware – all without diving into complex code. The platform even handles the intricate details of filter selection, feature extraction, window length optimization, and decision tree generation automatically. Ready to transform your IoT projects with embedded intelligence? Visit stcom, search for ST-IoT Craft, and discover how you can teach your sensors to think – no coding required. Send us a text Support the show Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    20 min
  8. SEP 23

    Stochastic Training for Side-Channel Resilient AI

    Protecting valuable AI models from theft is becoming a critical concern as more computation moves to edge devices. This fascinating exploration reveals how sophisticated attackers can extract proprietary neural networks directly from hardware through side-channel attacks - not as theoretical possibilities, but as practical demonstrations on devices from major manufacturers including Nvidia, ARM, NXP, and Google's Coral TPUs. The speakers present a novel approach to safeguarding existing hardware without requiring new chip designs or access to proprietary compilers. By leveraging the inherent randomness in neural network training, they demonstrate how training multiple versions of the same model and unpredictably switching between them during inference can significantly reduce vulnerability to these attacks. Most impressively, they overcome the limitations of edge TPUs by cleverly repurposing ReLU activation functions to emulate conditional logic on hardware that lacks native support for control flow. This allows implementation of security measures on devices that would otherwise be impossible to modify. Their technique achieves approximately 50% reduction in side-channel leakage with minimal impact on model accuracy. The presentation walks through the technical implementation details, showing how layer-wise parameter selection can provide quadratic security improvements compared to whole-model switching approaches. For anyone working with AI deployment on edge devices, this represents a critical advancement in protecting intellectual property and preventing system compromise through model extraction. Try implementing this stochastic training approach on your edge AI systems today to enhance security against physical attacks. Your valuable AI models deserve protection as they move closer to end users and potentially hostile environments. Send us a text Support the show Learn more about the EDGE AI FOUNDATION - edgeaifoundation.org

    19 min

Ratings & Reviews

4
out of 5
2 Ratings

About

Discover the cutting-edge world of energy-efficient machine learning, edge AI, hardware accelerators, software algorithms, and real-world use cases with this podcast feed from all things in the world's largest EDGE AI community.  These are shows like EDGE AI Talks, EDGE AI Blueprints as well as EDGE AI FOUNDATION event talks on a range of research, product and business topics. Join us to stay informed and inspired!

You Might Also Like