skip to content

The Agentic Edge: A New Operating System for the Physical World

Interactive Agentic Salon Recap

EdgeAI Salon May 2025

At the end of May, we hosted an EdgeAI Salon in New York to explore the next frontier of embedded intelligence. The event featured two leading-edge companies—Kibeam and Peridio—whose work exemplifies the transition from traditional devices to context-aware systems and is creating the new standard for EdgeAI infrastructure. This post is both a recap and a technical extension of the ideas discussed during our gathering.

The Shift from Screens to Sensors

For decades, we've been promised an intelligent, responsive, and proactive world through the Internet of Things—one where devices collaborate, adapt to our context, and solve problems with minimal human oversight. Yet despite huge strides in connectivity and hardware, this vision has largely remained unfulfilled. What we've been missing isn't better chips or faster modems—it's a new edge paradigm. To unlock the next generation of systems, we have to move beyond rigid, deterministic software and embrace a new approach: decentralized, context-aware, goal-seeking edge agents. This isn’t just an optimization strategy, it’s a fundamental shift in how we build systems for a world too complex, asynchronous, and dynamic for linear logic to scale.

A New Frontier for AI: The Edge

If artificial intelligence is to fulfill its promise to humanity, it must move beyond the computer screen. Today, the majority of AI models and experiences remain constrained to cloud platforms, smartphones, or desktop environments. But the most transformative applications of AI won’t happen behind glass—they’ll occur in the real world: in warehouses, classrooms, hospitals, factories, and homes.

Enter Edge AI. By embedding intelligence directly into devices at the periphery of the network, Edge AI enables localized decision-making, ultra-low latency response times, and resilience in environments where connectivity is unreliable or nonexistent.

According to IoT Analytics, the number of connected devices is expected to grow from 18.8 billion in 2024 to 40.1 billion by 2030. This explosion in hardware presents a staggering opportunity—but also a formidable challenge. Without rethinking how these devices process, act on, and share data, we risk being buried under a sea of sensors with no intelligence. The missing ingredient isn’t better hardware. It’s a shift in how we design systems:

From Deterministic Control to Adaptive Behavior: Traditional irrigation systems water plants on a timer regardless of weather. An adaptive edge agent integrates soil sensors and weather forecasts to dynamically schedule watering.

From Command-driven Interfaces to Context-driven Autonomy: A smart speaker that only responds to explicit commands is command-driven. A context-aware system dims lights, adjusts HVAC, and silences alerts automatically based on occupancy, time of day, and behavior patterns.

From Centralized Cloud Intelligence to Decentralized, Embedded Decision-making: A warehouse robot that depends on the cloud for navigation suffers latency or fails during connectivity issues. A decentralized version plans and re-routes locally while syncing episodically with cloud coordination.

EdgeAI represents the evolution of embedded computing into something more fluid, more capable, and more human-aligned.

Rethinking Programming at the Edge

The legacy software paradigm, in which devices execute predefined rules in siloed environments, is rapidly breaking down. Real-world edge scenarios demand adaptable, autonomous, and reactive agents. These systems must learn from context, prioritize competing goals, and operate in uncertain or partially observable environments.

EdgeAI is not simply about running models on small chips. It’s about enabling embedded systems to reason, to coordinate, and to evolve over time. It requires combining machine learning with modern distributed systems, new tools for managing heterogeneity, and dynamic orchestration across compute, network, and storage. Designing edge-native, autonomous agents involves:

Model Compression Techniques

  • Quantization (e.g., INT8, FP16)

  • Pruning for sparsity

  • On-device distillation

On-Device Learning and Adaptation

  • Online reinforcement learning

  • Lightweight retraining from sensor data

  • Federated learning across device fleets

Systems Architecture

  • Asynchronous messaging buses

  • Local policy interpreters (e.g., Wasm, Lua, or eBPF)

  • Hierarchical memory: RAM, flash, cloud-synced logs

Security and Verification

  • Hardware attestation

  • Runtime constraint checking

  • Sandboxed execution for third-party logic

Salon Spotlight: Two Companies Leading the Way

Kibeam: Intelligence Meets Imagination

Kibeam is pioneering EdgeAI in early childhood education. Their screen-free, AI-powered Kibeam Wand brings traditional books to life by embedding intelligence directly into the reading experience. By integrating real-time language modeling and interactive voice prompts, the device encourages children to read aloud, respond to questions, and engage with content actively— all without needing a screen. It represents a shift from passive consumption to embodied, conversational learning.

Under the hood, the Kibeam Wand is an EdgeAI device. It processes audio input locally, responds in real-time, and logs anonymized data for analysis by parents and educators. In doing so, it unlocks entirely new educational modalities that are both scalable and human-centered.

Peridio: A Control Plane for the Agentic Edge

Peridio is quietly building the orchestration layer for the future of embedded intelligence. Their open-source OS and device management platform streamline firmware deployment, observability, and secure communications for fleets of Linux-based edge devices. Rather than managing one device at a time, Peridio enables developers to manage thousands of devices simultaneously as a coherent system.

What sets Peridio apart is its focus on developer productivity and day 1 production readiness, enabling atomic updates, secure OTA pipelines, and visibility into distributed systems that operate in intermittently environments. This makes it possible to deploy, monitor, and continuously improve AI agents running on the edge—even in places where traditional DevOps tooling falls short.

EdgeAI: Why Context is the Next Frontier

The convergence of EdgeAI, decentralized systems, and autonomous software agents is quickly emerging as one of the defining technological shifts of this decade. With over 40 billion connected devices expected to come online by 2030, the sheer density of compute at the edge demands more than simple telemetry or rule-based automation. Intelligence must be embedded directly into the physical world. The most promising future for AI isn’t limited to multimodal models operating in the cloud—it’s distributed, embodied, and ambient. The edge is where autonomy meets context. It's where AI stops reacting to typed prompts and begins acting on real-time environmental signals. It’s where decisions happen in milliseconds, where connectivity can’t be assumed, and where adaptability matters more than precision. Systems that can reason locally, update incrementally, and coordinate across decentralized networks will outperform static logic in real-world environments. The companies and platforms that embrace this shift won’t just gain speed; they have the potential to build long-term defensibility by creating adaptive ecosystems that learn and evolve at the point of interaction.

The result isn’t just faster computation—it offers new behavior patterns. Edge intelligence allows systems to operate independently, recover from failure gracefully, and participate in network effects without always relying on the cloud. It’s not about scaling down—it’s about scaling out into billions of distinct, local, continuously changing contexts. Edge AI isn't about shrinking models for microcontrollers. It's about scaling intelligence to billions of contexts simultaneously. It demands:

  • New abstractions that reflect local autonomy and partial observability.

  • Flexible deployment models that prioritize latency, privacy, and bandwidth efficiency.

  • Agent coordination frameworks that support distributed learning and negotiation.

  • Observability and control tools built for asynchronous, edge-first systems.

  • Security and attestation primitives that function across trustless environments.

The next wave of computing won’t be defined by what happens on your screen—it’ll be defined by what happens all around you. From intelligent toys to self-healing factories, the world is waking up and the infrastructure to support it is just beginning to take shape.

You are leaving Galaxy.com

You are leaving the Galaxy website and being directed to an external third-party website that we think might be of interest to you. Third-party websites are not under the control of Galaxy, and Galaxy is not responsible for the accuracy or completeness of the contents or the proper operation of any linked site. Please note the security and privacy policies on third-party websites differ from Galaxy policies, please read third-party privacy and security policies closely. If you do not wish to continue to the third-party site, click “Cancel”. The inclusion of any linked website does not imply Galaxy’s endorsement or adoption of the statements therein and is only provided for your convenience.