Skip to main content

Command Palette

Search for a command to run...

Why I am Moving my AI "Agents" to the Edge (and Why You Should Too)

The "Offline" Advantage for First Responders

Published
3 min read
Why I am Moving my AI "Agents" to the Edge (and Why You Should Too)
K
I view technology through the lens of mission-critical resilience. As a Systems Engineer in public safety communications, I operate with the reality that UPTIME is a life-safety requirement (not just a KPI). This high-stakes mindset is the foundation of VertexOps. My engineering philosophy is a hybrid of professional infrastructure management and the "boots on the ground" pragmatism I’ve gained through CERT. I’m a firm believer in DIGITAL SOVEREIGNTY: knowing exactly where data lives, how it’s secured, and how it survives when things get chaotic. CURRENT TECHNICAL FOCUS: >LOCAL AI & SOVEREIGNTY: Deploying models like Gemma and Qwen on my Dell T3610 to prove that high-performance AI doesn't have to trade off privacy. >RESILIENT INFRASTRUCTURE: Managing the transition from legacy virtualization to hardened, self-hosted stacks using Nextcloud and Proxmox. >INFOSEC & RF: Hardening systems against modern threats while maintaining my roots in Amateur Radio (KO6JKE). There is a natural crossover between the "tinkerer" soul of a Ham operator and the discipline of a Systems Engineer. I’m here to document the builds, share the troubleshooting logs, and help other professionals bridge the gap between "it works" and "it’s resilient."

I have been watching the "Agentic AI" trend blow up on Hashnode lately. It seems like every other post is about how AI is moving from just answering questions to actually doing things—writing code, managing QA, and even handling incident triage. It is exciting stuff, but as someone who works in public safety, my first thought is always the same: What happens when the cloud goes dark?

In my line of work, we talk about "resilience" a lot. Whether it is my day job or volunteering with Sacramento CERT, you learn pretty quickly that if your tools depend on a perfect internet connection and a third-party server's uptime, you do not actually own those tools.

That is why I have been spending my nights in my home lab (shoutout to my trusty Dell T3610) moving away from the "cloud-first" mindset.

The Shift to the Edge

With the release of Gemma 4 and Qwen 3.5, the gap between "cloud AI" and "local AI" has basically evaporated for most practical tasks. I have been testing these models via Ollama, and the performance on consumer-grade hardware is getting insane.

Here is why this matters for those of us building infrastructure:

  1. Privacy is non-negotiable: If you are working with sensitive data—whether it is public safety info or just your own personal projects—sending that to a proprietary cloud model is a risk. Keeping it local means you keep the keys.

  2. True Resilience: If the grid goes sideways or the fiber gets cut, my local LLM keeps running. For an "Agent" to be useful in a real emergency, it has to be reachable.

  3. Latency: When you are running a local model on your own metal, you are not waiting on API calls or rate limits. It just works.

What is in my Stack?

I am currently leaning heavily on a self-hosted setup that looks something like this:

  • Hypervisor: VMware ESXi 8 (standard stuff, but rock solid).

  • Model Runner: Ollama, pulling the latest Qwen and Gemma weights.

  • Orchestration: Exploring how to use these local models for basic "agentic" tasks like automated log analysis and system hardening.

Why this matters

I have always liked platforms that focus on community and shared knowledge. The tech sector needs more of that "civic" mindset. We should be building systems that empower people, not just systems that make us dependent on a few giant corporations.

If you are just starting with local LLMs, my advice is to stop worrying about the benchmarks and just start building. Setup an old workstation, install Linux, and see what you can make it do without an internet connection. You might be surprised at how much power you actually have sitting under your desk.

I am curious—how many of you are actually running your "Agents" locally vs. relying on Claude or GPT-5? Let’s talk about it in the comments.

More from this blog

T

The Digital First Responder | Systems Engineering & Mission Critical IT

7 posts

The Digital First Responder is the technical authority for systems engineering in high-stakes environments. Managed by Kerry Kier, this blog documents the architecture, security, and deployment of mission-critical IT. We focus on bridging the gap between infrastructure resilience and emergency communications through local-first AI, self-hosted sovereignty, and high-availability logic. This is where professional systems engineering meets the front lines of public safety.