Back to Community
GuideAdvanced

Deploying Edge Infrastructure

Edge computing puts your infrastructure where your users are. This guide covers how to plan, deploy, and manage production compute at the edge, from selecting sites to serving traffic through a distributed mesh.

What is Edge Deployment?

Edge deployment means placing compute resources close to where data is generated or consumed, rather than in a centralized data center hundreds or thousands of miles away. This reduces latency, improves reliability, and keeps data local.

An "edge pod" is a small cluster of machines deployed at a specific location. It could be a server rack in an office, a mini PC in a retail store, a ruggedized box at a cell tower, or a few machines in a local colocation facility. The machines join your mesh network and operate as part of your larger infrastructure.

Edge is not about scale. It is about proximity. A single machine serving a local application at 2ms latency is more valuable than a 100-server cluster 1,000 miles away serving the same application at 40ms.

Planning Your Edge Pod

Before deploying, answer these questions:

What workloads will run here?

AI inference, web application serving, database replicas, IoT data collection, content caching. Each has different hardware requirements. AI inference needs GPU. Web serving needs fast networking. Databases need reliable storage and enough RAM.

What is the physical environment?

Temperature, power availability, physical security, internet connectivity. A climate-controlled server room is ideal. A closet with a good ethernet drop works. A garage in summer does not (heat kills hardware). Minimum: stable power, internet, and a temperature under 85F/30C.

What is the redundancy requirement?

For development and internal tools, a single machine is fine. For production serving, you want at least two machines so one can go down for maintenance without interruption. For critical workloads, three machines with automated failover.

Networking at the Edge

Edge networking has two layers: the local network (how machines at the site talk to each other) and the overlay network (how the site connects to the rest of your infrastructure).

Local networking

Machines at the same site should be on the same physical network. A managed switch with VLANs is ideal but a basic unmanaged switch works. The machines will discover each other on the LAN automatically during bootstrap.

Mesh overlay

WireGuard tunnels connect the edge pod to your control plane and other sites. Traffic is encrypted end-to-end. The overlay handles NAT traversal, so edge pods behind residential or office firewalls connect without port forwarding.

Public ingress

To serve public traffic from an edge pod, use Cloudflare Tunnels. A tunnel creates an outbound connection from your machine to Cloudflare's network, making your services reachable without exposing your IP or opening inbound ports. ProxAPI can provision tunnels, DNS records, and SSL automatically.

Security at the Edge

Edge deployments face unique security considerations because the hardware is physically distributed and may be in less controlled environments.

  • Encrypt everything in transit. WireGuard handles inter-node encryption. All API communication uses mutual TLS. There is no unencrypted traffic between nodes.
  • Encrypt data at rest. Use LUKS full-disk encryption on edge nodes. If a machine is physically stolen, the data is unreadable without the encryption key.
  • Minimal attack surface. Edge nodes should run only the workloads they need. No unnecessary services, no open ports, no SSH from the public internet. All management happens through the mesh overlay.
  • Automated patching. AI agents can handle security updates across distributed edge pods, applying patches during maintenance windows and rolling back if something breaks.

Use Cases

Local AI inference for retail or healthcare

A GPU-equipped mini PC at each location runs AI models for document processing, customer interactions, or diagnostic assistance. Patient data and customer data never leaves the premises. Latency is sub-10ms.

Multi-site web serving

Deploy your application to edge pods in multiple cities. Cloudflare routes users to the nearest pod automatically. You get CDN-like latency without paying CDN prices, and your application runs on hardware you own.

IoT and sensor data

Process sensor data locally instead of streaming it to a central cloud. An edge pod at a factory, farm, or building processes data in real-time and only sends aggregated results upstream. Reduces bandwidth costs and keeps raw data local.

Deployment Checklist

[ ]Hardware selected and tested (boot, memory check, disk check)
[ ]Physical site prepared (power, cooling, network, physical security)
[ ]Proxmox VE installed and accessible via web UI
[ ]Node joined to mesh overlay (WireGuard / Tailscale)
[ ]Full-disk encryption enabled (LUKS)
[ ]Cloudflare Tunnel configured for public-facing services
[ ]Monitoring and alerting connected to central dashboard
[ ]Backup strategy configured (local snapshots + off-site replication)
[ ]Workloads deployed and health-checked

Related Guides