← Back to Blog

September 30, 2025

Technical Overview

A deep dive into the architecture, ML models, and infrastructure that powers LinkRivers.

LinkRivers is built from the ground up to be a web operations platform that scales from a single portfolio site to thousands of enterprise domains. This document outlines the key architectural decisions, infrastructure choices, and ML systems that make that possible.

Architecture at a Glance

Our architecture follows a microservices pattern deployed across multiple availability zones. The core principles:

Design philosophy: We optimize for reliability first, then speed, then cost. A monitoring platform that goes down is worse than no monitoring at all.

Infrastructure Stack

Compute

Railway + Containers

Database

PostgreSQL

Cache

Redis

Queue

BullMQ

ML Runtime

Python + GPU inference

Frontend

React + Vite

Why These Choices

PostgreSQL handles our transactional workloads and time-series data. We use table partitioning for metrics data, keeping recent data hot while archiving older data to cold storage. This gives us fast queries on recent data while maintaining a full audit trail.

Redis serves multiple purposes: session storage, rate limiting, real-time feature flags, and as a caching layer for expensive database queries. Our cache invalidation strategy is event-driven - when underlying data changes, we publish an event that triggers cache eviction.

BullMQ powers our job queue system. Every monitoring check, notification dispatch, and ML inference job flows through the queue. This decouples our API layer from compute-intensive work and provides natural backpressure during traffic spikes.

Monitoring Pipeline

When you add a site to LinkRivers, here's what happens:

Check Distribution

We run checks from 30+ locations worldwide. Each check is assigned to 3-5 locations simultaneously to eliminate false positives from regional network issues. A site is only marked "down" when multiple locations confirm the failure.

Minimizing false positives: We use multi-location verification, retry logic, and ML-based confirmation to keep false positive rates low.

ML Models

LinkRivers uses machine learning for three primary functions:

1. Anomaly Detection

We train proprietary models on each site's historical data. The model learns normal patterns for response time, availability, and traffic. When metrics deviate significantly, we flag it as an anomaly - often catching issues before they become full outages.

2. Predictive Forecasting

After 2-4 weeks of training on your data, time-series forecasting models predict metrics 24-72 hours ahead. This powers features like "SSL expires in 14 days" and "Traffic spike expected Thursday based on historical patterns."

3. Root Cause Analysis

When an incident occurs, our classification model analyzes the symptoms and suggests probable causes. This accelerates diagnosis from minutes to seconds for common failure modes.

Model Training Pipeline

Agent Lite

Agent Lite is our client-side JavaScript that powers Real User Monitoring. Design goals:

The agent captures Core Web Vitals (LCP, FID, CLS), navigation timing, resource loading, JavaScript errors, and user interactions. All data is batched and sent in compressed payloads to minimize network overhead.

Autopilot System

Autopilot is our automated remediation engine. When an issue is detected:

Human oversight: Every Autopilot action is logged and reversible. Users can configure supervised mode where fixes are proposed but require approval before execution.

Security Model

Security is foundational, not an afterthought:

API Security

API keys are hashed before storage. Rate limiting is enforced at multiple layers: per-key, per-IP, and per-endpoint. All API calls are logged for audit purposes with automatic anomaly detection on access patterns.

Scalability

The architecture is designed to scale horizontally from day one. Key design principles:

This architecture allows us to grow with our customers without requiring fundamental changes.

What's Next

We're actively working on:

Questions about our architecture? Get in touch - we love talking about this stuff.