URL Monitor System

KeyValue
StatusActive
OwnerQA Automation
Updated2026-03-26
ScopeLink crawling, broken-link reporting, and URL-quality monitoring across CNC sites

The URL Monitor exists because broken links create a very different kind of failure from selector or journey regressions. The page can load. The UI can look fine. The user can still land on a dead page, a redirect loop, or a quietly expired campaign destination.

What The URL Monitor Does

CapabilityWhy It Matters
crawl pagesfinds the links the sites actually expose
extract links intelligentlykeeps the signal closer to real user navigation
check URLs asynchronouslymakes large checks practical
deduplicate resultsavoids alert floods
track source pageshelps people find where the bad link came from
keep historylets the team spot recurring URL issues

Crawl Depth Options

Crawl ModeBest For
basic crawlquick health check
nav-focused crawlmenu and structural link paths
full crawldeeper audits and larger sweeps

What Makes It Useful

The system is not just a raw link checker. It keeps enough context to help someone act on the result.

Useful ContextWhy It Helps
source trackingyou know where the bad URL came from
historyyou know whether this is new or repeating
deduplicationone broken endpoint does not become dozens of noisy alerts
scheduled operationissues surface without someone manually crawling the sites

Typical Problems It Catches

Problem TypeExample
broken navigation targetsmenu item or footer link goes nowhere
expired campaign pagesold marketing or editorial landing page now returns 404
backend 500slink exists but destination is broken
decommissioned subdomainshistorical links still live on the page

Slack And Reporting

The URL Monitor should feel like an operator tool, not a crawler dump.

Reporting GoalHow It Should Feel
clear severityis this one broken page or a broad issue?
source contextwhere was the link found?
trend awarenessis it recurring or brand new?
low noiseone incident, not fifty near-identical messages

Why It Belongs In This Repo

Links are part of product quality. For news sites, bad links affect:

  • reader trust
  • editorial campaigns
  • SEO
  • subscription journeys
  • archive content

That makes URL quality a real operational concern, not just a maintenance nicety.

[EXPAND: Good operator workflow]

When the URL monitor reports problems:

  1. check whether the failures cluster on one destination or one source surface
  2. confirm whether the breakage is editorial, backend, or infrastructure
  3. fix the source link if it is a content issue
  4. escalate the destination if the source is correct but the page is broken

[END EXPAND]

NeedGo To
general monitorsMonitoring
commandsCLI Reference