← Home
About · Astrognosy AI

Built on 15 Years
of Foundational Research

Positional Correlation Fields didn't emerge from a weekend hackathon. It's the result of a decade and a half of work on how machines understand language — and what happens when you apply that lens to everything else.

Where It Started

In 2008, Jake Westerbeck partnered with Jeff Spetalnick to build production NLP systems — practical tools for how machines handle language input, error correction, and statistical prediction from corpora. The research centered on a deceptively simple question: how does the structure of language — the patterns of which words follow which — encode meaning in a way that doesn't require brute-force compute?

That work produced two patents focused on frequency-based language efficiency: reducing keyboard input errors, improving OCR accuracy, enabling compression and translation through corpus-learned statistical likelihoods. These weren't academic exercises — they were built to run efficiently on the hardware of the time, without relying on large neural models that didn't yet exist at scale.

"The insight wasn't that language is complex. It's that the positional relationships between tokens — across a window of context — encode far more structure than anyone was using efficiently."

Jeff passed away before seeing where those ideas would lead. The research sat. But the fundamental insight — that positional co-occurrence patterns carry structural meaning that can be scored and compared without training — didn't go away. It waited for the right problem.

In Memory — Jeff Spetalnick

Jeff Spetalnick was a researcher and inventor whose work on statistical language modeling laid the groundwork for everything that followed. His approach — practical, efficiency-first, grounded in how real systems behave — is embedded in the DNA of the Pacific platform.

The foundational patents co-developed during our collaboration remain a testament to ideas that were ahead of their time:


The Evolution to PCF

The leap from frequency-based NLP to Positional Correlation Fields was not a pivot — it was a completion. The 2010 patents worked because positional patterns in token sequences are structurally meaningful. PCF formalizes that intuition into a general-purpose scoring mechanism: analyze co-occurrence patterns across positional offsets, normalize by calibrated thresholds, and you have a structural signature — a vector that encodes what a sequence looks like, independent of domain.

When applied to network traffic flows in 2025, the same math that once predicted the next word in a sentence detected cyberattacks without a signature database. When applied to LLM query verification, it routed queries to CPU with 25% GPU savings. The engine was always domain-agnostic. It just took years of problems — and 15 years of perspective — to see it clearly.

Astrognosy AI was founded to build the products that prove this out: not as research demos, but as real infrastructure that enterprises can deploy today.


The Founder

JW

Jake Westerbeck

Founder & Inventor · Astrognosy AI / Pacific Platform
Jake has spent 15+ years working at the intersection of language, statistics, and production systems. Starting with NLP efficiency research in 2008, he developed the frequency-based methods that became the 2010 and 2015 patents, then spent years working toward a more general formulation — one that could handle any sequential token stream, not just text. The result is Positional Correlation Fields, now powering the Pacific platform across LLM routing, cybersecurity, data center intelligence, chip validation, and robotic sensing.

Intellectual Property Timeline

From foundational frequency models to domain-agnostic positional fields — a 15-year arc of research, now protected across five product verticals.

2008–2010
Foundational NLP Research
Collaboration with Jeff Spetalnick on production NLP systems — statistical language modeling, corpus-based error reduction, and efficient frequency-driven text analysis.
Prior art establishing statistical token co-occurrence as structural signal
2010
US20100131900A1 — Frequency-Based Language Analysis
Methods and systems for improved data input, compression, recognition, correction, and translation through frequency-based language analysis. Predictive text input, OCR enhancement, and corpus-driven statistical likelihoods.
View on Google Patents ↗
2015
US9122318B2 — Reducing Keyboard Data Entry Errors
Dynamic visual distinction of likely next keys using corpus-learned statistical likelihoods — error minimization at the input layer through positional probability scoring.
View on Google Patents ↗
February 2026
Patent 1 — PCF Core Framework (Non-Provisional)
Positional Correlation Fields for domain-agnostic verification. CPU-native inference, structural signature computation, dual-signal anomaly detection, deterministic caching, and hybrid routing. 42 claims.
U.S. Prov. No. 63/978,633 · Filed Feb 9, 2026
March 2026
Patent 2 — Domain-Agnostic PCF + VASE Agent
Extends PCF to sequential token streams across NLP, cybersecurity, hardware telemetry, robotics, and industrial IoT. Adds autonomous RL corpus agent (VASE) and edge deployment. 66 claims.
U.S. Provisional · Filed March 2026
2026 — In Preparation
Patent 3 — Wharf Protocol
Multi-agent competitive verification marketplace using PCF as objective structural arbitration. Agents offload query cargo for selection — winner determined by PCF convergence, not authority.
In preparation

Where This Goes

The Pacific platform exists to prove a single thesis: that a domain-agnostic structural scoring engine — one that runs on CPU, requires no training data, and improves automatically as it processes more queries — is the right foundation for AI infrastructure at scale.

Every product on the platform is a domain-specific proof of that thesis. Compass proves it in LLM routing. Laminar proves it in network security. Stratum, Forge, and Beacon are proving it in physical infrastructure. Wharf will prove it in multi-agent coordination — the port where agents offload cargo for PCF selection.

CPU-Native

Every product runs without a GPU. Lower cost, lower power, edge-deployable — from day one.

Zero Training

No labeled data required to start. The engine calibrates on normal behavior and detects divergence.

Corpus Flywheel

More usage means a better corpus, means better detection — a structural moat that compounds.

Domain-Agnostic

One engine. Language, networks, silicon, sensors, agents — the math doesn't change between domains.

Let's Build Together

Pilots, partnerships, and investor conversations welcome. Based in San Antonio, Texas — open to remote and on-site collaboration.