SAIL

Applied Research · XEROTECH LTD

Sovereign AI Lab (SAIL)

Regulated organisations cannot use AI the way Silicon Valley does. Their data cannot leave the building. Their compliance obligations do not pause for innovation cycles. The tools being built for them are too expensive, too opaque, or too dependent on infrastructure they do not control. SAIL exists to close that gap.

How it works

Research that ships

Every research track connects directly to a commercial deployment. The chain is deliberate.

SAIL researches it

XEROTECH builds it

PULVINIR and CallGPT 6X deploy it

The patent-pending privacy filtering in CallGPT 6X came out of SAIL’s privacy-preserving research track. The epistemic confidence framework in PULVINIR — our regulatory intelligence platform for energy, legal, tax, health, and education — came out of the compliance intelligence track. Research here does not sit in a drawer.

Engineering Capabilities

Retrieval-augmented generation over domain-specific knowledge engines

Deterministic rules engines for regulatory penalty modelling and compliance gap scoring

Vector databases for semantic retrieval across regulatory corpora

Multi-tier LLM architectures with audited per-call cost and token telemetry

Automated document generation with epistemic confidence tagging on every claim

Workflow orchestration across cloud and on-premise infrastructure

Client-side data redaction before any information reaches an AI provider

Full audit trail with request-level telemetry across every LLM call

Academic research infrastructure at Artificial Intelligence University. Commercial deployment at XEROTECH.

7
Published papers
9
Patents cited by Apple,
Microsoft & Amazon
5
Regulated sectors
in production

What we research

AI for regulated organisations

The dominant assumption in AI is that deployment requires hyperscale compute and frontier-model APIs. For regulated organisations, that assumption is wrong. SAIL asks what actually works when data cannot leave the building.

Compliance intelligence

Deterministic rules engines produce auditable numbers — penalty bands, downtime costs, maturity scores — without any LLM involvement. AI generates the narrative, grounded in cited sources, with every claim tagged for confidence level. A board, a compliance function, and a regulator can each read the same document and know exactly what is sourced and what is inferred. Built for energy, tax, legal, health, and education sectors.

Small-scale language models

How small can a model be and still do the job? Not as an academic question — as an engineering constraint for organisations that will never rent GPU clusters. Our published ILM model runs inference on hardware architecture from 1999. On a narrowly scoped task, the optimiser choice contributed more measurable improvement than switching from LSTM to Transformer. Useful AI does not require frontier compute. We can prove it with numbers.

Privacy-preserving architecture

AI systems where sensitive data never reaches an external provider. Client-side redaction before any query is transmitted. On-premise inference on hardware the organisation already owns. The patent-pending privacy filtering in CallGPT 6X came directly from this track, as did the deployment patterns used across PULVINIR engagements.

AI governance for regulated environments

What technical and organisational structures make AI safe to deploy when a regulator is watching. Ofgem for energy. FCA and PRA for financial services. CQC for healthcare. SRA for legal. Ofsted and OfS for education. ICO for data protection. NCSC CAF 4.0 for cyber resilience. The EU AI Act across Europe. We study the regulatory instruments, model the enforcement patterns, and publish analysis through the SAIL newsletter.

Published research

Papers, models, and data

Preprints on Zenodo and TechRxiv. Open models and datasets on HuggingFace. Longer-form analysis through AI Business Review.

“On this corpus at this scale, the optimiser contributed more to the measurable improvement than attention did.”

ILM & ArfaLM — SAIL, April 2026

Language models

ILM and ArfaLM

A 2.3M parameter LSTM trained using only techniques available in 1999 versus an 8.4M parameter Transformer with modern training — on identical child-directed speech data. When the performance gap is decomposed, the optimiser change alone accounts for 1.56× of the improvement. The architecture change, tokenisation improvement, and 3.6× parameter increase together account for 1.38×. On a narrowly scoped task at small scale, how you train matters more than what you train. Part 1 published. Part 2 — on-hardware deployment, integer quantisation, Pentium II benchmarks — in progress.

Paper (Zenodo)
·
Models (HuggingFace)
·
Dataset (HuggingFace)

Attention & interpretability

The Persistent Tension Hypothesis

Endogenous attention capture in large language models — how internal representational competition shapes what a model attends to, independent of prompt design. With Dr. Atif Naseer.

Paper (Zenodo)

Attention steering

Focus Mirror

An interactive framework for steering large language model attention in real time during inference — directing and redirecting what the model focuses on without retraining. With Dr. Atif Naseer.

Paper (TechRxiv)

AI governance

Lifecycle Governance for Multi-Agent Systems

Principles and a research agenda for governing multi-agent AI systems from design through deployment, monitoring, and retirement — covering accountability, auditability, and human-oversight requirements across regulated environments.

Paper (Zenodo)

Disaster management

Edge-Enabled Generative AI and AR/VR in Disaster Management

A theoretical framework for integrating edge AI with augmented and virtual reality to support decision-making in disaster response — relevant to critical infrastructure operators and civil contingency planners. With A.A. Khan.

Paper (TechRxiv)

Beyond the brief

The Temporal Fluid Multiverse

A unified framework for causality and observation, submitted to Royal Society Proceedings A. Outside SAIL’s applied remit — included because we publish across disciplines.

Paper (Zenodo) →

All models and datasets: huggingface.co/nshah-fbcs
·
ORCID
·
Google Scholar

Intellectual property

Patents and trademarks

Patents

2 granted US patents
Granted

Cited by Apple, Microsoft, Amazon, IBM, Morgan Stanley, Lenovo, Citrix, and Sturm Ruger & Co.

7 applied US patents
Pending

Full portfolio on Justia →

UK Trademarks

Orion — UK00004116898
Registered

eXtended Intelligence (XI) — UK00004350265
Registered

MTRX — UK00004090712
Registered

Where we publish

Our Channels

Papers and models

Preprints on Zenodo and TechRxiv. Trained models and datasets on HuggingFace. Code where it matters.

HuggingFace →

AI Business Review

Longer-form research and analysis for a business and policy audience. Independent editorial publication under AIU.ac.

Read AIBR →

SAIL Newsletter

Fortnightly on LinkedIn. Sovereign AI policy, deployment analysis, and original research for decision-makers in regulated industries.

Subscribe →

People

The team behind the lab

Noman Shah, FBCS

Lab Director · Founder & CEO, XEROTECH LTD · Founder & President, AIU.ac

Fellow of the British Computer Society — a distinction held by the top 1% of 60,000 members. 9 patents cited by Apple, Microsoft, Amazon, and IBM. 7 published research papers. Executive education at Harvard Business School and Oxford Saïd. HBR Advisory Council. Judge for the NASA Conrad Challenge. Previously designed and led a $200M World Bank-funded technology programme across three countries. Creator of CallGPT 6X and the AIfu methodology.

Google Scholar
·
ORCID
·
HuggingFace
·
nomanshah.com

Dr. Atif Naseer

Co-Founder & CTO, XEROTECH LTD

PhD Machine Learning. 60+ publications, 433+ citations on Google Scholar. Co-author on the Persistent Tension Hypothesis and Focus Mirror. Built crowd analytics systems processing millions of data points in real time. Research faculty at Umm Al-Qura University since 2012.

LinkedIn
·
Google Scholar

Industrial arm

XEROTECH LTD · Company #14474495 · ICO ZC065188

Academic arm

Artificial Intelligence University · UKPRN 10095512 · BCS-approved centre

Backed by

Google for Startups · Microsoft for Startups · Barclays Eagle Labs · AWS Activate

Commercial products

CallGPT 6X (privacy-first AI workspace) ·
VisionXI (AI accessibility) ·
PULVINIR (regulatory intelligence)

The gap is real. We are closing it.

If your organisation handles regulated data and needs AI that works inside your walls, talk to us. If you are a researcher working on small models, deployment efficiency, or sector-specific governance, we are open to collaboration.

XEROTECH LTD, Company #14474495 · Artificial Intelligence University, UKPRN 10095512