AI engineering for safety-critical systems.

AXONVERTEX AI is an AI research and AI engineering studio focused on AI deployments in private & low resource settings.

We finetune tiny foundation models, build private-compute AI inference stacks, and design responsible AI workflows that can be deployed under real-world constraints.

Focus
Healthcare · Clinical Trials · Safety-critical AI
Stack
Private Compute · Ollama · vLLM · Tiny FMs
AI Engineering
Production pipelines · Evaluation · Tooling
Ethics
NIST-aligned ARIA & Responsible AI

Agents of S.E.A.L.E.D

An agentic framework for safer, auditable AI assistance in complex workflows. Built with strong guardrails and real-world constraints in mind.

See flagship initiatives →
Open research Clinical data Graph + AI Inference at the edge

What we do

AXONVERTEX AI sits at the intersection of independent research and deployed systems: designing, stress-testing, and shipping AI that can withstand real-world constraints.

Independent AI research portfolio

Self-directed research across agents, FHIR graphs, vector embeddings, and evaluation, captured in talks (Agents of S.E.A.L.E.D, FHIR in the W.H.O.L.E), NIST challenges, and open artifacts on Hugging Face.

Responsible AI systems

Design, evaluation, and governance aligned with NIST-oriented ARIA practices – from risk assessment and documentation to human-in-the-loop control.

Healthcare & clinical data

FHIR, graphs, and longitudinal patient context powering decision support and trial automation with stringent privacy guarantees.

Private compute ecosystems

Secure, low-latency inference on Ollama and vLLM, tuned for constrained hardware and sensitive environments where data cannot leave the boundary.

Flagship initiatives

A sample of the projects where AXONVERTEX AI combines responsible AI research with production-grade engineering.

Agents of S.E.A.L.E.D

Safety-Engineered Agentic Learning for Evidence-Driven Decisions

An agentic system architecture for safety-critical domains. Agents of S.E.A.L.E.D combines explicit policy constraints, memory, and oversight to keep complex AI workflows auditable and aligned with human operators.

  • Deterministic safety rails and traceable decision paths.
  • Configurable oversight levels for clinical and regulatory settings.
  • Composable tooling for integrating with trial platforms and data stores.
Watch Agents of S.E.A.L.E.D talk

FHIR in the W.H.O.L.E

Graphing clinical data for smarter AI systems

FHIR in the W.H.O.L.E explores how FHIR resources and knowledge graphs combine to create rich, queryable clinical contexts that modern AI systems can reason over safely.

  • Graph-native representation of FHIR resources.
  • Context-aware retrieval for RAG and agentic workflows.
  • Improved explainability through structured relationships.
View FHIR in the W.H.O.L.E session

TrialBridge

Built with HI10x Innovation & Transformation GmbH

TrialBridge focuses on connecting sponsors, sites, and patients through responsible AI that respects regulatory and ethical boundaries while increasing trial velocity.

  • Patient and site matching powered by explainable models.
  • Protocol-aware assistants with hard safety and privacy constraints.
  • Configurable deployment for private or hybrid compute environments.

Co-designed with HI10x to fit real clinical operations instead of abstract benchmarks.

Visit TrialBridge

Agent Guardrails Framework

Prompt injection + tool misuse · Local Linux workspace

An interactive demo deck showing how to turn agent autonomy into an enforceable, auditable system boundary: deterministic policy, scanner pipeline, and an observability ledger.

  • Least-privilege tool gateway with allow/deny rules.
  • Injection + patch-safety scanners with explainable decisions.
  • Audit-ready event timeline (events.jsonl / alerts.jsonl).
Open the guardrails demo deck →

Responsible AI and ARIA with NIST

AXONVERTEX AI’s work is grounded in NIST-aligned approaches to AI risk, assurance, and governance. We design systems where accountability is a feature, not an afterthought.

Assessment of risk in AI

Contributions to NIST’s 2024 ARIA challenge, focusing on practical methods to identify and measure risk in deployed AI systems.

Generative AI challenges (2024–2026)

Ongoing participation in NIST Generative AI challenges, helping shape evaluation approaches that connect model behaviour to system-level outcomes in safety-critical domains.

Governed deployment

Human-in-the-loop controls, clear escalation paths, and operational playbooks keep responsible AI principles active throughout the lifecycle, not just at launch, with ARIA-style evidence at every stage.

Private compute ecosystem inference

Sensitive workloads demand private, controllable infrastructure. AXONVERTEX AI builds inference stacks that keep data close to where it is generated, while still delivering state-of-the-art performance.

Ollama-first workflows

We leverage Ollama to run compact models on laptops, edge servers, and controlled clinical environments – enabling offline or near-offline operation with reproducible model stacks.

  • Simple packaging for tiny and domain-specific models.
  • On-device inference for low-connectivity and high-privacy scenarios.
  • Consistent developer ergonomics across environments.

vLLM for high-throughput serving

For larger models and multi-tenant workloads, we rely on vLLM to deliver efficient high-throughput serving with attention to latency, cost, and resource isolation.

  • Batching and caching tuned for real-world conversational patterns.
  • Per-tenant controls for quotas and safety configurations.
  • Integration with structured logging for compliance and audits.

End-to-end private ecosystems

From data ingestion to monitoring, we assemble ecosystems that treat privacy as a non-negotiable design constraint, not an optional add-on.

  • Preference for on-prem or VPC deployment where required.
  • Clear data residency and retention guarantees.
  • Guardrail layers that work across model families and runtimes.

Open research areas

We aim to publish methods, tools, and finetuning tiny foundational models openly, with a focus on reproducibility and practical impact.

Tiny foundational models

Architectures optimized for constrained environments that still deliver high task performance, with transparent training recipes and evaluation results.

RAG & RAFT

Retrieval-Augmented Generation and Retrieval-Augmented Fine-Tuning pipelines that combine structured and unstructured data to produce grounded, auditable outputs.

Vector embeddings

Representation learning research focused on embeddings that capture clinical and operational context while remaining efficient for search, clustering, and routing.

Secure AI pipelines

Hardening data and model pipelines with encryption, access control, and rigorous monitoring, suited for regulatory environments.

Bias and fairness

Techniques to identify, measure, and mitigate bias across datasets, models, and full systems, with emphasis on high-stakes use cases.

Common frameworks & standards

Alignment with emerging common frameworks for AI safety and governance, making systems easier to audit and integrate.

Contact

For collaborations, pilots, or research partnerships, reach out directly. We prioritise work where responsible AI and rigorous engineering materially improve outcomes.

Email

For general enquiries and partnership discussions, email [email protected].

LinkedIn

Follow updates and announcements on LinkedIn .

Research artifacts

Models, prompts, and experimental artifacts will be published on our Hugging Face organization .