Everything you need to know about how we work, what we deploy, and what it means for your organisation's data.
What exactly does JD Fortress AI provide?
We deploy advanced large language models (LLMs) and custom retrieval-augmented generation (RAG) pipelines entirely on your premises — on your hardware, in your VPC, or in fully air-gapped environments.
No internet connection is required for operation. Your data never leaves your network or touches any external cloud provider. This gives you AI comparable to leading tools, but built exclusively around your own documents, policies, contracts, procedures, and knowledge — ask detailed questions and get precise, cited answers from your own sources.
For teams ready for the next step, we can layer on agentic capabilities: proactive, always-on AI assistants that monitor channels, triage issues, execute routine tasks, and handle workflows autonomously — all running locally on your infrastructure with the same zero-leak guarantees.
What are typical use cases for a law firm?
Here are examples we’ve seen with law firms running our secure, local RAG system:
Pull and summarise relevant case law, precedents, or internal notes instantly when preparing advice, pleadings, or opinions — cutting research time significantly.
Upload new regulations, client contracts, or updates and have the system highlight potential risks, inconsistencies, or next steps for compliance review.
Produce first-draft letters, NDAs, engagement letters, or other standard documents in your firm’s style, referencing past examples from your repository.
Handle due-diligence requests during transactions by quickly searching and cross-referencing your full document set.
Give junior lawyers and paralegals reliable, context-aware explanations of clauses, procedures, or points of law — without sending anything externally.
With agentic mode enabled: an always-on assistant monitors incoming queries, retrieves context from your knowledge base, drafts compliant responses or flags issues proactively (often overnight), and escalates only when human review is needed — freeing senior staff for higher-value work.
How much does it cost?
Pricing depends on your setup: model size, number of users, storage volume, whether it’s air-gapped, and any professional services for data ingestion, customisation, or agentic extensions.
Under conservative assumptions, many clients see payback within 6 months through time saved and reduced risk exposure. Agentic features can accelerate ROI further by automating ongoing workflows. Try our ROI calculator to model the numbers for your firm.
Contact us for a tailored discussion and quote — no obligation.
How frequently are the underlying language models updated, and who handles this?
We refresh the core models roughly every six months, timed to major capability jumps in the LLM field that justify the update effort. Updates are fully offline: we deliver new model weights via secure transfer (encrypted drives or similar), and your team — or ours during onboarding/support — applies them.
Since the entire system has no internet exposure, there are no ongoing security patches or vulnerability scans needed for the deployment itself — no external attack surface. This holds true even when running proactive agents.
How does my company’s data get integrated into the AI system?
Data integration is handled end-to-end by our team as part of deployment. We work with you to:
Organise and securely export your documents and knowledge sources
Convert formats as needed
Create vector embeddings and build the searchable index (the RAG foundation)
Tune the retrieval and generation pipeline to match your workflows and terminology
You don’t manage the technical pipeline — we set it up, test it with your data, and hand over a working system. Ongoing additions or tweaks are straightforward.
What if much of our data still exists only on paper?
We can handle that. Through optional professional services, we coordinate secure scanning and OCR to turn physical files into searchable digital text ready for ingestion. This is quoted separately based on volume and complexity — get in touch for details and a realistic timeline. Digitised content then feeds seamlessly into both standard RAG queries and any proactive agent behaviours.
Can email archives be included in the knowledge base?
Yes — archived email exports (PST, EML, or similar) are one of the richest sources of institutional knowledge we incorporate.
For compliance, we stick to non-live, historical exports only — no live mailbox connections. We can set up a completely local, agent-triggered refresh schedule (e.g., weekly ingestions) that keeps the knowledge base current without ongoing risk.
Which file formats are supported?
We fully support the most common business formats, including:
PDF
Microsoft Word (.doc, .docx)
Microsoft Excel (.xls, .xlsx)
Microsoft PowerPoint (.ppt, .pptx)
Plain text, Markdown, emails, and scanned images via OCR
The pipeline manages both structured and unstructured content effectively — whether for on-demand queries or feeding into proactive agent tools.
How frequently is the ingested data refreshed or updated?
Refresh cadence is entirely up to you and defined in your service agreement. Options include:
Manual/on-demand (e.g., after major document updates)
Scheduled automated syncs (daily, weekly, monthly) where feasible within your security rules
For agentic deployments, more frequent controlled refreshes can keep proactive behaviours highly relevant. We discuss the right frequency during setup.
Since the system is fully isolated, how do I actually get answers on my everyday work computer?
In true air-gapped setups (common for our highest-security clients), the system runs isolated by design. Practical access options we help implement:
Run the interface on a dedicated secure workstation or thin client connected via internal LAN (browser-based or approved app).
Use controlled removable media (encrypted USB drives) to transfer queries in and responses out — sneakernet style.
For agentic features: configure internal triggers (e.g., file drops, scheduled checks, or approved messaging channels on a segmented network) so the AI monitors and acts without needing constant manual input.
Many clients find this deliberate separation actually improves focus, auditability, and compliance — especially when agents handle routine monitoring autonomously.
If my data is already stored securely in Google Cloud, why is a local AI solution more secure?
Even when data resides in a highly secure cloud environment, sending that data to a third-party LLM provider creates a material risk: your confidential information temporarily leaves your perimeter and is processed by someone else’s infrastructure.
With JD Fortress:
Your data never leaves your controlled environment to reach an external LLM.
You keep using your existing secure cloud storage as the source of truth.
All AI inference happens locally — so answers, insights, and generated content remain entirely within your fortress.
This gives you the best of both worlds: the convenience of cloud-hosted data with the ironclad privacy of offline, on-premises AI.
Still have questions?
We're always happy to have a plain-speaking conversation. No pitch, no obligation.