I
Impetora
Submit a project
Use case

Internal knowledge AI grounded in your own documents

Internal knowledge AI is the practice of using retrieval-augmented systems to answer employee questions, accelerate onboarding, and surface policy or compliance guidance from your own documents. Impetora ships these systems with citations on every answer, deflecting 92% of routine internal questions and saving 11 minutes per employee per day.

92%
Internal questions answered without handoff
11min
Saved per employee per day
3d
Onboarding time, down from 14 days
100%
Answers with source citations

01.What is internal knowledge AI?

Internal knowledge AI describes systems that answer employee questions by retrieving from your own corpus of policies, runbooks, contracts, training material, and historical decisions, then generating a grounded answer with the source clauses cited inline. The category covers employee Q&A assistants, onboarding accelerators, compliance look-ups, sales-enablement assistants for product and pricing, and policy search across HR, legal, and finance.

McKinsey's analysis of generative-AI economic potential places knowledge-management and employee Q&A among the highest-confidence value categories, contributing meaningfully to the USD 2.6 to 4.4 trillion annual opportunity the report describes. The reason these systems convert reliably is that the data is bounded, the success metric is measurable, and the user is your own employee, who can spot a wrong answer faster than an external customer can.

02.How does it traditionally work?

Without AI, internal knowledge lives in a fragmented stack: a corporate intranet with stale pages, a Confluence or Notion workspace with inconsistent ownership, an HR portal locked behind SSO, an internal Slack or Teams channel where institutional memory accumulates, and a queue of employees who DM the same five experts whenever a question gets hard. Average enterprise employees spend 1.8 to 2.4 hours per day searching for or recreating information they cannot find, depending on the organisation.

Onboarding amplifies the cost. IBM's onboarding analysis places average per-hire cost near USD 4,000 and the productivity-recovery window at 14 to 21 days in regulated industries. The traditional fix, training more internal experts and writing more wiki pages, has poor unit economics: the wiki goes stale faster than it can be written, and the experts get pulled into one-off questions that prevent them from doing the work they were hired for.

03.How does Impetora's TRACE methodology solve it?

Trust. All retrieval, inference, and conversation logs run in EU regions, with role-based access aligned to your existing SSO and ACL boundaries. An employee asking a finance question never sees retrieved chunks they would not have access to in the source system. EU AI Act Article 13 transparency obligations are satisfied: employees see capability and limitation disclosure built into the assistant surface.

Readiness. We sample real employee questions from existing channels (Slack threads, support tickets, helpdesk queues) before the model is selected, so the eval set reflects how employees actually phrase questions. Architecture. Versioned retrieval indexes per document domain, with refresh pipelines that update within minutes when source documents change. Citations and evidence. Every answer links back to the exact paragraph and document version it relied on. An employee can verify the answer, and a compliance team can prove which policy was in force at the time of the response.

04.What does the system architecture look like?

Four components in series. Ingest: connectors for your document repositories (SharePoint, Google Drive, Confluence, Notion, internal CMS, contract management) with ACL-aware indexing so retrieval respects the same access rules as the source system. Process: chunking, embedding, hybrid retrieval combining semantic and keyword search, with a re-ranker tuned to your evaluation set.

Review: the answer surface (Slack bot, Teams app, in-portal widget) shows the answer with cited source chunks expandable inline. Employees rate the answer with one click; the rating writes to the evaluation set. Deliver: a structured event lands in the audit log on every query, including the user identity, retrieved chunks, model version, and answer. The log is queryable by compliance and HR for incident-response or training-improvement work.

05.What measurable outcomes can you expect?

Four numbers we have validated against pilot baselines. 92% of routine internal questions answered without human handoff, in line with Stanford HAI's AI Index finding that grounded retrieval-augmented systems hit 90 to 95% answer accuracy when retrieval recall exceeds 85%. Time saved per employee runs 11 minutes per day on average, conservative against the published 30 to 60 minute figures because we count only verifiable, deflected interactions.

Onboarding time-to-productivity drops from a typical 14-day window to 3 days for the role-relevant policy, tooling, and process knowledge that the assistant covers. The remaining ramp time is the work-context familiarity that no assistant can shortcut. Audit coverage is 100%: every query, retrieval, and response lands in the log with full lineage, ready for compliance review.

06.How long does a deployment take?

A first pilot reaches production-grade behaviour on a single domain (HR, IT helpdesk, finance policy, or sales enablement) in 4 weeks. Phase one (weeks 1 to 2) is the readiness sprint: document inventory, ACL audit, employee-question sampling, scope sign-off. Phase two (weeks 3 to 4) is the build and shadow-mode rollout to a pilot group. Phase three (weeks 5 to 11) extends to additional domains and the full employee base, with each new domain requiring 1 to 2 weeks of evaluation work.

07.What does it cost?

Pilot engagements at this scope start at EUR 25,000 for a single domain and a defined employee population. Full production deployments across three to five domains with SSO, ACL-aware retrieval, and audit-log integration typically land between EUR 60,000 and EUR 150,000. Submit a project for a custom estimate, and we will quote against your document corpus, integration surface, and employee scale before any code is written.

Frequently asked questions

Does the system see documents employees should not have access to?+

No. The retrieval layer indexes documents with their source-system ACLs preserved as metadata, and every query filters retrieval to documents the asking employee can already access in the source system. If an HR record is restricted to the HR team in SharePoint, an engineer asking about it gets the same not-found response they would get if they navigated to the source. We do not train any model on your documents, and we do not aggregate retrieval results across access boundaries. The ACL audit during the readiness sprint is non-negotiable; we will not deploy without it.

How does it stay current when our documents change?+

Source documents are watched through their native APIs (SharePoint, Drive, Confluence, Notion), with an incremental refresh pipeline that re-indexes changed pages within minutes. The retrieval index versions documents, so the audit log can prove which version of a policy was in force when an employee asked a question. For high-stakes domains like compliance and finance, we hold a manual approval step on document updates before they reach the index, so a policy in draft does not surface to employees as guidance. The refresh cadence is configured per domain at scoping.

Can it integrate with Slack and Teams?+

Yes. Native integrations for Slack and Microsoft Teams, with conversation context carried across messages and the response surface respecting your existing permission and DLP policies. We also ship a web widget for in-portal embedding, an API for custom interfaces, and an SSO-protected web app for the cases where employees prefer a dedicated assistant surface. Audit logging is identical across surfaces, so compliance review does not depend on where the question was asked.

What about hallucinations?+

Three controls. First, retrieval is mandatory: if no source chunk passes the relevance threshold, the assistant returns a not-found response with a suggested human owner, never a guess. Second, the prompt enforces citation: the model is instructed to ground every claim in the retrieved chunks and to refuse to answer outside that scope. Third, employee feedback writes to the evaluation set; we run weekly drift reports comparing answer accuracy against a held-out gold set, and we re-tune retrieval or prompting whenever a domain drifts beyond the threshold agreed at scoping.

Is this just a chatbot, or something more?+

More. The system covers four interaction patterns: question-answer in chat surfaces, structured policy lookup with deterministic field extraction, onboarding flows that walk a new hire through role-relevant content with progress tracking, and audit-friendly compliance look-ups that return the cited clause and the version-as-of date. Different domains favour different patterns; we scope which patterns each domain needs during readiness. The shared infrastructure underneath is one retrieval and audit layer; the surfaces and prompts are tuned per use case.

How does it handle multiple languages?+

Native multilingual support across the major European languages, including Lithuanian, German, French, Spanish, and English. Each language has its own retrieval index per domain, an evaluation set in-language, and confidence thresholds calibrated against in-language test data. An employee can ask in Lithuanian against an English-language source corpus, and the assistant will retrieve, ground, and answer in Lithuanian with citations to the original-language source. Language tags carry into the audit log so locale-specific compliance review remains clean.

Submit a project for a custom estimate.