What powers factually's AI?

Are you looking for more information regarding Factually? Check out our FAQ!

Still have questions? Reach out!

Checked on December 15, 2025
Disclaimer: Factually can make mistakes. Please verify important information or breaking news. Learn more.

This fact-check may be outdated. Consider refreshing it to get the most current information.

Executive summary

Factually’s AI is not described in the provided search results; available sources do not mention what powers Factually’s AI. Reporting about the current AI landscape emphasizes that leading models are Gemini (by Google/DeepMind) and OpenAI’s GPT family, with industry debates over factuality highlighted by Google DeepMind’s FACTS Benchmark that found top models score roughly 69% on factual accuracy [1] [2]. Major vendors are pushing “more factual” model variants (e.g., Gemini 3 Pro) and agent architectures that combine models with retrieval and task orchestration [3] [4].

1. Missing direct evidence: no sourced description of Factually’s stack

None of the supplied articles identify the technical foundations, vendors, or chips behind a product called “Factually.” The search results focus on broader industry developments — benchmarks, Gemini, GPT releases and enterprise adoption — but do not mention Factually by name. Therefore any specific claim about what “powers Factually’s AI” would be unsupported by the current reporting (not found in current reporting).

2. Benchmarks and factuality: the industry bar is imperfect

Google DeepMind’s new FACTS Benchmark Suite is now in circulation and, according to Business Insider reporting derived from DeepMind, shows the best evaluated models achieve about 69% factual accuracy on its tests — a sobering marker for products that sell themselves on correctness [1] [2]. Journalists and enterprises are treating that 69% figure as a wake-up call: fluency is strong, but factual reliability remains a weak point when AI is used for niche or high-stakes tasks [1].

3. Who the market leaders are — and what they claim

Google’s Gemini series, especially commercial “Pro” variants, is explicitly marketed as being trained to minimize hallucinations and positioned as its “most factual” model; Google is also packaging agentic research products that use Gemini 3 Pro as a core component for deeper, multi-step information tasks [3]. OpenAI continues to push its GPT lineage (including rapid releases such as GPT‑5.2 in December 2025) and promotes enterprise integrations that customers report save time, though those claims come alongside internal competitive pressures [3] [5].

4. Agents and retrieval: the architecture most likely behind fact-focused products

Multiple outlets describe a shift from standalone chat models to agentic systems that combine a base model with retrieval, grounding, and orchestration layers. Google’s Deep Research agent and its Interactions API exemplify this trend: models are being embedded in agents that can call external tools, search, and structure multi-step queries to improve factual grounding [3] [4]. If a product called Factually exists as a fact-focused service, the prevailing industry pattern suggests it would likely use a retrieval-augmented architecture tied to a commercially available model — but that inference is not confirmed in the supplied sources (not found in current reporting).

5. Incentives and hidden agendas shaping claims of “factual” AI

Vendors actively brand certain model variants as “most factual” to win enterprise trust; Google touts Gemini 3 Pro’s factual improvements while OpenAI highlights enterprise time savings and client wins — both have clear commercial incentives to over-index on reliability claims [3] [5]. Benchmarks like FACTS are valuable but can be gamed by tuning, cherry-picking evaluation sets, or focusing on narrow improvements that don’t generalize to all domains [2] [1].

6. Remaining limitations and what to watch next

Reporting emphasizes systemic constraints beyond model architecture: compute, chips and infrastructure bottlenecks, and the limits of current evaluation methods. Analysts warn that top models still hallucinate on niche knowledge and complex reasoning, and that infrastructure (chips, energy, grid connections) constrains how broadly firms can scale more factual, larger systems [1] [6]. Watch for independent audits, cross‑vendor benchmarks, and whether vendors open-source evaluation methodology for FACTS-style benchmarks [2] [1].

Summary takeaway: available sources do not say what powers Factually’s AI. The best-supported industry facts in the supplied reporting are that Google’s Gemini (including Gemini 3 Pro) and OpenAI’s GPT family dominate headlines, that DeepMind’s FACTS benchmark puts top-model factual accuracy near 69%, and that agentic, retrieval-augmented architectures are the dominant approach to improving factuality — but none of these sources identify a product named Factually or its underlying technology [1] [3] [2] [4].

Want to dive deeper?
Who owns Factually and what investments fund its AI?
What large language model or proprietary architecture does Factually use?
How does Factually train and validate its AI for accuracy and bias?
What data sources and ingestion pipelines power Factually's AI knowledge?
What privacy, security, and transparency practices govern Factually's AI?