Elicit (AI Research Tool)
Elicit (AI Research Tool)
Elicit is an AI research assistant with 2M+ users that searches 138M papers and automates literature reviews, founded by AI alignment researchers from Ought and funded by Coefficient Giving (formerly Open Philanthropy) (\$31M total). The platform achieved 90%+ extraction accuracy and claims 80% time savings for systematic reviews, though its alignment theory (that decomposed reasoning is safer than end-to-end training) remains empirically unvalidated at scale.
Quick Assessment
| Dimension | Assessment |
|---|---|
| Type | AI research tool / public benefit corporation |
| Founded | 2020 (spun out from Ought in 2023) |
| Headquarters | Oakland, California, USA |
| Users | 2 million+ researchers in academia and industry |
| Database Coverage | 125-138 million academic papers |
| Funding | $31 million total ($9M seed 2023, $22M Series A 2025) |
| Key Innovation | Sentence-level citations and decomposed reasoning for literature review automation |
| AI Safety Connection | Founded by AI alignment researchers; funded by Coefficient Giving |
Key Links
| Source | Link |
|---|---|
| Official Website | elicit.com |
| Wikipedia | en.wikipedia.org |
Overview
Elicit is an AI-powered research assistant designed to automate scientific research tasks including literature search, summarization, data extraction, and synthesis across over 125 million academic papers.12 The platform distinguishes itself through its emphasis on scale (analyzing up to 1,000 papers and 20,000 data points), accuracy validation for scientific use, and transparency through sentence-level citations rather than simple chat interfaces.23 With over 2 million users including researchers from top academic institutions and enterprises, Elicit has grown organically through word-of-mouth while generating millions in annual revenue from tens of thousands of paying subscribers.4
Elicit emerged from Ought, a non-profit AI alignment research organization founded in 2017 to explore how machine learning could scale up good reasoning.56 The transition from alignment research tool to commercial product reflects the founders' belief that making good reasoning cheaper and more accessible—even for less careful actors—ultimately reduces existential risk by shifting machine learning development toward safer, more alignable paradigms based on task decomposition rather than end-to-end training.7 The platform supports workflows ranging from systematic literature reviews (claiming up to 80% time savings) to research reports, screening, and data extraction across disciplines including materials science, biotech, software development, and health technology.13
The platform draws from three major academic databases: Semantic Scholar (200+ million publications with partnerships to 50+ major publishers), OpenAlex (243 million publications from 260,000+ sources, approximately twice the coverage of Web of Science or Scopus), and PubMed (biomedical and life sciences literature).8 Recent developments include the $22 million Series A funding round in February 2025 aimed at expanding beyond academia into industries like pharmaceuticals, medical technology, and consumer goods, positioning Elicit as a potential "research operating system" for evidence-based decision-making.9
History
Founding and Early Development
Elicit was co-founded in 2018 by Andreas Stuhlmüller and Jungwon Byun, emerging from Ought, a non-profit organization Stuhlmüller had initiated in 2017 to explore how machine learning could assist with thinking, reflection, and evidence-based decision-making.510 Stuhlmüller, who holds a Ph.D. in Cognitive Science from MIT (Josh Tenenbaum's group) and completed postdoctoral work at Stanford (Noah Goodman's Computational Cognitive Science lab), had been focused on automating reasoning since his teenage years and co-created the WebPPL programming language for probabilistic machine learning.1112 Byun brought operational expertise from her role as Head of Growth at Upstart (where she helped scale revenue and originations fivefold pre-IPO), management consulting experience at Oliver Wyman, and leadership of the Elmseed Enterprise Fund.1112
The initial focus centered on transforming research processes using language models to automate workflows like literature reviews, with the project incubated within Ought's broader mission to scale good reasoning through machine learning.510 The founders' approach emphasized task decomposition—breaking complex reasoning into verifiable subtasks—as a potentially safer alternative to end-to-end AI training methods.13 By mid-2022, Elicit had gained substantial traction in automating research tasks, leading the founders to transition the project from a non-profit tool to an independent public benefit corporation in 2023.51014
Funding and Growth
In September 2023, shortly after spinning out from Ought, Elicit secured $9 million in seed funding.5 The company's growth was characterized by organic word-of-mouth adoption and high Net Promoter Scores (NPS), which the team tracked as their North Star metric and improved monthly through rapid iteration.4 By late 2024, Elicit had achieved tens of thousands of paying subscribers and millions in annual revenue while serving over 400,000 monthly active users among the 2 million total researchers using the platform.49
On February 26, 2025, Elicit announced a $22 million Series A funding round led by Spark Capital and Footwork, with participation from existing investors including Fifty Years, Basis Set, and Mythos.9 The funding was explicitly aimed at expanding beyond academia into evidence-based AI applications for industries, particularly in anticipation of AI-driven economic shifts potentially occurring by 2027.9 The company emphasized building "the most trusted AI platform for evidence-backed decisions" while maintaining its focus on high-stakes applications requiring rigorous validation.9
Product Evolution and Recent Developments
From its initial focus on literature reviews, Elicit expanded to offer features including Deep Research, automated research overviews, Elicit Reports, systematic reviews, and meta-analyses.514 The platform's product philosophy centered on reliability in complex research through task decomposition and depth of processing, directly reflecting the founders' AI alignment research background.14 Key 2025 developments included:
- Integration of Claude Opus 4.5, which Elicit evaluated as outperforming peer models in data extraction and report generation (December 17, 2025)15
- Introduction of Research Agents for competitive landscape analysis and topic exploration (December 9, 2025)15
- Launch of Strict Screening and 80-Paper Reports capabilities (December 19, 2025)15
- Integration of ClinicalTrials.gov, making 545,000 clinical studies searchable through the platform15
- Achievement of SOC 2 Type II certification (October 30, 2025), demonstrating security and privacy controls16
The platform's feature set emphasizes interactive workflows with large tables, multi-step processes, customizable reports, and the ability to upload custom documents or integrate company subscriptions for additional full-text access.317 Elicit's positioning as a systematic review tool with keyword search capabilities across Elicit, PubMed, and ClinicalTrials.gov, combined with claims of up to 80% time savings, reflects its evolution toward comprehensive research automation.1518
Key People and Organizational Structure
Leadership Team
Andreas Stuhlmüller serves as Co-founder and CEO, bringing deep expertise in cognitive science and probabilistic programming. His academic background includes doctoral work in Josh Tenenbaum's group at MIT and postdoctoral research in Noah Goodman's lab at Stanford, where he co-created WebPPL, a programming language for probabilistic machine learning.1112 His lifelong focus on automating reasoning, dating back to his teenage years, directly shaped Elicit's emphasis on decomposed, verifiable AI systems rather than opaque end-to-end models.11 He served as an advisor to the 2025 AI for Human Reasoning Fellowship.
Jungwon Byun serves as Co-founder and COO, responsible for operational strategy and growth. Before joining Ought in 2019, she led growth at Upstart, where she scaled revenue and loan originations fivefold prior to the company's IPO.1112 Her background also includes management consulting at Oliver Wyman and leadership of the Elmseed Enterprise Fund, as well as undergraduate work on microfinance in Africa during her time at Yale, where she earned her B.A. in Economics cum laude.1119
James Brady (also referred to as James Barry in some sources) heads engineering, bringing extensive experience from his role as VP of Technology at Spring, a creator commerce platform.1120 He previously founded startups in personal knowledge management and developer tools and built the Payroll product team at Square.1120
Kevin Bird serves as Head of Product, joining from Nubank where he was the first product manager and spent eight years building fintech products.20 Chad Thornton leads design with a background at major technology companies including Airtable, Dropbox, and Medium.20 Sarah Park heads operations, though specific background details were not provided in available sources.11
Engineering and Technical Team
The team includes Spruce Bondera, an AI Engineer who previously founded Continuum and worked on vehicle motion planning at Waymo.21 The company maintains a high hiring bar and emphasizes values aligned with scaling reasoning for science, rapid product shipping, and organic growth through product-market fit.20
Technology and Product Capabilities
Core Functionality
Elicit's primary functionality centers on searching, summarizing, and extracting data from academic literature. The platform provides semantic search capabilities that find relevant papers even without perfect keyword matches, searching across 138 million papers from Semantic Scholar, PubMed, and OpenAlex.817 This coverage includes over 200 million publications from Semantic Scholar (with direct partnerships to more than 50 major publishers) and 243 million from OpenAlex, approximately double the coverage of Web of Science or Scopus.8 The platform does not include books, dissertations, patents, or non-academic publications.8
The platform's research reports follow a mini systematic review process, capable of synthesizing up to 40-80 papers into a single report with sentence-level citations backing every claim.22215 This citation approach distinguishes Elicit from chat-based AI tools that only link to references without specifying which sentences support which claims.3 For systematic literature reviews, Elicit automates screening and data extraction steps, providing detailed rationale for screening decisions and supporting refinement of criteria before full-scale application.18 Researchers using Elicit for systematic reviews report time savings of up to 80%.218
Data Sources and Integration
Elicit pulls from three major academic databases:8
- Semantic Scholar: 200+ million publications with direct partnerships to 50+ major publishers
- OpenAlex: 243 million publications from 260,000+ sources, with approximately twice the coverage of Web of Science or Scopus
- PubMed: Biomedical and life sciences literature
The 2025 integration of ClinicalTrials.gov added 545,000 clinical studies to Elicit's searchable corpus, expanding its utility for medical and pharmaceutical research.15 Users can also upload custom documents or integrate company subscriptions for additional full-text access.17
Advanced Features and Workflows
Elicit provides interactive workflows through multi-step processes and large interactive tables that allow researchers to explore literature dynamically and refine searches in real-time.3 Research Agents, introduced in December 2025, enable automated competitive landscape analysis and topic exploration.15 The platform's Notebooks feature allows researchers to organize and annotate findings, while Alerts notify users of new relevant publications.15
The systematic review workflow includes keyword search capabilities across Elicit, PubMed, and ClinicalTrials.gov, with Strict Screening options for applying inclusion/exclusion criteria.1518 Reports can be customized and follow systematic review-inspired methodologies, supporting workflows in materials discovery, UX innovation, algorithmic performance analysis, and health technology assessment.115
Elicit's technical approach emphasizes accuracy validation for scientific use, claiming to be the most accurate AI tool for research applications.1 The platform's December 2025 evaluation of Claude Opus 4.5 demonstrated superior performance in data extraction and report generation compared to peer models, reflecting ongoing investment in model selection and validation.15
Connection to AI Safety and Alignment Research
Origins in AI Alignment Work
Elicit's development is deeply rooted in AI alignment research conducted at Ought, the non-profit organization founded by Andreas Stuhlmüller in 2017.56 Ought's research program explored how machine learning could scale up good reasoning, with a particular focus on "scalable oversight" methods that would allow humans to verify AI systems' reasoning processes even for tasks too complex for direct human evaluation.237 This work investigated debate-based alignment methods where AI systems would justify their conclusions through decomposed reasoning steps that humans could audit.2324
The theoretical foundation emphasized task decomposition as a potentially safer paradigm for AI development compared to end-to-end training. By breaking complex reasoning tasks into verifiable subtasks—such as separating literature search from summarization, extraction, and synthesis—Ought hypothesized that AI systems could be more alignable because humans could inspect intermediate steps rather than only evaluating final outputs.1314 A 2021 AI alignment literature review highlighted how Elicit tested compositional AI training by decomposing science questions into subtasks, potentially creating systems less prone to reward hacking or emergent misalignment.13
Theory of Change and X-Risk Considerations
Ought's theory of change, articulated in posts on the EA Forum and LessWrong, positioned Elicit as net positive for existential risk reduction through two primary mechanisms.7 First, by making good reasoning cheaper and more accessible, Elicit would increase the total amount of careful thinking about existential risks, including AI alignment itself. Second, by demonstrating the viability of decomposition-based approaches to AI development, Elicit could shift machine learning research efforts toward paradigms that are inherently more alignable than black-box, end-to-end systems.7
The founders acknowledged potential downsides, noting that making reasoning more accessible could benefit "less careful actors" and that scaling research automation could accelerate AI capabilities development without proportional safety work.725 However, they argued that the approach of making good reasoning cheaper would disproportionately benefit careful actors (who face higher opportunity costs for time spent on research) and that monitoring for misuse was part of their ongoing strategy.7 Coefficient Giving, which provided approximately $15 million in funding across Elicit's seed and Series A rounds, explicitly frames its support as contributing to AI alignment efforts aimed at reducing existential risks from misaligned artificial general intelligence.2526
Alignment Techniques in Practice
In practice, Elicit incorporates several alignment techniques that have become standard in language model deployment. The platform uses reinforcement learning from human feedback (RLHF) to ensure outputs align with researcher intentions and avoid harmful content.2728 Red-teaming processes test the system for potential failure modes, and the platform emphasizes transparency through sentence-level citations that allow users to verify AI claims against source material.228 The integration of models like Anthropic's Claude, which employs Constitutional AI for alignment, reflects ongoing attention to incorporating safety-focused language models.29
Elicit's accuracy validation, which the company claims makes it the most accurate AI for scientific research, includes rigorous testing of data extraction and summarization against ground truth research findings.1 This emphasis on verifiability and accuracy serves dual purposes: ensuring the tool is scientifically reliable while also addressing alignment concerns about AI systems producing plausible but incorrect information (a form of specification gaming where the system optimizes for sounding correct rather than being correct).30
Criticisms and Limitations
Inherited Alignment Risks from Foundation Models
Critics within the AI safety community note that Elicit's reliance on frontier large language models (such as GPT variants and Claude) means the platform inherits the general alignment risks present in those underlying systems.253031 These risks include potential reward hacking (where models optimize for proxy metrics rather than true objectives), emergent behaviors that weren't present during training, and the possibility of specification gaming where systems produce outputs that satisfy evaluation criteria without genuinely meeting user intent.2530 If Elicit's approach were scaled to AGI-level research tools without addressing these fundamental alignment challenges, it could potentially amplify existential risk by accelerating capabilities development faster than safety research advances.2530
Discussions on EA Forum and LessWrong have debated whether Elicit sufficiently prioritizes safety over productivity in its commercial deployment.25 Some argue that the shift from Ought's non-profit research focus to a for-profit company optimizing for revenue and user growth may dilute the original alignment-focused mission, potentially prioritizing features that increase engagement and subscriptions over those that ensure long-term safety.25 While no major scandals have emerged, the general concern about rare but potentially harmful outputs from language models applies to Elicit as it does to all LLM-based tools.27
Methodological and Coverage Limitations
Elicit's database coverage, while extensive at 138 million papers across Semantic Scholar, OpenAlex, and PubMed, explicitly excludes books, dissertations, patents, and non-academic publications.8 This limitation means researchers conducting comprehensive literature reviews in fields that rely heavily on gray literature, policy documents, or industry reports may find gaps in coverage. The platform's focus on academic papers also means it may miss practitioner knowledge published in conference proceedings, technical reports, or preprint servers not indexed by its three primary sources.
The platform's accuracy claims, while validated through internal testing showing 90%+ accuracy in paper extraction, lack comprehensive independent validation published in peer-reviewed venues.32 The December 2025 evaluation of Claude Opus 4.5 showing superior performance was conducted by Elicit itself rather than by independent researchers, raising questions about potential evaluation bias.15 Users report generally high satisfaction through Net Promoter Score metrics, but systematic studies comparing Elicit's output quality against human expert literature reviews remain limited in published literature.4
Scope and Accessibility Constraints
Elicit's pricing model, with tens of thousands of paying subscribers supporting millions in annual revenue, suggests that full access to advanced features requires paid subscriptions.4 While this is standard for commercial research tools, it creates accessibility barriers for researchers at institutions or in regions with limited funding, potentially concentrating the benefits of AI-augmented research among already well-resourced researchers. The platform's focus on English-language academic literature, typical of major scholarly databases, also limits its utility for research that requires multilingual coverage or access to non-English academic traditions.
The tool's emphasis on systematic review and literature synthesis workflows means it is optimized for specific research tasks rather than serving as a general-purpose research assistant. Researchers conducting highly specialized searches, working with niche databases, or requiring deep domain expertise in result interpretation may find Elicit's automated approach less suitable than traditional manual methods combined with specialist librarian assistance. The platform's ability to analyze up to 1,000 papers, while substantial, still represents a fraction of the corpus in some research domains where comprehensive reviews might require examining tens of thousands of studies.
Key Uncertainties
Long-term alignment trajectory under commercial pressures: It remains uncertain whether Elicit's for-profit structure will maintain the alignment-focused development approach that characterized Ought's non-profit research, particularly as the company scales and faces pressure to maximize revenue and market share. The balance between rapid feature development for user acquisition and careful safety validation will be tested as the platform expands into new industries and use cases beyond academic research.
Scalability of decomposition-based alignment: While Elicit demonstrates that task decomposition can work for literature review automation, it is unclear whether this approach will scale to more complex reasoning tasks or whether it can prevent emergent misalignment in systems approaching AGI-level capabilities. The founders' theory that decomposition creates more alignable systems remains a hypothesis requiring validation across increasingly difficult domains.
Impact on research quality and scientific culture: The net effect of widespread research automation on scientific rigor and reproducibility is uncertain. While Elicit may democratize access to literature review capabilities and save researcher time, it could also enable publication of lower-quality research if users rely too heavily on automated summaries without deep engagement with primary sources. The platform's influence on citation patterns, literature synthesis quality, and researcher skill development warrants ongoing empirical investigation.
Market positioning and competitive landscape: Elicit's expansion beyond academia into industrial applications faces competition from established enterprise research tools, general-purpose AI assistants, and specialized vertical solutions. Whether the platform can maintain its emphasis on accuracy, transparency, and sentence-level citations while competing on speed and user experience with less rigorous but potentially more convenient alternatives remains to be seen.
Effectiveness for existential risk reduction: Whether Elicit's approach—making good reasoning cheaper and shifting ML development toward decomposition—actually reduces existential risk on net is fundamentally uncertain. The accelerating-capabilities-without-proportional-safety concern raised by critics could outweigh the benefits of increasing careful thinking about risks, particularly if the tool enables faster AI development timelines without corresponding advances in alignment solutions.
Sources
Footnotes
-
Built In SF - Elicit Company Profile — Built In SF - Elicit Company Profile ↩ ↩2 ↩3 ↩4 ↩5
-
Elicit - NBT Substack Profile — Elicit - NBT Substack Profile ↩ ↩2 ↩3 ↩4 ↩5
-
Canvas Business Model - Brief History of Elicit — Canvas Business Model - Brief History of Elicit ↩ ↩2 ↩3 ↩4 ↩5 ↩6 ↩7
-
Fifty Years - Elicit Investment — Fifty Years - Elicit Investment ↩ ↩2
-
EA Forum - Ought's Theory of Change — EA Forum - Ought's Theory of Change ↩ ↩2 ↩3 ↩4 ↩5 ↩6
-
Elicit Support - Data Sources — Elicit Support - Data Sources ↩ ↩2 ↩3 ↩4 ↩5 ↩6
-
Elicit Blog - Series A Announcement — Elicit Blog - Series A Announcement ↩ ↩2 ↩3 ↩4 ↩5
-
Elicit Team Page — Elicit Team Page ↩ ↩2 ↩3 ↩4 ↩5 ↩6 ↩7 ↩8 ↩9
-
Cognitive Revolution - Unbounded AI-Assisted Research with Elicit Founders — Cognitive Revolution - Unbounded AI-Assisted Research with Elicit Founders ↩ ↩2 ↩3 ↩4
-
Alignment Forum - 2021 AI Alignment Literature Review — Alignment Forum - 2021 AI Alignment Literature Review ↩ ↩2 ↩3
-
Elicit Blog - AI for Human Reasoning — Elicit Blog - AI for Human Reasoning ↩ ↩2 ↩3 ↩4
-
Elicit Blog - Features Tag — Elicit Blog - Features Tag ↩ ↩2 ↩3 ↩4 ↩5 ↩6 ↩7 ↩8 ↩9 ↩10 ↩11 ↩12 ↩13
-
Elicit Solutions - Search — Elicit Solutions - Search ↩ ↩2 ↩3
-
Elicit Solutions - Systematic Review — Elicit Solutions - Systematic Review ↩ ↩2 ↩3 ↩4
-
TechCrunch - Elicit Building Tool to Automate Scientific Literature Review — TechCrunch - Elicit Building Tool to Automate Scientific Literature Review ↩
-
NBT Substack - Elicit (Team Details) — NBT Substack - Elicit (Team Details) ↩ ↩2 ↩3 ↩4 ↩5
-
Elicit Team Page - Spruce Bondera — Elicit Team Page - Spruce Bondera ↩
-
Alignment Forum - Newcomer's Guide to Technical AI Safety Field — Alignment Forum - Newcomer's Guide to Technical AI Safety Field ↩ ↩2
-
OpenAI - How We Think About Safety Alignment — OpenAI - How We Think About Safety Alignment ↩
-
Alignment Forum - Newcomer's Guide to Technical AI Safety Field (Funding Details) — Alignment Forum - Newcomer's Guide to Technical AI Safety Field (Funding Details) ↩ ↩2 ↩3 ↩4 ↩5 ↩6 ↩7
-
Coefficient Giving - Other Scientific Research Areas — Coefficient Giving - Other Scientific Research Areas ↩
-
Witness.ai - AI Alignment Blog — Witness.ai - AI Alignment Blog ↩
-
arXiv - AI Alignment Paper (2310.19852) — arXiv - AI Alignment Paper (2310.19852) ↩ ↩2 ↩3 ↩4
-
Stanford AI Alignment - What is AI Safety — Stanford AI Alignment - What is AI Safety ↩
-
University of Arizona LibGuides - AI for Researchers: Elicit — University of Arizona LibGuides - AI for Researchers: Elicit ↩
References
“We focus on scalable oversight mechanisms that evolve alongside the increasing capabilities of AI systems.”
The source does not mention Elicit, Andreas Stuhlmüller, Ought, or debate-based alignment methods.
“This is what I'm personally working on as CEO and co-founder of Elicit. We mostly help people understand and interpret the scientific literature, though we're trying to make that much broader—moving from the "evidence-based" part toward the "decision-making" part.”
unsupported: The source does not mention the initial focus being on transforming research processes using language models to automate workflows like literature reviews. unsupported: The source does not mention the project being incubated within Ought's broader mission to scale good reasoning through machine learning. unsupported: The source does not mention the founders' approach emphasizing task decomposition as a potentially safer alternative to end-to-end AI training methods. unsupported: The source does not mention Elicit gaining substantial traction in automating research tasks by mid-2022. unsupported: The source does not mention the founders transitioning the project from a non-profit tool to an independent public benefit corporation in 2023.
“Evaluating Claude Opus 4.5 Claude Opus 4.5 is better than Sonnet 4.5, Google Gemini 3 Pro, and OpenAI GPT 5 at data extraction and writing reports with fewer hallucinations.”
“Initially incubated within Ought, a non-profit, Elicit transitioned to an independent public benefit corporation in 2023.”
The claim states that Ought was founded in 2017, but the source says Andreas Stuhlmüller initiated Ought in 2017. The claim states that the platform supports workflows ranging from systematic literature reviews (claiming up to 80% time savings), but the source does not mention the 80% time savings. The claim mentions that Elicit supports workflows for research reports, screening, and data extraction across disciplines including materials science, biotech, software development, and health technology, but the source only mentions that Elicit focuses on automating literature reviews.
“Elicit, the AI research assistant, helps you automate time-consuming research tasks like summarizing papers, extracting data, and synthesizing your findings.”
WRONG NUMBERS: The source states that Elicit has 15 employees, while the claim mentions over 2 million users and tens of thousands of paying subscribers. FABRICATED DETAILS: The source does not mention the number of academic papers analyzed, the number of data points, accuracy validation, sentence-level citations, top academic institutions, enterprises, annual revenue, or organic growth through word-of-mouth. WRONG ATTRIBUTION: The claim describes Elicit as an AI-powered research assistant designed to automate scientific research tasks, which is a correct description, but the source doesn't provide the specific details about scale, accuracy validation, and transparency mentioned in the claim.
“Elicit enables you to combine multiple searches across many databases: Use AI-based semantic search to find papers that don't overlap”
The claim mentions "interactive workflows with large tables, multi-step processes, customizable reports", but the source only mentions customizable reports and data extraction from tables. The claim mentions keyword search capabilities across Elicit, PubMed, and ClinicalTrials.gov, but the source only mentions Elicit searches across ClinicalTrials.gov. The claim mentions Elicit's evolution toward comprehensive research automation, but the source does not explicitly mention this evolution.
“RLHF is a training method where human feedback helps guide the learning process of an AI model .”
The claim mentions sentence-level citations for verifying AI claims, but the source does not explicitly mention sentence-level citations. The claim mentions the integration of models like Anthropic's Claude, which employs Constitutional AI for alignment, but the source only mentions Anthropic in the context of organizations leading AI safety efforts.
“Elicit Reports follow a mini systematic review process, the gold standard for evidence synthesis.”
The claim mentions synthesizing 'up to 40-80 papers', while the source only mentions '80-Paper Reports'. The claim that 'Researchers using Elicit for systematic reviews report time savings of up to 80%' is not directly supported by the source.
“Elicit achieves SOC 2 Type II certification.”
“The spinout from Ought, a nonprofit research lab, launched the first version of its product two years ago and has by word-of-mouth alone built up a monthly user base of 200,000 researchers, ranging from MIT and Stanford scientists to researchers at industry leaders like Novartis, Genentech, and OpenAI.”
The claim that Ought was founded in 2017 is not explicitly stated in the source, although it does say that Andreas Stuhlmüller started Ought in 2017. The claim that Elicit supports workflows ranging from systematic literature reviews (claiming up to 80% time savings) is not supported by the source. The source mentions that researchers using Elicit halved the time and cost of data extraction and analysis in a pilot project, but it does not mention 80% time savings. The claim that Elicit supports workflows ranging from research reports, screening, and data extraction across disciplines including materials science, biotech, software development, and health technology is not explicitly stated in the source. The source mentions that Elicit is used by researchers ranging from MIT and Stanford scientists to researchers at industry leaders like Novartis, Genentech, and OpenAI, and that it is used in fields like semiconductors, energy, and materials science.
“Elicit has raised $22M in Series A funding at a $100M valuation led by Spark Capital and Footwork.”
unsupported: The claim mentions specific databases (Semantic Scholar, OpenAlex, PubMed) and their sizes, which are not mentioned in the provided source. unsupported: The claim mentions Elicit being positioned as a potential 'research operating system' which is not mentioned in the provided source. minor_issues: The claim mentions industries like pharmaceuticals, medical technology, and consumer goods, while the source lists Pharma, Medtech, and Consumer goods. minor_issues: The claim mentions the funding round in February 2025, which is mentioned in the source as Feb 26, 2025.
“Since launching last year, Elicit has grown through word-of-mouth to over 400,000 monthly active active users, tens of thousands of paying subscribers, and millions in annual revenue.”
WRONG NUMBERS: The source states "over 400,000 monthly active active users", while the claim states "over 2 million users". WRONG NUMBERS: The source is dated Feb 26, 2025, and states "Since launching last year...", implying Elicit launched in 2024. The claim does not specify a launch date. OVERCLAIMS: The claim that Elicit is used by researchers from top academic institutions and enterprises is not explicitly stated in the source, although it is implied. UNSUPPORTED: The claim that Elicit analyzes up to 1,000 papers and 20,000 data points is not supported by the source. UNSUPPORTED: The claim that Elicit distinguishes itself through accuracy validation for scientific use is not explicitly supported by the source, although it is implied through the description of Elicit Reports. UNSUPPORTED: The claim that Elicit distinguishes itself through transparency through sentence-level citations rather than simple chat interfaces is not explicitly supported by the source, although it is implied through the description of Elicit Reports.
“Andreas and Jungwon have held a very high bar for hiring, with folks on the team like Kevin Bird , head of product, who was the first PM at Nubank and had an incredible 8 year journey there, James Barry , head of engineering, who was previously VP Engineering at Teespring, which he joined after selling his prior company to Square, and Chad Thornton , head of design, who led design at Airtable, Dropbox, and Medium.”
The source mentions James Barry as head of engineering, not James Brady. The source states James Barry was previously VP Engineering at Teespring, not VP of Technology at Spring. The source does not mention that James Barry founded startups in personal knowledge management and developer tools or that he built the Payroll product team at Square, only that he sold his prior company to Square.
“BYOData Upload your own documents Integrate company subscriptions to access more full-text papers Supported files”
The claim mentions interactive workflows with large tables, multi-step processes, and customizable reports, but the source doesn't explicitly detail these features. The claim of 80% time savings is not mentioned in the source.
“in Cognitive Science from MIT (Josh Tenenbaum's group) and completed postdoctoral work at Stanford (Noah Goodman's Computational Cognitive Science lab), had been focused on automating reasoning since his teenage years and co-created the WebPPL programming language for probabilistic machine learning.”
“Elicit searches across over 138 million academic papers from Semantic Scholar, PubMed, and OpenAlex. Our Data Sources Elicit pulls papers from these major sources: Semantic Scholar 200+ million publications across all fields Direct partnerships with 50+ major publishers Learn more about Semantic Scholar's coverage OpenAlex 243 million publications from 260,000+ sources About 2x broader coverage than Web of Science or Scopus Learn more about OpenAlex's coverage PubMed 33+ million citations Focuses on biomedical literature, life sciences, and clinical medicine Learn more about PubMed”
WRONG NUMBERS: PubMed has 33+ million citations, not an unspecified amount. UNSUPPORTED: The source does not mention a $22 million Series A funding round or expansion into industries like pharmaceuticals, medical technology, and consumer goods. UNSUPPORTED: The source does not mention Elicit being positioned as a potential 'research operating system'.
“Academic papers Search over 138 million academic papers and conference proceedings Updates weekly Included sources”
WRONG NUMBERS: The wiki claim states 'over 125 million academic papers' while the source says '138 million academic papers'. WRONG NUMBERS: The wiki claim states 'over 2 million users' while the source says 'over 5 million researchers'. UNSUPPORTED: The claim that Elicit generates 'millions in annual revenue from tens of thousands of paying subscribers' is not supported by the source.
“Our mission is to automate and scale open-ended reasoning. If we can improve the world’s ability to reason, we’ll unlock positive impact across many domains including AI governance & alignment, psychological well-being, economic development, and climate change.”
The claim that Ought was founded in 2017 is unsupported by the source. The claim that Elicit supports workflows across disciplines including materials science, biotech, software development, and health technology is unsupported by the source. The claim that Elicit provides up to 80% time savings is unsupported by the source.
“Spruce Bondera AI Engineer Before Elicit, Spruce was a founder at Continuum, an LLM coding assistant that focused on creating and traversing knowledge graphs on large codebases. Prior to that he worked on vehicle motion planning at Waymo as they first tested and launched in San Francisco.”
unsupported: The company maintains a high hiring bar and emphasizes values aligned with scaling reasoning for science, rapid product shipping, and organic growth through product-market fit.
“Before starting Elicit, Andreas was a postdoctoral researcher in Noah Goodman's Computational Cognitive Science lab at Stanford. He co-created WebPPL, a programming language for probabilistic machine learning. He holds a Ph.D. in Cognitive Science from MIT, where he worked in Josh Tenenbaum's group.”
The claim that Andreas Stuhlmüller had been focused on automating reasoning since his teenage years is not supported by the source. The claim that Byun helped scale revenue and originations fivefold pre-IPO is not supported by the source.
“Scale Elicit can find up to 1,000 relevant papers and analyze up to 20,000 data points at once. Accuracy Elicit is the most accurate AI product for scientific research. Learn about how we validated Elicit's accuracy. Transparency Elicit supports all AI-generated claims with sentence-level citations from the underlying sources. More than chat Elicit goes beyond chat to provide rich, interactive tables and multi-step workflows.”
The source states "TRUSTED BY OVER 5 MILLION RESEARCHERS", not 2 million users. The source states that Elicit searches "over 138 million academic papers", not 125 million. The source does not mention that Elicit generates millions in annual revenue from tens of thousands of paying subscribers.