What are foundation models?
Foundation models — also called large language models (LLMs) — are general-purpose AI systems trained on massive datasets. OpenAI's GPT-4 (ChatGPT), Anthropic's Claude, Google's Gemini, xAI's Grok, and DeepSeek are the leading foundation models in 2026. They can summarize documents, answer questions, draft text, and reason through complex problems. They are the most significant technology breakthrough in a generation.
Can you use ChatGPT for PE due diligence?
You can — and many individual investors already do — but it breaks down at the institutional level. ChatGPT can summarize a CIM if you paste it in. Claude can analyze a financial statement. Gemini can draft an investment memo. The problem is what happens next: there is no audit trail showing where numbers came from, no structured Excel or PowerPoint output, no connection to your CRM or data room, no persistent memory across deals, and no compliance infrastructure. For an individual analyst doing quick research, foundation models are excellent. For a firm that needs auditable, repeatable, team-wide workflows, they are incomplete.
What's missing when using AI models directly?
Foundation models were designed for general-purpose conversations, not institutional investment workflows. When PE, VC, and growth equity firms try to use ChatGPT or Claude directly, they run into the same gaps every time.
- No source tracing — outputs cannot be traced to a specific page in a source document
- No structured outputs — models produce text, not formatted Excel models or branded PowerPoint decks
- No CRM integration — deal context is siloed from your pipeline in Affinity, DealCloud, or Salesforce
- No persistent deal memory — each conversation starts from scratch with no knowledge of prior analysis
- No team workflows — one analyst's conversation is invisible to the rest of the deal team
- No compliance infrastructure — SOC 2, data retention policies, and access controls are absent
- No domain-specific pipelines — no built-in logic for due diligence, portfolio monitoring, or LP reporting
What happens to your data after the chat ends?
This is the question that separates tools from infrastructure. ChatGPT is stateless — every conversation starts from zero. Upload a CIM on Monday, and by Tuesday the model has no memory of it. For a firm evaluating 200 deals per year and monitoring 30 portfolio companies, this means re-uploading, re-explaining, and re-contextualizing every single interaction. Emblem maintains a persistent institutional knowledge graph. Every document, every metric, every deal memo, every analyst note accumulates into a firm-wide intelligence layer that compounds over time. Your 50th deal analysis is fundamentally better than your first — because the system carries forward pattern recognition across your entire deal history. A chatbot gives you answers. An operating system gives you institutional memory.
How does Emblem handle compliance and audit requirements?
When an LP asks how a valuation was derived, or a regulator requests documentation of the diligence process, a ChatGPT conversation history is not a compliance record. Emblem maintains a complete audit chain: every claim is traced to a source document and page number, every model assumption links back to its origin, every analyst interaction is logged with timestamps and user attribution. For firms operating under SEC, FCA, or MAS oversight — or simply meeting LP expectations for transparency — this is not a feature. It is the baseline requirement that foundation models do not meet.
How does Emblem use foundation models?
Emblem is not a competitor to ChatGPT, Claude, or Gemini. Emblem uses them. The platform orchestrates multiple AI models — including OpenAI, Anthropic's Claude, Google's Gemini, xAI's Grok, and DeepSeek — selecting the right model for each task. A financial extraction task might use one model; a memo drafting task might use another. This model-agnostic architecture means Emblem always uses the best available AI for each workflow step, and firms are never locked into a single provider. On top of this multi-model backbone, Emblem adds the infrastructure institutional investors need: RAG-powered source tracing, structured output generation (Excel, PowerPoint, Word), CRM and cloud storage integrations, persistent deal context, team collaboration, and SOC 2 Type II compliance.
- Model-agnostic: Orchestrates OpenAI, Claude, Gemini, Grok, and DeepSeek
- RAG pipeline: Every output is source-traced to the original document and page
- Structured outputs: Generates real Excel models, branded PowerPoint decks, and Word memos
- Integrations: Connects to Affinity, DealCloud, Salesforce, HubSpot, Attio, Box, Egnyte
- Compliance: SOC 2 Type II certified with enterprise data controls
Why do firms choose Emblem over direct model access?
The analogy is ERP software. Every enterprise uses databases, but no one builds their own ERP from raw PostgreSQL. The database is the foundation; the ERP is the application layer that makes it useful for business operations. Foundation models are the database. Emblem is the application layer that makes them useful for investment operations. Firms choose Emblem because they need outputs they can trust (source tracing), outputs in formats they can use (Excel, PowerPoint), connections to systems they already run (CRM, data rooms), and a platform their compliance team can approve (SOC 2). You would not build your own ERP. Do not build your own AI workflow.
Official Integration Partners
Also Integrates With
Frequently Asked Questions
Does Emblem use ChatGPT?
Can ChatGPT generate Excel financial models?
Is ChatGPT secure enough for PE firms?
Why not just build internal tools on top of ChatGPT's API?
What does model-agnostic mean?
Related Solutions
Related Resources
Bring your firm to the future
See how Emblem automates due diligence, portfolio monitoring, and LP reporting with 100% source-traced accuracy.
Get your time back