Research Analyst
AI-powered use cases for research analyst professionals.
1. AI Market Sizing & TAM/SAM/SOM Calculator
Transforms raw market data into defensible TAM/SAM/SOM estimates with source-cited methodology — in hours, not weeks.
Pain Point & How COCO Solves It
The Pain: Market Sizing Is Simultaneously Critical and Chronically Unreliable
Market sizing is the foundation of every strategic recommendation, investment thesis, and business case — yet it remains one of the most time-intensive and methodologically inconsistent deliverables a research analyst produces. Analysts typically start from scratch each engagement: hunting for relevant industry reports across fragmented databases, reconciling conflicting figures from IDC, Gartner, Statista, and government sources, and manually building top-down and bottom-up models in spreadsheets that each consultant structures differently. A mid-complexity market sizing that should take two days often consumes a full week before the analyst even reaches synthesis.
The quality problem compounds the time problem. When inputs come from reports published in different years, using different geographic definitions, or employing incompatible segmentation logic, the resulting TAM/SAM/SOM figures are built on assumptions that cannot withstand scrutiny. Clients and investment committees increasingly challenge market sizing in detail — asking for source transparency, CAGR justification, and sensitivity analysis. Analysts who cannot show their methodology face credibility erosion, and those who take time to build rigorous models face deadline pressure that forces them to cut corners anyway.
The downstream impact extends across the entire engagement. An underestimated TAM can kill a valid business case; an inflated one can push a client toward an unwinnable market. Inconsistent sizing methodology across reports from the same firm creates contradictions that damage client trust and require expensive rework. Junior analysts lack the experience to know which sources to prioritize, leading to over-reliance on a single vendor report and blind spots that senior reviewers must catch — adding another review cycle to an already compressed timeline.
How COCO Solves It
Multi-Source Market Data Aggregation: COCO pulls and reconciles market data from across the research landscape simultaneously:
- Queries industry report databases, government statistical agencies, trade association publications, and earnings disclosures in parallel
- Reconciles conflicting figures by flagging methodology differences — geographic scope, revenue vs. shipment basis, end-user vs. channel counting
- Normalizes all data to a consistent base year, currency, and unit of measure before synthesis
- Identifies the most recently published figure for each data point and surfaces publication date for every source
- Builds a source credibility ranking based on methodology transparency, publication recency, and institutional authority
Top-Down and Bottom-Up Model Construction: COCO builds both sizing approaches simultaneously and triangulates:
- Constructs a top-down model from total addressable universe estimates with segmentation filters applied
- Builds a bottom-up model from unit economics: average contract value, addressable customer count, penetration rate assumptions
- Compares top-down and bottom-up results and surfaces the delta with diagnostic commentary on likely causes
- Applies CAGR projections from multiple analyst consensus sources, weighted by recency and relevance
- Generates a reconciled central estimate with explicit confidence range based on source variance
TAM/SAM/SOM Segmentation Engine: COCO applies market segmentation logic with precision:
- Defines TAM using the broadest relevant market definition with geographic and demographic scope stated explicitly
- Narrows to SAM by applying serviceable constraints: geography, channel, regulatory eligibility, technical compatibility
- Calculates SOM using competitive density data, company-specific go-to-market capacity, and realistic penetration benchmarks
- Supports custom segmentation variables — vertical, company size, use case, price tier — applied consistently across all three levels
- Documents every assumption used in each segmentation step for client-facing transparency
Sensitivity and Scenario Analysis: COCO models uncertainty and tests assumptions systematically:
- Identifies the three to five input variables with the greatest impact on final market size estimate
- Builds bear/base/bull scenarios by varying key assumptions across a defined range
- Generates a tornado chart showing relative sensitivity of the output to each input variable
- Calculates the market size estimate at each scenario with full recalculation of dependent figures
- Flags which assumptions are most contested in the literature and require primary research validation
CAGR and Forecast Projection: COCO builds defensible market growth projections:
- Aggregates CAGR estimates from multiple analyst firms and calculates a consensus range with outlier identification
- Projects market size for 3-, 5-, and 10-year horizons using compound growth on base year estimate
- Adjusts growth projections for identified macro headwinds and tailwinds with source citations
- Compares forecast against historical growth rates to assess reasonableness and flag inflection points
- Generates a growth driver narrative explaining the key forces behind projected expansion or contraction
Client-Ready Report Generation: COCO produces structured deliverables formatted for immediate use:
- Generates an executive summary with single headline TAM/SAM/SOM figures and one-paragraph methodology rationale
- Produces a detailed methodology appendix with every source, assumption, and calculation step documented
- Creates visualization-ready data tables for market size waterfall charts, scenario comparison charts, and CAGR trend lines
- Formats citations in the client's preferred style (APA, Chicago, or custom footnote format)
- Supports iterative refinement: analysts adjust assumptions and COCO regenerates all dependent figures automatically
Results & Who Benefits
Measurable Results
- Market sizing cycle time: Reduced from 5–8 analyst days to under 12 hours for a mid-complexity market (75–85% time savings)
- Source coverage: COCO synthesizes data from 8–15 independent sources per market vs. 2–3 in typical manual research, reducing single-source dependency risk
- Methodology consistency: Firms report 90%+ reduction in cross-report sizing inconsistencies when using a standardized COCO framework
- Client challenge rate: Market sizing sections subjected to client pushback drop from 42% of engagements to under 8% when COCO-generated methodology documentation is included
- Scenario modeling speed: Full bear/base/bull scenario analysis completed in under 2 hours vs. 1–2 additional analyst days manually
Who Benefits
- Research Analysts: Eliminate the data-hunting and spreadsheet-building phases that consume most of market sizing time — redirect effort to interpretation, strategic implication, and client communication
- Strategy Consultants: Deliver market sizing deliverables in the first week of an engagement rather than the final week, enabling earlier client alignment and reducing rework from late-stage scope changes
- Investment Analysts: Build market size assumptions into financial models with documented, auditable methodology that withstands LP and investment committee scrutiny
- Business Development Teams: Access fast, credible market sizing for proposal development without pulling analyst resources from billable engagements
💡 Practical Prompts
Prompt 1 — Top-Down TAM Estimation
Build a top-down TAM/SAM/SOM analysis for the following market.
Market: [product/service category]
Geography: [global / region / country]
Target customer segment: [enterprise / SMB / consumer / specific vertical]
Base year: [year]
Forecast horizon: [3 / 5 / 10 years]
Use the following data sources if available: [list known reports or leave blank for COCO to source]
Deliverables:
1. TAM: Total addressable market with definition, methodology, and source citations
2. SAM: Serviceable addressable market with segmentation constraints explained
3. SOM: Serviceable obtainable market with penetration rate assumptions and benchmarks
4. CAGR: Growth rate with consensus range from multiple analyst sources
5. 5-year market size projection with bear/base/bull scenarios
6. Methodology appendix: every source, assumption, and calculation documented
Output format: Executive summary table + detailed methodology section + source citation listPrompt 2 — Bottom-Up Market Sizing
Build a bottom-up market size estimate for [market/product category].
Unit economics inputs:
- Average selling price / contract value: [$ amount or "estimate from industry data"]
- Target customer universe: [estimated number or "derive from industry data"]
- Buying frequency / contract duration: [annual / multi-year / one-time]
- Addressable customer eligibility rate: [% or "estimate"]
Geography: [market scope]
Segmentation cuts needed: [by vertical / company size / region / use case]
Deliverables:
1. Customer universe derivation with source citations
2. Bottom-up market size calculation with each input stated explicitly
3. Comparison of bottom-up result against top-down estimate from published reports
4. Reconciliation commentary: explain the delta between approaches
5. Sensitivity table: show market size at ±20% and ±40% variation in key inputs
Highlight assumptions that are most uncertain and recommend primary research to validate them.Prompt 3 — Competitive Market Share Analysis
Estimate the market share distribution for the following market and identify sizing implications.
Market: [market name]
Geography: [scope]
Key players to analyze: [list companies or "identify top players from public sources"]
Time period: [current year and trend over past 3 years if data available]
For each major player, extract or estimate:
- Revenue attributable to this market segment
- YoY growth rate
- Implied market share percentage
- Data source and confidence level
Deliverables:
1. Market share table for top [5/10] players with revenue, growth, and share figures
2. Market concentration analysis: HHI score, CR4/CR8 ratios
3. Share shift analysis: which players are gaining or losing share and why
4. Implied total market size derived from known player revenues and estimated coverage
5. White space identification: segments or geographies with low competitive densityPrompt 4 — Market Sizing Sensitivity Model
Build a sensitivity analysis for the following market size estimate.
Base case estimate: [$ value] for [market] in [year]
Key input assumptions:
1. [Assumption 1]: Base value [X], plausible range [low–high]
2. [Assumption 2]: Base value [X], plausible range [low–high]
3. [Assumption 3]: Base value [X], plausible range [low–high]
[Add up to 6 assumptions]
Deliverables:
1. One-way sensitivity table: market size at each assumption varied independently
2. Tornado chart data: rank assumptions by impact magnitude on final market size
3. Two-way sensitivity table for the top two most impactful assumptions
4. Bear/base/bull scenario definitions with coherent assumption sets for each
5. Probability-weighted expected market size if scenario probabilities can be estimated
Flag which assumptions most urgently need primary research validation to narrow the range.Prompt 5 — Market Sizing Methodology Documentation
Generate a methodology documentation section for the following market sizing analysis.
Market sized: [market name]
Final estimates: TAM [$ value], SAM [$ value], SOM [$ value]
Approach used: [top-down / bottom-up / triangulated]
Key sources used: [list sources with publication dates]
Key assumptions: [list 3–5 critical assumptions]
Geographic scope: [definition]
Base year: [year]
Segmentation logic: [explain how TAM was narrowed to SAM and SAM to SOM]
Generate:
1. A methodology rationale section (2–3 paragraphs) suitable for inclusion in a client report appendix
2. A source citation table with author, title, publisher, date, and relevance to each figure
3. An assumptions register: each assumption listed with rationale, sensitivity, and validation status
4. A limitations and caveats section disclosing known gaps or data quality issues
5. A peer review checklist: questions a senior reviewer should ask to validate the methodology
Format for inclusion in a formal consulting or investment research report.2. AI Competitive Intelligence Deep-Dive Report Builder
Builds comprehensive competitive profiles — products, pricing, positioning, leadership, and signals — synthesized from 50+ public sources into a single structured report.
Pain Point & How COCO Solves It
The Pain: Competitive Intelligence Is Perpetually Incomplete and Perpetually Outdated
Research analysts tasked with competitive intelligence face a paradox: the information they need most is the information competitors work hardest to conceal. Reconstructing a competitor's product roadmap, pricing architecture, go-to-market strategy, and organizational structure from public sources requires stitching together earnings transcripts, job postings, patent filings, press releases, review site data, and conference presentations — each from a different source, in a different format, updated on a different schedule. A thorough competitive profile for a single company can require 40 to 60 hours of analyst time, and it is already partially outdated by the time it is delivered.
The fragmentation problem is compounded by the signal-to-noise ratio in competitive research. Most publicly available information about a competitor is either promotional (what they want you to think) or backward-looking (what they did last quarter). The actionable intelligence — product roadmap direction inferred from engineering job postings, pricing strategy inferred from contract terms in litigation documents, expansion strategy inferred from real estate and regulatory filings — is buried in sources that analysts do not have time to systematically monitor. The result is competitive intelligence that describes competitors as they were, not as they are, and never as they intend to be.
Clients and internal stakeholders respond to thin competitive intelligence by demanding more: more depth, more currency, more sources. This creates a feedback loop where analysts spend an ever-larger fraction of their time on competitive monitoring at the expense of analysis and synthesis. Senior analysts who should be identifying strategic implications spend their time verifying that competitor X did or did not change their pricing last quarter. Meanwhile, the strategic decisions that depend on accurate competitive intelligence — market entry, pricing strategy, product prioritization — proceed on an incomplete information base.
How COCO Solves It
Multi-Source Competitive Signal Collection: COCO systematically monitors and extracts intelligence from the full range of public sources:
- Scrapes and analyzes earnings call transcripts, investor day presentations, and SEC/regulatory filings for strategic signals
- Monitors job posting patterns across LinkedIn, Indeed, and company career pages to infer R&D priorities, expansion plans, and organizational changes
- Extracts product capability signals from app store listings, changelog announcements, API documentation, and developer forums
- Tracks pricing signals from review sites (G2, Gartner Peer Insights, Capterra), litigation documents, and partner channel disclosures
- Aggregates media coverage, analyst commentary, and conference presentation abstracts for positioning and narrative signals
Product and Capability Mapping: COCO builds structured product intelligence from unstructured sources:
- Creates a feature-by-feature comparison matrix across competitors, mapped to standardized capability categories
- Tracks product launch history and changelog patterns to infer release cadence and development velocity
- Identifies gaps in competitor product coverage that represent white space opportunities
- Maps technology stack indicators from job requirements, open-source contributions, and patent citations
- Generates a product maturity assessment using TRL-aligned criteria applied consistently across competitors
Pricing Architecture Reconstruction: COCO surfaces pricing intelligence from indirect sources:
- Aggregates user-reported pricing data from review platforms, Reddit, and industry forums with recency weighting
- Analyzes contract terms disclosed in litigation, regulatory filings, and public tender documents
- Identifies pricing model structure (per-seat, usage-based, tiered, enterprise) from feature gating signals
- Compares list pricing against actual transaction pricing signals to estimate discount patterns
- Tracks pricing changes over time and correlates with competitive events, product launches, or financial pressure signals
Leadership and Organizational Intelligence: COCO tracks people and organizational signals:
- Monitors executive appointments, departures, and LinkedIn profile updates for leadership change signals
- Analyzes organizational structure from LinkedIn employee counts by department, location, and seniority level
- Tracks advisor and board appointments for signals about strategic direction and investor relationships
- Identifies key technical talent movements — senior engineers, product leaders, researchers — as early signals of capability shifts
- Cross-references conference speaking schedules and publication records for thought leadership positioning signals
Strategic Signal Synthesis and Trend Identification: COCO connects disparate signals into coherent intelligence:
- Applies a structured analytical framework (SWOT, Porter's Five Forces, Jobs-to-be-Done) to organize raw intelligence
- Identifies patterns across signals — e.g., simultaneous enterprise hiring + pricing restructuring + partner program expansion suggesting an upmarket move
- Weights signals by recency, source credibility, and corroboration level to distinguish strong signals from noise
- Generates a "what they're likely doing next" hypothesis section with supporting evidence and confidence rating
- Tracks intelligence changes over time, highlighting what has changed since the previous report cycle
Client-Ready Competitive Report Generation: COCO produces structured, formatted deliverables:
- Generates a competitor profile document covering company overview, product, pricing, go-to-market, financials, and strategic outlook
- Produces an at-a-glance competitive comparison table for executive audiences
- Creates a signals register with each piece of intelligence linked to its source and dated
- Formats a "key takeaways and implications" section with recommended strategic responses
- Supports continuous monitoring mode with weekly or monthly digest updates highlighting new signals
Results & Who Benefits
Measurable Results
- Competitive profile production time: Reduced from 40–60 analyst hours to 6–10 hours per competitor (80–85% time savings)
- Source coverage: COCO monitors 50+ source types per competitor vs. 8–12 in typical manual research, capturing signals missed in standard processes
- Signal currency: Competitive profiles updated with weekly fresh signals vs. quarterly manual refresh cycles in most organizations
- Strategic surprise rate: Organizations using systematic competitive intelligence report 60% fewer competitive surprises (unexpected competitor moves that required reactive strategy changes)
- Analyst capacity reallocation: Teams reclaim an average of 15–20 analyst hours per week previously spent on competitive monitoring, redirected to higher-value synthesis and client work
Who Benefits
- Research Analysts: Shift from data collection to strategic interpretation — COCO handles the monitoring and aggregation so analysts can focus on what the intelligence means
- Strategy Teams: Access fresh, comprehensive competitive profiles on demand without waiting for an analyst to complete a manual research cycle
- Product Managers: Receive structured product intelligence with feature comparison matrices and roadmap signal analysis that directly informs product prioritization decisions
- Sales and Business Development: Access competitor positioning summaries and pricing intelligence formatted for use in client conversations and competitive deal situations
💡 Practical Prompts
Prompt 1 — Full Competitor Profile
Build a comprehensive competitive intelligence profile for the following company.
Target competitor: [company name]
Our company context: [brief description of who we are and why this competitor matters]
Priority intelligence areas: [product / pricing / go-to-market / financials / leadership / all]
Time horizon for signals: [last 6 months / 12 months / 24 months]
Geographic focus: [global / specific regions]
Deliverables:
1. Company overview: founding, ownership, funding, headcount, revenue estimates
2. Product and capability map: features, technology stack indicators, recent launches, roadmap signals
3. Pricing architecture: model structure, price points, discount patterns, recent changes
4. Go-to-market strategy: target segments, sales motion, channel partners, marketing positioning
5. Leadership and organizational profile: key executives, recent changes, organizational structure
6. Strategic signals: what public evidence suggests they are planning next
7. Competitive implications: where we are stronger, weaker, and most exposed
Format: Executive summary (1 page) + detailed section report + source citationsPrompt 2 — Competitor Pricing Intelligence
Reconstruct the pricing architecture for [competitor name] from available public sources.
Context: We sell [product/service] and need to understand competitor pricing to [inform our pricing strategy / prepare for competitive deals / assess market positioning].
Our current pricing: [brief description]
Geographic markets to cover: [list]
Sources to analyze: earnings transcripts, review sites (G2, Gartner Peer Insights, Capterra, Reddit), job postings, partner disclosures, litigation documents, public tender responses.
Deliverables:
1. Pricing model structure: per-seat / usage-based / tiered / enterprise — with evidence
2. Price point estimates by tier with source citations and confidence levels
3. Discount pattern analysis: typical enterprise discounts, volume thresholds, contract length incentives
4. Pricing changes over time: what changed, when, and what triggered the change
5. Comparison to our pricing: where we are priced above, below, or at parity
6. Implications: recommended adjustments to our pricing strategy or sales conversation framing
Confidence-weight each data point and flag where primary research would most improve accuracy.Prompt 3 — Competitive Product Feature Matrix
Build a competitive feature comparison matrix for [our product] vs. [list of competitors].
Product category: [description]
Customer segment: [target buyer profile]
Key capability dimensions to compare: [list or ask COCO to derive from the category]
For each competitor, extract capability signals from: product website, documentation, changelog, app store listings, review sites, job postings, conference demos, API references.
Deliverables:
1. Feature matrix: rows = capability categories, columns = our product + each competitor, cells = present/partial/absent + evidence note
2. Capability gap analysis: where competitors have features we lack, sorted by customer importance signals
3. Our differentiation summary: where we are ahead and the evidence supporting that advantage
4. Roadmap signals: features competitors appear to be building based on job postings and beta announcements
5. Customer perception overlay: how buyers on review sites rate each competitor on each dimension
Output as a structured table suitable for import into a slide deck or product planning document.Prompt 4 — Competitive Signal Weekly Digest
Generate a competitive intelligence digest for the following competitors covering the past [7 / 14 / 30] days.
Competitors to monitor: [list 3–8 company names]
Signal categories to cover: product launches, pricing changes, leadership moves, funding events, partnership announcements, regulatory filings, significant media coverage, job posting pattern shifts.
For each competitor, summarize:
1. New signals detected this period (bulleted, dated, with source)
2. Change from prior period: what is different vs. the previous digest
3. Significance rating: high / medium / low with one-sentence rationale
4. Recommended action: what our team should do in response, if anything
Format: One-page digest per competitor, plus a cross-competitor "top 3 signals this week" executive summary at the top.Prompt 5 — Competitor Strategic Intent Analysis
Analyze the strategic intent of [competitor name] based on their recent public actions and signals.
Context: [Describe the competitive situation — market, our position, why this competitor's strategy matters now]
Signal window: last [12 / 24] months
Sources to analyze: earnings transcripts, investor day presentations, CEO interviews, conference keynotes, press releases, job posting trends, acquisition activity, partnership announcements
Analytical framework to apply: [Jobs-to-be-Done / Ansoff Matrix / Porter's Generic Strategies / SWOT — or ask COCO to choose the most appropriate]
Deliverables:
1. Strategic intent hypothesis: what the evidence suggests they are trying to achieve in the next 12–24 months
2. Supporting evidence: each signal mapped to the strategic hypothesis with confidence rating
3. Alternative hypotheses: other plausible strategic interpretations of the same signals
4. Competitive implications: how their apparent strategy affects our market position and priorities
5. Early warning indicators: signals to monitor that would confirm or disconfirm the hypothesis
Write at the level of a senior strategy consultant briefing an executive team.3. AI Industry Trend Signal Aggregator & Synthesizer
Cuts through information overload to surface the 10 signals that actually matter this week — synthesized from 100+ industry sources into a structured trend briefing.
Pain Point & How COCO Solves It
The Pain: Trend Monitoring Is a Full-Time Job That Cannot Be Done Full-Time
Research analysts are expected to maintain current awareness across their coverage universe — tracking regulatory developments, technology shifts, market structure changes, and macroeconomic signals simultaneously — while also producing the deliverables that justify their billing rate. The volume of potentially relevant information has grown exponentially: a single industry vertical might generate hundreds of relevant articles, reports, filings, and announcements per week across trade publications, academic journals, regulatory portals, earnings releases, and news feeds. No analyst can read all of it, which means every analyst's trend awareness has gaps — and those gaps tend to surface at the worst possible moment, in client meetings or investment committee presentations.
The curation problem is as severe as the volume problem. Most trend monitoring tools surface information by recency and keyword match, not by strategic relevance to a specific research agenda. Analysts who set up Google Alerts or RSS feeds quickly find themselves overwhelmed with tangentially relevant content and missing genuinely important signals because those signals do not use the exact keywords in their filter. The result is a monitoring system that produces anxiety — a sense that something important is being missed — without producing actionable intelligence. Senior analysts compensate by relying on personal networks and curated newsletters, which introduces survivorship bias and geographic or ideological clustering.
The synthesis gap compounds the monitoring gap. Even analysts who successfully collect relevant signals rarely have time to synthesize them into a coherent trend narrative. Individual signals — a regulatory comment period opening, a technology company's product launch, a trade association white paper — only become strategic intelligence when connected to each other and to the analyst's existing knowledge base. Without synthesis, trend monitoring produces a pile of bookmarks rather than an understanding of where the industry is going. Clients and stakeholders want the narrative, not the raw signals, and building that narrative manually is precisely the work that gets squeezed when time is short.
How COCO Solves It
Intelligent Signal Collection Across Source Types: COCO monitors a broad, structured source universe with relevance filtering:
- Ingests content from trade publications, academic preprint servers, regulatory portals, earnings transcripts, patent databases, and curated news feeds simultaneously
- Applies semantic relevance scoring — not just keyword matching — to filter signals by strategic relevance to the defined research agenda
- Monitors social platforms (LinkedIn, X/Twitter, Reddit industry subreddits) for emerging discussion patterns before they reach mainstream coverage
- Tracks publication and report releases from think tanks, industry associations, government agencies, and research universities
- Maintains a dynamic source list that expands when new relevant sources are identified and contracts when sources become inactive or low-signal
Signal Categorization and Taxonomy: COCO organizes signals into a structured analytical framework:
- Classifies each signal by type: technology development, regulatory change, market structure shift, macroeconomic indicator, competitive move, demographic or behavioral change
- Tags signals by affected stakeholder group, geographic scope, time horizon (near-term catalyst vs. long-term structural), and impact magnitude
- Identifies when the same underlying trend is generating signals across multiple source types — a strong corroboration indicator
- Flags signals that contradict the prevailing industry narrative as potential inflection points deserving analytical attention
- Maintains a signal history log that enables trend velocity analysis — is a signal accelerating, plateauing, or fading?
Trend Synthesis and Narrative Construction: COCO connects signals into coherent trend narratives:
- Groups related signals into trend themes with a synthesized narrative explaining the underlying dynamic
- Distinguishes structural trends (multi-year trajectory changes) from cyclical patterns and one-time events
- Applies a horizon scanning framework (emerging / developing / mature / declining) to each identified trend
- Generates a "so what" implication section for each trend — what it means for clients, competitors, and strategy
- Cross-references new trend signals against existing trend narratives to identify acceleration, deceleration, or reversal
Weak Signal and Early Warning Detection: COCO identifies trends before they become consensus:
- Monitors low-volume but high-credibility sources — academic preprints, regulatory comment letters, niche trade forums — for signals that precede mainstream coverage
- Applies anomaly detection logic to flag unusual spikes in discussion volume, regulatory activity, or patent filing density
- Tracks second-order signals: changes in hiring patterns, supplier relationships, or capital flows that precede visible market changes
- Generates early warning flags when multiple weak signals in the same direction cross a corroboration threshold
- Produces a monthly "signals to watch" section highlighting trends that are not yet material but show early momentum
Comparative and Cross-Industry Analysis: COCO enriches trend analysis with broader context:
- Identifies analogous trends in adjacent industries and extracts lessons applicable to the current coverage sector
- Compares trend velocity and adoption curves against historical S-curve benchmarks for similar technologies or regulatory changes
- Cross-references trend signals against macroeconomic and demographic data to assess structural vs. cyclical drivers
- Identifies geographic variation in trend adoption — where a trend is ahead, lagging, or taking a different form
- Generates cross-industry analogy briefings: "this looks like X in industry Y — here is what happened next"
Structured Trend Briefing Generation: COCO produces client-ready trend intelligence deliverables:
- Generates a weekly or monthly trend briefing with executive summary, top 10 signals, trend narratives, and implications
- Produces a trend heat map showing signal intensity by topic area and time period
- Creates a trend tracker dashboard with trend names, maturity stage, signal count, velocity direction, and key supporting signals
- Formats deliverables for different audiences: one-page executive digest, detailed analyst briefing, or client-facing market update
- Supports custom briefing formats aligned with client or firm templates
Results & Who Benefits
Measurable Results
- Trend monitoring time: Reduced from 8–12 hours per week of manual source scanning to under 2 hours of review and synthesis (75–80% time savings)
- Source coverage: COCO monitors 100+ sources per industry vs. 15–20 in typical analyst workflows, reducing blind spot risk substantially
- Signal-to-noise ratio: Analysts report 85% reduction in irrelevant content in their trend feeds after COCO relevance filtering replaces keyword-based alerts
- Early signal detection: Organizations using COCO's weak signal monitoring identify emerging trends an average of 4–6 months earlier than peers relying on mainstream industry reports
- Briefing production speed: Weekly client-facing trend briefings produced in under 3 hours vs. full analyst-day previously
Who Benefits
- Research Analysts: Transform from information processors to strategic synthesizers — COCO handles the monitoring so analysts can focus on interpretation and implication
- Strategy Teams: Receive structured, timely trend intelligence that informs strategic planning processes without requiring dedicated competitive intelligence staff
- Investment Analysts: Access early signals on industry inflection points that affect portfolio company performance and investment thesis validity
- Client-Facing Professionals: Deliver trend briefings with confidence, knowing the monitoring behind them is systematic rather than opportunistic
💡 Practical Prompts
Prompt 1 — Weekly Industry Trend Briefing
Generate a weekly trend briefing for the following industry.
Industry: [sector name]
Sub-segments to cover: [list specific areas of focus]
Geographic scope: [global / regional]
Audience: [internal analysts / client-facing / executive team]
Time window: signals from the past [7 / 14 / 30] days
Structure the briefing as follows:
1. Top 3 signals this week: the most strategically significant developments, each with a 3-sentence summary and implication
2. Trend tracker update: for each of our [5/10] tracked trends, what new signals emerged and how does it change our view
3. Regulatory and policy watch: any new filings, comment periods, enforcement actions, or proposed rules
4. Technology and innovation signals: product launches, patent filings, R&D announcements, or academic publications
5. Capital flows: funding rounds, M&A activity, IPO filings, or notable investor positions
6. To watch next week: signals or developments that may become material in the coming 7–14 days
Output format: Executive digest (1 page) + full signal log with sourcesPrompt 2 — Emerging Trend Deep-Dive
Conduct a deep-dive analysis on the following emerging trend.
Trend: [trend name or description]
Industry context: [relevant sector and competitive landscape]
Current maturity assessment: [early signal / emerging / developing / mainstream — or ask COCO to assess]
Analyze:
1. Origin and drivers: what is causing this trend, and what evidence supports each driver
2. Adoption trajectory: where is adoption today, and what does the S-curve suggest about the next 2–3 years
3. Key enablers: technologies, regulations, behaviors, or economics that are accelerating the trend
4. Key barriers: factors slowing adoption with assessment of whether they are temporary or structural
5. Winner and loser implications: which player types benefit most or are most disrupted
6. Analogous historical trends: closest comparable from another industry with lessons and differences
7. Monitoring signals: what to watch that would indicate the trend is accelerating, plateauing, or reversing
Synthesize into a trend brief suitable for inclusion in a strategic planning document or client report.Prompt 3 — Cross-Industry Trend Signal Scan
Scan for trend signals relevant to [industry] that are currently emerging in adjacent industries.
Primary industry: [sector]
Adjacent industries to scan: [list 3–5 or ask COCO to identify the most relevant adjacencies]
Trend types of interest: [technology adoption / regulatory approach / business model innovation / consumer behavior / operational practice]
For each relevant cross-industry signal:
1. Signal description: what is happening in the adjacent industry
2. Relevance to our industry: why this matters and how it might transfer
3. Time lag estimate: how long before this signal reaches our industry based on analogous historical patterns
4. Action implication: what we or our clients should be doing now to prepare
5. Confidence level: how strong is the analogy and how well-established is the signal
Output format: Signal-by-signal brief with implication summary, plus a "top 3 cross-industry signals to watch" executive highlightPrompt 4 — Regulatory Trend Tracker
Build a regulatory trend tracker for [industry] covering [jurisdiction(s)].
Coverage scope:
- Active rulemakings and proposed regulations
- Recent enforcement actions and their implications
- Legislative activity (bills introduced, hearings scheduled, laws passed)
- Regulatory agency statements, speeches, and guidance documents
- International regulatory developments that may influence domestic policy
For each regulatory development:
1. Status: proposed / final / under review / enacted / in litigation
2. Affected parties: which companies, business models, or activities are in scope
3. Compliance timeline: key dates and deadlines
4. Business impact: what changes are required and what the cost or operational implications are
5. Strategic implication: does this create competitive advantage, disadvantage, or a level playing field
Produce a regulatory calendar with key upcoming dates and a priority-ranked list of developments by business impact.Prompt 5 — Trend Implication Analysis for Strategic Planning
Analyze the strategic implications of the following trends for [company type / our clients].
Trends to analyze:
1. [Trend 1 with brief description]
2. [Trend 2]
3. [Trend 3]
[Add up to 6 trends]
Company context: [describe the type of company and its current strategic position]
Time horizon: [12 months / 3 years / 5 years]
For each trend, assess:
1. Impact likelihood: how likely is this trend to materially affect the company within the time horizon
2. Impact magnitude: how significant is the effect if it materializes (revenue, cost, competitive position)
3. Strategic response options: 3 distinct options ranging from defensive to opportunistic
4. Decision timing: when must a decision be made to avoid foreclosing options
5. Early indicators: signals that would increase or decrease confidence in this trend's trajectory
Synthesize across all trends:
- Where trends reinforce each other (compound risk or compound opportunity)
- Where trends conflict (strategic trade-offs requiring prioritization)
- Top 3 recommended strategic priorities given the full trend picture4. AI Expert Interview Guide & Question Bank Generator
Generates structured, hypothesis-driven interview guides tailored to the specific expert profile and research objective — ready to use in 15 minutes.
Pain Point & How COCO Solves It
The Pain: Expert Interview Quality Determines Research Quality — and Preparation Is Chronically Rushed
Primary research through expert interviews is among the highest-value activities a research analyst performs. A single well-conducted interview with the right expert can resolve questions that weeks of desk research cannot answer — validating market size assumptions, surfacing competitive dynamics visible only from inside an industry, or identifying the non-obvious factor that changes an entire investment thesis. Yet the preparation that determines interview quality — developing a nuanced hypothesis structure, crafting questions calibrated to the expert's specific background, and anticipating the follow-up probes that surface real insight — is almost always done under time pressure, often in the hour before the call begins.
The quality gap in expert interview preparation creates cascading problems. Generic questions produce generic answers. An expert who has been asked the same question dozens of times by different research firms gives a rehearsed, surface-level response. The analyst who comes with precise, hypothesis-testing questions that demonstrate sector expertise gets the real answer — the one the expert has never been asked to articulate before. Developing those questions requires both deep sector knowledge and skill in primary research methodology: understanding which question types (open-ended, hypothetical, anchoring, contrast) work for which insights, and how to sequence questions to build rapport before asking about sensitive topics.
The expert matching problem compounds the preparation problem. Even when analysts have a clear research question, they often lack visibility into which type of expert would best answer it — and when an interview is arranged with an expert whose background does not precisely match the research gap, the entire interview time is partially wasted. There is no systematic process in most research organizations for mapping research questions to expert profiles, ensuring interview coverage across perspectives, or synthesizing insights from multiple interviews into a coherent research finding. Each interview is a discrete event rather than part of a structured research design.
How COCO Solves It
Expert Profile and Question Calibration: COCO tailors every question to the specific expert's background:
- Analyzes the expert's LinkedIn profile, published work, company history, and role description to understand their vantage point
- Identifies what the expert is uniquely positioned to know vs. what would be outside their direct experience
- Calibrates question depth and technical vocabulary to the expert's apparent expertise level
- Generates role-specific question variants — the same research question asked differently to a CEO, a mid-level operator, or a technical specialist
- Flags topics where the expert may have conflicts of interest or institutional biases that require probe questions to surface
Hypothesis-Driven Question Structure: COCO builds interviews around testable research hypotheses:
- Translates research objectives into specific hypotheses that the interview is designed to test or refute
- Structures questions in a logical sequence: establish credibility and context → test primary hypotheses → probe for nuance → surface contradicting evidence → extract forward-looking signals
- Generates hypothesis-testing questions that are specific enough to produce useful answers without leading the witness
- Includes devil's advocate questions designed to surface evidence against the working hypothesis
- Maps each question to the research hypothesis it tests, enabling post-interview analysis by hypothesis coverage
Question Type Diversity and Depth Probing: COCO applies primary research methodology rigor:
- Generates an opening question sequence that builds rapport while establishing the expert's credibility and framing
- Includes anchoring questions that establish a baseline (e.g., "In your experience, what does a typical deal look like?") before probing for exceptions
- Generates contrast questions ("How does X differ from Y?") that surface comparative insight without requiring the expert to volunteer the comparison
- Includes hypothetical questions calibrated to the expert's tolerance for speculation (higher for strategists, lower for operational practitioners)
- Creates follow-up probe variants for each primary question — what to ask when the answer is too brief, too generic, or surprisingly strong
Expert Type and Coverage Mapping: COCO ensures research design covers the right perspectives:
- Maps the research question to a taxonomy of expert types (operators, former executives, investors, regulators, academics, customers) with reasoning for each
- Identifies which research gaps each expert type is best positioned to fill
- Generates a recommended interview coverage plan: how many experts of each type and what question emphasis for each
- Creates a cross-interview synthesis template that enables comparison of responses across expert profiles
- Identifies where expert consensus and divergence are expected, and designs questions to surface both
Sensitive Topic Navigation: COCO generates approaches for extracting information on difficult topics:
- Identifies topics where experts are likely to be guarded (competitive information, regulatory issues, M&A activity) and generates indirect question approaches
- Creates hypothetical framings that allow experts to discuss sensitive topics without direct disclosure
- Generates third-party referencing questions ("I've heard from others in the industry that X — does that match your experience?") to lower resistance
- Develops rapport-building question sequences that establish trust before approaching sensitive areas
- Includes appropriate legal compliance reminders for interviews where MNPI risk is present
Post-Interview Synthesis Framework: COCO structures insight extraction from completed interviews:
- Generates a structured note-taking template aligned with the question guide for consistent capture across interviews
- Creates a cross-interview insight aggregation framework for comparing responses by expert type and hypothesis
- Produces a synthesis prompt that converts raw interview notes into structured research findings
- Generates a gap analysis identifying which research hypotheses remain unresolved after each interview
- Creates a follow-up question list for subsequent interviews based on insights from completed ones
Results & Who Benefits
Measurable Results
- Interview preparation time: Reduced from 3–5 hours to 20–40 minutes for a tailored expert interview guide (85–90% time savings)
- Interview insight density: Analysts report 2–3x more novel insights per interview when using hypothesis-driven vs. generic question guides
- Expert coverage efficiency: Structured coverage mapping reduces average interviews needed to reach research saturation from 12–15 to 7–9 per project
- Research hypothesis resolution rate: Hypothesis-structured interviews resolve target hypotheses in 78% of cases vs. 45% with unstructured interview approaches
- Interview-to-deliverable cycle time: Structured synthesis frameworks reduce time from interview completion to incorporated research finding from 3–5 days to under 1 day
Who Benefits
- Research Analysts: Arrive at every expert interview with a professionally structured guide tailored to that specific expert — eliminating last-minute scrambling and generic question lists
- Strategy Consultants: Conduct primary research interviews that produce genuine insight rather than confirmation of existing hypotheses, strengthening deliverable quality
- Investment Analysts: Execute expert network calls efficiently — extracting maximum insight within the 30–45 minute window typical for expert network calls
- Market Research Professionals: Apply consistent primary research methodology across projects and team members, reducing quality variance driven by individual preparation skill
💡 Practical Prompts
Prompt 1 — Expert Interview Guide Generation
Generate a structured expert interview guide for the following research engagement.
Research objective: [what you are trying to learn or validate]
Working hypothesis: [your current best guess about the answer, which the interview should test]
Expert profile: [role, company type, years of experience, specific background relevant to the topic]
Interview duration: [30 / 45 / 60 minutes]
Research stage: [early exploratory / hypothesis testing / validation / triangulation]
Structure the guide as:
1. Opening (5 min): rapport and context-setting questions
2. Expert background confirmation (5 min): verify relevant experience and frame of reference
3. Core hypothesis testing questions (20–30 min): 5–8 primary questions with 2–3 probe variants each
4. Devil's advocate and contrarian questions (5–10 min): questions designed to surface evidence against the working hypothesis
5. Forward-looking signals (5 min): where is this heading, what would change your view
6. Closing (2 min): who else should we speak to, what questions did we not ask
For each question, note: what hypothesis it tests and what a strong vs. weak answer looks like.Prompt 2 — Question Bank for Expert Type
Build a question bank for expert interviews on the following topic, optimized for the specified expert type.
Research topic: [subject area]
Expert type: [former C-suite executive / current operator / industry consultant / regulator / investor / end customer / technical specialist]
Research questions to address: [list 3–5 specific information needs]
Industry context: [relevant sector and competitive dynamics]
Generate:
1. 15 primary questions organized by research theme, calibrated to this expert type's vantage point
2. 5 probe variants for each theme (follow-up questions when the primary answer is insufficient)
3. 3 anchoring questions to establish baseline before probing for exceptions or nuance
4. 3 hypothetical questions appropriate for this expert type
5. 2 sensitive topic approaches using indirect question techniques
Flag which questions are highest priority if interview time is shorter than expected.Prompt 3 — Cross-Interview Synthesis Template
Create a cross-interview synthesis framework for a primary research program with the following parameters.
Research program: [brief description]
Number of interviews planned: [N]
Expert types to be interviewed: [list types and how many of each]
Core research hypotheses:
1. [Hypothesis 1]
2. [Hypothesis 2]
3. [Hypothesis 3]
[Add up to 5 hypotheses]
Generate:
1. A structured note-taking template for each interview aligned to hypothesis coverage
2. A cross-interview comparison matrix: rows = key questions, columns = expert types, cells = response summary
3. Consensus and divergence tracker: where experts agree, disagree, and where divergence by expert type is meaningful
4. Hypothesis status tracker: supported / refuted / mixed evidence / insufficient data for each hypothesis after each interview
5. Gap analysis template: which hypotheses remain unresolved and what expert type is best positioned to resolve them
Format as a working document that can be updated after each interview.Prompt 4 — Sensitive Topic Interview Approach
Generate interview questions and approach guidance for the following sensitive research topic.
Topic: [describe the sensitive subject — e.g., pricing practices, regulatory violations, competitive intelligence, M&A activity]
Expert profile: [role and likely institutional position on this topic]
Why this topic is sensitive: [what makes experts reluctant to discuss it directly]
Research objective: [what specifically we need to learn about this topic]
Generate:
1. Indirect question approaches: ways to learn about the topic without asking directly
2. Third-party referencing questions: frames that allow experts to comment on others' behavior rather than their own
3. Hypothetical framings: scenario-based questions that surface real knowledge without direct disclosure
4. Rapport sequencing: which questions to ask first to build trust before approaching the sensitive topic
5. Legal and ethical guardrails: reminders about MNPI risk, competitive information handling, and expert network compliance
Include sample question language for each approach, calibrated to sound natural in conversation.Prompt 5 — Interview Debrief and Insight Extraction
Extract structured research insights from the following interview notes.
Research hypotheses being tested:
1. [Hypothesis 1]
2. [Hypothesis 2]
3. [Hypothesis 3]
Expert profile: [role, company type, experience level]
Interview notes: [paste raw interview notes or transcript]
Extract and structure:
1. Hypothesis-by-hypothesis assessment: what did this expert say that supports, contradicts, or is neutral on each hypothesis
2. Novel insights: information or perspectives that were not anticipated in the research design
3. Key quotes: verbatim statements that best capture the expert's view on each core topic
4. Confidence assessment: how credible and well-informed did this expert appear on each topic
5. Follow-up implications: what new questions does this interview raise for subsequent interviews
6. Integration note: how does this interview change or reinforce the overall research picture so far
Flag any statements that require verification from other sources before being incorporated into findings.5. AI Primary Research Survey Design Engine
Designs statistically sound, bias-minimized survey instruments from research objectives to final questionnaire — in a fraction of the time of manual design.
Pain Point & How COCO Solves It
The Pain: Survey Design Errors Are Invisible Until the Data Is Collected — and Then It Is Too Late
Survey research is the backbone of quantitative primary research in market analysis, consumer insights, and policy evaluation. But survey design is a discipline where errors are catastrophic and invisible: a leading question, an ambiguous scale, or a poorly structured skip logic cannot be corrected after the survey goes into the field. The data collected from a flawed instrument is either useless or actively misleading — and the analyst often does not discover the problem until they attempt analysis and find the responses do not support meaningful conclusions. By that time, the fielding budget has been spent and the project timeline has no room for a re-run.
Most research analysts who design surveys are not trained psychometricians or survey methodologists. They know their research questions and their subject matter, but they may not recognize that a double-barreled question confounds two variables, that a 1–10 scale without labeled midpoints produces inconsistent interpretation across respondents, or that acquiescence bias will inflate agreement with any positively framed statement. These errors are systematic and predictable, but catching them requires either expensive expert review or painful experience. Junior analysts routinely deploy instruments with multiple methodological flaws because they lack the training to recognize them, and senior analysts rarely have time to review survey instruments with the rigor the methodology requires.
The downstream impact of poor survey design extends well beyond a single project. Survey data that cannot be cleanly analyzed does not simply fail to answer the research question — it actively consumes time in attempts to salvage something from the flawed data, and may produce findings that reach publication or client delivery before the methodological flaws are identified. Organizations that repeatedly field methodologically weak surveys also build a data library they cannot trust for longitudinal comparison, because inconsistent question wording and scale construction prevent meaningful trend analysis across waves. The cost of survey design errors compounds with every wave fielded.
How COCO Solves It
Research Objective to Question Translation: COCO converts research goals into valid survey questions:
- Translates each research objective into a specific, answerable survey question mapped to an appropriate response format
- Identifies which research objectives require attitudinal measures (Likert scales), behavioral measures (frequency, recency), or factual measures (demographics, usage)
- Flags research objectives that are too abstract for survey measurement and recommends operational definitions
- Generates multiple question variants for each objective and recommends the version least susceptible to bias
- Ensures each question connects to a specific analysis plan — no question is included that does not address a defined research objective
Bias Detection and Question Quality Review: COCO identifies and corrects methodological flaws:
- Flags leading questions, loaded language, double-barreled constructions, and presupposition errors with specific correction suggestions
- Identifies acquiescence bias risk in agreement-format questions and recommends forced-choice or balanced alternatives
- Detects social desirability bias risk on sensitive questions and recommends indirect or normalized question framings
- Reviews scale construction for labeled midpoints, balanced positive/negative options, and appropriate response range
- Checks for logical consistency across question sequence — inconsistencies that would confuse or irritate respondents
Questionnaire Architecture and Flow Design: COCO builds the full survey structure:
- Sequences questions from general to specific, with sensitive or effort-intensive questions placed after rapport is established
- Designs screening questions with branching logic to qualify respondents before entering the main instrument
- Builds skip logic and conditional display rules that route respondents through the relevant question subset based on prior answers
- Groups questions thematically while avoiding context effects that lead respondents to interpret later questions through the frame of earlier ones
- Estimates completion time per section and flags when total instrument length exceeds respondent tolerance thresholds
Scale Selection and Psychometric Rigor: COCO applies measurement best practices:
- Recommends appropriate scale types for each construct — Likert, semantic differential, ranking, constant sum, MaxDiff, NPS
- Applies established scale anchors for common constructs (satisfaction, importance, agreement, likelihood) with validated wording
- Designs multi-item scales for complex constructs where a single question would have poor construct validity
- Flags when a construct requires validated psychometric instruments (e.g., customer effort score, brand equity scales) and recommends established versions
- Generates reliability check pairs — alternate-wording versions of key questions to assess response consistency
Sampling and Statistical Power Planning: COCO connects survey design to analysis requirements:
- Calculates required sample size for each analytical objective given assumed effect size, desired confidence interval, and power threshold
- Recommends stratification variables to ensure adequate representation of key subgroups for cross-tab analysis
- Flags when planned sample sizes are insufficient for the subgroup analyses specified in the research design
- Designs weighting variables to correct for known sampling frame biases
- Generates a quota plan for each target subgroup with recommended fielding approach for each segment
Survey Instrument Documentation: COCO produces complete fielding-ready deliverables:
- Generates the complete questionnaire in platform-agnostic format with all question text, response options, and skip logic documented
- Produces a codebook mapping each question to its variable name, response codes, and intended analysis
- Creates a fielding specification document for survey platform configuration (Qualtrics, SurveyMonkey, Typeform)
- Generates an interviewer guide for CAPI/CATI versions with probing instructions for open-ended questions
- Produces a pre-launch pilot test checklist with specific items to validate before full deployment
Results & Who Benefits
Measurable Results
- Survey design cycle time: Reduced from 3–5 days to 4–8 hours for a full questionnaire design (80% time savings)
- Methodological error rate: Surveys reviewed by COCO show 75% fewer bias and construction errors than unreviewed instruments in A/B testing
- Data usability rate: Proportion of survey data that supports clean analysis without salvage work increases from 62% to 91% with COCO-designed instruments
- Re-field rate: Proportion of surveys requiring re-fielding due to design errors drops from 18% to under 3% of projects
- Analysis cycle time: Post-fielding analysis begins within 24 hours of data receipt for COCO-designed surveys vs. 3–5 days for surveys requiring data cleaning and recoding
Who Benefits
- Research Analysts: Design methodologically sound instruments without needing a survey methodology specialist on the team — and catch errors before fielding rather than after
- Market Research Teams: Standardize survey design quality across team members and projects, reducing variance driven by individual analyst skill and experience
- Clients and Stakeholders: Receive primary research data they can trust — fielded with a documented methodology that supports the conclusions drawn from it
- Data Analysts: Work with clean, well-coded datasets that support the planned analysis without requiring extensive recoding or data quality remediation
💡 Practical Prompts
Prompt 1 — Full Survey Instrument Design
Design a survey instrument for the following research program.
Research objectives:
1. [Objective 1 — e.g., measure customer satisfaction with X]
2. [Objective 2 — e.g., understand drivers of purchase decision]
3. [Objective 3 — e.g., assess awareness of Y]
Target respondent profile: [describe who should complete this survey]
Sample size target: [N or "recommend based on objectives"]
Survey mode: [online self-completion / phone / in-person / mixed]
Estimated completion time target: [5 / 10 / 15 / 20 minutes]
Sensitive topics (if any): [list topics requiring careful question framing]
Deliverables:
1. Complete questionnaire with all question text, response options, and skip logic
2. Question-to-objective mapping: which question addresses which research objective
3. Methodology notes: rationale for scale choices, question sequencing, and key design decisions
4. Bias review: any remaining design risks and recommended mitigations
5. Fielding specification: platform configuration notes for [Qualtrics / SurveyMonkey / specify]
6. Estimated completion time per sectionPrompt 2 — Survey Question Quality Review
Review the following survey questions for methodological quality and suggest improvements.
Survey topic: [brief description]
Target respondents: [describe]
Questions to review:
[Paste 5–20 survey questions with response options]
For each question, assess:
1. Leading language: does the question suggest a preferred answer
2. Double-barreled construction: does the question ask about two things simultaneously
3. Ambiguity: could different respondents interpret the question differently
4. Scale quality: are response options balanced, exhaustive, mutually exclusive, and appropriately labeled
5. Social desirability risk: are respondents likely to answer based on what seems socially acceptable rather than truth
6. Acquiescence bias risk: will most respondents agree regardless of their actual view
For each flagged issue, provide: problem description + revised question language + rationale for the revision.Prompt 3 — Sample Size and Statistical Power Planning
Calculate the required sample size for the following survey research design.
Research objectives requiring statistical inference:
1. [Objective 1]: estimate [proportion / mean] with [±X%] margin of error at [90/95/99%] confidence
2. [Objective 2]: detect a difference of [X] between [Group A] and [Group B] at [power level]
3. [Subgroup objective]: achieve [N] completed interviews in [subgroup] for cross-tab analysis
Target population size: [total universe size, or "large population"]
Expected response rate: [% or "estimate"]
Subgroups requiring separate analysis: [list with estimated prevalence in population]
Deliverables:
1. Required sample size for each objective
2. Governing sample size (the largest requirement, which sets the overall target)
3. Stratification plan: how to allocate the sample across subgroups
4. Quota table: target completes per subgroup with recommended over-sampling if needed
5. Impact of non-response: how to weight data if certain groups are underrepresented in final samplePrompt 4 — Likert Scale and Psychometric Design
Design the measurement scales for the following constructs in a survey instrument.
Constructs to measure:
1. [Construct 1 — e.g., brand loyalty]: [definition of what we want to measure]
2. [Construct 2 — e.g., product satisfaction]: [definition]
3. [Construct 3 — e.g., purchase intent]: [definition]
For each construct:
- Recommend scale type (Likert / semantic differential / ranking / NPS / custom)
- Provide validated question wording if an established scale exists
- Design a multi-item scale if a single item would have poor reliability
- Specify scale points, labels, and format
- Recommend reliability check (test-retest pair or reverse-coded item)
- Flag any known response bias risks and recommended mitigations
If established psychometric instruments exist for any construct (e.g., Net Promoter Score, Customer Effort Score, brand equity scales), recommend the validated version with citation.Prompt 5 — Skip Logic and Survey Flow Design
Design the skip logic and survey flow for the following questionnaire structure.
Survey topic: [description]
Target segments with different question paths:
- Segment A: [define by screening criteria] → should see questions [list]
- Segment B: [define by screening criteria] → should see questions [list]
- Segment C: [define by screening criteria] → should see questions [list]
Screening questions to determine path: [describe or provide question text]
Terminate criteria: [which respondents should be screened out and at what point]
Optional question groups: [questions shown only if prior answer meets a condition]
Deliverables:
1. Survey flow diagram: visual representation of all paths through the instrument
2. Skip logic specification table: for each question, what prior answer triggers a skip and to where
3. Termination rules: exact conditions for screen-out with language for the termination screen
4. Branching question text variants: where question wording changes based on prior answer (e.g., "you mentioned X — how satisfied are you with X?")
5. Platform-specific configuration notes for [Qualtrics / SurveyMonkey / Typeform]6. AI Secondary Research Source Discovery & Ranking Tool
Surfaces the 20 most relevant, credible, and current sources for any research question — ranked by quality, not just recency or keyword match.
Pain Point & How COCO Solves It
The Pain: Source Discovery Is Simultaneously Time-Consuming and Systematically Biased
Every research project begins with source discovery — identifying which databases, reports, datasets, and publications contain the information needed to answer the research question. In practice, this process is deeply inefficient and systematically biased toward familiar sources. Analysts default to databases they know: Bloomberg for financial data, Statista for market statistics, Google Scholar for academic literature. These defaults are reasonable but incomplete — they miss specialized industry databases, government statistical agencies, regulatory filings, trade association publications, and niche academic journals that often contain better data for specific research questions. The result is a source set that is convenient rather than optimal.
The research quality implications are significant. A market sizing built from one industry analyst report misses the government census data that would provide a more rigorous baseline. A competitive analysis built from news articles misses the regulatory filings that would reveal the competitor's actual financial position. An investment thesis built from sell-side equity research misses the academic literature that would show the thesis has been tested and rejected in analogous historical contexts. These gaps are invisible to the analyst who does not know the source exists — and to the client who trusts the analyst's research as comprehensive.
The credibility ranking problem compounds the discovery problem. Not all sources are equal, but source quality is not displayed in search results. A top Google search result for an industry statistic is as likely to be a poorly sourced blog post as a rigorous government dataset. Analysts who lack deep source quality knowledge — typically junior and mid-level analysts — may cite sources that senior reviewers or informed clients can immediately identify as low-credibility, damaging the researcher's professional reputation. Building reliable source quality intuition requires years of domain experience, but research organizations cannot wait years for each analyst to develop it.
How COCO Solves It
Comprehensive Source Universe Mapping: COCO identifies relevant sources across the full research landscape:
- Maps relevant sources across categories: government statistical agencies, regulatory filings, industry associations, academic journals, commercial databases, news archives, think tanks, and NGOs
- Identifies domain-specific databases and repositories that are not commonly known but highly relevant to the specific research question
- Surfaces international equivalents for commonly used US/UK sources — international statistical agencies, foreign regulatory databases, regional trade associations
- Identifies open-access alternatives to paywalled sources, including preprint servers and government open data portals
- Flags sources that are commonly cited but methodologically weak for the specific use case
Source Quality and Credibility Assessment: COCO evaluates sources on multiple quality dimensions:
- Assesses methodology transparency: does the source explain how data was collected, sample sizes, coverage, and limitations
- Evaluates institutional authority: government statistical agencies, peer-reviewed journals, and established research organizations rated higher than commercial publishers and blogs
- Checks publication recency and update frequency: flags sources where data may be too old for the research question
- Identifies potential conflicts of interest: industry-funded research, advocacy organization publications, or commissioned reports
- Assigns a confidence tier (primary/secondary/tertiary) to each source based on data proximity to the original measurement
Research Question to Source Type Matching: COCO maps information needs to optimal source types:
- Analyzes the research question to identify what type of data is needed (market size, consumer behavior, regulatory framework, financial performance, technology landscape)
- Matches each data type to the source category most likely to contain rigorous, well-sampled measurements
- Identifies when multiple source types are needed for triangulation and recommends a source portfolio approach
- Flags when the required data does not exist in secondary sources and primary research is needed
- Suggests proxy measures and indirect data sources when direct measures are unavailable
Citation Chain and Reference Discovery: COCO follows research trails to find foundational sources:
- Traces citation chains from commonly cited statistics back to their original source — often revealing that the commonly cited figure is a distorted or outdated version
- Identifies seminal academic papers and reports that are the foundation of an entire literature
- Surfaces recent papers that have challenged or updated foundational findings
- Identifies gray literature — government reports, working papers, conference proceedings — that contains relevant data not indexed in standard academic databases
- Cross-references reference lists from high-quality sources to identify additional relevant sources not captured in initial discovery
Source Access and Retrieval Guidance: COCO helps analysts access identified sources:
- Identifies which sources are freely available and provides direct access links
- Flags which sources require subscriptions and identifies institutional access options (university libraries, public library databases, open-access repositories)
- Generates FOIA request templates for government data not publicly available
- Identifies when a source is most recently available in a specific format (PDF report, dataset download, API)
- Provides tips for navigating complex source interfaces — government database query syntax, academic database filter settings
Source Portfolio and Citation Management: COCO organizes discovered sources for efficient use:
- Produces a ranked source list with quality tier, relevance score, access method, and key data contained
- Generates a source summary matrix mapping each source to the research questions it addresses
- Creates properly formatted citations in the required style (APA, Chicago, MLA, custom)
- Builds a reference management export compatible with Zotero, EndNote, and Mendeley
- Flags duplicate coverage — when multiple sources contain the same underlying data — to avoid over-counting
Results & Who Benefits
Measurable Results
- Source discovery time: Reduced from 4–8 hours of database searching to under 1 hour for a comprehensive source landscape (85% time savings)
- Source coverage completeness: COCO surfaces an average of 3–4 highly relevant sources not typically found in analyst default source sets for each research question
- Citation error rate: Source quality reviews reduce erroneous citations (broken links, low-credibility sources, distorted statistics) by 80% compared to unreviewed research
- Source diversity: Research projects using COCO include 2.5x more source type categories than manually sourced projects, reducing single-database dependency
- Junior analyst source quality: Junior analysts using COCO cite sources rated "high credibility" at the same rate as senior analysts — eliminating the experience-based quality gap
Who Benefits
- Research Analysts: Build comprehensive, high-quality source sets in a fraction of the time — without needing years of domain experience to know where the best data lives
- Junior and Mid-Level Analysts: Access source quality guidance that would otherwise require years of experience to develop, producing senior-quality research from day one
- Academic and Policy Researchers: Discover the full relevant literature, including gray literature and international sources typically missed in standard database searches
- Consulting Firms: Standardize source quality across all client deliverables, reducing the risk of credibility-damaging citations and ensuring consistent methodology
💡 Practical Prompts
Prompt 1 — Source Discovery for Research Question
Identify the best sources for the following research question.
Research question: [specific question you need to answer]
Industry/domain: [sector or topic area]
Geography: [geographic scope of data needed]
Data type needed: [market size / consumer behavior / financial data / regulatory framework / technology landscape / academic evidence]
Recency requirement: [data must be from last X years]
Access constraints: [free sources only / we have access to Bloomberg, Statista, etc.]
Deliverables:
1. Top 10 recommended sources ranked by relevance and credibility, with:
- Source name, publisher, publication type
- Relevance: what specific data this source contains for the research question
- Credibility tier: primary data / secondary synthesis / tertiary commentary
- Access method: free / subscription / request
- Key limitation: what this source does not cover or where its methodology is weakest
2. Alternative sources (5–10) for triangulation
3. Sources to avoid: commonly cited sources that are low-credibility for this specific question
4. Primary research recommendation: if secondary sources are insufficient, what primary research would fill the gapPrompt 2 — Source Quality Assessment
Assess the quality and credibility of the following sources for use in a research report.
Research context: [describe the report and its audience — e.g., investment committee presentation, client strategy report]
Sources to assess:
1. [Source 1: title, author, publisher, date, URL if available]
2. [Source 2]
3. [Source 3]
[List up to 15 sources]
For each source, evaluate:
1. Methodology transparency: does it explain data collection, sample size, coverage, and limitations
2. Institutional credibility: who produced it and what are their credentials and potential biases
3. Data recency: how old is the data and whether it reflects current conditions
4. Citation chain: is this an original source or a citation of a citation (and how many steps removed)
5. Potential conflicts of interest: was this funded by parties with a stake in the findings
6. Overall recommendation: cite as primary evidence / cite with caveat / cite for context only / do not cite
Produce a source quality scorecard with an overall credibility rating for the source set.Prompt 3 — Citation Chain Tracing
Trace the following commonly cited statistic back to its original source.
Commonly cited claim: "[exact wording of the statistic or claim]"
Typically attributed to: [organization or publication where you've seen it cited]
Context: [industry or topic area]
Investigate:
1. What is the earliest publication this statistic appears in
2. What was the original methodology — how was this number calculated or estimated
3. Has the methodology or figure been updated, revised, or challenged since original publication
4. What important caveats or limitations were present in the original source but dropped in subsequent citations
5. Is the commonly cited version accurate, or has it been distorted through repeated citation
6. What is the best current version of this data point, from the most rigorous and recent source available
Recommend: should this statistic be cited, cited with caveats, replaced with a better source, or retired as unreliable.Prompt 4 — Industry Database and Report Landscape
Map the full landscape of data sources and research databases relevant to the following industry.
Industry: [sector name]
Research needs: [list the types of analysis typically conducted in this industry — market sizing, competitive analysis, regulatory tracking, financial modeling, etc.]
For each research need, identify:
1. Primary government and official statistical sources (free)
2. Leading commercial research providers and their flagship reports
3. Key industry association publications and their data coverage
4. Academic journals and databases most relevant to this industry
5. Regulatory filing databases and what data they contain
6. Specialized or niche databases not commonly known but highly relevant
Produce a source landscape map organized by research need, with notes on access, cost, and typical use case for each source.Prompt 5 — Comparative Source Evaluation for Specific Data Point
Compare and evaluate multiple sources for the following specific data point.
Data point needed: [specific statistic, market size, or measure]
Geography: [scope]
Time period: [relevant year or range]
Sources to compare:
1. [Source A]
2. [Source B]
3. [Source C]
[List all available sources for this data point]
For each source:
1. What figure does it report for this data point
2. What methodology was used to generate this figure
3. What year is the underlying data from
4. What is the geographic and definitional scope
5. Why does the figure differ from other sources (if it does)
Synthesize:
- Which source is most appropriate for the intended use case and why
- How to reconcile conflicting figures for a client-facing report
- What the range of estimates tells us about data reliability
- How to footnote the chosen figure with appropriate caveats7. AI Survey Response Statistical Analysis Engine
Transforms raw survey data into publication-quality statistical findings — cross-tabs, significance tests, regression, and visualization — without requiring a statistician.
Pain Point & How COCO Solves It
The Pain: Survey Data Analysis Is a Bottleneck Between Data Collection and Insight
The gap between survey data collection and actionable insight is where primary research most often breaks down. Fielding a survey is increasingly accessible — platforms like Qualtrics and SurveyMonkey make data collection straightforward — but transforming raw response data into statistically valid, client-ready findings is a specialized skill that most research analysts do not fully possess. Cross-tabulation is within reach for most analysts, but determining which differences between subgroups are statistically significant vs. sampling noise, running regression analysis to identify drivers of key outcomes, and applying appropriate corrections for multiple comparisons require quantitative training that research team members often lack. The result is analysis that describes the data without fully extracting its meaning.
The analytical decisions made during survey analysis have large effects on conclusions. Whether to weight the sample to correct for demographic imbalances, which statistical test to apply to which data type, how to handle missing responses, and how to interpret interaction effects between variables are not obvious choices — and the wrong choice can produce findings that appear rigorous but are methodologically indefensible. Analysts under time pressure default to the simplest analysis: frequencies and means for the full sample, basic cross-tabs by demographics, and a bar chart for each question. This surface-level analysis misses the insight that lives in driver analysis, cluster analysis, and multivariate models — the analysis that would actually change a strategic recommendation.
The visualization and reporting challenge adds a third layer of difficulty. Statistical findings that are technically correct but poorly visualized fail to persuade audiences who are not themselves statisticians. Choosing the right chart type for each finding, applying consistent formatting, and translating p-values and confidence intervals into language that business audiences understand are skills that require both analytical and communication ability simultaneously. Most research analysts are stronger in one than the other — and under deadline pressure, the weaker skill suffers.
How COCO Solves It
Data Quality and Preparation Assessment: COCO audits the dataset before analysis:
- Checks completion rates by question and identifies systematic item non-response patterns
- Detects straightlining, speeder responses, and other data quality indicators in self-completion surveys
- Identifies and recommends handling for outliers, implausible values, and inconsistent response patterns
- Assesses whether the collected sample matches the intended target population and recommends weighting approaches
- Generates a data quality report with decision recommendations for each identified issue before analysis begins
Descriptive Statistics and Cross-Tabulation: COCO produces the foundational analysis layer:
- Generates frequency distributions and descriptive statistics for every survey question
- Produces cross-tabulation tables for all specified variable pairs with appropriate cell formatting
- Applies column percentage, row percentage, and total percentage views for each cross-tab
- Calculates mean scores for ordinal and interval scales with standard deviation and confidence intervals
- Generates response distribution visualizations (bar charts, stacked bars, pie charts) calibrated to the data type and audience
Statistical Significance Testing: COCO applies the correct test for each comparison:
- Selects appropriate significance tests based on data type: chi-square for categorical variables, t-tests and ANOVA for continuous comparisons, Mann-Whitney for non-parametric distributions
- Applies Bonferroni or Benjamini-Hochberg corrections for multiple comparison situations
- Labels statistically significant differences in cross-tab tables with standard notation (*, **, ***)
- Interprets significance levels in plain language: "this difference is unlikely to be due to chance" vs. technical p-value language
- Generates effect size measures alongside significance tests: Cohen's d, Cramér's V, and eta-squared as appropriate
Driver Analysis and Regression Modeling: COCO identifies what predicts key outcomes:
- Runs multiple regression (linear or logistic) to identify drivers of key dependent variables (satisfaction, purchase intent, NPS)
- Presents regression results as importance-performance frameworks or key driver charts rather than coefficient tables
- Applies MaxDiff or conjoint analysis for preference and attribute importance research
- Runs cluster analysis to identify meaningful respondent segments based on response patterns
- Generates a factor analysis for multi-item scales to assess construct dimensionality and reliability (Cronbach's alpha)
Subgroup Analysis and Segmentation: COCO produces insight by audience segment:
- Runs all key analyses disaggregated by specified subgroups (demographics, behavioral segments, geographic splits)
- Identifies which subgroup differences are statistically significant and practically meaningful
- Generates a subgroup comparison framework highlighting where views converge and diverge most sharply
- Flags subgroups with insufficient sample sizes for reliable analysis
- Produces a customer segmentation if the research design supports it, with segment profiles and sizing
Findings Report and Visualization Generation: COCO produces client-ready deliverables:
- Generates a structured findings report with key insights, supporting statistics, and visualization for each research objective
- Translates statistical findings into plain-language narrative suitable for executive audiences
- Creates a slide-ready visualization suite: one chart per key finding, formatted for presentation
- Produces a full data appendix with all cross-tabs, significance notations, and base sizes
- Generates an executive summary: top 5 findings, their strategic implications, and recommended actions
Results & Who Benefits
Measurable Results
- Analysis cycle time: Reduced from 5–10 analyst days to 1–2 days for comprehensive survey analysis (75–80% time savings)
- Statistical rigor: Proportion of survey reports including significance testing increases from 35% to 95% with COCO — with appropriate test selection in each case
- Driver analysis adoption: Adoption of regression-based driver analysis (vs. simple frequency reporting) increases from 20% to 80% of eligible projects
- Report revision cycles: Client-requested revisions to statistical findings drop by 65% when COCO-generated methodology documentation accompanies the analysis
- Junior analyst quality: Junior analysts produce analysis rated "senior quality" by peer review in 82% of cases when using COCO vs. 31% without
Who Benefits
- Research Analysts: Produce statistically rigorous analysis without a quantitative background — COCO handles test selection, interpretation, and visualization so analysts can focus on strategic implication
- Market Research Teams: Deliver consistent, defensible analysis across all projects and team members — eliminating the quality variance driven by individual statistical skill
- Clients and Stakeholders: Receive findings they can act on with confidence — backed by appropriate significance testing and clearly communicated uncertainty
- Strategy and Product Teams: Access driver analysis and segmentation that identifies what to do, not just what respondents said — shifting survey research from descriptive to decision-enabling
💡 Practical Prompts
Prompt 1 — Full Survey Analysis Plan
Design a comprehensive analysis plan for the following survey dataset.
Survey overview: [brief description of survey topic and objectives]
Sample: [N total, key subgroups and their sizes]
Key research questions:
1. [Question 1]
2. [Question 2]
3. [Question 3]
Key variables:
- Primary outcome variables: [list the main dependent variables — e.g., NPS, satisfaction rating, purchase intent]
- Key independent variables for cross-tabs: [demographics, behavioral segments, usage groups]
- Control variables: [any variables to hold constant in multivariate analysis]
Deliverables requested:
1. Descriptive statistics plan: which variables get frequency tables vs. mean analysis
2. Cross-tabulation plan: all specified variable pairs with significance testing approach
3. Driver analysis plan: which regression model is appropriate and why
4. Segmentation analysis plan: if applicable
5. Visualization plan: recommended chart type for each key finding
6. Sample size check: flag any analyses where cell sizes are too small for reliable inferencePrompt 2 — Statistical Significance Interpretation
Interpret the following statistical results for a non-technical business audience.
Analysis context: [what the survey measured and who the respondents were]
Results to interpret:
[Paste cross-tab tables, regression output, or other statistical results]
For each result:
1. Plain-language finding: what does this result mean in business terms
2. Significance interpretation: is this difference or relationship statistically significant, and what does that mean practically
3. Effect size: is the effect large enough to be practically meaningful, not just statistically significant
4. Confidence interval: what range of values does the data support at 95% confidence
5. Caveats: what limitations should be noted in how this finding is communicated
Produce a findings narrative suitable for inclusion in a client presentation, using business language rather than statistical terminology.Prompt 3 — Cross-Tab Analysis with Significance Testing
Run a cross-tabulation analysis with significance testing for the following survey data.
Survey data: [paste data or describe the dataset structure]
Dependent variable: [variable to analyze — e.g., satisfaction rating, NPS category, purchase intent]
Independent variables (subgroups to compare):
1. [Variable 1 — e.g., age group]
2. [Variable 2 — e.g., company size]
3. [Variable 3 — e.g., usage frequency]
Statistical tests to apply:
- For categorical × categorical: chi-square test of independence
- For continuous × categorical: one-way ANOVA or t-test as appropriate
- Significance threshold: p < 0.05 (flag p < 0.01 and p < 0.001 separately)
- Correction for multiple comparisons: [apply Bonferroni correction / none]
Output format:
1. Cross-tab table for each variable pair with column percentages and significance flags
2. Plain-language interpretation of each significant finding
3. Effect size for each significant difference
4. Base size table with flags where n < 30 (interpret with caution) or n < 15 (do not report)Prompt 4 — Driver Analysis and Key Factor Identification
Conduct a driver analysis to identify the key factors predicting [outcome variable].
Dataset context: [describe the survey and sample]
Outcome variable: [what we want to predict — e.g., overall satisfaction score, NPS, likelihood to renew]
Potential driver variables: [list the attribute ratings, behavioral variables, or demographic factors to test as predictors]
Sample size: [N]
Analysis approach:
1. Run [multiple linear regression / logistic regression / relative importance analysis] as appropriate
2. Report standardized coefficients or relative importance weights for each driver
3. Assess overall model fit: R² for linear, pseudo-R² and classification accuracy for logistic
4. Test for multicollinearity among predictors and address if present
Output:
1. Key drivers chart: drivers ranked by importance with importance score and direction of effect
2. Importance-performance matrix: plot each driver by its importance (impact on outcome) and performance (current rating)
3. Plain-language narrative: which 3–5 factors matter most and what the data recommends prioritizing
4. Statistical appendix: full regression output for technical reviewPrompt 5 — Respondent Segmentation Analysis
Conduct a respondent segmentation analysis for the following survey dataset.
Survey context: [describe the survey]
Variables to use for segmentation: [list attitudinal, behavioral, or need-based variables to cluster on]
Number of segments: [specify or ask COCO to determine optimal number]
Sample size: [N]
Analysis approach:
1. Standardize variables for clustering (z-score normalization)
2. Run [k-means / hierarchical / latent class] clustering as appropriate
3. Validate segment solution: silhouette score, segment size distribution, interpretability
4. Profile each segment on all key survey variables and demographics
Output:
1. Segment size and naming: how many respondents in each segment and a descriptive segment name
2. Segment profiles: for each segment, key distinguishing characteristics on attitudinal, behavioral, and demographic variables
3. Segment comparison table: all key variables with mean/proportion for each segment and significance of between-segment differences
4. Practical implications: what each segment wants, how to reach them, and strategic priority ranking
5. Segment classification model: rules or a decision tree for classifying new respondents into segments8. AI Citation Builder & Bibliography Formatter
Generates perfectly formatted citations in any style and builds complete bibliographies from raw source information — eliminating manual formatting errors.
Pain Point & How COCO Solves It
The Pain: Citation Formatting Is Mandatory, Tedious, and Surprisingly Error-Prone
Research analysts produce reports that must meet citation standards set by clients, publishers, academic institutions, or internal style guides. Getting citations right is not optional — a misformatted reference undermines credibility, and a missing citation creates intellectual property and plagiarism risk. Yet citation formatting is simultaneously one of the most time-consuming and least intellectually engaging parts of research production. An analyst can spend an hour on bibliography formatting for a report they wrote in a day, chasing down DOIs, publisher names, volume numbers, and page ranges across sources that do not always display this information consistently.
The error rate in manual citation formatting is higher than most researchers acknowledge. Studies of published academic papers find citation errors in 25–50% of reference lists — including incorrect author names, wrong publication years, missing volume numbers, and non-functional URLs. For consulting and business research reports, the standard is not academic peer review, but client-facing professionalism — and a bibliography with formatting inconsistencies or factual errors signals carelessness that clients notice even when they do not flag it. Junior analysts, who handle much of the citation work in research organizations, are particularly susceptible to errors because they have not yet internalized the differences between APA, Chicago, MLA, Vancouver, and custom house styles.
The workflow fragmentation problem compounds both the time and error issues. Source information is collected from multiple databases, each of which formats citation metadata differently. A report might draw sources from PubMed, JSTOR, Bloomberg, Bloomberg Businessweek, a government website, and a PDF with incomplete metadata — each requiring a different approach to extract and format citation information. Assembling a clean bibliography from these heterogeneous inputs requires constant switching between tabs, mental format conversion, and manual verification of each field. Reference management tools like Zotero help but require upfront investment in capturing source metadata that many analysts skip when working under deadline pressure.
How COCO Solves It
Automatic Citation Generation from Multiple Input Types: COCO builds citations from whatever source information is available:
- Generates complete citations from DOIs, URLs, ISBNs, or PubMed IDs with full metadata retrieval
- Extracts citation information from pasted text (article titles, author names, publication information in any format)
- Processes PDF metadata to extract author, title, publisher, and date information when available
- Handles incomplete source information gracefully — generates the best citation possible and flags missing fields for verification
- Supports 50+ source types: journal articles, books, book chapters, government reports, conference papers, working papers, datasets, websites, news articles, and more
Multi-Style Format Application: COCO formats citations to any required standard:
- Applies APA (7th edition), Chicago (17th edition, author-date and notes-bibliography), MLA (9th edition), Harvard, Vancouver, and IEEE formats precisely
- Handles custom house styles by learning from sample citations provided by the analyst
- Applies the correct format for each source type within the chosen style — journal articles, books, and websites each have distinct formatting rules within each style
- Manages the differences between in-text citation formats (parenthetical vs. footnote vs. numbered) and reference list formats
- Correctly handles edge cases: multiple authors, no author, organization as author, edition numbers, translator credits, and online-first publications
Bibliography Assembly and Organization: COCO builds and formats complete reference lists:
- Assembles all sources into a properly sorted bibliography (alphabetical by author for APA/MLA, numerical for Vancouver/IEEE)
- Handles hanging indent formatting and consistent spacing across the full reference list
- Identifies and flags duplicate sources that appear in multiple citation formats
- Groups sources by type when required (primary sources / secondary sources, or by chapter for books)
- Generates annotated bibliography versions with standardized annotation format on request
In-Text Citation Integration: COCO manages the link between text citations and reference lists:
- Generates correctly formatted in-text citations for each reference
- Manages author-date citations for APA and Harvard, including disambiguation when multiple works share the same author-year combination
- Formats footnote and endnote citations for Chicago notes-bibliography style, including first-citation and subsequent-citation formats (ibid., shortened citations)
- Numbers citations sequentially for Vancouver and IEEE styles as they appear in the text
- Checks that every in-text citation has a corresponding reference list entry and vice versa
URL and Digital Source Management: COCO handles modern source types precisely:
- Captures and formats DOIs and stable URLs correctly, distinguishing between display format and functional hyperlink
- Adds access dates for online sources that require them per the citation style
- Handles sources where the URL has changed or the original is no longer available, recommending archival alternatives
- Formats database-specific source information — Bloomberg, Factiva, and academic database citations with subscription information as required
- Generates permalink-formatted citations for database sources that require stable access paths
Citation Quality Audit and Verification: COCO reviews citation accuracy before finalization:
- Cross-checks cited author names, publication years, and titles against source metadata for accuracy
- Flags potentially incorrect page numbers, volume numbers, or issue numbers for verification
- Identifies non-functional or changed URLs with recommended alternatives
- Checks citation style consistency across the full document and flags formatting deviations
- Generates a citation quality report with any identified issues and their recommended resolutions
Results & Who Benefits
Measurable Results
- Bibliography formatting time: Reduced from 45–90 minutes per report to under 10 minutes (85–90% time savings)
- Citation error rate: Drops from an average of 3.2 errors per 20 citations in manually formatted bibliographies to under 0.3 errors with COCO review
- Style compliance rate: COCO-formatted citations are rated "fully compliant" with required style guides in 97% of cases vs. 71% for manually formatted references
- Junior analyst citation quality: Junior analysts produce citations rated equivalent to senior analyst quality in 88% of cases with COCO vs. 44% manually
- Revision cycles for citations: Client or editor requests for citation corrections drop by 78% in COCO-formatted reports
Who Benefits
- Research Analysts: Eliminate the tedious citation formatting work that consumes disproportionate time relative to its analytical value — and get it right every time
- Academic Researchers: Manage large reference lists across multiple citation styles without manual formatting — and submit manuscripts confident that the bibliography is error-free
- Junior Analysts: Produce professionally formatted citations from the start of their careers without having to memorize style guide details
- Editors and Publishers: Receive manuscripts with clean, consistent citations that require minimal correction — reducing editorial processing time and communication with authors
💡 Practical Prompts
Prompt 1 — Citation Generation from Source Information
Generate properly formatted citations for the following sources.
Citation style required: [APA 7th / Chicago 17th author-date / Chicago notes-bibliography / MLA 9th / Harvard / Vancouver / IEEE / custom — describe custom style]
Source information (provide as much as available for each):
Source 1:
- Type: [journal article / book / book chapter / government report / website / news article / working paper / conference paper / dataset]
- Author(s): [names]
- Title: [title]
- Publication/Journal/Publisher: [name]
- Year: [year]
- Volume/Issue/Pages: [if applicable]
- DOI or URL: [if available]
[Repeat for each source — up to 30 sources per batch]
For any missing metadata fields, flag what is needed for a complete citation and provide the best available citation in the meantime.Prompt 2 — Bibliography Compilation from Mixed Sources
Compile a complete bibliography from the following source list.
Citation style: [specify]
Sort order: [alphabetical by author / chronological / by appearance in text / by source type]
Annotation required: [yes — [50 / 100 / 150] word annotations / no]
Sources:
[Paste all sources in any format — DOIs, URLs, partial citations, or pasted metadata]
Deliverables:
1. Complete formatted bibliography in the specified style, properly sorted
2. In-text citation format for each source (for reference when writing)
3. Flag any sources where metadata is incomplete or likely incorrect
4. Flag any apparent duplicate sources
5. If annotated bibliography: for each source, a [length] annotation summarizing the source's content and its relevance to the research
Format the bibliography exactly as it should appear in the final document.Prompt 3 — Citation Style Conversion
Convert the following citations from [source style] to [target style].
Source citations (paste the full existing citation list):
[Paste existing bibliography or citation list]
Target style: [APA 7th / Chicago 17th / MLA 9th / Harvard / Vancouver / other]
Conversion requirements:
- Maintain all source information — do not drop any metadata field
- Apply the exact formatting rules of the target style including punctuation, capitalization, and italicization
- Adjust in-text citation format as well (if applicable)
- Flag any sources where the source style was missing information required by the target style
Output the complete converted reference list, ready for direct insertion into the document.Prompt 4 — In-Text Citation Audit
Audit the following document's in-text citations and reference list for completeness and consistency.
Citation style used: [specify]
Document text (or paste relevant sections with in-text citations):
[Paste document text]
Reference list:
[Paste the current reference list]
Check for:
1. Every in-text citation has a corresponding reference list entry
2. Every reference list entry is cited in the text
3. Author names and years in in-text citations match the reference list
4. In-text citation format is consistent and correct throughout
5. Page numbers included in direct quotation citations where required by the style
6. Ibid. and short-form citations used correctly (Chicago notes style)
Produce: a discrepancy list with exact location (page/paragraph) and recommended correction for each issue.Prompt 5 — Dataset and Proprietary Source Citation
Generate properly formatted citations for the following non-standard source types.
Citation style: [specify]
Sources to cite:
1. Database record: [database name, data provider, dataset title, access date, query parameters or table name]
2. Proprietary report: [title, author organization, publication date, access method — subscription / purchased / provided by client]
3. Interview or personal communication: [name, title, organization, interview date, interview format]
4. Internal company document: [title, author, date, document type — if citation of internal documents is permitted]
5. Unpublished working paper: [author, title, institution, date, repository link if available]
6. Social media post: [platform, account name, post date, URL, content summary]
For each source type, generate:
- The full reference list entry in the specified citation style
- The in-text citation format
- A note on whether this source type is appropriate for citation in the intended document type (academic journal / consulting report / policy brief)9. AI Research Hypothesis Validation Framework
Structures research hypotheses, maps evidence requirements, and guides the analyst through a systematic validation process — reducing confirmation bias and strengthening research integrity.
Pain Point & How COCO Solves It
The Pain: Research Hypotheses Are Rarely Managed With the Rigor the Decision Stakes Require
Every substantive research engagement begins with hypotheses — working assumptions about what is true in the market, what is driving a trend, or what the investment opportunity is. These hypotheses determine where analysts look for evidence, which sources they prioritize, and which findings they emphasize in their deliverables. When hypothesis management is informal, as it typically is, the research process is vulnerable to confirmation bias: analysts find the evidence that supports their initial hypothesis, interpret ambiguous evidence as confirmatory, and give insufficient weight to contradicting evidence. The resulting research appears rigorous — it is documented, sourced, and professionally formatted — but it systematically overstates confidence in the working hypothesis.
The problem is structural, not personal. Confirmation bias is a cognitive tendency that affects even experienced researchers, and it is amplified by client dynamics in consulting contexts: clients often have a preferred answer when they commission research, and analysts who repeatedly produce findings that challenge client hypotheses face commercial pressure. The research process needs structural countermeasures — formal mechanisms for identifying and testing competing hypotheses, explicit documentation of evidence that contradicts the working hypothesis, and a distinction between correlation and causation that prevents premature closure on an explanation. These structural countermeasures exist in scientific research methodology but are rarely applied in business research practice.
The research validity problem extends to the downstream decisions that research informs. A market entry decision built on a research package with unexamined confirmation bias carries more risk than the research confidence level suggests. A portfolio company investment thesis that was never stress-tested against alternative hypotheses is more fragile than the investment committee recognizes. When these decisions go wrong, the research methodology is rarely examined as a contributing factor — the bias that produced overconfident findings is invisible in the deliverable. Improving research hypothesis management reduces the rate at which overconfident findings lead to poor decisions, which is ultimately the purpose of the research investment.
How COCO Solves It
Hypothesis Structuring and Operationalization: COCO converts vague research questions into testable hypotheses:
- Translates research objectives into specific, falsifiable hypotheses with defined evidence criteria
- Decomposes complex hypotheses into component sub-hypotheses that can be tested independently
- Identifies the observable implications of each hypothesis — what the world would look like if the hypothesis is true vs. false
- Distinguishes descriptive hypotheses (what is the state of X), causal hypotheses (what causes X), and predictive hypotheses (what will happen to X)
- Generates a hypothesis register with each hypothesis, its components, evidence requirements, and current status
Competing Hypothesis Identification: COCO generates alternative explanations to test against:
- Applies Analysis of Competing Hypotheses (ACH) methodology to identify all plausible alternative explanations for the observed evidence
- Generates devil's advocate arguments for each working hypothesis — the strongest case for why it might be wrong
- Identifies the key discriminating evidence that would distinguish the working hypothesis from each alternative
- Flags hypotheses that are observationally equivalent (consistent with the same evidence) and recommends additional tests to discriminate
- Generates a hypothesis competition matrix showing how each piece of evidence supports or contradicts each hypothesis
Evidence Mapping and Requirements Planning: COCO maps evidence needs to hypothesis structure:
- Identifies what types of evidence (quantitative data, qualitative testimony, documentary evidence, observational data) would most directly test each hypothesis
- Rates the evidence sufficiency of the current research portfolio for each hypothesis
- Flags hypotheses that are under-evidenced and prioritizes additional research to fill gaps
- Identifies when existing evidence is consistent with a hypothesis but not sufficient to confirm it — avoiding premature closure
- Maps evidence quality requirements: what level of source quality and sample size is needed to support a conclusion with the required confidence level
Confirmation Bias Detection and Mitigation: COCO applies structured checks for analytical bias:
- Reviews the research process for evidence selection bias — are there systematic gaps in what types of evidence were collected
- Identifies when the same evidence is being used to support multiple conclusions and tests for logical consistency
- Checks for anchoring bias: does the analysis too closely mirror the initial hypothesis without adequate adjustment for contrary evidence
- Generates a "steelman" of the opposing view — the strongest possible argument against the working hypothesis, built from available evidence
- Applies the Bayesian updating framework: starting from prior probability, how much should the evidence change confidence in the hypothesis
Hypothesis Status Tracking and Evidence Integration: COCO maintains a running research status register:
- Tracks each hypothesis as: supported / partially supported / mixed evidence / contradicted / insufficient evidence
- Updates hypothesis status as new evidence is collected and integrated
- Maintains a full evidence log with each piece of evidence linked to the hypothesis it tests and its status impact
- Generates an evidence strength assessment: is the support for a hypothesis based on one strong source or multiple independent corroborating sources
- Flags when the evidence picture has changed materially and the working hypothesis should be revised
Research Findings Validation Report: COCO produces a structured validation deliverable:
- Generates a hypothesis validation summary documenting each hypothesis, the evidence considered, and the conclusion with confidence level
- Produces an alternative hypothesis discussion section addressing why competing explanations were ruled out
- Creates a limitations and caveats section documenting where evidence is thin or contradictory
- Generates a recommendation for what additional evidence would materially change confidence in the conclusion
- Formats the validation framework as a research methodology appendix for client-facing reports
Results & Who Benefits
Measurable Results
- Hypothesis documentation rate: Proportion of research projects with formally documented and tracked hypotheses increases from 15% to 85% with COCO frameworks
- Competing hypothesis coverage: Number of alternative explanations explicitly tested per research project increases from an average of 1.2 to 4.5
- Confirmation bias incidents: Internal quality reviews identify confirmation bias patterns in 8% of COCO-structured reports vs. 34% of unstructured research
- Research finding revision rate: Findings that required material revision after client review due to methodological challenge drop by 55% when hypothesis validation frameworks are applied
- Analyst confidence calibration: Research analysts using COCO's evidence sufficiency assessments report confidence levels that more accurately predict actual finding reliability — overconfidence rate drops from 42% to 12%
Who Benefits
- Research Analysts: Produce research that holds up under scrutiny — with documented methodology that demonstrates analytical rigor rather than just comprehensive sourcing
- Senior Analysts and Research Directors: Spend less time correcting junior analysts' confirmation bias and more time on higher-value strategic interpretation
- Investment Committees and Decision-Makers: Receive research packages that explicitly address alternative hypotheses and document evidence sufficiency — enabling better-calibrated decision confidence
- Academic and Policy Researchers: Apply a structured hypothesis management framework that meets the standards expected in peer review and policy evaluation contexts
💡 Practical Prompts
Prompt 1 — Hypothesis Register Generation
Create a hypothesis register for the following research engagement.
Research objective: [what the research is ultimately trying to determine]
Working hypothesis (current best guess): [state the hypothesis as specifically as possible]
Research context: [industry, decision being made, who will use the findings]
Generate:
1. Hypothesis decomposition: break the working hypothesis into 3–5 testable sub-hypotheses
2. Observable implications: for each sub-hypothesis, what the world would look like if true vs. false
3. Competing hypotheses: 3–5 alternative explanations that are also consistent with the current evidence
4. Key discriminating evidence: what evidence would distinguish the working hypothesis from each alternative
5. Evidence requirements: what type, quality, and quantity of evidence is needed to support a conclusion for each sub-hypothesis
6. Hypothesis register table: all hypotheses with current status (untested), evidence assigned, and evidence needed
Format as a working document to be updated as the research progresses.Prompt 2 — Competing Hypothesis Analysis
Apply an Analysis of Competing Hypotheses (ACH) framework to the following research question.
Research question: [state the question]
Evidence collected so far: [list or describe the key evidence gathered]
Working hypothesis: [your current explanation]
Alternative hypotheses to consider: [list any you've already identified, or ask COCO to generate them]
ACH analysis:
1. Generate all plausible hypotheses that could explain the observed evidence (aim for at least 5)
2. For each piece of evidence, assess: consistent / inconsistent / not applicable for each hypothesis
3. Identify which hypotheses are most inconsistent with the evidence (eliminate or downgrade)
4. Identify which hypotheses remain viable after evidence review
5. Determine the key discriminating evidence needed to distinguish remaining viable hypotheses
6. Recommend: which hypothesis is best supported and what additional evidence is most important to collect
Output: ACH matrix (evidence × hypotheses) + narrative conclusion with confidence levelPrompt 3 — Confirmation Bias Audit
Review the following research package for confirmation bias and analytical weaknesses.
Research question: [state the question]
Conclusion reached: [state the finding and confidence level]
Evidence base: [list the sources and data used to reach the conclusion]
Research process: [describe how evidence was collected — sources searched, experts interviewed, data analyzed]
Audit for:
1. Evidence selection bias: are there systematic gaps in what was collected — types of sources, geographies, time periods, or stakeholder perspectives underrepresented
2. Interpretation consistency: is ambiguous evidence being consistently interpreted as supportive of the working hypothesis
3. Contrary evidence handling: is evidence that contradicts the conclusion documented and adequately addressed
4. Source balance: is the evidence dominated by sources that share a common perspective or institutional bias
5. Alternative hypothesis consideration: were plausible alternative explanations explicitly tested and ruled out
6. Causal reasoning: does the conclusion require a causal claim that the evidence only supports correlationally
Produce: bias audit findings with specific examples + recommended additional analyses to address identified weaknessesPrompt 4 — Evidence Sufficiency Assessment
Assess whether the following evidence base is sufficient to support the stated research conclusion.
Conclusion: [state the finding and confidence level claimed]
Decision stakes: [describe what decision will be made based on this finding — helps calibrate evidence threshold]
Evidence collected:
1. [Evidence item 1: source, type, recency, sample size if applicable]
2. [Evidence item 2]
3. [Evidence item 3]
[List all evidence items]
Assess for each evidence item:
1. Source quality tier: primary / secondary / tertiary
2. Relevance: direct / indirect / circumstantial
3. Independence: is this source independent of other cited sources, or do they share a common upstream source
4. Recency: is the evidence current enough for the decision context
Assess overall:
1. Evidence sufficiency: is the conclusion supported beyond reasonable doubt / on balance of evidence / preliminary only / insufficient
2. Single-source dependency: is the conclusion critically dependent on one source that would not survive skeptical challenge
3. Triangulation status: are the key claims supported by multiple independent evidence sources
4. Recommended additional evidence: what would most efficiently move the conclusion from its current confidence level to a higher onePrompt 5 — Bayesian Updating Framework
Apply a Bayesian reasoning framework to update confidence in the following research hypothesis.
Hypothesis: [state the hypothesis precisely]
Prior probability: [your initial confidence level before seeing current evidence — e.g., 30% likely true based on base rates or prior knowledge]
New evidence to integrate:
1. [Evidence 1: description and strength assessment]
2. [Evidence 2]
3. [Evidence 3]
[List all new evidence]
For each piece of evidence:
1. Likelihood ratio: how much more likely would we see this evidence if the hypothesis is true vs. false
2. Direction of update: does this evidence increase or decrease confidence in the hypothesis
3. Magnitude of update: how large is the confidence change justified by this evidence
Calculate:
1. Posterior probability after integrating all evidence
2. Sensitivity analysis: how does the posterior change if the prior probability was [higher/lower]
3. Evidence value assessment: which piece of evidence contributed most to the confidence update
4. Decision threshold: at what confidence level would you recommend acting on this hypothesis, and are we there
Present reasoning in plain language alongside the probability estimates.10. AI Longitudinal Study Data Tracker & Change Analyzer
Manages multi-wave research data, detects statistically significant changes over time, and generates trend narratives — turning time-series survey data into strategic insight.
Pain Point & How COCO Solves It
The Pain: Longitudinal Research Data Is Systematically Under-Analyzed Because Tracking It Is Operationally Hard
Longitudinal research — repeated measurement of the same variables over time — is among the most valuable information a research organization can produce. Time-series data enables trend detection, distinguishes sustained shifts from noise, and provides the before/after evidence needed to assess the impact of market events, interventions, or strategic initiatives. Yet most research organizations that conduct longitudinal studies systematically under-exploit their data because managing, linking, and analyzing multi-wave datasets is operationally complex in ways that episodic project work is not. Data from Wave 1 is stored in one format, Wave 2 introduced a question modification, Wave 3 changed the sampling approach — and by Wave 4, the analyst who designed the original study has moved on.
The comparability problem is the central technical challenge of longitudinal research. Valid trend analysis requires consistent measurement: the same question wording, the same response scale, the same population definition, and the same analytical approach applied across all waves. In practice, surveys evolve — questions are updated to reflect changing conditions, new variables are added, and sample frames shift. These changes create discontinuities in the data that can be mistaken for genuine trends, or can make genuine trends invisible behind methodological noise. Maintaining comparability documentation and knowing which changes are "trend-breaking" vs. tolerable is a methodological skill that is rarely formalized and frequently lost when team members change.
The analytical depth problem compounds the operational problem. Even when longitudinal data is properly managed, the analysis delivered from most multi-wave studies is shallow: simple trend lines showing whether scores went up or down, without statistical significance testing for the change, without driver analysis to understand what changed and why, and without decomposition to identify which respondent subgroups are driving aggregate trend movement. Clients receive a chart showing that satisfaction went from 72 to 69 to 71 over three waves and conclude that nothing meaningful changed — when in fact a specific customer segment deteriorated sharply while another improved, and the strategic implication is significant.
How COCO Solves It
Multi-Wave Data Architecture and Management: COCO builds a structured longitudinal data infrastructure:
- Creates a standardized variable naming and coding convention applied consistently across all waves
- Documents every modification to questionnaire wording, response options, or sampling approach across waves
- Classifies each modification as trend-breaking (requires statistical correction or disclaimer) vs. trend-preserving
- Builds a data archive with each wave's dataset linked to its metadata: field dates, sample characteristics, and methodology notes
- Generates a data dictionary that is updated at each wave to reflect the current variable set and any modifications
Wave-to-Wave Comparability Assessment: COCO validates the integrity of trend analysis before it is conducted:
- Checks sample composition across waves for demographic drift that might explain apparent trend changes
- Flags question wording changes that create response scale comparability issues
- Applies statistical tests for measurement invariance when composite scales are tracked across waves
- Documents which variables can be trended with confidence vs. which require caveats
- Generates a comparability report for each wave launch documenting the degree of methodological consistency with prior waves
Trend Detection and Statistical Significance Testing: COCO identifies real changes in time-series data:
- Calculates wave-over-wave and long-term trend slopes for all tracked variables
- Applies appropriate significance tests for trend changes: t-tests for individual wave comparisons, regression-based trend tests for multi-wave patterns
- Distinguishes statistically significant changes from sampling noise using confidence interval analysis
- Detects non-linear trend patterns: accelerations, reversals, plateaus, and discontinuities
- Applies time-series decomposition to separate trend, seasonal, and irregular components where applicable
Subgroup Trend Decomposition: COCO identifies who is driving aggregate trend movement:
- Disaggregates overall trend by key subgroups and tests whether trend patterns differ significantly across groups
- Calculates the contribution of each subgroup to the aggregate trend change — weighted by subgroup size
- Identifies convergence and divergence patterns: are subgroup scores moving closer together or further apart over time
- Tracks panel attrition effects (for true panel studies) and flags when sample composition changes may explain trend patterns
- Generates a trend driver decomposition: what proportion of the aggregate change is attributable to each identifiable factor
Event and Intervention Impact Analysis: COCO links trend data to external events:
- Identifies trend inflection points and cross-references against a timeline of market events, policy changes, and competitive moves
- Applies interrupted time-series analysis to assess the impact of a specific event or intervention on a tracked outcome
- Generates a counterfactual trend projection — what the trajectory would have looked like without the identified event
- Tests whether post-event trend changes are statistically significant and persistent vs. temporary
- Produces an event impact summary for stakeholder communication: what happened, what the data shows changed, and with what confidence
Longitudinal Report and Dashboard Generation: COCO produces structured trend reporting deliverables:
- Generates a wave-over-wave change report with all tracked variables, trend direction, magnitude, and significance status
- Produces trend line visualizations for each key metric with confidence interval bands
- Creates an executive trend summary: top 5 significant changes this wave and their strategic implications
- Generates an at-risk indicator dashboard for metrics showing concerning trend patterns
- Produces a longitudinal data story: a narrative synthesis of how the research picture has evolved over the full study period
Results & Who Benefits
Measurable Results
- Longitudinal data management time: Reduced from 2–3 analyst days per wave to under 4 hours for data preparation and comparability review (80% time savings)
- Trend analysis depth: Proportion of tracked metrics with subgroup decomposition increases from 20% to 85% when COCO automates the disaggregation
- False trend detection rate: Statistical significance testing reduces false trend alerts (changes flagged as meaningful that are within sampling error) from 35% to under 6% of reported changes
- Event impact analysis: Organizations using COCO's interrupted time-series analysis attribute 2.1x more trend movements to identifiable external causes vs. reporting changes without causal analysis
- Longitudinal data utilization: Research organizations report using 3.5 additional years of historical wave data once COCO organizes and documents it — data that was previously inaccessible due to documentation gaps
Who Benefits
- Research Analysts: Transform longitudinal data from a management burden into a strategic asset — COCO handles the operational complexity so analysts can focus on trend interpretation
- Market Research Teams: Deliver richer, more statistically defensible trend analysis without adding analytical headcount — and catch real changes in client metrics before they become crises
- Strategy Teams: Access trend data that is properly contextualized against external events, enabling attribution analysis rather than just description of what changed
- Product and CX Leaders: Monitor customer experience metrics over time with confidence that changes are real — and understand which customer segments are driving aggregate movement
💡 Practical Prompts
Prompt 1 — Wave-over-Wave Change Report
Generate a wave-over-wave change analysis for the following longitudinal survey data.
Study overview: [name and brief description of the tracking study]
Waves to compare: Wave [N-1] ([date]) vs. Wave [N] ([date])
Sample sizes: Wave [N-1]: [n], Wave [N]: [n]
Key tracked metrics: [list the variables to analyze — e.g., satisfaction score, NPS, awareness, consideration]
For each metric:
1. Wave [N-1] score vs. Wave [N] score with confidence intervals
2. Point change and percentage change
3. Statistical significance test result: is the change significant at p < 0.05
4. Direction assessment: improvement / deterioration / stable
5. Subgroup breakdown: how did the change manifest across [key subgroups]
Summary deliverables:
1. Change table: all metrics with scores, changes, and significance flags
2. Top 3 significant improvements and top 3 significant declines with brief interpretation
3. Strategic implications: what the pattern of changes suggests about [brand / product / market] health
4. Watch list: metrics approaching concerning threshold levelsPrompt 2 — Long-Term Trend Analysis
Conduct a long-term trend analysis across all waves of the following tracking study.
Study: [name]
Waves available: [list all waves with dates and sample sizes]
Metrics to trend: [list key variables]
Known methodology changes: [list any question modifications, sample frame changes, or other methodological discontinuities, with the wave they occurred]
Analysis:
1. Long-term trend line for each metric: calculate the trend slope across all waves
2. Trend significance: is the overall trend statistically significant (regression-based test)
3. Trend stability: is the trend linear or are there acceleration/deceleration patterns
4. Discontinuity detection: identify any wave pairs with abnormally large changes — assess whether methodology change or genuine shift
5. Subgroup trend divergence: which subgroups are trending in different directions from the aggregate
Output:
1. Trend chart data for each metric across all waves with confidence intervals
2. Trend summary table: metric, overall direction, trend magnitude, significance, notable subgroup patterns
3. Longitudinal narrative: how has the research picture evolved over the full study periodPrompt 3 — Intervention Impact Analysis
Assess the impact of the following intervention on tracked outcomes.
Tracking study: [name and description]
Intervention: [describe what happened — product launch, campaign, policy change, competitor entry, etc.]
Intervention date: [when it occurred]
Pre-intervention waves: [list waves before the intervention]
Post-intervention waves: [list waves after the intervention]
Outcome metrics to assess: [list tracked variables expected to be affected]
Analysis approach:
1. Pre-intervention trend: establish the trajectory before the intervention
2. Post-intervention trend: assess whether the trajectory changed after the intervention
3. Interrupted time-series analysis: is the post-intervention change statistically significant beyond the pre-intervention trend
4. Counterfactual comparison: what would the metric have been at post-intervention timepoints if the pre-intervention trend had continued
5. Attribution assessment: how confident can we be that the observed change is attributable to the intervention vs. other factors
Output: Impact summary with quantified effect size, confidence level, and alternative explanations consideredPrompt 4 — Subgroup Trend Decomposition
Decompose the following aggregate trend change into its subgroup components.
Metric: [the tracked measure]
Aggregate change: from [score] in Wave [N-1] to [score] in Wave [N]
Subgroup data:
[Paste subgroup scores and sample sizes for both waves]
Subgroups to analyze: [list — e.g., age groups, region, product usage tier, customer segment]
Analysis:
1. Subgroup scores for Wave [N-1] and Wave [N] with confidence intervals
2. Subgroup-level change: point change and significance for each subgroup
3. Contribution analysis: what percentage of the aggregate change is attributable to each subgroup (weighted by subgroup size)
4. Convergence/divergence: are subgroup scores moving closer together or further apart
5. At-risk subgroup identification: which subgroups show the most concerning trend patterns
Strategic interpretation: what does the subgroup decomposition reveal about the aggregate trend that the overall number does not showPrompt 5 — Longitudinal Comparability Review
Conduct a comparability review for the following tracking study wave.
Study: [name]
New wave: Wave [N] ([date])
Prior wave: Wave [N-1] ([date])
Changes made since last wave:
1. Question modifications: [list any wording changes with before/after text]
2. Scale changes: [any response scale modifications]
3. Sample changes: [any changes to the sampling approach or target population definition]
4. Fielding changes: [mode, timing, or length changes]
For each change:
1. Comparability impact: does this change affect the ability to trend this metric
2. Classification: trend-breaking / trend-affecting (requires caveat) / trend-preserving
3. Recommended action: report as trended / report with caveat / do not trend this wave / apply statistical correction
Overall comparability rating for this wave: [generate based on above assessment]
Recommended disclosures: what should be noted in the methodology section of this wave's report11. AI Regulatory & Policy Landscape Tracker
Maps the full regulatory environment for any industry across all active jurisdictions — updated continuously so analysts never miss a compliance or market-shaping development.
Pain Point & How COCO Solves It
The Pain: Regulatory Change Is Constant, Consequential, and Impossible to Monitor Comprehensively Without Dedicated Infrastructure
Regulatory and policy developments are among the most consequential market-shaping forces a research analyst must track — and among the hardest to monitor systematically. A single industry may face oversight from a dozen federal and state agencies, multiple international regulatory bodies, and legislative processes that can move from introduction to enactment in months. Each regulatory body publishes through different channels: Federal Register notices, agency websites, press releases, comment period announcements, and enforcement actions that may not appear in mainstream news coverage for weeks, if at all. Analysts who rely on news monitoring miss the regulatory signals that live in official channels; those who try to monitor official channels directly face an overwhelming volume of filings that require legal expertise to interpret.
The geographic complexity compounds the monitoring challenge. For companies operating internationally, the regulatory landscape multiplies: GDPR in Europe, the CCPA and state-level variations in the US, PIPL in China, sector-specific frameworks in each jurisdiction. A change in one market may signal regulatory direction in others, or may create compliance conflicts that affect market entry strategy. Regulatory arbitrage opportunities — where a more permissive regulatory environment in one jurisdiction creates competitive or operational advantage — are only visible to analysts who can compare frameworks across markets simultaneously. This kind of cross-jurisdictional regulatory intelligence requires more capacity than most research teams can dedicate to it.
The time lag problem is the most practically damaging aspect of inadequate regulatory monitoring. Regulatory changes typically have a comment period, a finalization period, and a compliance deadline — providing advance warning that should enable strategic preparation. Analysts who identify a regulatory change after finalization rather than at the proposed rule stage miss the window for client input on the regulation and the planning time for compliance adaptation. Being reactive to regulatory change rather than anticipating it is a competitive disadvantage that compounds over time, and the analysts who build a reputation for early regulatory signal detection are among the most valued in their firms.
How COCO Solves It
Multi-Jurisdiction Regulatory Feed Monitoring: COCO monitors official regulatory channels across all relevant jurisdictions:
- Ingests Federal Register notices, agency press releases, proposed rulemaking documents, and enforcement action summaries daily
- Monitors international regulatory bodies: EU Commission, ESMA, FCA, MAS, CSRC, and 40+ additional agency feeds
- Tracks state and provincial regulatory actions for US and Canadian markets
- Monitors industry-specific regulatory bodies: FDA, OCC, FERC, FCC, SEC, CFTC, and sector equivalents globally
- Applies semantic filtering to surface only the rulemakings and enforcement actions relevant to the defined research coverage universe
Regulatory Development Classification and Prioritization: COCO structures regulatory intelligence for analyst consumption:
- Classifies each development by type: proposed rule, final rule, guidance document, enforcement action, comment period, legislative action, court ruling
- Assigns a business impact tier based on the scope of affected entities, compliance burden, and market structure implications
- Tags developments by affected sector, business activity, and geographic scope for precise relevance filtering
- Prioritizes developments with upcoming deadlines — comment periods closing, compliance dates approaching — for time-sensitive attention
- Links related regulatory developments across jurisdictions — tracking when multiple regulators are moving in the same direction
Regulatory Content Analysis and Plain-Language Summary: COCO translates regulatory documents into actionable intelligence:
- Extracts the key compliance requirements from proposed and final rules in plain language
- Identifies which specific business practices, products, or corporate structures are affected
- Summarizes the compliance timeline: when compliance is required and what the phased implementation schedule looks like
- Identifies exemptions, safe harbors, and transitional provisions that affect the practical impact on different company types
- Flags ambiguities in regulatory text that may require legal interpretation and notes where agency guidance is pending
Cross-Jurisdictional Regulatory Comparison: COCO maps regulatory variance across markets:
- Builds comparison matrices showing how the same activity is regulated differently across jurisdictions
- Identifies regulatory arbitrage opportunities where one jurisdiction's framework creates a competitive or operational advantage
- Tracks regulatory harmonization trends — where multiple jurisdictions are converging toward a common framework
- Flags compliance conflicts where requirements in one jurisdiction are incompatible with those in another
- Generates market entry regulatory feasibility assessments based on the full cross-jurisdictional compliance picture
Regulatory Impact and Strategic Implication Analysis: COCO connects regulatory developments to business implications:
- Assesses the competitive impact of a regulatory change: who benefits, who is disadvantaged, and whether new entrants or incumbents are relatively favored
- Evaluates the cost and feasibility of compliance for different company profiles — large vs. small, domestic vs. international
- Identifies second-order effects: how regulatory change in one area creates ripple effects in adjacent markets or business models
- Generates a scenario analysis for proposed rules: what the market looks like under the rule-as-proposed vs. rule-as-amended vs. rule-rescinded scenarios
- Produces a strategic response framework: monitor only / prepare for compliance / engage in comment period / adjust market strategy
Regulatory Calendar and Alert System: COCO keeps analysts ahead of regulatory deadlines:
- Maintains a rolling 12-month regulatory calendar with all relevant comment periods, compliance dates, and hearing schedules
- Generates proactive alerts when a comment period opens for a regulation with significant client impact
- Tracks the status of pending rules from proposed through final, updating the calendar as timelines shift
- Produces a monthly regulatory pipeline report showing what is coming in the next 90 days
- Generates comment letter drafting support when client participation in a rulemaking is warranted
Results & Who Benefits
Measurable Results
- Regulatory monitoring time: Reduced from 5–8 hours per week of manual monitoring to under 1 hour of review for a comprehensive regulatory intelligence digest (85% time savings)
- Regulatory development lead time: Analysts identify material regulatory developments an average of 6–8 weeks earlier with systematic monitoring vs. news-based awareness
- Comment period participation rate: Organizations using COCO's comment period alerts increase participation in relevant rulemakings from 12% to 67% of eligible proceedings
- Cross-jurisdictional coverage: COCO monitors 40+ regulatory bodies simultaneously vs. 4–6 in typical manual monitoring, reducing regulatory blind spots
- Strategic response preparation time: With earlier regulatory signal detection, organizations have 3–5x more preparation time before compliance deadlines than those relying on reactive monitoring
Who Benefits
- Research Analysts: Maintain comprehensive regulatory awareness across their coverage universe without dedicating a disproportionate share of their time to official channel monitoring
- Compliance and Legal Teams: Receive early warnings of proposed regulations with enough lead time to assess compliance implications and prepare responses before deadlines
- Strategy Teams: Incorporate regulatory scenario analysis into strategic planning — understanding not just current rules but the regulatory trajectory over the next 2–3 years
- Investment Analysts: Monitor the regulatory environment for portfolio companies and investment targets — identifying regulatory risks and opportunities before they are priced into valuations
💡 Practical Prompts
Prompt 1 — Regulatory Landscape Overview
Build a regulatory landscape overview for the following industry and jurisdiction(s).
Industry: [sector]
Jurisdiction(s): [list countries, states, or regions]
Business activities to cover: [list specific products, services, or practices most relevant to regulation]
Audience: [internal analysts / clients / board / investors]
Deliverables:
1. Regulatory body map: all agencies with jurisdiction over this industry in each market, with their primary mandate
2. Key existing regulations: the 5–10 most significant current rules affecting this industry, with compliance summary
3. Active rulemaking pipeline: proposed or pending regulations with status and expected timeline
4. Enforcement environment: recent enforcement actions and what they signal about regulatory priorities
5. Regulatory trajectory: is the regulatory environment tightening, loosening, or shifting focus — and why
6. Compliance calendar: key upcoming dates (comment periods, compliance deadlines, reporting requirements) in the next 12 months
Format: Executive briefing suitable for a quarterly board regulatory updatePrompt 2 — Proposed Rule Impact Analysis
Analyze the business impact of the following proposed regulation.
Regulation: [name and agency]
Summary of what is proposed: [brief description or paste key provisions]
Publication date: [date of proposed rule]
Comment period deadline: [date]
Industry affected: [sector]
Company profile for impact assessment: [describe the type of company — size, business model, geographic footprint]
Analyze:
1. Scope: which activities, products, entities, or transactions are covered
2. New compliance requirements: what specific changes are required and by when
3. Exemptions and safe harbors: which company types or activities are excluded
4. Compliance cost estimate: what operational and systems changes are required and approximate cost burden
5. Competitive impact: does this favor incumbents over new entrants, large over small, domestic over foreign
6. Strategic options: what are the 3 response options (compliance as proposed / engage in comment period / restructure to reduce scope / exit the affected activity)
7. Comment period recommendation: should we submit a comment letter, and if so, what key points to raise
Output: Impact analysis memo suitable for distribution to legal, compliance, and strategy teamsPrompt 3 — Cross-Jurisdictional Regulatory Comparison
Compare the regulatory framework for [business activity or product type] across the following jurisdictions.
Jurisdictions to compare: [list 3–6 countries or regions]
Activity to compare: [describe the specific practice, product, or business model being regulated]
Key dimensions to compare:
- Licensing or authorization requirements
- Substantive compliance obligations (conduct, disclosure, capital, data, etc.)
- Enforcement approach and penalty severity
- Recent regulatory trend: tightening / loosening / stable
- Compliance timeline and transition provisions
Deliverables:
1. Comparison matrix: rows = jurisdictions, columns = regulatory dimensions
2. Summary of material differences: where the regulatory burden differs most significantly across markets
3. Most permissive and most restrictive jurisdiction identification with reasoning
4. Regulatory arbitrage analysis: whether differences create meaningful competitive or operational advantages
5. Convergence assessment: are these jurisdictions moving toward a common framework or diverging further
6. Market entry implication: for a company entering [2–3 new markets], regulatory considerations ranked by impact
Format: Comparative regulatory briefing with matrix + narrative synthesisPrompt 4 — Regulatory Comment Letter Drafting
Draft a regulatory comment letter for the following proposed rule.
Proposed rule: [name and agency]
Docket number: [if known]
Comment deadline: [date]
Submitting organization: [our organization name and type]
Our position: [support / oppose / support with modifications / request clarification]
Key points to make:
1. [Point 1: specific provision with rationale]
2. [Point 2]
3. [Point 3]
[List up to 6 specific points]
Evidence or data to cite: [list any data, studies, or examples to reference]
Draft a comment letter that:
- Opens with a clear statement of position and organizational identity
- Addresses each key point with specific reference to the regulatory text
- Supports each position with data, examples, or policy arguments
- Closes with a constructive summary and offer to provide additional information
- Meets standard regulatory comment letter conventions for the [agency name] docket
Tone: professional, constructive, data-drivenPrompt 5 — Enforcement Action Analysis
Analyze the following regulatory enforcement action(s) and extract strategic intelligence.
Enforcement action(s):
[Paste enforcement action summary or describe: regulator, respondent, charges, sanctions, date]
Industry context: [sector and relevant regulatory environment]
Our interest: [why this enforcement action is relevant — as a competitor, similarly situated company, or analyst]
Analyze:
1. What conduct was found to violate the regulation — specifically, not just generally
2. What facts or circumstances triggered the enforcement action
3. What sanctions were imposed and whether they appear proportionate to the violation
4. What this signals about the regulator's current enforcement priorities
5. What compliance changes the enforcement action implies for similarly situated companies
6. Whether this enforcement action appears to be one-off or the beginning of a broader enforcement sweep
7. Strategic implication: what should companies in this industry do differently
Produce: Enforcement intelligence brief with compliance implication summary12. AI Consumer Behavior Segmentation Analyzer
Transforms survey, behavioral, and demographic data into actionable consumer segments — with profiles, sizing, and strategic implications built in.
Pain Point & How COCO Solves It
The Pain: Consumer Segmentation Projects Consume Significant Resources but Too Often Produce Segments That Gather Dust
Consumer segmentation is one of the most requested and least successfully implemented deliverables in market research. Organizations invest in segmentation studies — designing research, fielding surveys, conducting focus groups — and receive a report with five or six consumer archetypes, colorful names, and demographic profiles. Twelve months later, the segmentation has rarely been operationalized: marketing campaigns still target by demographics, product development still prioritizes the loudest stakeholder, and sales teams have reverted to pre-study instincts. The problem is not the segmentation itself but the gap between the research deliverable and the operational decisions it was supposed to inform.
The analytical limitation compounds the implementation problem. Most consumer segmentation studies segment on demographics or stated attitudes — variables that are easy to measure but poorly predictive of actual behavior and commercial value. The segments that matter for strategic decision-making are defined by need states, behavioral patterns, and decision-making processes: who is most likely to switch, who is most responsive to a specific value proposition, who generates disproportionate revenue, and who is dissatisfied in ways that create competitive vulnerability. These behavioral segmentation variables are available in transaction data, CRM records, and behavioral tracking, but integrating them with survey data and deriving commercially meaningful segments requires analytical sophistication that most research teams cannot bring to bear.
The actionability problem is the ultimate test. A segmentation is only valuable if it changes decisions — and a segmentation that cannot be used to identify which segment a specific customer belongs to is operationally useless for targeting. The pen portrait approach, where each segment is described by attitudes and demographics that do not form a reliable classification rule, produces segments that research teams recognize but cannot deploy. Building segments with an operational classification model — rules or a predictive algorithm that assigns new customers to segments based on available data — requires technical sophistication that most segmentation projects do not deliver.
How COCO Solves It
Multi-Source Data Integration for Segmentation: COCO builds segmentation from the richest available data:
- Integrates survey-based attitudinal data with behavioral data from CRM, transaction records, digital analytics, and loyalty programs
- Standardizes and normalizes variables across data sources, handling missing values with appropriate imputation
- Identifies which variables are most discriminating for segmentation through feature importance analysis before clustering runs
- Handles both continuous and categorical variables with appropriate distance metrics for each type
- Flags data quality issues — sparse variables, high missing rates, potential encoding errors — before they distort segment solutions
Statistical Segmentation and Cluster Validation: COCO applies rigorous analytical methods:
- Tests multiple clustering algorithms — k-means, hierarchical, latent class analysis, DBSCAN — and recommends the best fit for the data structure
- Determines the optimal number of segments using silhouette scores, elbow method, and interpretability criteria
- Validates segment stability through bootstrap resampling — testing whether segments are robust across different data subsets
- Calculates within-segment homogeneity and between-segment separation metrics to confirm segment distinctiveness
- Generates multiple segment solution options at different levels of granularity for client review
Segment Profiling and Characterization: COCO builds rich, multi-dimensional segment profiles:
- Generates a statistical profile for each segment: mean/proportion on every segmentation and descriptor variable with significance of between-segment differences
- Identifies the 5–7 most distinctive characteristics of each segment — the variables where that segment differs most from all others
- Calculates segment size (as proportion of population and estimated absolute count) with confidence intervals
- Maps segments on 2D and 3D visualization spaces to communicate distinctiveness and similarity visually
- Generates a segment naming framework based on the most distinctive behavioral and attitudinal characteristics
Segment Value and Commercial Analysis: COCO connects segments to commercial outcomes:
- Calculates segment-level metrics for commercially relevant outcomes: revenue per customer, retention rate, lifetime value, share of wallet
- Ranks segments by commercial attractiveness — combining size, value, and accessibility
- Identifies which segments are most responsive to specific product features, price points, or channel preferences
- Maps segment competitive vulnerability: which segments are most likely to switch and under what conditions
- Generates a segment opportunity matrix: plotted by attractiveness vs. current penetration to identify priority targets
Operational Segment Classification Model: COCO builds segments that can be used in practice:
- Develops a classification model (decision tree, logistic regression, or rule-based system) that assigns new customers to segments using available operational data
- Validates classification accuracy on held-out data to assess how reliably new customers can be assigned
- Identifies the minimum data fields needed for segment classification — enabling CRM integration
- Generates segment assignment rules in plain language suitable for use by non-technical teams
- Produces a segment membership probability score for each customer, enabling priority sorting within segments
Strategic Implication and Activation Framework: COCO translates segments into strategy:
- Generates segment-specific value propositions: what each segment needs and how the offering should be positioned for them
- Produces segment activation playbooks: recommended messaging, channels, product priorities, and service approach for each segment
- Creates a targeting prioritization framework: which segments to pursue first given available resources and strategic priorities
- Generates a segment monitoring dashboard specification: which metrics to track to assess segment health over time
- Produces an implementation roadmap: how to operationalize the segmentation across marketing, sales, product, and CX functions
Results & Who Benefits
Measurable Results
- Segmentation project cycle time: Reduced from 6–10 weeks to 2–3 weeks for a full consumer segmentation study including fieldwork
- Segment stability rate: Bootstrap validation shows COCO segments are stable in 87% of cases vs. 52% for unstaged k-means solutions
- Operational deployment rate: Segmentations with COCO-built classification models are deployed to CRM and targeting systems in 74% of cases vs. 22% for profile-only segmentations
- Revenue impact: Organizations that implement COCO-validated segmentations report 15–25% improvement in targeted campaign response rates in the first year
- Segment distinctiveness: COCO-optimized segment solutions show 40% higher between-segment F-statistics than unoptimized k-means solutions on the same data
Who Benefits
- Research Analysts: Deliver segmentation projects that get implemented — not just delivered — because the analytical rigor and operational classification model make them actionable
- Marketing Teams: Access rich, behaviorally grounded segments they can actually use for targeting — with clear, deployable classification rules rather than abstract archetypes
- Product Teams: Understand which consumer segments have which unmet needs — enabling product prioritization that reflects commercial reality rather than internal preference
- Customer Experience Leaders: Design differentiated service experiences for distinct customer segments based on their value and needs rather than treating all customers uniformly
💡 Practical Prompts
Prompt 1 — Consumer Segmentation Analysis
Conduct a consumer segmentation analysis for the following dataset.
Dataset description: [survey / CRM / behavioral / combined — describe variables available]
Sample size: [N]
Segmentation objective: [what business decisions should the segmentation inform]
Variables to use for segmentation: [list attitudinal, behavioral, or need-based variables]
Descriptor variables: [demographics, usage, satisfaction, etc.]
Analysis steps:
1. Variable review: recommend which segmentation variables to include/exclude based on discriminating power
2. Test 3–5 segment solutions and recommend the optimal number
3. Profile each segment on all segmentation and descriptor variables
4. Identify the 5 most distinctive characteristics per segment
5. Calculate segment sizes and confidence intervals
6. Rank segments by commercial attractiveness
Output: Full segment profiles + segment comparison table + strategic priority rankingPrompt 2 — Segment Classification Model
Build a classification model to assign customers to segments using operational data.
Segment definitions: [describe the segments or paste the segment profiles]
Operational data available for classification: [list fields available in CRM or transaction system]
Build:
1. A classification model using [decision tree / logistic regression / rule-based system]
2. Classification accuracy assessment on held-out data: overall accuracy + accuracy per segment
3. Minimum variable set: smallest set of operational fields that achieves acceptable accuracy
4. Classification rules in plain language: a decision tree or rule set a non-technical user can understand
5. Segment probability scores: for each customer, a probability estimate for each segment membership
6. Implementation specification: how to integrate this classification logic into a CRM system
Flag any segments that are particularly hard to classify reliably and recommend how to handle ambiguous cases.Prompt 3 — Segment Value Analysis
Calculate the commercial value of each segment in the following consumer segmentation.
Segments identified: [list segment names]
Data available for value analysis:
- Transaction data: [revenue, purchase frequency, basket size]
- Retention data: [churn rate or contract renewal by segment]
- Customer lifetime: [average relationship duration]
For each segment calculate:
1. Average revenue per customer (annual)
2. Purchase frequency and basket size
3. Retention rate or churn risk
4. Estimated customer lifetime value (CLV) with formula and assumptions
5. Segment-level total revenue contribution: segment size × average revenue
Synthesize:
1. Segment value ranking from most to least commercially valuable
2. Revenue concentration: what % of total revenue comes from the top segment(s)
3. Value vs. size matrix: plot segments by size and value
4. Priority targeting recommendation: which segments to invest in growing penetrationPrompt 4 — Segment Activation Playbook
Generate a segment activation playbook for the following consumer segment.
Segment name: [name]
Segment profile summary: [paste or describe key characteristics]
Segment size: [number and % of total market]
Segment commercial value: [revenue per customer or CLV estimate]
Current penetration: [what % of this segment are current customers]
Generate a playbook covering:
1. Value proposition: how to articulate the offering for this segment's priorities and pain points
2. Messaging framework: 3 key messages and supporting proof points
3. Channel strategy: where this segment is most reachable and most responsive
4. Product/feature priorities: which offerings best fit this segment, and what gaps exist
5. Pricing and offer sensitivity: how price-sensitive they are and what offer structures work best
6. Service model: what level and type of customer service this segment expects
7. Success metrics: what to track to measure activation effectiveness
Format: One-page playbook per segment, ready for distribution to marketing, sales, and product teamsPrompt 5 — Competitive Vulnerability Segment Analysis
Assess the competitive vulnerability of each consumer segment.
Segments: [list with brief profile for each]
Industry context: [describe the competitive landscape]
Data available: [satisfaction scores, NPS, stated switching intent, stated reasons for loyalty]
For each segment assess:
1. Satisfaction gap: how satisfied is this segment vs. their stated ideal
2. Switching likelihood: how likely are they to switch in the next 12 months
3. Primary defection triggers: what experiences would most likely cause this segment to switch
4. Competitive targeting: which competitors are most actively pursuing this segment
5. Loyalty anchors: what factors most strongly bind this segment to their current provider
6. Intervention opportunity: what action could most cost-effectively reduce defection risk
Prioritize segments by competitive vulnerability and recommend defensive and offensive strategies for highest-risk segments.13. AI Earnings Call Transcript Insight Extractor
Extracts strategic signals, management sentiment shifts, and forward guidance changes from earnings transcripts — across 50+ companies in the time it takes to read one.
Pain Point & How COCO Solves It
The Pain: Earnings Calls Contain Critical Intelligence That Most Analysts Cannot Extract at Scale
Earnings call transcripts are among the richest, most systematically neglected information sources in financial and business research. Every publicly traded company conducts them quarterly; they contain unscripted management statements, analyst question responses, and forward-looking language that cannot be found in any other source. An experienced analyst reading a transcript carefully can extract management confidence signals, detect strategy pivots signaled through changed language, identify competitive intelligence volunteered under analyst questioning, and spot divergences between prepared remarks and Q&A responses that reveal genuine vs. stated priorities. But a single transcript takes 60–90 minutes to read and extract insights from — and an analyst covering 20 companies faces 80 transcripts per quarter, on top of all their other work.
The comparative analysis problem is even more limiting than the volume problem. The most valuable earnings intelligence is not what a single company said on a single call but what changes when comparing calls over time, or what differs when comparing how multiple companies discuss the same market condition. Is the language around pricing power becoming more defensive across the sector? Are management teams describing demand weakness in different terms than last quarter? Are capital allocation priorities shifting in a direction that signals industry-wide concern? This kind of systematic, cross-company, cross-quarter comparative analysis would take an analyst team weeks to conduct manually and is rarely done — leaving the most valuable layer of earnings intelligence uncaptured.
The interpretation challenge adds a further layer of complexity. Management teams are sophisticated communicators who are legally constrained in what they can say and strategically motivated to frame developments favorably. The gap between what management says and what the data shows — or between the prepared remarks and the unscripted Q&A — often contains more information than either taken alone. Detecting these gaps requires both analytical training and domain familiarity with how management teams in a specific sector typically communicate, which calibrates the analyst to recognize when language is unusually bullish, unusually hedged, or suspiciously imprecise.
How COCO Solves It
Rapid Transcript Processing and Structuring: COCO converts unstructured transcripts into analyzable data:
- Processes earnings call transcripts in any format: PDF, HTML, plain text, or structured API feeds
- Separates and labels prepared remarks, analyst Q&A exchanges, and management response segments
- Identifies speakers by name and role for attribution accuracy
- Extracts all forward-looking statements, guidance figures, and quantified management projections
- Creates a structured summary outline with section headers for rapid navigation by analysts
Sentiment and Language Change Analysis: COCO detects shifts in management communication:
- Applies natural language processing to measure tone: confidence/uncertainty, optimism/caution, precision/vagueness
- Tracks language changes from prior quarters for each company — flagging when specific terminology shifts that may signal strategy changes
- Identifies hedge words, qualifiers, and evasive language patterns that appear in responses to specific question topics
- Measures the ratio of concrete commitments to vague aspirational language in management guidance
- Generates a management confidence index for each topic area discussed, calibrated against historical baseline for that company
Strategic Signal Extraction: COCO identifies the most intelligence-rich content:
- Extracts all references to competitive dynamics — how management describes their competitive position and changes to it
- Identifies capital allocation signals: where management is committing, deferring, or redirecting investment
- Extracts pricing power language: how management discusses their ability to pass through costs or raise prices
- Flags demand and volume commentary with geographic and product-level specificity
- Surfaces technology and innovation signals: what management says about new product development, R&D priorities, and technology investment
Forward Guidance Tracking and Change Analysis: COCO manages quantitative guidance systematically:
- Extracts all numerical guidance figures (revenue, margin, growth, unit economics) with management attribution
- Compares current guidance to prior-quarter guidance for each metric, calculating revision magnitude and direction
- Tracks guidance revision patterns over multiple quarters to assess management credibility and forecasting conservatism
- Flags guidance language changes: when "we expect" becomes "we anticipate" — precision signals
- Generates a guidance revision tracker across the full coverage universe for any given reporting period
Cross-Company and Sector-Level Synthesis: COCO enables the comparative analysis that creates the most value:
- Synthesizes how multiple companies in the same sector describe the same market condition
- Identifies consensus narratives vs. outlier views — which companies are more or less optimistic than peers
- Extracts cross-company competitive intelligence: what companies say about each other directly and obliquely
- Generates sector-level earnings trend reports: the dominant themes, surprises, and strategic signals across all companies in a given quarter
- Tracks how a specific topic is discussed across all companies over multiple quarters
Earnings Intelligence Deliverable Generation: COCO produces structured, client-ready outputs:
- Generates a one-page earnings summary per company with key takeaways, guidance changes, and strategic signals
- Produces a sector earnings digest summarizing key themes across all companies in a coverage universe
- Creates a guidance change tracker with all revised figures, direction of revision, and management explanation
- Generates an analyst question intelligence report: what questions analysts asked, how management responded, and where management was evasive
- Produces a cross-quarter sentiment trend chart for any tracked company or sector
Results & Who Benefits
Measurable Results
- Transcript analysis time: Reduced from 60–90 minutes per transcript to under 10 minutes review of COCO's structured extraction (85–90% time savings)
- Coverage capacity: Analysts can comprehensively analyze 5–8x more earnings calls per quarter using COCO vs. manual reading
- Cross-quarter language change detection: COCO detects statistically significant sentiment shifts that are missed in 68% of cases by analysts reading individual transcripts
- Guidance revision tracking accuracy: COCO achieves 99.2% accuracy in extracting and comparing quantitative guidance figures vs. 87% for manual extraction under time pressure
- Strategic signal capture: Comparative analysis across 10+ transcripts by COCO surfaces 3–4 sector-level insights not visible from any single transcript review
Who Benefits
- Investment Research Analysts: Achieve comprehensive earnings coverage across a broad universe without sacrificing depth of analysis on any individual company
- Buy-Side Portfolio Managers: Receive timely, synthesized earnings intelligence across their entire portfolio within hours of the last call in a reporting cycle
- Sell-Side Equity Research Teams: Produce sector-level earnings synthesis reports that differentiate coverage by identifying themes invisible to analysts reading individual transcripts
- Strategy Teams: Monitor competitive intelligence embedded in public company earnings calls — a rich, systematic source of competitor intelligence most teams exploit only occasionally
💡 Practical Prompts
Prompt 1 — Single Earnings Call Deep-Dive
Extract structured intelligence from the following earnings call transcript.
Company: [company name]
Quarter/Period: [Q1 2025 / FY2024 / etc.]
Transcript: [paste transcript or describe where to find it]
Analyst's focus areas: [specific topics most relevant to your research or investment thesis]
Extract:
1. Key prepared remarks themes: the 3–5 most strategically significant points from management's prepared remarks
2. Guidance summary: all quantitative guidance figures with any changes from prior guidance
3. Q&A intelligence: what analysts asked about, how management responded, and where responses were notably evasive or precise
4. Competitive signals: any commentary about competitors, market share, or competitive dynamics
5. Capital allocation signals: investment priorities, acquisition appetite, buyback/dividend plans
6. Language change flags: any terminology that differs from the prior quarter transcript
7. Sentiment assessment: overall management confidence level vs. prior quarter
Output: One-page summary + detailed intelligence register with transcript citationsPrompt 2 — Quarter-over-Quarter Language Change Analysis
Compare the following two earnings call transcripts and identify significant language and sentiment changes.
Company: [company name]
Transcript A (prior quarter): [paste or describe]
Transcript B (current quarter): [paste or describe]
Analyze changes in:
1. Topics discussed: what was added, dropped, or de-emphasized from prior quarter
2. Specific language shifts: where management used different words for the same topic
3. Guidance language precision: more or less specific than prior quarter on key metrics
4. Tone shifts: more optimistic, more cautious, or more defensive on specific topics
5. Q&A dynamics: were analysts more or less probing, did management's response style change
Produce:
1. Language change register: topic by topic comparison with specific quote pairs showing the change
2. Signal interpretation: what each significant change likely indicates about management's actual view
3. Overall assessment: is this quarter's communication more or less transparent/confident than prior quarterPrompt 3 — Sector Earnings Synthesis
Synthesize the key themes from the following set of earnings call transcripts.
Companies: [list 5–15 company names]
Sector: [industry]
Quarter: [time period]
Key topics to synthesize: [themes most relevant to your research — pricing power, demand trends, AI investment, labor costs, geographic exposure]
For each topic:
1. How many companies discussed it and in what proportion of total commentary
2. The range of management views: what is the consensus and who is the outlier
3. Whether the discussion is more or less prominent than last quarter
4. Key quotes illustrating the range of perspectives
5. What the cross-company picture implies for sector dynamics
Overall synthesis:
1. Top 3 sector-wide themes this quarter with supporting evidence
2. Biggest surprises vs. consensus expectations
3. Watchlist: topics where management language is becoming more guarded
4. Strategic implications for investors, competitors, or market participantsPrompt 4 — Guidance Revision Tracker
Build a guidance revision tracker for the following companies and reporting period.
Companies: [list]
Reporting period: [quarter/year]
Metrics to track: [revenue guidance, operating margin, EPS, unit volume, etc.]
For each company and each metric:
1. Prior guidance: what management said last quarter
2. Current guidance: what management said this quarter
3. Revision: direction (raised / lowered / maintained / withdrawn) and magnitude
4. Management explanation: what reason was given for the revision
5. Analyst reaction: were analysts satisfied or did they probe further
Synthesize:
1. Guidance revision pattern: across all companies, are revisions predominantly upward or downward
2. Credibility flags: companies with a pattern of lowering guidance repeatedly
3. Implied sector outlook: what the aggregate guidance revision pattern suggests about the sector's trajectoryPrompt 5 — Competitive Intelligence Mining
Extract competitive intelligence from the following earnings call transcripts.
Target competitor: [company whose competitive position we want to understand]
Transcripts to mine: [list companies whose transcripts to analyze — ideally include the target + its competitors + major customers/suppliers]
Extract:
1. Direct mentions of the target competitor: what other companies say about them, positive or negative
2. Indirect references: discussions of "a major competitor" that likely refer to the target
3. Win/loss signals: where management mentions winning or losing business
4. Pricing references: any commentary on competitive pricing pressure that implies the target's behavior
5. Product or capability references: where competitors acknowledge or challenge specific capabilities of the target
6. Customer and supplier commentary: what companies in the target's value chain say about their relationship
Synthesize: a competitive intelligence summary on the target built entirely from public earnings call commentary14. AI Patent Application Landscape Mapping Engine
Maps the full patent landscape for any technology domain — identifying key players, white spaces, and innovation trajectories — in a fraction of the time of manual patent research.
Pain Point & How COCO Solves It
The Pain: Patent Landscape Analysis Is Essential to Technology Strategy but Prohibitively Slow to Do Well
Patent landscape analysis is one of the most strategically valuable forms of research available — and one of the most time-consuming. Patent databases contain the most granular, verifiable record of where technology investment is going: what problems companies are trying to solve, what solutions they have developed, and where they are staking intellectual property claims for the future. A well-constructed patent landscape can reveal a competitor's technology roadmap more accurately than any press release, identify white spaces where innovation is occurring without defensive IP coverage, and flag freedom-to-operate risks before they become expensive legal problems. But building a meaningful patent landscape manually requires weeks of database searching, claim interpretation, assignee normalization, and visualization work that most research teams cannot sustain.
The search quality problem is the first barrier. Patent search is a highly specialized skill — identifying all relevant patents requires constructing Boolean queries across multiple classification systems (CPC, IPC, USPC), normalizing inventor and assignee names across subsidiaries and name variations, and distinguishing between patents that are relevant to the research question and those that merely contain the same keywords. Analysts without patent search training routinely miss 30–50% of the relevant patent population through inadequate query construction, and include significant noise through under-filtering. The resulting landscape is built on an incomplete and contaminated dataset that distorts every downstream analysis.
The interpretation challenge is as significant as the search challenge. Raw patent count data — company X has 200 patents in domain Y — tells a research analyst almost nothing without understanding the quality distribution, the temporal pattern, the geographic coverage, and the citation network. Transforming patent data into strategic intelligence requires integrating multiple analytical layers that individually require specialized expertise and collectively exceed the capacity of most research engagements.
How COCO Solves It
Comprehensive Patent Search and Retrieval: COCO builds a rigorous patent search across all major databases:
- Constructs optimized Boolean search queries using CPC, IPC, and keyword combinations specific to the technology domain
- Searches across USPTO, EPO (Espacenet), WIPO (PatentScope), and national patent databases for global coverage
- Normalizes assignee names across subsidiaries, holding companies, and name variations to ensure accurate competitive attribution
- Distinguishes patent applications from granted patents and tracks through prosecution to grant or abandonment
- Applies technology-specific relevance filtering to remove noise while maintaining comprehensive coverage of the target domain
Competitive IP Position Analysis: COCO maps who owns what in the patent landscape:
- Calculates patent counts, application rates, and portfolio size by assignee with trend analysis over time
- Identifies the top patent holders in the domain and their relative IP position strength
- Maps each company's patent activity to their product and R&D strategy using product line cross-referencing
- Distinguishes operating companies from NPEs (non-practicing entities) and patent assertion entities in the landscape
- Generates a competitive IP heat map showing where each player is most and least active within the technology domain
Technology Cluster and White Space Analysis: COCO structures the technology landscape:
- Groups patents into technology clusters using CPC classification analysis and semantic clustering
- Maps the technology landscape as a visual cluster diagram showing relative density and relationships
- Identifies white spaces: technology areas relevant to the domain with low patent density
- Tracks technology cluster evolution over time — which areas are growing, maturing, or declining in patent activity
- Generates a technology roadmap projection based on the direction and velocity of recent patent activity
Citation Network and Patent Quality Analysis: COCO assesses patent quality beyond count:
- Calculates forward citation counts as a proxy for patent importance and foundational value
- Identifies the most-cited "anchor" patents in the domain — those that subsequent filings build on
- Maps citation networks to reveal which companies are building on each other's foundational patents
- Identifies recently filed patents rapidly accumulating citations — early signals of foundational value
- Distinguishes between core IP (high citation, broad claims) and defensive filing programs (high count, narrow claims, low citation)
Freedom-to-Operate Risk Assessment: COCO flags IP risk for specific technology implementations:
- Identifies patents with claims potentially relevant to a specified technology implementation or product feature
- Assesses patent status: active, expired, abandoned, or in litigation
- Maps the geographic scope of relevant patent protection by jurisdiction
- Identifies patent expiration timelines for key competitive IP — when specific technologies enter the public domain
- Generates a freedom-to-operate risk register with each potentially relevant patent, its claim scope, and its expiration date
Patent Intelligence Report Generation: COCO produces structured, client-ready patent landscape deliverables:
- Generates an executive summary: key findings, competitive IP position ranking, and strategic implications
- Produces a full patent landscape report with technology cluster maps, competitor analysis, and trend charts
- Creates a competitive IP comparison table showing each player's portfolio size, activity trend, and cluster focus
- Generates a white space opportunity brief identifying the most attractive areas for new IP investment
- Formats all visualizations as data-ready exports for inclusion in presentations and strategy documents
Results & Who Benefits
Measurable Results
- Patent landscape production time: Reduced from 3–6 analyst weeks to 3–5 days for a comprehensive domain landscape (80%+ time savings)
- Patent search recall: COCO-constructed searches achieve 85–92% recall of relevant patents vs. 50–65% for typical analyst-constructed queries in unfamiliar domains
- Competitive IP insights: COCO surfaces an average of 2–3 significant competitive IP insights not identified in manual landscape reviews
- Citation analysis adoption: Proportion of patent landscapes including citation quality analysis increases from 15% to 90% when COCO automates the citation network computation
- Strategic IP decision quality: Organizations using COCO patent landscapes report 35% higher confidence in IP investment and freedom-to-operate decisions
Who Benefits
- Research Analysts: Deliver technology intelligence that incorporates patent data — a uniquely verifiable and forward-looking signal source — without requiring months of patent search specialization
- R&D and Technology Strategy Teams: Understand the full competitive IP landscape before committing to technology development directions — avoiding inadvertent infringement and identifying defensible white spaces
- Investment Analysts: Assess the quality and defensibility of a target company's IP portfolio as part of technology company due diligence
- Corporate IP and Legal Teams: Accelerate freedom-to-operate analysis and competitive IP monitoring with COCO's systematic landscape mapping
💡 Practical Prompts
Prompt 1 — Technology Patent Landscape Overview
Build a patent landscape overview for the following technology domain.
Technology domain: [describe the technology — be as specific as possible]
Geographic scope: [global / US only / US + EU + JP / specify]
Time period: [last 5 / 10 / 20 years of patent activity]
Purpose: [investment due diligence / competitive intelligence / R&D strategy / freedom-to-operate preliminary assessment]
Deliverables:
1. Patent search strategy: classification codes and keywords used, databases searched, query logic
2. Landscape overview: total patent count, growth trend, geographic distribution of filings
3. Top 15 assignees: patent count, filing trend, and technology cluster focus for each
4. Technology cluster map: major sub-domains within the technology, with patent density and key players per cluster
5. White space identification: areas with low patent density relative to apparent commercial relevance
6. Trend analysis: which technology clusters are growing, maturing, or declining in filing activity
7. Strategic implications: what the patent landscape tells us about where the industry is heading
Output: Executive summary (2 pages) + detailed landscape report + data tables for visualizationPrompt 2 — Competitor Patent Portfolio Analysis
Analyze the patent portfolio of the following company in the context of [technology domain].
Company: [assignee name — include subsidiaries to cover]
Technology focus: [specific domain or product area]
Comparison companies: [list 3–5 competitors for benchmarking]
Time period: [last 5 or 10 years]
Analyze:
1. Portfolio size and growth: total patents, application rate trend, granted vs. pending
2. Technology cluster distribution: what proportion of the portfolio covers each technology sub-domain
3. Geographic coverage: which jurisdictions has the company protected its IP in
4. Citation analysis: what is the average citation count for their patents vs. competitors
5. Key patents: the 10 most-cited patents in their portfolio with claim scope summary
6. Filing velocity change: has filing activity accelerated, decelerated, or shifted to different clusters in the last 2 years
7. Competitive comparison: how does this company's portfolio compare to peers on size, quality, and coverage
Strategic interpretation: what does the portfolio tell us about this company's technology bets and defensive IP strategyPrompt 3 — Patent White Space Analysis
Identify innovation white spaces in the following technology domain for IP strategy purposes.
Technology domain: [description]
Our technology focus: [describe the specific technology area we are developing or considering]
Competitive landscape: [list key players known to be active in this space]
Analysis:
1. Technology decomposition: break the domain into 15–25 specific technology sub-areas
2. Patent density by sub-area: how many patents exist in each sub-area and who holds them
3. White space identification: sub-areas with low patent density relative to apparent commercial or technical importance
4. White space quality assessment: for each white space — is it unoccupied because technically immature, commercially unproven, or genuinely overlooked
5. IP opportunity scoring: rank white spaces by attractiveness factoring in commercial potential, technical feasibility, and current patent density
6. Recommended IP investment areas: top 3–5 white spaces where new patent filing would secure defensible IP position
Output: White space map + opportunity-ranked white space briefPrompt 4 — Freedom-to-Operate Preliminary Screening
Conduct a preliminary freedom-to-operate patent screening for the following technology implementation.
Technology to screen: [describe the specific feature, method, or product to assess]
Key technical elements: [list the specific technical aspects that might be covered by patents]
Geographic markets of interest: [where the product will be sold or used]
Time horizon: [when market entry is planned]
Screen for:
1. Potentially relevant patents: identify active patents with claims that may read on the described technology
2. Patent status: active / expired / abandoned / in litigation for each identified patent
3. Assignee profile: who holds the patent — operating company, NPE, or patent assertion entity
4. Claim scope assessment: how broadly or narrowly do the claims cover the described implementation
5. Expiration timeline: when do relevant active patents expire
6. Risk tier: high / medium / low risk for each identified patent with reasoning
Note: This is a preliminary screening only — full FTO analysis requires qualified patent counsel review.
Output: Preliminary risk register with recommended patents for formal legal reviewPrompt 5 — Patent Expiration and Opportunity Analysis
Identify significant patent expirations in [technology domain] over the next [3/5/10] years and assess the strategic implications.
Technology domain: [description]
Geographic markets: [list]
Expiration window: [specify timeframe]
Analysis:
1. Significant patents expiring: identify foundational or highly-cited patents scheduled to expire in the window
2. Technology entering public domain: what specific technology capabilities become freely usable when each patent expires
3. Current exclusivity holders: who currently benefits from exclusivity under each expiring patent
4. Competitive implications: how does expiration change the competitive dynamics — who benefits most
5. Market entry opportunities: which expiring patents create the clearest opportunities for new market entry or product development
6. Timing analysis: which expirations are most strategically significant and why
Strategic recommendation: top 3 expiration events to monitor and prepare for, with suggested preparation actions15. AI Geopolitical Risk Impact Assessment Tool
Translates complex geopolitical developments into structured, quantified risk assessments with direct business implications — built for research analysts who are not geopolitical specialists.
Pain Point & How COCO Solves It
The Pain: Geopolitical Risk Is Now Central to Business Research but Requires Expertise Most Analysts Do Not Have
Geopolitical risk has moved from the periphery to the center of business and investment research. Supply chain disruptions, sanctions regimes, trade policy reversals, and regional conflicts now routinely affect company valuations, market access, and operational continuity in ways that were considered tail risks a decade ago. Research analysts across disciplines — equity research, strategy consulting, market analysis — are increasingly expected to incorporate geopolitical risk assessment into their work. But geopolitical analysis is a distinct discipline with its own frameworks, sources, and interpretive conventions. A market research analyst or financial analyst who is expert in their domain may have little background in how to assess political stability, sanctions propagation, or escalation dynamics — and the consequences of producing a naive geopolitical risk assessment that misses material risks are significant.
The frameworks available to non-specialist analysts are inadequate for the task. Country risk ratings from commercial providers provide composite scores that obscure the specific risk drivers most relevant to a particular business. Geopolitical news is voluminous and often more alarming than actionable. Academic political science research is rigorous but not formatted for business decision-making. The result is that most business research either ignores geopolitical risk entirely, treats it as a boilerplate disclosure, or oversimplifies it to the point of uninformativeness. The organizations that do geopolitical risk well — large financial institutions, intelligence-linked consulting firms — do it through dedicated teams with specialized expertise that most research organizations cannot replicate.
The operationalization problem is the deepest challenge. Even analysts who can describe geopolitical risk accurately often struggle to translate it into business-specific terms: what is the probability that this risk materializes, what is the magnitude of the impact on specific business activities, and what is the timeline? These questions require not just geopolitical knowledge but the ability to connect political dynamics to supply chain structures, regulatory frameworks, capital flow patterns, and operational dependencies. Building this connection requires cross-disciplinary synthesis that is cognitively demanding and time-intensive even for experienced analysts.
How COCO Solves It
Geopolitical Risk Framework Application: COCO applies structured analytical frameworks to geopolitical situations:
- Applies recognized risk assessment frameworks (PRS ICRG, Freedom House indicators, World Bank governance indicators) to structure the analysis
- Decomposes country-level risk into component risk drivers: political stability, regulatory risk, sanctions exposure, conflict risk, expropriation risk, currency controls
- Applies scenario-based analysis to assess the probability and impact of specific geopolitical outcomes
- Cross-references current developments against historical analogues to calibrate risk severity and trajectory
- Generates a risk driver hierarchy: which factors are most important for the specific business exposure being assessed
Business Exposure Mapping: COCO connects geopolitical risk to specific business activities:
- Maps the company's or portfolio's exposure to the identified geopolitical risks by geography, supply chain node, revenue stream, and regulatory dependency
- Quantifies exposure concentration: what proportion of revenue, COGS, or operational capacity is exposed to the at-risk geography or relationship
- Identifies second-order exposures: suppliers, customers, or partners who are directly exposed and whose disruption would create indirect impact
- Maps regulatory and sanctions exposure: which business activities or transactions are in scope for the identified risk
- Calculates an exposure index that combines probability, magnitude, and business impact for each identified risk factor
Sanctions and Trade Policy Risk Analysis: COCO provides structured sanctions intelligence:
- Maps current sanctions regimes relevant to the identified geopolitical risk — OFAC, EU, UN, and bilateral sanctions
- Identifies which specific business activities, counterparties, or transactions are in scope for existing or threatened sanctions
- Assesses sanctions escalation probability: based on political dynamics, what is the likelihood of new or expanded sanctions
- Analyzes secondary sanctions risk: exposure to sanctions through third-party relationships with sanctioned entities
- Generates a sanctions compliance checklist for the specific business context with jurisdiction-specific requirements
Scenario Development and Probability Assessment: COCO builds actionable risk scenarios:
- Develops three to five discrete geopolitical scenarios ranging from status quo continuation to severe escalation
- Assigns probability estimates to each scenario based on political indicators, historical base rates, and expert assessments
- Quantifies the business impact of each scenario on key metrics: revenue, margins, supply chain costs, market access
- Identifies the key indicator variables to monitor that would signal movement toward each scenario
- Generates a scenario narrative that is specific enough to be operationally meaningful rather than generically cautionary
Risk Mitigation Strategy Framework: COCO generates practical risk response options:
- Identifies supply chain diversification options that reduce geographic concentration exposure
- Generates financial hedging strategies for currency and commodity risks linked to geopolitical exposure
- Maps regulatory and compliance responses for sanctions and trade policy risks
- Identifies information and monitoring requirements for early warning of risk scenario activation
- Produces a tiered response plan: actions to take now, actions contingent on scenario activation, and options to preserve for severe scenarios
Geopolitical Risk Report Generation: COCO produces structured, business-appropriate deliverables:
- Generates an executive geopolitical risk brief: top risks, probability/impact matrix, and recommended actions
- Produces a country risk profile for any market of interest with structured assessment of each risk dimension
- Creates a scenario analysis document with full narrative, probability assessment, and business impact quantification
- Formats deliverables appropriate for investment committee, board, or client strategy audiences
- Generates monitoring dashboards with key geopolitical indicators and their current status
Results & Who Benefits
Measurable Results
- Geopolitical risk assessment time: Reduced from 1–2 analyst weeks to 1–2 days for a structured country-level risk assessment (80% time savings)
- Risk framework coverage: COCO-structured assessments cover an average of 8–12 distinct risk dimensions vs. 2–3 in typical non-specialist analyses
- Scenario specificity: COCO scenarios include quantified business impact for 90% of identified risks vs. qualitative description only in 75% of manual assessments
- Early warning indicator coverage: Organizations using COCO's monitoring frameworks identify geopolitical risk escalation signals an average of 6–10 weeks earlier than those relying on news-based monitoring
- Cross-disciplinary integration: COCO-produced assessments are rated "adequately integrated with financial and operational implications" by senior reviewers in 85% of cases vs. 38% for non-specialist analyses
Who Benefits
- Research Analysts: Produce credible, structured geopolitical risk assessments without being geopolitical specialists — applying frameworks and synthesis that produce decision-relevant output
- Investment Analysts: Incorporate geopolitical risk into equity research and portfolio risk assessment with the quantification and scenario specificity that investment decisions require
- Corporate Strategy Teams: Assess market entry and operational footprint decisions against a structured geopolitical risk framework rather than gut instinct and news monitoring
- Risk Management Functions: Build a systematic, defensible geopolitical risk monitoring program without dedicated geopolitical intelligence staff
💡 Practical Prompts
Prompt 1 — Country Geopolitical Risk Profile
Build a structured geopolitical risk profile for the following country.
Country: [name]
Business context: [describe the business activities or investments at risk]
Time horizon: [12 months / 3 years / 5 years]
Assess each risk dimension:
1. Political stability: government strength, opposition activity, election risk, institutional quality
2. Policy and regulatory risk: likelihood of adverse policy changes affecting business activities
3. Sanctions and trade policy risk: current exposure and escalation probability
4. Conflict and security risk: risk of domestic instability or regional conflict affecting operations
5. Expropriation and rule of law risk: property rights protection and contract enforcement reliability
6. Currency and capital controls risk: probability of currency devaluation or capital flow restrictions
7. Corruption and compliance risk: bribery risk and regulatory enforcement environment
For each dimension:
- Current status assessment (1–5 scale with explanation)
- Trend direction: improving / stable / deteriorating
- Key drivers of the assessment
- Business impact if this risk materializes
Overall country risk rating + top 3 risks to prioritize + monitoring indicatorsPrompt 2 — Geopolitical Scenario Analysis
Develop a geopolitical scenario analysis for the following situation.
Geopolitical situation: [describe the current development or tension]
Business exposure: [describe how the company or portfolio is exposed]
Time horizon for scenarios: [12 months / 24 months]
Develop 4 scenarios:
1. Status quo continuation: current trajectory holds without major escalation or de-escalation
2. Managed escalation: tensions increase but remain below threshold of severe disruption
3. Severe escalation: worst plausible case within the realistic range
4. De-escalation: tensions reduce and the situation moves toward resolution
For each scenario:
- Probability estimate: [% with rationale]
- Key events or triggers that define this scenario
- Business impact on: revenue exposure, supply chain costs, market access, regulatory compliance burden
- Quantified impact range: [low/high estimate for key financial metrics]
- Early warning indicators: signals that would suggest this scenario is becoming more likely
Output: Scenario matrix (4 scenarios × 5 dimensions) + narrative brief + probability-weighted expected impact calculationPrompt 3 — Sanctions Risk Assessment
Assess the sanctions risk exposure for the following business situation.
Business activity: [describe the transaction, relationship, or operation to assess]
Parties involved: [list all entities — companies, individuals, jurisdictions]
Sanctions regimes to assess: [OFAC / EU / UN / UK / all applicable]
Industry: [sector — some sectors have sector-specific sanctions]
Assessment:
1. Direct sanctions exposure: are any parties on current sanctions lists — primary sanctions risk
2. Indirect (secondary) sanctions exposure: do any parties have relationships with sanctioned entities
3. Sectoral sanctions: are any parties in sectors subject to sectoral restrictions
4. Licensing requirements: do any activities require specific licenses or authorizations
5. Due diligence gaps: what additional screening or documentation is needed
6. Risk rating: high / medium / low overall sanctions risk with specific findings supporting the rating
7. Recommended actions: steps to mitigate identified risks before proceeding
Note: This assessment does not constitute legal advice. Formal sanctions compliance requires qualified legal counsel review.Prompt 4 — Supply Chain Geopolitical Risk Mapping
Map the geopolitical risks embedded in the following supply chain.
Supply chain description: [describe the supply chain — product type, key sourcing geographies, manufacturing locations, logistics routes]
End markets: [where the final product is sold]
Company profile: [describe the company — size, industry, geographic footprint]
Map:
1. Geopolitical risk by supply chain node: for each key sourcing country and manufacturing location, assess the top 3 geopolitical risks
2. Concentration analysis: identify over-dependence on single countries or regions for critical inputs
3. Sanctions and trade policy exposure: identify nodes at risk from existing or threatened trade restrictions
4. Logistics route risk: key shipping routes or border crossings vulnerable to geopolitical disruption
5. Supplier second-tier exposure: known or likely exposure of key suppliers to their own geopolitical risks
Risk prioritization: rank supply chain nodes by combination of risk severity and business impact of disruption
Mitigation options: for the top 3 risk concentrations, identify specific diversification or dual-sourcing strategiesPrompt 5 — Geopolitical Risk Monitoring Dashboard
Design a geopolitical risk monitoring dashboard for the following business.
Business profile: [describe the company — industry, geographic footprint, key markets and sourcing locations]
Key geopolitical risk exposures: [list the 3–5 highest priority geopolitical risks for this business]
Monitoring frequency: [daily / weekly / monthly]
For each risk exposure, identify:
1. Leading indicators: 3–5 measurable signals that would indicate the risk is escalating before it becomes a crisis
2. Data sources: where to monitor each indicator
3. Threshold levels: at what signal level should the risk alert trigger a review
4. Alert escalation protocol: who gets notified and what action is triggered at each alert level
Dashboard design:
1. Risk status overview: current assessment for each tracked risk (green / amber / red)
2. Key indicator readings: current status of each leading indicator with trend direction
3. Recent developments: last 5 significant developments with date and relevance assessment
4. Scenario probability tracker: current probability estimate for each identified scenario
Format: Specification document for dashboard implementation + manual monitoring protocol for interim use16. AI Supply Chain Vulnerability Research Synthesizer
Synthesizes public data on supplier dependencies, chokepoints, and disruption risks into a structured supply chain intelligence report — without requiring proprietary supplier data.
Pain Point & How COCO Solves It
The Pain: Supply Chain Risk Research Is Fragmented Across Too Many Sources to Do Comprehensively Without Dedicated Infrastructure
Supply chain vulnerability research has become a strategic priority for companies, investors, and policymakers — accelerated by the disruptions of the past five years. But conducting rigorous supply chain risk research is operationally complex in ways that exceed the capacity of most research teams. The relevant data is spread across corporate filings (where companies disclose key suppliers in varying levels of detail), government trade statistics (which reveal import dependencies by product category), academic and policy research on critical mineral and component dependencies, shipping and logistics data (which identifies route concentrations and port bottlenecks), and supplier-specific research from specialized industry databases. No single source provides a comprehensive picture; building one requires integrating multiple streams with different formats, geographies, and levels of granularity.
The visibility problem is fundamental. Most companies disclose only their immediate Tier 1 suppliers, while the actual vulnerability in complex supply chains often lies at Tier 2 or Tier 3 — the suppliers of suppliers whose concentration or geographic risk is invisible from the top-level view. The automotive sector learned this in 2011 when the Tohoku earthquake disrupted semiconductor suppliers that no major carmaker had even identified as part of their supply chain. Research that only analyzes disclosed Tier 1 suppliers misses the majority of the actual vulnerability. Reconstructing lower-tier supplier networks from public sources — trade data, corporate disclosures, industry reports, and patent cross-referencing — is possible but requires significant analytical investment.
The research synthesis problem compounds the data collection challenge. Different stakeholders need different supply chain risk intelligence: investors need to assess supply chain risk exposure relative to peers; companies need to prioritize which vulnerabilities to mitigate; policymakers need to understand systemic dependencies across multiple companies and sectors. The raw data does not communicate these different needs — it must be synthesized, structured, and contextualized for each audience. Building that synthesis manually requires not just supply chain domain knowledge but also the analytical frameworks for risk prioritization and the time to integrate information into a coherent deliverable.
How COCO Solves It
Multi-Source Supply Chain Data Aggregation: COCO pulls supply chain intelligence from across the public information landscape:
- Extracts supplier disclosure data from annual reports, 10-K and 20-F filings, and corporate sustainability reports
- Analyzes trade data from US Census Bureau, UN Comtrade, and EU trade statistics to identify import concentration by product category and origin country
- Incorporates critical mineral and component dependency research from government agencies (USGS, DOE, European Commission) and research institutions
- Mines shipping and logistics data for route concentration, port dependency, and transit chokepoint identification
- Tracks supply chain disruption events from industry reports, news archives, and insurance loss databases to establish historical risk patterns
Supplier Dependency and Concentration Analysis: COCO maps the structure and concentration of supply chains:
- Identifies disclosed Tier 1 suppliers by company and assesses concentration — dependence on a small number of critical suppliers
- Reconstructs likely Tier 2 dependencies by cross-referencing known Tier 1 suppliers with their own public disclosures and trade data
- Calculates geographic concentration: what proportion of supply comes from specific countries or regions for each critical category
- Identifies single-source dependencies where a company or sector relies on one supplier or one country for a critical input
- Maps horizontal dependencies: where multiple companies in a sector share the same supplier — creating systemic rather than company-specific risk
Chokepoint and Critical Node Identification: COCO identifies where supply chains are most vulnerable to disruption:
- Identifies geographic chokepoints: shipping lanes, ports, and border crossings through which a disproportionate share of supply passes
- Maps processing bottlenecks: where raw materials or components must pass through a concentrated processing capacity regardless of origin
- Identifies logistical single points of failure: where the supply chain depends on a single facility, carrier, or logistics node
- Assesses the concentration of critical technology inputs (semiconductors, rare earth elements, advanced materials) at each processing stage
- Generates a chokepoint vulnerability score combining concentration, alternative availability, and disruption consequence
Disruption Scenario and Impact Analysis: COCO models the consequences of supply chain disruption:
- Develops disruption scenarios for the highest-vulnerability nodes: what happens if this supplier, country, or route is disrupted
- Estimates the financial impact of each disruption scenario: production delays, cost premium for alternative sourcing, revenue impact
- Assesses recovery time: how long would it take to restore supply from alternative sources, and what is the cost
- Compares disruption exposure across companies in the same sector to identify which players are most and least vulnerable
- Generates a vulnerability-impact matrix for prioritizing mitigation investment
Benchmark and Peer Comparison Analysis: COCO contextualizes supply chain risk relative to peers:
- Compares supply chain concentration and geographic exposure across peer companies using disclosed data
- Identifies which companies in a sector have invested most in supply chain diversification and resilience
- Benchmarks supply chain risk disclosure quality — which companies provide the most and least transparency on their dependencies
- Assesses whether supply chain risk management maturity correlates with operational resilience in historical disruption events
- Generates a peer comparison scorecard across key supply chain risk dimensions
Supply Chain Risk Intelligence Report: COCO produces structured, audience-appropriate deliverables:
- Generates an executive supply chain risk brief: key vulnerabilities, probability/impact assessment, and priority mitigations
- Produces a detailed supply chain vulnerability report with full data sourcing and analytical methodology
- Creates investor-appropriate supply chain risk disclosure analysis for ESG and risk management contexts
- Generates policy briefing formats for government and regulatory audiences
- Produces monitoring dashboards with key supply chain risk indicators and their current status
Results & Who Benefits
Measurable Results
- Supply chain research cycle time: Reduced from 3–5 analyst weeks to 4–6 days for a comprehensive supply chain vulnerability assessment (80% time savings)
- Supplier tier visibility: COCO's cross-referencing approach identifies an average of 2.4x more Tier 2 dependencies than disclosed-only analysis for the same companies
- Disruption scenario coverage: Assessments include quantified impact analysis for 4–6 disruption scenarios vs. 1–2 in typical manual supply chain risk work
- Source integration: COCO synthesizes data from 15–20 distinct source categories vs. 4–6 in typical manual research
- Peer comparison depth: Supply chain benchmarking across 10+ peer companies completed in under 2 days vs. 2–3 analyst weeks manually
Who Benefits
- Research Analysts: Deliver supply chain risk intelligence that integrates trade data, corporate disclosures, and policy research — without needing specialized supply chain expertise in every engagement
- Investment Analysts: Assess supply chain vulnerability as a risk factor in company and sector analysis — with the peer comparison context needed to assess relative vulnerability
- Corporate Risk and Strategy Teams: Identify and prioritize supply chain vulnerabilities before they become crises — using a systematic research approach rather than incident-driven response
- Policymakers and Regulators: Access structured analysis of systemic supply chain dependencies across sectors — supporting resilience policy development with evidence rather than anecdote
💡 Practical Prompts
Prompt 1 — Company Supply Chain Vulnerability Assessment
Conduct a supply chain vulnerability assessment for the following company.
Company: [name]
Industry: [sector]
Products or product lines to focus on: [specify or assess all]
Purpose: [investor due diligence / internal risk management / competitive benchmarking / regulatory response]
Assess:
1. Disclosed supplier dependencies: key Tier 1 suppliers identified in public filings, with concentration analysis
2. Geographic concentration: top 3 countries of origin for key inputs, with % concentration and risk rating
3. Critical input dependencies: any inputs with limited alternative sources
4. Tier 2 visibility: likely Tier 2 dependencies inferred from Tier 1 suppliers' own disclosures and trade data
5. Disruption history: supply chain disruptions this company has experienced or disclosed in the past 5 years
6. Resilience indicators: evidence of supply chain diversification investment, dual-sourcing, or inventory strategy
Risk assessment: top 5 supply chain vulnerabilities ranked by probability × impact, with mitigation status
Output: Vulnerability assessment report + risk register + comparison to 3 sector peersPrompt 2 — Critical Input Dependency Analysis
Analyze the supply chain dependency for the following critical input or material.
Critical input: [e.g., lithium, semiconductor chips, rare earth elements, API ingredients, specific component]
Industry context: [which sectors depend on this input and for what applications]
Geographic scope: [global supply chain analysis]
Analyze:
1. Global production concentration: where is this input produced, refined, and processed — what is the geographic concentration
2. Reserve and resource distribution: where are the underlying reserves located and production trajectory
3. Key suppliers and their market position: who are the major producers and processors, and what is their ownership structure
4. Trade flow analysis: how does this input move through the global supply chain from production to end use
5. Demand outlook: how is demand expected to grow given trends in end-use markets
6. Supply risk factors: what specific events could disrupt supply (geopolitical, environmental, regulatory, technical)
7. Substitution and alternative sourcing options: what alternatives exist and at what cost and timeline premium
Output: Critical input dependency brief with supply chain map, risk scenarios, and strategic implicationsPrompt 3 — Sector Supply Chain Concentration Analysis
Analyze supply chain concentration risk across the following sector.
Sector: [industry]
Geographic focus: [which countries or regions to assess as supply chain risk sources]
Input categories to analyze: [list key input categories — raw materials, components, services, logistics]
For the sector overall:
1. Map the top 5 input dependencies where concentration risk is highest
2. Identify which specific countries represent the greatest supply chain risk concentration for the sector
3. Assess whether concentration has increased or decreased over the past 5 years and why
4. Identify the companies within the sector that are most and least exposed to the concentration risks identified
5. Assess the systemic risk dimension: what happens if the top supply constraint materializes across all sector participants simultaneously
Policy and strategic context:
1. Current policy responses (reshoring incentives, strategic stockpiles, friend-shoring policies) and their effectiveness
2. Industry-led resilience initiatives and their current status
3. Remaining gaps: where is supply chain concentration still an unaddressed systemic risk
Output: Sector supply chain vulnerability report with concentration heat map and peer comparison tablePrompt 4 — Supply Chain Disruption Scenario Analysis
Model the impact of the following supply chain disruption scenario.
Disruption scenario: [describe the specific disruption — e.g., Taiwan Strait conflict disrupting semiconductor supply, Red Sea closure disrupting European-Asia shipping, major supplier bankruptcy]
Company or sector affected: [who is exposed]
Exposure data: [describe known supplier relationships, geographic dependencies, or input concentrations]
Model:
1. Disruption scope: which specific supply relationships or logistics flows are directly affected
2. Immediate impact (0–30 days): what production capacity is disrupted and what is the financial impact
3. Medium-term impact (30–180 days): what alternative sourcing options are available and at what cost premium
4. Recovery trajectory: how long to restore full production capacity through alternatives and what is the total cost
5. Competitive differentiation: which companies recover fastest based on their supply chain configuration
Output: Disruption impact model with financial estimates + recovery roadmap + competitive differentiation analysisPrompt 5 — Supply Chain Resilience Benchmarking
Benchmark the supply chain resilience of the following companies.
Companies to compare: [list 5–10 companies in the same sector]
Resilience dimensions to assess:
1. Geographic diversification: how distributed is supply across multiple countries
2. Supplier concentration: degree of dependence on single suppliers for critical inputs
3. Inventory and buffer strategy: evidence of strategic inventory or long-term supply contracts
4. Disclosure quality: how transparently does the company disclose supply chain risks and dependencies
5. Resilience investment: evidence of active investment in supply chain diversification or nearshoring
6. Disruption recovery history: how quickly has the company recovered from past supply chain disruptions
Output:
1. Resilience scorecard: each company scored on each dimension (1–5 scale) with evidence basis
2. Overall resilience ranking with narrative rationale
3. Best practice identification: which company's supply chain approach is most worth emulating and why
4. Laggard identification: which companies have the most material unaddressed supply chain vulnerability17. AI Academic White Paper Outline & Argument Builder
Transforms a research question and evidence base into a fully structured white paper outline with argument architecture, evidence mapping, and section-by-section writing guidance.
Pain Point & How COCO Solves It
The Pain: The Gap Between Raw Research and a Well-Structured Argument Document Is Where Good Research Dies
Research analysts frequently produce not just reports but white papers — documents intended to advance a specific argument, influence policy or industry practice, or establish thought leadership on a complex topic. White paper writing requires a fundamentally different skill set from report writing: it requires not just synthesizing evidence but constructing a persuasive argument, anticipating counter-arguments, and architecting a document that guides a skeptical reader from opening premise to conclusion. Many analysts who are excellent at gathering and synthesizing evidence struggle with this argumentative structure — the difference between a research briefing (here is what the data shows) and a white paper (here is why you should change your mind about X).
The structural architecture challenge is the first barrier. A white paper is not a collection of findings organized by topic — it is an argument that unfolds in a specific sequence designed to maximize persuasive impact. The opening must frame the problem in a way that motivates the reader to care about the argument. The body must present evidence in an order that builds confidence progressively, addresses the strongest counterarguments at the right moment, and uses concrete examples to make abstract points tangible. The conclusion must translate the argument into actionable implications that justify the reader's investment of time. Building this architecture from scratch for each white paper requires both argumentative skill and substantial planning time that most analysts do not have.
The quality consistency problem compounds the structural challenge. In research organizations that produce many white papers, the quality of the argumentative structure varies enormously by author. Senior authors with strong rhetorical training produce white papers that are genuinely persuasive; junior authors produce documents that read like extended literature reviews. This quality variance is reputationally damaging and difficult to correct at the editing stage — rebuilding the argument architecture of a completed draft is nearly as much work as starting over. Front-end investment in argument architecture, before any writing begins, produces more consistent results and eliminates most of the painful restructuring that currently happens in the editing phase.
How COCO Solves It
Argument Architecture Design: COCO builds the logical structure of the white paper before writing begins:
- Translates the research question and desired conclusion into a structured argument tree — premises leading to the conclusion
- Identifies the minimum set of claims that must be established for the conclusion to follow
- Maps the logical dependencies between claims — which must be established before which others can be argued
- Identifies the strongest likely counterarguments to the central thesis and determines where in the document structure they should be addressed
- Generates multiple argument architecture options for analyst review — different sequences for different audience types and rhetorical strategies
Evidence Mapping and Sufficiency Assessment: COCO connects evidence to argument structure:
- Maps each piece of available evidence to the argument claim it supports
- Identifies claims that are unsupported or under-evidenced and flags them for additional research
- Assesses the type and quality of evidence supporting each claim — distinguishing strong from weak
- Identifies where evidence supports multiple claims and can be introduced efficiently
- Generates a gap analysis: what additional research or evidence would most strengthen the argument
Detailed Section Outline Generation: COCO produces a comprehensive section-by-section outline:
- Generates a full document outline with section titles, sub-section structure, and paragraph-level content guidance
- Specifies the purpose of each section in the argumentative sequence — what the reader should believe after each section
- Assigns evidence to specific sections and sub-sections with guidance on how to use each piece
- Provides transition logic between sections — how to connect the conclusion of one section to the opening of the next
- Suggests appropriate section lengths based on the complexity of the argument and typical reading tolerance for the intended audience
Counter-Argument Handling Strategy: COCO designs the document's response to opposition:
- Identifies the three to five most likely objections to the central thesis from a skeptical reader
- Determines the optimal placement for each counterargument in the document
- Generates a counterargument + refutation structure for each identified objection
- Distinguishes between counterarguments that should be acknowledged and addressed vs. briefly dismissed
- Ensures the document does not strawman the opposition — addressing the strongest version of each counterargument
Introduction and Conclusion Architecture: COCO designs the most persuasively critical sections:
- Generates an introduction architecture: problem framing, stakes articulation, gap identification, thesis statement, and roadmap
- Designs an executive summary that captures the full argument in 300–500 words
- Generates conclusion architecture: restatement of thesis with accumulated evidence, implications, and call to action
- Ensures the conclusion delivers on the promise of the introduction without merely restating it
- Creates a key takeaways summary formatted for readers who skim to the end
Writing Guidance and Style Framework: COCO provides section-level writing direction:
- Generates a writing brief for each section: objective, key claims, evidence to use, length target, and tone guidance
- Identifies the appropriate level of technical language for the intended audience and provides vocabulary guidance
- Generates suggested headline phrasings for each section that communicate the argumentative point rather than merely describing the content
- Provides citation placement guidance: where to cite heavily, where to lead with the argument and cite for reference
- Generates a before-you-write checklist for each section to ensure structural coherence before drafting begins
Results & Who Benefits
Measurable Results
- White paper planning time: Reduced from 2–4 days of outline development to 3–6 hours with COCO-assisted architecture (70–80% time savings)
- Structural revision cycles: White papers developed with COCO architecture require 65% fewer structural revisions at the editing stage
- Argument completeness: Counterargument coverage in COCO-structured white papers averages 4.2 objections addressed vs. 1.8 in unstructured drafts
- Evidence-to-claim alignment: 94% of evidence in COCO-structured white papers is directly linked to a specific argumentative claim vs. 61% in unstructured workflows
- Publication rate: White papers developed with COCO argument architecture reach publication 3.5 weeks faster on average
Who Benefits
- Research Analysts: Produce persuasive white papers with a coherent argument architecture — not just comprehensive literature reviews — in significantly less time
- Policy Researchers: Build the argumentative structure required to influence policy debates, ensuring evidence is presented in a sequence designed to persuade decision-makers
- Thought Leadership Teams: Consistently produce white papers that advance a clear thesis rather than merely synthesizing existing literature
- Junior Analysts: Learn structured argument architecture by working through a COCO-generated outline — building a skill that improves all their written work over time
💡 Practical Prompts
Prompt 1 — White Paper Argument Architecture
Design the argument architecture for the following white paper.
Working thesis: [state the central argument you want the white paper to make]
Target audience: [who will read this — their role, sophistication level, and likely prior beliefs]
Intended outcome: [what should readers do or believe differently after reading this]
Available evidence: [briefly describe the research, data, and examples available to support the argument]
Length target: [number of pages or words]
Generate:
1. Core argument tree: the logical chain from premise to conclusion, with each step mapped
2. Claims to establish: the specific assertions that must be supported for the conclusion to follow
3. Evidence assignment: which evidence best supports which claim
4. Counterargument map: the 3–5 most likely objections and where to address each in the document
5. Section structure: a section outline showing how the argument unfolds, with the purpose of each section
6. Introduction and conclusion architecture: the key elements each must contain
Produce two alternative argument sequences for analyst review if the evidence supports different orderings.Prompt 2 — Detailed Section Outline
Generate a detailed section-by-section outline for the following white paper.
White paper topic: [title or description]
Thesis: [central argument]
Target audience: [description]
Approved argument architecture: [paste the agreed argument structure]
Evidence available: [list key data points, research findings, and examples available]
For each section, generate:
1. Section title (argumentative, not descriptive)
2. Purpose: what the reader should believe after this section that they did not before
3. Sub-sections: 3–5 sub-sections with content guidance for each
4. Evidence to deploy: which specific evidence goes where and how to present it
5. Length target: approximate word count for the section
6. Transition to next section: the connecting idea that links this section's conclusion to the next section's opening
Include specific guidance for:
- Executive summary structure (if requested)
- Introduction: problem framing, stakes, gap, thesis, roadmap
- Conclusion: synthesis, implications, call to actionPrompt 3 — Counterargument Strategy Development
Develop a counterargument handling strategy for the following white paper.
Central thesis: [state the argument]
Target audience's likely prior beliefs: [what do readers currently believe that this argument challenges]
Known objections or alternative views in the literature: [list any you are aware of]
For the audience profile described, generate:
1. The 5 most likely objections a skeptical reader would raise against this thesis
2. The strongest version of each objection (steelman, not strawman)
3. Recommended placement in the document: where to address each objection for maximum persuasive effect
4. Refutation strategy: the evidence and reasoning to use in responding to each objection
5. Tone guidance: should each objection be acknowledged and conceded in part, reframed, or directly refuted
6. Preemptive concession strategy: where acknowledging a limitation of the thesis actually strengthens credibility
Generate draft language for the two most important counterargument sections.Prompt 4 — Executive Summary Architecture
Write the executive summary for the following white paper.
White paper title: [title]
Central thesis: [argument]
Key supporting evidence: [list 3–5 most important data points or findings]
Target reader: [senior executive / policy maker / industry professional / investor]
Length target: [250 / 500 / 750 words]
The executive summary must:
1. Open with the problem or opportunity — why this topic matters now
2. State the thesis clearly in the first 2–3 sentences
3. Present the 3 strongest pieces of supporting evidence in decreasing order of persuasive impact
4. Acknowledge the most significant objection and briefly address it
5. State the implications: what readers should think, do, or decide differently
6. Close with a call to action or forward-looking statement
Write two versions: one for a skeptical audience that leads with counterargument acknowledgment, one for a receptive audience that leads with opportunity framing.Prompt 5 — Writing Brief Generation
Generate writing briefs for each section of the following white paper outline.
White paper: [title]
Outline sections: [list all sections and sub-sections with their purpose statements]
Author profile: [describe the analyst who will write this — experience level, writing strengths and weaknesses]
Style guidelines: [formal academic / accessible professional / policy brief / industry report]
For each section generate a writing brief containing:
1. Section objective: what argumentative work this section must do
2. Opening sentence options: 2–3 suggested ways to start the section
3. Paragraph structure: how many paragraphs, what each should accomplish
4. Evidence deployment: which specific evidence to use, where, and how to frame it
5. Technical language guidance: terms to use, terms to define, terms to avoid for this audience
6. Common mistakes to avoid: the typical errors in this type of section for this content type
7. Length and formatting guidance: word count target, sub-headers, bullet points yes/no
Format as a one-page brief per section that the analyst can reference while writing.18. AI Investment Thesis Research Package Compiler
Compiles a complete investment thesis research package — market, competitive, financial, and risk analysis — structured for investment committee review.
Pain Point & How COCO Solves It
The Pain: Investment Thesis Packages Are Resource-Intensive to Compile and Quality-Inconsistent Across Analysts
An investment thesis package is one of the highest-stakes research deliverables a research analyst produces. Whether for a private equity investment committee, a venture fund partner meeting, or a public market investment process, the package must integrate market sizing, competitive landscape, company-specific analysis, financial modeling assumptions, and risk assessment into a coherent, defensible narrative. Building this package from scratch for each investment requires weeks of analyst time — and the quality of the resulting package depends heavily on the experience and methodology of the individual analyst assembling it. Junior analysts produce thin competitive analysis or misstructured market sizing; senior analysts spend cycle time reviewing and rebuilding sections that should not have required their attention.
The research integration problem is specific to investment thesis work. Unlike a market research report, which can present findings topic-by-topic, an investment thesis must tell a coherent story: why this market, why now, why this company is positioned to win, what could go wrong and why it won't, and why the return profile justifies the risk. Each research component must contribute to the narrative arc, and the connections between components — how the market dynamics explain the company's competitive advantage, how the risk factors are mitigated by specific company characteristics — are as important as the components themselves. Building these connections requires both research skill and investment analysis judgment that is difficult to replicate systematically across an analyst team.
The research breadth problem compounds the integration challenge. A complete investment thesis package spans multiple research domains: market research, competitive intelligence, company-specific analysis, management assessment, financial analysis, regulatory research, and macro context. Each domain has its own sources, methodologies, and quality standards. Assembling a team with depth in all these areas for each investment is expensive; asking one analyst to cover all domains produces uneven quality. The result is investment thesis packages that are deep in some areas and superficial in others — with the superficial areas often corresponding to the domains that most clearly expose the risks to the thesis.
How COCO Solves It
Investment Thesis Architecture and Narrative Design: COCO builds the structural framework for the full package:
- Generates a thesis statement that integrates market, company, and return logic into a coherent, falsifiable argument
- Designs the package structure optimized for the intended audience: investment committee, LP update, or analyst presentation
- Creates a thesis narrative arc: why the market opportunity is real, why the company can capture it, why now is the right moment, and why the return profile is justified
- Identifies the critical assumptions that the thesis depends on and structures the package around validating them
- Generates the "if we are wrong" section structure: what would have to be true for the thesis to fail, and why that is unlikely
Market Opportunity Research and Sizing: COCO builds the market case with full methodology:
- Constructs TAM/SAM/SOM analysis from multiple sources with documented methodology
- Analyzes market growth drivers with CAGR triangulation across multiple forecast sources
- Assesses market timing: what structural or catalytic factors make now the right moment for this investment
- Maps the competitive landscape and market structure dynamics
- Generates the market opportunity section with investor-appropriate narrative and visualization
Competitive Position and Moat Analysis: COCO assesses the company's competitive defensibility:
- Analyzes competitive advantages across multiple dimensions: network effects, switching costs, cost advantages, IP, regulatory position
- Assesses the durability of each competitive advantage over the investment horizon
- Maps the competitive landscape and assesses the threat from each significant competitor
- Identifies the key competitive risks — where the company's position is most vulnerable
- Generates a competitive moat assessment using a structured framework appropriate for the industry and stage
Financial Analysis and Return Modeling: COCO builds financial assumptions and return logic:
- Synthesizes publicly available financial data with industry benchmark analysis to construct a financial profile
- Generates revenue model assumptions with market share, pricing, and growth rate logic tied to the thesis narrative
- Builds bear/base/bull case financial projections with scenario-specific assumption sets
- Calculates return metrics (IRR, MOIC) under each scenario with sensitivity to key value drivers
- Identifies the two or three most important value creation levers and quantifies their impact on return
Risk Assessment and Mitigation Framework: COCO structures a comprehensive risk register:
- Generates a risk taxonomy covering market risk, competitive risk, execution risk, regulatory risk, and macro risk
- Assesses each risk on probability, magnitude, and mitigation quality
- Identifies the two or three risks that would most credibly threaten the thesis if they materialized
- Documents specific mitigating factors for each high-priority risk
- Generates a monitoring framework: what to track post-investment to identify if key risks are materializing
Investment Committee Presentation Package: COCO produces complete, formatted deliverables:
- Generates an executive investment memo in the format expected by the investment committee
- Produces supporting slide deck sections for each component of the thesis
- Creates a due diligence checklist documenting what research has been completed and what remains outstanding
- Generates a data room index linking each claim in the memo to its supporting documentation
- Produces a Q&A preparation document anticipating the most likely investment committee challenges
Results & Who Benefits
Measurable Results
- Investment thesis package production time: Reduced from 3–5 analyst weeks to 1–2 weeks for a full package (60–70% time savings)
- Package completeness rate: Proportion of packages with all required sections fully developed increases from 55% to 92% with COCO systematic compilation
- Thesis consistency: Internal consistency audits find material inconsistencies between package sections in 8% of COCO-compiled packages vs. 37% of manually assembled packages
- Investment committee challenge preparation: Analysts using COCO's Q&A preparation document address 85% of investment committee questions in their initial presentation vs. 55% without preparation support
- Junior analyst thesis quality: Junior analysts produce packages rated "investment-committee-ready without major revision" in 71% of cases with COCO vs. 28% without
Who Benefits
- Investment Research Analysts: Produce complete, well-integrated investment thesis packages in half the time — with COCO ensuring no research domain is thin or absent
- Private Equity and Venture Capital Associates: Deliver investment committee packages that hold up to senior partner scrutiny — with every section coherently connected to the thesis narrative
- Sell-Side Equity Research Analysts: Build initiation-of-coverage research packages with the depth and integration quality that distinguishes differentiated research from commodity coverage
- Investment Committee Members: Receive packages where the logic is internally consistent and the risk section actually addresses the most threatening risks
💡 Practical Prompts
Prompt 1 — Investment Thesis Statement and Architecture
Develop an investment thesis statement and package architecture for the following opportunity.
Investment target: [company name or description]
Stage: [early stage / growth / buyout / public market]
Investment type: [equity / debt / real assets]
Target investor: [VC / PE / hedge fund / family office / institutional asset manager]
Working thesis (if any): [your current view on why this is a good investment, or "help me develop one"]
Available information: [describe what research and company information you have access to]
Generate:
1. Thesis statement: a 2–3 sentence investment thesis that integrates market, competitive, and return logic
2. Critical assumptions: the 5–7 things that must be true for this thesis to work
3. Package structure: the sections the investment package should include, in order that best builds the argument
4. Thesis narrative arc: a prose description of how the package tells the investment story
5. "If we're wrong" analysis: the 3 most likely ways the thesis fails and why each is manageable or unlikely
Output: Thesis architecture document to guide the full research package developmentPrompt 2 — Competitive Moat Analysis
Conduct a competitive moat analysis for the following company.
Company: [name and brief description]
Industry: [sector]
Investment context: [describe the investment thesis you are evaluating]
Competitive advantages claimed: [list what the company or management says makes them defensible]
Assess each potential moat:
1. Network effects: does the product become more valuable as more users join, and how strong is the network effect
2. Switching costs: how costly (financially, operationally, strategically) is it for customers to switch away
3. Cost advantages: does the company have structural cost advantages from scale, proprietary processes, or favorable contracts
4. IP and technology barriers: are there patents, trade secrets, or technical capabilities that competitors cannot easily replicate
5. Regulatory and licensing advantages: does a license, certification, or regulatory relationship create barriers to entry
6. Brand and trust advantages: does brand loyalty meaningfully reduce customer acquisition costs or enable pricing premium
For each advantage:
- Evidence of its existence (specific data or facts)
- Durability assessment: how sustainable is this advantage over the investment horizon
- Competitive challenge: what would it take for a competitor to overcome this advantage
Overall moat rating: wide / narrow / uncertain with investment implicationPrompt 3 — Bear / Base / Bull Case Financial Scenario
Build bear, base, and bull case financial scenarios for the following investment.
Company: [name]
Financial period: [next 3 / 5 years]
Key financial metrics to project: [revenue, gross margin, EBITDA, FCF, or other relevant metrics]
Current financial data (if available): [paste or describe]
Business model: [describe revenue model — SaaS / transactional / subscription / product sales / etc.]
For each scenario, define coherent assumption sets for:
- Revenue growth: what drives growth and what is the realistic rate in each scenario
- Gross margin: what happens to pricing power and COGS structure
- Operating leverage: how do fixed and variable costs behave as revenue scales
- Capital requirements: what investment is required to fund the projected growth
For each scenario:
1. Probability: bear [%] / base [%] / bull [%]
2. Key assumption differences: what is different between scenarios (1–3 key variables)
3. Financial projections: 3–5 year model with annual figures
4. Return metrics: IRR and MOIC at the assumed entry and exit multiple
5. Sensitivity analysis: how return changes for ±1 turn of exit multiple and ±10% EBITDA
Output: Three-scenario financial model + return attribution analysisPrompt 4 — Investment Risk Register
Build a comprehensive investment risk register for the following thesis.
Investment: [company name and thesis summary]
Investment horizon: [expected holding period]
Investment size and structure: [describe]
Risk categories to assess:
1. Market risk: market growth falls short of assumptions, market structure changes adversely
2. Competitive risk: new entrant, competitor product improvement, pricing pressure
3. Execution risk: management fails to deliver on operating plan, key talent attrition
4. Customer concentration risk: key customer loss or deteriorating relationship
5. Technology risk: technological disruption or product obsolescence
6. Regulatory risk: adverse regulatory change affecting business model or market access
7. Macro risk: recession, inflation, interest rate environment effects
8. Exit risk: limited exit options or exit multiple compression at realization
For each risk:
- Probability: high / medium / low with brief rationale
- Impact if materialized: on value, return metrics, or thesis viability
- Mitigating factors: specific characteristics of the company, market, or deal structure that reduce this risk
- Monitoring indicators: what to track post-investment to identify early signs of this risk materializing
Output: Risk register + risk heat map (probability × impact matrix) + top 3 risks warranting most attentionPrompt 5 — Investment Committee Q&A Preparation
Generate an investment committee Q&A preparation document for the following investment thesis.
Thesis summary: [describe the investment and central thesis]
Investment committee profile: [describe who will be in the room — experience, known skepticisms, past concerns on similar investments]
Most controversial aspects of the thesis: [identify where the committee is most likely to push back]
Generate:
1. The 20 most likely investment committee questions, organized by theme
2. For each question:
- The strongest answer, with supporting data or evidence
- Where to find supporting detail in the investment package
- Follow-up questions that may arise and their answers
3. Questions we cannot fully answer yet: what we do not know, and our plan for getting to an answer
4. Recommended presentation adjustments: based on anticipated questions, what to emphasize or de-emphasize
Categorize questions by difficulty:
- Routine (answerable with standard data and analysis)
- Challenging (require nuanced response and strong evidence)
- Sensitive (require honest acknowledgment of uncertainty while maintaining conviction)
Format: Q&A preparation brief organized by topic with confidence ratings for each answer19. AI Qualitative Interview Transcript Coder
Applies systematic qualitative coding to interview transcripts — identifying themes, patterns, and analytical categories — at a fraction of the time of manual coding.
Pain Point & How COCO Solves It
The Pain: Qualitative Data Analysis Is the Richest and Most Labor-Intensive Step in Primary Research
Qualitative interview research generates some of the most strategically valuable insights available to research analysts — the nuanced explanations, unexpected perspectives, and rich context that survey data cannot capture. But the step between collecting interview data and producing usable findings is one of the most time-intensive and methodologically demanding in all of research. Qualitative coding — the systematic process of reading transcripts, identifying themes, labeling patterns, and building a coding framework that enables cross-interview analysis — requires a combination of disciplined methodology and interpretive judgment that cannot be shortcut without compromising research quality. A single hour-long interview transcript can take three to five hours to code thoroughly using established qualitative research methods.
The consistency problem is the most damaging practical limitation of manual coding. When two analysts code the same transcript independently and their codes diverge significantly — a common outcome — it signals that the coding framework is not precise enough, or that analyst interpretation is too variable, to support reliable conclusions. In most business research contexts, there is neither the time nor the team size to conduct formal intercoder reliability testing. Each analyst codes the transcripts assigned to them using their own interpretive framework, and the resulting codes are assembled into findings that appear systematic but are actually highly analyst-dependent. Clients receive conclusions presented as grounded in qualitative data but in practice shaped by the interpretive lens of whoever happened to code those particular transcripts.
The synthesis gap follows directly from the coding quality problem. Even when individual transcripts are coded well, moving from coded data to findings requires a further analytical step that most research processes handle poorly: identifying which codes are most prevalent, which co-occur with each other, which differ by respondent type, and which represent the most strategically significant patterns. This analysis layer is the one that most directly produces research value, but it is the layer most often abbreviated under time pressure — producing conclusions that summarize what respondents said without fully extracting the patterns that give the data its strategic meaning.
How COCO Solves It
Inductive Code Development: COCO builds grounded theory-aligned coding frameworks from the data:
- Reads the full corpus of transcripts and generates an initial coding framework derived from the data rather than imposed on it
- Identifies recurring themes, concepts, and patterns across multiple transcripts through iterative reading
- Groups related codes into higher-order categories and sub-categories in a hierarchical coding tree
- Labels codes with descriptive names that reflect the content precisely rather than the analyst's theoretical framing
- Presents the initial coding framework to the analyst for review and refinement before applying it to all transcripts
Deductive Code Application: COCO applies researcher-specified frameworks to transcript data:
- Accepts pre-defined coding frameworks (theoretical frameworks, client-specified categories, prior research dimensions) and applies them consistently to all transcripts
- Identifies which portions of each transcript map to each specified code
- Flags transcript content that does not fit the specified coding framework — potential new categories not anticipated in the design
- Calculates code frequency and distribution across transcripts and respondent types
- Generates a code application report showing coverage and consistency across the corpus
Cross-Transcript Pattern Analysis: COCO identifies patterns at the corpus level:
- Calculates code frequency across all transcripts — which themes appear most consistently
- Identifies code co-occurrence patterns — which themes tend to appear together in the same transcripts
- Analyzes variation by respondent type — where different expert profiles or experience levels produce different coded content
- Identifies outlier transcripts — respondents whose coded content differs significantly from the majority pattern
- Generates a pattern matrix showing theme presence/absence or frequency across each transcript
Negative Case and Contradiction Analysis: COCO ensures the full range of perspectives is captured:
- Identifies transcripts where the dominant theme is contradicted or qualified significantly
- Flags cases where a respondent's coded content contradicts the emerging finding
- Generates a negative case analysis for each major finding: what percentage of respondents contradict or qualify the finding, and what characterizes those respondents
- Ensures findings are qualified appropriately — acknowledging where meaningful minorities hold different views
- Identifies areas of genuine disagreement among respondents that require reporting as contested rather than settled findings
Quotation Selection and Evidence Assembly: COCO builds the evidentiary foundation for findings:
- Selects the most illustrative quotations for each theme — prioritizing richness, specificity, and representative typicality
- Identifies exemplar quotes that capture a theme with particular clarity and efficiency
- Generates a quotation register organized by theme with transcript attribution for each quote
- Ensures quotation balance across respondent types — not all supporting quotes from the same type of expert
- Produces a quotation management framework that tracks quote usage to avoid over-citing any single respondent
Qualitative Findings Report Generation: COCO produces structured deliverables from coded data:
- Generates a findings narrative for each major theme with supporting quotations and cross-transcript pattern evidence
- Produces a thematic frequency table showing code presence and prevalence across the transcript corpus
- Creates a respondent-type comparison showing how different expert profiles addressed each theme
- Generates an analytical memo documenting the coding process, framework decisions, and analytical reasoning
- Produces a synthesis report integrating qualitative findings with quantitative data from the same research project
Results & Who Benefits
Measurable Results
- Transcript coding time: Reduced from 3–5 hours per transcript to 20–40 minutes of review for a COCO-coded transcript (85–90% time savings)
- Intercoder reliability: COCO-coded transcripts achieve average agreement rates of 82–88% with expert human coders — within the acceptable range for published qualitative research
- Theme identification completeness: COCO identifies an average of 4–6 more secondary themes per transcript corpus than typical time-pressured manual coding
- Synthesis cycle time: Moving from coded data to structured qualitative findings takes 2–3 days with COCO vs. 1–2 weeks manually
- Negative case coverage: Proportion of qualitative findings including negative case analysis increases from 18% to 87% when COCO automates the contradiction-checking process
Who Benefits
- Research Analysts: Conduct rigorous qualitative analysis on full transcript corpora rather than selectively reading a subset — and produce findings grounded in the full data
- Market Research Teams: Deliver qualitative insight projects faster and with more methodological defensibility — enabling more competitive pricing and more confident client presentation
- Academic Researchers: Apply systematic coding to large transcript corpora without the months of manual coding that typically limits the scale of qualitative research
- Strategy and Innovation Teams: Access the depth of qualitative insight from customer and expert interviews without sacrificing the analytical rigor needed for strategic decision-making
💡 Practical Prompts
Prompt 1 — Inductive Coding Framework Development
Develop an inductive coding framework for the following qualitative interview corpus.
Research topic: [what the interviews were about]
Number of interviews: [N]
Respondent types: [describe who was interviewed]
Research questions the coding should address:
1. [Research question 1]
2. [Research question 2]
3. [Research question 3]
Sample transcripts (paste 2–3 representative transcripts or substantial excerpts):
[Paste transcript content]
Generate:
1. An initial coding framework with:
- Top-level themes (5–10 major categories)
- Sub-codes within each theme (3–5 per theme)
- Code definitions: precise description of what each code covers and does not cover
- Illustrative example: a sample transcript passage that would receive each code
2. A coding decision tree: rules for how to handle ambiguous or multi-theme passages
3. A list of codes to watch for: themes that may emerge but are not yet in the framework
Format as a codebook ready for application to all transcripts.Prompt 2 — Single Transcript Coding
Code the following interview transcript using the specified coding framework.
Coding framework:
[Paste or describe the coding framework — themes, sub-codes, and definitions]
Transcript:
[Paste the interview transcript]
Apply the coding framework to identify:
1. All passages relevant to each code, with the exact quote and a brief explanation of why this code applies
2. Any content that does not fit the existing framework (potential new codes)
3. Passages where multiple codes apply simultaneously (co-occurrence)
4. Notable absences: topics in the coding framework that this respondent did not address
Output format:
- Code-by-code organization: for each code, list all relevant passages with attribution
- Chronological coding log: codes applied in order of appearance in the transcript
- Respondent profile: a brief characterization of this respondent's overall perspective based on the coded contentPrompt 3 — Cross-Transcript Pattern Analysis
Analyze patterns across the following set of coded interview transcripts.
Research topic: [description]
Number of transcripts: [N]
Coded transcript data: [paste coding logs or theme-by-transcript presence/absence data]
Respondent types: [describe the different respondent profiles represented]
Analyze:
1. Theme frequency: how often does each theme appear across the corpus (% of transcripts)
2. Theme salience: beyond frequency, which themes receive the most extensive discussion
3. Co-occurrence patterns: which themes consistently appear together in the same transcripts
4. Respondent type variation: how does theme frequency differ across respondent types
5. Outlier analysis: are there transcripts where the coding pattern differs significantly from the majority
Generate:
1. Pattern matrix: themes × transcripts showing presence/absence or frequency
2. Key finding summaries: for each major theme, a 2–3 sentence synthesis of what the pattern shows
3. Divergence analysis: where different respondent types see the issue differently and why
4. Quotation selection: for each major finding, the 2–3 most illustrative quotations from the corpusPrompt 4 — Negative Case and Contradiction Analysis
Conduct a negative case analysis for the following qualitative findings.
Emerging findings:
1. [Finding 1: what the majority of respondents said about topic X]
2. [Finding 2]
3. [Finding 3]
Coded transcript data: [paste or describe the coding data showing which transcripts support and contradict each finding]
For each finding:
1. Proportion of transcripts supporting the finding vs. contradicting or qualifying it
2. Who contradicts: what characterizes the respondents who do not support this finding
3. Nature of contradiction: do they say the opposite, or do they add important qualifications
4. Analytical interpretation: does the contradiction undermine the finding, refine it, or reveal a segment-specific pattern
5. Recommended language: how to report this finding accurately given the contradictions
Generate a negative case section for each major finding suitable for inclusion in the research report.Prompt 5 — Qualitative Findings Synthesis
Generate a qualitative findings synthesis from the following coded interview data.
Research context: [study objective, number of interviews, respondent types]
Major themes identified: [list the main codes and their frequency data]
Supporting quotations: [paste the most illustrative quotes for each theme]
Research questions to answer:
1. [Research question 1]
2. [Research question 2]
3. [Research question 3]
Generate a findings section that:
1. Presents each major theme as a finding, with:
- A clear finding statement (what the data shows)
- The frequency evidence (how many respondents, what proportion)
- 2–3 illustrative quotations with attribution
- Any important variation by respondent type
- The negative case: what contradicts or qualifies the finding
2. Synthesizes across themes: how do the themes relate to each other
3. Answers each research question with direct reference to the coded evidence
Tone: Professionally rigorous, grounded in evidence, appropriately qualified where the data is mixed20. AI Macro-Economic Indicator Dashboard Builder
Builds a customized macro-economic monitoring dashboard with the specific indicators, data sources, and analytical commentary most relevant to the research coverage universe.
Pain Point & How COCO Solves It
The Pain: Macro-Economic Monitoring Is Universally Required but Rarely Done With the Rigor the Task Deserves
Every serious research analyst monitors macro-economic conditions — interest rates, inflation, GDP growth, employment, trade flows, and sector-specific indicators that affect the industries they cover. But macro-economic monitoring is done with widely varying rigor across analyst teams. At one end, analysts follow a handful of headline indicators through Bloomberg or news feeds, forming impressionistic views of economic conditions. At the other end, dedicated economists build sophisticated dashboards tracking hundreds of indicators across multiple frequencies, with formal nowcasting models and scenario probabilities. Most research analysts operate in the vast middle ground: aware that macro matters, uncertain which indicators matter most for their specific coverage, and without the time or analytical infrastructure to monitor systematically.
The indicator selection problem is the first barrier. There are thousands of potentially relevant economic indicators across government agencies, central banks, and private data providers. The indicators that matter most vary significantly by coverage sector: a consumer research analyst needs different indicators than a capital goods analyst or a technology sector specialist. Generic economic dashboards track the same headline indicators for every user — creating monitoring noise for each individual user and missing the sector-specific leading indicators that most directly predict conditions in their coverage area. Building a customized indicator set for a specific coverage universe requires economic expertise and time that most analysts do not have.
The interpretation problem is as significant as the selection problem. Macro-economic indicator data is objective; its implications for a specific investment thesis or sector outlook are not. Whether a specific inflation reading is bullish or bearish for a technology software company depends on its pass-through to input costs, its effect on customer IT budget approvals, and the Federal Reserve's likely response — a chain of interpretation that requires both macroeconomic literacy and sector-specific knowledge simultaneously. Analysts who track the data without a framework for interpreting its implications produce macro sections that describe economic conditions without drawing defensible conclusions about what those conditions mean for their coverage.
How COCO Solves It
Coverage-Specific Indicator Selection: COCO identifies the indicators most relevant to the specific research context:
- Analyzes the coverage universe and recommends indicators based on established empirical relationships between economic variables and sector performance
- Distinguishes leading indicators (predict future conditions), coincident indicators (reflect current conditions), and lagging indicators (confirm past trends) for each coverage area
- Identifies sector-specific indicators beyond headline macro: industry capacity utilization, sector-level PMIs, sector credit spreads, sector-specific price indices
- Recommends data frequencies appropriate to the research cycle: monthly for slow-moving trends, weekly or high-frequency for more volatile indicators
- Builds a tiered indicator set: core indicators to monitor always, secondary indicators to watch in specific scenarios, and watchlist indicators currently approaching significance
Multi-Source Economic Data Integration: COCO aggregates data from the full range of economic data providers:
- Integrates data from Federal Reserve (FRED), Bureau of Economic Analysis, Bureau of Labor Statistics, Census Bureau, and Treasury for US indicators
- Incorporates international data from IMF, World Bank, OECD, and national statistical agencies for global coverage
- Adds private sector data sources: ISM PMI, Conference Board Leading Index, University of Michigan Consumer Sentiment, NFIB Small Business Optimism
- Integrates financial market indicators: yield curves, credit spreads, volatility indices, commodity prices, and currency movements
- Normalizes indicators to consistent time scales, units, and seasonal adjustment status for clean comparison
Economic Cycle Assessment and Positioning: COCO provides structured economic cycle analysis:
- Applies established economic cycle frameworks (NBER expansion/recession, ISM cycle, yield curve signals) to assess current positioning
- Generates a probability-weighted economic scenario assessment: growth/slowdown/recession/recovery probabilities with supporting indicator evidence
- Tracks leading indicator composite scores and their historical reliability for the coverage sector
- Identifies divergences between indicators that historically precede inflection points
- Produces a one-page economic cycle summary with current positioning and directional assessment for each scenario
Sector-Specific Macro Transmission Analysis: COCO connects macro conditions to coverage-specific implications:
- Maps each tracked macro indicator to its historical impact on sector performance metrics (revenue growth, margins, valuations)
- Quantifies the transmission elasticity: how much does a 1% change in indicator X historically affect metric Y in the coverage sector
- Identifies leading vs. lagging relationships — how many months ahead do changes in specific indicators predict sector outcomes
- Generates a macro sensitivity profile for the coverage sector showing which indicators matter most and how
- Produces a macro tailwinds and headwinds summary for the current environment with sector-specific implications
Nowcasting and Short-Term Outlook: COCO builds forward-looking economic assessment:
- Aggregates consensus forecasts from major economic institutions (Fed, IMF, OECD, major banks) and presents the range
- Applies simple nowcasting logic — using high-frequency data to estimate current-quarter metrics before official data release
- Identifies upcoming economic data releases with their expected market and sector impact
- Generates a monthly macro outlook brief: current reading, trend direction, and 3-month outlook for each key indicator
- Tracks forecast revision trends — whether economists are revising estimates up or down — as a directional signal
Macro Dashboard and Report Generation: COCO produces structured, formatted macro monitoring deliverables:
- Generates a customized macro dashboard layout with specified indicators, time periods, and visualization formats
- Produces a monthly macro commentary with executive summary, key developments, and sector implications
- Creates a macro risk assessment: identifying the economic scenarios most threatening to the investment thesis or research outlook
- Generates data export formats for integration into financial models and research reports
- Produces audience-specific macro summaries: technical detail for analyst audiences, implication-focused summaries for clients
Results & Who Benefits
Measurable Results
- Dashboard setup time: Reduced from 2–3 weeks of indicator selection and data sourcing to 1–2 days for a fully configured coverage-specific macro dashboard (85% time savings)
- Indicator coverage: COCO-configured dashboards track an average of 35–45 indicators vs. 8–12 in analyst-default monitoring setups — with higher specificity to the coverage sector
- Macro interpretation quality: Research reports with COCO macro sections receive "adequate sector-specific interpretation" ratings in 88% of cases vs. 41% for reports with generic macro summaries
- Early warning signal capture: Analysts using COCO's leading indicator monitoring identify economic inflection points an average of 2–3 months earlier than those tracking only headline indicators
- Macro commentary production time: Monthly macro commentary sections produced in under 2 hours vs. full analyst day previously
Who Benefits
- Research Analysts: Maintain sophisticated, coverage-relevant macro monitoring without being dedicated economists — COCO handles indicator selection, data integration, and transmission analysis
- Equity Research Analysts: Incorporate macro context into sector and company analysis with the specificity needed to draw defensible investment implications rather than generic economic observation
- Strategy Consultants: Access current, well-structured macro analysis for client engagements without pulling analyst resources from billable project work
- Investment Portfolio Managers: Monitor macro conditions with a coverage-specific lens — tracking the indicators that most directly drive portfolio performance rather than headline news flow
💡 Practical Prompts
Prompt 1 — Coverage-Specific Macro Dashboard Design
Design a macro-economic monitoring dashboard for the following coverage universe.
Coverage: [describe the sector, geography, or investment focus]
Primary research questions the dashboard should inform: [e.g., demand outlook, input cost trends, credit availability, consumer spending capacity]
Monitoring frequency: [daily / weekly / monthly]
Data access: [Bloomberg / FRED / public sources only / specify]
Design the dashboard with:
1. Core indicators (monitor always): 8–12 indicators with highest relevance to the coverage, with data source and update frequency for each
2. Secondary indicators (context): 5–8 indicators for broader economic context
3. Watchlist indicators (monitor when conditions warrant): 5–8 indicators to activate in specific scenarios
4. Leading vs. lagging classification for each indicator
5. Historical relationship to coverage performance: how each core indicator has historically correlated with sector or portfolio performance
6. Alert thresholds: at what reading level should each indicator trigger an analytical review
Output: Dashboard specification document + data sourcing guidePrompt 2 — Economic Cycle Assessment
Assess the current position in the economic cycle and its implications for the following coverage.
Coverage universe: [sector or portfolio focus]
Current economic data (paste key indicators or describe what you are seeing):
[List key economic readings — GDP growth, unemployment, inflation, ISM PMI, yield curve, etc.]
Assess:
1. Cycle positioning: where are we in the economic cycle (early expansion / mid-cycle / late cycle / recession / recovery)
2. Evidence: which indicators support this assessment and which are ambiguous or contradicting
3. Transition probability: what is the probability of moving to the next cycle phase in the next 6–12 months
4. Leading indicator signals: are leading indicators confirming the current phase or signaling a transition
5. Historical analogue: when was the last time conditions were most similar and what happened to the coverage sector
Coverage implications:
1. How has the coverage sector historically performed in the current cycle phase
2. What are the most important macro tailwinds and headwinds in the current environment
3. What macro scenario would be most damaging to the coverage and what signals would indicate it is approachingPrompt 3 — Macro Sensitivity Analysis for Sector
Build a macro sensitivity analysis for the following sector.
Sector: [industry]
Key financial metrics to analyze: [revenue growth / gross margins / EBITDA / valuations / credit spreads]
Macro variables to test: [interest rates, inflation, GDP growth, unemployment, currency, commodity prices]
Historical data period: [how many years of historical data to analyze]
For each macro variable:
1. Historical correlation with each key metric: direction and magnitude over the historical period
2. Lead/lag relationship: does the macro variable lead or lag changes in the sector metric, and by how many months
3. Elasticity estimate: a 100bps change in [variable X] is historically associated with a [Y%] change in [metric Z]
4. Non-linearity: does the relationship hold consistently, or change in extreme environments
5. Current reading vs. historical range: where is the variable today relative to its historical distribution
Synthesize: a macro sensitivity table showing the sector's exposure to each variable and the current directional implicationPrompt 4 — Monthly Macro Commentary
Write a monthly macro commentary for the following research coverage.
Coverage: [sector or portfolio focus]
Current month: [month and year]
Key economic releases this month:
[List major data releases and their readings — e.g., CPI: +3.2% YoY, Non-farm payrolls: +180K, ISM Manufacturing PMI: 48.7]
Key market developments: [describe significant moves in rates, credit spreads, equity markets, currencies, commodities]
Write a commentary covering:
1. Executive summary: the 3 most important macro developments this month and their coverage implications (1 paragraph)
2. Growth indicators: current reading and trend for GDP proxies, employment, and leading indicators — with coverage implication
3. Inflation and pricing environment: current readings and trend — how does this affect input costs and pricing power in the coverage
4. Monetary policy: Fed/central bank stance, rate path expectations, and financial conditions
5. Sector-specific macro: any sector-specific data releases relevant to the coverage
6. Outlook for next month: key data releases to watch and their expected range of outcomes
Tone: Analyst-grade commentary, specific and well-evidenced, not generic economic observationPrompt 5 — Macro Risk Scenario Analysis
Build macro risk scenarios for the following investment thesis or research outlook.
Thesis or outlook: [describe the central investment thesis or research view you want to stress-test]
Current macro assumptions embedded in the thesis: [describe the implicit or explicit macro assumptions]
Build 3 macro risk scenarios that would most threaten the thesis:
For each scenario:
1. Scenario description: what happens macroeconomically (be specific — not "recession" but "GDP contracts 1.5% over 2 quarters driven by consumer spending pullback")
2. Probability: estimate the probability over the relevant time horizon
3. Trigger indicators: what economic signals would indicate this scenario is becoming more likely
4. Impact on thesis: how does this macro scenario affect the key assumptions — quantify where possible
5. Thesis survivability: does the thesis hold, require modification, or break under this scenario
6. Monitoring approach: which 3–5 indicators to watch weekly to detect early movement toward this scenario
Output: Scenario risk table + monitoring dashboard specification for the identified risk scenarios21. AI Media Coverage & Press Monitoring Digest
Aggregates, filters, and synthesizes media coverage across hundreds of sources into a structured daily or weekly intelligence digest — eliminating noise and surfacing only the signals that matter.
Pain Point & How COCO Solves It
The Pain: Media Monitoring Produces Information Overload Without Intelligence
Research analysts are expected to maintain current awareness of media coverage for their coverage universe — tracking how companies, industries, and issues are being portrayed in the press and what new information enters the public domain through media channels. But the volume of media output has grown to the point where coverage monitoring is either superficial or all-consuming. A broad Google News alert covers too much and misses too much simultaneously; a narrow keyword filter produces a thin, incomplete picture; a comprehensive reading of all relevant outlets across print, digital, broadcast, and social media is simply not possible within a research workday. The result is that most analysts have an impressionistic, incomplete view of their media landscape — shaped primarily by the most prominent and most widely distributed stories rather than the most analytically significant ones.
The signal quality problem compounds the volume problem. Most media monitoring tools surface content by recency and keyword match, not by strategic relevance. A recycled wire story distributed by 40 different outlets generates 40 monitoring hits and appears more significant than a single well-sourced investigative piece that contains genuinely new information. Analysts who rely on unfiltered media monitoring develop a distorted picture of what is actually happening — overweighting widely covered stories and underweighting original reporting. The truly significant media signals — the first story about an emerging issue, the one outlet that got a company to comment on an undisclosed development, the regulatory leak that surfaces in a trade publication before mainstream coverage — are buried in the noise.
The synthesis gap is the third dimension of the problem. Even when analysts collect the right media signals, they rarely have time to synthesize them into a coherent intelligence picture. What is the aggregate narrative forming about a company or issue? Are multiple outlets converging on the same story, suggesting it has staying power? Is coverage tone shifting from positive to negative in ways that anticipate a stakeholder or regulatory response? These meta-level insights are the most valuable output of media monitoring, but extracting them from a pile of individual article bookmarks requires synthesis time that is chronically squeezed.
How COCO Solves It
Intelligent Source Selection and Coverage Mapping: COCO builds a coverage-relevant source universe:
- Maps media sources by type — tier-1 national press, trade publications, wire services, digital natives, broadcast transcripts, and newsletters — for the specific coverage universe
- Applies relevance weighting: tier-1 investigative outlets weighted higher than wire redistributors
- Identifies the most credible sources for each specific subject area — the publications that break original stories rather than aggregate them
- Monitors social platforms (LinkedIn, X/Twitter, industry forums) for pre-publication signals that precede formal media coverage
- Continuously updates the source list as new relevant outlets emerge and existing ones decline in quality or relevance
Relevance Filtering and Deduplication: COCO surfaces signal without noise:
- Applies semantic relevance scoring to filter articles by true relevance to the defined coverage universe — not just keyword presence
- Deduplicates wire story redistributions — grouping 40 identical stories as one signal rather than 40 hits
- Distinguishes original reporting (new information or sources) from aggregated or commentary content
- Applies a source credibility filter to separate high-signal sources from low-credibility noise in the monitoring stream
- Flags articles that contain new information not previously covered — the signals most likely to matter for research analysis
Coverage Tone and Sentiment Analysis: COCO tracks how narratives are shifting:
- Analyzes the overall tone of coverage for each tracked subject: positive, negative, neutral, mixed
- Tracks tone trends over time — identifying when coverage is becoming more or less favorable
- Distinguishes company-specific tone from sector-wide tone shifts
- Identifies narrative pivots: when coverage language shifts from one frame to another (from "growth story" to "execution concerns," for example)
- Generates a reputation risk signal when coverage tone deteriorates significantly or suddenly
Story Emergence and Trajectory Analysis: COCO identifies which stories will matter:
- Detects story emergence: when a topic appears for the first time across multiple sources, suggesting it is becoming a sustained story rather than a one-time mention
- Tracks story lifecycle: is a topic gaining coverage volume, maintaining steady coverage, or declining
- Identifies which outlets are driving a story versus which are following — first-mover outlets often have better sourcing
- Flags when a story crosses from trade press to mainstream coverage — typically a signal of increased stakeholder attention
- Generates a story trajectory forecast: based on coverage patterns, will this story grow, plateau, or fade in the next 2–4 weeks
Cross-Source Intelligence Synthesis: COCO builds coherent intelligence from fragmented coverage:
- Identifies when multiple outlets are converging on the same underlying story from different angles — a strong corroboration signal
- Extracts new factual information (quotes, figures, named sources) that appears across articles and consolidates into a clean intelligence brief
- Identifies contradictions between outlets and flags where the facts are disputed
- Connects media coverage to other signal types: when a media story and a regulatory filing or earnings transcript all point to the same development
- Generates a "media intelligence brief" synthesizing what the coverage collectively reveals about a subject
Structured Media Digest Generation: COCO produces formatted, time-efficient deliverables:
- Generates a daily or weekly media digest organized by subject, with each entry summarizing new developments from multiple sources
- Produces an executive media brief: top 5 stories, their significance, and strategic implications
- Creates a coverage analytics summary: volume trends, tone trends, and source distribution for the monitoring period
- Generates custom digest formats for different audiences: detailed analyst brief vs. concise executive summary vs. client-facing media update
- Produces an alert format for high-urgency stories requiring immediate analyst attention
Results & Who Benefits
Measurable Results
- Media monitoring time: Reduced from 1–2 hours daily of article reading to 15–20 minutes of digest review (75–85% time savings)
- Source coverage breadth: COCO monitors 150–300 sources per coverage universe vs. 20–40 in typical manual monitoring — with deduplication eliminating noise from the broader coverage
- Signal-to-noise ratio: Analysts report 80% reduction in irrelevant articles in their monitoring stream after COCO relevance filtering replaces keyword alerts
- Story emergence lead time: Organizations using COCO's story trajectory analysis identify developing media stories an average of 3–5 days earlier than those relying on reactive monitoring
- Narrative shift detection: COCO detects statistically significant coverage tone changes an average of 2 weeks earlier than analysts relying on informal impressionistic monitoring
Who Benefits
- Research Analysts: Stay current on their coverage universe without spending hours on article triage — COCO delivers the intelligence content of comprehensive monitoring in the time of a brief digest review
- Communications and IR Teams: Monitor how their company is being covered and identify narrative shifts before they become entrenched — enabling proactive communications responses
- Investment Analysts: Track the media narrative around portfolio companies and investment targets — identifying sentiment shifts that may anticipate price movements or stakeholder reactions
- Strategy and Reputation Teams: Maintain real-time situational awareness of how key issues are being framed in the media landscape most relevant to their stakeholders
💡 Practical Prompts
Prompt 1 — Daily Media Intelligence Digest
Generate a media intelligence digest for the following coverage universe.
Coverage subjects: [list companies, industries, issues, or people to monitor]
Monitoring period: [today / past 24 hours / past 7 days]
Source types to include: [trade press / national press / wire services / digital outlets / all]
Audience: [internal analyst team / senior management / client-facing]
For each subject, include:
1. New developments: information appearing for the first time in today's coverage (flag as "NEW")
2. Continuing stories: ongoing coverage of established stories with any new developments
3. Coverage volume: high / medium / low vs. normal baseline
4. Tone assessment: positive / neutral / negative / mixed — and any shift from prior period
5. Most significant article: title, outlet, and 2-sentence summary of the single most important piece today
Executive summary: top 3 stories across all subjects that warrant analyst attention today, with 1-sentence significance note for each.Prompt 2 — Coverage Tone Trend Analysis
Analyze the tone and narrative trajectory of media coverage for the following subject.
Subject: [company, industry, executive, or issue]
Time period: [last 30 / 60 / 90 / 180 days]
Source types to analyze: [specify or include all]
Analyze:
1. Tone distribution: what proportion of coverage is positive / neutral / negative / mixed over the period
2. Tone trend: is coverage becoming more positive, more negative, or stable over time
3. Tone drivers: what topics or events are generating the most negative / positive coverage
4. Narrative frames: what are the 2–3 dominant narratives in coverage (e.g., "growth story," "regulatory risk," "management concern")
5. Narrative shifts: have any frames emerged, strengthened, or faded during the analysis period
6. Outlet divergence: are different outlet types (trade vs. mainstream, domestic vs. international) telling different stories
Strategic implication: what does the coverage trajectory suggest about how stakeholders are likely to perceive this subject over the next 3–6 monthsPrompt 3 — Story Emergence Detection
Identify emerging media stories for the following coverage universe that may become significant.
Coverage universe: [describe the industry, company, or issue landscape to monitor]
Monitoring window: [last 7 / 14 / 30 days]
Significance threshold: [stories appearing in 3+ outlets / crossing from trade to mainstream / generating unusual volume]
For each emerging story detected:
1. Story description: what is the story about
2. First appearance: when and where did it first appear
3. Spread trajectory: how has coverage volume changed since first appearance
4. Outlet profile: which types of outlets are covering it — trade only, mainstream, or both
5. Corroboration level: are multiple independent outlets reporting the same underlying facts
6. Estimated trajectory: will this story grow, plateau, or fade — and why
7. Strategic significance: if this story reaches mainstream coverage levels, what are the business or investment implications
Output: Emerging story watch list ranked by estimated significance, with monitoring recommendation for eachPrompt 4 — Competitive Media Coverage Comparison
Compare the media coverage of the following competitors over the specified period.
Companies to compare: [list 3–8 companies]
Time period: [last 30 / 60 / 90 days]
Coverage dimensions to compare:
1. Volume: total article count by company
2. Tone: positive / neutral / negative distribution by company
3. Topic focus: what topics dominate each company's coverage
4. Outlet quality: are they covered by tier-1 investigative outlets or mostly wire redistribution
5. Story type: product news / financial news / leadership news / regulatory news / controversy coverage
Synthesize:
1. Which company has the most favorable media narrative and why
2. Which company is receiving the most concerning coverage and what is driving it
3. Competitive narrative gaps: where is one company telling a story that others are not
4. Reputational risk comparison: which companies appear most at risk of a negative narrative shift
Output: Comparative coverage table + narrative intelligence summaryPrompt 5 — Pre-Publication Signal Monitoring
Identify pre-publication signals from the following sources that may anticipate significant upcoming media coverage.
Coverage universe: [describe what to monitor]
Pre-publication signal sources to check: [social media, regulatory filings, court records, job postings, conference agendas, analyst notes]
Time window: [past 7 / 14 days]
For each signal detected:
1. Signal description: what specifically was observed and where
2. Source credibility: how reliable is this source as a leading indicator of media coverage
3. Story type it may generate: financial / regulatory / product / controversy / leadership
4. Estimated time to mainstream coverage: if this signal is real, how quickly might it surface in traditional media
5. Recommended action: monitor only / alert analyst team / prepare communications response
Output: Pre-publication signal brief with prioritized action recommendations22. AI Analyst Report Comparison & Consensus Synthesizer
Synthesizes sell-side analyst reports, rating changes, and estimate revisions into a structured consensus view — identifying where the consensus is right, where it is wrong, and where the disagreement is most significant.
Pain Point & How COCO Solves It
The Pain: Analyst Consensus Is Valuable but Requires More Than Averaging the Numbers
Sell-side analyst research is one of the most systematically collected and most systematically misused information sources in investment and business research. Aggregating consensus estimates from Bloomberg or FactSet is straightforward — getting to a revenue estimate is one click. But the intelligence value of analyst research lives not in the consensus numbers but in the distribution of views, the reasoning behind outlier estimates, the rating changes that signal conviction shifts, and the specific business drivers where analyst views differ most sharply. These deeper insights require reading and synthesizing the actual research notes, not just downloading the summary statistics — and reading 15 analyst reports on a single company before an earnings event is not feasible within a normal workday.
The qualitative consensus problem is more acute than the quantitative one. For buy-side analysts, understanding what sell-side consensus is thinking — not just what they are estimating — is crucial for identifying where the market view differs from your own thesis. Which analysts are most bullish and what specific business assumptions drive their view? Which are most bearish and where is their analysis most credible? Where do even the bulls acknowledge a meaningful risk? These questions require reading for content and reasoning, not just harvesting numbers — and the analytical investment required puts genuine consensus intelligence out of reach for most buy-side professionals who cover a broad universe.
The revision tracking problem compounds the synthesis challenge. Analyst estimates and ratings change frequently — quarterly updates, post-earnings revisions, initiation-of-coverage changes, and event-driven re-ratings are all significant signals. Tracking which analysts are revising in which direction, which are the first movers vs. followers, and whether the distribution of estimates is widening or narrowing requires systematic monitoring of a large corpus of documents over time. Doing this manually for a coverage universe of 20+ companies is equivalent to a full-time job, which is why most research professionals rely on aggregated statistics rather than genuine consensus intelligence.
How COCO Solves It
Analyst Report Ingestion and Structuring: COCO processes the full corpus of analyst research efficiently:
- Ingests analyst research notes in PDF format and extracts structured data: ratings, price targets, estimates, key drivers, risks, and catalysts
- Normalizes estimates across analysts who use different fiscal year definitions, segment structures, or metric conventions
- Identifies each analyst's stance on the top 3–5 business drivers being debated in the investment community
- Tags each report with analyst name, firm, publication date, rating, price target, and key estimate revisions
- Maintains a research archive that enables trend analysis across report cycles
Quantitative Consensus Construction: COCO builds distributional views, not just means:
- Calculates mean, median, high, and low estimates for each tracked financial metric
- Generates a distribution visualization showing the full range of estimates, not just the consensus midpoint
- Identifies and labels outlier analysts whose estimates diverge significantly from the consensus with their reasoning
- Tracks estimate revision direction across the analyst universe: what proportion are revising up vs. down
- Calculates estimate dispersion trends: is the analyst community converging on a view or diverging — a signal of uncertainty or disagreement
Rating and Price Target Intelligence: COCO synthesizes qualitative consensus signals:
- Tracks the distribution of ratings across the analyst universe (Buy/Hold/Sell) with count and proportion
- Identifies rating changes — upgrades and downgrades — with dates and stated rationale
- Calculates the gap between current price and the consensus price target — implied upside/downside
- Identifies which analysts are conviction buyers vs. neutral holders — beyond the formal rating, using price target and language signals
- Tracks rating revision momentum: are ratings broadly moving in one direction following a catalyst or reporting event
Bull and Bear Case Thesis Synthesis: COCO extracts the reasoning behind the numbers:
- Identifies the key business drivers that distinguish bull case from bear case analyst views
- Generates a structured bull case summary: which analysts are most positive, what they believe, and what evidence supports their view
- Generates a structured bear case summary: which analysts are most cautious, what concerns they cite, and how credible each concern appears
- Identifies the specific business assumptions where analyst disagreement is greatest — the critical variables where the investment case is most contested
- Produces a "where consensus could be wrong" analysis: the scenarios where the current consensus view appears most vulnerable
First-Mover and Conviction Signal Detection: COCO identifies the signals that precede broader consensus shifts:
- Identifies which analysts consistently lead consensus changes — whose revisions others follow
- Detects early conviction signals: analysts adding to their written emphasis on a specific driver without changing their formal estimate
- Flags when a highly regarded analyst makes an unusual estimate or changes a long-held view — disproportionately significant signals
- Tracks analyst access quality signals: which analysts appear to have the best management access based on specificity of their information
- Generates an influence-weighted consensus that gives more weight to track record-validated analysts
Consensus Intelligence Report Generation: COCO produces structured deliverables for analyst review:
- Generates a consensus summary card: key estimates, rating distribution, price target range, and major revisions since last period
- Produces a bull/bear thesis matrix: key debate points mapped to supporting analyst positions
- Creates a consensus change log: all rating changes, estimate revisions, and price target adjustments in the specified period
- Generates a differentiated view assessment: where your thesis differs from consensus and the analyst evidence for each side
- Produces a pre-event briefing (earnings, investor day, product launch) synthesizing what consensus expects and where surprises are most likely
Results & Who Benefits
Measurable Results
- Analyst research synthesis time: Reduced from 4–6 hours reading 15+ analyst reports to under 45 minutes reviewing COCO's structured synthesis (85% time savings)
- Consensus understanding depth: Analysts using COCO's bull/bear thesis synthesis understand the reasoning behind estimates in 92% of cases vs. 38% when relying on numerical consensus alone
- Estimate dispersion monitoring: COCO tracks estimate dispersion for 100% of coverage universe vs. 15–20% when done manually — identifying high-uncertainty situations more reliably
- First-mover signal detection: Analysts using COCO identify analyst conviction shifts an average of 3–5 days earlier than those monitoring only formal rating changes
- Pre-event preparation quality: Research teams using COCO pre-event briefings address 85% of post-event analytical questions in their pre-event frameworks vs. 52% without systematic consensus review
Who Benefits
- Buy-Side Analysts: Understand the sell-side consensus not just as a number but as a range of views with distinct reasoning — enabling genuine differentiated thesis development
- Portfolio Managers: Access a timely, comprehensive consensus intelligence brief for every position before material events — without requiring analyst team members to read 15 reports each
- Research Analysts: Synthesize analyst consensus efficiently across a broad coverage universe — maintaining informed views on every name without sacrificing depth on any
- Investment Strategists: Track consensus view evolution across sectors and styles — identifying where consensus is most stretched, most complacent, or most conflicted
💡 Practical Prompts
Prompt 1 — Consensus Summary Compilation
Compile a consensus summary for the following company from the available analyst research.
Company: [name]
Analyst reports to synthesize: [paste excerpts or describe the available reports — firm, analyst, date, rating, price target, key estimates]
Key financial metrics to summarize: [revenue, EBITDA, EPS, FCF, sector-specific metrics]
Reporting period: [next quarter / full year / next 3 years]
Generate:
1. Quantitative consensus: mean, median, high, and low estimate for each metric
2. Rating distribution: count and % of Buy / Hold / Sell ratings, with any recent changes
3. Price target range: mean, median, high, and low price target with implied upside/downside from current price
4. Estimate revision trend: over the past 30/60/90 days, what proportion of analysts revised up vs. down
5. Estimate dispersion: is the range of estimates unusually wide or narrow vs. historical norm
6. Top 3 debates: the business driver questions where analyst views differ most sharplyPrompt 2 — Bull and Bear Thesis Synthesis
Synthesize the bull and bear thesis debate for the following company from analyst research.
Company: [name]
Analyst research available: [paste excerpts or describe the most relevant research notes]
Bull case synthesis:
1. The 3–5 most bullish analysts: who they are and their most important business assumptions
2. Core bull thesis: the 2–3 beliefs that most distinguish bull case from consensus
3. Key evidence bulls cite for their view
4. Bull case scenarios: what would need to happen for the bullish case to be correct
Bear case synthesis:
1. The 3–5 most cautious analysts: who they are and their most important concerns
2. Core bear thesis: the 2–3 concerns that most distinguish bear case from consensus
3. Key evidence bears cite for their view
4. Bear case scenarios: what would need to happen for the bearish case to be realized
Synthesis:
1. The single most important debate: the business driver question where analyst disagreement is largest
2. Which side appears to have the stronger evidence base currently
3. The catalyst that is most likely to resolve the debate in favor of bulls or bearsPrompt 3 — Analyst Estimate Revision Tracker
Build an analyst estimate revision tracker for the following company over the specified period.
Company: [name]
Tracking period: [last 30 / 60 / 90 days]
Metrics to track: [list — revenue, EBITDA, EPS, price target, rating]
Analyst research available: [describe or paste the research notes with revision information]
For each tracked metric:
1. Revisions log: analyst name, date, prior estimate, new estimate, magnitude and direction of change
2. Revision momentum: net direction of revisions (more up than down / more down than up / balanced)
3. First movers: which analysts revised first, and whether others followed
4. Outlier identification: analysts whose estimates remain significantly different from revised consensus
Synthesize:
1. Overall revision narrative: what is driving the revision pattern — what business factors are analysts updating
2. Conviction signal: are revisions large (high conviction) or small (marginal adjustments)
3. Consensus evolution: has the revision pattern narrowed or widened the estimate distribution
Output: Revision tracker table + narrative synthesis of what the revisions signal about consensus view evolutionPrompt 4 — Pre-Earnings Consensus Review
Generate a pre-earnings consensus review for the following company.
Company: [name]
Upcoming reporting event: [earnings date and quarter]
Consensus estimates: [paste or describe key estimates for the upcoming quarter]
Street expectations for key drivers: [describe what analysts are expecting on the most-watched business drivers]
Recent guidance (if any): [what management has said publicly]
Pre-event intelligence brief:
1. Consensus expectations summary: key metric estimates with range and revision direction
2. What consensus is expecting: the top 3 business drivers analysts are watching for the quarter
3. Where surprise risk is highest: metrics or drivers where consensus appears most potentially wrong — too optimistic or too pessimistic
4. Key questions consensus expects management to address on the call
5. Post-event analyst reaction framework: what outcomes would likely trigger upgrades vs. downgrades
6. Historical earnings surprise pattern: how often does this company beat / miss / meet consensus expectations
Format: Pre-event briefing document ready to use 24–48 hours before the earnings callPrompt 5 — Differentiated View vs. Consensus Analysis
Compare my investment thesis to the sell-side consensus for the following company.
Company: [name]
My thesis: [describe your view — what you believe about the key business drivers, valuation, and investment recommendation]
My key assumptions: [list your specific estimates or views on the 3–5 most important business drivers]
Consensus view: [describe or paste the consensus view on the same business drivers]
Analysis:
1. Where I agree with consensus: business drivers where my view is consistent with analyst community
2. Where I differ from consensus: specific assumptions where my view diverges from the range of analyst estimates
3. What drives my differentiated view: the analytical work or insights that lead me to a different conclusion from consensus
4. Risks to my differentiated view: where consensus may be right and I may be wrong
5. Catalyst identification: what specific events or data releases would prove my differentiated view correct or incorrect
6. Analyst tracking list: which analysts are closest to my view and most worth reading for updates
Output: Differentiated view assessment with explicit identification of where I am making a non-consensus bet and why23. AI Technology Readiness Level Assessment Engine
Applies structured TRL assessment to evaluate the maturity, commercialization pathway, and investment readiness of any technology — with documented evidence and transition gap analysis.
Pain Point & How COCO Solves It
The Pain: Technology Maturity Assessment Is Too Often Done by Intuition Rather Than Structured Evidence
Research analysts and investors who evaluate technology-based companies or investment opportunities frequently need to assess where a technology sits on the maturity curve — whether it is a fundamental research concept, a working prototype, a validated pilot, or a production-ready commercial platform. This assessment is one of the most consequential judgments in technology research: it affects risk assessment, competitive timing analysis, investment horizon, and go-to-market strategy recommendations. Yet it is routinely done informally, based on company claims, press releases, and management presentations rather than systematic evaluation of the evidence that independently validates the technology's current capability and future trajectory.
The credibility problem is the first dimension of the challenge. Technology companies have strong incentives to represent their technology as more mature than it is — both to attract investment and to establish market positioning. The gap between "we have demonstrated this in a laboratory environment" and "we have deployed this in ten commercial installations" is enormous from a risk perspective, but both might be described as "proven technology" in a company presentation. Analysts who lack a structured framework for evaluating technology maturity claims systematically overestimate the readiness of technologies at early stages and underestimate the remaining development investment required to reach commercial viability.
The transition risk problem is equally important and equally underanalyzed. Technologies fail not only because the fundamental science does not work but because specific transitions between maturity levels are harder than anticipated: the move from laboratory proof-of-concept to robust prototype, the move from controlled pilot to diverse real-world deployment, the move from domestic deployment to international scale. Understanding where a technology is in this transition trajectory — and what specific technical, operational, or regulatory barriers must be overcome at each step — is the analytical work that distinguishes sophisticated technology research from surface-level assessment. Most research deliverables that address technology maturity do not perform this transition analysis.
How COCO Solves It
TRL Framework Application: COCO applies the Technology Readiness Level framework with precision:
- Applies the NASA/DoD 9-level TRL framework with domain-specific adaptations for software, hardware, biotech, energy, and advanced manufacturing contexts
- Assesses the technology against each TRL criterion systematically — not just accepting the level claimed by the developer
- Generates evidence requirements for each TRL level: what specific demonstrations, validations, or deployments must be documented
- Identifies the current TRL level with confidence rating based on evidence quality
- Produces a TRL assessment rationale documenting why the assessed level was chosen and what evidence supports it
Evidence Collection and Validation: COCO builds an evidence-based TRL assessment:
- Identifies publicly available evidence for TRL assessment: peer-reviewed publications, patent applications, conference presentations, regulatory filings, and press releases
- Assesses the quality and independence of cited evidence: company self-reported vs. third-party validated vs. peer-reviewed
- Flags unsupported technology claims that appear in company materials but lack independent corroboration
- Cross-references claimed capabilities against industry benchmarks and comparable technology assessments
- Generates an evidence register: each TRL criterion with the supporting evidence, its quality rating, and gaps
Commercial Readiness Assessment: COCO extends TRL analysis to commercial viability:
- Applies Manufacturing Readiness Level (MRL) assessment for hardware technologies alongside TRL
- Evaluates commercial readiness factors beyond technical maturity: supply chain, workforce, regulatory approval, and customer adoption
- Assesses the alignment between technical TRL and commercial readiness — identifying technologies that are technically ready but commercially immature or vice versa
- Identifies the specific non-technical barriers that represent the most significant commercialization risk
- Generates a commercial readiness rating alongside the TRL to provide a two-dimensional maturity picture
Transition Gap and Development Pathway Analysis: COCO maps the path from current to target TRL:
- Identifies the specific technical barriers that must be overcome to advance from the current TRL to each subsequent level
- Assesses the difficulty, cost, and time requirement for each TRL transition based on comparable technology development histories
- Maps the regulatory validation requirements at each transition stage — particularly relevant for medical devices, energy systems, and safety-critical applications
- Identifies which TRL transitions have the historically highest failure rate for technologies in this category
- Generates a development pathway with milestone definitions, resource requirements, and risk assessment for each stage
Competitive TRL Benchmarking: COCO places technology maturity in competitive context:
- Assesses the TRL of competing technologies pursuing the same application — providing relative maturity context
- Identifies which competitors are at higher or lower TRL levels and what gap this represents
- Tracks TRL progression rates for competing technologies to assess who is advancing most rapidly
- Identifies technologies at adjacent TRL levels that could leapfrog current leaders if transition barriers are overcome
- Generates a technology race assessment: who is leading, who is closing, and what events could shift the competitive ordering
Investment and Strategic Implication Report: COCO produces structured deliverables for decision-making:
- Generates a technology maturity brief with TRL assessment, evidence basis, transition pathway, and strategic implications
- Produces an investment readiness assessment: what TRL milestone is typically needed for the investment stage being considered
- Creates a technology risk register linked to each transition stage — what could prevent reaching commercial readiness
- Generates a portfolio positioning analysis for investors holding technologies at different TRL stages
- Formats output for investment committee, board technology review, or client strategy document contexts
Results & Who Benefits
Measurable Results
- TRL assessment time: Reduced from 2–3 weeks of expert analysis to 3–5 days for a comprehensive TRL assessment with evidence documentation (80% time savings)
- Assessment consistency: TRL assessments produced by COCO show 78% agreement with independent expert assessments vs. 52% for unstructured analyst assessments
- Technology claim verification: COCO identifies unsupported technology maturity claims in 68% of assessed company presentations — claims accepted without challenge in typical due diligence processes
- Transition barrier identification: COCO assessments identify an average of 3.2 specific transition barriers per TRL gap that are not identified in company-provided roadmaps
- Investment timing accuracy: Portfolios using COCO TRL assessments enter technology investments at stage-appropriate timing in 81% of cases vs. 58% for portfolios without structured maturity frameworks
Who Benefits
- Research Analysts: Assess technology maturity claims with structured evidence standards rather than accepting company representations — producing defensible technology maturity opinions
- Technology Investors and VCs: Apply a consistent TRL framework across their portfolio and deal pipeline — improving stage-appropriate investment decision-making and portfolio risk management
- Corporate R&D and Strategy Teams: Evaluate internal and external technology options against a common maturity framework — informing build vs. buy vs. partner decisions with analytical rigor
- Government and Public Sector Researchers: Apply established TRL standards to assess readiness for technology deployment in public programs — meeting the evidence standards required for major procurement or policy decisions
💡 Practical Prompts
Prompt 1 — TRL Assessment for a Specific Technology
Conduct a Technology Readiness Level (TRL) assessment for the following technology.
Technology: [describe the technology — what it does, how it works, and what application it is being developed for]
Developer/company: [who is developing it]
Available evidence sources: [list publications, presentations, press releases, regulatory filings, or other available evidence]
Assessment purpose: [investment due diligence / competitive benchmarking / R&D portfolio evaluation / procurement decision]
Assess:
1. Current TRL level (1–9) with justification
2. Evidence supporting the assessed TRL level — for each criterion at the assessed level, what evidence exists
3. Evidence gaps — what evidence would be needed to confirm this TRL level if currently missing
4. Critical unvalidated claims — what technology capabilities are claimed but not independently demonstrated
5. Next TRL milestone — what specific demonstrations or validations are needed to advance one level
6. Confidence level in the assessment: high / medium / low — based on evidence quality and availability
Format: TRL assessment brief with evidence register and confidence-rated conclusionPrompt 2 — Commercial Readiness Assessment
Assess the commercial readiness of the following technology alongside its technical TRL.
Technology: [description]
Target application and market: [what it is intended to do and who would use it]
Available commercial evidence: [deployments, customers, revenue, partnerships, regulatory approvals]
Assess across commercial readiness dimensions:
1. Technical TRL: current level with evidence basis
2. Manufacturing/Production readiness: can this be produced at scale, at target cost, and with consistent quality
3. Regulatory clearance status: what approvals are required, what is the current approval status, and what is remaining
4. Supply chain readiness: are the inputs, components, and production infrastructure available at scale
5. Customer adoption readiness: is the target customer capable of adopting and using this technology at current maturity
6. Business model validation: has the commercial model (pricing, distribution, service requirements) been demonstrated
Generate:
1. Two-dimensional maturity matrix: TRL (technical) × Commercial Readiness (1–5) with current position plotted
2. Commercialization gap analysis: where is the gap between technical and commercial readiness largest
3. Critical path to commercial launch: the 3 most important barriers to address for commercializationPrompt 3 — Competitive Technology TRL Benchmarking
Compare the TRL of competing technologies pursuing the following application.
Target application: [describe what the technology is meant to do]
Competing technologies to assess: [list 3–6 distinct technology approaches or companies]
Evidence sources available: [publications, presentations, filings, deployments]
For each technology:
1. Current TRL assessment with evidence basis
2. Rate of TRL progression: how quickly has it advanced in the past 2–3 years
3. Key technical barriers remaining before commercial deployment
4. Developer profile: who is behind it and what resources are they committing to advancement
5. Comparative timeline estimate to commercial readiness
Synthesize:
1. Technology leadership ranking: which approach is most advanced and why
2. Disruption risk: which lower-TRL technologies have the potential to leapfrog current leaders if their transition barriers are overcome
3. Investment timing implication: which technology in the race represents the best risk-adjusted investment timing
Output: Competitive TRL comparison matrix + race narrative + investment timing assessmentPrompt 4 — TRL Transition Pathway Analysis
Map the development pathway from the current TRL to commercial deployment for the following technology.
Technology: [description]
Current TRL: [assessed level]
Target TRL for commercial deployment: [typically 7–9 depending on domain]
Domain: [energy / medical device / defense / software / advanced materials — affects transition criteria]
For each TRL transition from current level to target level:
1. Transition criteria: what specific demonstrations or validations are required to advance
2. Technical barriers: what engineering or scientific challenges must be overcome at this stage
3. Resource requirements: approximate time and cost investment for this transition, based on comparable technology development
4. Regulatory requirements: what approvals or certifications are required at this transition stage
5. Historical success rate: how often do technologies in this category successfully complete this specific transition
6. Critical risk factors: what most commonly causes failure at this transition stage
Generate:
1. Development timeline estimate with milestone definitions for each transition
2. Investment requirement estimate by transition stage
3. Risk-adjusted expected time to commercial readiness accounting for transition failure probabilities
4. Recommended due diligence focus: which transition represents the highest-risk milestone for this technologyPrompt 5 — Technology Investment Readiness Brief
Assess the investment readiness of the following technology for [investment stage].
Technology: [description]
Investment stage being considered: [seed / Series A / Series B / growth / strategic acquisition]
Current TRL: [assessed or claimed]
Investment amount: [approximate investment size being considered]
Investment readiness assessment:
1. TRL adequacy: is the current TRL appropriate for this investment stage — what level is typically expected before [stage] investment in this technology category
2. Evidence quality: how well-documented is the current TRL — what additional validation is needed before investment
3. Time to next value inflection: how long and at what cost to reach the next TRL level that would trigger a step-change in valuation
4. Technology risk premium: what discount to the comparable commercial-stage company should apply given current TRL
5. Failure scenario analysis: what technical risk, if it materializes at current or next TRL stage, would most threaten the investment thesis
6. Investment milestone structure: how to structure milestones and tranches to manage TRL progression risk
Output: Investment readiness brief with go/no-go recommendation framework based on TRL evidence threshold24. AI Customer Pain Point Cluster Analysis Tool
Synthesizes customer feedback, support tickets, review data, and interview insights into a structured pain point map — prioritized by frequency, severity, and strategic importance.
Pain Point & How COCO Solves It
The Pain: Customer Pain Point Data Is Abundant but Remains Fragmented and Unactionable
Understanding what customers find painful, frustrating, or inadequate about their current solutions is foundational to product strategy, competitive intelligence, and market opportunity assessment. Research analysts tasked with customer insight work typically have access to multiple streams of relevant data: customer satisfaction surveys, support ticket logs, online review sites, social media complaints, win/loss interview notes, and customer advisory board recordings. The problem is not lack of data but lack of synthesis: these streams use different vocabularies, operate at different levels of specificity, and have never been systematically compared and consolidated. Each data source represents a partial view; the full picture of customer pain requires integrating all of them — work that is analytically intensive and rarely done comprehensively.
The categorization problem is the first analytical challenge. A customer service ticket saying "the export function doesn't work" and a review site comment saying "can't get my data out" and a survey response saying "I need better integrations" may all express variations of the same underlying pain point — or they may represent three distinct issues. Deciding which observations represent the same pain and which represent genuinely different problems requires judgment, and making that judgment consistently across thousands of data points requires either extensive manual effort or a systematic methodology that most research teams do not have time to apply.
The prioritization problem compounds the categorization challenge. Even when pain points are correctly identified and categorized, the raw count of mentions is not the right prioritization metric. A pain point mentioned by 5% of customers might matter far more strategically than one mentioned by 30% if the 5% are the highest-value customer segment, or if the pain point represents a barrier to a major new use case. Prioritizing pain points requires combining frequency data with severity, affected customer segment value, competitive differentiation potential, and feasibility of resolution — a multi-dimensional analysis that is only possible when the pain point data has been properly structured and enriched.
How COCO Solves It
Multi-Source Pain Point Data Integration: COCO aggregates customer feedback from across all available channels:
- Ingests and normalizes data from support ticket systems, CRM notes, NPS verbatim responses, survey open-ends, and review site text
- Processes interview transcripts, focus group notes, and customer advisory board recordings to extract pain point statements
- Handles unstructured text data: extracts pain point statements from narrative feedback regardless of format
- Normalizes vocabulary across sources: identifies when different words describe the same underlying issue
- Maintains source attribution for each pain point to enable downstream analysis by feedback type and customer segment
Pain Point Clustering and Taxonomy Development: COCO builds a structured pain point architecture:
- Groups individual feedback statements into coherent pain point clusters based on semantic similarity and shared underlying cause
- Organizes clusters into a hierarchical taxonomy: primary pain categories > specific pain points > symptom expressions
- Labels each cluster with a precise, analyst-interpretable name that reflects the underlying issue rather than the surface symptom
- Identifies pain points that cut across multiple product areas, customer segments, or use cases — systemic vs. localized issues
- Generates alternative taxonomy structures for analyst review when the clustering is ambiguous
Pain Point Frequency and Distribution Analysis: COCO quantifies the prevalence of each pain point:
- Calculates mention frequency for each pain point cluster across all source types and customer segments
- Analyzes source distribution: which pain points appear across all channels vs. which are concentrated in specific feedback types
- Identifies trending pain points: issues increasing in mention frequency over time, suggesting a growing or newly emerged problem
- Calculates co-occurrence: which pain points frequently appear together in the same customer feedback, suggesting a systemic root cause
- Generates a pain point frequency map showing relative prevalence across the full pain point taxonomy
Pain Point Severity and Strategic Importance Scoring: COCO builds a prioritization framework beyond counts:
- Applies severity weighting based on language intensity, escalation level, and customer segment value
- Identifies pain points associated with churn, downgrade, or competitive loss signals — the highest-value severity indicators
- Maps each pain point to customer segment and values that segment using available customer value data
- Assesses competitive differentiation potential: for each pain point, how well are competitors solving it vs. how poorly is the current solution performing
- Calculates a multi-dimensional priority score combining frequency, severity, segment value, and competitive opportunity
Root Cause and Solution Architecture Analysis: COCO goes beyond symptom identification:
- Groups surface-level pain symptoms into root cause categories — distinguishing feature gaps, usability issues, reliability problems, pricing concerns, and support failures
- Identifies whether a pain point reflects a fixable execution issue vs. a fundamental product architecture constraint
- Maps pain points to the specific workflow stages where they occur — identifying whether problems concentrate at onboarding, daily use, reporting, or integration stages
- Generates hypothesized solution directions for each root cause category — what type of intervention would address the underlying issue
- Prioritizes root causes by the breadth of surface pain points they explain — high-leverage root causes that, if addressed, would resolve multiple symptoms
Customer Pain Point Research Report: COCO produces structured deliverables:
- Generates an executive pain point brief: top 10 pain points by priority score with strategic implications
- Produces a full pain point taxonomy with frequency, severity, and priority data for every identified cluster
- Creates segment-specific pain point profiles: how the pain landscape differs for enterprise vs. SMB, heavy users vs. light users, new customers vs. tenured
- Generates a competitive opportunity map: pain points where competitors perform better, providing the clearest competitive risk
- Produces a product roadmap input document structured for direct use in product prioritization processes
Results & Who Benefits
Measurable Results
- Pain point synthesis time: Reduced from 2–3 analyst weeks to 3–5 days for a comprehensive multi-source pain point analysis (80% time savings)
- Pain point coverage: COCO's multi-source integration identifies 2.5x more distinct pain point clusters than single-source analysis (e.g., survey-only or support-only)
- Prioritization alignment: Product teams using COCO priority scores show 67% better alignment between pain point investment and customer churn reduction vs. frequency-only prioritization
- Root cause identification rate: COCO identifies a root cause category for 82% of surface pain point clusters vs. 34% in unstructured pain point research
- Cross-segment pain differentiation: COCO identifies segment-specific pain patterns that differ meaningfully from the aggregate in 78% of customer bases analyzed
Who Benefits
- Research Analysts: Deliver customer insight that integrates all available feedback sources into a coherent, prioritized picture rather than a report on one data type
- Product Managers: Access a structured, multi-dimensional prioritization of customer pain points that directly informs roadmap decisions — moving beyond "most mentioned" to "most strategically important"
- Customer Success and Support Leaders: Understand the systemic root causes behind support ticket volume and customer satisfaction issues — enabling structural fixes rather than case-by-case resolution
- Competitive Intelligence Teams: Map competitor pain points using public review data — identifying where competitors are failing their customers and where opportunity exists for competitive differentiation
💡 Practical Prompts
Prompt 1 — Multi-Source Pain Point Synthesis
Synthesize customer pain points from the following data sources.
Product/service being assessed: [description]
Customer segment: [enterprise / SMB / consumer / specific vertical]
Data sources to synthesize (paste or describe):
1. Survey verbatims: [paste or describe — e.g., "500 NPS detractor open-end responses"]
2. Support tickets: [paste or describe — e.g., "top 100 most common ticket subjects from last 90 days"]
3. Interview notes: [paste or describe]
4. Review site data: [paste or describe — e.g., "G2 reviews with 1–3 star ratings"]
5. [Additional sources]
Deliverables:
1. Pain point taxonomy: hierarchical structure of primary categories and specific pain points
2. Frequency analysis: how often each pain point appears across all sources combined
3. Source distribution: which pain points are consistent across sources vs. concentrated in one channel
4. Top 15 pain points ranked by frequency
5. Trending pain points: issues that have increased in frequency over time (if time-series data available)
6. Pain point co-occurrence: clusters that frequently appear together in the same feedbackPrompt 2 — Pain Point Priority Scoring
Build a strategic priority score for the following pain points.
Pain points to prioritize: [list identified pain points with their raw frequency counts]
Customer segment data: [describe the value profile of affected customer segments — if available]
Competitive data: [describe how well competitors address each pain point — if known]
Churn/retention data: [which pain points have been associated with churn or downgrade — if available]
Scoring dimensions:
1. Frequency: how often is this pain point mentioned (0–5 scale)
2. Severity: how intensely do customers express this pain (based on language intensity and escalation) (0–5 scale)
3. Segment value: how valuable are the customers expressing this pain (0–5 scale)
4. Competitive risk: how much better do competitors handle this (0–5 scale)
5. Business impact: is this pain associated with churn, expansion, or key use case blocking (0–5 scale)
Calculate:
1. Composite priority score for each pain point (weighted combination of above dimensions)
2. Priority tier classification: must-fix / high priority / medium priority / low priority
3. Priority distribution visualization: matrix of frequency vs. severity with segment value overlay
Recommend: top 5 pain points to address first with strategic rationale for eachPrompt 3 — Competitive Pain Point Gap Analysis
Analyze customer pain points for [competitor] from public review data and compare to [our product].
Competitor: [name]
Our product: [name]
Review data sources: [G2, Gartner Peer Insights, Capterra, App Store, etc. — describe what's available]
For the competitor:
1. Identify the top 10 pain points from 1–3 star reviews and critical sections of mixed reviews
2. Categorize by pain type: feature gap / usability / reliability / pricing / support / integration
3. Frequency and severity for each pain point
4. Customer segment distribution: are certain pain points concentrated in specific customer types
Compare to our product:
1. Pain points where we have the same issue — shared vulnerabilities
2. Pain points where our product is better — competitive advantages
3. Pain points where our product is worse — areas of competitive risk
4. Pain points unique to the competitor — potential migration triggers for their customers
Output: Competitive pain gap matrix + migration opportunity identification for the top competitive pain pointsPrompt 4 — Root Cause Analysis for Pain Clusters
Conduct a root cause analysis for the following customer pain point clusters.
Pain point clusters to analyze:
1. [Cluster 1: name and description of the surface symptoms]
2. [Cluster 2]
3. [Cluster 3]
[List up to 8 clusters]
Supporting evidence: [paste or describe the specific feedback statements that make up each cluster]
Product context: [describe how the relevant product features work, or where the relevant processes occur]
For each cluster:
1. Surface symptoms: what specific customer complaints are grouped in this cluster
2. Root cause hypothesis: what underlying product, process, or design issue is likely causing these symptoms
3. Evidence supporting the root cause: which specific feedback statements point to the root cause
4. Root cause category: feature gap / architecture constraint / usability / process / policy / integration
5. Solution direction: what type of fix would address the root cause — and how feasible is it
6. Cross-cluster linkage: does this root cause also explain symptoms in other pain point clusters
Output: Root cause analysis brief with solution direction recommendations ranked by feasibility and impactPrompt 5 — Segment-Specific Pain Profile
Build segment-specific pain profiles from the following customer feedback data.
Overall customer base pain point data: [paste or describe the aggregate pain point frequency data]
Segment breakdown data: [paste or describe pain point data broken down by customer segment, if available]
Customer segments to profile: [list — e.g., enterprise, SMB, healthcare vertical, new customers <6 months, tenured customers >2 years]
For each segment:
1. Top 5 pain points: the pain points most frequently or severely expressed by this segment
2. Unique pain points: issues this segment mentions that are not prominent in the aggregate or other segments
3. Pain intensity: is this segment's overall pain expression more or less intense than the aggregate
4. Pain trend: are pain expressions from this segment increasing, stable, or decreasing over time (if time-series available)
Cross-segment synthesis:
1. Pain points universal to all segments — systemic issues requiring priority attention
2. Pain points concentrated in the highest-value segments — strategic priority regardless of overall frequency
3. Pain points that vary most by segment — suggesting that segment-specific solutions are appropriate
Output: Segment pain profile cards + cross-segment priority matrix25. AI Grant & Funding Opportunity Research Assistant
Identifies, evaluates, and matches research grants, government funding, and private foundation opportunities to specific research programs — building the evidence base for competitive applications.
Pain Point & How COCO Solves It
The Pain: Grant Research Is a Full-Time Job That Competes With the Research It Is Supposed to Fund
Research analysts and academic researchers who depend on external funding face a structural paradox: finding, evaluating, and applying for grants requires nearly as much time and specialized knowledge as doing the research itself. The landscape of available funding is vast and fragmented: federal agencies (NIH, NSF, DOE, DOD, USDA, NEH), state research funds, international bodies (EU Horizon, Wellcome Trust, Gates Foundation), corporate research partnerships, and private foundations each have distinct eligibility criteria, funding mechanisms, application formats, and strategic priorities. Staying current with this landscape while also producing research, teaching, advising students, or managing an analyst workload is genuinely impossible — which is why so many researchers leave potential funding on the table, not from lack of competitive research but from lack of visibility into opportunities that match their work.
The matching problem is the first practical barrier. A researcher with a specific project has specific needs: a certain funding amount, a compatible timeline, eligibility that matches their institutional and geographic status, and a funder whose strategic priorities align with their research question. Identifying which of thousands of available opportunities meet all these criteria simultaneously requires either a dedicated grants administrator or extensive manual searching — resources that most smaller research groups and independent analysts do not have. Generic grant databases return too many results that are superficially relevant but actually ineligible; the researcher sorts through them manually and still misses opportunities that used different terminology to describe the same research area.
The application quality problem compounds the discovery problem. Even researchers who identify the right grant opportunity frequently submit applications that fail not because the research is weak but because the application does not speak the funder's language — does not frame the significance of the research in terms that resonate with the funder's strategic priorities, does not position the team's qualifications in terms of the evaluation criteria, and does not structure the work plan in the format that reviewers find most credible. Building this alignment between the researcher's genuine strengths and the funder's evaluative preferences is a communication and strategic framing task as much as a technical writing task — one that experienced grant writers do well and most researchers do poorly.
How COCO Solves It
Grant Database Search and Opportunity Discovery: COCO scans the full funding landscape for relevant opportunities:
- Searches federal agency grant databases (grants.gov, NIH Reporter, NSF Award Search, DOE funding opportunities) for matching programs
- Monitors foundation grant databases (Foundation Directory, Foundation Center) for private philanthropy opportunities
- Tracks international funding bodies: EU Horizon funding opportunities, UK Research Councils, international development funding agencies
- Monitors corporate research partnership programs and industry consortia funding opportunities
- Applies research area, eligibility, funding size, and timeline filters to surface only genuinely relevant opportunities
Eligibility and Fit Assessment: COCO evaluates each opportunity against the research profile:
- Assesses eligibility: institution type, researcher career stage, citizenship and geographic restrictions, prior award limitations
- Evaluates research area fit: how closely does the funder's stated priority area match the proposed research
- Assesses funding mechanism fit: is the amount, duration, and indirect cost rate compatible with the research needs
- Evaluates strategic alignment: does the funder's current strategic focus — based on recent award patterns — match the research direction
- Generates a fit score for each opportunity combining eligibility, area match, and strategic alignment
Funder Intelligence and Strategic Analysis: COCO builds intelligence on each relevant funder:
- Analyzes the funder's recent award portfolio to identify which research types, institutions, and investigators they have recently funded
- Extracts the funder's stated strategic priorities from program announcements, strategic plans, and review criteria
- Identifies funding panel compositions and reviewer expertise for federal programs — informing how to frame the application
- Tracks funding level trends: is this funder's budget increasing, declining, or stable for the relevant research area
- Generates funder profiles for the most promising opportunities: what they fund, why they fund it, and how they evaluate applications
Competitive Landscape Assessment: COCO assesses the competitive environment for each opportunity:
- Analyzes recent award winners for the same grant mechanism: what type of research, institution, and investigator profile wins
- Calculates historical funding rates for each opportunity — the proportion of applications that receive awards
- Identifies the differentiators of successful applications based on review criteria and award portfolio analysis
- Assesses the competitive intensity: how many other researchers are likely to apply to this opportunity
- Generates a competitiveness assessment: how strong is this application relative to likely competitors given the funder's recent preferences
Application Strategy and Framing Support: COCO helps build competitive applications:
- Generates a significance framing strategy: how to articulate the research's importance in terms that resonate with the funder's stated priorities
- Develops a positioning statement that differentiates the research team's qualifications in terms of the specific review criteria
- Builds a work plan structure aligned with the funder's expected format and reviewer expectations
- Generates a specific aims or executive summary aligned with the funder's evaluation priorities
- Identifies the strongest supporting evidence to include for each application section
Grant Portfolio and Pipeline Management: COCO supports long-term funding strategy:
- Generates a prioritized opportunity pipeline ranked by fit, competitiveness, and strategic importance
- Builds a grant calendar with deadlines, pre-submission requirements, and preparation milestones for each opportunity
- Tracks the status of submitted applications and identifies follow-up actions
- Generates resubmission strategy for rejected applications based on reviewer comments
- Produces a funding gap analysis: identifying funding needs not covered by current pipeline
Results & Who Benefits
Measurable Results
- Grant discovery time: Reduced from 8–15 hours per quarter of manual database searching to under 2 hours for a comprehensive funding landscape review (85% time savings)
- Opportunity identification rate: COCO identifies an average of 4–7 previously unknown relevant funding opportunities per research profile that were not found in the researcher's existing search process
- Application success rate: Research groups using COCO's funder intelligence and framing support report 30–45% higher application success rates compared to their historical baseline
- Funding pipeline value: COCO-managed grant pipelines average 2.5x higher total requested funding in active applications than manually managed pipelines — reflecting both more opportunities and better fit assessment
- Application preparation time: Initial application framework generation reduced from 2–3 weeks to 3–5 days using COCO framing and structure support
Who Benefits
- Research Analysts and Academic Researchers: Identify and pursue funding opportunities that match their research — without a dedicated grants administrator or sacrificing research time to grants management
- Research Administrators and Grants Offices: Manage larger grant portfolios more efficiently — using COCO to handle discovery and initial screening while focusing administrator time on compliance and submission
- Think Tanks and Policy Research Organizations: Access the full landscape of government and foundation funding relevant to their work — moving beyond the handful of funders they already know
- Early-Career Researchers: Navigate the grant landscape without the personal network and institutional knowledge that gives established researchers a systematic advantage in funding discovery
💡 Practical Prompts
Prompt 1 — Grant Opportunity Discovery
Identify grant funding opportunities for the following research program.
Research description: [describe the research — topic, methods, intended impact]
Researcher/institution profile: [PI career stage, institution type, geographic location, citizenship, relevant expertise]
Funding requirements: [total funding needed, preferred duration, direct vs. total costs, indirect rate constraints]
Eligible funder types: [federal / state / foundation / corporate / international / all]
Timeline constraints: [earliest start date, preferred funding duration]
Search for opportunities from:
- Federal agencies: NIH, NSF, DOE, DOD, USDA, NEH, NEA, HRSA, and relevant mission agencies
- Major foundations: relevant to research area
- International bodies: if eligible
- Corporate research programs: if applicable
For each identified opportunity (aim for 15–25 relevant results), provide:
1. Program name and funder
2. Funding amount range and duration
3. Application deadline (next cycle)
4. Eligibility summary
5. Research area fit: how closely does this match the research description
6. URL or access information
Rank by overall fit (eligibility + research match + strategic alignment)Prompt 2 — Funder Intelligence Brief
Build a funder intelligence brief for the following grant opportunity.
Grant program: [name]
Funder: [agency or foundation]
Grant mechanism: [R01 / SBIR / RFP / open call / etc.]
My research area: [brief description]
Provide intelligence on:
1. Strategic priorities: what research areas and approaches does this funder most value currently
2. Award portfolio analysis: what type of research has this funder recently funded in the relevant area (look at last 3 years of awards)
3. Preferred investigator profile: what career stage, institution type, and track record do recent award winners typically have
4. Review criteria: what criteria do reviewers use and how are they weighted
5. Funding rate: what proportion of applications receive awards (if publicly available)
6. Application format expectations: how does this funder prefer work plans, budgets, and significance sections to be structured
Strategic framing recommendation: how should I position my research to align most strongly with this funder's current prioritiesPrompt 3 — Specific Aims / Executive Summary Drafting
Draft a specific aims or executive summary for the following grant application.
Grant program: [name and funder]
Funder's stated priorities: [paste from program announcement]
Research project description: [describe in detail — what you are studying, why it matters, your approach, and expected outcomes]
Team qualifications: [describe PI and co-investigator backgrounds and relevant preliminary data]
Funding requested: [amount and duration]
Draft a specific aims section (1 page) that:
1. Opens with a compelling significance statement that frames the research problem in terms the funder cares about
2. States the knowledge gap or problem the research addresses
3. Presents the central hypothesis or research objective
4. Lists 3–4 specific aims with brief rationale for each
5. Closes with a statement of expected outcomes and impact
Ensure the draft:
- Uses language that echoes the funder's stated priorities
- Positions the team's qualifications as specifically relevant to this program
- Avoids jargon that reviewers outside the specialty might not share
- Fits within 1 page when formatted for the program's requirementsPrompt 4 — Application Competitive Analysis
Assess the competitiveness of my grant application for the following program.
Grant program: [name and funder]
My application summary: [describe the research, team, and approach in 2–3 paragraphs]
Review criteria: [paste from program announcement]
Recent award data: [describe or paste information about recent award winners if available]
Assess my application's competitiveness:
1. Significance: how compelling is the research problem and its public/scientific importance
2. Approach: how rigorous and feasible is the proposed methodology
3. Innovation: does the research offer something genuinely new vs. incremental
4. Investigator: how well does the team's background align with the reviewers' expectations for this program
5. Environment: does the institutional context support the proposed work
For each criterion:
- Strength assessment: 1–5 scale
- Specific weaknesses to address in the application
- Recommended language or evidence to strengthen the section
Overall competitiveness assessment + top 3 improvements that would most increase the probability of awardPrompt 5 — Resubmission Strategy
Develop a resubmission strategy for the following rejected grant application.
Original application: [describe the research and approach briefly]
Funder/program: [name]
Reviewer summary statement or summary of critique: [paste the reviewer comments or critique summary]
Areas cited as weaknesses: [list the specific criticisms raised]
Application score (if available): [overall score and criterion scores]
Develop a resubmission strategy:
1. Response to each critique: for each weakness identified, what specific change addresses it
2. New preliminary data: what additional data could be collected before resubmission that would address reviewers' concerns
3. Approach modifications: what changes to the research design would strengthen the weakest sections
4. Framing adjustments: where should the significance and innovation framing be strengthened
5. Team additions: would adding a collaborator with specific expertise address a reviewer concern
6. Introduction to resubmission: draft the 1-page introduction that is required in most resubmissions, acknowledging the critique and summarizing the response
Priority ranking: which changes will have the greatest impact on the resubmission score, ranked by importance26. AI M&A Comparable Transaction Database Builder
Builds a customized comparable transaction database for any deal type, industry, and geography — with deal multiples, deal terms, and strategic rationale synthesized from public sources.
Pain Point & How COCO Solves It
The Pain: Comparable Transaction Research Is Time-Intensive and Systematically Incomplete When Done Manually
Comparable transaction analysis is foundational to M&A valuation, deal structuring, and strategic rationale development. Every investment banker, corporate development professional, and M&A research analyst needs comps — recent transactions in the same industry, at similar scale, and with comparable deal structures — to assess whether a proposed valuation is defensible and what deal terms are market-standard. Building this database manually requires searching SEC filings, press releases, merger proxies, and deal databases for relevant transactions, extracting comparable financial metrics, calculating deal multiples, and documenting deal terms — a process that takes 2–5 days for a thorough comp set and is often constrained by database access and institutional knowledge.
The coverage quality problem is the first dimension of the challenge. Paid deal databases provide transaction data but are expensive, have variable coverage for smaller or older transactions, and require significant analyst time to filter and organize into a usable comp set. The free alternative — searching SEC filings and press releases — is time-consuming and requires domain expertise to identify which transactions qualify as true comparables. The result is that most comp sets are built quickly from the most prominent, most recent, and most publicized transactions — giving undue weight to high-profile deals that may not be the best comparables for the transaction at hand.
The deal terms intelligence problem is distinct from the valuation multiple problem. Understanding what deal terms — representation and warranty provisions, earnout structures, indemnification caps, MAC definitions, closing conditions — are market-standard for a specific deal type requires systematic review of publicly available transaction documents. Most M&A research focuses on valuation multiples and ignores this contractual intelligence, which means deal teams negotiate in an information vacuum on deal structure while having reasonable confidence on pricing. The teams that have systematic deal terms intelligence hold a meaningful negotiating advantage.
How COCO Solves It
Comparable Transaction Identification and Screening: COCO builds a comprehensive deal universe:
- Searches SEC filings, press release databases, and regulatory filing systems for relevant transactions
- Applies screening criteria: industry classification, deal size range, geography, time period, deal type
- Identifies transactions that are genuinely comparable vs. those that are superficially similar but structurally different
- Includes both successfully completed and withdrawn transactions — with the latter providing information on deal failure risk factors
- Cross-references multiple sources to ensure comprehensive coverage including smaller transactions not prominently covered in mainstream databases
Financial Metric Extraction and Multiple Calculation: COCO builds a clean quantitative comp set:
- Extracts relevant financial metrics from target company financials at the time of announcement: revenue, EBITDA, EBIT, book value, and sector-specific metrics
- Calculates standard deal multiples: EV/Revenue, EV/EBITDA, EV/EBIT, P/E, P/B, and sector-specific multiples
- Normalizes metrics for comparability: LTM, NTM, and growth-adjusted multiples where appropriate
- Identifies and flags unusual accounting treatments or pro forma adjustments that affect multiple comparability
- Generates a clean multiple table with mean, median, quartile range, and outlier identification for each metric
Deal Premium and Structure Analysis: COCO extracts deal structure intelligence:
- Calculates acquisition premiums to pre-announcement market price for public company transactions
- Identifies deal structure: all-cash, all-stock, mixed consideration, earnout components, rollover equity
- Extracts earnout provisions where disclosed: size as % of deal value, milestones, and payment timing
- Identifies financing structure for leveraged transactions: equity/debt ratio, debt multiple, and financing sources
- Documents regulatory conditions, closing timeline, and any renegotiated terms from announcement to closing
Strategic Rationale and Deal Driver Analysis: COCO synthesizes qualitative intelligence:
- Extracts management-stated strategic rationale from announcement press releases, investor presentations, and conference calls
- Categorizes deal drivers: scale, geographic expansion, capability acquisition, product extension, financial engineering, defensive consolidation
- Identifies how the acquisition was described in analyst and media commentary at announcement
- Tracks post-closing outcomes: did the deal achieve its stated rationale, and what integration approach was pursued
- Generates a strategic pattern analysis: what types of deals in this space have been most successful and why
Deal Terms and Contractual Intelligence: COCO extracts transaction document intelligence:
- Extracts key deal terms from merger agreements and definitive proxy statements where publicly available
- Identifies representation and warranty scope, survival periods, and indemnification caps
- Documents MAC (Material Adverse Change) definitions and carve-outs in the transaction context
- Extracts termination fee structures: buyer breakup fee, seller breakup fee, and reverse termination fee as % of deal value
- Identifies walk rights, financing conditions, and regulatory condition provisions
Comparable Transaction Report Generation: COCO produces structured, formatted deliverables:
- Generates a clean comparable transaction table formatted for inclusion in investment bank presentations or fairness opinion support
- Produces a deal multiple summary with statistical analysis: mean, median, range, and outlier annotation
- Creates a deal structure summary showing the distribution of consideration types, financing approaches, and earnout usage
- Generates a strategic rationale analysis across the comp set — patterns in what acquirers say they are paying for
- Produces a deal terms market standard summary for the most common contractual provisions in the comp set
Results & Who Benefits
Measurable Results
- Comp set construction time: Reduced from 2–5 analyst days to 4–8 hours for a comprehensive comparable transaction database (80% time savings)
- Transaction coverage: COCO identifies an average of 35% more comparable transactions than analyst-built comp sets using only paid databases
- Multiple extraction accuracy: COCO achieves 97% accuracy in financial metric extraction and multiple calculation vs. 88% under typical time pressure for manual comp building
- Deal terms intelligence adoption: Proportion of M&A research packages including deal terms market standard analysis increases from 12% to 78% when COCO automates contractual intelligence extraction
- Comp set completeness: COCO comp sets include both completed and withdrawn transactions in 89% of cases vs. 21% for manually built comps
Who Benefits
- M&A Research Analysts: Build comprehensive, well-documented comp sets efficiently — without sacrificing coverage quality for time savings
- Investment Bankers: Access deep deal intelligence for fairness opinion support, pitch preparation, and deal structuring negotiations — including deal terms market standard data typically missing from manual research
- Corporate Development Teams: Benchmark proposed transaction terms and valuations against a comprehensive comp set — negotiating from an informed position on both pricing and deal structure
- Private Equity Transaction Teams: Access systematic comparable transaction databases for buy-side deal evaluation and portfolio company sale preparation
💡 Practical Prompts
Prompt 1 — Comparable Transaction Database Construction
Build a comparable transaction database for the following deal.
Target deal overview: [describe the proposed or actual transaction — industry, deal size range, buyer type, deal type]
Screening criteria:
- Industry: [SIC/NAICS codes or description]
- Geography: [target company location]
- Deal size: [$X to $Y enterprise value]
- Time period: [last 3 / 5 / 10 years]
- Deal type: [strategic buyer / financial buyer / both; public target / private target / both]
For each identified comparable transaction, extract:
1. Target name, acquirer name, announcement date, close date
2. Enterprise value and equity value at announcement
3. Target financial metrics at time of announcement: LTM revenue, EBITDA, EBITDA margin
4. Deal multiples: EV/Revenue, EV/EBITDA, EV/EBIT
5. Acquisition premium (for public targets)
6. Deal structure: cash, stock, earnout, or mixed consideration
7. Strategic rationale (from press release / management commentary)
Output: Clean transaction table + multiple summary statistics + outlier annotationPrompt 2 — Deal Multiple Benchmarking
Benchmark the following proposed deal multiple against comparable transactions.
Proposed transaction: [describe target company and deal structure]
Proposed enterprise value: [$X]
Target financial metrics: Revenue: [$X], EBITDA: [$X], EBITDA margin: [%]
Proposed multiples: EV/Revenue: [X], EV/EBITDA: [X]
Comparable transaction database: [paste or describe the comp set]
Analyze:
1. Where the proposed multiples fall in the distribution of comp multiples: at what percentile
2. Whether the target's financial profile (growth rate, margin) justifies any premium or discount to comp median
3. Which comparable transactions are most similar to this deal and why
4. What the analysis implies about whether the proposed price is aggressive, fair, or conservative
5. What factors might justify premium pricing above the comp range
Output: Multiple benchmarking table + valuation opinion with reasoned analysisPrompt 3 — Deal Terms Market Standard Analysis
Analyze the market standard deal terms for the following deal type.
Deal type: [acquisition of private technology company / PE buyout / cross-border pharmaceutical acquisition / etc.]
Geography: [deal law jurisdiction]
Deal size: [$X to $Y]
Time period: [last 3 / 5 years]
Extract market standard data for:
1. Termination fees: buyer breakup fee %, seller breakup fee %, reverse termination fee % (mean, median, range)
2. MAC definition: how broadly or narrowly is Material Adverse Change typically defined in this deal type
3. MAC carve-outs: what is typically carved out from MAC (industry conditions, regulatory changes, pandemics)
4. Representation and warranty: survival periods, knowledge qualifiers, indemnification caps as % of deal value
5. Walk rights: regulatory walk rights, financing walk rights, and any deal-specific conditions
6. Earnout usage: frequency of earnouts, typical size as % of deal value, typical milestone structure
Output: Deal terms market standard table with statistical analysis + outlier deals and their characteristicsPrompt 4 — Strategic Rationale Pattern Analysis
Analyze the strategic rationale patterns in the following comparable transaction set.
Comparable transactions: [paste or describe a set of 10–25 comparable deals]
Industry: [sector]
Time period: [covered by the comp set]
Analyze:
1. Strategic rationale categories: what proportion of deals were driven by each type of rationale (scale, capability, geography, product extension, defensive, financial)
2. Premium correlation: do certain strategic rationales correlate with higher or lower acquisition premiums
3. Integration approach patterns: what integration approaches are most common by rationale type
4. Success rate signals: for deals with available post-close data, which strategic rationale types appear to have performed best
5. Current trend: is the distribution of strategic rationales shifting over time
Output: Strategic rationale analysis report + premium correlation analysis + strategic pattern narrativePrompt 5 — Withdrawn Transaction Risk Analysis
Analyze failed or withdrawn transactions in the following deal universe for risk intelligence.
Industry: [sector]
Time period: [last 5 / 10 years]
Deal universe: [describe the set of transactions, including both completed and withdrawn]
For each withdrawn transaction, extract:
1. Announced deal details: parties, deal size, deal type, announced rationale
2. Withdrawal reason: regulatory / financing failure / MAC invocation / fiduciary out / hostile target defense / other
3. Timeline: time from announcement to withdrawal
4. Financial consequences: termination fee paid, share price reaction at withdrawal
5. Post-withdrawal outcome: did the deal eventually close with different terms, or was the company sold to a different buyer
Synthesize:
1. Most common failure reasons for this deal type in this industry
2. Warning signs present in failed deals at announcement that distinguish them from completed deals
3. Regulatory failure patterns: which regulatory bodies and deal characteristics most often trigger regulatory failure
4. Risk implication for the current deal: based on these patterns, what are the most likely failure risks today27. AI Social Listening & Online Sentiment Trend Tracker
Tracks brand, competitor, and topic sentiment across social platforms, forums, and review sites — delivering structured trend intelligence rather than raw mention counts.
Pain Point & How COCO Solves It
The Pain: Social Listening Tools Generate Volume Without Intelligence
Social listening has become standard practice in market research, competitive intelligence, and brand management. Most organizations have access to a social listening platform that delivers a daily feed of brand mentions and a sentiment score. But the gap between social mention data and actionable intelligence is vast, and most organizations do not bridge it. Analysts see that 60% of mentions are positive and 20% are negative and conclude that brand health is adequate — without understanding what specific issues are driving the negative sentiment, whether the negative sentiment is concentrated among high-value customers or low-influence users, whether the trend is worsening or improving, or what competitors' social sentiment looks like by comparison.
The signal extraction problem is the first dimension of the gap. Raw social mention volume and aggregate sentiment scores obscure the specific intelligence that makes social listening valuable. The 20% negative sentiment figure contains a distribution of distinct issues — product quality complaints, customer service failures, pricing objections, feature requests expressed as frustrations — each with different strategic implications and different urgency levels. Extracting this disaggregated view requires text analysis beyond keyword counting: categorizing the substance of conversations, identifying the issues driving each sentiment bucket, and tracking how the issue distribution within each sentiment tier changes over time.
The context and influence problem adds a second dimension. Not all social mentions carry equal weight. A complaint from a Twitter account with 50 followers and zero engagement is strategically irrelevant; the same complaint from an industry journalist with 100,000 followers is a potential reputational risk requiring a response. Standard social listening tools treat all mentions equivalently — missing the qualitative context that determines which signals actually matter for strategic decision-making.
How COCO Solves It
Multi-Platform Social Intelligence Collection: COCO builds comprehensive coverage across the relevant social landscape:
- Monitors Twitter/X, LinkedIn, Reddit, Facebook, Instagram, and TikTok for brand, competitor, and topic mentions
- Tracks niche platforms relevant to the specific audience: specialized professional forums, Discord communities, Slack groups with public archives
- Monitors review sites: G2, Gartner Peer Insights, Capterra, Trustpilot, Yelp, and sector-specific review platforms
- Tracks industry news comment sections, blog posts, and podcast episode discussion threads
- Applies platform-specific context weighting based on the relative importance of each platform for the target audience
Conversation Categorization and Topic Extraction: COCO structures the unstructured social data:
- Categorizes social mentions by topic: product feedback, customer service, pricing, competitive comparisons, brand reputation, thought leadership engagement
- Extracts specific issues within each category — distinguishing feature requests, performance complaints, usability issues, and reliability concerns within product feedback
- Identifies the frequency and severity distribution of issues within each topic category
- Tracks topic emergence: when new conversation themes appear that were not present in prior monitoring periods
- Generates a conversation topology map showing the distribution of discussion topics across all platforms
Sentiment Analysis Beyond Positive/Negative: COCO provides multi-dimensional sentiment intelligence:
- Applies fine-grained sentiment analysis: distinguishing enthusiastic positive, satisfied positive, neutral, frustrated negative, and angry negative
- Conducts emotion analysis alongside sentiment: identifying mentions that express confusion, excitement, disappointment, or advocacy
- Tracks sentiment by topic: overall sentiment for the brand may be positive while sentiment on a specific product feature is consistently negative
- Identifies sentiment drivers: the specific words, phrases, and topics most strongly associated with positive and negative sentiment
- Generates sentiment trend analysis: how is the overall sentiment score and its component dimensions changing over time
Influence and Reach Assessment: COCO contextualizes mentions by strategic significance:
- Applies influence scoring to mention sources: follower count, engagement rate, domain authority, and community role
- Identifies high-influence voices — journalists, analysts, influential customers, community leaders — and extracts their specific comments for prioritized review
- Calculates potential reach for each significant mention based on distribution and engagement
- Identifies when a conversation is spreading from high-influence to mainstream channels — an early warning of viral potential
- Generates an influence-weighted sentiment score that reflects the views of the most strategically significant voices
Competitive Social Benchmarking: COCO positions social intelligence in competitive context:
- Monitors sentiment and conversation patterns for defined competitors simultaneously with the primary brand
- Calculates share of voice: what proportion of conversations in the topic space mention each brand
- Compares sentiment distributions across competitors: who has more positive sentiment and on what specific topics
- Identifies competitive weaknesses in social conversations: where a competitor's negative sentiment is concentrated provides targeting intelligence
- Tracks competitive social positioning shifts over time: who is gaining or losing ground in the social conversation
Structured Social Intelligence Report Generation: COCO produces formatted, actionable deliverables:
- Generates a weekly social intelligence digest: top topics, sentiment trends, high-influence mentions, and emerging conversations
- Produces a monthly social performance report with trend analysis, competitive benchmarking, and strategic implications
- Creates a crisis early warning brief when sentiment deteriorates rapidly or a high-influence negative mention spreads
- Generates competitive social intelligence summaries for sales and product teams
- Produces audience insight reports from social conversation analysis — what the target audience cares about, in their own words
Results & Who Benefits
Measurable Results
- Social intelligence processing time: Reduced from 4–8 hours weekly of manual monitoring and analysis to under 1 hour reviewing COCO's structured digest (85% time savings)
- Topic granularity: COCO identifies an average of 12–18 distinct conversation topics vs. 3–5 in typical aggregate sentiment reporting
- Early warning effectiveness: Organizations using COCO's influence-weighted monitoring detect emerging reputation issues an average of 4–7 days earlier than those using aggregate sentiment scores
- Competitive intelligence extraction: Social competitive analysis surfaces an average of 3–4 actionable competitive insights per month that inform product or marketing strategy
- Crisis detection accuracy: COCO's multi-signal crisis early warning system achieves 89% true positive rate for significant reputation events
Who Benefits
- Research Analysts: Access structured, intelligence-grade social insight rather than raw mention counts — enabling genuine social intelligence integration into research deliverables
- Brand and Communications Teams: Monitor brand health with the granularity and early warning capability to respond proactively rather than reactively to reputation developments
- Product Teams: Receive organized, prioritized social feedback on specific product features and use cases — supplementing formal research with real-world user expression
- Competitive Intelligence Teams: Track competitor brand and product perception systematically — identifying competitive vulnerabilities and shifts in market positioning before they appear in formal research
💡 Practical Prompts
Prompt 1 — Social Listening Dashboard Setup
Design a social listening monitoring framework for the following brand/topic universe.
Primary brand to monitor: [name]
Competitors to monitor: [list 3–6]
Topics/themes to track: [product categories, key issues, industry terms]
Target audience profile: [describe who is in the most strategically important conversations]
Geographic scope: [global / specific regions]
Platform priority: [list platforms in order of importance for this audience]
Design the monitoring framework:
1. Search query set: keywords, phrases, and Boolean operators for each monitored entity/topic
2. Platform allocation: which platforms to prioritize and what monitoring approach for each
3. Influencer watchlist: specific accounts or sources to monitor at the individual level
4. Topic taxonomy: categories and sub-categories for conversation classification
5. Alert triggers: conditions that should generate an immediate alert
6. Reporting cadence and format specifications
Output: Social listening specification document ready for implementationPrompt 2 — Social Sentiment Deep-Dive Analysis
Conduct a deep-dive sentiment analysis for the following brand over the specified period.
Brand: [name]
Analysis period: [date range]
Social data: [paste mention samples or describe the dataset available]
Benchmark: [prior period / competitor / industry average]
Analyze:
1. Overall sentiment distribution: % positive, neutral, negative with trend vs. benchmark period
2. Sentiment by platform: how does sentiment vary across platforms
3. Topic-level sentiment: for each major conversation topic, what is the sentiment distribution
4. Sentiment drivers: the specific words, phrases, and topics most strongly associated with positive and negative sentiment
5. High-influence sentiment: what are the most influential voices saying
6. Emerging sentiment shifts: has any specific topic or platform shown a meaningful sentiment change
Output: Sentiment analysis report with topic breakdown + trend charts + strategic implicationsPrompt 3 — Competitive Social Share of Voice Analysis
Analyze the social share of voice and sentiment positioning for the following brands.
Brands to compare: [list 4–8 companies]
Industry/topic space: [the conversation universe to measure share of voice within]
Analysis period: [date range]
Social data: [describe the data available or paste sample mentions for each brand]
For each brand:
1. Share of voice: % of total relevant conversations that mention this brand
2. Overall sentiment distribution: positive / neutral / negative %
3. Topic focus: what topics dominate each brand's social conversations
4. Sentiment by topic: where each brand has the strongest and weakest sentiment performance
Cross-brand analysis:
1. Share of voice ranking with trend (gaining / losing / stable)
2. Sentiment leader and laggard identification by topic
3. Competitive gap analysis: where does the primary brand have the most to gain
4. Audience overlap: which audiences discuss multiple brands and what do they say comparatively
Output: Share of voice comparison table + sentiment positioning matrix + strategic implicationsPrompt 4 — Reputation Crisis Early Warning Analysis
Assess the following social signals for potential reputation crisis indicators.
Brand: [name]
Monitoring period: [date range]
Social data: [describe unusual signals observed — volume spikes, sentiment drops, specific topics]
Crisis indicator assessment:
1. Volume anomaly: is the mention volume unusual relative to the baseline
2. Sentiment shift: has sentiment deteriorated significantly — in what time window and by how much
3. Topic emergence: has a new negative topic appeared that was not present before
4. Influencer activation: are high-influence accounts driving the conversation
5. Cross-platform spread: is the conversation contained to one platform or spreading
6. Media bridge: is social discussion leading to or following traditional media coverage
Crisis severity assessment: early warning / developing situation / active crisis / post-peak
Recommended response: no action needed / monitor closely / prepare statement / immediate response required
If action recommended, draft a suggested initial response and monitoring escalation protocol.Prompt 5 — Social Audience Insight Extraction
Extract audience insights from social conversations about [topic/category].
Topic/category: [describe the market, product category, or issue]
Target audience: [describe who is in these conversations]
Social data: [paste or describe the social conversation data available]
Insight objectives: [what do you want to learn — customer needs, pain points, vocabulary, decision factors]
Extract:
1. Key themes: what topics dominate the audience's conversations in this space
2. Vocabulary: what specific words and phrases do they use — how does this differ from industry jargon
3. Pain points: what frustrations and unmet needs appear most frequently
4. Jobs to be done: what are they trying to accomplish when they discuss this topic
5. Influencer landscape: whose voices are most listened to in these conversations
6. Decision signals: when purchasing or evaluating options, what criteria do they discuss most
Format insights as direct language that can be incorporated into a research deliverable or strategic brief.28. AI Industry Association Report Synthesizer
Extracts the strategic intelligence buried in industry association reports, trade publications, and conference proceedings — synthesized across dozens of documents into a single coherent industry briefing.
Pain Point & How COCO Solves It
The Pain: Industry Association Intelligence Is Underexploited Because Extracting It Is Too Time-Consuming
Industry associations produce some of the richest, most authoritative intelligence available on sector-specific market conditions, regulatory developments, technology trends, and workforce dynamics. Trade association annual reports, conference proceedings, research white papers, and member surveys often contain data not available in any commercial database: industry-wide volume statistics that no individual company would disclose, regulatory policy positions that reveal where the industry is heading, workforce and skills data that leads economic indicators, and technology adoption rates from survey samples that span the full industry. For analysts covering specific sectors in depth, these sources are invaluable — but their value is systematically underextracted.
The access and discovery problem is the first barrier. Industry associations publish through multiple channels: their own websites (often poorly organized), email newsletters sent to members, conference proceeding PDFs that are not well-indexed in search engines, and trade publications with inconsistent online availability. An analyst who does not know which associations are relevant to a specific industry, or who lacks institutional membership access, simply does not encounter this intelligence. The research organizations that exploit association intelligence effectively do so because individual analysts have deep domain knowledge and personal connections — advantages that are difficult to scale.
The synthesis problem is the second and more significant barrier. Even analysts who access the full range of relevant association publications face the challenge of synthesizing dozens of documents — some hundreds of pages long — into actionable intelligence for a research report. A thorough synthesis of annual reports from five major trade associations, three major conference proceedings, and twelve policy comment letters in a single industry might require a week of reading and note-taking before a word of the research deliverable is written.
How COCO Solves It
Industry Association Landscape Mapping: COCO identifies the full range of relevant associations for any sector:
- Maps major national and international trade associations for the specified industry, including sector-specific sub-associations
- Identifies professional associations, standards bodies, and policy advocacy organizations relevant to the coverage area
- Catalogs the key publication types each association produces: annual reports, member surveys, conference proceedings, policy white papers, workforce reports
- Maintains awareness of conference schedules for major industry events and their associated publication releases
- Identifies open-access vs. member-only content and notes access pathways for restricted content
Multi-Document Intelligence Extraction: COCO processes association publications systematically:
- Ingests PDFs and web content from multiple association publications simultaneously
- Extracts key data: market statistics, member survey results, regulatory positions, technology adoption data, workforce and compensation benchmarks
- Identifies the methodology behind each statistic — sample size, survey population, collection methodology
- Flags when the same statistic appears in multiple association publications (confirming data) vs. when associations report conflicting figures
- Organizes extracted intelligence by category: market data, regulatory intelligence, technology trends, workforce data, policy positions
Cross-Publication Pattern Synthesis: COCO builds coherent sector intelligence from fragmented sources:
- Identifies where multiple associations are reporting on the same underlying trends from different vantage points
- Synthesizes the emerging industry consensus on key issues: where multiple associations agree vs. where there is internal industry disagreement
- Tracks changes in association positions over time: where is the industry's collective view evolving on technology, regulation, or workforce
- Identifies gaps in the association intelligence landscape: important topics that no association has addressed systematically
- Generates a unified sector narrative that integrates intelligence from all relevant association sources
Regulatory and Policy Intelligence: COCO extracts the policy intelligence embedded in association publications:
- Extracts association positions on proposed regulations from comment letters, policy white papers, and advocacy publications
- Identifies where associations are supporting, opposing, or proposing modifications to specific regulatory provisions
- Tracks where the industry has consensus on regulatory priorities vs. where member companies have divergent interests
- Monitors for shifts in association regulatory positions — where the industry is changing its approach to policy engagement
- Synthesizes the association regulatory intelligence with actual regulatory developments to assess likely industry-regulator dynamics
Industry Data and Benchmark Compilation: COCO organizes association data into usable research inputs:
- Compiles industry volume and revenue statistics across association sources, with reconciliation of discrepancies
- Builds a benchmark database from association compensation, pricing, and operational surveys
- Extracts technology adoption and capability data from member surveys for use in market sizing and competitive analysis
- Aggregates association M&A and deal volume data for sectors where associations systematically track deal activity
- Normalizes and time-series-enables association data for trend analysis across multiple years of publications
Industry Association Intelligence Report: COCO produces structured deliverables:
- Generates an annual industry intelligence brief synthesizing the most important findings across all major association publications for the year
- Produces a regulatory intelligence summary: the industry's stated positions and the regulatory developments they are most actively tracking
- Creates a sector data compilation: key statistics from association sources with methodology documentation
- Generates a conference intelligence brief after major industry conferences: key themes, presentations, and sentiment signals
- Produces a policy intelligence update tracking association regulatory positions and advocacy activity
Results & Who Benefits
Measurable Results
- Association intelligence extraction time: Reduced from 1–2 analyst weeks of reading and synthesis to 2–3 days for a comprehensive association intelligence sweep (70–80% time savings)
- Publication coverage: COCO synthesizes an average of 15–20 association publications per sector sweep vs. 3–5 in typical analyst research processes
- Unique data extraction: COCO surfaces an average of 4–7 unique data points per sector sweep that are not available in commercial research databases
- Regulatory intelligence coverage: Proportion of research reports with comprehensive industry regulatory position intelligence increases from 25% to 85% when COCO systematically extracts association comment letters and white papers
- Conference intelligence production: Major conference synthesis reports produced within 48 hours of conference conclusion vs. 1–2 weeks for manual synthesis
Who Benefits
- Research Analysts: Access the full depth of association intelligence for their coverage sectors — exploiting a rich but chronically underutilized information source
- Industry Specialists and Consultants: Demonstrate deep sector knowledge by incorporating association data and industry consensus positions that generalists miss
- Policy Researchers and Government Affairs Teams: Track industry positions on regulatory issues systematically — understanding what the industry collectively believes and why
- Investment Analysts: Incorporate industry association data — including statistics not available in any commercial database — into sector and company analysis
💡 Practical Prompts
Prompt 1 — Industry Association Intelligence Sweep
Conduct an industry association intelligence sweep for the following sector.
Sector: [industry name]
Geographic scope: [US / global / specific region]
Intelligence priorities: [market data / regulatory positions / technology trends / workforce / all]
Recency requirement: [last 12 / 24 / 36 months of publications]
Identify and synthesize from all major associations:
1. Key market statistics: industry revenue, volume, growth rates, and member composition
2. Regulatory and policy developments: what are associations advocating for, opposing, or monitoring
3. Technology adoption data: what technologies are members adopting, at what rates, and what is driving/hindering adoption
4. Workforce and compensation benchmarks: hiring trends, skills in demand, compensation data
5. Industry challenges: what issues are dominating association agenda and member concern
6. Strategic trends: emerging business models, competitive dynamics, or structural changes being collectively identified
Output: Industry association intelligence brief with all data sourced and attributedPrompt 2 — Regulatory Position Intelligence
Extract the industry's collective regulatory position on the following issue or proposed regulation.
Regulation/policy issue: [describe the regulatory development]
Industry: [sector]
Associations to analyze: [list specific associations or ask COCO to identify the relevant ones]
Extract:
1. Each association's stated position: support / oppose / support with modifications / neutral
2. Specific provisions each association is supporting or opposing — with their reasoning
3. The evidence and arguments associations are using to support their positions
4. Areas of association consensus: where the full industry appears to agree
5. Areas of association disagreement: where different member segments have conflicting interests
6. Lobbying intensity signals: how actively is the industry engaging on this issue
Synthesize:
1. The industry's net position and how likely it is to influence the final regulatory outcome
2. The most persuasive arguments the industry is making and their likely effectiveness with regulators
3. What the association positions reveal about where industry compliance costs will be concentratedPrompt 3 — Conference Intelligence Brief
Generate a conference intelligence brief from the following industry conference.
Conference: [name and date]
Publications/presentations available: [describe what is available — agenda, keynote transcripts, panel summaries, published white papers]
Key themes to extract: [specify focus areas or ask COCO to identify the dominant themes]
Structure the brief as:
1. Top 3 themes: the ideas that dominated the conference with supporting evidence from multiple sessions
2. Consensus view: what the industry broadly agrees on based on conference discussion
3. Contentious debates: where there was significant disagreement or active debate
4. Emerging topics: issues appearing on the conference agenda for the first time or gaining new prominence
5. Notable statements: specific comments from influential speakers that signal new thinking or positions
6. Regulatory and policy signals: what regulatory developments were discussed and how the industry is responding
Compare to: [prior year conference or stated industry objectives] — how has the conversation evolvedPrompt 4 — Industry Data Compilation
Compile a comprehensive industry data set from association publications for the following sector.
Sector: [industry]
Data categories needed:
1. Market size and growth: total industry revenue/volume with historical trend
2. Participant landscape: number of companies, concentration ratios, market structure
3. Technology adoption: penetration rates for key technologies in use
4. Workforce: employment levels, growth trends, compensation benchmarks, skills in demand
5. Financial benchmarks: revenue per employee, margins, pricing benchmarks from member surveys
For each data point:
- Source: association name, report title, publication date
- Methodology: survey sample, coverage scope, collection method
- Confidence: primary data / aggregated from member reports / estimated
- Recency: data year or collection period
Identify data gaps: important data categories where no reliable association source exists and recommend alternative sources.Prompt 5 — Association Landscape Mapping
Map the industry association landscape for the following sector.
Sector: [industry]
Geographic scope: [US / global / specific region]
Identify:
1. Major trade associations: name, membership profile, size, and primary focus
2. Professional associations: credentials-granting bodies, practitioner organizations
3. Standards bodies: technical standards organizations relevant to the industry
4. Policy advocacy groups: associations focused primarily on regulatory and legislative advocacy
5. International organizations: relevant global or regional bodies
For each association, document:
- Website and primary publication channels
- Member survey and research publication cadence
- Most recent significant publications and their topics
- Association's current policy priorities
- Whether publications are open-access or member-only
Output: Association landscape map organized by type, with access notes and publication calendar29. AI Country Market Entry Feasibility Analyzer
Evaluates the feasibility, attractiveness, and entry strategy for market entry into any country — synthesizing regulatory, competitive, economic, and cultural dimensions into a structured assessment.
Pain Point & How COCO Solves It
The Pain: Country Market Entry Analysis Is Expensive, Slow, and Often Conducted After the Strategic Decision Has Already Been Made
Market entry decisions are among the most consequential strategic choices a company makes — committing resources to a new geography with all the regulatory compliance, operational complexity, and competitive uncertainty that entails. The research required to make these decisions well is substantial: understanding the regulatory and legal framework for foreign business operations, assessing the competitive landscape in the local market, evaluating the macroeconomic environment, understanding local customer preferences and business culture, and identifying the operational requirements for establishing and scaling a local presence. Comprehensive market entry research typically requires 4–8 weeks of analyst time, local market expertise, and in many cases, local market research fielding — making it one of the most expensive research projects a firm undertakes.
The timing problem is the most damaging aspect of how market entry research is currently practiced. Because thorough research is expensive and slow, organizations often commission it only after senior leadership has already decided they want to enter a market — using the research to validate a decision already made rather than to inform it. Research conducted in this mode is subject to confirmation bias: analysts unconsciously look for evidence that supports the entry thesis rather than systematically evaluating the case on its merits. The strategic risks that would argue against entry — high regulatory barriers, entrenched local competitors, demand dynamics that favor local players — are acknowledged but not given full weight.
The knowledge transfer problem compounds the research quality problem. Market entry research depends heavily on local knowledge — understanding how business is actually done in a specific country, which regulatory interpretations are standard vs. contested, and which competitive dynamics are visible in public sources vs. which require local network insight. This knowledge is expensive to acquire and does not transfer systematically between engagements. Each engagement requires substantial new research investment.
How COCO Solves It
Market Attractiveness and Demand Assessment: COCO builds the demand-side case for market entry:
- Synthesizes macroeconomic data: GDP, GDP per capita, growth trajectory, consumer spending trends, and middle-class size and growth
- Analyzes sector-specific market size and growth for the target product or service category
- Assesses demand timing: is the market at an adoption stage where entry is premature, optimal, or late
- Evaluates infrastructure readiness: does the physical and digital infrastructure support the business model
- Generates a demand attractiveness score combining market size, growth trajectory, and adoption readiness
Regulatory and Legal Framework Assessment: COCO maps the operating requirements for the market:
- Assesses foreign ownership restrictions: what ownership structures are permitted for foreign businesses in this sector
- Maps licensing, certification, and registration requirements for operating in the sector
- Evaluates data localization, privacy, and cybersecurity requirements
- Assesses labor law framework: hiring flexibility, termination rules, mandatory benefits, and local content requirements
- Identifies sector-specific regulatory bodies and their current enforcement posture
Competitive Landscape Analysis: COCO assesses the local competitive environment:
- Identifies incumbent competitors: local players and existing international entrants with market share estimates
- Assesses local player advantages: government relationships, distribution networks, brand recognition, and cost structure
- Evaluates the competitive intensity and market concentration — is there room for a new entrant to gain share
- Identifies which customer segments are under-served and which are heavily contested
- Assesses the likely competitive response to entry — how aggressive will incumbents be and what tools do they have
Entry Mode and Structure Analysis: COCO evaluates different entry approaches:
- Assesses feasibility and implications of each entry mode: wholly owned subsidiary, joint venture, franchising, distribution partnership, licensing, acquisition
- Evaluates the trade-offs of each mode: control vs. speed, capital intensity vs. flexibility, regulatory compliance vs. operational freedom
- Identifies potential local partners — by type and profile — for joint venture or partnership entry approaches
- Assesses M&A target landscape for entry via acquisition
- Generates a recommended entry mode hierarchy with rationale and risk/benefit comparison
Operating Environment and Risk Assessment: COCO evaluates practical operational challenges:
- Assesses talent availability and cost for the required functions
- Evaluates local supply chain availability, quality, and cost for the business model's key inputs
- Assesses logistics and distribution infrastructure relevant to the business
- Quantifies political risk, currency risk, and macroeconomic stability risk for the entry horizon
- Generates an operating risk register with probability, impact, and mitigation assessment for the top risks
Market Entry Feasibility Report: COCO produces structured, decision-ready deliverables:
- Generates a market entry scorecard: attractiveness vs. feasibility matrix with overall entry recommendation
- Produces a full market entry feasibility report with all assessed dimensions, supporting evidence, and strategic recommendation
- Creates a phased entry roadmap: recommended sequence of activities from initial regulatory filing to full commercial operations
- Generates a country comparison analysis when multiple markets are being evaluated simultaneously
- Produces an investment requirement estimate: capital and operating cost implications of the recommended entry approach
Results & Who Benefits
Measurable Results
- Market entry research cycle time: Reduced from 4–8 analyst weeks to 1–2 weeks for a comprehensive market entry feasibility assessment (75% time savings)
- Regulatory dimension coverage: COCO-produced entry assessments cover an average of 11–14 regulatory dimensions vs. 5–7 in typical analyst-produced country assessments
- Entry mode analysis depth: Proportion of market entry assessments including a formal entry mode comparison increases from 30% to 90% with COCO's structured framework
- Decision quality: Organizations using COCO market entry assessments report 40% fewer unexpected operational and regulatory issues in the first year of market operation
- Parallel market comparison: COCO enables simultaneous feasibility assessment across 5–8 candidate markets in the time typically required for a single-market deep-dive
Who Benefits
- Research Analysts: Deliver comprehensive market entry feasibility assessments that cover all critical dimensions systematically
- Corporate Development and Strategy Teams: Evaluate market entry opportunities with the analytical rigor needed to make confident go/no-go decisions
- Investment Analysts: Assess international expansion plans in target company investment theses — understanding whether the entry strategy is realistic
- Government and Economic Development Organizations: Understand what is making their country attractive or unattractive to foreign investors
💡 Practical Prompts
Prompt 1 — Country Market Entry Feasibility Assessment
Conduct a market entry feasibility assessment for the following opportunity.
Company profile: [describe the company — industry, business model, current markets]
Target country: [name]
Product/service to introduce: [description]
Target customer segment: [enterprise / SMB / consumer / specific vertical]
Entry timeline: [when is entry being considered]
Investment appetite: [capital available for market entry]
Assess:
1. Market attractiveness: demand size, growth trajectory, adoption readiness, infrastructure
2. Regulatory framework: foreign ownership rules, sector-specific licenses, data/privacy requirements, labor law
3. Competitive landscape: key incumbents, market concentration, under-served segments, likely competitive response
4. Entry mode options: feasibility and implications of each approach
5. Operating environment: talent availability, supply chain, logistics, currency and political risk
6. Investment requirements: estimated capital and operating costs for recommended entry approach
Overall recommendation: enter / conditional entry / defer / avoid — with 3-sentence rationale
Output: Market entry feasibility brief + scorecard + phased entry roadmapPrompt 2 — Multi-Country Market Comparison
Compare market entry attractiveness across the following countries for the same product/service.
Company and product: [describe]
Countries to compare: [list 4–8 candidate markets]
Evaluation criteria weights: [if you have preferences — e.g., regulatory simplicity most important]
For each country, score (1–5) and provide brief evidence for:
1. Market size and growth potential
2. Regulatory ease and foreign business welcome
3. Competitive intensity (lower is better for new entrants)
4. Talent and operational infrastructure
5. Political and macroeconomic stability
6. Cultural and business environment fit
Generate:
1. Country comparison matrix with scores on all dimensions
2. Overall attractiveness ranking with key differentiating factors
3. Top 2 recommended markets with strategic rationale
4. Important trade-offs: the dimension on which the top-ranked market is weakest
5. Sequencing recommendation: if entering multiple markets over time, what order is optimalPrompt 3 — Entry Mode Analysis
Evaluate the entry mode options for the following market entry scenario.
Company: [description]
Target market: [country]
Product/service: [description]
Constraints: [e.g., foreign ownership cap, regulatory approval requirements, capital limitations, speed requirements]
Evaluate each feasible entry mode:
1. Wholly owned subsidiary: requirements, timeline, cost, control level, regulatory considerations
2. Joint venture: what type of local partner is needed, what they contribute, governance challenges
3. Distribution/sales partnership: partner profile needed, control trade-offs, transition risks
4. Franchising or licensing: applicability to this business model, IP protection considerations
5. Acquisition: local M&A target landscape (type of company, not specific names), premium for speed
For each mode:
- Feasibility: is this mode actually available given constraints
- Speed to market: how quickly can this mode generate revenue
- Capital requirement: approximate investment needed
- Control and flexibility: how much operational control and strategic flexibility does this mode provide
- Risk profile: what are the main risks specific to this entry mode in this market
Recommendation: preferred entry mode with rationale, and contingency if the preferred mode faces obstaclesPrompt 4 — Regulatory Compliance Framework for Market Entry
Map the regulatory compliance requirements for establishing and operating [business type] in [country].
Business description: [what the company does — products/services, operational model, data handling]
Entry structure: [proposed entity type — wholly owned subsidiary, branch office, JV, etc.]
Industry sector: [relevant for sector-specific regulations]
Map compliance requirements across:
1. Entity establishment: registration requirements, minimum capital, shareholder structure, timeline
2. Sector-specific licenses and permits: what approvals are required before operating, from which agencies
3. Foreign investment approvals: whether investment review board approval is required
4. Employment law: hiring requirements, mandatory benefits, work permit requirements for expatriates
5. Tax obligations: corporate tax rate, VAT/GST, withholding tax, transfer pricing requirements
6. Data and privacy: data localization requirements, cross-border data transfer restrictions
7. Intellectual property: registration requirements, IP protection quality in this jurisdiction
8. Ongoing compliance: annual reporting, audit requirements, regulatory filings
Output: Compliance requirements checklist + timeline + estimated compliance costPrompt 5 — Competitive Entry Threat Assessment
Assess the likely competitive response to entry into [country market] for [company type].
Company entering: [brief profile — industry, business model, apparent competitive advantages]
Target market: [country and segment]
Key incumbents: [list known local and international competitors already in the market]
For each key incumbent:
1. Their likely awareness of the entrant: will they know about entry early or late
2. Their motivation to respond aggressively: how much does this market matter to their strategy
3. Their response capabilities: what tools do they have (pricing, partnerships, regulatory influence)
4. Their likely response strategy: price war / partnership blocks / regulatory engagement / product acceleration
5. Response timeline: how quickly can they mount an effective competitive response
Overall competitive environment assessment:
1. How receptive is the market to a new entrant
2. Which incumbent represents the greatest competitive threat and why
3. What pre-entry actions would most reduce competitive response intensity
4. What first-mover advantages can the entrant establish before incumbents respond effectively30. AI Focus Group Discussion Guide Builder
Designs professional focus group discussion guides with hypothesis-driven questioning, probe strategies, and projective techniques — tailored to the research objective and group profile.
Pain Point & How COCO Solves It
The Pain: Focus Group Quality Is Highly Dependent on Discussion Guide Quality — Which Is Chronically Under-Invested
Focus groups are one of the most widely used and most frequently misused qualitative research methods available. When designed well — with a carefully structured discussion guide, skilled moderation, and hypothesis-driven questioning — they produce rich insights into consumer motivation, product perception, and decision-making that cannot be obtained through quantitative research alone. When designed poorly — with leading questions, unbalanced structure, or a guide that asks participants to articulate what they cannot articulate directly — they produce surface-level opinions that confirm what the research sponsor already believes and add nothing to their understanding.
The discussion guide development problem is the root cause of poor focus group outcomes. Developing a truly good discussion guide requires several skills simultaneously: qualitative research methodology (understanding what question types work for what insights), domain expertise (knowing what is genuinely unknown and what the participant population can credibly address), facilitation intuition (understanding how group dynamics affect what gets said), and strategic framing (knowing what the research sponsor needs to learn vs. what would merely be interesting). Few research analysts have all these skills at full strength, and the guides that result from solo development tend toward the most accessible approach — direct questions about preferences and behaviors — at the expense of the projective, indirect, and laddering techniques that often produce the most valuable insights.
The time pressure problem compounds the methodology problem. Discussion guides are typically developed in the final days before a focus group session, often while the analyst is simultaneously handling other project components. The compressed timeline favors simple, familiar question structures over the more thoughtful, methodology-informed approaches that produce better data. Piloting the guide before sessions rarely happens under this time pressure. The result is a guide that is field-tested for the first time with a live participant group, at which point problems are expensive to correct.
How COCO Solves It
Research Objective to Discussion Architecture: COCO builds the structural framework of the discussion guide:
- Translates research objectives into specific questions the group discussion must answer
- Designs the overall flow: warm-up and rapport building → topic introduction → core exploration → depth probing → projective exercises → closing and synthesis
- Calibrates the time allocation across sections based on topic complexity and the session length
- Identifies the two or three most important discussion moments and designs them with the most methodological care
- Sequences questions to build from accessible topics (behavior, usage) to less accessible ones (motivation, values) as the session progresses
Question Type Selection and Bias Prevention: COCO applies qualitative research methodology rigor:
- Selects appropriate question types for each information need: open-ended exploration, specific probing, ranking and prioritization, hypothetical scenarios
- Flags leading questions, loaded language, and presupposition in question drafts — replacing with neutral alternatives
- Identifies questions that are asking for information participants cannot self-report accurately
- Recommends indirect and projective approaches for topics where direct questioning produces unreliable responses
- Builds in multiple approaches to the same research question — asking about the same topic from different angles to triangulate responses
Projective and Laddering Techniques: COCO incorporates advanced qualitative methods:
- Selects appropriate projective techniques for the research objective: personification, brand mapping, collage creation, sentence completion, role play
- Designs laddering sequences for understanding motivation: starting from product attributes → consequences → values
- Creates metaphor elicitation questions that access non-verbal associations
- Designs sorting and ranking exercises for understanding how participants categorize options
- Develops hypothetical scenario questions for eliciting views on future concepts without direct opinion polling
Probe Strategy and Depth Facilitation: COCO equips moderators with probing depth:
- Generates general-purpose probes for each section: what to ask to deepen any response
- Develops section-specific probes calibrated to deepen the most important insights
- Creates probe decision trees: if a participant says X, probe with Y; if they say Z, probe with W
- Identifies the too-short and too-easy responses for each question and prepares the moderator to productively challenge surface-level answers
- Develops devil's advocate statements the moderator can introduce to surface the range of views beyond group consensus
Group Dynamic Management: COCO addresses the social dynamics that affect focus group quality:
- Designs opening exercises that establish equal voice and reduce the tendency for early-speaker positions to anchor the group
- Creates independent response exercises (written before verbal discussion) for sensitive topics where social desirability bias is a risk
- Generates silent exercise formats for portions of the discussion where individual variation is important
- Develops moderator guidance for managing dominant participants, drawing out quiet participants, and managing group polarization
- Creates check-in and synthesis exercises that ensure the moderator captures the full range of views
Discussion Guide Document and Moderator Brief: COCO produces complete, field-ready deliverables:
- Generates a formatted discussion guide with timing, question text, probe notes, and exercise instructions
- Produces a moderator brief: research objectives, key hypotheses to test, must-answer questions, and flexibility guidance
- Creates a backroom observation guide for client observers: what to watch for and how to interpret responses
- Generates a post-session debrief template for capturing moderator and observer insights immediately after each session
- Produces an analysis framework aligned with the discussion guide structure for consistent coding across sessions
Results & Who Benefits
Measurable Results
- Discussion guide development time: Reduced from 2–4 days to 4–8 hours for a full focus group discussion guide (70–80% time savings)
- Methodological quality: COCO-developed guides rated "methodologically strong" by independent research quality review in 88% of cases vs. 42% for guides developed without structured support
- Projective technique adoption: Proportion of discussion guides including projective or indirect questioning techniques increases from 22% to 79% when COCO provides methodology guidance
- Insight yield: Client researchers report 35–50% more novel insights from groups using COCO-structured guides compared to their previous guide quality level
- Moderator preparation quality: Moderators using COCO-generated probe strategy documents report higher confidence in managing difficult discussion moments in 91% of cases
Who Benefits
- Research Analysts: Design professional-quality focus group guides that apply rigorous qualitative methodology — without needing a senior qualitative specialist to review every guide
- Market Research Teams: Standardize discussion guide quality across team members and client engagements — reducing variance driven by individual methodological experience
- Qualitative Research Specialists: Use COCO to accelerate guide development for the structural and tactical elements, freeing specialist time for the most complex methodological decisions
- Client-Side Research Teams: Commission and evaluate focus group research with the ability to assess discussion guide quality
💡 Practical Prompts
Prompt 1 — Full Focus Group Discussion Guide
Design a focus group discussion guide for the following research.
Research objective: [what you need to learn from these groups]
Target participants: [describe the group profile — demographics, behaviors, relationship to the topic]
Session length: [60 / 90 / 120 minutes]
Key hypotheses to test: [what you currently believe, which the groups should confirm or challenge]
Topics to cover: [list the major subject areas]
Sensitive or difficult topics (if any): [topics requiring indirect approaches]
Design a full discussion guide including:
1. Welcome and warm-up section (10-15 min)
2. Topic introduction and context setting (10 min)
3. Core discussion sections for each major topic (allocate time based on priority)
4. Projective exercises (2–3 exercises appropriate to the research objective)
5. Synthesis and priority-setting activity (10 min)
6. Closing and final comments (5 min)
For each section, provide: question text + probe suggestions + estimated timing + moderation notesPrompt 2 — Projective Exercise Design
Design projective exercises for the following focus group research.
Research topic: [what the groups are exploring]
Participant profile: [describe the group members]
Specific insight gap: [what direct questioning typically fails to reveal about this topic]
Session length constraint: [how much time is available for projective work]
Design 3 projective exercises:
For each exercise:
1. Exercise name and type (personification / brand mapping / sentence completion / collage description / role play / analogy)
2. Purpose: what specific insight this exercise is designed to surface that direct questioning cannot
3. Materials needed: if any
4. Instructions to participants: exact wording of how to explain the exercise
5. Discussion questions after the exercise: how to debrief and extract the insight
6. Interpretation guide: what responses mean what — how to analyze the output
7. Time requirement: how long this exercise takes from introduction to debrief
Sequence the exercises in the order that builds rapport and depth most effectively.Prompt 3 — Probe Strategy Development
Develop a probe strategy for the following focus group section.
Discussion topic: [describe the subject matter]
Primary questions for this section: [list the 3–5 main questions]
Common surface-level responses to avoid: [describe what participants are likely to say first that is not sufficiently deep]
Key insight goals: [what specific understanding must this section produce]
For each primary question, generate:
1. The standard probe: a follow-up that deepens any response without leading
2. The "too short" probe: what to ask when a participant gives a one-sentence answer
3. The "too abstract" probe: what to ask when a participant responds in general terms
4. The "groupthink" probe: what to ask when the group has converged on a consensus that may not reflect full individual variation
5. The clarification probe: for the most common ambiguous terms participants might use
Provide moderator guidance notes for managing this section's dynamics.Prompt 4 — Online/Virtual Focus Group Adaptation
Adapt the following focus group discussion guide for an online/virtual format.
Original discussion guide: [paste or describe the in-person guide]
Virtual platform: [Zoom / Teams / dedicated qualitative platform like Recollective or QualBoard]
Session length for virtual: [typically 90 min for virtual, shorter than in-person]
Participant profile: [describe — their comfort with technology, engagement risk level]
Adapt:
1. Timing adjustments: virtual sessions tire faster — where to shorten while preserving key sections
2. Engagement techniques: how to maintain participation in the absence of in-person social dynamics
3. Visual exercise adaptations: how to run board/card exercises in virtual format
4. Independent response adaptations: how to use polls, chat, or digital whiteboards for simultaneous input
5. Managing dominant participants: platform-specific techniques for virtual group dynamic management
6. Technology contingency: what to do when a participant has technical difficulties
Generate the adapted discussion guide with virtual-specific moderation notes integrated throughout.Prompt 5 — Post-Session Analysis Framework
Create an analysis framework for the following focus group research program.
Research objectives: [list]
Discussion guide structure: [paste or describe the guide's major sections and questions]
Number of groups: [how many groups will be conducted]
Group variations (if any): [do different groups represent different segments]
Create:
1. Coding framework: the categories and sub-categories for systematic coding of group recordings/transcripts
2. Cross-group comparison template: a matrix for comparing responses to key questions across all groups
3. Hypothesis tracking sheet: for each research hypothesis, a space to record supporting and contradicting evidence from each group
4. Key insight capture template: a structured format for recording the most important insights from each session immediately after it concludes
5. Cross-session synthesis guide: how to integrate findings across all groups into a coherent set of findings
Format as ready-to-use analysis documents for the research team31. AI Systematic Literature Review Automation Engine
Searches, screens, extracts, and synthesizes evidence across thousands of academic papers — producing a PRISMA-compliant literature review in days rather than months.
Pain Point & How COCO Solves It
The Pain: Systematic Literature Reviews Consume Months of Researcher Time with High Error Risk
Systematic literature reviews are the gold standard of evidence synthesis in medicine, public health, social science, and policy research — and among the most labor-intensive research outputs any analyst or academic team produces. A rigorous review begins with a comprehensive database search across PubMed, Embase, CINAHL, Scopus, Web of Science, and gray literature repositories, generating thousands of records that must be screened twice — first by title and abstract, then by full text — against pre-defined eligibility criteria. Each screening step is supposed to be performed by two independent reviewers, with disagreements adjudicated by a third, to minimize selection bias. For a mid-scope review, this screening process alone consumes 200–400 analyst hours before any data extraction begins.
The data extraction phase amplifies the burden. Every included study must be read in full, and specific data elements — study design, population characteristics, intervention details, outcome measures, statistical results, and risk-of-bias assessments — must be extracted into a standardized form. Errors introduced during extraction propagate directly into the synthesis and can invalidate the review's conclusions. Risk-of-bias assessment requires applying validated tools such as Cochrane's RoB 2, ROBINS-I, or the Newcastle-Ottawa Scale to each study — a further hour per paper for an experienced reviewer. And when the database search must be updated to capture recent publications, the entire process restarts from the screening stage, consuming another cycle of reviewer time.
The synthesis stage, where patterns across studies are identified and conclusions drawn, often receives the least time because the earlier stages consumed the budget. Quantitative meta-analysis requires statistical software expertise and careful heterogeneity assessment; narrative synthesis requires the analytical skill to resolve conflicting findings without cherry-picking. The result is that most systematic reviews are either underpowered — based on an incomplete search — or outdated by the time they are published, given the 18–24 month timeline from protocol registration to final report. Research funders and journals increasingly demand living reviews that update continuously, an expectation that current manual workflows cannot meet.
How COCO Solves It
Multi-Database Parallel Search Execution: COCO executes comprehensive literature searches across all relevant databases simultaneously:
- Queries PubMed, Embase, Cochrane Library, Scopus, Web of Science, PsycINFO, and CINAHL in parallel using a structured Boolean search strategy
- Searches gray literature sources — ClinicalTrials.gov, WHO ICTRP, government report repositories, and preprint servers — for unpublished and non-peer-reviewed evidence
- Translates the search strategy across database-specific syntax and MeSH/Emtree controlled vocabulary systems automatically
- Deduplicates results across databases using DOI, PMID, title-author matching, and fuzzy string comparison to eliminate duplicate records
- Generates a PRISMA search flow diagram with record counts at each stage for immediate inclusion in the review report
Dual-Reviewer-Equivalent Title and Abstract Screening: COCO replicates the accuracy of independent dual screening at scale:
- Applies eligibility criteria to each title and abstract using structured natural language understanding of population, intervention, comparator, outcome, and study design (PICOS) elements
- Flags borderline records for human reviewer attention rather than making binary include/exclude decisions on ambiguous cases
- Calibrates screening sensitivity against a gold-standard set of known-include studies to verify the tool is not missing eligible records
- Generates a screening decision log with the specific criterion cited for each exclusion, enabling audit and inter-rater reliability calculation
- Accelerates full-text screening by pre-extracting key eligibility-relevant passages from each paper for reviewer verification
Structured Data Extraction Template Completion: COCO extracts and populates the review's data extraction form for every included study:
- Reads full-text PDFs and extracts population characteristics, intervention details, comparator descriptions, outcome definitions, and reported statistical results
- Maps extracted variables to the review protocol's pre-specified extraction fields, flagging studies where a variable is not reported or must be calculated from available data
- Identifies tables, figures, and supplementary appendices as extraction sources and integrates them with the main text analysis
- Flags inconsistencies within a single paper — e.g., Methods and Results sections reporting incompatible sample sizes — for reviewer resolution
- Produces a completed extraction table in the reviewer's specified format (Excel, Covidence, RevMan, or custom) ready for synthesis
Risk-of-Bias Assessment Application: COCO applies validated quality assessment tools to each included study:
- Applies Cochrane RoB 2 criteria to randomized controlled trials, ROBINS-I to non-randomized studies, and Newcastle-Ottawa Scale to observational research
- Extracts the specific methodological information needed to judge each domain — randomization method, allocation concealment, blinding status, outcome reporting completeness
- Generates a judgment (low/some concerns/high risk) for each domain with the supporting text passage cited as evidence
- Produces risk-of-bias summary and traffic-light plots in PRISMA-compliant format for immediate insertion into the review manuscript
- Identifies patterns across studies — e.g., systematic reporting bias in a specific journal or research group — that affect the overall evidence quality assessment
Quantitative Meta-Analysis and Heterogeneity Assessment: COCO performs statistical synthesis where appropriate:
- Calculates pooled effect sizes (odds ratios, risk ratios, mean differences, standardized mean differences) using fixed-effects and random-effects models
- Tests for statistical heterogeneity using Cochran's Q test and I² statistic, with visual Forest plots generated automatically
- Performs subgroup analyses and meta-regression on pre-specified moderator variables to explore sources of heterogeneity
- Conducts sensitivity analyses — excluding high-risk-of-bias studies, imputing missing data, varying outcome definitions — to test result robustness
- Assesses publication bias using funnel plots, Egger's test, and trim-and-fill analysis, with results integrated into the evidence quality narrative
Living Review Continuous Update Engine: COCO maintains the review currency over time:
- Reruns the full search strategy on a scheduled basis and compares new records against the existing screening decisions
- Alerts reviewers when a new eligible study is identified that would affect the pooled estimates or change the evidence quality assessment
- Integrates new studies into the meta-analysis with automatic recalculation of pooled effects and heterogeneity statistics
- Generates an update summary report documenting what changed since the last version, which studies were added, and how conclusions were affected
- Supports continuous publication to a living review platform with version-controlled manuscript updates
Results & Who Benefits
Measurable Results
- Screening time: Reduced from 200–400 reviewer hours to under 20 hours of human review for a 3,000-record search — 90–95% time savings
- Search comprehensiveness: Average of 23% more eligible studies identified per review compared to single-database manual searches, due to parallel multi-database coverage
- Data extraction accuracy: Extraction error rate reduced from an average of 8–12% in manual extraction to under 2% with COCO-assisted extraction and human verification
- Time to first draft: Complete review manuscript draft — including PRISMA diagram, evidence tables, risk-of-bias plots, and Forest plots — produced within 5 business days of search completion
- Living review update cycle: Database searches updated and new studies screened within 48 hours of scheduled run, enabling quarterly living review publication vs. 18–24 months for a new manual review
Who Benefits
- Academic Researchers and PhD Candidates: Eliminate the months-long screening and extraction phases that consume the majority of systematic review time — redirect intellectual effort to synthesis, interpretation, and contribution to knowledge
- Health Technology Assessment Teams: Produce evidence summaries for reimbursement decisions and clinical guideline development on compressed regulatory timelines without sacrificing rigor
- Government Policy Research Units: Maintain living evidence bases on active policy questions without requiring permanent dedicated reviewer teams for each topic area
- Research Funders and Systematic Review Centers: Scale systematic review production across multiple topics simultaneously without proportional increases in reviewer headcount
💡 Practical Prompts
Prompt 1: Database Search Strategy Development
Develop a comprehensive systematic literature search strategy for the following review.
Review topic: [describe the research question using PICOS framework]
Population: [who are the subjects of interest]
Intervention or exposure: [what intervention, exposure, or factor is being studied]
Comparator: [what is being compared to, if applicable]
Outcomes: [primary and secondary outcomes of interest]
Study designs: [RCTs only / observational studies / all designs]
Date limits: [e.g., 2000 to present]
Language limits: [English only / all languages]
Databases to search: [list or confirm which databases to include]
Deliverables:
1. Boolean search string for each specified database with database-specific syntax and controlled vocabulary (MeSH, Emtree)
2. Translation notes explaining any syntax differences across databases
3. Estimated record yield per database based on test search
4. Gray literature search strategy: which registries, repositories, and websites to search
5. PRISMA-compliant search reporting text for the Methods section of the review
Confirm the search strategy retrieves all studies in the following known-include set before finalizing: [list 3–5 known relevant studies]Prompt 2: Title and Abstract Screening
Screen the following batch of records against the eligibility criteria for our systematic review.
Eligibility criteria:
- Include: [describe what must be true for a study to be included]
- Exclude: [describe what automatically excludes a study]
- Uncertain — flag for full-text review: [describe borderline situations]
Records to screen: [paste title/abstract list or upload file]
For each record:
1. Decision: Include / Exclude / Uncertain — flag for full-text
2. Reason: cite the specific eligibility criterion applied (one primary reason for exclusions)
3. Confidence: High / Moderate / Low — flag low-confidence decisions for reviewer attention
Output format: Screening decision table with Record ID, Title, Decision, Reason, Confidence columns
Flag any records where the abstract does not provide enough information to make a reliable eligibility decision.Prompt 3: Data Extraction
Extract data from the following included study using our pre-specified extraction form.
Study citation: [author, year, journal, DOI]
Full text: [attach PDF or paste relevant sections]
Extract the following fields:
1. Study design: [RCT / cohort / case-control / cross-sectional / other]
2. Setting and country
3. Population: sample size, age, sex distribution, inclusion/exclusion criteria used in the study
4. Intervention details: [specific variables relevant to your review]
5. Comparator details
6. Follow-up duration
7. Primary outcome: measure, assessment method, results with confidence intervals
8. Secondary outcomes: [list]
9. Subgroup analyses reported
10. Funding source and conflicts of interest
Flag any fields where the information is not reported, is ambiguous, or appears inconsistent with another section of the paper.
Output: Completed extraction form + list of flagged items requiring reviewer clarificationPrompt 4: Risk-of-Bias Assessment
Perform a risk-of-bias assessment for the following study using the specified tool.
Study citation: [author, year, journal]
Study design: [randomized controlled trial / non-randomized study / observational study]
Assessment tool: [Cochrane RoB 2 / ROBINS-I / Newcastle-Ottawa Scale]
Full text: [attach PDF or paste relevant Methods and Results sections]
For each domain of the specified tool:
1. Judgment: Low risk / Some concerns / High risk
2. Supporting evidence: quote the specific text passage from the paper that informs this judgment
3. Rationale: explain how the evidence leads to the judgment
Generate:
- Completed domain-level risk-of-bias form
- Overall risk-of-bias judgment with justification
- Traffic-light plot data for visual summary generation
- Notes on any methodological information that was not reported and would be needed for a definitive judgmentPrompt 5: Evidence Synthesis and Conclusion Drafting
Synthesize the evidence from our completed systematic review for the following outcome.
Outcome: [specify the outcome being synthesized]
Included studies: [number of studies, total participants]
Pooled effect estimate (if meta-analysis was performed): [effect size, CI, I² value]
Risk-of-bias profile: [summarize the overall quality of the evidence body]
Subgroup findings (if any): [describe any important subgroup differences]
Draft:
1. Results narrative: describe what the evidence shows, including effect size, direction, precision, and consistency across studies
2. Evidence quality assessment: apply GRADE criteria to rate the overall certainty of evidence (high/moderate/low/very low) with justification for each rating decision
3. Limitations paragraph: key methodological limitations of the evidence base that affect confidence in the findings
4. Implications for practice: what the evidence means for practitioners or policymakers, calibrated to the certainty level
5. Implications for future research: specific gaps that future studies should address, framed as research questions
Write in formal academic register suitable for journal submission. Flag any interpretive judgments where reviewer input is needed.32. AI Patent Landscape Analysis Engine
Maps the complete patent ecosystem for any technology domain — identifying filing trends, key assignees, whitespace opportunities, and freedom-to-operate risks — in a fraction of the time of manual patent searches.
Pain Point & How COCO Solves It
The Pain: Patent Landscape Analysis Is Slow, Expensive, and Often Incomplete
Understanding the patent landscape for a technology domain is essential for R&D strategy, competitive intelligence, investment due diligence, and freedom-to-operate assessment — yet it remains one of the most specialized and time-consuming research tasks an analyst can undertake. A comprehensive patent landscape begins with constructing a search strategy that captures all relevant patents without being overwhelmed by noise: selecting the right International Patent Classification (IPC) codes, Cooperative Patent Classification (CPC) codes, and keyword combinations across national and international databases. This construction alone can take days, and getting it wrong — either missing a key classification or using an overly broad term — invalidates the entire analysis. Most organizations lack in-house patent analysts and rely on expensive external IP counsel who charge by the hour for work that is inherently database-search-intensive.
The volume problem is severe. A technology domain like CRISPR gene editing, solid-state batteries, or large language model inference has tens of thousands of relevant patents across USPTO, EPO, WIPO, CNIPA, and national IP offices in Japan, South Korea, and other jurisdictions. Reading even the abstracts of 10,000 patents is impractical; reading full claims — where the actual scope of protection is defined — is impossible at that scale without automation. Patent families must be identified to avoid double-counting, since the same invention is typically filed in multiple jurisdictions simultaneously. Claims must be read and categorized by technology sub-domain, application area, and protective scope. Prior art must be assessed for each claim to evaluate patent strength. None of this analysis can be delegated to junior staff without extensive patent training.
The strategic value of landscape analysis — identifying whitespace where no one has claimed protection, spotting competitors who are filing aggressively in a new area before it becomes apparent from product announcements, or detecting licensing leverage — is directly proportional to the comprehensiveness and timeliness of the analysis. Yet the manual process almost always produces a point-in-time snapshot that is already partially outdated by the time it is delivered, because patent offices publish new filings on an 18-month lag from priority date. Analysts who want to track a technology domain continuously must rebuild the landscape repeatedly — a cost that most organizations cannot absorb.
How COCO Solves It
Multi-Jurisdiction Patent Database Search: COCO conducts comprehensive patent searches across all major IP offices simultaneously:
- Queries USPTO, EPO (esp@cenet), WIPO PatentScope, CNIPA, J-PlatPat, KIPRIS, and national IP office databases in parallel
- Constructs a balanced search strategy combining IPC/CPC classification codes, keyword strings in multiple languages, and assignee name variations
- Applies patent family deduplication to group related filings (priority application, national phase entries, continuations) into single invention records
- Filters by filing date, publication date, grant status, legal status (active, expired, lapsed, pending), and jurisdiction
- Generates a search transparency report documenting the full search strategy, record counts, and deduplication methodology for reproducibility
Technology Clustering and Sub-Domain Mapping: COCO organizes the patent landscape by technology structure:
- Reads abstracts, claims, and descriptions to classify each patent into technology sub-domains using machine-assisted categorization
- Builds a hierarchical technology taxonomy for the domain, surfacing sub-areas that are patent-dense and those that are sparse
- Identifies cross-domain patents that claim protection across multiple technology areas — often indicators of platform inventions with broad blocking potential
- Generates a heat map of patent density across the technology taxonomy, showing where IP activity is concentrated and where whitespace exists
- Flags technology sub-domains with recent acceleration in filing rates — a leading indicator of competitive focus shifting before it appears in products or publications
Assignee and Inventor Intelligence: COCO profiles the competitive landscape at the organizational level:
- Identifies all corporate, academic, and individual assignees active in the domain, with filing volume and trend data for each
- Tracks assignee name variations, subsidiary filings, and post-acquisition portfolio transfers to attribute patents accurately to parent organizations
- Profiles key inventors: researchers filing in multiple organizations, solo inventors who may represent licensing opportunities, and inventor networks indicating collaboration patterns
- Compares assignee filing trends over time to identify organizations accelerating, decelerating, or newly entering the technology domain
- Maps geographic concentration of assignees to identify national innovation clusters and potential jurisdiction-specific competitive threats
Claims Analysis and Scope Assessment: COCO reads and categorizes patent claims for strategic value assessment:
- Parses independent and dependent claims for each patent to characterize the scope of protection claimed
- Identifies broad foundational patents with wide independent claims that could block multiple downstream applications
- Flags continuation and continuation-in-part filings that extend protection timelines through claims refinement
- Assesses claim language for key terms that define the scope boundary — terms that could be designed around vs. terms with broad coverage
- Groups patents by claimed application area to identify where specific use cases are densely claimed vs. open for development
Freedom-to-Operate Risk Identification: COCO surfaces patents that may present blocking risks:
- Identifies active patents with unexpired terms and claims covering the specific technology implementation described
- Assesses assignee licensing history and litigation track record to estimate enforcement probability
- Flags patents from non-practicing entities (NPEs) or patent assertion entities (PAEs) in the relevant domain
- Identifies patents with lapsed maintenance fee payments that have entered the public domain despite originally blocking claims
- Generates a risk-tiered patent list — high/medium/low FTO risk — with the specific claims and claim elements driving each rating
Landscape Report and Monitoring: COCO produces structured deliverables and maintains the analysis over time:
- Generates a complete patent landscape report: executive summary, filing trend analysis, assignee leaderboard, technology heat map, whitespace analysis, and FTO risk register
- Produces visualization-ready data for filing trend charts, assignee comparison matrices, and geographic distribution maps
- Sets up automated monitoring for new filings in the technology domain with weekly or monthly digest alerts
- Alerts when a specific competitor files in a new technology sub-domain or when a patent covering a key claim area is granted
- Maintains a living landscape database that integrates new publications on a rolling basis without requiring a full rebuild
Results & Who Benefits
Measurable Results
- Landscape construction time: Reduced from 6–12 weeks of IP analyst time to under 5 business days for a comprehensive multi-jurisdiction landscape — 85–92% time savings
- Patent coverage: Landscapes built with COCO identify an average of 3.1x more relevant patent families compared to single-database searches limited to USPTO and EPO
- Whitespace identification accuracy: Technology sub-domains identified as whitespace by COCO confirmed as unprotected in 91% of cases upon subsequent IP counsel verification
- FTO risk surfacing: High-risk patents flagged by COCO identified in 87% of cases where subsequent full FTO opinions confirmed a relevant blocking claim
- Ongoing monitoring cost: Continuous landscape monitoring delivered at less than 15% of the cost of quarterly manual rebuilds commissioned from external IP counsel
Who Benefits
- Corporate R&D Strategy Teams: Identify technology whitespace for patent filing programs, detect competitor IP activity before it reaches product announcements, and prioritize licensing negotiations with comprehensive landscape intelligence
- IP Counsel and Patent Attorneys: Accelerate freedom-to-operate searches and prior art identification with AI-assisted pre-analysis, focusing legal expertise on claim interpretation rather than database search construction
- Investment Analysts and VC Due Diligence Teams: Assess IP moat strength and freedom-to-operate risk for technology companies before investment, with a structured landscape deliverable rather than anecdotal IP commentary
- Academic Technology Transfer Offices: Evaluate commercial patent potential and competitor landscape for faculty inventions quickly and cost-effectively, enabling more informed licensing and spin-out decisions
💡 Practical Prompts
Prompt 1: Patent Landscape Scoping
Build a comprehensive patent landscape for the following technology domain.
Technology domain: [describe the technology area — be specific about scope]
Core technical concepts: [list 5–10 key technical terms defining the domain]
Known key players (if any): [list companies or research institutions known to be active]
Geographic scope: [global / specific jurisdictions]
Time period: [e.g., 2010 to present / last 5 years]
Purpose: [R&D strategy / FTO assessment / investment due diligence / licensing negotiation]
Deliverables:
1. Search strategy: IPC/CPC codes and keyword strings used, with rationale for each included term
2. Filing volume trend: patents filed per year in this domain from [start year] to present
3. Top 20 assignees by filing volume with trend data (growing / stable / declining)
4. Technology sub-domain breakdown: heat map of filing density across the technology taxonomy
5. Whitespace analysis: technology sub-areas with fewer than [threshold] patents — potential filing opportunities
6. Most-cited patents in the domain — the foundational prior art every practitioner should know
Exclude: design patents, plant patents, purely mechanical implementations unrelated to [specify].Prompt 2: Assignee Competitive Intelligence
Analyze the patent filing activity of the following companies in [technology domain].
Companies to analyze: [list up to 10 competitors or benchmark organizations]
Technology domain: [describe]
Time period for trend analysis: [e.g., 2015 to present]
For each company:
1. Total patent families filed in domain — overall and by sub-domain
2. Filing trend: are they accelerating, maintaining, or decelerating filings in the last 3 years?
3. Technology focus: which sub-domains does this company's portfolio concentrate in?
4. Key inventors: top 3–5 inventors by filing volume — are they still active at this company?
5. Geographic strategy: which jurisdictions does this company prioritize for protection?
6. Licensing activity: any known licensing programs, standard-essential patent declarations, or litigation in this domain?
Output: Competitive IP intelligence matrix comparing all companies across the above dimensions + narrative on the company with the strongest strategic IP position and why.Prompt 3: Freedom-to-Operate Pre-Screen
Perform a preliminary freedom-to-operate screening for the following product implementation.
Product description: [describe the specific technical implementation — what it does and how]
Key technical features: [list the specific technical elements that should be screened]
Launch jurisdictions: [countries where the product will be sold or used]
Launch timeline: [when commercialization is planned]
Screen for:
1. Active patents (granted, not expired, not lapsed) with claims that read on the described implementation
2. Pending applications in the same technology area that could mature into blocking patents before launch
3. High-risk assignees: known NPEs, patent assertion entities, or aggressive litigants in this domain
4. Design-around options: for any high-risk patent, describe the claim elements that define the blocking scope and potential technical workarounds
Output: Risk-tiered patent list (High / Medium / Low) with claim language quoted for each high-risk patent + recommendation for which patents warrant full IP counsel FTO opinion
Note: This is a preliminary screening, not a legal FTO opinion — flag for referral to patent counsel.Prompt 4: Whitespace and Filing Opportunity Analysis
Identify patent filing opportunities in the following technology domain for our R&D pipeline.
Technology domain: [describe]
Our R&D focus areas: [list the technical problems or application areas our R&D is targeting]
Our known innovations (do not disclose to third parties): [describe at a general level — or keep confidential]
Analyze:
1. Technology sub-areas in [domain] with fewer than 10 active patent families — potential whitespace
2. Application areas where patent density is low relative to commercial activity — indicating under-protection of valuable implementations
3. Geographic gaps: technology areas where US/EU protection is dense but Asian jurisdiction filing is sparse — or vice versa
4. Expiring foundational patents: key patents expiring in the next 3–5 years that will open design space currently blocked
5. Continuation opportunities: published applications by competitors where claims have not yet been fully prosecuted — monitoring-worthy
Output: Whitespace opportunity matrix ranked by commercial relevance × filing feasibility + recommended patent strategy for each opportunity areaPrompt 5: Patent Portfolio Valuation Framework
Assess the relative strategic value of the following patent portfolio.
Portfolio owner: [company or entity name]
Portfolio scope: [describe or attach patent list with patent numbers]
Technology domain: [describe]
Valuation purpose: [acquisition / licensing negotiation / litigation strategy / investor reporting]
Assess each patent (or patent family) on:
1. Claim breadth: broad pioneering claims vs. narrow implementation-specific claims
2. Claim validity risk: strength of prior art that could challenge the patent
3. Remaining term: years until expiration (considering any terminal disclaimers)
4. Jurisdiction coverage: which markets are protected
5. Commercial relevance: how central is the claimed technology to current commercial products?
6. Licensing potential: evidence of prior licensing, SEP declarations, or litigation — indicates established licensing value
Output: Portfolio heatmap scoring each family on the above dimensions + top 10 highest-value patents with justification + portfolio-level strategic assessment (blocking value / licensing revenue potential / defensive value)33. AI Competitive Intelligence Synthesis Engine
Aggregates signals from earnings calls, product releases, job postings, patent filings, pricing changes, and regulatory submissions to produce a unified competitor intelligence picture — updated continuously, not quarterly.
Pain Point & How COCO Solves It
The Pain: Competitive Intelligence Is Fragmented, Reactive, and Always Behind
Research analysts responsible for competitive intelligence face a structurally impossible task: monitoring dozens of competitors across hundreds of information channels simultaneously, while still producing the synthesis work that makes the intelligence actionable. Competitor intelligence is scattered across earnings call transcripts, press releases, product documentation, regulatory filings, job boards, patent databases, LinkedIn announcements, trade conference presentations, customer review platforms, and social media — each requiring a different monitoring approach and generating a different signal type. Building a comprehensive view of even a single competitor requires an analyst to pull from at least ten different source categories, none of which automatically integrate with each other. Doing this for a competitive set of eight to twelve companies, updated weekly, is beyond what any individual analyst can sustain alongside their other research responsibilities.
The result is a characteristic failure mode: competitive intelligence reports that are comprehensive when first built but stale within weeks, covering only the competitors that made public announcements recently and missing the quieter moves — gradual price adjustments, steady talent hiring in a new capability area, incremental patent filing campaigns — that often matter more strategically than headline announcements. Analysts default to monitoring the sources that are easiest to track, which typically means formal press releases and quarterly earnings calls, while missing the higher-signal indicators that require effort to find. When a competitor launches a significant product feature or enters a new market segment, the intelligence team often learns about it from the sales force — who encountered it in a competitive deal — rather than from proactive monitoring.
The synthesis problem is as significant as the monitoring problem. Raw competitive signals — an earnings call comment about international expansion, a job posting for a specific engineering role, a patent application in an adjacent technology area — only become actionable intelligence when interpreted together, against a baseline of prior knowledge about the competitor's strategy and capabilities. Maintaining that interpretive baseline requires institutional memory that lives in individual analysts' heads, not in structured systems. When analysts turn over, the competitive intelligence function degrades. When multiple analysts split coverage of different competitors, synthesis across the competitive set becomes inconsistent. And when leadership asks "what does the competitive landscape look like right now?", the honest answer is usually that no one knows comprehensively — they know what happened publicly in the last quarter.
How COCO Solves It
Multi-Channel Signal Collection and Normalization: COCO monitors the full information landscape for every tracked competitor:
- Tracks earnings call transcripts, investor day presentations, SEC/regulatory filings, and annual reports from all monitored companies
- Monitors product announcement pages, changelogs, developer documentation, API updates, and pricing pages for product intelligence
- Scans job posting databases (LinkedIn, Indeed, Glassdoor, Greenhouse, Lever) for hiring signals — headcount by function, role requirements indicating new capability development, location patterns
- Watches patent filings and publication schedules at USPTO, EPO, and CNIPA for R&D direction signals 18 months ahead of product announcements
- Aggregates customer review signals from G2, Gartner Peer Insights, Capterra, and Trustpilot to track satisfaction trends and emerging product weaknesses
Signal Classification and Strategic Interpretation: COCO translates raw signals into structured intelligence:
- Classifies each signal by type — product, pricing, talent, partnership, geographic expansion, M&A, regulatory, technology — for systematic tracking
- Interprets signals against each competitor's known strategy, positioning, and historical patterns to distinguish noise from meaningful moves
- Identifies signal clusters — multiple concurrent indicators pointing in the same strategic direction — that warrant elevated attention
- Flags signals that contradict the competitor's stated strategy or prior behavior, which often indicate strategic pivots before they are publicly announced
- Assigns significance ratings (high/medium/low) to each signal based on its potential impact on the competitive dynamic in your market
Cross-Competitor Pattern Recognition: COCO identifies patterns across the competitive set, not just within individual competitors:
- Detects when multiple competitors are making similar moves simultaneously — indicating a market-wide response to a shared threat or opportunity
- Identifies competitive sequencing patterns: which competitor typically leads and which follow, and on what timeline
- Spots differentiation erosion: cases where competitor product features, pricing, or messaging are converging on your offering, reducing your advantage
- Surfaces asymmetric competitive dynamics — competitors who are strong in specific segments or geographies but weak in others — for targeting recommendations
- Tracks competitive response patterns to your own moves: price changes, product launches, or partnership announcements and how competitors historically react
Competitor Profile Maintenance and Deep Dives: COCO builds and maintains structured competitor profiles:
- Maintains a structured profile for each competitor covering strategy, product portfolio, pricing structure, target segments, key customers, technology stack, and go-to-market motion
- Updates each profile element when a relevant signal is detected, with the source and date of each update recorded
- Generates a confidence score for each profile element — high confidence from multiple consistent sources, low confidence from single or potentially unreliable signals
- Produces deep-dive competitor analysis on demand: a comprehensive profile refresh integrating all recent signals against the competitor's historical baseline
- Identifies information gaps in each competitor profile — areas where signal coverage is sparse — and recommends targeted research to fill them
Win/Loss Intelligence Integration: COCO integrates field competitive intelligence with market signal monitoring:
- Analyzes win/loss interview transcripts and CRM deal notes for competitive patterns that don't appear in public signals
- Correlates field competitive observations with external signals: does the sales team's experience with a competitor's pricing changes match what public signals suggest?
- Identifies competitive intelligence gaps that field teams consistently encounter — questions about competitor capabilities that public signals can't answer, indicating primary research priorities
- Tracks which competitor features, pricing moves, or partnership announcements correlate with deal losses, quantifying their competitive impact
- Generates a competitive battlecard for each competitor that integrates monitored signals with field intelligence for sales enablement
Executive Competitive Intelligence Briefing Generation: COCO produces structured deliverables at cadenced intervals:
- Generates a weekly competitive intelligence digest: significant signals from the past week, interpretation, and recommended responses, formatted for leadership review
- Produces a monthly competitive landscape report: all competitor profile updates, trend analysis across the competitive set, and strategic implications
- Creates quarterly competitive deep-dive reports for one or two competitors, integrating all signals from the quarter into a comprehensive strategic assessment
- Supports ad-hoc competitive intelligence requests: "What do we know about Competitor X's plans in the enterprise segment?" answered in minutes from the maintained intelligence base
- Tracks intelligence-to-decision linkages: which competitive insights led to strategic decisions, for demonstrating CI function ROI
Results & Who Benefits
Measurable Results
- Competitor monitoring coverage: Increased from an average of 3–5 actively monitored competitors with manual processes to 15–25 competitors monitored continuously with the same analyst headcount
- Signal-to-insight lag: Average time from a competitive signal occurring to analyst synthesis reduced from 3–6 weeks to under 48 hours for high-significance signals
- Signal channel coverage: Firms using COCO monitor an average of 14 concurrent signal channels per competitor vs. 3–4 with manual monitoring, surfacing 4.2x more relevant signals per competitor per month
- Competitive surprise rate: Significant competitive moves that caught leadership completely off-guard reduced from 62% of material events to under 18% in organizations with comprehensive CI programs
- Sales win rate vs. key competitors: Organizations integrating CI with sales enablement through COCO-maintained battlecards report 12–18% improvement in win rates against their top two competitors within 12 months
Who Benefits
- Research Analysts and CI Teams: Transform from reactive report writers responding to ad-hoc requests into proactive intelligence professionals maintaining a living competitive picture with strategic foresight value
- Product Management: Access continuous competitive product intelligence — feature updates, pricing changes, customer complaints — to inform roadmap decisions with current competitive context rather than quarterly snapshots
- Sales and Business Development: Receive continuously updated battlecards and deal-specific competitive intelligence drawn from a maintained knowledge base, reducing loss rates from competitive blind spots
- Executive Leadership: Receive structured weekly and monthly competitive briefings that enable proactive strategic decisions rather than reactive responses to competitive moves already in market
💡 Practical Prompts
Prompt 1: Competitor Signal Sweep
Perform a comprehensive competitive signal sweep for the following competitor.
Competitor: [company name]
Our industry/market context: [describe the market and competitive relationship]
Time period: [e.g., last 30 days / last quarter]
Collect and synthesize signals from:
1. Public statements: earnings calls, investor presentations, press releases, executive interviews
2. Product signals: new feature announcements, pricing changes, product page updates, API/changelog updates
3. Talent signals: significant new hires, departures of key executives, job posting patterns indicating new capability investment
4. Partnership signals: announced integrations, reseller agreements, strategic alliances, customer wins
5. Technology signals: patent filings, developer conference presentations, open-source contributions
6. Customer sentiment: recent review trends on G2, Gartner Peer Insights, or equivalent platforms
For each signal: source, date, signal type, what it means strategically, significance rating (High / Medium / Low)
Output: Ranked signal list by significance + 3-paragraph strategic interpretation summaryPrompt 2: Competitive Positioning Analysis
Analyze the competitive positioning of the following companies in [market segment].
Companies to analyze: [list your company and key competitors]
Market segment: [describe the specific segment or use case being analyzed]
Customer decision criteria: [what factors do customers in this segment prioritize when choosing a vendor]
For each company:
1. Stated positioning: how do they describe themselves and their differentiation?
2. Actual differentiation: based on product capabilities, pricing, and customer feedback, what do they actually do better or worse?
3. Target customer: who are they winning with — company size, industry, use case?
4. Pricing strategy: what is the pricing model and relative price positioning?
5. Key weaknesses: where do customers complain, where do they lose deals?
Output: Competitive positioning matrix + narrative on the most dangerous competitor in [segment] and why + white space where no competitor currently dominatesPrompt 3: Job Posting Intelligence Analysis
Analyze the following competitor's recent job postings for strategic intelligence.
Competitor: [company name]
Job postings: [attach list or describe how to access — e.g., link to careers page]
Analysis period: [e.g., postings from the last 90 days]
Extract intelligence on:
1. Headcount growth direction: which functions are growing (engineering, sales, marketing, operations)?
2. Geographic expansion signals: are they hiring in new cities, regions, or countries?
3. New capability investment: what specific technical skills or domain expertise are they building that they didn't previously advertise?
4. Seniority signals: are they hiring leaders (VP, Director) in functions that suggest new strategic priorities?
5. Product direction clues: what product areas or technologies appear in engineering job descriptions?
6. Go-to-market changes: sales role descriptions that indicate new segments, channels, or pricing model changes
Output: Intelligence summary with each finding, confidence level, and strategic implication for our positioning + 90-day watch list of developments to monitorPrompt 4: Earnings Call Competitive Analysis
Analyze the following earnings call transcript for competitive intelligence.
Company: [competitor name]
Transcript: [attach or paste earnings call transcript]
Quarter: [Q? FY?]
Extract:
1. Explicit competitive mentions: any named competitors referenced — what was said and in what context?
2. Implicit competitive signals: product wins, new customer segments, geographic expansions that affect our competitive position
3. Guidance and outlook: what are they expecting for growth — does this match or contradict what we're seeing in our own pipeline?
4. Key metrics called out: what KPIs did management emphasize — what does this tell us about where they are strong or struggling?
5. Analyst questions: what did analysts probe on — often reveals concerns not explicitly addressed by management
6. Strategic language changes: compare with prior quarter transcript — what new themes appeared, what old themes disappeared?
Output: CI intelligence brief from this earnings call + updated competitor profile elements with date-stamped changesPrompt 5: Competitive Battlecard Update
Update the competitive battlecard for [competitor name] based on the latest intelligence.
Current battlecard: [attach existing battlecard or describe current content]
New intelligence to integrate: [describe recent signals — new features, pricing changes, customer wins/losses, executive hires]
Sales context: [what deals are currently in play where this competitor appears]
Update the battlecard with:
1. Why customers choose us over [competitor]: updated with any new differentiators or strengthened existing ones
2. Why customers choose [competitor] over us: updated with any new competitor strengths or improved weaknesses
3. Landmine questions: questions that expose [competitor's] weaknesses when asked in a sales conversation
4. Objection responses: when a prospect says "[competitor] offers X", how do we respond — updated with current factual counterpoints
5. Deal-specific intelligence: what do we know about [competitor's] current sales motion, discount patterns, and pressure points
Format as a one-page sales enablement card suitable for AE use in active deals.34. AI Regulatory Impact Assessment Tool
Analyzes proposed or enacted regulatory changes and quantifies their operational, financial, and strategic implications for a business or investment portfolio — producing a structured impact assessment in hours, not weeks.
Pain Point & How COCO Solves It
The Pain: Regulatory Change Is Constant, Multi-Jurisdictional, and Analytically Demanding
Research analysts supporting strategy, compliance, or investment functions must continuously track regulatory developments across multiple jurisdictions and translate them into business-relevant impact assessments — a task that is simultaneously urgent and time-intensive. A single regulatory change can require reading hundreds of pages of proposed rulemaking, public comments, agency guidance documents, and legislative history before an analyst can even begin assessing impact. In global industries — financial services, pharmaceuticals, technology, energy, food and agriculture — relevant regulations span dozens of jurisdictions simultaneously, each with different timelines, implementation periods, and exemption structures. Tracking the regulatory calendar and maintaining awareness of developments across all of these is a full-time monitoring function that most analyst teams simply cannot staff.
The impact assessment itself is analytically demanding in ways that go beyond reading comprehension. Understanding the financial impact of a new capital requirement, an emissions standard, or a data localization mandate requires integrating regulatory text with operational data, financial models, and sector benchmarks — a cross-disciplinary synthesis that takes experienced analysts days and often requires input from multiple subject matter experts. When regulators release proposals for public comment, the window to analyze the proposal, model its impact, and formulate a strategic response is typically 30–90 days — a timeline that industry participants frequently miss because their analysis capacity is stretched across multiple concurrent regulatory developments. The result is that organizations either respond reactively after implementation — incurring compliance costs they could have mitigated — or devote disproportionate resources to regulation tracking at the expense of other strategic research.
The compounding challenge is regulatory interdependence. A new data privacy regulation may affect both technology procurement and cross-border data transfer compliance. A new ESG disclosure requirement may interact with existing financial reporting standards and affect capital allocation decisions. A sector-specific rule may align with, conflict with, or create arbitrage opportunities relative to rules in adjacent sectors. Analysts who assess each regulation in isolation miss these interdependencies, producing impact assessments that are technically accurate within their narrow scope but strategically incomplete. Senior leaders need cross-regulatory synthesis to make good decisions, but building that synthesis manually requires coordination across multiple analyst silos — a coordination cost that most organizations absorb inefficiently.
How COCO Solves It
Regulatory Calendar Monitoring and Alert System: COCO maintains continuous awareness of the regulatory environment:
- Tracks regulatory agendas, rulemaking dockets, and legislative calendars across federal agencies (SEC, EPA, FDA, CFPB, FTC, CFTC), EU regulatory bodies, and equivalent national authorities
- Monitors Official Journals, Federal Register, regulatory agency websites, and parliamentary publications for new proposals, final rules, and guidance documents
- Classifies each development by industry relevance, jurisdiction, implementation timeline, and regulatory type (mandatory rule / guidance / proposed rulemaking / enforcement action)
- Generates customized regulatory alert digests filtered by industry, geography, and business function — delivering only the developments relevant to the analyst's coverage area
- Maintains a forward-looking regulatory pipeline showing expected proposals, consultation deadlines, and implementation dates for the next 12–24 months
Regulatory Text Analysis and Plain-Language Interpretation: COCO translates regulatory documents into actionable analysis:
- Reads full regulatory text — including preambles, definitions, operative provisions, exemptions, and transition rules — and produces a structured plain-language summary
- Identifies the specific obligations, prohibitions, and definitions that affect the covered business, distinguishing mandatory requirements from guidance and best practices
- Extracts key dates: effective dates, compliance deadlines, phase-in periods, reporting deadlines, and public comment windows
- Highlights areas of regulatory ambiguity where the text does not clearly resolve how it applies to specific business activities — flagging for legal counsel clarification
- Compares the final rule against the proposed rule to identify changes made in response to public comment, and against prior regulation to identify what has changed from the previous framework
Operational Impact Modeling: COCO quantifies the business impact of regulatory requirements:
- Maps each regulatory requirement to the specific business processes, systems, data assets, or products it affects
- Estimates the operational cost of compliance: process redesign, system implementation, personnel training, ongoing reporting, and third-party audit costs
- Benchmarks compliance cost estimates against public disclosures from peers and regulatory impact analyses published by the issuing agency
- Identifies operational changes required before the compliance deadline and builds a reverse timeline from the effective date to current date
- Models alternative compliance pathways where the regulation provides options, comparing cost and risk profiles of each approach
Financial and Investment Impact Analysis: COCO assesses the financial implications of regulatory changes:
- Quantifies direct financial impacts: cost of compliance, revenue changes from restricted product offerings, capital requirement changes, and potential fine exposure for non-compliance
- Models indirect financial effects: customer behavior changes in response to the regulation, competitive dynamics shifts as regulations favor certain business models, and supply chain cost pass-through
- Assesses portfolio-level exposure for investment analysts: which holdings are positively or negatively positioned relative to regulatory changes, and what the earnings impact range is
- Evaluates regulatory arbitrage opportunities: cases where regulatory differences across jurisdictions create competitive advantage for entities structured to optimize across them
- Stress-tests financial projections for investments or business units against regulatory scenarios: what happens if the strict interpretation prevails vs. the industry-favorable interpretation?
Cross-Regulatory Interdependency Mapping: COCO synthesizes impact across the regulatory landscape:
- Identifies interactions between the analyzed regulation and existing regulations that affect the same business activities, processes, or assets
- Flags conflicts between regulations — cases where compliance with one requirement makes compliance with another more difficult or costly
- Detects complementary requirements — regulations that share similar data collection or reporting obligations, enabling shared compliance infrastructure
- Maps regulatory developments across jurisdictions to identify harmonization trends and divergence risks for international operations
- Synthesizes a unified view of the total regulatory burden facing a business function or portfolio, accounting for all concurrent requirements
Regulatory Response Strategy and Public Comment Support: COCO supports strategic engagement with the regulatory process:
- Analyzes public comments filed by industry peers and trade associations on proposed regulations to identify the industry consensus position and dissenting views
- Drafts structured public comment submissions: identifying the most impactful provisions to comment on, building the evidentiary case for proposed modifications, and forecasting regulatory receptiveness to each argument
- Generates a regulatory engagement strategy: which proposals warrant direct engagement, which are better addressed through trade association coordination, and which are likely final regardless of industry input
- Tracks regulatory outcomes against industry comment positions to calibrate regulatory influence effectiveness over time
- Produces regulatory scenario analyses for planning purposes: impact assessment under the proposed text vs. modified text incorporating likely accommodations based on comment patterns
Results & Who Benefits
Measurable Results
- Regulatory monitoring coverage: Analysts using COCO maintain awareness of 3–5x more regulatory developments relevant to their coverage universe without increasing monitoring time
- Impact assessment turnaround: Structured regulatory impact assessment produced within 2–3 business days of a new rule publication vs. 2–4 weeks with manual analysis
- Public comment response rate: Organizations using COCO assistance submit comments on 68% more regulatory proposals that materially affect their business, vs. 31% without AI assistance
- Compliance cost accuracy: Regulatory impact cost estimates produced with COCO modeling align within ±15% of actual implementation costs vs. ±40–60% for initial manual estimates
- Regulatory surprise incidents: Material compliance failures attributable to undetected regulatory changes reduced by 74% in organizations with comprehensive COCO-assisted monitoring programs
Who Benefits
- Research Analysts and Policy Teams: Transform regulatory tracking from an overwhelming reactive function into a managed, comprehensive monitoring program — with structured impact assessments produced on compressed timelines
- Compliance and Legal Departments: Receive pre-analyzed regulatory summaries and preliminary impact assessments that focus legal review on interpretation questions rather than initial document comprehension
- Investment Managers and Analysts: Assess regulatory risk and opportunity exposure across portfolios with structured, comparable impact assessments rather than case-by-case ad-hoc analysis
- C-Suite and Board: Access consolidated regulatory risk dashboards showing the full regulatory burden facing the organization, prioritized by financial impact and implementation urgency
💡 Practical Prompts
Prompt 1: New Regulation Impact Assessment
Perform a comprehensive impact assessment for the following regulatory development.
Regulation: [name and reference number if known]
Issuing authority: [regulatory agency / legislative body]
Publication date: [date]
Effective / compliance date: [date]
Full text: [attach PDF or paste key provisions]
Our organization context:
- Industry: [describe]
- Relevant business activities: [describe the activities that may be affected]
- Jurisdictions of operation: [list]
Assess:
1. Applicability: does this regulation apply to our organization and activities — and under which specific provisions?
2. Key obligations: what must we do, stop doing, or change — with specific provision references for each
3. Key exemptions: are any of our activities exempted — and what conditions must be met to qualify?
4. Compliance timeline: what must be done by when — build a reverse timeline from the effective date
5. Estimated compliance cost: operational, system, personnel, and ongoing reporting costs
6. Strategic implications: competitive dynamics changes, new market opportunities, or business model impacts
Output: Structured impact assessment with executive summary + operational compliance checklist + financial impact range estimatePrompt 2: Regulatory Horizon Scan
Perform a regulatory horizon scan for [industry] covering the next 12–24 months.
Industry: [describe]
Jurisdictions to cover: [list — e.g., US federal, EU, UK, Singapore, China]
Business functions most exposed to regulatory change: [e.g., data handling, product approval, capital management, environmental compliance]
Identify:
1. Regulations currently in proposed/consultation stage that are likely to be finalized in the next 12 months — with current status and expected timeline
2. Regulatory reviews underway that may result in significant changes to existing frameworks
3. Enforcement priority signals: areas where regulators have signaled increased enforcement attention through speeches, guidance, or recent enforcement actions
4. International regulatory harmonization or divergence trends: are major jurisdictions moving in the same or different directions on key topics?
5. Legislative proposals with meaningful probability of enactment that would significantly affect [industry]
Output: Regulatory horizon calendar (12-month + 24-month view) with each development rated by probability, impact magnitude, and implementation urgency + recommended preparation steps for high-probability, high-impact itemsPrompt 3: Public Comment Drafting
Draft a public comment submission for the following proposed regulation.
Proposed rule: [name and docket number]
Issuing agency: [name]
Comment deadline: [date]
Proposed rule summary: [attach or describe the key provisions]
Our position:
- Provisions we support: [list with rationale]
- Provisions we oppose or seek modification of: [list with rationale]
- Alternative approaches we prefer: [describe if applicable]
Draft a public comment that:
1. Opens with a clear statement of our identity and our support/opposition to the proposal overall
2. Addresses each provision of concern with: (a) description of the specific provision, (b) our concern and its basis, (c) proposed alternative language or approach, (d) evidence supporting our proposed modification
3. Cites supporting data, peer research, or prior regulatory precedent for key arguments
4. Maintains a constructive, collaborative tone consistent with productive regulator-industry dialogue
5. Closes with a summary of our specific requested modifications
Length target: [e.g., 5–10 pages]. Formal regulatory comment style.Prompt 4: Cross-Jurisdiction Regulatory Comparison
Compare the following regulatory requirement across the specified jurisdictions.
Regulatory topic: [e.g., data breach notification / capital adequacy / emissions disclosure / product liability]
Jurisdictions to compare: [list — e.g., US (federal), EU, UK, Canada, Australia, Singapore]
For each jurisdiction:
1. Applicable law or regulation: name, reference, effective date
2. Scope: who is covered, what activities are regulated
3. Key requirements: what specifically must be done, by whom, on what timeline
4. Penalties: maximum fines and enforcement mechanisms
5. Recent enforcement trends: notable cases or regulatory guidance in the last 2 years
6. Expected changes: any pending proposals or announced reviews
Synthesis:
- Strictest jurisdiction on each key dimension
- Areas of harmonization: where requirements are largely consistent across jurisdictions
- Areas of conflict: where compliance with one jurisdiction's requirements creates tension with another's
- Recommended compliance approach for an organization operating across all listed jurisdictions
Output: Comparison matrix + strategic compliance recommendations for multi-jurisdictional operationsPrompt 5: Regulatory Scenario Analysis
Model the business impact of the following regulatory scenarios for our organization.
Regulatory topic: [describe the regulation under analysis]
Organization context: [describe the business, its relevant activities and scale]
Scenarios to model:
- Scenario A (Base case): [describe the most likely regulatory outcome — e.g., the proposed rule as written]
- Scenario B (Favorable): [describe an industry-accommodating outcome — e.g., delayed effective date, key exemption granted]
- Scenario C (Adverse): [describe a stricter-than-expected outcome — e.g., expanded scope, accelerated timeline, stronger penalty framework]
For each scenario:
1. Key differences from Scenario A
2. Probability assessment: how likely is this outcome based on regulatory signals and comment patterns?
3. Operational impact: what additional or reduced changes are required vs. base case?
4. Financial impact: estimated cost / revenue / capital impact delta vs. base case
5. Strategic implications: does this scenario open or close strategic options?
Output: Scenario comparison table + recommended preparation strategy that builds optionality across all three scenarios35. AI Research Dataset Quality Auditing Engine
Audits research datasets for completeness, consistency, validity, and bias — producing a structured data quality report with remediation recommendations before analysis begins.
Pain Point & How COCO Solves It
The Pain: Garbage-In-Garbage-Out Corrupts Research Findings at the Source
Research analysts working with primary or secondary data face a fundamental risk that is rarely given adequate systematic attention: the data they are analyzing is wrong before they start. Survey datasets contain duplicate responses, satisficing patterns, and bot-generated records. Administrative datasets drawn from CRM systems, government registries, or partner feeds have inconsistent coding conventions, missing values systematically distributed by geography or time period, and unit measurement changes buried in methodology notes from three database updates ago. Panel data accumulated over years has attrition patterns correlated with the outcome variables being studied. Market data from commercial providers has survivorship bias, backcasting revisions, and coverage gaps that affect cross-sectional comparisons. Most research reports treat data quality as a one-paragraph methods section limitation rather than a quantified and remediated pre-analysis step.
The consequences of undetected data quality problems cascade through the entire research output. Statistical analyses performed on data with systematic missing values produce biased estimates. Segmentation analyses run on datasets with duplicate records produce inflated segment sizes. Trend analyses based on data with undocumented methodology changes confuse structural breaks for real-world phenomena. When findings are later challenged — in peer review, in client questions, in regulatory scrutiny — the inability to demonstrate systematic data quality assessment undermines confidence in the entire analysis. Experienced researchers develop intuitions about where data quality problems commonly lurk, but those intuitions are not systematically applied, not documented, and not transferable to junior analysts who may not know what to look for.
The scale of modern research data compounds the problem. A survey with 5,000 responses, a transaction database with 2 million records, or a panel dataset spanning 15 years across 50 countries cannot be visually inspected for quality problems — the problems that matter most are often invisible without systematic statistical testing. Standard summary statistics — mean, standard deviation, sample size — tell you nothing about whether the data has heaped responses at round numbers, whether the same respondent submitted twenty records under different email addresses, whether a specific geographic cluster has implausibly identical values suggesting imputation without disclosure, or whether the distribution of missing values is random or correlated with key analytical variables. Catching these problems requires a systematic quality audit protocol that most analysts don't have time to run and most organizations don't have standardized procedures for executing.
How COCO Solves It
Structural Completeness Assessment: COCO evaluates data coverage and structural integrity:
- Calculates missing value rates for every variable in the dataset, with breakdowns by time period, geography, respondent segment, and other key dimensions
- Tests whether missingness is missing completely at random (MCAR), missing at random (MAR), or missing not at random (MNAR) — distinguishing ignorable from non-ignorable patterns
- Identifies variables with systematically missing values that correlate with the outcome variable — the most dangerous form of data quality problem for causal claims
- Detects truncated distributions suggesting a data extraction cutoff problem — cases where values below a threshold are missing due to collection limitations rather than true absence
- Produces a completeness scorecard with traffic-light ratings for each variable and a prioritized list of remediation options: imputation, reweighting, or exclusion with documented justification
Duplicate and Consistency Detection: COCO identifies redundancy and internal contradictions:
- Detects exact duplicates and near-duplicates — records identical except for minor variations in timestamp, identifier, or non-critical fields — with configurable similarity thresholds
- Identifies within-respondent inconsistencies in survey data: contradictory answers to logically related questions, impossible combinations of responses, and response pattern reversals across waves
- Checks referential integrity in relational datasets: foreign key mismatches, orphaned records, and ID values referenced but not defined
- Detects temporal inconsistencies: dates in impossible sequences, event timestamps preceding the events they record, and timestamps outside the stated data collection period
- Flags records where the same entity appears under multiple identifiers — common in administrative data where entity matching has not been applied — producing estimated over-count factors
Distribution and Outlier Analysis: COCO tests data distributions for credibility:
- Applies Benford's Law analysis to financial and count variables — deviations from expected first-digit distributions indicate potential data fabrication or transcription errors
- Tests for heaping: artificial concentration of responses at round numbers, multiples of 5, or psychologically salient values that indicate estimation rather than measurement
- Identifies multivariate outliers using Mahalanobis distance and isolation forest methods — records that are not outliers on any single variable but are implausible in combination
- Tests distributional assumptions required by the planned analysis — normality, homoscedasticity, independence — and reports violations with quantified severity
- Compares distributions against known reference distributions from validated sources to flag values implausible on domain knowledge grounds
Systematic Bias Detection: COCO identifies sampling and measurement biases that threaten external validity:
- Tests survey respondent demographics against the target population frame to quantify representation gaps by age, gender, geography, education, and other stratification variables
- Detects social desirability bias patterns: systematic response differences on sensitive questions that suggest respondents are answering what they believe is expected rather than their true view
- Identifies acquiescence bias (systematic "yes" tendency), extreme response style (overuse of scale endpoints), and central tendency bias (overuse of midpoints) in Likert-scale data
- Detects panel attrition patterns in longitudinal data: whether respondents who drop out differ systematically from those who remain — a key threat to longitudinal validity
- Flags geographic or temporal clusters of implausible uniformity — regions or time windows where values are suspiciously similar, suggesting data imputation or collection method changes without documentation
Variable Coding and Unit Consistency Validation: COCO enforces coding coherence across the dataset:
- Identifies inconsistent coding of categorical variables: the same category represented by multiple codes across different records or time periods due to coding convention changes
- Detects unit of measurement inconsistencies: currency figures mixing pre- and post-denomination-change values, distance data mixing metric and imperial, percentage figures mixing decimal and percentage-point scales
- Validates date and time formatting consistency across all temporal variables
- Flags free-text fields with systematic entry patterns suggesting automated or template-based responses rather than genuine unique input
- Cross-validates derived variables against their source variables: calculated fields that don't match manual recalculation from their components indicate transformation errors
Remediation Recommendation and Data Quality Report: COCO produces a structured quality improvement plan:
- Generates a complete data quality report: summary scorecard, issue registry, severity ratings, and estimated impact on planned analyses for each identified problem
- Provides specific remediation recommendations for each issue class: imputation methods appropriate for the missingness mechanism, deduplication rules for specific duplicate patterns, reweighting targets for sampling gaps
- Produces a data quality attestation statement documenting all checks performed, issues found, remediation applied, and residual limitations — suitable for inclusion in research methods sections
- Generates cleaned dataset variants applying each recommended remediation, preserving the original for comparison and auditing the transformation applied to each record
- Tracks data quality metrics over time when the same dataset is updated periodically, flagging deterioration in quality dimensions that indicate upstream collection or processing problems
Results & Who Benefits
Measurable Results
- Data quality issues identified: COCO systematic audits surface an average of 3.4x more data quality issues than analyst visual inspection and ad-hoc checks on the same datasets
- Pre-analysis audit time: Comprehensive data quality audit completed in under 4 hours for datasets up to 500,000 records vs. 2–5 analyst days for manual quality checking
- Finding replication rate: Research findings based on COCO-audited and remediated datasets replicate successfully at a rate 27% higher than findings from non-audited datasets in the same organization
- Methods challenge rate: Research reports with COCO-generated data quality attestation sections experience 71% fewer data methodology challenges from clients, reviewers, or regulators
- Duplicate record reduction: Deduplication protocols identify and remove an average of 4–12% of records as duplicates in survey and administrative datasets — significantly affecting segmentation and frequency analyses
Who Benefits
- Research Analysts and Data Scientists: Gain systematic protection against the most common source of invalidated research findings — poor data quality — with documented evidence of due diligence that satisfies peer review and client scrutiny
- Survey Research Teams: Detect and remove bot responses, duplicate submissions, and low-quality speedster responses before analysis, improving the accuracy of survey-based market research and public opinion measurement
- Financial Research Teams: Validate the completeness and consistency of market data, financial statement data, and alternative data inputs before incorporating them into quantitative models where data errors compound
- Regulatory and Audit Functions: Produce data quality attestations that meet evidentiary standards for regulatory submissions, compliance certifications, and external audit review
💡 Practical Prompts
Prompt 1: Comprehensive Dataset Quality Audit
Perform a comprehensive data quality audit on the following dataset before we begin analysis.
Dataset description: [describe what the data is — survey, transaction records, panel data, etc.]
Number of records: [approximate count]
Number of variables: [count]
Data collection period: [date range]
Planned analysis: [describe what we intend to do with this data — regression, segmentation, trend analysis, etc.]
Dataset: [attach file or describe how to access]
Audit for:
1. Completeness: missing value rates by variable and by key dimensions (time, geography, segment)
2. Duplicates: exact and near-duplicate records
3. Outliers: univariate and multivariate outliers
4. Distribution issues: heaping, truncation, impossible values, Benford's Law violations
5. Coding consistency: variable coding changes over time, unit inconsistencies
6. Bias indicators: systematic patterns in missingness or response that threaten validity
Output: Data quality report with: (a) summary scorecard, (b) issue registry with severity ratings, (c) estimated impact on planned analysis for each issue, (d) prioritized remediation recommendationsPrompt 2: Survey Response Quality Screening
Screen the following survey dataset for low-quality responses before analysis.
Survey context: [describe the survey topic, length, respondent recruitment method]
Number of responses: [count]
Survey completion date range: [dates]
Dataset: [attach]
Screen for:
1. Speedsters: responses completed in under [X] minutes — implausibly fast given survey length
2. Straight-liners: respondents who selected the same response option for all matrix/grid questions
3. Duplicate submissions: same respondent submitting multiple times (check email, IP, device fingerprint where available)
4. Bot responses: implausible response patterns, timestamps at impossible speeds, generic open-text answers
5. Contradictory responses: respondents giving logically inconsistent answers to related questions
6. Out-of-quota responses: respondents who passed screening but whose profile data contradicts their screening answers
For each category: count of flagged responses, percentage of total sample, and recommended disposition (exclude / downweight / include with flag).
Final output: Cleaned dataset with exclusions applied + data quality memo documenting screening methodology and results for methods section.Prompt 3: Missing Value Pattern Analysis
Analyze the missing value patterns in the following dataset and recommend an imputation strategy.
Dataset: [attach or describe]
Variables with significant missingness: [list variables or confirm COCO should identify them]
Planned analysis: [describe — determines which variables must be complete]
Analyze:
1. Missing value rates: by variable, by time period, by geography, by respondent segment
2. Missingness mechanism: test each variable for MCAR (missing completely at random), MAR (missing at random conditional on observed variables), or MNAR (missing not at random)
3. Impact on planned analysis: for each high-missing variable used in the analysis, quantify how the missingness pattern will affect the findings if not addressed
4. Imputation options: for each variable, recommend: (a) complete case analysis (if MCAR and <5% missing), (b) single imputation method, (c) multiple imputation — with justification for each choice
5. Sensitivity check: recommend a sensitivity analysis approach to test whether findings change materially under different missing data assumptions
Output: Missing value analysis report + recommended imputation code/approach for each variable + text for methods section documentationPrompt 4: Financial Data Integrity Check
Validate the integrity of the following financial dataset before it is used in quantitative analysis.
Data description: [describe — e.g., equity prices, company financials, alternative data feed]
Source: [data provider name]
Coverage: [time period, geographic scope, asset universe]
Dataset: [attach or describe]
Validate:
1. Survivorship bias: does the dataset include delisted, bankrupt, or acquired entities — or only survivors? Quantify the survivorship bias if present
2. Look-ahead bias: are any variables derived using information not available at the point-in-time the observation represents?
3. Backcasting: identify any variables where the data provider applies retrospective revisions to historical data
4. Corporate action adjustments: confirm price data is appropriately adjusted for splits, dividends, and spin-offs
5. Coverage consistency: are there periods or segments where coverage drops significantly below the mean — indicating data gaps that could bias time-series or cross-sectional comparisons?
6. Unit and currency consistency: confirm all financial figures use consistent units and currency throughout the observation window
Output: Integrity assessment with each issue found, the time period and entity segment affected, estimated bias impact on analytical conclusions, and recommended data treatmentPrompt 5: Panel Data Attrition Analysis
Analyze attrition patterns in the following longitudinal panel dataset.
Dataset: [attach or describe — panel of respondents/entities tracked across multiple waves]
Number of waves: [count]
Wave dates: [list]
Initial sample size: [count at wave 1]
Analyze:
1. Attrition rate: overall and by wave — what percentage of the initial sample is present at each subsequent wave?
2. Attrition pattern: is attrition random or systematically related to key variables measured at baseline?
3. Attritors vs. completers: compare baseline characteristics of those who dropped out at each wave vs. those who remained — are they systematically different on key outcome-relevant dimensions?
4. Missing at random test: can the attrition pattern be explained by variables observed in the dataset — or does it appear to be driven by unobservable factors?
5. Impact on analytical validity: given the attrition pattern found, which planned analyses may produce biased estimates — and how large is the potential bias?
6. Remediation options: inverse probability weighting, multiple imputation for missing waves, bounding analysis for sensitivity
Output: Attrition analysis report + recommended weighting or imputation approach + methods section text documenting the attrition pattern and the correction applied36. AI Citation Network Mapping Engine
Maps the citation relationships among academic papers, patents, and reports in any research domain — identifying foundational works, emerging clusters, and knowledge flow pathways to guide literature strategy and research positioning.
Pain Point & How COCO Solves It
The Pain: Citation Networks Encode the Intellectual History of a Field — and Are Nearly Impossible to Navigate Manually
Understanding how knowledge in a field develops, which papers have had the most influence, and where the current frontier of research is located requires navigating citation networks — the web of references that connects papers to each other across time, institutions, and subfields. This navigation is essential for researchers positioning their own work, for funders assessing the impact of past investments, for analysts evaluating the maturity of a technology or evidence base, and for strategists understanding where intellectual communities are forming around emerging ideas. But citation network analysis requires both the data and the analytical tools to make sense of it, and most researchers have access to neither in a practical form. Citation counts in Google Scholar tell you a paper is highly cited; they don't tell you why, by whom, in what context, or whether the citations represent endorsement, critique, or incidental reference.
The structure of citation networks reveals information that is not available from any other source. Co-citation analysis — identifying papers cited together — reveals the intellectual community that researchers implicitly form around a shared set of foundational works. Bibliometric coupling — identifying papers that share many common references — reveals which research groups are working on closely related problems even if they don't cite each other directly. Citation burst detection identifies papers that experienced sudden spikes in citation rate, often indicating a field-defining finding that reoriented subsequent research. Forward citation tracing reveals the intellectual descendants of a foundational paper, showing which research directions it enabled and which it influenced most strongly. These analyses require processing tens of thousands of citation relationships — a task completely inaccessible to manual analysis.
For research strategy purposes, the most valuable use of citation network analysis is negative space: identifying the papers that ought to be cited together but aren't, the research groups that are working on related problems without engaging with each other's work, and the foundational assumptions underlying a field that have never been empirically tested because every paper in the tradition takes them as given. These structural gaps represent both opportunities for synthesis contributions and risks of duplicated effort. The researcher who publishes a paper connecting two previously isolated intellectual communities can achieve outsized influence with relatively straightforward analysis — but identifying those communities and their disconnection requires citation network analysis that most researchers lack the tools and time to perform.
How COCO Solves It
Citation Data Collection and Network Construction: COCO builds the complete citation graph for any research domain:
- Queries citation databases — Semantic Scholar, OpenCitations, CrossRef, Web of Science, Scopus, and Microsoft Academic — to collect citation relationships for all papers in the defined domain
- Constructs a directed citation network where each node is a paper and each edge is a citation relationship, with edge direction indicating the citing-cited relationship
- Resolves citation ambiguities: papers cited under multiple title variations, conference and journal versions of the same work, and preprints later published in edited form
- Enriches each node with metadata: authors, institutions, publication venue, publication date, abstract, open-access status, and altmetric indicators
- Calculates standard network metrics for each node: in-degree (times cited), out-degree (references cited), betweenness centrality, PageRank, and eigenvector centrality
Foundational Paper and Intellectual Ancestry Identification: COCO identifies the papers that built the field:
- Ranks papers by multiple centrality metrics to surface the most influential works — those most cited, those cited by the most important papers, and those that serve as bridges between subfields
- Traces the intellectual ancestry of current research: given any set of recent papers, identifies the foundational chain of prior work each one builds upon
- Identifies seminal papers that generated citation bursts — sudden large increases in citation rate indicating field-defining impact — with the burst time window and papers driving the burst
- Produces a chronological intellectual map showing how foundational ideas evolved over time and which papers represented genuine turning points in the field's development
- Generates an annotated reading list of foundational works, ordered by intellectual importance rather than recency
Research Cluster and Community Detection: COCO identifies intellectual communities within the citation network:
- Applies community detection algorithms (Louvain, Leiden, modularity optimization) to identify clusters of papers more densely connected to each other than to the rest of the network
- Characterizes each cluster by its dominant themes, leading institutions, key authors, and flagship papers
- Identifies inter-cluster connector papers and authors — researchers whose work bridges multiple intellectual communities and carries disproportionate influence on the field's integration
- Detects emerging clusters: communities of papers that are growing rapidly in citation density, indicating a new research direction forming before it has been labeled or widely recognized
- Maps institutional and national concentration within each cluster: are certain communities dominated by specific universities or national research systems?
Knowledge Gap and White Space Analysis: COCO identifies missing connections and under-explored areas:
- Identifies pairs of research clusters with high topical similarity but low citation density — communities working on related problems without engaging with each other's literature
- Detects assumptions foundational to the field that are widely cited but rarely tested empirically — potential targets for impactful replication or challenge studies
- Surfaces topics that receive extensive citation in reviews and meta-analyses but have sparse primary research — indicating demand for evidence that has not yet been produced
- Identifies geographic and institutional gaps in research coverage: domains where research production is concentrated in a few institutions or countries, leaving systematic blind spots
- Maps the age distribution of citations in current papers: fields heavily relying on old foundational work without recent empirical challenge may have stale assumptions ripe for revisiting
Citation Impact Trend and Disruption Analysis: COCO tracks how influence flows and changes over time:
- Calculates citation half-life for the domain: how quickly do papers become obsolete vs. remain influential across decades
- Identifies papers with consistently high citation rates over long time periods vs. one-time citation spikes that did not sustain — distinguishing lasting impact from momentary relevance
- Applies disruption index metrics to identify papers that replaced vs. built upon prior work — genuine paradigm shifts vs. incremental contributions
- Detects declining citation trajectories for previously dominant papers — indicating a research direction has been superseded or its findings challenged
- Forecasts expected citation trajectory for recent papers based on early citation velocity and the citation patterns of historically similar papers
Research Positioning and Strategy Report: COCO translates network analysis into actionable research strategy:
- Produces a complete citation network visualization with cluster coloring, node sizing by influence, and key papers labeled — suitable for presentations and grant applications
- Generates a research positioning analysis: given a set of papers representing the analyst's or research group's prior work, where do they sit in the network — which communities are they connected to and disconnected from?
- Recommends citation strategy: which high-influence papers are under-cited in the analyst's work, and which clusters should be engaged to increase cross-community visibility
- Identifies the journals and conferences where papers with the highest ongoing citation velocity are published — guiding venue selection for maximum impact
- Produces a gap analysis memo suitable for grant proposal background sections: documenting the state of knowledge, its limitations, and the specific gap the proposed research will fill
Results & Who Benefits
Measurable Results
- Citation network construction time: Complete network for a 10,000-paper domain built in under 6 hours vs. weeks of manual data collection and processing
- Foundational paper identification: COCO-identified foundational papers match expert panel consensus lists at a 93% overlap rate, while including an average of 4 high-influence papers that expert panels initially missed
- Research gap identification: Analysts using COCO citation network analysis identify an average of 2.8 actionable research gaps per domain review vs. 0.6 from unassisted literature reading
- Grant proposal success rate: Research proposals that include a citation-network-supported gap analysis and positioning statement show 22% higher funding success rates in competitive grant programs
- Literature review comprehensiveness: Systematic reviews using COCO citation network seeding identify 31% more eligible studies than reviews relying solely on keyword database searches
Who Benefits
- Academic Researchers: Map the intellectual landscape of their field with precision — identify what has been done, what hasn't, and where their work makes the most distinctive contribution, enabling stronger positioning in publications and grants
- Research Funding Agencies: Assess the impact of past investments by tracing how funded research has propagated through the citation network, and identify under-resourced intellectual communities that would benefit from targeted funding
- Technology Analysts and Innovation Scout Teams: Use citation network analysis of patent and academic literature simultaneously to track where scientific knowledge is flowing into applied technology, identifying emerging commercialization opportunities
- University Libraries and Research Support Teams: Provide researchers with citation network intelligence services that enhance the competitiveness of institutional research proposals and the impact of publications
💡 Practical Prompts
Prompt 1: Citation Network Construction and Overview
Build and analyze the citation network for the following research domain.
Research domain: [describe the field or sub-field — be specific]
Seed papers (known key works): [list 5–10 papers you know are central to this domain]
Time period: [e.g., 1990 to present / 2010 to present]
Network depth: [1-hop (direct citations only) / 2-hop (citations of citations) / full domain]
Deliverables:
1. Network statistics: total nodes (papers), total edges (citation relationships), network density, average degree
2. Top 20 most influential papers: ranked by PageRank / eigenvector centrality, with one-sentence description of each paper's contribution
3. Research cluster summary: how many distinct intellectual communities exist in this network, and what characterizes each?
4. Network visualization data: node list with coordinates and attributes for visualization in VOSviewer or Gephi
5. Temporal analysis: how has the network grown over time — citation volume by year and emergence of new clusters
Output: Citation network analysis report + visualization-ready data filesPrompt 2: Intellectual Ancestry Tracing
Trace the intellectual ancestry of the following paper(s) to identify the foundational works they build upon.
Target papers: [list paper titles, authors, and DOIs]
Analysis depth: [trace back 2 / 3 / 4 generations of citation chains]
For each target paper:
1. Direct references: which papers does this paper cite most critically (not incidentally)?
2. Second-generation ancestors: which papers do the direct references themselves draw on?
3. Foundational convergence: which papers appear in the ancestry of multiple target papers — the shared intellectual heritage of this research area?
4. Landmark papers: which papers in the ancestry chain represent genuine turning points — work that enabled the target paper to be written?
5. Intellectual lineage narrative: describe in 2–3 paragraphs how the ideas in the target paper developed from prior work
Output: Ancestry citation chain diagram + annotated bibliography of foundational works + intellectual lineage narrative suitable for a literature review introductionPrompt 3: Research Gap Identification
Identify research gaps in [domain] using citation network analysis.
Domain: [describe]
Seed paper set representing the current frontier: [list 10–20 recent high-citation papers]
Analyze:
1. Topically similar but citation-disconnected clusters: research communities working on related problems without citing each other — where are the missing bridges?
2. Review-demand vs. primary evidence gap: topics that frequently appear in review papers' "future research needed" sections but have limited primary studies published
3. Methods monopoly: topics where evidence is dominated by a single methodology — areas where alternative methods would add new insight
4. Geographic/population gaps: research domains where study populations are concentrated in specific regions, leaving generalizability gaps for other contexts
5. Temporal gaps: topics with foundational papers from 10+ years ago that have not been revisited with modern data, methods, or theoretical frameworks
Output: Gap analysis memo with each gap described, supporting evidence from the network analysis, and a suggested research question framing for each gap — suitable for grant proposal background sectionsPrompt 4: Research Cluster Characterization
Characterize the research clusters in the following citation network.
Citation network data: [attach node/edge files or describe domain for COCO to construct]
Cluster detection: [apply Louvain algorithm / use pre-identified clusters: list]
For each identified cluster:
1. Thematic identity: what is this cluster about — what research question or approach unifies the papers in it?
2. Key papers: the 5 most central papers in this cluster — titles, authors, citation count, one-sentence contribution summary
3. Leading authors and institutions: who is most active and influential in this cluster?
4. Temporal trajectory: is this cluster growing, stable, or declining in citation activity?
5. Connections to other clusters: which other clusters does this one cite most — what is the knowledge flow pattern?
6. Entry points: the 2–3 papers a new researcher should read first to enter this intellectual community
Output: Cluster profile document for each identified community + inter-cluster connection matrix showing citation flows between communitiesPrompt 5: Research Positioning Analysis
Analyze the citation network positioning of the following researcher or research group.
Researcher/group: [name]
Publication list: [attach or list key papers with DOIs]
Target domain network: [describe the domain or provide network data]
Analyze:
1. Current position: where do this researcher's papers sit in the domain's citation network — which clusters are they part of?
2. Citation in-flow: which intellectual communities cite this researcher's work most — who is building on it?
3. Citation out-flow: which intellectual communities does this researcher draw from most — who do they build on?
4. Bridge opportunities: which clusters are topically related to this researcher's work but not currently connected by citations — potential new audiences to engage?
5. Influence trajectory: is this researcher's citation impact growing, stable, or declining in recent years — and what does the trend suggest about research direction alignment with the field's evolution?
6. Strategic recommendations: which 3–5 papers in the domain should this researcher cite in future work to improve positioning, and which 2–3 clusters represent the highest-value audience for new research to target?
Output: Positioning analysis report with network visualization showing researcher's position relative to all clusters + strategic recommendations for maximizing research impact37. AI Grant Proposal Optimization Engine
Analyzes funder priorities, reviewer expectations, and successful proposal patterns to strengthen grant narratives, align evaluation criteria, and improve competitive positioning before submission.
Pain Point & How COCO Solves It
The Pain: Grant Writing Is High-Stakes, Time-Consuming, and Systematically Disadvantages Researchers Without Institutional Support
Grant writing is one of the most consequential and least taught skills in research. A successful NIH R01, NSF grant, Wellcome Trust award, or European Research Council project grant can fund a research program for years — and rejection means not just deferred research but deferred careers for junior investigators and research staff whose positions depend on funding. Yet the skills required to write a competitive proposal — translating complex technical research into compelling narratives aligned with funder priorities, structging specific aims pages that tell a coherent story, writing significance sections that clearly articulate the gap in knowledge being filled — are rarely taught systematically and are almost entirely acquired through trial, error, and mentorship that is unequally distributed across institutions. Researchers at R1 universities with strong grants administration offices have access to professional editors, institutional support for budget development, and colleagues who have succeeded with the same mechanism. Researchers at smaller institutions, in developing countries, or in early career stages often write their first competitive proposals in near-complete isolation.
The alignment problem between what researchers want to propose and what funders want to fund is a persistent source of avoidable rejection. Funding agencies publish detailed program announcements, strategic priorities, and in many cases the review criteria and scoring rubrics by which proposals will be evaluated — yet researchers frequently write proposals that address questions that interest them without carefully translating those questions into the language and framing that reviewer panels respond to. Program officers at NSF and NIH describe the same recurring weaknesses in unfunded proposals: insufficient justification of significance, unclear conceptual innovation beyond prior work, unrealistic aims that exceed what the budget and timeline can plausibly accomplish, preliminary data that doesn't adequately support the proposed approach, and research plans that read like laboratory notebooks rather than scientific narratives. Each of these weaknesses is identifiable in advance and correctable before submission — if the researcher has access to systematic feedback from someone who knows what a competitive proposal looks like.
The time pressure compounds the challenge. Grant applications involve coordinating multiple components — research plan narrative, specific aims, human subjects protocol, budget justification, letters of support, biosketches, facilities and resources — often under a deadline that arrives while the researcher is simultaneously running their laboratory, teaching, and supervising trainees. The narrative sections that most determine competitiveness are the last to be written because they require the most original thinking, and they receive the least time because the administrative requirements consumed the early preparation window. Researchers routinely submit proposals they know are not ready because the next deadline is 12 months away and the project cannot wait. The result is unnecessary rejection of scientifically meritorious research that a better-written, better-aligned proposal could have funded.
How COCO Solves It
Funder Priority Alignment Analysis: COCO maps the proposal against funder priorities with precision:
- Reads current funding opportunity announcements, strategic plans, prior year funding portfolios, and program officer statements to extract explicit and implicit funder priorities
- Identifies the specific language, concepts, and framing that this funder uses to describe the kind of research they want to fund — and maps the proposal's language against it
- Scores each section of the proposal on alignment with stated funder priorities, flagging misalignments and recommending specific language adjustments
- Analyzes the portfolio of recently funded grants from this mechanism to identify thematic patterns, methodological preferences, and investigator profile characteristics that predict success
- Compares the proposal's focus area, methodology, and innovation claims against the distribution of currently active grants to identify whether the proposed work would fill a gap or duplicate existing investment
Specific Aims Page Optimization: COCO strengthens the most critical one-page section:
- Evaluates the logical flow of the specific aims page: does the problem statement lead clearly to the gap, the gap to the objective, the objective to the aims, and the aims to the expected outcomes?
- Tests each aim for feasibility within the stated budget, timeline, and team: are the proposed experiments achievable in the grant period, or do they assume resources or preliminary data not yet available?
- Checks aim interdependence: are the aims designed so that failure of one aim doesn't prevent the others from being completed — or does the entire project collapse if a single aim encounters obstacles?
- Evaluates the innovation claims: is the proposed work genuinely novel relative to the prior literature, or does the novelty claim depend on not citing contradicting prior work?
- Benchmarks the aims page structure against high-scoring funded proposals in the same mechanism, identifying structural and language patterns associated with top scores
Significance and Innovation Section Strengthening: COCO sharpens the scientific rationale:
- Evaluates whether the significance section clearly and specifically articulates the gap in knowledge — not just the importance of the topic, but the specific missing piece that this proposal fills
- Tests the logical chain: problem → impact of problem → current limitation → why prior approaches failed → how this proposal's approach overcomes the limitation
- Strengthens innovation framing: distinguishes conceptual innovation (new theoretical framework), methodological innovation (new tools or approaches), and empirical innovation (new data or populations studied) — ensuring the innovation claim is specific and defensible
- Identifies places where significance claims are asserted without citation to supporting evidence, and recommends specific citations to strengthen each claim
- Compares significance framing against the review criteria language to ensure the section directly answers the questions reviewers are asked to score
Research Strategy Narrative Improvement: COCO improves scientific rigor and reviewability:
- Reads the full research strategy and evaluates each aim's experimental approach for completeness: hypothesis, design, methods, expected results, interpretation, and contingency plans for negative results
- Identifies missing elements that reviewers consistently flag as weaknesses: no power calculation, no statement of potential limitations, no alternative approach for high-risk experiments
- Evaluates the preliminary data section: does the data presented directly support the feasibility of the proposed approach, or does it address adjacent questions that don't reduce the scientific risk?
- Checks the narrative coherence of the research plan: does each section build logically on the prior one, or does the proposal read as disconnected experiments assembled under a single title?
- Identifies where the proposal can strengthen reviewer confidence through additional specificity — exact reagents, validated assay protocols, specific analysis pipelines — vs. where brevity is appropriate
Budget and Timeline Optimization: COCO ensures the financial and project plan is credible:
- Reviews budget justification for completeness: is every line item justified with specific, quantified rationale rather than generic descriptions?
- Checks budget against the research plan: are all the experiments described in the research strategy covered by the proposed budget, or are there resource gaps?
- Evaluates the project timeline for realism: are the milestones achievable given typical experiment lead times, data analysis cycles, and manuscript preparation?
- Flags budget items that are commonly challenged by program officers: excessive consultant rates, equipment requests for items that should be available through institutional facilities, or personnel allocations inconsistent with the proposed work scope
- Compares the budget to publicly available funding data on similar mechanisms to identify whether the request is within the competitive range for grants of this type
Reviewer Simulation and Pre-Submission Review: COCO simulates the reviewer experience before submission:
- Reads the complete proposal from the perspective of a reviewer assigned to score it on the applicable review criteria (Significance, Investigators, Innovation, Approach, Environment for NIH; or equivalent for other funders)
- Generates a mock review summary statement with a preliminary score estimate and specific strengths and weaknesses formatted to match the actual review template
- Identifies the top three weaknesses most likely to drive a low score, with specific revision recommendations for each
- Checks the proposal against common rejection reasons documented by the funding agency — does this proposal have the warning signs?
- Produces a revision priority list: the changes that would most improve the competitive score, ordered by impact and feasibility within the remaining preparation time
Results & Who Benefits
Measurable Results
- First-submission success rate: Researchers using COCO-assisted proposal optimization report first-submission funding rates of 34–41% vs. a field-wide average of 18–25% for competitive mechanisms like NIH R01 and NSF CAREER
- Review score improvement: Proposals revised using COCO reviewer simulation feedback show average percentile improvement of 12–18 points on resubmission relative to the prior summary statement weaknesses
- Preparation time reduction: Researchers report spending 30–40% less time on narrative revision cycles when COCO-assisted alignment analysis identifies misalignments early in the drafting process
- Funder language alignment score: Proposals optimized with COCO score 2.1x higher on blind funder-priority alignment assessments by program officers compared to unassisted proposals from the same research group
- Aims page revision cycles: Average number of specific aims page revision rounds reduced from 6–8 cycles with peer feedback alone to 3–4 cycles when COCO structural analysis is incorporated after the second draft
Who Benefits
- Early-Career Researchers: Gain access to systematic proposal development support that levels the playing field with researchers at better-resourced institutions, improving first-submission competitive rates during the career stage when grant success is most career-defining
- Research Administrators and Grants Offices: Provide higher-quality pre-submission review support to larger investigator portfolios without proportional increases in grants office staffing
- Principal Investigators with Multiple Active Grants: Efficiently manage proposal development across multiple concurrent submissions with systematic alignment analysis that reduces the time investment per proposal
- Research Institutions and Funding Agencies: Increase the overall quality and rigor of submitted proposals, reducing the noise-to-signal ratio in competitive review processes and improving the probability that funded research represents genuinely high-quality science
💡 Practical Prompts
Prompt 1: Funder Priority Alignment Check
Analyze the alignment between my grant proposal and the target funding opportunity.
Funding opportunity: [title, number, link or attach announcement]
Funder: [NIH / NSF / ERC / Wellcome / foundation name]
Mechanism: [R01 / R21 / CAREER / ERC Starting Grant / etc.]
My proposal: [attach or paste the specific aims and significance sections]
Analyze:
1. Priority alignment: identify the explicit priorities in the funding announcement and score each section of my proposal on how directly it addresses them (1–5 scale with commentary)
2. Language mapping: what terms, concepts, and framing does the funder use that I should adopt or reference more explicitly?
3. Portfolio fit: based on the description of currently funded grants in this program, does my proposal fill a gap or duplicate existing investment?
4. Misalignments: where does my proposal address questions that this funder is unlikely to prioritize — should these sections be de-emphasized or reframed?
5. Missing elements: what does the funder emphasize that my proposal doesn't address — what should I add?
Output: Alignment scorecard by proposal section + specific revision recommendations with example reframing languagePrompt 2: Specific Aims Page Review
Review and strengthen my specific aims page.
Specific aims page: [attach or paste — should be approximately 1 page]
Funding mechanism: [name]
Research area: [describe the field and research question briefly]
Evaluate:
1. Narrative flow: does the page tell a coherent story from problem → gap → objective → aims → expected outcome?
2. Problem statement: is the problem clearly defined and its significance established in the first paragraph?
3. Gap identification: is the specific gap in knowledge that this proposal fills clearly articulated — not just the importance of the topic?
4. Aims structure: are the aims parallel in construction, appropriately scoped, and designed so the project succeeds even if one aim encounters obstacles?
5. Innovation: is the proposed innovation specific and defensible — or is it generic ("this has never been done before")?
6. Feasibility signals: does the page convey that this team can do this work within the proposed budget and timeline?
Output: Detailed critique with specific revision recommendations for each element + a revised version of the aims page incorporating the most critical improvementsPrompt 3: Reviewer Simulation
Simulate a peer review of my grant proposal.
Funding mechanism: [name and funder]
Review criteria: [paste from the funding announcement, or confirm COCO should apply standard criteria for this mechanism]
Proposal: [attach full proposal document]
Simulate a peer review:
1. Apply each review criterion on the applicable scoring scale
2. Write a summary statement formatted to match the actual review template for this mechanism, including:
- Overall impact / priority statement
- Scored strengths for each criterion
- Scored weaknesses for each criterion
- Additional comments
- Preliminary score estimate
3. Identify the top 3 weaknesses most likely to drive a lower-than-competitive score
4. For each weakness: describe the specific problem, explain why a reviewer would penalize it, and provide a concrete revision recommendation
Output: Mock summary statement + revision priority list with specific, actionable recommendations for each weaknessPrompt 4: Budget Justification Strengthening
Review and strengthen the budget justification for my grant proposal.
Budget: [attach budget pages or paste budget summary]
Budget justification narrative: [attach or paste]
Research plan summary: [describe the proposed experiments and team]
Funding mechanism and budget limits: [e.g., NIH R01, modular $250K/year limit]
Review:
1. Completeness: is every budget line item present in the justification with specific, quantified rationale?
2. Research plan alignment: are all experiments described in the research strategy covered by the budget — and are there items in the budget that don't correspond to described activities?
3. Personnel justification: is the effort allocation for each team member consistent with their described role in the proposed work?
4. Common challenge flags: identify any line items likely to be questioned by a program officer or grants management specialist
5. Comparability: is the overall budget request consistent with funded grants of similar scope in this mechanism?
Output: Annotated budget justification with specific revision recommendations for each weakness + suggested rewritten justification text for the most critical itemsPrompt 5: Resubmission Response Planning
Plan the resubmission strategy for the following unfunded proposal.
Previous submission score: [percentile or score received]
Summary statement from reviewers: [attach or paste]
Original proposal: [attach or paste]
Resubmission deadline: [date]
Plan:
1. Score the reviewer concerns by impact: which weaknesses drove the most score damage — prioritize addressing these first
2. Categorize each concern: (a) legitimate scientific weakness requiring substantive revision, (b) misunderstanding by reviewers requiring clarification, (c) perceived weakness addressable by additional preliminary data
3. Develop the Introduction to the resubmission: how to acknowledge reviewer concerns, describe what has changed, and signal the improvements — in the required 1-page limit
4. Revision plan: specific changes to be made to each section, ordered by priority
5. New preliminary data recommendations: if additional data would strengthen the resubmission, what experiments should be prioritized in the available time?
6. Risk assessment: are there core concerns from the reviewers that cannot be fully addressed in a resubmission — and does the overall competitive potential justify resubmission to this mechanism?
Output: Resubmission strategy memo + draft Introduction page + prioritized revision checklist38. AI Meta-Analysis Workflow Engine
Automates the statistical and methodological workflow of quantitative meta-analysis — from data extraction standardization through heterogeneity testing, publication bias assessment, and Forest plot generation.
Pain Point & How COCO Solves It
The Pain: Meta-Analysis Is the Highest-Value Research Synthesis Method and the Most Technically Demanding
Meta-analysis — the statistical combination of effect sizes from multiple independent studies — occupies the apex of the evidence hierarchy. A well-conducted meta-analysis can resolve contradictions in a literature that decades of individual studies could not settle, provide definitive effect size estimates with narrow confidence intervals, and establish the conditions under which an effect is large, small, or absent. Clinical guidelines, regulatory decisions, and investment theses in evidence-based domains all look first to meta-analytic evidence. Yet conducting a rigorous meta-analysis requires expertise across multiple specialized domains simultaneously: systematic review methodology to ensure the underlying literature search is comprehensive, biostatistics for effect size calculation and pooled estimation, research methods for appropriate subgroup analysis and moderator specification, and scientific communication for presenting results in ways that reviewers and practitioners can interpret and apply. Very few researchers have training across all of these domains, and those who do are in high demand.
The technical workflow of meta-analysis is riddled with decision points where incorrect choices invalidate the entire analysis. Effect sizes must be extracted from heterogeneous statistical reporting: some papers report means and standard deviations, others report t-statistics, others report odds ratios, correlation coefficients, or hazard ratios — and these must all be converted to a common metric using formulas that vary by study design and outcome type. Missing data — studies that report p-values but not effect size magnitude, or confidence intervals without sample sizes — must be handled through imputation methods that are valid only under specific assumptions. The choice between fixed-effects and random-effects models is not merely a statistical preference but a theoretical claim about whether all studies are estimating the same true effect or a distribution of effects — a claim with major implications for how the pooled estimate should be interpreted and applied.
Subgroup analysis and meta-regression — the most valuable parts of a meta-analysis for understanding when and for whom effects are large — are also the most easily mishandled. Pre-specifying moderators in a registered protocol before conducting the analysis is essential for avoiding fishing; conducting post-hoc subgroup analyses until a significant finding appears is statistically invalid but common under publication pressure. Heterogeneity must be tested using Cochran's Q and I², but these statistics are widely misinterpreted — a high I² does not mean the studies are incomparable; a low I² does not mean they are homogeneous. Publication bias assessment requires funnel plot interpretation and Egger's test that are poorly understood outside of biostatistics training. And the GRADE framework for rating overall evidence quality requires systematically applying complex judgment rules to the entire evidence body — a task that takes experienced systematic reviewers hours per outcome.
How COCO Solves It
Effect Size Extraction and Harmonization: COCO standardizes statistical data from heterogeneous reporting formats:
- Extracts effect sizes, sample sizes, and variance measures from each included study regardless of the statistical form in which they are reported — means/SDs, t-statistics, F-statistics, p-values, odds ratios, correlation coefficients, or hazard ratios
- Converts all effect sizes to the specified common metric (Cohen's d, Hedges' g, log odds ratio, Fisher's z-transformed r, or log hazard ratio) using appropriate formulas for each conversion
- Calculates standard errors and confidence intervals for each converted effect size, applying continuity corrections for zero-cell counts in binary outcome studies
- Handles multi-arm studies appropriately: splitting the control group when two active treatments are compared, or using appropriate methods for three-arm trials
- Flags studies where the reported statistics are insufficient to calculate or convert an effect size, specifying exactly what additional information is needed
Fixed-Effects and Random-Effects Model Specification: COCO applies appropriate pooling models with full transparency:
- Implements inverse-variance weighted fixed-effects pooling and DerSimonian-Laird random-effects pooling simultaneously, comparing results and flagging where the two approaches meaningfully diverge
- Tests the theoretical appropriateness of fixed-effects vs. random-effects framing for the specific research question, and recommends the more defensible model with justification
- Applies the Hartung-Knapp-Sidik-Jonkman correction for random-effects confidence intervals when study counts are small (fewer than 25 studies), improving coverage probability
- Implements Bayesian meta-analysis as an alternative or complement to frequentist pooling when specified, with prior distribution specification and posterior summary generation
- Generates a sensitivity analysis comparing pooled estimates across model specifications — fixed-effects, random-effects, robust variance estimation — to test conclusion robustness
Heterogeneity Assessment and Source Exploration: COCO quantifies and explains between-study variation:
- Calculates Cochran's Q statistic, I² index, H² statistic, and Tau² (between-study variance) with confidence intervals for all heterogeneity measures
- Interprets heterogeneity magnitude using benchmarks appropriate to the research domain, avoiding the common misapplication of generic I² thresholds
- Conducts pre-specified subgroup analyses: stratifying by study design, population characteristics, intervention features, or outcome definition to identify sources of heterogeneity
- Runs meta-regression on continuous moderator variables — year of publication, sample size, intervention duration, mean age — to identify variables that explain between-study variation
- Generates L'Abbé plots for binary outcome data and bubble plots for meta-regression to visualize heterogeneity sources
Publication Bias Detection and Correction: COCO applies multiple complementary bias assessment methods:
- Generates funnel plots with Egger's test for asymmetry, reporting the regression intercept, standard error, and p-value with appropriate interpretation caveats for small study numbers
- Applies Begg and Mazumdar rank correlation test as a non-parametric complement to Egger's regression test
- Implements trim-and-fill analysis to estimate the number of missing studies and recalculate the bias-adjusted pooled estimate
- Applies the Vevea and Hedges selection model as a more sophisticated alternative to trim-and-fill when study characteristics support its assumptions
- Calculates fail-safe N (Rosenthal's file drawer number): how many null-result studies would need to exist and be unpublished to reduce the pooled effect to a non-significant level
Forest Plot and Visualization Generation: COCO produces publication-quality output:
- Generates Forest plots with study-level effect sizes, confidence intervals, weights, and the pooled diamond — formatted to journal specifications including JAMA, BMJ, Lancet, or custom styles
- Produces subgroup Forest plots with separate diamonds for each subgroup and a test of subgroup differences
- Generates funnel plots, Galbraith plots, and Baujat plots for heterogeneity visualization
- Creates L'Abbé plots for binary outcomes and meta-regression bubble plots
- Exports all plots in vector formats (SVG, EPS) and raster formats (TIFF 300+ DPI) suitable for journal submission alongside underlying data tables
GRADE Evidence Quality Rating: COCO applies the GRADE framework to rate overall evidence certainty:
- Applies the five GRADE downgrading factors — risk of bias, inconsistency, indirectness, imprecision, publication bias — to the body of evidence for each outcome
- Applies the three GRADE upgrading factors — large effect size, dose-response gradient, plausible confounding in the opposing direction — for non-randomized evidence
- Generates GRADE evidence profile tables and Summary of Findings tables in the standard format required by Cochrane reviews and clinical guideline development
- Provides written rationale for each GRADE judgment, suitable for inclusion in the review text and for transparency in replication
- Produces an overall evidence certainty rating (high/moderate/low/very low) with a plain-language interpretation of what the certainty level means for practice and policy
Results & Who Benefits
Measurable Results
- Effect size extraction time: Standardized effect size extraction for a 50-study meta-analysis completed in under 3 hours vs. 2–3 analyst days for manual extraction and conversion
- Statistical error rate: Effect size calculation and conversion errors reduced from an average of 11% in manual extraction to under 1.5% with COCO-assisted extraction and formula verification
- Forest plot production time: Publication-quality Forest plots generated in under 30 minutes from completed extraction data vs. 4–8 hours using statistical software by an analyst without advanced programming skills
- GRADE rating agreement: COCO-generated GRADE ratings agree with expert panel ratings at 89% concordance on the overall certainty level and 94% on individual domain judgments
- Time to submission-ready manuscript: Complete meta-analysis results section — including all statistical tests, plots, and GRADE tables — produced in under 2 weeks from the completion of study screening, vs. 2–4 months with fully manual statistical workflow
Who Benefits
- Academic Meta-Analysts: Eliminate the most technically demanding and error-prone stages of the meta-analysis workflow — effect size extraction, model specification, and heterogeneity analysis — while maintaining scientific ownership of design decisions and interpretation
- Clinical Guideline Development Groups: Produce rigorous, GRADE-compliant evidence syntheses on accelerated timelines to support active clinical practice guideline updates without compromising evidence quality standards
- Health Technology Assessment Organizations: Generate meta-analytic evidence summaries for reimbursement decisions, comparative effectiveness reviews, and benefit-risk assessments within compressed regulatory and budget cycles
- Research Synthesis Centers and Cochrane Review Groups: Scale the number of high-quality meta-analyses produced annually without proportional growth in biostatistician headcount, by enabling research teams with substantive expertise to handle the statistical workflow with systematic AI support
💡 Practical Prompts
Prompt 1: Effect Size Extraction and Standardization
Extract and standardize effect sizes from the following included studies for our meta-analysis.
Outcome of interest: [describe — e.g., depression symptom severity, 6-month abstinence rate, mean difference in systolic blood pressure]
Target effect size metric: [Cohen's d / Hedges' g / log odds ratio / correlation r / log hazard ratio]
Included studies: [attach extraction table or list studies with their statistical results]
For each study:
1. Identify the reported statistic type (means/SDs, t-statistic, F-statistic, p-value, OR, HR, correlation)
2. Convert to [target metric] using the appropriate formula — show the conversion formula and inputs used
3. Calculate the standard error and 95% confidence interval for the converted effect size
4. Note the sample size used in the conversion (total N, or N per group for between-subjects designs)
5. Flag any studies where insufficient statistics are reported to calculate the effect size — specify what is missing
Output: Completed extraction table with: Study ID, N, converted effect size, SE, 95% CI, conversion method, and flags for any problems or approximations madePrompt 2: Meta-Analysis Statistical Synthesis
Conduct the meta-analytic statistical synthesis for the following dataset.
Effect size data: [attach table with study IDs, effect sizes, standard errors, and sample sizes]
Analysis plan:
- Primary model: [fixed-effects / random-effects / both for comparison]
- Subgroup analyses: [list pre-specified subgroups — e.g., by study design, population age, intervention duration]
- Moderator variables for meta-regression: [list continuous variables to test]
- Publication bias methods: [funnel plot + Egger's test / trim-and-fill / selection model]
Conduct:
1. Pooled effect size estimation under specified model(s) with 95% CI and p-value
2. Heterogeneity statistics: Q, df, p-value, I², Tau², and their 95% CIs
3. All specified subgroup analyses with test of subgroup differences
4. Meta-regression results for each continuous moderator
5. Publication bias assessment using all specified methods
Output: Complete statistical results table + interpretive narrative for the Results section + all plot data files for Forest plot, funnel plot, and meta-regression visualizationsPrompt 3: Forest Plot Specification
Generate a publication-quality Forest plot for the following meta-analysis.
Data: [attach or paste — study labels, effect sizes, confidence intervals, weights]
Plot specifications:
- Effect size metric label: [e.g., "Standardized Mean Difference (Hedges' g)"]
- Null value for vertical line: [0 for SMD/MD, 1 for OR/RR/HR]
- Study ordering: [by year / by effect size magnitude / by study weight / by subgroup]
- Subgroup display: [yes/no — if yes, specify subgroup variable]
- Weight display: [show percentage weights in a column]
- Diamond position: [at the bottom / at each subgroup + overall]
- Target journal style: [JAMA / BMJ / Lancet / APA / custom]
- Output format: [SVG + TIFF 300DPI]
Generate:
- Forest plot in specified format
- Underlying data table confirming the values plotted
- Figure caption text following the target journal's style guidePrompt 4: Heterogeneity Source Exploration
Investigate the sources of heterogeneity in the following meta-analysis.
Meta-analysis results: [paste pooled results including I² and Tau² estimates]
Study-level data: [attach extraction table with study characteristics]
Observed heterogeneity: I² = [value]%, Tau² = [value]
Investigate:
1. Subgroup analyses by each of the following study characteristics: [list — e.g., study design, population age range, intervention dose, country income level, risk-of-bias rating]
2. For each subgroup: pooled effect within subgroup, I² within subgroup, and test of subgroup differences
3. Meta-regression on continuous moderators: [list variables]
4. Influence analysis: identify any single study whose exclusion materially changes the pooled estimate or I² — confirm robustness
5. Outlier detection: identify studies with effect sizes falling outside the 95% prediction interval — do they share common characteristics?
Output: Complete heterogeneity exploration results table + narrative interpretation identifying the most plausible sources of observed heterogeneity + recommendations for how to describe residual unexplained heterogeneity in the DiscussionPrompt 5: GRADE Evidence Quality Assessment
Apply the GRADE framework to rate the certainty of evidence for the following outcome.
Outcome: [describe — e.g., reduction in depressive symptoms at 6 months]
Number of studies: [count]
Total participants: [count]
Study designs: [RCTs / observational / mixed]
Pooled effect estimate: [effect size, CI, p-value]
Heterogeneity: I² = [value]%, Tau² = [value]
Assess each GRADE domain:
1. Risk of bias: overall RoB profile of included studies — downgrade if most studies are at serious/critical risk
2. Inconsistency: degree of heterogeneity — downgrade if substantial and unexplained
3. Indirectness: how directly does the evidence apply to our population, intervention, and outcome of interest?
4. Imprecision: are the confidence intervals around the pooled estimate wide enough to be compatible with both important benefit and no effect?
5. Publication bias: is there evidence of selective reporting — downgrade if strongly suspected
Starting certainty: [High (RCTs) / Low (observational)]
Apply any applicable upgrading factors (large effect, dose-response, opposing confounding).
Output: GRADE evidence profile table in standard format + written rationale for each domain judgment + overall certainty rating (High/Moderate/Low/Very Low) with plain-language interpretation39. AI Research Collaboration Matching Engine
Identifies the most productive potential research collaborators based on complementary expertise, methodological fit, publication network overlap, and funding compatibility — accelerating team formation for grant applications and large-scale projects.
Pain Point & How COCO Solves It
The Pain: Research Collaboration Formation Is Inefficient, Proximity-Dependent, and Systematically Reproduces Existing Networks
Finding the right research collaborator is one of the most consequential decisions a researcher makes, yet it remains almost entirely dependent on informal networks, conference encounters, and institutional proximity — mechanisms that are inefficient, geographically constrained, and systematically biased toward researchers who already know each other or share the same social circles. A researcher at a major university in an under-resourced department may be unaware that the most methodologically compatible collaborator for their next grant exists at a smaller institution across the country or in another country entirely. The researcher whose work would most benefit from a biostatistics partner doesn't find that partner until the grant deadline forces them to ask whoever happens to be available in their department — often not the best fit. Junior researchers who lack established networks are at a particularly severe disadvantage: their collaboration opportunities are disproportionately shaped by their PhD advisor's network, reproducing intellectual genealogies that limit scientific diversity.
The misalignment problem is structural. Large multi-investigator grants — NIH P01 programs, NSF Research Coordination Networks, EU Horizon collaborative projects — explicitly require multidisciplinary teams with complementary expertise that no single researcher or single institution can provide. Assembling these teams requires identifying not just subject matter experts but the specific combination of methodological skill sets, patient/data access, and regulatory expertise that the project demands. A team missing a biostatistician with experience in longitudinal multilevel modeling, or a computational chemist with specific software expertise, or a regulatory affairs expert with a specific agency relationship, may have a technically excellent scientific plan that reviewers assess as unfeasible for lack of the right team. Researchers without existing relationships in needed expertise areas typically resort to cold outreach based on publication searches — identifying candidates but having no systematic way to assess fit, availability, or the likelihood of a productive collaboration before investing significant time in relationship-building.
The collaboration formation process itself consumes resources that most research teams can't afford. A meaningful assessment of a potential collaborator's methodological approach, publication quality, grant history, and scientific perspective requires reading a substantial body of their work — not just counting publications or citations. Understanding whether two researchers' methodological philosophies are compatible requires deeper engagement than a title or abstract scan. And when a research team needs to document the qualifications and collaboration rationale in a grant application — justifying why this specific combination of investigators is optimal for this project — the synthesis of that case requires analysis that is currently done ad-hoc and inconsistently.
How COCO Solves It
Expertise Profile Construction: COCO builds comprehensive research profiles for potential collaborators:
- Analyzes the full publication record of each candidate collaborator: research topics, methodological approaches, data types used, populations studied, and scientific questions addressed
- Extracts expertise dimensions from publication metadata and abstract content: substantive domains, statistical methods, laboratory techniques, clinical specializations, and policy areas
- Profiles grant funding history: funding agencies, mechanism types, award sizes, and active vs. prior funding — indicating research capacity and funding compatibility
- Maps institutional resources: facilities, datasets, patient populations, and equipment mentioned in publications and grant abstracts that indicate unique research capabilities
- Constructs a structured expertise vector for each researcher that enables systematic comparison against the requesting researcher's profile and project needs
Complementarity Analysis and Gap Matching: COCO identifies collaborators who fill specific gaps:
- Analyzes the requesting researcher's own expertise profile to identify what skills, methods, data access, and domain knowledge the proposed project requires but they don't currently have
- Searches the global research population for candidates whose expertise profiles complement the identified gaps — not just broad thematic overlap, but specific methodological or domain fit
- Weights candidates by the specificity of their complementarity: a biostatistician with direct experience in the exact analytical method the project requires ranks above a general biostatistician
- Identifies whether each candidate's contribution would be primary scientific leadership, methodological support, data access, or regulatory/translational expertise — enabling role-appropriate matching
- Presents candidates in ranked order with an explicit statement of what each would contribute to the project and how they address the identified gap
Collaboration Network Analysis: COCO assesses collaboration fit using network-level signals:
- Maps the existing collaboration networks of both the requesting researcher and each candidate: who have they successfully published with before, and in what configurations?
- Identifies second-degree connections: candidates who have collaborated with someone in the requesting researcher's network, providing a warm introduction pathway
- Detects collaboration bridge opportunities: candidates who sit at the intersection of two previously disconnected research communities that the proposed project needs to bring together
- Assesses collaboration track record: does the candidate have a history of completing multi-investigator publications, or do their collaborations typically remain at the co-authorship planning stage?
- Flags candidates whose recent collaboration patterns suggest they are actively seeking new partnerships vs. those who appear fully committed to existing research programs
Publication Quality and Impact Assessment: COCO evaluates research output quality beyond simple citation counts:
- Assesses the quality tier of each candidate's publication venues: journal impact factors, acceptance rates, and field-normalized citation rates
- Evaluates the candidate's citation impact relative to their field and career stage — distinguishing genuinely high-impact researchers from those with high counts in high-volume fields
- Identifies the candidate's most-cited papers and assesses whether the cited work is methodologically or substantively relevant to the proposed project
- Analyzes the candidate's authorship patterns: are they typically first or last author (indicating intellectual leadership) or middle author (indicating collaborative contributor roles)?
- Reviews recent publication trajectory: is the candidate's research output accelerating, stable, or declining — indicating available bandwidth and research momentum
Grant Funding Compatibility Analysis: COCO assesses collaborative grant potential:
- Reviews each candidate's grant history for funding agency preferences, mechanism experience, and budget scale — identifying candidates who can contribute as PIs, co-Is, or consultants
- Identifies candidates with existing funded grants in complementary areas who could contribute as co-investigators without requiring new funding for their participation
- Flags candidates with prior experience on the specific funding mechanism the project targets — experience with NIH P01, NSF center grants, or EU Horizon proposals is a significant practical asset
- Assesses potential conflicts: do any candidates have existing grants that would create budget overlap or commitment conflicts?
- Identifies candidates who have previously received funding from the target agency, providing insight into alignment with funder priorities and reviewer receptiveness to their participation
Collaboration Opportunity Report and Outreach Support: COCO produces structured deliverables to accelerate team formation:
- Generates a ranked collaboration candidate report with profile summary, complementarity rationale, connection pathway, and collaboration fit assessment for each candidate
- Drafts personalized outreach messages for each candidate: framing the collaboration opportunity from the candidate's perspective, highlighting the project's alignment with their research interests
- Produces a team composition rationale memo for grant applications: documenting why this specific combination of investigators is optimal for the project, with each team member's contribution and unique qualification articulated
- Identifies alternative candidates for each role in case primary candidates are unavailable or unresponsive — ensuring backup options at each position
- Generates a team building timeline: the sequence of outreach, negotiation, and commitment needed to assemble the team before the grant submission deadline
Results & Who Benefits
Measurable Results
- Collaboration identification time: High-quality collaboration candidates identified in under 4 hours vs. 2–4 weeks of informal network inquiries and publication searches
- Geographic diversity: Teams assembled using COCO matching include investigators from an average of 2.8 more institutions than teams assembled through existing personal networks alone
- Complementarity fit score: Collaborators identified by COCO match the requesting researcher's identified methodological gaps at a rate rated "high fit" by independent expert assessment in 81% of cases vs. 43% for informally recruited collaborators
- Grant application quality: Multi-investigator grant applications with COCO-supported team rationale narratives receive reviewer ratings of "team is well-qualified and complementary" in 76% of reviews vs. 51% for comparable applications without systematic team justification
- Collaboration completion rate: Research collaborations initiated through COCO-supported matching result in at least one co-authored publication within 3 years at a rate of 68% vs. an estimated 35–45% for cold-contact collaborations initiated without structured compatibility analysis
Who Benefits
- Research Analysts and Principal Investigators: Find methodologically and substantively compatible collaborators efficiently, expanding the scope of feasible projects and the quality of multi-investigator grant applications
- Research Administration and Grants Offices: Assist investigators in assembling grant-competitive teams faster and with better documented rationale, improving multi-investigator application quality
- Funding Agencies and Research Councils: Enable better-matched multidisciplinary teams on large collaborative grants by providing researchers with tools to look beyond their immediate networks
- Early-Career Researchers: Access collaboration opportunities that are normally gated by established networks, leveling the playing field and enabling independent career development beyond their PhD advisor's collaboration circle
💡 Practical Prompts
Prompt 1: Collaboration Candidate Search
Find potential research collaborators for the following project.
Project description: [describe the research question, proposed methodology, and scientific goals in 2–3 paragraphs]
Principal investigator background: [describe your own expertise, methods, and current research program]
Collaboration gaps: [what expertise, methods, data access, or domain knowledge does the project need that you don't currently have?]
Grant mechanism target (if applicable): [e.g., NIH R01, NSF collaborative grant, EU Horizon partnership]
Geographic constraints: [any institutional or geographic requirements for the collaboration]
Find candidates who:
1. Have published research in [describe substantive area] using [describe methods needed]
2. Have experience with [describe specific capabilities — datasets, patient populations, equipment, regulatory experience]
3. Have a track record of completing multi-author collaborative projects
4. Are at career stage [junior / mid-career / senior] — specify if important for team balance
For each candidate: name, institution, expertise summary, most relevant publications, collaboration track record, and why they are a strong fit for this project
Rank by complementarity fit, then by publication quality, then by collaboration network proximity.Prompt 2: Potential Collaborator Profile Analysis
Analyze the research profile of the following potential collaborator to assess fit for our project.
Candidate: [name, institution, Google Scholar or Semantic Scholar link]
Our project: [describe the research question and the specific role this person would play]
Specific capabilities needed: [describe what we need from this collaborator]
Analyze:
1. Expertise match: what are the candidate's primary research topics, methods, and domains — and how directly do they match what we need?
2. Publication quality: what are their 5 most impactful publications — are they methodologically and substantively relevant to our project?
3. Grant history: what agencies, mechanisms, and scales have they been funded at — are they eligible and positioned for the role we're proposing?
4. Collaboration record: who have they published with — do they have a track record of completing collaborative projects?
5. Availability signals: are they currently active in publishing — or do patterns suggest they may have reduced bandwidth?
6. Red flags: any signals of research integrity concerns, productivity gaps, or network factors that would affect collaboration risk?
Output: Collaboration fit assessment (Strong / Moderate / Weak) with rationale + specific questions to ask in an initial conversation to assess fit furtherPrompt 3: Team Composition Rationale for Grant
Draft the team composition and expertise rationale section for our grant application.
Grant mechanism: [name and funder]
Research plan summary: [attach or describe]
Team members:
- PI: [name, institution, expertise summary]
- Co-I 1: [name, institution, expertise summary]
- Co-I 2: [name, institution, expertise summary]
- [Additional team members as applicable]
Draft:
1. Team overview narrative: how does this team combine the expertise required to execute the proposed research?
2. Individual contribution statements: for each team member, describe their specific role and unique qualification for that role — linking their expertise directly to project activities
3. Collaboration history: where team members have prior collaboration experience, document it; where they don't, articulate why the proposed collaboration is nonetheless well-founded
4. Institutional resources: what facilities, datasets, or infrastructure does each institution contribute that is essential to project success?
5. Team diversity and balance: how does the team achieve appropriate disciplinary, career stage, and institutional diversity?
Length: [1–2 pages]. Match the style and specificity of funded proposals in [funder name] mechanisms.Prompt 4: Collaboration Outreach Message Drafting
Draft outreach messages to the following potential collaborators for our research project.
Project overview: [describe in 3–4 sentences — accessible to someone outside your specialty]
Our team and institution: [describe briefly]
The role we're seeking: [what would this person contribute and at what commitment level]
Funding status: [unfunded exploratory / planning for [specific grant] / already funded]
Timeline: [when we would need a commitment]
Candidates to contact:
1. [Name, institution, relevant connection or context]
2. [Name, institution, relevant connection or context]
For each candidate, draft a message that:
1. Opens with a specific reference to their work that shows genuine engagement (not generic praise)
2. Describes the project in terms of its scientific opportunity, not just our need for their expertise
3. States the collaboration role and time commitment clearly and honestly
4. Includes a low-commitment next step (a 30-minute call) rather than immediately asking for a commitment
5. Is appropriately brief (under 300 words) for an unsolicited outreach
Tone: professional and collegial. Do not oversell or create urgency pressure.Prompt 5: Research Partnership Portfolio Analysis
Analyze our institution's research collaboration portfolio to identify patterns and opportunities.
Institution: [name]
Analysis period: [e.g., last 5 years]
Data source: [faculty publication list / grant portfolio / both]
Analyze:
1. Collaboration intensity: what proportion of publications and grants involve external collaborators vs. single-institution work?
2. Partner institution map: which external institutions appear most frequently as collaboration partners — geographic and institutional type distribution
3. Disciplinary breadth: are collaborations crossing disciplinary boundaries or concentrated within similar fields?
4. Network gaps: which research areas at our institution have the weakest external collaboration networks — departments most working in isolation
5. Collaboration success rates: do collaborative publications and grants show higher citation impact and funding success rates than single-institution work?
6. Strategic opportunities: which external institutions or research groups would be highest-value new collaboration partners based on our faculty's research profile and their complementary capabilities?
Output: Collaboration portfolio report + network visualization data + strategic partnership recommendations for the next 3-year planning cycle
