Marvin Labs
Implementing AI in Equity Research: Step-by-Step Team Rollout Guide

Implementing AI in Equity Research: Step-by-Step Team Rollout Guide

20 min readJames Yerkess, Senior Strategic Advisor

Introduction

Implementing AI in equity research isn't a technology challenge, it's a change management challenge. The technology works. The proven time savings are 40% for analysts who adopt systematically. The failure mode isn't tool quality, it's organizational resistance, poor pilot design, and forced adoption without demonstrated value.

This guide provides a practical framework for implementation, from initial one-week pilots through team-wide scaling. Whether you're an individual analyst exploring tools independently, a research director planning team adoption, or a CIO evaluating strategic investment in research automation, this guide creates a structured path from experimentation to systematic automation.

The best approach to AI adoption isn't a six-month implementation project. It's picking one repetitive workflow, trying a tool for one week, measuring time saved, and then expanding systematically. Start small, prove value, scale deliberately.

For workflow-specific implementation details, see Automated Equity Research Workflows. For technology background, see AI Technologies for Equity Research.

The One-Week Pilot Framework

The fastest way to validate automation's value is a one-week pilot on a single, high-frequency workflow. This approach minimizes risk (one week of time), provides quantifiable results (hours saved), and builds conviction for broader adoption.

Quick Start

The One-Week Pilot Framework

Monday: Setup (30 minutes)

  1. Choose one repetitive workflow that you perform weekly or more frequently:

    • Best starter workflows: Earnings call analysis, document search across filings, competitive tracking
    • Avoid for pilots: Complex financial modeling, one-off special situations, workflows requiring IT integration
  2. Select a tool with instant signup:

    • Requirements: Free trial available, no sales call required, productive within 30 minutes
    • Recommendation: Marvin Labs offers self-service signup with all features available immediately
  3. Baseline your current process:

    • Time the manual workflow for 2-3 recent examples
    • Record: Total time, breakdown by sub-task, quality assessment
    • Example: "Earnings call analysis: 3.5 hours (reading transcript: 1.5h, extracting themes: 1h, writing summary: 1h)"

Tuesday-Thursday: Pilot Testing (Real Work)

  1. Perform the workflow using automation for the next 3-5 real examples (not retrospective):

    • Use the tool as you would naturally in your workflow
    • Note friction points, time saved, quality differences
    • Track: Setup time, analysis time, editing time
  2. Compare outcomes:

    • Time: Did it actually save time, or just shift work around?
    • Quality: Did you catch more insights, miss anything, maintain rigor?
    • Experience: Was it easier, more frustrating, neutral?

Friday: Decision (30 minutes)

  1. Calculate ROI:

    • Time saved per instance × frequency per year × hourly cost
    • Example: 2 hours saved × 50 earnings calls/year × $150/hour = $15,000/year value
    • Tool cost: Typically $1,200-6,000/year = 2.5-12x ROI
  2. Decision framework:

    • If time savings ≥ 50% and quality maintained: Adopt permanently
    • If time savings 25-50% or quality improved: Adopt for trial period, expand to other workflows
    • If time savings < 25%: Either tool misfit (try different tool) or workflow not suitable for automation

Example Pilot: Earnings Call Analysis

Monday Setup:

  • Tool: Marvin Labs
  • Workflow: Earnings call analysis for portfolio companies
  • Baseline: 3.5 hours per company (measured across last 3 calls)
  • Test plan: Next 5 earnings calls over the week

Tuesday-Thursday Results:

  • Company A: 45 minutes (74% reduction) - Material Summary + AI Chat for follow-ups
  • Company B: 60 minutes (71% reduction) - More complex call, needed deeper analysis via AI Chat
  • Company C: 50 minutes (76% reduction) - Standard call, summary sufficient
  • Company D: 40 minutes (77% reduction) - Light quarter, minimal changes
  • Company E: 55 minutes (74% reduction) - Multiple segments, AI Chat for segment-specific questions
  • Average: 50 minutes vs. 3.5 hours traditional (76% time savings)

Friday Decision:

  • Time savings: 2.5 hours × 50 calls/year = 125 hours/year
  • Value: 125 hours × $150/hour = $18,750
  • Tool cost: $2,988/year (Marvin Labs Team plan)
  • ROI: 6.3x in year one
  • Quality: Caught guidance change in footnote that analyst missed in manual review
  • Decision: Adopt permanently, expand to 8-K monitoring

Quick Start

Quick Win: Start with Earnings Season

The highest-impact pilot timing is during earnings season, when you're already overwhelmed and time savings are most tangible. If you're reading this in week 1-2 of an earnings season, sign up for a tool TODAY and use it for the next 10 earnings calls. You'll know within 48 hours if it's working.

Assessing Current Workflows

Before adopting automation, understand where your time goes. Not all research tasks benefit equally: document reading and data extraction are perfect automation candidates, while relationship-building and strategic thinking remain human-only.

Conduct a Time Audit

Track your time for one week in 30-minute increments using a spreadsheet or notes file. The goal is pattern recognition, not precision.

Example Time Audit Results:

CategoryHours/Week% of TotalAutomation Potential
Document Analysis1830%⭐⭐⭐ High
Financial Modeling1017%⭐⭐ Medium
Writing1017%⭐⭐ Medium
Meetings1220%❌ Low
Strategic Analysis712%❌ Low
Administrative35%⭐⭐⭐ High
Total60100%52% automatable

Insight: This analyst has 31 hours/week in high and medium automation categories. If automation delivers 60% time savings on those hours, that's 18.6 hours/week freed for strategic work.

Prioritization Matrix

Map tasks by time spent (x-axis) and automation potential (y-axis). Focus on upper-right quadrant for first implementations:

Prioritization Matrix - Getting Started with AI
Prioritization Matrix - Getting Started with AI

Vendor Selection

Tool selection determines adoption success. The wrong tool can fail due to poor workflow fit or excessive complexity. Evaluate systematically without analysis paralysis.

Evaluation Criteria

Critical Criteria:

  1. Time to Value: Productive in under 30 minutes? Self-service signup? (Red flag: "6-week implementation timeline")
  2. Workflow Fit: Automates YOUR specific workflows with clear use case documentation? (Red flag: "We can customize that")
  3. Output Quality: Accurate summaries with source citations? Catches nuances vs. obvious points only?

Important Criteria:

  1. Cost and Pricing: Transparent pricing with monthly subscriptions and free evaluation plans? (Red flag: "contact sales")
  2. Feature Relevance: Handles YOUR document types and workflows with clear examples?

Nice to Have:

  1. Integration: Exports to Excel and Bloomberg or APIs available? (Often overrated, most start standalone)

Three-Phase Evaluation

Phase 1: Desk Research (2 hours)

  • Review websites, pricing, and analyst reviews
  • Shortlist 2-3 tools

Phase 2: Testing (1 week per tool)

  • Sign up for evaluation plans
  • Test with identical real-world scenarios
  • Measure setup time, quality, and time savings

Phase 3: Decision (1 hour)

  • Compare results
  • Calculate ROI
  • Choose best combination of time savings, quality, and cost

Common Vendor Selection Mistakes

  1. Feature lists over workflows: Choose tools that automate your top workflows, not those with the most features
  2. IT-driven selection: Analysts should trial tools and recommend based on usage, not IT requirements alone
  3. Analysis paralysis: Trial 2-3 tools for 2 weeks and decide, not 15 tools over 6 months
  4. Cheapest option: If a tool saves 40% of your time, $500/month is cheap
  5. Integration delays: Start standalone, prove value, integrate later if needed

Team Rollout Strategy

Once you've proven value with a pilot, scaling to multiple workflows and team members requires intentional change management. The goal: replicate your success across the team while avoiding common scaling pitfalls.

Individual Analyst: Expanding Workflows

Expand systematically rather than automating everything simultaneously:

Month 1: Master your pilot workflow until second nature Month 2: Add adjacent workflow (e.g., if you automated earnings calls, add 8-K monitoring) Month 3: Add complex workflow (coverage initiation, competitive analysis) Month 4+: Systematically automate remaining high-ROI workflows from your time audit

Track time saved weekly and monitor quality. If a workflow isn't saving time, diagnose why.

Research Team: Rolling Out Across Analysts

Structured rollout prevents chaos and builds momentum:

Phase 1: Proof of Concept (1 analyst, 1 month)

  • Champion analyst proves 40%+ time savings
  • Presents results to team

Phase 2: Early Adopters (3-5 analysts, 2 months)

  • Volunteers use same tools
  • Weekly check-ins build use case library and best practices

Phase 3: Team Rollout (All analysts, 3 months)

  • Mandate automation for 1-2 workflows
  • Champion-led training
  • Allocate subscriptions for all analysts
  • Target 90%+ adoption

Phase 4: Optimization (Ongoing)

  • Monthly sharing of new workflows
  • Track team-wide time savings and coverage expansion

Change Management Best Practices

  1. Lead with volunteers: Early phases opt-in. Mandate specific workflows only after value is proven.
  2. Share wins: Weekly team meetings for discoveries. Build wiki of use cases and prompts.
  3. Provide support: Dedicated Slack channel, champion office hours, training resources.
  4. Track and communicate ROI: Monthly reports showing hours saved and team-wide impact.

Stop the Hype

Hype: "Roll out AI to the entire team on day one for maximum impact!"

Reality: Forced adoption without proof of value creates resistance. Start with volunteers, prove ROI, then scale. The champion model works: one successful analyst becomes the internal advocate who drives team adoption organically.

Common Scaling Challenges

Challenge 1: "Didn't save time"

  • Solution: Pair with champion for hands-on session. Try 3 times; if still slow, try different tool.

Challenge 2: "Not accurate enough"

  • Solution: AI accelerates, doesn't replace verification. Teach source citation checking.

Challenge 3: "Don't trust AI"

  • Solution: Emphasize verification: AI proposes, analyst verifies. Show examples of AI catching missed details.

Challenge 4: "Too many tools"

  • Solution: Create decision tree: "Earnings calls → Marvin Labs; Data extraction → Daloopa"

Challenge 5: "My workflow is unique"

  • Solution: Break into steps. Identify mechanical parts (automatable) vs. judgment (human-only).

Implementation by Organization Type

While fundamental workflows remain consistent, different organization types face unique constraints and priorities that shape how automation delivers value.

Buy-Side: Hedge Funds and Asset Managers

Buy-side equity research focuses on generating alpha through differentiated insights, superior analysis, and conviction-driven portfolio construction. Automation's value proposition isn't producing more reports, it's finding insights competitors miss and moving faster on market-moving events.

High-Conviction Deep-Dive: Finding Variant Perception

Challenge: Analysts tracking 100 potential investments (80% passive monitoring, 20% active research) can't deeply analyze every company continuously.

AI-Augmented Workflow:

Phase 1: Broad Monitoring with Deep Research Agents

  • Configure agents to monitor 100 companies for thesis-relevant triggers
  • Management tone shifts on key initiatives
  • Changes in competitive positioning rhetoric
  • Emerging risks in footnotes or risk factors
  • Agents operate continuously, alerting only on material developments
  • Analyst Time: 2 hours weekly vs. 10-15 hours manual monitoring

Phase 2: Rapid Deep-Dive When Opportunity Identified

  • Agent flags: "Management tone on international expansion has shifted from confident to cautious across last 3 quarters"
  • AI Chat: "Show me every mention of international expansion challenges, customer metrics, and competitive dynamics over 2 years"
  • Within 30 minutes, analyst has comprehensive view
  • Analyst decides: Worth 40-hour deep-dive with primary research?

Phase 3: Proprietary Primary Research

  • If promising, analyst conducts traditional primary research: expert networks, management meetings, field checks
  • AI handles all secondary document analysis
  • Result: Differentiated view based on primary research + comprehensive secondary foundation

ROI for Buy-Side:

  • Coverage expansion: Monitor 100 companies effectively vs. 40 traditionally
  • Signal detection: Identify emerging thesis changes 1-2 quarters earlier
  • Alpha potential: 1-2 additional high-conviction positions per year

For a hedge fund generating 15% annual returns, adding 2 high-conviction positions (2-3% portfolio weight each at 25% excess returns) contributes 0.3-0.5% to fund returns, worth $3-5M annually on a $1B fund.

Real-World Implementation: Long-Short Equity Hedge Fund

Context: $2B long-short equity fund, 4 sector analysts, 40 high-conviction positions + 60 companies in watchlist.

Challenge: Analysts spending 60% of time maintaining existing coverage, only 40% prospecting new ideas. Missing emerging opportunities in watchlist companies.

Solution: Deploy Marvin Labs Deep Research Agents for continuous watchlist monitoring.

Implementation:

  • Each analyst configures agents for 60 watchlist companies (240 total)
  • Agents monitor earnings, filings, presentations for thesis-relevant developments
  • Weekly digest of material changes requiring attention
  • Deep-dive analysis available on-demand via AI Chat

Results (6 months):

  • Analyst time on maintenance: 60% → 35% (automated monitoring)
  • Analyst time on new idea generation: 40% → 65%
  • New high-conviction positions initiated: 4 (vs. 1-2 typically)
  • One position (3% weight) returned 40% in 6 months: 1.2% contribution to fund returns
  • ROI: $24M additional returns on $2B fund; automation cost: $24K/year (1,000:1 return)

Event-Driven Investing: Speed to Insight Advantage

Traditional Challenge: Market-moving event announced at 7:00 AM. By 9:30 AM market open, analyst needs comprehensive view. Manual analysis requires 6-8 hours, by which time market has priced in consensus view.

AI-Augmented Speed Advantage:

7:00 AM: Deal announced

  • Automated Data Import detects press release
  • Agent processes announcement, extracts key terms
  • Agent generates initial summary (5 minutes)

7:15 AM: Analyst reviews agent summary

  • AI Chat: "Summarize target company's last 5 years: revenue growth, margin trends, competitive position, key risks"
  • AI processes 5 years of 10-Ks, transcripts (2 minutes)
  • Analyst reviews synthesis (15 minutes)

7:45 AM: Deep-dive analysis

  • AI Chat: "What is acquirer's M&A history? Success rate, integration challenges?"
  • AI Chat: "What regulatory concerns might this deal face?"
  • Analyst synthesizes AI insights with judgment (30 minutes)

8:30 AM: Investment decision ready

  • Pro forma valuation model (20 minutes with automated data extraction)
  • Deal probability assessment (human judgment: 15 minutes)

9:00 AM: Positioned ahead of market open

Total time: 2 hours vs. 6-8 hours traditional, enabling entry at better prices before consensus forms.

Sell-Side Research

Sell-side equity research operates under different economics: analysts publish research for external clients rather than internal portfolio managers, and MiFID II has unbundled research payments from trading, creating direct budget pressure.

Broad Coverage Mandate: 60+ Companies Per Analyst

Sell-side analysts typically cover 60-80 companies across a sector, double typical buy-side coverage. During peak earnings season, 60 companies × 3 hours each = 180 hours of work compressed into 3 weeks.

Traditional Result: Analysts triage ruthlessly, deep coverage for 10-15 "core" names, superficial coverage for the rest. Clients notice inconsistent research quality.

AI-Augmented Broad Coverage:

With Material Summaries and Deep Research Agents:

  • Each company: 45 minutes analyst time (vs. 3-4 hours traditional)
  • 50 companies × 45 minutes = 37.5 hours over 2 weeks
  • Remaining time: 60 hours - 37.5 = 22.5 hours for strategic analysis

Coverage Quality Improvement:

  • Traditional: 10-15 names deeply covered, 35-45 names superficially covered
  • AI-Augmented: All 60 names comprehensively covered with consistent quality

Morning Meeting Preparation: Daily Publication Deadline

Sell-side analysts feed the morning meeting machine: by 7:00 AM daily, sales force needs talking points on overnight earnings and market-moving news.

Traditional Workflow (Earnings After-Hours): 5 hours

  • 6:00 PM: Earnings call ends, transcript published
  • 6:00-8:00 PM: Read transcript, extract key points (2 hours)
  • 8:00-9:30 PM: Write comprehensive research note (90 minutes)
  • 9:30-11:00 PM: Prepare morning meeting slides, client email, sales talking points (90 minutes)
  • 11:00 PM: Submit materials

AI-Augmented Workflow: 1.5 hours

  • 6:10 PM: Material Summaries generated automatically
  • 6:10-6:45 PM: Analyst reviews summary, adds strategic perspective (35 minutes)
  • 6:45-7:30 PM: Analyst writes research note using AI first-draft assistance (45 minutes)
  • 7:30-7:45 PM: AI repurposes report into morning meeting slides, client email (15 minutes)
  • 7:45 PM: Submit materials, go home

Quality of Life Impact: Sell-side research faces high attrition rates due to burnout. Automation that reduces peak-season hours from 80+ to 50-60 improves retention, reducing recruiting and training costs.

Client Communication: Responding to 50+ Daily Inquiries

Sell-side analysts serve 100+ institutional clients. Responding to 50+ daily inquiries during earnings season consumes 2-3 hours.

Traditional Challenge:

  • Bloomberg message: "Quick, what did MSFT say about Azure growth vs. last quarter?"
  • Analyst response: Find transcript (2 minutes), search mentions (3 minutes), read context (4 minutes), craft response (3 minutes) = 12 minutes
  • 50 inquiries × 12 minutes = 10 hours daily (impossible)

AI-Augmented Client Service:

  • Bloomberg message: "Quick, what did MSFT say about Azure growth vs. last quarter?"
  • Analyst uses AI Chat: Query takes 30 seconds, returns comprehensive answer
  • Analyst forwards with 1-2 sentences of strategic perspective (1 minute)
  • Total time: 90 seconds vs. 12 minutes (87% reduction)
  • 50 inquiries × 90 seconds = 75 minutes daily (manageable)

Stop the Hype

Hype: "AI replaces sell-side research, clients can query AI directly!"

Reality: Clients pay for differentiated perspectives, judgment, and access to analyst expertise, not just information retrieval. AI handles retrieval so analysts spend more time on judgment. The sell-side analysts who thrive are those who automate mechanics and differentiate on insights.

Wealth Management

Wealth managers and financial advisors must deliver institutional-quality investment guidance to individual clients, translating complex sell-side research into accessible advice for dozens or hundreds of clients.

Client Education: Translating Institutional Research

Wealth management clients often lack financial expertise to interpret sell-side research. Advisors must translate institutional research into accessible language, time-consuming at scale.

AI-Augmented Translation:

Input: Sell-side research note (15 pages, technical language, valuation models)

AI Prompt: "Translate this research for an individual investor. Explain: (1) what the analyst believes will happen, (2) why they believe it, (3) what it means for someone who owns the stock, (4) key risks. Use accessible language, avoid jargon."

AI Output (2 minutes):

  • Plain-English summary of investment thesis
  • Key catalysts explained simply
  • Risk factors in client-friendly terms
  • Actionable implications for portfolio

Advisor Review: 5 minutes to verify accuracy and add personal recommendations

Result: Clients receive institutional insights in accessible format, improving financial literacy and advisor value proposition.

Personalized Client Communications

Creating personalized investment commentary for 50+ clients based on individual portfolios is time-intensive.

Traditional Challenge:

  • Weekly market update: Customize for 50 clients based on their holdings (3-4 hours)
  • Earnings season: Tailor stock-specific updates for affected clients (2 hours per stock)
  • Result: Advisors choose between generic mass emails or unsustainable personalization

AI-Augmented Personalization:

Example Use Cases:

  • "Convert this tech sector update into a one-page summary for a client heavily invested in FAANG stocks, emphasizing portfolio implications"
  • "Rewrite this market commentary for a conservative retiree focused on capital preservation and income"
  • "Create five-bullet updates on Q3 earnings for clients holding MSFT, explaining impact on their portfolios"

Workflow:

  • Upload research report or market commentary
  • Specify client profile, portfolio holdings, communication preferences
  • AI generates personalized version (30-60 seconds)
  • Advisor reviews and adds personal touch (2-3 minutes)
  • Total time: 3-4 minutes per client vs. 15-20 minutes (80% reduction)

Impact: Scale personalization to 100+ clients without overnight work.

Corporate Strategy and Investor Relations

Public company strategy teams and investor relations departments conduct equity research from the opposite perspective: analyzing competitors and understanding how investors perceive them.

Competitive Intelligence

Strategy teams must maintain continuous awareness of competitor positioning, tracking 5-10 competitors across earnings, filings, presentations, news.

Traditional Challenge:

  • 8 competitors × quarterly earnings = 32 earnings events annually
  • Manual analysis: 3 hours per competitor per quarter
  • Total: 96 hours quarterly just on earnings
  • Additional monitoring: 8-Ks, presentations, news (5+ hours weekly)

AI-Augmented Competitive Intelligence:

Quarterly Competitive Summary:

  • AI Chat: "Compare strategic priorities emphasized by all 8 competitors this quarter vs. last quarter. Identify emerging themes."
  • AI generates comprehensive competitive landscape summary (5 minutes)
  • Analyst reviews, adds strategic interpretation (2 hours)

Time Reduction: 8 hours weekly → 1.5 hours weekly (81% reduction)

Investor Perception Analysis

Investor relations teams need to understand how investors perceive their company: Which themes do analysts emphasize? What concerns appear repeatedly?

Traditional Workflow (22 hours):

  • IR team reads sell-side research notes from 15 covering analysts (8 hours)
  • Manually tracks common themes, concerns, valuation approaches (4 hours)
  • Analyzes Q&A patterns from own earnings call (2 hours)
  • Compares to competitor earnings calls (4 hours)
  • Synthesizes findings for management (4 hours)

AI-Augmented Workflow (5 hours):

  • Upload all sell-side research notes to Marvin Labs
  • AI Chat: "Analyze all analyst research. What are the most common investment themes? What concerns appear in 5+ notes?"
  • Sentiment Analysis: Compare management tone to competitor calls
  • AI generates comprehensive perception summary (10 minutes)
  • IR team reviews, prepares management recommendations (4 hours)

Strategic Output:

  • Data-driven messaging recommendations for next quarter
  • Early identification of investor concerns requiring proactive address

Real-World Implementation: Mid-Cap Technology Company

Context: $8B market cap software company, 2-person strategy team, 6 public competitors, quarterly board meetings.

Challenge: Strategy team spending 30 hours weekly on competitive intelligence and investor perception tracking, leaving insufficient time for strategic initiatives.

Solution: Implemented Marvin Labs for automated competitor monitoring and investor analysis.

Results (3 months):

  • Competitive intelligence time: 30 hours weekly → 6 hours weekly
  • Time freed for strategic projects: 24 hours weekly per team member = 48 hours weekly team capacity
  • New strategic initiatives launched: 2 (previously backlogged)
  • Board feedback: Competitive intelligence more comprehensive and timely than previous manual approach
  • ROI: $200K in strategic consulting fees saved (no longer needed external help for competitive analysis)

Training and Ongoing Support

Successful implementation requires ongoing training and support infrastructure, not just initial tool deployment.

Training Program Structure

Week 1: Orientation (2 hours)

  • Platform overview and core capabilities
  • First workflow automation (hands-on)
  • Common use cases demonstration

Week 2-4: Hands-On Practice

  • Analysts automate workflows independently
  • Weekly office hours with champion
  • Slack channel for quick questions

Month 2+: Advanced Training

  • Monthly best practices sessions
  • Advanced workflow tutorials
  • Cross-team knowledge sharing

Building Internal Knowledge Base

Create internal documentation:

  • Workflow playbooks (step-by-step guides)
  • Prompt library (effective queries for common tasks)
  • Troubleshooting guide
  • ROI tracking templates

Measuring Success

Track key metrics:

  • Adoption rate (% of analysts using tools weekly)
  • Time savings (hours per analyst per week)
  • Quality metrics (errors caught, insights generated)
  • Coverage expansion (companies per analyst)
  • User satisfaction (quarterly surveys)

Conclusion

Implementing AI in equity research succeeds when approached as a change management challenge, not a technology deployment. The proven path:

  1. Start with one-week pilots on high-frequency workflows
  2. Prove quantifiable ROI (40%+ time savings)
  3. Scale through champions, not mandates
  4. Expand systematically across workflows and team members
  5. Measure and communicate ongoing value

The organizations achieving 40% research cost reduction share common traits:

  • Started with volunteer pilots, not forced rollouts
  • Led with workflow-focused tools, not feature-rich platforms
  • Championed by analysts, not imposed by IT
  • Measured and communicated ROI continuously
  • Scaled over months, not overnight

For individual analysts, the decision is simpler: try one tool for one week on one workflow. If you don't save 50%+ time, try a different tool or workflow. If you do save time, expand systematically and become your team's champion.

The technology works. The workflows are proven. The remaining challenge is organizational: proving value through disciplined pilots and scaling through demonstrated results.

For comprehensive workflow details, see our guide to Automated Equity Research Workflows. For technology background, see AI Technologies for Equity Research. For complete coverage, see our Complete Guide to Equity Research Automation.

James Yerkess
by James Yerkess

James is a Senior Strategic Advisor to Marvin Labs. He spent 10 years at HSBC, most recently as Global Head of Transaction Banking & FX. He served as an executive member responsible for the launch of two UK neo banks.

Get Started

Experience professional-grade AI for equity research, validate insights for yourself, and see how it fits into your workflow.