Skip to main content

What You’ll Learn

This guide shows you how to integrate Firecrawl with Junis using the Firecrawl MCP Server. Your agents will be able to:
  • Scrape single web pages or batch process multiple URLs
  • Crawl entire websites recursively
  • Search the web and extract content from results
  • Extract structured data using JSON schemas
  • Map website structures and discover all pages
Prerequisites:
  • Firecrawl account and API key (Get one here)
  • Admin role in Junis (for organization-level setup)
  • Basic understanding of web scraping concepts

Quick Setup (5 Minutes)

1

Get Firecrawl API Key

  1. Visit Firecrawl Dashboard
  2. Sign up or log in
  3. Click “Create API Key”
  4. Copy your API key (starts with fc-)
2

Add Firecrawl Platform to Junis

Navigate to Team > MCP Skills in Junis and find the Firecrawl card.If Firecrawl is already configured (globe icon 🌍), skip to Step 3.Otherwise, click “Connect” and fill in:
  • Platform Name: Firecrawl
  • MCP Server URL: https://api.firecrawl.dev/mcp/
  • Transport Type: Streamable HTTP
3

Add Your Credentials

Click “Add Auth” on the Firecrawl card and paste your API key.Click “Test Connection” to verify it works.
4

Enable for Your Agents

Go to Admin > Agents, edit the agent you want to connect, and check “Firecrawl” in the MCP Platforms section.Save and test by asking: “Scrape the homepage of https://example.com

Available Tools (8 Tools)

Firecrawl MCP provides 8 powerful web research tools:

Content Extraction

firecrawl_scrape

Single Page ScrapingExtract content from a single URL in markdown or HTML format.Example: “Scrape the content of https://blog.example.com/post

firecrawl_batch_scrape

Batch ScrapingScrape multiple URLs simultaneously.Example: “Scrape these 5 product pages: URL1, URL2, …”

Site Discovery

firecrawl_map

Website MappingDiscover all indexed URLs on a website.Example: “Map all pages on example.com”

firecrawl_search

Web Search + ExtractSearch the web and extract content from top results.Example: “Search for ‘AI trends 2024’ and summarize top 5 results”

Advanced Operations

firecrawl_crawl

Recursive CrawlingCrawl an entire website recursively.Example: “Crawl all product pages on example.com”

firecrawl_extract

Structured Data ExtractionExtract data using JSON schemas for consistent formatting.Example: “Extract price, stock, and reviews from this product page”

Status Monitoring

firecrawl_check_batch_status

Batch StatusCheck progress of batch scraping jobs.

firecrawl_check_crawl_status

Crawl StatusMonitor progress of crawling operations.

Common Workflows

1. Single Page Analysis

Use Case: Analyze a blog post or article Workflow:
1. User: "Analyze the content of this blog post: [URL]"
2. Agent uses firecrawl_scrape to get the content
3. Agent summarizes and analyzes the text
Agent Instruction:
When asked to analyze a web page:
1. Use firecrawl_scrape to get the content
2. Focus on the main content (set onlyMainContent: true)
3. Provide a concise summary and key insights

2. Competitor Research

Use Case: Analyze multiple competitor websites Workflow:
1. User: "Research our top 5 competitors' pricing pages"
2. Agent uses firecrawl_batch_scrape with 5 URLs
3. Agent uses firecrawl_extract to get structured pricing data
4. Agent compares and creates a report
Agent Instruction:
For competitor research:
1. Use batch_scrape for multiple URLs
2. Extract key information like pricing, features, target audience
3. Create a comparison table
4. Highlight our advantages and areas for improvement

3. Content Discovery

Use Case: Find all blog posts on a website Workflow:
1. User: "Find all blog posts on example.com"
2. Agent uses firecrawl_map to discover all URLs
3. Agent filters URLs containing "/blog/"
4. Agent uses batch_scrape to get content
5. Agent categorizes and summarizes posts
Agent Instruction:
To discover content on a website:
1. Use firecrawl_map to get all URLs
2. Filter relevant URLs based on patterns
3. Use batch_scrape for multiple pages
4. Organize content by topic or date

4. Structured Data Collection

Use Case: Collect product information from e-commerce sites Workflow:
1. User: "Extract product details from these pages"
2. Agent uses firecrawl_extract with JSON schema:
   {
     "name": "string",
     "price": "number",
     "stock": "boolean",
     "rating": "number"
   }
3. Agent returns structured data
Agent Instruction:
For structured data extraction:
1. Define a clear JSON schema for the data
2. Use firecrawl_extract with the schema
3. Validate extracted data for completeness
4. Format results in a table or JSON

Tool Parameters Explained

firecrawl_scrape

ParameterTypeDescriptionExample
urlstringRequired. URL to scrapehttps://example.com
formatsarrayOutput format(s)["markdown", "html"]
onlyMainContentbooleanExtract only main contenttrue
includeTagsarrayHTML tags to include["article", "main"]
excludeTagsarrayHTML tags to exclude["nav", "footer"]

firecrawl_batch_scrape

ParameterTypeDescriptionExample
urlsarrayRequired. List of URLs["url1", "url2", "url3"]
optionsobjectScraping optionsSame as firecrawl_scrape
ParameterTypeDescriptionExample
querystringRequired. Search query"AI trends 2024"
limitnumberMax results to return5
langstringLanguage code"en" or "ko"
countrystringCountry code"US" or "KR"

firecrawl_crawl

ParameterTypeDescriptionExample
urlstringRequired. Starting URLhttps://example.com
maxDepthnumberMax crawl depth3
limitnumberMax pages to crawl100
allowExternalLinksbooleanFollow external linksfalse
includePathsarrayURL patterns to include["/blog/*"]
excludePathsarrayURL patterns to exclude["/admin/*"]

firecrawl_extract

ParameterTypeDescriptionExample
urlsarrayRequired. URLs to extract from["url1", "url2"]
promptstringExtraction instructions"Extract product details"
schemaobjectJSON Schema for data{"name": "string", "price": "number"}
enableWebSearchbooleanUse web searchfalse

Example Use Cases

Web Research Agent

Agent Prompt:
You are a web research specialist. When users ask you to research a topic:
1. Use firecrawl_search to find relevant sources
2. Use firecrawl_scrape or batch_scrape to get full content
3. Analyze and synthesize information
4. Provide citations with URLs
5. Summarize key findings in bullet points
Connected MCP: Firecrawl + Notion (for saving research) Example Interaction:
User: "Research the latest developments in quantum computing"
Agent: [Uses firecrawl_search("quantum computing 2024", limit=10)]
       [Uses firecrawl_batch_scrape on top 5 results]
       "Here are the latest developments in quantum computing:
       - IBM announced a 1000+ qubit processor (Source: [URL])
       - Google achieved quantum advantage in optimization problems (Source: [URL])
       - New error correction methods show promise (Source: [URL])
       ..."

Content Monitoring Agent

Agent Prompt:
You monitor competitor websites and news sources for updates.
- Check specific URLs daily
- Compare new content with previous versions
- Highlight changes and new information
- Alert team if significant updates detected
Connected MCP: Firecrawl + Slack (for notifications) Example Interaction:
User: "Check if competitor X has updated their pricing page"
Agent: [Uses firecrawl_scrape on competitor pricing URL]
       [Compares with previously stored data]
       "Changes detected on competitor X pricing page:
       - Basic plan increased from $49 to $59/month
       - New Enterprise tier added at $299/month
       - Free trial extended from 7 to 14 days"

E-commerce Data Collector

Agent Prompt:
You collect product data from e-commerce websites.
- Extract product name, price, availability, ratings
- Use structured extraction (firecrawl_extract)
- Store data in a consistent format
- Track price changes over time
Connected MCP: Firecrawl + PostgreSQL (for storage) Example Interaction:
User: "Collect product data from these 20 URLs"
Agent: [Uses firecrawl_extract with schema:
        {
          "product_name": "string",
          "price": "number",
          "currency": "string",
          "in_stock": "boolean",
          "rating": "number",
          "review_count": "number"
        }]
       "Extracted data from 20 products:
       [Displays formatted table with all product information]"

Troubleshooting

Symptom: Connection fails with authentication errorCause: Invalid or expired Firecrawl API keySolution:
  1. Check your API key at https://www.firecrawl.dev/app/api-keys
  2. Generate a new key if needed
  3. Update credentials in Junis (Team > MCP Skills > Firecrawl)
Symptom: Scraping operations fail after many requestsCause: Firecrawl account has run out of creditsSolution:
  1. Check your usage at https://www.firecrawl.dev/app
  2. Purchase additional credits or upgrade plan
  3. Implement rate limiting in your agent logic
Symptom: firecrawl_crawl or batch_scrape timeoutCause: Large websites or many URLsSolution:
  • Reduce maxDepth or limit parameters
  • Use includePaths to filter specific sections
  • Check status periodically with check_crawl_status
  • Break large jobs into smaller batches
Symptom: firecrawl_extract misses some fieldsCause: JSON schema doesn’t match page structureSolution:
  1. Test with firecrawl_scrape first to see raw content
  2. Refine your JSON schema based on actual page structure
  3. Use more descriptive prompt to guide extraction
  4. Try extracting fewer fields per request
Symptom: Agent connected but Firecrawl tools don’t appearSolution:
  1. Verify connection: Team > MCP Skills > Firecrawl > Test Connection
  2. Check agent configuration: Admin > Agents > [Your Agent] > MCP Platforms
  3. Restart agent (edit and save without changes)
  4. Check logs: Admin > Dashboard > Recent Activity

Performance Tips

✅ Optimization Best Practices:
  • Use onlyMainContent: true to reduce noise and speed up scraping
  • Filter URLs with includePaths and excludePaths before crawling
  • Batch process multiple URLs instead of sequential single scrapes
  • Cache scraped content to avoid redundant API calls
  • Set reasonable limit and maxDepth for crawling operations
  • Use firecrawl_map first to plan your scraping strategy

Rate Limits & Costs

Firecrawl Pricing Tiers

TierCredits/MonthBest For
Free500 creditsTesting and small projects
Starter10,000 creditsRegular scraping tasks
Growth50,000 creditsMedium-scale operations
EnterpriseCustomLarge-scale data collection

Credit Usage

  • Scrape: 1 credit per page
  • Batch Scrape: 1 credit per page
  • Map: 5 credits per site
  • Search: 2 credits per query
  • Crawl: 1 credit per page discovered
  • Extract: 2 credits per page
Monitor your usage at https://www.firecrawl.dev/app to avoid unexpected credit depletion.

Advanced Configuration

Custom Scraping Options

{
  "formats": ["markdown", "html"],
  "onlyMainContent": true,
  "includeTags": ["article", "main", "section"],
  "excludeTags": ["nav", "footer", "aside", "advertisement"],
  "waitFor": 3000,
  "timeout": 30000
}

Crawling Strategy

For large websites, use a phased approach: Phase 1: Discovery
Use firecrawl_map to get all URLs → Filter relevant sections
Phase 2: Selective Crawling
Use firecrawl_crawl with includePaths: ["/products/*", "/blog/*"]
Phase 3: Data Extraction
Use firecrawl_extract on discovered URLs with structured schema

What’s Next?


Additional Resources

Pro Tip: Combine Firecrawl with Notion MCP to automatically save research findings, or with Slack MCP to get real-time alerts when monitored websites change.