xpay tools logo
  • Tools
  • Collections
  • Skills
  • Docs
  • xpay.sh
  1. Collections
  2. /
  3. Competitive Intelligence

Competitive Intelligence

Use Case
Know what your competitors are doing

Stay ahead of the competition with tools for brand analysis, employee tracking, keyword research, and web scraping. Monitor competitor activity across web, social, and search.

SKILL.md
22

tools

6

providers

100+ calls
​
​

22 tools

Keywords Everywhere logo
Keywords Everywhere
Get Credits
get_credits

Get your account's credit balance

$0.01/call

Keywords Everywhere logo
Keywords Everywhere
Get Countries
get_countries

Get list of supported countries

$0.01/call

Keywords Everywhere logo
Keywords Everywhere
Get Currencies
get_currencies

Get list of supported currencies

$0.01/call

Keywords Everywhere logo
Keywords Everywhere
Get Keyword Data
get_keyword_data

Get Volume, CPC and competition for a set of keywords

$0.01/call

Keywords Everywhere logo
Keywords Everywhere
Get Related Keywords
get_related_keywords

Get related keywords based on a seed keyword

$0.01/call

Keywords Everywhere logo
Keywords Everywhere
Get Pasf Keywords
get_pasf_keywords

Get 'People Also Search For' keywords based on a seed keyword

$0.01/call

Keywords Everywhere logo
Keywords Everywhere
Get Domain Keywords
get_domain_keywords

Get keywords that a domain ranks for

$0.01/call

Keywords Everywhere logo
Keywords Everywhere
Get Url Keywords
get_url_keywords

Get keywords that a URL ranks for

$0.01/call

Keywords Everywhere logo
Keywords Everywhere
Get Domain Traffic
get_domain_traffic

Get traffic metrics for a domain

$0.01/call

Keywords Everywhere logo
Keywords Everywhere
Get Url Traffic
get_url_traffic

Get traffic metrics for a URL

$0.01/call

Keywords Everywhere logo
Keywords Everywhere
Get Domain Backlinks
get_domain_backlinks

Get backlinks for a domain

$0.01/call

Keywords Everywhere logo
Keywords Everywhere
Get Unique Domain Backlinks
get_unique_domain_backlinks

Get unique domain backlinks

$0.01/call

Keywords Everywhere logo
Keywords Everywhere
Get Page Backlinks
get_page_backlinks

Get backlinks for a specific URL

$0.01/call

Keywords Everywhere logo
Keywords Everywhere
Get Unique Page Backlinks
get_unique_page_backlinks

Get unique backlinks for a specific URL

$0.01/call

Exa Search logo
Exa Search
Web Search Exa
web_search_exa

Search the web for any topic and get clean, ready-to-use content. Best for: Finding current information, news, facts, people, companies, or answering questions about any topic. Returns: Clean text content from top search results. Query tips: describe the ideal page, not keywords. "blog post comparing React and Vue performance" not "React vs Vue". Use category:people / category:company to search through Linkedin profiles / companies respectively. If highlights are insufficient, follow up with web_fetch_exa on the best URLs.

100+ calls

$0.02/call

Firecrawl logo
Firecrawl
Firecrawl Scrape
firecrawl_scrape

Scrape content from a single URL with advanced options. This is the most powerful, fastest and most reliable scraper tool, if available you should always default to using this tool for any web scraping needs. **Best for:** Single page content extraction, when you know exactly which page contains the information. **Not recommended for:** Multiple pages (call scrape multiple times or use crawl), unknown page location (use search). **Common mistakes:** Using markdown format when extracting specific data points (use JSON instead). **Other Features:** Use 'branding' format to extract brand identity (colors, fonts, typography, spacing, UI components) for design analysis or style replication. **CRITICAL - Format Selection (you MUST follow this):** When the user asks for SPECIFIC data points, you MUST use JSON format with a schema. Only use markdown when the user needs the ENTIRE page content. **Use JSON format when user asks for:** - Parameters, fields, or specifications (e.g., "get the header parameters", "what are the required fields") - Prices, numbers, or structured data (e.g., "extract the pricing", "get the product details") - API details, endpoints, or technical specs (e.g., "find the authentication endpoint") - Lists of items or properties (e.g., "list the features", "get all the options") - Any specific piece of information from a page **Use markdown format ONLY when:** - User wants to read/summarize an entire article or blog post - User needs to see all content on a page without specific extraction - User explicitly asks for the full page content **Handling JavaScript-rendered pages (SPAs):** If JSON extraction returns empty, minimal, or just navigation content, the page is likely JavaScript-rendered or the content is on a different URL. Try these steps IN ORDER: 1. **Add waitFor parameter:** Set `waitFor: 5000` to `waitFor: 10000` to allow JavaScript to render before extraction 2. **Try a different URL:** If the URL has a hash fragment (#section), try the base URL or look for a direct page URL 3. **Use firecrawl_map to find the correct page:** Large documentation sites or SPAs often spread content across multiple URLs. Use `firecrawl_map` with a `search` parameter to discover the specific page containing your target content, then scrape that URL directly. Example: If scraping "https://docs.example.com/reference" fails to find webhook parameters, use `firecrawl_map` with `{"url": "https://docs.example.com/reference", "search": "webhook"}` to find URLs like "/reference/webhook-events", then scrape that specific page. 4. **Use firecrawl_agent:** As a last resort for heavily dynamic pages where map+scrape still fails, use the agent which can autonomously navigate and research **Usage Example (JSON format - REQUIRED for specific data extraction):** ```json { "name": "firecrawl_scrape", "arguments": { "url": "https://example.com/api-docs", "formats": ["json"], "jsonOptions": { "prompt": "Extract the header parameters for the authentication endpoint", "schema": { "type": "object", "properties": { "parameters": { "type": "array", "items": { "type": "object", "properties": { "name": { "type": "string" }, "type": { "type": "string" }, "required": { "type": "boolean" }, "description": { "type": "string" } } } } } } } } } ``` **Prefer markdown format by default.** You can read and reason over the full page content directly — no need for an intermediate query step. Use markdown for questions about page content, factual lookups, and any task where you need to understand the page. **Use JSON format when user needs:** - Structured data with specific fields (extract all products with name, price, description) - Data in a specific schema for downstream processing **Use query format only when:** - The page is extremely long and you need a single targeted answer without processing the full content - You want a quick factual answer and don't need to retain the page content **Usage Example (markdown format - default for most tasks):** ```json { "name": "firecrawl_scrape", "arguments": { "url": "https://example.com/article", "formats": ["markdown"], "onlyMainContent": true } } ``` **Usage Example (branding format - extract brand identity):** ```json { "name": "firecrawl_scrape", "arguments": { "url": "https://example.com", "formats": ["branding"] } } ``` **Branding format:** Extracts comprehensive brand identity (colors, fonts, typography, spacing, logo, UI components) for design analysis or style replication. **Performance:** Add maxAge parameter for 500% faster scrapes using cached data. **Returns:** JSON structured data, markdown, branding profile, or other formats as specified. **Safe Mode:** Read-only content extraction. Interactive actions (click, write, executeJavascript) are disabled for security.

100+ calls

$0.03/call

Firecrawl logo
Firecrawl
Firecrawl Search
firecrawl_search

Search the web and optionally extract content from search results. This is the most powerful web search tool available, and if available you should always default to using this tool for any web search needs. The query also supports search operators, that you can use if needed to refine the search: | Operator | Functionality | Examples | ---|-|-| | `""` | Non-fuzzy matches a string of text | `"Firecrawl"` | `-` | Excludes certain keywords or negates other operators | `-bad`, `-site:firecrawl.dev` | `site:` | Only returns results from a specified website | `site:firecrawl.dev` | `inurl:` | Only returns results that include a word in the URL | `inurl:firecrawl` | `allinurl:` | Only returns results that include multiple words in the URL | `allinurl:git firecrawl` | `intitle:` | Only returns results that include a word in the title of the page | `intitle:Firecrawl` | `allintitle:` | Only returns results that include multiple words in the title of the page | `allintitle:firecrawl playground` | `related:` | Only returns results that are related to a specific domain | `related:firecrawl.dev` | `imagesize:` | Only returns images with exact dimensions | `imagesize:1920x1080` | `larger:` | Only returns images larger than specified dimensions | `larger:1920x1080` **Best for:** Finding specific information across multiple websites, when you don't know which website has the information; when you need the most relevant content for a query. **Not recommended for:** When you need to search the filesystem. When you already know which website to scrape (use scrape); when you need comprehensive coverage of a single website (use map or crawl. **Common mistakes:** Using crawl or map for open-ended questions (use search instead). **Prompt Example:** "Find the latest research papers on AI published in 2023." **Sources:** web, images, news, default to web unless needed images or news. **Scrape Options:** Only use scrapeOptions when you think it is absolutely necessary. When you do so default to a lower limit to avoid timeouts, 5 or lower. **Optimal Workflow:** Search first using firecrawl_search without formats, then after fetching the results, use the scrape tool to get the content of the relevantpage(s) that you want to scrape **Usage Example without formats (Preferred):** ```json { "name": "firecrawl_search", "arguments": { "query": "top AI companies", "limit": 5, "sources": [ { "type": "web" } ] } } ``` **Usage Example with formats:** ```json { "name": "firecrawl_search", "arguments": { "query": "latest AI research papers 2023", "limit": 5, "lang": "en", "country": "us", "sources": [ { "type": "web" }, { "type": "images" }, { "type": "news" } ], "scrapeOptions": { "formats": ["markdown"], "onlyMainContent": true } } } ``` **Returns:** Array of search results (with optional scraped content).

100+ calls

$0.03/call

Firecrawl logo
Firecrawl
Firecrawl Crawl
firecrawl_crawl

Starts a crawl job on a website and extracts content from all pages. **Best for:** Extracting content from multiple related pages, when you need comprehensive coverage. **Not recommended for:** Extracting content from a single page (use scrape); when token limits are a concern (use map + batch_scrape); when you need fast results (crawling can be slow). **Warning:** Crawl responses can be very large and may exceed token limits. Limit the crawl depth and number of pages, or use map + batch_scrape for better control. **Common mistakes:** Setting limit or maxDiscoveryDepth too high (causes token overflow) or too low (causes missing pages); using crawl for a single page (use scrape instead). Using a /* wildcard is not recommended. **Prompt Example:** "Get all blog posts from the first two levels of example.com/blog." **Usage Example:** ```json { "name": "firecrawl_crawl", "arguments": { "url": "https://example.com/blog/*", "maxDiscoveryDepth": 5, "limit": 20, "allowExternalLinks": false, "deduplicateSimilarURLs": true, "sitemap": "include" } } ``` **Returns:** Operation ID for status checking; use firecrawl_check_crawl_status to check progress. **Safe Mode:** Read-only crawling. Webhooks and interactive actions are disabled for security.

100+ calls

$0.05/call

Scrape Creators logo
Scrape Creators
Scrapecreators Linkedin
scrapecreators_linkedin

LinkedIn

100+ calls

$0.04/call

Scrape Creators logo
Scrape Creators
Scrapecreators Company Page
scrapecreators_company_page

Company Page

100+ calls

$0.04/call

Scrape Creators logo
Scrape Creators
Scrapecreators Twitter
scrapecreators_twitter

Twitter

100+ calls

$0.04/call

Scrape Creators logo
Scrape Creators
Scrapecreators User Tweets
scrapecreators_user_tweets

User Tweets

100+ calls

$0.04/call

Connect Competitive Intelligence
​
​

Cursor

Claude Code

Claude Desktop

Windsurf

VS Code

Cline

Roo Code

ChatGPT

Gemini CLI

Amazon Q

Goose

Augment

n8n

API / cURL

AI SDK

TypeScript SDK

Add to ~/.cursor/mcp.json
{
  "mcpServers": {
    "compete": {
      "url": "https://compete.mcp.xpay.sh/mcp?key=YOUR_API_KEY"
    }
  }
}
Collection

22 tools from 6 providers

Pay per tool call. No subscriptions.
xpay tools logo

Marketplace for AI Capabilities. Run agents, tools & prompts with pay-per-use micropayments.

or ask your AI app
Popular Tools
FirecrawlTavilyExaHunterPerplexityRecraftIdeogramFlux
Collections
AI Image & MediaContent & ResearchDev ToolsFinanceLead GenWeb ScrapingCompetitive Intel
Developers
DocumentationAPI ReferenceQuickstartSDKsSkillsskill.mdllms.txt
Resources
The Agentic EconomyAgentic CommerceAgent Frameworksx402 ProtocolBlog
Company
About xpayxpay.shGitHubDiscordTwitter / X

© 2026 Agentically Inc. All rights reserved.
Privacy PolicyTerms of Service
✦
xpay Assistant
Competitive Intelligence Collection

Ask me about tools, pricing, or how to connect

What tools are in this collection?
How much do these tools cost?
Which one is best for getting started?
DocsCollectionsExplore