mirror of
https://github.com/infiniflow/ragflow.git
synced 2025-12-08 12:32:30 +08:00
## 🚀 Firecrawl Integration for RAGFlow This PR implements the Firecrawl integration for RAGFlow as requested in issue https://github.com/firecrawl/firecrawl/issues/2167 ### ✅ Features Implemented - **Data Source Integration**: Firecrawl appears as a selectable data source in RAGFlow - **Configuration Management**: Users can input Firecrawl API keys through RAGFlow's interface - **Web Scraping**: Supports single URL scraping, website crawling, and batch processing - **Content Processing**: Converts scraped content to RAGFlow's document format with chunking - **Error Handling**: Comprehensive error handling for rate limits, failed requests, and malformed content - **UI Components**: Complete UI schema and workflow components for RAGFlow integration ### 📁 Files Added - `intergrations/firecrawl/` - Complete integration package - `intergrations/firecrawl/integration.py` - RAGFlow integration entry point - `intergrations/firecrawl/firecrawl_connector.py` - API communication - `intergrations/firecrawl/firecrawl_config.py` - Configuration management - `intergrations/firecrawl/firecrawl_processor.py` - Content processing - `intergrations/firecrawl/firecrawl_ui.py` - UI components - `intergrations/firecrawl/ragflow_integration.py` - Main integration class - `intergrations/firecrawl/README.md` - Complete documentation - `intergrations/firecrawl/example_usage.py` - Usage examples ### 🧪 Testing The integration has been thoroughly tested with: - Configuration validation - Connection testing - Content processing and chunking - UI component rendering - Error handling scenarios ### 📋 Acceptance Criteria Met - ✅ Integration appears as selectable data source in RAGFlow's data source options - ✅ Users can input Firecrawl API keys through RAGFlow's configuration interface - ✅ Successfully scrapes content from provided URLs and imports into RAGFlow's document store - ✅ Handles common edge cases (rate limits, failed requests, malformed content) - ✅ Includes basic documentation and README updates - ✅ Code follows RAGFlow's existing patterns and coding standards ### �� Related Issue https://github.com/firecrawl/firecrawl/issues/2167 --------- Co-authored-by: AB <aj@Ajays-MacBook-Air.local>
262 lines
7.5 KiB
Python
262 lines
7.5 KiB
Python
"""
|
|
Example usage of the Firecrawl integration with RAGFlow.
|
|
"""
|
|
|
|
import asyncio
|
|
import logging
|
|
|
|
from .ragflow_integration import RAGFlowFirecrawlIntegration, create_firecrawl_integration
|
|
from .firecrawl_config import FirecrawlConfig
|
|
|
|
|
|
async def example_single_url_scraping():
|
|
"""Example of scraping a single URL."""
|
|
print("=== Single URL Scraping Example ===")
|
|
|
|
# Configuration
|
|
config = {
|
|
"api_key": "fc-your-api-key-here", # Replace with your actual API key
|
|
"api_url": "https://api.firecrawl.dev",
|
|
"max_retries": 3,
|
|
"timeout": 30,
|
|
"rate_limit_delay": 1.0
|
|
}
|
|
|
|
# Create integration
|
|
integration = create_firecrawl_integration(config)
|
|
|
|
# Test connection
|
|
connection_test = await integration.test_connection()
|
|
print(f"Connection test: {connection_test}")
|
|
|
|
if not connection_test["success"]:
|
|
print("Connection failed, please check your API key")
|
|
return
|
|
|
|
# Scrape a single URL
|
|
urls = ["https://httpbin.org/json"]
|
|
documents = await integration.scrape_and_import(urls)
|
|
|
|
for doc in documents:
|
|
print(f"Title: {doc.title}")
|
|
print(f"URL: {doc.source_url}")
|
|
print(f"Content length: {len(doc.content)}")
|
|
print(f"Language: {doc.language}")
|
|
print(f"Metadata: {doc.metadata}")
|
|
print("-" * 50)
|
|
|
|
|
|
async def example_website_crawling():
|
|
"""Example of crawling an entire website."""
|
|
print("=== Website Crawling Example ===")
|
|
|
|
# Configuration
|
|
config = {
|
|
"api_key": "fc-your-api-key-here", # Replace with your actual API key
|
|
"api_url": "https://api.firecrawl.dev",
|
|
"max_retries": 3,
|
|
"timeout": 30,
|
|
"rate_limit_delay": 1.0
|
|
}
|
|
|
|
# Create integration
|
|
integration = create_firecrawl_integration(config)
|
|
|
|
# Crawl a website
|
|
start_url = "https://httpbin.org"
|
|
documents = await integration.crawl_and_import(
|
|
start_url=start_url,
|
|
limit=5, # Limit to 5 pages for demo
|
|
scrape_options={
|
|
"formats": ["markdown", "html"],
|
|
"extractOptions": {
|
|
"extractMainContent": True,
|
|
"excludeTags": ["nav", "footer", "header"]
|
|
}
|
|
}
|
|
)
|
|
|
|
print(f"Crawled {len(documents)} pages from {start_url}")
|
|
|
|
for i, doc in enumerate(documents):
|
|
print(f"Page {i+1}: {doc.title}")
|
|
print(f"URL: {doc.source_url}")
|
|
print(f"Content length: {len(doc.content)}")
|
|
print("-" * 30)
|
|
|
|
|
|
async def example_batch_processing():
|
|
"""Example of batch processing multiple URLs."""
|
|
print("=== Batch Processing Example ===")
|
|
|
|
# Configuration
|
|
config = {
|
|
"api_key": "fc-your-api-key-here", # Replace with your actual API key
|
|
"api_url": "https://api.firecrawl.dev",
|
|
"max_retries": 3,
|
|
"timeout": 30,
|
|
"rate_limit_delay": 1.0
|
|
}
|
|
|
|
# Create integration
|
|
integration = create_firecrawl_integration(config)
|
|
|
|
# Batch scrape multiple URLs
|
|
urls = [
|
|
"https://httpbin.org/json",
|
|
"https://httpbin.org/html",
|
|
"https://httpbin.org/xml"
|
|
]
|
|
|
|
documents = await integration.scrape_and_import(
|
|
urls=urls,
|
|
formats=["markdown", "html"],
|
|
extract_options={
|
|
"extractMainContent": True,
|
|
"excludeTags": ["nav", "footer", "header"]
|
|
}
|
|
)
|
|
|
|
print(f"Processed {len(documents)} URLs")
|
|
|
|
for doc in documents:
|
|
print(f"Title: {doc.title}")
|
|
print(f"URL: {doc.source_url}")
|
|
print(f"Content length: {len(doc.content)}")
|
|
|
|
# Example of chunking for RAG processing
|
|
chunks = integration.processor.chunk_content(doc, chunk_size=500, chunk_overlap=100)
|
|
print(f"Number of chunks: {len(chunks)}")
|
|
print("-" * 30)
|
|
|
|
|
|
async def example_content_processing():
|
|
"""Example of content processing and chunking."""
|
|
print("=== Content Processing Example ===")
|
|
|
|
# Configuration
|
|
config = {
|
|
"api_key": "fc-your-api-key-here", # Replace with your actual API key
|
|
"api_url": "https://api.firecrawl.dev",
|
|
"max_retries": 3,
|
|
"timeout": 30,
|
|
"rate_limit_delay": 1.0
|
|
}
|
|
|
|
# Create integration
|
|
integration = create_firecrawl_integration(config)
|
|
|
|
# Scrape content
|
|
urls = ["https://httpbin.org/html"]
|
|
documents = await integration.scrape_and_import(urls)
|
|
|
|
for doc in documents:
|
|
print(f"Original document: {doc.title}")
|
|
print(f"Content length: {len(doc.content)}")
|
|
|
|
# Chunk the content
|
|
chunks = integration.processor.chunk_content(
|
|
doc,
|
|
chunk_size=1000,
|
|
chunk_overlap=200
|
|
)
|
|
|
|
print(f"Number of chunks: {len(chunks)}")
|
|
|
|
for i, chunk in enumerate(chunks):
|
|
print(f"Chunk {i+1}:")
|
|
print(f" ID: {chunk['id']}")
|
|
print(f" Content length: {len(chunk['content'])}")
|
|
print(f" Metadata: {chunk['metadata']}")
|
|
print()
|
|
|
|
|
|
async def example_error_handling():
|
|
"""Example of error handling."""
|
|
print("=== Error Handling Example ===")
|
|
|
|
# Configuration with invalid API key
|
|
config = {
|
|
"api_key": "invalid-key",
|
|
"api_url": "https://api.firecrawl.dev",
|
|
"max_retries": 3,
|
|
"timeout": 30,
|
|
"rate_limit_delay": 1.0
|
|
}
|
|
|
|
# Create integration
|
|
integration = create_firecrawl_integration(config)
|
|
|
|
# Test connection (should fail)
|
|
connection_test = await integration.test_connection()
|
|
print(f"Connection test with invalid key: {connection_test}")
|
|
|
|
# Try to scrape (should fail gracefully)
|
|
try:
|
|
urls = ["https://httpbin.org/json"]
|
|
documents = await integration.scrape_and_import(urls)
|
|
print(f"Documents scraped: {len(documents)}")
|
|
except Exception as e:
|
|
print(f"Error occurred: {e}")
|
|
|
|
|
|
async def example_configuration_validation():
|
|
"""Example of configuration validation."""
|
|
print("=== Configuration Validation Example ===")
|
|
|
|
# Test various configurations
|
|
test_configs = [
|
|
{
|
|
"api_key": "fc-valid-key",
|
|
"api_url": "https://api.firecrawl.dev",
|
|
"max_retries": 3,
|
|
"timeout": 30,
|
|
"rate_limit_delay": 1.0
|
|
},
|
|
{
|
|
"api_key": "invalid-key", # Invalid format
|
|
"api_url": "https://api.firecrawl.dev"
|
|
},
|
|
{
|
|
"api_key": "fc-valid-key",
|
|
"api_url": "invalid-url", # Invalid URL
|
|
"max_retries": 15, # Too high
|
|
"timeout": 500, # Too high
|
|
"rate_limit_delay": 15.0 # Too high
|
|
}
|
|
]
|
|
|
|
for i, config in enumerate(test_configs):
|
|
print(f"Test configuration {i+1}:")
|
|
errors = RAGFlowFirecrawlIntegration(FirecrawlConfig.from_dict(config)).validate_config(config)
|
|
|
|
if errors:
|
|
print(" Errors found:")
|
|
for field, error in errors.items():
|
|
print(f" {field}: {error}")
|
|
else:
|
|
print(" Configuration is valid")
|
|
print()
|
|
|
|
|
|
async def main():
|
|
"""Run all examples."""
|
|
# Set up logging
|
|
logging.basicConfig(level=logging.INFO)
|
|
|
|
print("Firecrawl RAGFlow Integration Examples")
|
|
print("=" * 50)
|
|
|
|
# Run examples
|
|
await example_configuration_validation()
|
|
await example_single_url_scraping()
|
|
await example_batch_processing()
|
|
await example_content_processing()
|
|
await example_error_handling()
|
|
|
|
print("Examples completed!")
|
|
|
|
|
|
if __name__ == "__main__":
|
|
asyncio.run(main())
|