Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

🤖 AI-powered web scraping editor with visual workflow builder. Build, test & deploy web scrapers using natural language. Powered by ScrapeGraphAI & LangGraph.

License

NotificationsYou must be signed in to change notification settings

ScrapeGraphAI/scrapecraft

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

ScrapeCraft is a web-based scraping editor similar to Cursor but specialized for web scraping. It uses AI assistance to help users build scraping pipelines with the ScrapeGraphAI API.

scrapecraft.mp4

Features

  • 🤖 AI-powered assistant using OpenRouter (Kimi-k2 model)
  • 🔗 Multi-URL bulk scraping support
  • 📋 Dynamic schema definition with Pydantic
  • 💻 Python code generation with async support
  • 🚀 Real-time WebSocket streaming
  • 📊 Results visualization (table & JSON views)
  • 🔄 Auto-updating deployment with Watchtower

Tech Stack

  • Backend: FastAPI, LangGraph, ScrapeGraphAI
  • Frontend: React, TypeScript, Tailwind CSS
  • Database: PostgreSQL
  • Cache: Redis
  • Deployment: Docker, Docker Compose, Watchtower

Prerequisites

  • Docker and Docker Compose
  • OpenRouter API key (Get it fromOpenRouter)
  • ScrapeGraphAI API key (Get it fromScrapeGraphAI)

Quick Start with Docker

  1. Clone the repository

    git clone https://github.com/ScrapeGraphAI/scrapecraft.gitcd scrapecraft
  2. Set up environment variables

    cp .env.example .env

    Edit the.env file and add your API keys:

  3. Start the application with Docker

    docker compose up -d
  4. Access the application

  5. Stop the application

    docker compose down

Development Mode

If you want to run the application in development mode without Docker:

Backend Development

cd backendpip install -r requirements.txtuvicorn app.main:app --reload --host 0.0.0.0 --port 8000

Frontend Development

cd frontendnpm installnpm start

Usage

  1. Create a Pipeline: Click "New Pipeline" to start
  2. Add URLs: Use the URL Manager to add websites to scrape
  3. Define Schema: Create fields for data extraction
  4. Generate Code: Ask the AI to generate scraping code
  5. Execute: Run the pipeline to scrape data
  6. Export Results: Download as JSON or CSV

Remote Updates

The application includes Watchtower for automatic updates:

  1. Push new Docker images to your registry
  2. Watchtower will automatically detect and update containers
  3. No manual intervention required

API Endpoints

  • POST /api/chat/message - Send message to AI assistant
  • GET /api/pipelines - List all pipelines
  • POST /api/pipelines - Create new pipeline
  • PUT /api/pipelines/{id} - Update pipeline
  • POST /api/pipelines/{id}/run - Execute pipeline
  • WS /ws/{pipeline_id} - WebSocket connection

Environment Variables

VariableDescriptionHow to Get
OPENROUTER_API_KEYYour OpenRouter API keyGet API Key
SCRAPEGRAPH_API_KEYYour ScrapeGraphAI API keyGet API Key
JWT_SECRETSecret key for JWT tokensGenerate a random string
DATABASE_URLPostgreSQL connection stringAuto-configured with Docker
REDIS_URLRedis connection stringAuto-configured with Docker

License

MIT

About

🤖 AI-powered web scraping editor with visual workflow builder. Build, test & deploy web scrapers using natural language. Powered by ScrapeGraphAI & LangGraph.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages


[8]ページ先頭

©2009-2025 Movatter.jp