ChatBees
  • 👋Welcome
  • Concepts
    • 📖Architecture
    • 📖Security
    • 📖Serverless RAG
    • 📖Namespace and Collection
    • 📖Access Control
    • 🚀Ticket AI Agent
  • Ticket AI Agent
    • Installation
  • WEB APP REFERENCES
    • 🔑Sign-in and Sign-out
    • ⛓️Manage Connectors
    • 🌏Manage Collections
      • 💿Data Sources
        • 📖Configure Periodic Import
      • ❓Chat with collection
      • ⏱️View Q/A history
      • 📖Publish a Collection
    • 🏛️Manage Users
    • 💰[Flex] Billing and Payment
    • 💰[Enterprise] Billing and Payment
    • 📈Account Usage
    • 🗝️API Keys
    • 🖥️Generated Code Sample
  • ChatBots
    • 🪄AI Chat for Confluence
    • 🪄ChatBees Slack Bot
    • 🪄ChatBees Website ChatBot
    • 🪄Pnyx Discord Bot
  • Snowflake Native App
  • API References
    • 📖API Key
    • 📖Collection Operations
      • 📖Create Collection
      • 📖Configure Collection
      • 📖List Collections
      • 📖Delete Collection
    • 📖Document Operations
      • 📖Upload Document
      • 📖Summarize Document
      • 📖Get Document Outlines and FAQs
      • 📖Ask
      • 📖Chat
      • 📖Search
      • 📖Personalize Response
      • 📖List Documents
      • 📖Delete Document
    • 📖Crawl Operations
      • 📖Create Crawl
      • 📖Get Crawl
      • 📖Index Crawl
      • 📖Delete Crawl
    • 📖Ingest Data Sources
      • 📖Create Ingestion
      • 📖Get Ingestion
      • 📖Index Ingestion
      • 📖Delete Ingestion
Powered by GitBook
On this page
  1. API References
  2. Crawl Operations

Create Crawl

Specify the website root url and the collection for the website content. The service will automatically crawl it. This API returns a crawl_id, that you can use to track the crawl status.

POST /docs/create_crawl HTTP/1.1
Api-Key: my_api_key
Content-Type: application/json
Host: my_account_id.us-west-2.aws.chatbees.ai

{
  "namespace_name": "string",
  "collection_name": "string",
  "root_url": "string"
}

Response:
{
  "crawl_id": "string"
}
import chatbees as cb

# Configure API key
cb.init(api_key="my_api_key", account_id="my_account_id")

col = cb.collection('llm_research')

# create crawl, currently crawl up to 200 pages. 
# a crawl_id is returned, and you could use it to get the crawl status.
crawl_id = col.create_crawl(root_url)
PreviousCrawl OperationsNextGet Crawl

Last updated 10 months ago

📖
📖