Crawlr is an open-source CLI AI agent built to streamline the process of ingesting web-based information into structured knowledge bases. Utilizing OpenAI's GPT-3.5/4 models, it traverses specified URLs, cleans and chunks raw HTML into meaningful text segments, generates concise summaries, and creates vector embeddings for efficient semantic search. The tool supports configuration of crawl depth, domain filters, and chunk sizes, allowing users to tailor ingestion pipelines to project needs. By automating link discovery and content processing, Crawlr reduces manual data collection efforts, accelerates creation of FAQ systems, chatbots, and research archives, and seamlessly integrates with vector databases like Pinecone, Weaviate, or local SQLite setups. Its modular design enables easy extension for custom parsers and embedding providers.