init: scripts diversos (crawlers, conversores, scrapers)

This commit is contained in:
2026-03-05 20:38:36 +00:00
commit 6ac6f4be2a
925 changed files with 850330 additions and 0 deletions

23
crawl4all/README.md Normal file
View File

@@ -0,0 +1,23 @@
# crawl4all - Async Web Crawler
Crawler assincrono baseado em crawl4ai com profundidade configuravel.
## Setup
```bash
python3 -m venv venv
source venv/bin/activate
pip install -r requirements.txt
```
## Uso
```bash
# Editar config.py com URL alvo
python r1.py
```
## Configuracao
Editar `config.py`:
- `BASE_URL` - URL alvo
- `MAX_DEPTH` - Profundidade maxima (default: 2)
- `MAX_CONCURRENCY` - Workers paralelos (default: 10)
- `OUTPUT_DIR` - Pasta de output (default: "output")