mirror of
https://github.com/mendableai/firecrawl.git
synced 2024-11-16 03:32:22 +08:00
8d467c8ca7
* feat: use strictNullChecking * feat: switch logger to Winston * feat(scrapeURL): first batch * fix(scrapeURL): error swallow * fix(scrapeURL): add timeout to EngineResultsTracker * fix(scrapeURL): report unexpected error to sentry * chore: remove unused modules * feat(transfomers/coerce): warn when a format's response is missing * feat(scrapeURL): feature flag priorities, engine quality sorting, PDF and DOCX support * (add note) * feat(scrapeURL): wip readme * feat(scrapeURL): LLM extract * feat(scrapeURL): better warnings * fix(scrapeURL/engines/fire-engine;playwright): fix screenshot * feat(scrapeURL): add forceEngine internal option * feat(scrapeURL/engines): scrapingbee * feat(scrapeURL/transformars): uploadScreenshot * feat(scrapeURL): more intense tests * bunch of stuff * get rid of WebScraper (mostly) * adapt batch scrape * add staging deploy workflow * fix yaml * fix logger issues * fix v1 test schema * feat(scrapeURL/fire-engine/chrome-cdp): remove wait inserts on actions * scrapeURL: v0 backwards compat * logger fixes * feat(scrapeurl): v0 returnOnlyUrls support * fix(scrapeURL/v0): URL leniency * fix(batch-scrape): ts non-nullable * fix(scrapeURL/fire-engine/chromecdp): fix wait action * fix(logger): remove error debug key * feat(requests.http): use dotenv expression * fix(scrapeURL/extractMetadata): extract custom metadata * fix crawl option conversion * feat(scrapeURL): Add retry logic to robustFetch * fix(scrapeURL): crawl stuff * fix(scrapeURL): LLM extract * fix(scrapeURL/v0): search fix * fix(tests/v0): grant larger response size to v0 crawl status * feat(scrapeURL): basic fetch engine * feat(scrapeURL): playwright engine * feat(scrapeURL): add url-specific parameters * Update readme and examples * added e2e tests for most parameters. Still a few actions, location and iframes to be done. * fixed type * Nick: * Update scrape.ts * Update index.ts * added actions and base64 check * Nick: skipTls feature flag? * 403 * todo * todo * fixes * yeet headers from url specific params * add warning when final engine has feature deficit * expose engine results tracker for ScrapeEvents implementation * ingest scrape events * fixed some tests * comment * Update index.test.ts * fixed rawHtml * Update index.test.ts * update comments * move geolocation to global f-e option, fix removeBase64Images * Nick: * trim url-specific params * Update index.ts --------- Co-authored-by: Eric Ciarla <ericciarla@yahoo.com> Co-authored-by: rafaelmmiller <8574157+rafaelmmiller@users.noreply.github.com> Co-authored-by: Nicolas <nicolascamara29@gmail.com>
81 lines
3.1 KiB
Plaintext
81 lines
3.1 KiB
Plaintext
# ===== Required ENVS ======
|
|
NUM_WORKERS_PER_QUEUE=8
|
|
PORT=3002
|
|
HOST=0.0.0.0
|
|
REDIS_URL=redis://redis:6379 #for self-hosting using docker, use redis://redis:6379. For running locally, use redis://localhost:6379
|
|
REDIS_RATE_LIMIT_URL=redis://redis:6379 #for self-hosting using docker, use redis://redis:6379. For running locally, use redis://localhost:6379
|
|
PLAYWRIGHT_MICROSERVICE_URL=http://playwright-service:3000/html
|
|
|
|
## To turn on DB authentication, you need to set up supabase.
|
|
USE_DB_AUTHENTICATION=true
|
|
|
|
# ===== Optional ENVS ======
|
|
|
|
# SearchApi key. Head to https://searchapi.com/ to get your API key
|
|
SEARCHAPI_API_KEY=
|
|
# SearchApi engine, defaults to google. Available options: google, bing, baidu, google_news, etc. Head to https://searchapi.com/ to explore more engines
|
|
SEARCHAPI_ENGINE=
|
|
|
|
# Supabase Setup (used to support DB authentication, advanced logging, etc.)
|
|
SUPABASE_ANON_TOKEN=
|
|
SUPABASE_URL=
|
|
SUPABASE_SERVICE_TOKEN=
|
|
|
|
# Other Optionals
|
|
# use if you've set up authentication and want to test with a real API key
|
|
TEST_API_KEY=
|
|
# set if you'd like to test the scraping rate limit
|
|
RATE_LIMIT_TEST_API_KEY_SCRAPE=
|
|
# set if you'd like to test the crawling rate limit
|
|
RATE_LIMIT_TEST_API_KEY_CRAWL=
|
|
# set if you'd like to use scraping Be to handle JS blocking
|
|
SCRAPING_BEE_API_KEY=
|
|
# add for LLM dependednt features (image alt generation, etc.)
|
|
OPENAI_API_KEY=
|
|
BULL_AUTH_KEY=@
|
|
# set if you have a llamaparse key you'd like to use to parse pdfs
|
|
LLAMAPARSE_API_KEY=
|
|
# set if you'd like to send slack server health status messages
|
|
SLACK_WEBHOOK_URL=
|
|
# set if you'd like to send posthog events like job logs
|
|
POSTHOG_API_KEY=
|
|
# set if you'd like to send posthog events like job logs
|
|
POSTHOG_HOST=
|
|
|
|
STRIPE_PRICE_ID_STANDARD=
|
|
STRIPE_PRICE_ID_SCALE=
|
|
STRIPE_PRICE_ID_STARTER=
|
|
STRIPE_PRICE_ID_HOBBY=
|
|
STRIPE_PRICE_ID_HOBBY_YEARLY=
|
|
STRIPE_PRICE_ID_STANDARD_NEW=
|
|
STRIPE_PRICE_ID_STANDARD_NEW_YEARLY=
|
|
STRIPE_PRICE_ID_GROWTH=
|
|
STRIPE_PRICE_ID_GROWTH_YEARLY=
|
|
|
|
# set if you'd like to use the fire engine closed beta
|
|
FIRE_ENGINE_BETA_URL=
|
|
|
|
# Proxy Settings for Playwright (Alternative you can can use a proxy service like oxylabs, which rotates IPs for you on every request)
|
|
PROXY_SERVER=
|
|
PROXY_USERNAME=
|
|
PROXY_PASSWORD=
|
|
# set if you'd like to block media requests to save proxy bandwidth
|
|
BLOCK_MEDIA=
|
|
|
|
# Set this to the URL of your webhook when using the self-hosted version of FireCrawl
|
|
SELF_HOSTED_WEBHOOK_URL=
|
|
|
|
# Resend API Key for transactional emails
|
|
RESEND_API_KEY=
|
|
|
|
# LOGGING_LEVEL determines the verbosity of logs that the system will output.
|
|
# Available levels are:
|
|
# NONE - No logs will be output.
|
|
# ERROR - For logging error messages that indicate a failure in a specific operation.
|
|
# WARN - For logging potentially harmful situations that are not necessarily errors.
|
|
# INFO - For logging informational messages that highlight the progress of the application.
|
|
# DEBUG - For logging detailed information on the flow through the system, primarily used for debugging.
|
|
# TRACE - For logging more detailed information than the DEBUG level.
|
|
# Set LOGGING_LEVEL to one of the above options to control logging output.
|
|
LOGGING_LEVEL=INFO
|