diff --git a/.github/archive/js-sdk.yml b/.github/archive/js-sdk.yml index c84bb8b1..7ef096d4 100644 --- a/.github/archive/js-sdk.yml +++ b/.github/archive/js-sdk.yml @@ -8,7 +8,6 @@ env: FLY_API_TOKEN: ${{ secrets.FLY_API_TOKEN }} HOST: ${{ secrets.HOST }} LLAMAPARSE_API_KEY: ${{ secrets.LLAMAPARSE_API_KEY }} - LOGTAIL_KEY: ${{ secrets.LOGTAIL_KEY }} POSTHOG_API_KEY: ${{ secrets.POSTHOG_API_KEY }} POSTHOG_HOST: ${{ secrets.POSTHOG_HOST }} NUM_WORKERS_PER_QUEUE: ${{ secrets.NUM_WORKERS_PER_QUEUE }} @@ -21,7 +20,6 @@ env: SUPABASE_SERVICE_TOKEN: ${{ secrets.SUPABASE_SERVICE_TOKEN }} SUPABASE_URL: ${{ secrets.SUPABASE_URL }} TEST_API_KEY: ${{ secrets.TEST_API_KEY }} - HYPERDX_API_KEY: ${{ secrets.HYPERDX_API_KEY }} HDX_NODE_BETA_MODE: 1 jobs: diff --git a/.github/archive/python-sdk.yml b/.github/archive/python-sdk.yml index 27449888..bdefeab6 100644 --- a/.github/archive/python-sdk.yml +++ b/.github/archive/python-sdk.yml @@ -8,7 +8,6 @@ env: FLY_API_TOKEN: ${{ secrets.FLY_API_TOKEN }} HOST: ${{ secrets.HOST }} LLAMAPARSE_API_KEY: ${{ secrets.LLAMAPARSE_API_KEY }} - LOGTAIL_KEY: ${{ secrets.LOGTAIL_KEY }} POSTHOG_API_KEY: ${{ secrets.POSTHOG_API_KEY }} POSTHOG_HOST: ${{ secrets.POSTHOG_HOST }} NUM_WORKERS_PER_QUEUE: ${{ secrets.NUM_WORKERS_PER_QUEUE }} @@ -21,7 +20,6 @@ env: SUPABASE_SERVICE_TOKEN: ${{ secrets.SUPABASE_SERVICE_TOKEN }} SUPABASE_URL: ${{ secrets.SUPABASE_URL }} TEST_API_KEY: ${{ secrets.TEST_API_KEY }} - HYPERDX_API_KEY: ${{ secrets.HYPERDX_API_KEY }} HDX_NODE_BETA_MODE: 1 jobs: diff --git a/.github/archive/rust-sdk.yml b/.github/archive/rust-sdk.yml index 62deeaab..792e06c2 100644 --- a/.github/archive/rust-sdk.yml +++ b/.github/archive/rust-sdk.yml @@ -8,7 +8,6 @@ env: FLY_API_TOKEN: ${{ secrets.FLY_API_TOKEN }} HOST: ${{ secrets.HOST }} LLAMAPARSE_API_KEY: ${{ secrets.LLAMAPARSE_API_KEY }} - LOGTAIL_KEY: ${{ secrets.LOGTAIL_KEY }} POSTHOG_API_KEY: ${{ secrets.POSTHOG_API_KEY }} POSTHOG_HOST: ${{ secrets.POSTHOG_HOST }} NUM_WORKERS_PER_QUEUE: ${{ secrets.NUM_WORKERS_PER_QUEUE }} @@ -21,7 +20,6 @@ env: SUPABASE_SERVICE_TOKEN: ${{ secrets.SUPABASE_SERVICE_TOKEN }} SUPABASE_URL: ${{ secrets.SUPABASE_URL }} TEST_API_KEY: ${{ secrets.TEST_API_KEY }} - HYPERDX_API_KEY: ${{ secrets.HYPERDX_API_KEY }} HDX_NODE_BETA_MODE: 1 diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index 8a9a74cc..ef7d1cba 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -12,7 +12,6 @@ env: FLY_API_TOKEN: ${{ secrets.FLY_API_TOKEN }} HOST: ${{ secrets.HOST }} LLAMAPARSE_API_KEY: ${{ secrets.LLAMAPARSE_API_KEY }} - LOGTAIL_KEY: ${{ secrets.LOGTAIL_KEY }} POSTHOG_API_KEY: ${{ secrets.POSTHOG_API_KEY }} POSTHOG_HOST: ${{ secrets.POSTHOG_HOST }} NUM_WORKERS_PER_QUEUE: ${{ secrets.NUM_WORKERS_PER_QUEUE }} @@ -25,7 +24,6 @@ env: SUPABASE_SERVICE_TOKEN: ${{ secrets.SUPABASE_SERVICE_TOKEN }} SUPABASE_URL: ${{ secrets.SUPABASE_URL }} TEST_API_KEY: ${{ secrets.TEST_API_KEY }} - HYPERDX_API_KEY: ${{ secrets.HYPERDX_API_KEY }} HDX_NODE_BETA_MODE: 1 FIRE_ENGINE_BETA_URL: ${{ secrets.FIRE_ENGINE_BETA_URL }} USE_DB_AUTHENTICATION: ${{ secrets.USE_DB_AUTHENTICATION }} diff --git a/.github/workflows/deploy-image-staging.yml b/.github/workflows/deploy-image-staging.yml new file mode 100644 index 00000000..e74aba9a --- /dev/null +++ b/.github/workflows/deploy-image-staging.yml @@ -0,0 +1,32 @@ +name: STAGING Deploy Images to GHCR + +env: + DOTNET_VERSION: '6.0.x' + +on: + push: + branches: + - mog/webscraper-refactor + workflow_dispatch: + +jobs: + push-app-image: + runs-on: ubuntu-latest + defaults: + run: + working-directory: './apps/api' + steps: + - name: 'Checkout GitHub Action' + uses: actions/checkout@main + + - name: 'Login to GitHub Container Registry' + uses: docker/login-action@v1 + with: + registry: ghcr.io + username: ${{github.actor}} + password: ${{secrets.GITHUB_TOKEN}} + + - name: 'Build Inventory Image' + run: | + docker build . --tag ghcr.io/mendableai/firecrawl-staging:latest + docker push ghcr.io/mendableai/firecrawl-staging:latest \ No newline at end of file diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index 73ccf0e6..b8c1f0a5 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -41,7 +41,6 @@ TEST_API_KEY= # use if you've set up authentication and want to test with a real SCRAPING_BEE_API_KEY= #Set if you'd like to use scraping Be to handle JS blocking OPENAI_API_KEY= # add for LLM dependednt features (image alt generation, etc.) BULL_AUTH_KEY= @ -LOGTAIL_KEY= # Use if you're configuring basic logging with logtail PLAYWRIGHT_MICROSERVICE_URL= # set if you'd like to run a playwright fallback LLAMAPARSE_API_KEY= #Set if you have a llamaparse key you'd like to use to parse pdfs SLACK_WEBHOOK_URL= # set if you'd like to send slack server health status messages diff --git a/SELF_HOST.md b/SELF_HOST.md index 78228485..46e08db9 100644 --- a/SELF_HOST.md +++ b/SELF_HOST.md @@ -62,7 +62,6 @@ TEST_API_KEY= # use if you've set up authentication and want to test with a real SCRAPING_BEE_API_KEY= # use if you'd like to use as a fallback scraper OPENAI_API_KEY= # add for LLM-dependent features (e.g., image alt generation) BULL_AUTH_KEY= @ -LOGTAIL_KEY= # Use if you're configuring basic logging with logtail PLAYWRIGHT_MICROSERVICE_URL= # set if you'd like to run a playwright fallback LLAMAPARSE_API_KEY= #Set if you have a llamaparse key you'd like to use to parse pdfs SLACK_WEBHOOK_URL= # set if you'd like to send slack server health status messages diff --git a/apps/api/.env.example b/apps/api/.env.example index 6ba49daa..d54c696d 100644 --- a/apps/api/.env.example +++ b/apps/api/.env.example @@ -33,8 +33,6 @@ SCRAPING_BEE_API_KEY= # add for LLM dependednt features (image alt generation, etc.) OPENAI_API_KEY= BULL_AUTH_KEY=@ -# use if you're configuring basic logging with logtail -LOGTAIL_KEY= # set if you have a llamaparse key you'd like to use to parse pdfs LLAMAPARSE_API_KEY= # set if you'd like to send slack server health status messages @@ -54,9 +52,6 @@ STRIPE_PRICE_ID_STANDARD_NEW_YEARLY= STRIPE_PRICE_ID_GROWTH= STRIPE_PRICE_ID_GROWTH_YEARLY= -HYPERDX_API_KEY= -HDX_NODE_BETA_MODE=1 - # set if you'd like to use the fire engine closed beta FIRE_ENGINE_BETA_URL= diff --git a/apps/api/jest.setup.js b/apps/api/jest.setup.js index c158ca42..0b3b09b7 100644 --- a/apps/api/jest.setup.js +++ b/apps/api/jest.setup.js @@ -1 +1 @@ -global.fetch = require('jest-fetch-mock'); +// global.fetch = require('jest-fetch-mock'); diff --git a/apps/api/package.json b/apps/api/package.json index a0f9cf8e..bb4ea268 100644 --- a/apps/api/package.json +++ b/apps/api/package.json @@ -32,9 +32,11 @@ "@tsconfig/recommended": "^1.0.3", "@types/body-parser": "^1.19.2", "@types/cors": "^2.8.13", + "@types/escape-html": "^1.0.4", "@types/express": "^4.17.17", "@types/jest": "^29.5.12", "@types/node": "^20.14.1", + "@types/pdf-parse": "^1.1.4", "body-parser": "^1.20.1", "express": "^4.18.2", "jest": "^29.6.3", @@ -53,9 +55,7 @@ "@bull-board/api": "^5.20.5", "@bull-board/express": "^5.20.5", "@devil7softwares/pos": "^1.0.2", - "@dqbd/tiktoken": "^1.0.13", - "@hyperdx/node-opentelemetry": "^0.8.1", - "@logtail/node": "^0.4.12", + "@dqbd/tiktoken": "^1.0.16", "@nangohq/node": "^0.40.8", "@sentry/cli": "^2.33.1", "@sentry/node": "^8.26.0", @@ -78,6 +78,7 @@ "date-fns": "^3.6.0", "dotenv": "^16.3.1", "dotenv-cli": "^7.4.2", + "escape-html": "^1.0.3", "express-rate-limit": "^7.3.1", "express-ws": "^5.0.2", "form-data": "^4.0.0", @@ -92,6 +93,7 @@ "languagedetect": "^2.0.0", "logsnag": "^1.0.0", "luxon": "^3.4.3", + "marked": "^14.1.2", "md5": "^2.3.0", "moment": "^2.29.4", "mongoose": "^8.4.4", @@ -114,6 +116,8 @@ "typesense": "^1.5.4", "unstructured-client": "^0.11.3", "uuid": "^10.0.0", + "winston": "^3.14.2", + "winston-transport": "^4.8.0", "wordpos": "^2.1.0", "ws": "^8.18.0", "xml2js": "^0.6.2", diff --git a/apps/api/pnpm-lock.yaml b/apps/api/pnpm-lock.yaml index 095b507c..3350c74e 100644 --- a/apps/api/pnpm-lock.yaml +++ b/apps/api/pnpm-lock.yaml @@ -24,14 +24,8 @@ importers: specifier: ^1.0.2 version: 1.0.2 '@dqbd/tiktoken': - specifier: ^1.0.13 - version: 1.0.15 - '@hyperdx/node-opentelemetry': - specifier: ^0.8.1 - version: 0.8.1 - '@logtail/node': - specifier: ^0.4.12 - version: 0.4.21 + specifier: ^1.0.16 + version: 1.0.16 '@nangohq/node': specifier: ^0.40.8 version: 0.40.8 @@ -98,6 +92,9 @@ importers: dotenv-cli: specifier: ^7.4.2 version: 7.4.2 + escape-html: + specifier: ^1.0.3 + version: 1.0.3 express-rate-limit: specifier: ^7.3.1 version: 7.3.1(express@4.19.2) @@ -140,6 +137,9 @@ importers: luxon: specifier: ^3.4.3 version: 3.4.4 + marked: + specifier: ^14.1.2 + version: 14.1.2 md5: specifier: ^2.3.0 version: 2.3.0 @@ -206,6 +206,12 @@ importers: uuid: specifier: ^10.0.0 version: 10.0.0 + winston: + specifier: ^3.14.2 + version: 3.14.2 + winston-transport: + specifier: ^4.8.0 + version: 4.8.0 wordpos: specifier: ^2.1.0 version: 2.1.0 @@ -237,6 +243,9 @@ importers: '@types/cors': specifier: ^2.8.13 version: 2.8.17 + '@types/escape-html': + specifier: ^1.0.4 + version: 1.0.4 '@types/express': specifier: ^4.17.17 version: 4.17.21 @@ -246,6 +255,9 @@ importers: '@types/node': specifier: ^20.14.1 version: 20.14.1 + '@types/pdf-parse': + specifier: ^1.1.4 + version: 1.1.4 body-parser: specifier: ^1.20.1 version: 1.20.2 @@ -494,42 +506,22 @@ packages: resolution: {integrity: sha512-IchNf6dN4tHoMFIn/7OE8LWZ19Y6q/67Bmf6vnGREv8RSbBVb9LPJxEcnwrcwX6ixSvaiGoomAUvu4YSxXrVgw==} engines: {node: '>=12'} + '@dabh/diagnostics@2.0.3': + resolution: {integrity: sha512-hrlQOIi7hAfzsMqlGSFyVucrx38O+j6wiGOf//H2ecvIEqYN4ADBSS2iLMh5UFyDunCNniUIPk/q3riFv45xRA==} + '@devil7softwares/pos@1.0.2': resolution: {integrity: sha512-49Ke26+++Ix8C5LChi4uga7aWgMuc5zV1NqjGxXxE7DralZwe+hvUuSJmBDWS+HHZaK9rFzLNdufV4HAvvOxPA==} engines: {node: '>=0'} deprecated: This package has been renamed to `fast-tag-pos` - '@dqbd/tiktoken@1.0.15': - resolution: {integrity: sha512-a6I67K1xUkuqcuwulobIJiLikkoE7egMaviI1Jg5bxSn2V7QGqXsGE3jTKr8UIOU/o74mAAd5TkeXFNBtaKF4A==} + '@dqbd/tiktoken@1.0.16': + resolution: {integrity: sha512-4uIrs5qxAwFVFFEP507HZIZhGOsgfaEMEWDXWalr+v+XP+wJwP60EVmkZtQyQe70IsKGVkx5umBxw4NfmU0pPg==} '@flydotio/dockerfile@0.4.11': resolution: {integrity: sha512-L52UAfrOhmAn3T4TxpeRofQOSO+Kctg+uraB4nLzo4mvvh+4Z7HYxSi7Dnq0Kirz+xx6fDIc4OMNT1EdaORecA==} engines: {node: '>=16.0.0'} hasBin: true - '@grpc/grpc-js@1.10.10': - resolution: {integrity: sha512-HPa/K5NX6ahMoeBv15njAc/sfF4/jmiXLar9UlC2UfHFKZzsCVLc3wbe7+7qua7w9VPh2/L6EBxyAV7/E8Wftg==} - engines: {node: '>=12.10.0'} - - '@grpc/proto-loader@0.7.13': - resolution: {integrity: sha512-AiXO/bfe9bmxBjxxtYxFAXGZvMaN5s8kO+jBHAJCON8rJoB5YS/D6X7ZNc6XQkuHNmyl4CYaMI1fJ/Gn27RGGw==} - engines: {node: '>=6'} - hasBin: true - - '@hyperdx/instrumentation-exception@0.1.0': - resolution: {integrity: sha512-Jgk7JY5J07Mq9fgXApGVhSkS4+WdzzRcWLReAZhxgo46KShxE6w614mFqUSnuo+z6ghlehsy4ForViUfxrFyew==} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@hyperdx/instrumentation-sentry-node@0.1.0': - resolution: {integrity: sha512-n8d/K/8M2owL2w4FNfV+lSVW6yoznEj5SdRCysV/ZIfyrZwpijiiSn7gkRcrOfKHmrxrupyp7DVg5L19cGuH6A==} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@hyperdx/node-opentelemetry@0.8.1': - resolution: {integrity: sha512-wNw0yQf54j/9KXVWeEOu8G6C5FT5EFlrz4dcmscTkwCvo6fQOLRZa/NbGcqugt0LSFMc0/6/Q5RDWVqDpEn0LQ==} - hasBin: true - '@ioredis/commands@1.2.0': resolution: {integrity: sha512-Sx1pU8EM64o2BrqNpEO1CNLtKQwyhuXuqyfH7oGKCk+1a33d2r5saW8zNwm3j6BTExtjrv2BxTgzzkMwts6vGg==} @@ -636,9 +628,6 @@ packages: '@jridgewell/trace-mapping@0.3.9': resolution: {integrity: sha512-3Belt6tdc8bPgAtbcmdtNJlirVoTmEb5e2gC94PnkwEW9jI6CAHUeoG85tjWP5WquqfavoMtMwiG4P926ZKKuQ==} - '@js-sdsl/ordered-map@4.4.2': - resolution: {integrity: sha512-iUKgm52T8HOE/makSxjqoWhe95ZJA1/G1sYsGev2JDKUSS14KAgg1LHb+Ba+IPow0xflbnSkOsZcO08C7w1gYw==} - '@langchain/core@0.2.12': resolution: {integrity: sha512-zaKvUcWU1Cxcpd/fxklygY6iUrxls10KTRzyHZGBAIKJq1JD/B10vX59YlFgBs7nqqVTEvaChfIE0O0e2qBttA==} engines: {node: '>=18'} @@ -651,28 +640,12 @@ packages: resolution: {integrity: sha512-cXWgKE3sdWLSqAa8ykbCcUsUF1Kyr5J3HOWYGuobhPEycXW4WI++d5DhzdpL238mzoEXTi90VqfSCra37l5YqA==} engines: {node: '>=18'} - '@logtail/core@0.4.21': - resolution: {integrity: sha512-QDq194+24bwi4e+a/pxyf4X67NewhTvBmh9iwM2NhbSVSQz4Fo8xQn1Ul8zuUrXETycu/Od2D8wT2tZFNFx/7A==} - - '@logtail/node@0.4.21': - resolution: {integrity: sha512-zpwkhJgcYaM+vsjotHRJthc0ot1vP0CAVy+fwrkL8XjfdC3NHiWb6f0agQpHlqdRX8RTsAbcYpWNXKPpFB5U9Q==} - - '@logtail/tools@0.4.21': - resolution: {integrity: sha512-xIaolScUwJEikllopGphxBX0lVlN/rA8pLAZiNCMNJXpPbwitoFKLW3w4qRuYdKoFCCJZKwOdwEqU2Fv0i9Cuw==} - - '@logtail/types@0.4.20': - resolution: {integrity: sha512-nYsum10eJMTo+ySBlYXvSrvgD1NDCVUeOlxLBbelq3XUmHu9L48VNR3P0BOmhLamYCTEgjatTj0PyPLfjL1W9g==} - '@mixmark-io/domino@2.2.0': resolution: {integrity: sha512-Y28PR25bHXUg88kCV7nivXrP2Nj2RueZ3/l/jdx6J9f8J4nsEGcgX0Qe6lt7Pa+J79+kPiJU3LguR6O/6zrLOw==} '@mongodb-js/saslprep@1.1.7': resolution: {integrity: sha512-dCHW/oEX0KJ4NjDULBo3JiOaK5+6axtpBbS+ao2ZInoAL9/YRQLhXzSNAFz7hP4nzLkIqsfYAK/PDE3+XHny0Q==} - '@msgpack/msgpack@2.8.0': - resolution: {integrity: sha512-h9u4u/jiIRKbq25PM+zymTyW6bhTzELvOoUd+AvYriWOAKpLGnIamaET3pnHYoI5iYphAHBI4ayx0MehR+VVPQ==} - engines: {node: '>= 10'} - '@msgpackr-extract/msgpackr-extract-darwin-arm64@3.0.3': resolution: {integrity: sha512-QZHtlVgbAdy2zAqNA9Gu1UpIuI8Xvsd1v8ic6B2pZmeFnFcMWiPLfWXh7TVw4eGEZ/C9TH281KwhVoeQUKbyjw==} cpu: [arm64] @@ -710,10 +683,6 @@ packages: '@one-ini/wasm@0.1.1': resolution: {integrity: sha512-XuySG1E38YScSJoMlqovLru4KTUNSjgVTIjyh7qMX6aNN5HY5Ct5LhRJdxO79JtTzKfzV/bnWpz+zquYrISsvw==} - '@opentelemetry/api-logs@0.51.1': - resolution: {integrity: sha512-E3skn949Pk1z2XtXu/lxf6QAZpawuTM/IUEXcAzpiUkTd73Hmvw26FiN3cJuTmkpM5hZzHwkomVdtrh/n/zzwA==} - engines: {node: '>=14'} - '@opentelemetry/api-logs@0.52.1': resolution: {integrity: sha512-qnSqB2DQ9TPP96dl8cDubDvrUyWc0/sK81xHTK8eSUspzDM3bsewX903qclQFvVhgStjRWdC5bLb3kQqMkfV5A==} engines: {node: '>=14'} @@ -722,567 +691,148 @@ packages: resolution: {integrity: sha512-3giAOQvZiH5F9bMlMiv8+GSPMeqg0dbaeo58/0SlA9sxSqZhnUtxzX9/2FzyhS9sWQf5S0GJE0AKBrFqjpeYcg==} engines: {node: '>=8.0.0'} - '@opentelemetry/auto-instrumentations-node@0.46.1': - resolution: {integrity: sha512-s0CwmY9KYtPawOhV5YO2Gf62uVOQRNvT6Or8IZ0S4gr/kPVNhoMehTsQvqBwSWQfoFrkmW3KKOHiKJEp4dVGXg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.4.1 - - '@opentelemetry/context-async-hooks@1.24.1': - resolution: {integrity: sha512-R5r6DO4kgEOVBxFXhXjwospLQkv+sYxwCfjvoZBe7Zm6KKXAV9kDSJhi/D1BweowdZmO+sdbENLs374gER8hpQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': '>=1.0.0 <1.9.0' - '@opentelemetry/context-async-hooks@1.25.1': resolution: {integrity: sha512-UW/ge9zjvAEmRWVapOP0qyCvPulWU6cQxGxDbWEFfGOj1VBBZAuOqTo3X6yWmDTD3Xe15ysCZChHncr2xFMIfQ==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': '>=1.0.0 <1.10.0' - '@opentelemetry/core@1.24.1': - resolution: {integrity: sha512-wMSGfsdmibI88K9wB498zXY04yThPexo8jvwNNlm542HZB7XrrMRBbAyKJqG8qDRJwIBdBrPMi4V9ZPW/sqrcg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': '>=1.0.0 <1.9.0' - '@opentelemetry/core@1.25.1': resolution: {integrity: sha512-GeT/l6rBYWVQ4XArluLVB6WWQ8flHbdb6r2FCHC3smtdOAbrJBIv35tpV/yp9bmYUJf+xmZpu9DRTIeJVhFbEQ==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': '>=1.0.0 <1.10.0' - '@opentelemetry/exporter-logs-otlp-http@0.51.1': - resolution: {integrity: sha512-cd6GZ9IqCrmvOJwi1HjRR7o9ihF7xhZTekgxUsoyTsPF+SjKMsLF9ur6HeBYkYhk+YjZ1ken3XUMH47oUTvu8Q==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/exporter-metrics-otlp-http@0.51.1': - resolution: {integrity: sha512-oFXvif9iksHUxrzG3P8ohMLt7xSrl+oDMqxD/3XXndU761RFAKSbRDpfrQs25U5D+A2aMV3qk+4kfUWdJhZ77g==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/exporter-metrics-otlp-proto@0.51.1': - resolution: {integrity: sha512-jhj8xD6S4cydXGCuf2tp56+4QI0DbDH6g+0MiPPJVdXjxLj+iycQuqB2cwljWpByblFaOjyUsL/VKtm8C7sQ9A==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/exporter-trace-otlp-grpc@0.51.1': - resolution: {integrity: sha512-P9+Hkszih95ITvldGZ+kXvj9HpD1QfS+PwooyHK72GYA+Bgm+yUSAsDkUkDms8+s9HW6poxURv3LcjaMuBBpVQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/exporter-trace-otlp-http@0.51.1': - resolution: {integrity: sha512-n+LhLPsX07URh+HhV2SHVSvz1t4G/l/CE5BjpmhAPqeTceFac1VpyQkavWEJbvnK5bUEXijWt4LxAxFpt2fXyw==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/exporter-trace-otlp-proto@0.51.1': - resolution: {integrity: sha512-SE9f0/6V6EeXC9i+WA4WFjS1EYgaBCpAnI5+lxWvZ7iO7EU1IvHvZhP6Kojr0nLldo83gqg6G7OWFqsID3uF+w==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/exporter-zipkin@1.24.1': - resolution: {integrity: sha512-+Rl/VFmu2n6eaRMnVbyfZx1DqR/1KNyWebYuHyQBZaEAVIn/ZLgmofRpXN1X2nhJ4BNaptQUNxAstCYYz6dKoQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/instrumentation-amqplib@0.37.0': - resolution: {integrity: sha512-XjOHeAOreh0XX4jlzTTUWWqu1dIGvMWM8yvd43JJdRMAmTZisezjKsxLjMEMIvF0PzQdoXwh9DiS9nYE4/QmpA==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-aws-lambda@0.41.1': - resolution: {integrity: sha512-/BLG+0DQr2tCILFGJKJH2Fg6eyjhqOlVflYpNddUEXnzyQ/PAhTdgirkqbICFgeSW2XYcEY9zXpuRldrVNw9cA==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-aws-sdk@0.41.0': - resolution: {integrity: sha512-7+8WMY0LQeqv6KIObXK+Py44qNFLeCU0ZLLxSZtXEbZ2wJlQISP1St65jRto0NV7isnZoyuOxb2+ZpypPPNv7Q==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-bunyan@0.38.0': - resolution: {integrity: sha512-ThNcgTE22W7PKzTzz5qfGxb5Gf7rA3EORousYo2nJWHHcF6gqiMNv2+GXY3MdpjLBr8IgCfhtvbQdD6rlIPUpA==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-cassandra-driver@0.38.0': - resolution: {integrity: sha512-ML4Vw0it2uIpETfX6skuSIGLHF9D3TUKOfdfrk9lnrzzWSzg2aS6pl3UeepkQX4wXHdzlxVRB0USrUqsmxMd5Q==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-connect@0.36.1': - resolution: {integrity: sha512-xI5Q/CMmzBmHshPnzzjD19ptFaYO/rQWzokpNio4QixZYWhJsa35QgRvN9FhPkwgtuJIbt/CWWAufJ3egJNHEA==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-connect@0.38.0': resolution: {integrity: sha512-2/nRnx3pjYEmdPIaBwtgtSviTKHWnDZN3R+TkRUnhIVrvBKVcq+I5B2rtd6mr6Fe9cHlZ9Ojcuh7pkNh/xdWWg==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-cucumber@0.6.0': - resolution: {integrity: sha512-90eAF2JPSbPAsOuGfYyctYaoYXqy4Clbxt0j/uUgg6dto4oqwUw3AvTyHQEztLGxeXwEzC1EQigDtVPg5ZexYA==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/instrumentation-dataloader@0.9.0': - resolution: {integrity: sha512-fiyCOAw+tlbneok1x7P5UseoGW5nS60CWWx7NXzYW+WOexpSmDQQW7olttGa8fqE6/sVCoi1l+QdfVoETZi/NQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-dns@0.36.1': - resolution: {integrity: sha512-NWRbQ7q0E3co/CNTWLZZvUzZoKhB1iTitY282IM8HDTXkA6VRssCfOcvaHw5ezOh23TJbAeYxmmpVj4hFvDPYQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-express@0.39.0': - resolution: {integrity: sha512-AG8U7z7D0JcBu/7dDcwb47UMEzj9/FMiJV2iQZqrsZnxR3FjB9J9oIH2iszJYci2eUdp2WbdvtpD9RV/zmME5A==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-express@0.41.1': resolution: {integrity: sha512-uRx0V3LPGzjn2bxAnV8eUsDT82vT7NTwI0ezEuPMBOTOsnPpGhWdhcdNdhH80sM4TrWrOfXm9HGEdfWE3TRIww==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-fastify@0.36.1': - resolution: {integrity: sha512-3Nfm43PI0I+3EX+1YbSy6xbDu276R1Dh1tqAk68yd4yirnIh52Kd5B+nJ8CgHA7o3UKakpBjj6vSzi5vNCzJIA==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-fastify@0.38.0': resolution: {integrity: sha512-HBVLpTSYpkQZ87/Df3N0gAw7VzYZV3n28THIBrJWfuqw3Or7UqdhnjeuMIPQ04BKk3aZc0cWn2naSQObbh5vXw==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-fs@0.12.0': - resolution: {integrity: sha512-Waf+2hekJRxIwq1PmivxOWLdMOtYbY22hKr34gEtfbv2CArSv8FBJH4BmQxB9o5ZcwkdKu589qs009dbuSfNmQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-fs@0.14.0': resolution: {integrity: sha512-pVc8P5AgliC1DphyyBUgsxXlm2XaPH4BpYvt7rAZDMIqUpRk8gs19SioABtKqqxvFzg5jPtgJfJsdxq0Y+maLw==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-generic-pool@0.36.0': - resolution: {integrity: sha512-CExAEqJvK8jYxrhN8cl6EaGg57EGJi+qsSKouLC5lndXi68gZLOKbZIMZg4pF0kNfp/D4BFaGmA6Ap7d5WoPTw==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-graphql@0.40.0': - resolution: {integrity: sha512-LVRdEHWACWOczv2imD+mhUrLMxsEjPPi32vIZJT57zygR5aUiA4em8X3aiGOCycgbMWkIu8xOSGSxdx3JmzN+w==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-graphql@0.42.0': resolution: {integrity: sha512-N8SOwoKL9KQSX7z3gOaw5UaTeVQcfDO1c21csVHnmnmGUoqsXbArK2B8VuwPWcv6/BC/i3io+xTo7QGRZ/z28Q==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-grpc@0.51.1': - resolution: {integrity: sha512-coRTugFL7De/VNH/1NqPlxnfik87jS+jBXsny+Y/lMhXIA3x8t71IyL9ihuewkD+lNtIxIz6Y7Sq6kPuOqz5dQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-hapi@0.38.0': - resolution: {integrity: sha512-ZcOqEuwuutTDYIjhDIStix22ECblG/i9pHje23QGs4Q4YS4RMaZ5hKCoQJxW88Z4K7T53rQkdISmoXFKDV8xMg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-hapi@0.40.0': resolution: {integrity: sha512-8U/w7Ifumtd2bSN1OLaSwAAFhb9FyqWUki3lMMB0ds+1+HdSxYBe9aspEJEgvxAqOkrQnVniAPTEGf1pGM7SOw==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-http@0.51.1': - resolution: {integrity: sha512-6b3nZnFFEz/3xZ6w8bVxctPUWIPWiXuPQ725530JgxnN1cvYFd8CJ75PrHZNjynmzSSnqBkN3ef4R9N+RpMh8Q==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-http@0.52.1': resolution: {integrity: sha512-dG/aevWhaP+7OLv4BQQSEKMJv8GyeOp3Wxl31NHqE8xo9/fYMfEljiZphUHIfyg4gnZ9swMyWjfOQs5GUQe54Q==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-ioredis@0.40.0': - resolution: {integrity: sha512-Jv/fH7KhpWe4KBirsiqeUJIYrsdR2iu2l4nWhfOlRvaZ+zYIiLEzTQR6QhBbyRoAbU4OuYJzjWusOmmpGBnwng==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-ioredis@0.42.0': resolution: {integrity: sha512-P11H168EKvBB9TUSasNDOGJCSkpT44XgoM6d3gRIWAa9ghLpYhl0uRkS8//MqPzcJVHr3h3RmfXIpiYLjyIZTw==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-knex@0.36.1': - resolution: {integrity: sha512-6bEuiI+yMf3D0+ZWZE2AKmXhIhBvZ0brdO/0A8lUqeqeS+sS4fTcjA1F2CclsCNxYWEgcs8o3QyQqPceBeVRlg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-koa@0.40.0': - resolution: {integrity: sha512-dJc3H/bKMcgUYcQpLF+1IbmUKus0e5Fnn/+ru/3voIRHwMADT3rFSUcGLWSczkg68BCgz0vFWGDTvPtcWIFr7A==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-koa@0.42.0': resolution: {integrity: sha512-H1BEmnMhho8o8HuNRq5zEI4+SIHDIglNB7BPKohZyWG4fWNuR7yM4GTlR01Syq21vODAS7z5omblScJD/eZdKw==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-lru-memoizer@0.37.0': - resolution: {integrity: sha512-dHLrn55qVWsHJQYdForPWPUWDk2HZ2jjzkT+WoQSqpYT1j4HxfoiLfBTF+I3EbEYFAJnDRmRAUfA6nU5GPdCLQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-memcached@0.36.0': - resolution: {integrity: sha512-5efkT8ZfN8il5z+yfKYFGm2YR3mhlhaJoGfNOAylKE/6tUH3WDTTWaP7nrURtWGc+fuvDktcEch18Se8qsGS7w==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-mongodb@0.43.0': - resolution: {integrity: sha512-bMKej7Y76QVUD3l55Q9YqizXybHUzF3pujsBFjqbZrRn2WYqtsDtTUlbCK7fvXNPwFInqZ2KhnTqd0gwo8MzaQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-mongodb@0.46.0': resolution: {integrity: sha512-VF/MicZ5UOBiXrqBslzwxhN7TVqzu1/LN/QDpkskqM0Zm0aZ4CVRbUygL8d7lrjLn15x5kGIe8VsSphMfPJzlA==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-mongoose@0.38.1': - resolution: {integrity: sha512-zaeiasdnRjXe6VhYCBMdkmAVh1S5MmXC/0spet+yqoaViGnYst/DOxPvhwg3yT4Yag5crZNWsVXnA538UjP6Ow==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-mongoose@0.40.0': resolution: {integrity: sha512-niRi5ZUnkgzRhIGMOozTyoZIvJKNJyhijQI4nF4iFSb+FUx2v5fngfR+8XLmdQAO7xmsD8E5vEGdDVYVtKbZew==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-mysql2@0.38.1': - resolution: {integrity: sha512-qkpHMgWSDTYVB1vlZ9sspf7l2wdS5DDq/rbIepDwX5BA0N0068JTQqh0CgAh34tdFqSCnWXIhcyOXC2TtRb0sg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-mysql2@0.40.0': resolution: {integrity: sha512-0xfS1xcqUmY7WE1uWjlmI67Xg3QsSUlNT+AcXHeA4BDUPwZtWqF4ezIwLgpVZfHOnkAEheqGfNSWd1PIu3Wnfg==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-mysql@0.38.1': - resolution: {integrity: sha512-+iBAawUaTfX/HAlvySwozx0C2B6LBfNPXX1W8Z2On1Uva33AGkw2UjL9XgIg1Pj4eLZ9R4EoJ/aFz+Xj4E/7Fw==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-mysql@0.40.0': resolution: {integrity: sha512-d7ja8yizsOCNMYIJt5PH/fKZXjb/mS48zLROO4BzZTtDfhNCl2UM/9VIomP2qkGIFVouSJrGr/T00EzY7bPtKA==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-nestjs-core@0.37.1': - resolution: {integrity: sha512-ebYQjHZEmGHWEALwwDGhSQVLBaurFnuLIkZD5igPXrt7ohfF4lc5/4al1LO+vKc0NHk8SJWStuRueT86ISA8Vg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-nestjs-core@0.39.0': resolution: {integrity: sha512-mewVhEXdikyvIZoMIUry8eb8l3HUjuQjSjVbmLVTt4NQi35tkpnHQrG9bTRBrl3403LoWZ2njMPJyg4l6HfKvA==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-net@0.36.0': - resolution: {integrity: sha512-rZlbSgwAJys8lpug+xIeAdO98ypYMAPVqrHqc4AHuUl5S4MULHEcjGLMZLoE/guEGO4xAQ5XUezpRFGM1SAnsg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-pg@0.41.0': - resolution: {integrity: sha512-BSlhpivzBD77meQNZY9fS4aKgydA8AJBzv2dqvxXFy/Hq64b7HURgw/ztbmwFeYwdF5raZZUifiiNSMLpOJoSA==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-pg@0.43.0': resolution: {integrity: sha512-og23KLyoxdnAeFs1UWqzSonuCkePUzCX30keSYigIzJe/6WSYA8rnEI5lobcxPEzg+GcU06J7jzokuEHbjVJNw==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-pino@0.39.0': - resolution: {integrity: sha512-uA17F2iP77o3NculB63QD2zv3jkJ093Gfb0GxHLEqTIqpYs1ToJ53ybWwjJwqFByxk7GrliaxaxVtWC23PKzBg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-redis-4@0.39.0': - resolution: {integrity: sha512-Zpfqfi83KeKgVQ0C2083GZPon3ZPYQ5E59v9FAbhubtOoUb9Rh7n111YD8FPW3sgx6JKp1odXmBmfQhWCaTOpQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-redis-4@0.41.0': resolution: {integrity: sha512-H7IfGTqW2reLXqput4yzAe8YpDC0fmVNal95GHMLOrS89W+qWUKIqxolSh63hJyfmwPSFwXASzj7wpSk8Az+Dg==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation-redis@0.39.1': - resolution: {integrity: sha512-HUjTerD84jRJnSyDrRPqn6xQ7K91o9qLflRPZqzRvq0GRj5PMfc6TJ/z3q/ayWy/2Kzffhrp7HCIVp0u0TkgUg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-restify@0.38.0': - resolution: {integrity: sha512-VYK47Z9GBaZX5MQLL7kZDdzQDdyUtHRD4J/GSr6kdwmIpdpUQXLsV3EnboeB8P+BlpucF57FyJKE8yWTOEMfnA==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-router@0.37.0': - resolution: {integrity: sha512-+OPcm7C9I5oPqnpStE+1WkdPWjRx0k5XKratxQmIDFZrmhRcqvMte3vrrzE/OBPg9iqh2tKrSe0y7+0sRfTJyQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-runtime-node@0.4.0': - resolution: {integrity: sha512-/NOgUF5gf3T5c3GMyy6fnQxaVzbOf9j2xcetgymIIX2HSN3Gk7o64G7KDvwHwhaa20ZiF0QDLb3m4AT+tn9eRg==} - engines: {node: '>=14.10.0'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-socket.io@0.39.0': - resolution: {integrity: sha512-4J2ehk5mJyDT6j2yJCOuPxAjit5QB1Fwzhx0LID5jjvhI9LxzZIGDNAPTTHyghSiaRDeNMzceXKkkEQJkg2MNw==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-tedious@0.10.1': - resolution: {integrity: sha512-maSXMxgS0szU52khQzAROV4nWr+3M8mZajMQOc3/7tYjo+Q3HlWAowOuagPvp4pwROK4x6oDaFYlY+ZSj1qjYA==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - - '@opentelemetry/instrumentation-undici@0.2.0': - resolution: {integrity: sha512-RH9WdVRtpnyp8kvya2RYqKsJouPxvHl7jKPsIfrbL8u2QCKloAGi0uEqDHoOS15ZRYPQTDXZ7d8jSpUgSQmvpA==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.7.0 - - '@opentelemetry/instrumentation-winston@0.37.0': - resolution: {integrity: sha512-vOx55fxdNjo2XojJf8JN4jP7VVvQCh7UQzzQ2Q2FpGJpt8Z3EErKaY8xOBkOuJH0TtL/Q72rmIn9c+mRG46BxA==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation@0.46.0': resolution: {integrity: sha512-a9TijXZZbk0vI5TGLZl+0kxyFfrXHhX6Svtz7Pp2/VBlCSKrazuULEyoJQrOknJyFWNMEmbbJgOciHCCpQcisw==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation@0.51.1': - resolution: {integrity: sha512-JIrvhpgqY6437QIqToyozrUG1h5UhwHkaGK/WAX+fkrpyPtc+RO5FkRtUd9BH0MibabHHvqsnBGKfKVijbmp8w==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.3.0 - '@opentelemetry/instrumentation@0.52.1': resolution: {integrity: sha512-uXJbYU/5/MBHjMp1FqrILLRuiJCs3Ofk0MeRDk8g1S1gD47U8X3JnSwcMO1rtRo1x1a7zKaQHaoYu49p/4eSKw==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': ^1.3.0 - '@opentelemetry/otlp-exporter-base@0.51.1': - resolution: {integrity: sha512-UYlnOYyDdzo1Gw559EHCzru0RwhvuXCwoH8jGo9J4gO1TE58GjnEmIjomMsKBCym3qWNJfIQXw+9SZCV0DdQNg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/otlp-grpc-exporter-base@0.51.1': - resolution: {integrity: sha512-ZAS+4pq8o7dsugGTwV9s6JMKSxi+guIHdn0acOv0bqj26e9pWDFx5Ky+bI0aY46uR9Y0JyXqY+KAEYM/SO3DFA==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/otlp-proto-exporter-base@0.51.1': - resolution: {integrity: sha512-gxxxwfk0inDMb5DLeuxQ3L8TtptxSiTNHE4nnAJH34IQXAVRhXSXW1rK8PmDKDngRPIZ6J7ncUCjjIn8b+AgqQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/otlp-transformer@0.51.1': - resolution: {integrity: sha512-OppYOXwV9LQqqtYUCywqoOqX/JT9LQ5/FMuPZ//eTkvuHdUC4ZMwz2c6uSoT2R90GWvvGnF1iEqTGyTT3xAt2Q==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': '>=1.3.0 <1.9.0' - - '@opentelemetry/propagation-utils@0.30.10': - resolution: {integrity: sha512-hhTW8pFp9PSyosYzzuUL9rdm7HF97w3OCyElufFHyUnYnKkCBbu8ne2LyF/KSdI/xZ81ubxWZs78hX4S7pLq5g==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/propagator-aws-xray@1.25.1': - resolution: {integrity: sha512-soZQdO9EAROMwa9bL2C0VLadbrfRjSA9t7g6X8sL0X1B8V59pzOayYMyTW9qTECn9uuJV98A7qOnJm6KH6yk8w==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': '>=1.0.0 <1.10.0' - - '@opentelemetry/propagator-b3@1.24.1': - resolution: {integrity: sha512-nda97ZwhpZKyUJTXqQuKzNhPMUgMLunbbGWn8kroBwegn+nh6OhtyGkrVQsQLNdVKJl0KeB5z0ZgeWszrYhwFw==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': '>=1.0.0 <1.9.0' - - '@opentelemetry/propagator-jaeger@1.24.1': - resolution: {integrity: sha512-7bRBJn3FG1l195A1m+xXRHvgzAOBsfmRi9uZ5Da18oTh7BLmNDiA8+kpk51FpTsU1PCikPVpRDNPhKVB6lyzZg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': '>=1.0.0 <1.9.0' - '@opentelemetry/redis-common@0.36.2': resolution: {integrity: sha512-faYX1N0gpLhej/6nyp6bgRjzAKXn5GOEMYY7YhciSfCoITAktLUtQ36d24QEWNA1/WA1y6qQunCe0OhHRkVl9g==} engines: {node: '>=14'} - '@opentelemetry/resource-detector-alibaba-cloud@0.28.10': - resolution: {integrity: sha512-TZv/1Y2QCL6sJ+X9SsPPBXe4786bc/Qsw0hQXFsNTbJzDTGGUmOAlSZ2qPiuqAd4ZheUYfD+QA20IvAjUz9Hhg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/resource-detector-aws@1.5.1': - resolution: {integrity: sha512-+IUh4gAwJf49vOJM6PIjmgOapRH5zr21ZpFnNU0QZmxRi52AXVhZN7A89pKW6GAQheWnVQLD7iUN87ieYt70tw==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/resource-detector-azure@0.2.9': - resolution: {integrity: sha512-16Z6kyrmszoa7J1uj1kbSAgZuk11K07yEDj6fa3I9XBf8Debi8y4K8ex94kpxbCfEraWagXji3bCWvaq3k4dRg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/resource-detector-container@0.3.11': - resolution: {integrity: sha512-22ndMDakxX+nuhAYwqsciexV8/w26JozRUV0FN9kJiqSWtA1b5dCVtlp3J6JivG5t8kDN9UF5efatNnVbqRT9Q==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/resource-detector-gcp@0.29.10': - resolution: {integrity: sha512-rm2HKJ9lsdoVvrbmkr9dkOzg3Uk0FksXNxvNBgrCprM1XhMoJwThI5i0h/5sJypISUAJlEeJS6gn6nROj/NpkQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': ^1.0.0 - - '@opentelemetry/resources@1.24.1': - resolution: {integrity: sha512-cyv0MwAaPF7O86x5hk3NNgenMObeejZFLJJDVuSeSMIsknlsj3oOZzRv3qSzlwYomXsICfBeFFlxwHQte5mGXQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': '>=1.0.0 <1.9.0' - '@opentelemetry/resources@1.25.1': resolution: {integrity: sha512-pkZT+iFYIZsVn6+GzM0kSX+u3MSLCY9md+lIJOoKl/P+gJFfxJte/60Usdp8Ce4rOs8GduUpSPNe1ddGyDT1sQ==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': '>=1.0.0 <1.10.0' - '@opentelemetry/sdk-logs@0.51.1': - resolution: {integrity: sha512-ULQQtl82b673PpZc5/0EtH4V+BrwVOgKJZEB7tYZnGTG3I98tQVk89S9/JSixomDr++F4ih+LSJTCqIKBz+MQQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': '>=1.4.0 <1.9.0' - '@opentelemetry/api-logs': '>=0.39.1' - - '@opentelemetry/sdk-metrics@1.24.1': - resolution: {integrity: sha512-FrAqCbbGao9iKI+Mgh+OsC9+U2YMoXnlDHe06yH7dvavCKzE3S892dGtX54+WhSFVxHR/TMRVJiK/CV93GR0TQ==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': '>=1.3.0 <1.9.0' - '@opentelemetry/sdk-metrics@1.25.1': resolution: {integrity: sha512-9Mb7q5ioFL4E4dDrc4wC/A3NTHDat44v4I3p2pLPSxRvqUbDIQyMVr9uK+EU69+HWhlET1VaSrRzwdckWqY15Q==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': '>=1.3.0 <1.10.0' - '@opentelemetry/sdk-node@0.51.1': - resolution: {integrity: sha512-GgmNF9C+6esr8PIJxCqHw84rEOkYm6XdFWZ2+Wyc3qaUt92ACoN7uSw5iKNvaUq62W0xii1wsGxwHzyENtPP8w==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': '>=1.3.0 <1.9.0' - - '@opentelemetry/sdk-trace-base@1.24.1': - resolution: {integrity: sha512-zz+N423IcySgjihl2NfjBf0qw1RWe11XIAWVrTNOSSI6dtSPJiVom2zipFB2AEEtJWpv0Iz6DY6+TjnyTV5pWg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': '>=1.0.0 <1.9.0' - '@opentelemetry/sdk-trace-base@1.25.1': resolution: {integrity: sha512-C8k4hnEbc5FamuZQ92nTOp8X/diCY56XUTnMiv9UTuJitCzaNNHAVsdm5+HLCdI8SLQsLWIrG38tddMxLVoftw==} engines: {node: '>=14'} peerDependencies: '@opentelemetry/api': '>=1.0.0 <1.10.0' - '@opentelemetry/sdk-trace-node@1.24.1': - resolution: {integrity: sha512-/FZX8uWaGIAwsDhqI8VvQ+qWtfMNlXjaFYGc+vmxgdRFppCSSIRwrPyIhJO1qx61okyYhoyxVEZAfoiNxrfJCg==} - engines: {node: '>=14'} - peerDependencies: - '@opentelemetry/api': '>=1.0.0 <1.9.0' - - '@opentelemetry/semantic-conventions@1.24.1': - resolution: {integrity: sha512-VkliWlS4/+GHLLW7J/rVBA00uXus1SWvwFvcUDxDwmFxYfg/2VI6ekwdXS28cjI8Qz2ky2BzG8OUHo+WeYIWqw==} - engines: {node: '>=14'} - '@opentelemetry/semantic-conventions@1.25.1': resolution: {integrity: sha512-ZDjMJJQRlyk8A1KZFCc+bCbsyrn1wTwdNt56F7twdfUfnHUZUq77/WfONCj8p72NZOyP7pNTdUWSTYC3GTbuuQ==} engines: {node: '>=14'} @@ -1306,36 +856,6 @@ packages: '@prisma/instrumentation@5.17.0': resolution: {integrity: sha512-c1Sle4ji8aasMcYfBBHFM56We4ljfenVtRmS8aY06BllS7SoU6SmJBwG7vil+GHiR0Yrh+t9iBwt4AY0Jr4KNQ==} - '@protobufjs/aspromise@1.1.2': - resolution: {integrity: sha512-j+gKExEuLmKwvz3OgROXtrJ2UG2x8Ch2YZUxahh+s1F2HZ+wAceUNLkvy6zKCPVRkU++ZWQrdxsUeQXmcg4uoQ==} - - '@protobufjs/base64@1.1.2': - resolution: {integrity: sha512-AZkcAA5vnN/v4PDqKyMR5lx7hZttPDgClv83E//FMNhR2TMcLUhfRUBHCmSl0oi9zMgDDqRUJkSxO3wm85+XLg==} - - '@protobufjs/codegen@2.0.4': - resolution: {integrity: sha512-YyFaikqM5sH0ziFZCN3xDC7zeGaB/d0IUb9CATugHWbd1FRFwWwt4ld4OYMPWu5a3Xe01mGAULCdqhMlPl29Jg==} - - '@protobufjs/eventemitter@1.1.0': - resolution: {integrity: sha512-j9ednRT81vYJ9OfVuXG6ERSTdEL1xVsNgqpkxMsbIabzSo3goCjDIveeGv5d03om39ML71RdmrGNjG5SReBP/Q==} - - '@protobufjs/fetch@1.1.0': - resolution: {integrity: sha512-lljVXpqXebpsijW71PZaCYeIcE5on1w5DlQy5WH6GLbFryLUrBD4932W/E2BSpfRJWseIL4v/KPgBFxDOIdKpQ==} - - '@protobufjs/float@1.0.2': - resolution: {integrity: sha512-Ddb+kVXlXst9d+R9PfTIxh1EdNkgoRe5tOX6t01f1lYWOvJnSPDBlG241QLzcyPdoNTsblLUdujGSE4RzrTZGQ==} - - '@protobufjs/inquire@1.1.0': - resolution: {integrity: sha512-kdSefcPdruJiFMVSbn801t4vFK7KB/5gd2fYvrxhuJYg8ILrmn9SKSX2tZdV6V+ksulWqS7aXjBcRXl3wHoD9Q==} - - '@protobufjs/path@1.1.2': - resolution: {integrity: sha512-6JOcJ5Tm08dOHAbdR3GrvP+yUUfkjG5ePsHYczMFLq3ZmMkAD98cDgcT2iA1lJ9NVwFd4tH/iSSoe44YWkltEA==} - - '@protobufjs/pool@1.1.0': - resolution: {integrity: sha512-0kELaGSIDBKvcgS4zkjz1PeddatrjYcmMWOlAuAPwAeccUrPHdUqo/J6LiymHHEiJT5NrF1UVwxY14f+fy4WQw==} - - '@protobufjs/utf8@1.1.0': - resolution: {integrity: sha512-Vvn3zZrhQZkkBE8LSuW3em98c0FwgO4nxzv6OdSxPKJIEKY2bGbHn+mhGIPerzI4twdxaP8/0+06HBpwf345Lw==} - '@puppeteer/browsers@2.2.3': resolution: {integrity: sha512-bJ0UBsk0ESOs6RFcLXOt99a3yTDcOKlzfjad+rhFwdaG1Lu/Wzq58GHYCDTlZ9z6mldf4g+NTb+TXEfe0PpnsQ==} engines: {node: '>=18'} @@ -1423,10 +943,6 @@ packages: engines: {node: '>= 10'} hasBin: true - '@sentry/core@8.13.0': - resolution: {integrity: sha512-N9Qg4ZGxZWp8eb2eUUHVVKgjBLtFIjS805nG92s6yJmkvOpKm6mLtcUaT/iDf3Hta6nG+xRkhbE3r+Z4cbXG8w==} - engines: {node: '>=14.18'} - '@sentry/core@8.26.0': resolution: {integrity: sha512-g/tVmTZD4GNbLFf++hKJfBpcCAtduFEMLnbfa9iT/QEZjlmP+EzY+GsH9bafM5VsNe8DiOUp+kJKWtShzlVdBA==} engines: {node: '>=14.18'} @@ -1450,18 +966,10 @@ packages: engines: {node: '>=14.18'} hasBin: true - '@sentry/types@8.13.0': - resolution: {integrity: sha512-r63s/H5gvQnQM9tTGBXz2xErUbxZALh4e2Lg/1aHj4zIvGLBjA2z5qWsh6TEZYbpmgAyGShLDr6+rWeUVf9yBQ==} - engines: {node: '>=14.18'} - '@sentry/types@8.26.0': resolution: {integrity: sha512-zKmh6SWsJh630rpt7a9vP4Cm4m1C2gDTUqUiH565CajCL/4cePpNWYrNwalSqsOSL7B9OrczA1+n6a6XvND+ng==} engines: {node: '>=14.18'} - '@sentry/utils@8.13.0': - resolution: {integrity: sha512-PxV0v9VbGWH9zP37P5w2msLUFDr287nYjoY2XVF+RSolyiTs1CQNI5ZMUO3o4MsSac/dpXxjyrZXQd72t/jRYA==} - engines: {node: '>=14.18'} - '@sentry/utils@8.26.0': resolution: {integrity: sha512-xvlPU9Hd2BlyT+FhWHGNwnxWqdVRk2AHnDtVcW4Ma0Ri5EwS+uy4Jeik5UkSv8C5RVb9VlxFmS8LN3I1MPJsLw==} engines: {node: '>=14.18'} @@ -1613,12 +1121,6 @@ packages: '@tsconfig/recommended@1.0.6': resolution: {integrity: sha512-0IKu9GHYF1NGTJiYgfWwqnOQSlnE9V9R7YohHNNf0/fj/SyOZWzdd06JFr0fLpg1Mqw0kGbYg8w5xdkSqLKM9g==} - '@types/accepts@1.3.7': - resolution: {integrity: sha512-Pay9fq2lM2wXPWbteBsRAGiWH2hig4ZE2asK+mm7kUzlxRTfL961rj89I6zV/E3PcIkDqyuBEcMxFT7rccugeQ==} - - '@types/aws-lambda@8.10.122': - resolution: {integrity: sha512-vBkIh9AY22kVOCEKo5CJlyCgmSWvasC+SWUxL/x/vOwRobMpI/HG1xp/Ae3AqmSiZeLUbOhW0FCD3ZjqqUxmXw==} - '@types/babel__core@7.20.5': resolution: {integrity: sha512-qoQprZvz5wQFJwMDqeseRXWv3rqMvhgpbXFfVyWhbx9X47POIA6i/+dXefEmZKoAgOaTdaIgNSMqMIU61yRyzA==} @@ -1634,24 +1136,18 @@ packages: '@types/body-parser@1.19.5': resolution: {integrity: sha512-fB3Zu92ucau0iQ0JMCFQE7b/dv8Ot07NI3KaZIkIUNXq82k4eBAqUaneXfleGY9JWskeS9y+u0nXMyspcuQrCg==} - '@types/bunyan@1.8.9': - resolution: {integrity: sha512-ZqS9JGpBxVOvsawzmVt30sP++gSQMTejCkIAQ3VdadOcRE8izTyW66hufvwLeH+YEGP6Js2AW7Gz+RMyvrEbmw==} - '@types/connect@3.4.36': resolution: {integrity: sha512-P63Zd/JUGq+PdrM1lv0Wv5SBYeA2+CORvbrXbngriYY0jzLUWfQMQQxOhjONEz/wlHOAxOdY7CY65rgQdTjq2w==} '@types/connect@3.4.38': resolution: {integrity: sha512-K6uROf1LD88uDQqJCktA4yzL1YYAK6NgfsI0v/mTgyPKWsX1CnJ0XPSDhViejru1GcRkLWb8RlzFYJRqGUbaug==} - '@types/content-disposition@0.5.8': - resolution: {integrity: sha512-QVSSvno3dE0MgO76pJhmv4Qyi/j0Yk9pBp0Y7TJ2Tlj+KCgJWY6qX7nnxCOLkZ3VYRSIk1WTxCvwUSdx6CCLdg==} - - '@types/cookies@0.9.0': - resolution: {integrity: sha512-40Zk8qR147RABiQ7NQnBzWzDcjKzNrntB5BAmeGCb2p/MIyOE+4BVvc17wumsUqUw00bJYqoXFHYygQnEFh4/Q==} - '@types/cors@2.8.17': resolution: {integrity: sha512-8CGDvrBj1zgo2qE+oS3pOCyYNqCPryMWY2bGfwA0dcfopWGgxs+78df0Rs3rc9THP4JkOhLsAa+15VdpAqkcUA==} + '@types/escape-html@1.0.4': + resolution: {integrity: sha512-qZ72SFTgUAZ5a7Tj6kf2SHLetiH5S6f8G5frB2SPQ3EyF02kxdyBFf4Tz4banE3xCgGnKgWLt//a6VuYHKYJTg==} + '@types/express-serve-static-core@4.19.3': resolution: {integrity: sha512-KOzM7MhcBFlmnlr/fzISFF5vGWVSvN6fTd4T+ExOt08bA/dA5kpSzY52nMsI1KDFmUREpJelPYyuslLRSjjgCg==} @@ -1664,9 +1160,6 @@ packages: '@types/graceful-fs@4.1.9': resolution: {integrity: sha512-olP3sd1qOEe5dXTSaFvQG+02VdRXcdytWLAZsAq1PecU8uqQAhkrnbli7DagjtXKW/Bl7YJbUsa8MPcuc8LHEQ==} - '@types/http-assert@1.5.5': - resolution: {integrity: sha512-4+tE/lwdAahgZT1g30Jkdm9PzFRde0xwxBNUyRsCitRvCQB90iuA2uJYdUnhnANRcqGXaWOGY4FEoxeElNAK2g==} - '@types/http-errors@2.0.4': resolution: {integrity: sha512-D0CFMMtydbJAegzOyHjtiKPLlvnm3iTZyZRSZoLq2mRhDdmLfIWOCYPfQJ4cu2erKghU++QvjcUjp/5h7hESpA==} @@ -1682,21 +1175,6 @@ packages: '@types/jest@29.5.12': resolution: {integrity: sha512-eDC8bTvT/QhYdxJAulQikueigY5AsdBRH2yDKW3yveW7svY3+DzN84/2NUgkw10RTiJbWqZrTtoGVdYlvFJdLw==} - '@types/keygrip@1.0.6': - resolution: {integrity: sha512-lZuNAY9xeJt7Bx4t4dx0rYCDqGPW8RXhQZK1td7d4H6E9zYbLoOtjBvfwdTKpsyxQI/2jv+armjX/RW+ZNpXOQ==} - - '@types/koa-compose@3.2.8': - resolution: {integrity: sha512-4Olc63RY+MKvxMwVknCUDhRQX1pFQoBZ/lXcRLP69PQkEpze/0cr8LNqJQe5NFb/b19DWi2a5bTi2VAlQzhJuA==} - - '@types/koa@2.14.0': - resolution: {integrity: sha512-DTDUyznHGNHAl+wd1n0z1jxNajduyTh8R53xoewuerdBzGo6Ogj6F2299BFtrexJw4NtgjsI5SMPCmV9gZwGXA==} - - '@types/koa__router@12.0.3': - resolution: {integrity: sha512-5YUJVv6NwM1z7m6FuYpKfNLTZ932Z6EF6xy2BbtpJSyn13DKNQEkXVffFVSnJHxvwwWh2SAeumpjAYUELqgjyw==} - - '@types/memcached@2.2.10': - resolution: {integrity: sha512-AM9smvZN55Gzs2wRrqeMHVP7KE8KWgCJO/XL5yCly2xF6EKa4YlbpK+cLSAH4NG/Ah64HrlegmGqW8kYws7Vxg==} - '@types/mime@1.3.5': resolution: {integrity: sha512-/pyBZWSLD2n0dcHE3hq8s8ZvcETHtEuF+3E7XVt0Ig2nvsVQXdghHVcEkIWjy9A0wKfTn97a/PSDYohKIlnP/w==} @@ -1712,6 +1190,9 @@ packages: '@types/node@20.14.1': resolution: {integrity: sha512-T2MzSGEu+ysB/FkWfqmhV3PLyQlowdptmmgD20C6QxsS8Fmv5SjpZ1ayXaEC0S21/h5UJ9iA6W/5vSNU5l00OA==} + '@types/pdf-parse@1.1.4': + resolution: {integrity: sha512-+gbBHbNCVGGYw1S9lAIIvrHW47UYOhMIFUsJcMkMrzy1Jf0vulBN3XQIjPgnoOXveMuHnF3b57fXROnY/Or7eg==} + '@types/pg-pool@2.0.4': resolution: {integrity: sha512-qZAvkv1K3QbmHHFYSNRYPkRjOWRLBYrL4B9c+wG0GSVGBw0NtJwPcgx/DSddeDJvRGMHCEQ4VMEVfuJ/0gZ3XQ==} @@ -1739,15 +1220,9 @@ packages: '@types/shimmer@1.0.5': resolution: {integrity: sha512-9Hp0ObzwwO57DpLFF0InUjUm/II8GmKAvzbefxQTihCb7KI6yc9yzf0nLc4mVdby5N4DRCgQM2wCup9KTieeww==} - '@types/stack-trace@0.0.29': - resolution: {integrity: sha512-TgfOX+mGY/NyNxJLIbDWrO9DjGoVSW9+aB8H2yy1fy32jsvxijhmyJI9fDFgvz3YP4lvJaq9DzdR/M1bOgVc9g==} - '@types/stack-utils@2.0.3': resolution: {integrity: sha512-9aEbYZ3TbYMznPdcdr3SmIrLXwC/AKZXQeCf9Pgao5CKb8CyHuEX5jzWPTkvregvhRJHcpRO6BFoGW9ycaOkYw==} - '@types/tedious@4.0.14': - resolution: {integrity: sha512-KHPsfX/FoVbUGbyYvk1q9MMQHLPeRZhRJZdO45Q4YjvFkv4hMNghCWTvy7rdKessBsmtz4euWCWAB6/tVpI1Iw==} - '@types/triple-beam@1.3.5': resolution: {integrity: sha512-6WaYesThRMCl19iryMYP7/x2OVgCtbIVflDGFpWnb9irXI3UjYE4AzmYuiUKY1AJstGijoY+MgUszMgRxIYTYw==} @@ -1974,9 +1449,6 @@ packages: resolution: {integrity: sha512-4Bcg1P8xhUuqcii/S0Z9wiHIrQVPMermM1any+MX5GeGD7faD3/msQUDGLol9wOcz4/jbg/WJnGqoJF6LiBdtg==} engines: {node: '>=10.0.0'} - bignumber.js@9.1.2: - resolution: {integrity: sha512-2/mKyZH9K85bzOEfhXDBFZTGd1CTs+5IHpeFQo9luiBG7hghdC851Pj2WAhb6E3R6b9tZj/XKhbg4fum+Kepug==} - bin-links@4.0.4: resolution: {integrity: sha512-cMtq4W5ZsEwcutJrVId+a/tjt8GSbS+h0oNkdl6+6rBuEv8Ot33Bevj5KPm40t309zuhVic8NjpuL42QCiJWWA==} engines: {node: ^14.17.0 || ^16.13.0 || >=18.0.0} @@ -1988,9 +1460,6 @@ packages: binary-search@1.3.6: resolution: {integrity: sha512-nbE1WxOTTrUWIfsfZ4aHGYu5DOuNkbxGokjV6Z2kxfJK3uaAb8zNK1muzOeipoLHZjInT4Br88BHpzevc681xA==} - bl@4.1.0: - resolution: {integrity: sha512-1W07cM9gS6DcLperZfFSj+bWLtaPGSOHWhPiGzXmvVJbRLdG82sH/Kn8EtW1VqWVA54AKf2h5k5BbnIbwF3h6w==} - bluebird@3.4.7: resolution: {integrity: sha512-iD3898SR7sWVRHbiQv+sHUtHnMvC1o3nW5rAcqnq3uOn07DSAppZYUkIGslDz6gXC7HfunPe7YVBgoEJASPcHA==} @@ -2118,22 +1587,10 @@ packages: cjs-module-lexer@1.3.1: resolution: {integrity: sha512-a3KdPAANPbNE4ZUv9h6LckSl9zLsYOP4MBmhIPkRaeyybt+r4UghLvq+xw/YwUcC1gqylCkL4rdVs3Lwupjm4Q==} - cli-cursor@3.1.0: - resolution: {integrity: sha512-I/zHAwsKf9FqGoXM4WWRACob9+SNukZTd94DWF57E4toouRulbCxcUh6RKUEOQlYTHJnzkPMySvPNaaSLNfLZw==} - engines: {node: '>=8'} - - cli-spinners@2.9.2: - resolution: {integrity: sha512-ywqV+5MmyL4E7ybXgKys4DugZbX0FC6LnwrhjuykIjnK9k8OQacQ7axGKnjDXWNhns0xot3bZI5h55H8yo9cJg==} - engines: {node: '>=6'} - cliui@8.0.1: resolution: {integrity: sha512-BSeNnyus75C4//NQ9gQt1/csTXyo/8Sb+afLAkzAptFuMsod9HFokGNudZpi/oQV73hnVK+sR+5PVRMd+Dr7YQ==} engines: {node: '>=12'} - clone@1.0.4: - resolution: {integrity: sha512-JQHZ2QMW6l3aH/j6xCqQThY/9OH4D/9ls34cgkUBiEeocRTU04tHfKPBsUK1PqZCUQM7GiA0IIXJSuXHI64Kbg==} - engines: {node: '>=0.8'} - cluster-key-slot@1.1.2: resolution: {integrity: sha512-RMr0FhtfXemyinomL4hrWcYJxmX6deFdCxpJzhDttxgO1+bcCnkk+9drydLVDmAMG7NE6aN/fl4F7ucU/90gAA==} engines: {node: '>=0.10.0'} @@ -2165,6 +1622,15 @@ packages: color-name@1.1.4: resolution: {integrity: sha512-dOy+3AuW3a2wNbZHIuMZpTcgjGuLU/uBL/ubcZF9OXbDo8ff4O8yVp5Bf0efS8uEoYo5q4Fx7dY9OgQGXgAsQA==} + color-string@1.9.1: + resolution: {integrity: sha512-shrVawQFojnZv6xM40anx4CkoDP+fZsw/ZerEMsW/pyzsRbElpsL/DBVW7q3ExxwusdNXI3lXpuhEZkzs8p5Eg==} + + color@3.2.1: + resolution: {integrity: sha512-aBl7dZI9ENN6fUGC7mWpMTPNHmWUSNan9tuWN6ahh5ZLNk9baLJOnSMlrQkHcrfFgz2/RigjUVAjdx36VcemKA==} + + colorspace@1.1.4: + resolution: {integrity: sha512-BgvKJiuVu1igBUF2kEjRCZXol6wiiGbY5ipL/oVPwm0BL9sIpMIzM8IK7vwuxIIzOXMV3Ey5w+vxhm0rR/TN8w==} + combined-stream@1.0.8: resolution: {integrity: sha512-FQN4MRfuJeHf7cBbBMJFXhKSDq+2kAArBlmRBvcvFE5BB1HZKXtSFASDhdlz9zOYwxh8lDdnvmMOe/+5cdoEdg==} engines: {node: '>= 0.8'} @@ -2323,17 +1789,10 @@ packages: resolution: {integrity: sha512-3sUqbMEc77XqpdNO7FRyRog+eW3ph+GYCbj+rK+uYyRMuwsVy0rMiVtPn+QJlKFvWP/1PYpapqYn0Me2knFn+A==} engines: {node: '>=0.10.0'} - defaults@1.0.4: - resolution: {integrity: sha512-eFuaLoy/Rxalv2kr+lqMlUnrDWV+3j4pljOIJgLIhI058IQfWJ7vXhyEIHu+HtC738klGALYxOKDO0bQP3tg8A==} - define-data-property@1.1.4: resolution: {integrity: sha512-rBMvIzlpA8v6E+SJZoo++HAYqsLrkg7MSfIinMPFhmkorw7X+dOXVJQs+QT69zGkzMyfDnIMN2Wid1+NbL3T+A==} engines: {node: '>= 0.4'} - define-lazy-prop@2.0.0: - resolution: {integrity: sha512-Ds09qNh8yw3khSjiJjiUInaGX9xlqZDY7JVryGxdxV7NPeuqQfplOpQ66yJFZut3jLa5zOwkXw1g9EI2uKh4Og==} - engines: {node: '>=8'} - degenerator@5.0.1: resolution: {integrity: sha512-TllpMR/t0M5sqCXfj85i4XaAzxmS5tVA16dqvdkMwGmzI+dXLXnw3J+3Vdv7VKw+ThlTMboK6i9rnZ6Nntj5CQ==} engines: {node: '>= 14'} @@ -2440,6 +1899,9 @@ packages: emoji-regex@9.2.2: resolution: {integrity: sha512-L18DaJsXSUk2+42pv8mLs5jJT2hqFkFE4j21wOmgbUqsZ2hL72NsUU785g9RXgo3s0ZNgVl42TiHp3ZtOv/Vyg==} + enabled@2.0.0: + resolution: {integrity: sha512-AKrN98kuwOzMIdAizXGI86UFBoo26CL21UM763y1h/GMSJ4/OHU9k2YlsmBpyScFo/wbLzWQJBMCW4+IO3/+OQ==} + encodeurl@1.0.2: resolution: {integrity: sha512-TPJXq8JqFaVYm2CWmPvnP2Iyo4ZSM7/QKcSmuMLDObfpH5fi7RUGmd/rTDf+rut/saiDiQEeVTNgAmJEdAOx0w==} engines: {node: '>= 0.8'} @@ -2542,9 +2004,6 @@ packages: resolution: {integrity: sha512-5T6nhjsT+EOMzuck8JjBHARTHfMht0POzlA60WV2pMD3gyXw2LZnZ+ueGdNxG+0calOJcWKbpFcuzLZ91YWq9Q==} engines: {node: '>= 0.10.0'} - extend@3.0.2: - resolution: {integrity: sha512-fjquC59cD7CyW6urNXK0FBufkZcoiGG80wTuPujX590cB5Ttln20E2UB4S/WARVqhXffZl2LNgS+gQdPIIim/g==} - extract-zip@2.0.1: resolution: {integrity: sha512-GDhU9ntwuKyGXdZBUgTIe+vXnWj0fppUEtMDL0+idd5Sta8TGpHssn/eusA9mrPr9qNDym6SxAYZjNvCn/9RBg==} engines: {node: '>= 10.17.0'} @@ -2597,6 +2056,9 @@ packages: resolution: {integrity: sha512-b6suED+5/3rTpUBdG1gupIl8MPFCAMA0QXwmljLhvCUKcUvdE4gWky9zpuGCcXHOsz4J9wPGNWq6OKpmIzz3hQ==} hasBin: true + fn.name@1.1.0: + resolution: {integrity: sha512-GRnmB5gPyJpAhTQdSZTSp9uaPSvl09KoYcMQtsB9rQoOmzs9dH6ffeccH+Z+cv6P68Hu5bC6JjRh4Ah/mHSNRw==} + follow-redirects@1.15.6: resolution: {integrity: sha512-wWN62YITEaOpSK584EZXJafH1AGpO8RVgElfkuXbTOrPX4fIfOyEpW/CsiNd8JdYrAoOvafRTOEnvsO++qCqFA==} engines: {node: '>=4.0'} @@ -2651,14 +2113,6 @@ packages: function-bind@1.1.2: resolution: {integrity: sha512-7XHNxH7qX9xG5mIwxkhumTox/MIRNcOgDrxWsMt2pAr23WHp6MrRlN7FBSFpCpr+oVO0F744iUgR82nJMfG2SA==} - gaxios@6.7.0: - resolution: {integrity: sha512-DSrkyMTfAnAm4ks9Go20QGOcXEyW/NmZhvTYBU2rb4afBB393WIMQPWPEDMl/k8xqiNN9HYq2zao3oWXsdl2Tg==} - engines: {node: '>=14'} - - gcp-metadata@6.1.0: - resolution: {integrity: sha512-Jh/AIwwgaxan+7ZUUmRLCjtchyDiqh4KjBJ5tW3plBZb5iL/BPcso8A5DlzeD9qlw0duCamnNdpFjxwaT0KyKg==} - engines: {node: '>=14'} - generic-pool@3.9.0: resolution: {integrity: sha512-hymDOu5B53XvN4QT9dBmZxPX4CWhBPPLguTZ9MMFeFa/Kg0xWVfylOVNlJji/E7yTZWFd/q9GO5TxDLq156D7g==} engines: {node: '>= 4'} @@ -2831,9 +2285,6 @@ packages: import-in-the-middle@1.7.1: resolution: {integrity: sha512-1LrZPDtW+atAxH42S6288qyDFNQ2YCty+2mxEPRtfazH6Z5QwkaBSTS2ods7hnVJioF6rkRfNoA6A/MstpFXLg==} - import-in-the-middle@1.7.4: - resolution: {integrity: sha512-Lk+qzWmiQuRPPulGQeK5qq0v32k2bHnWrRPFgqyvhw7Kkov5L6MOLOIU3pcWeujc9W4q54Cp3Q2WV16eQkc7Bg==} - import-local@3.1.0: resolution: {integrity: sha512-ASB07uLtnDs1o6EHjKpX34BKYDSqnFerfTOJL2HvMqF70LnxpjkzDB8J44oT9pu4AMPkQwf8jl6szgvNd2tRIg==} engines: {node: '>=8'} @@ -2871,6 +2322,9 @@ packages: is-arrayish@0.2.1: resolution: {integrity: sha512-zz06S8t0ozoDXMG+ube26zeCTNXcKIPJZJi8hBrF4idCLms4CG9QtK7qBl1boi5ODzFpjswb5JPmHCbMpjaYzg==} + is-arrayish@0.3.2: + resolution: {integrity: sha512-eVRqCvVlZbuw3GrM63ovNSNAeA1K16kaR/LRY/92w0zxQ5/1YzwblUX652i4Xs9RwAGjW9d9y6X88t8OaAJfWQ==} + is-binary-path@2.1.0: resolution: {integrity: sha512-ZMERYes6pDydyuGidse7OsHxtbI7WVeUEozgR/g7rd0xUimYNlvZRE/K2MgZTjWy725IfelLeVcEM97mmtRGXw==} engines: {node: '>=8'} @@ -2881,11 +2335,6 @@ packages: is-core-module@2.13.1: resolution: {integrity: sha512-hHrIjvZsftOsvKSn2TRYl63zvxsgE0K+0mYMoH6gD4omR5IWB2KynivBQczo3+wF1cCkjzvptnI9Q0sPU66ilw==} - is-docker@2.2.1: - resolution: {integrity: sha512-F+i2BKsFrH66iaUFc0woD8sLy8getkwTwtOBjvs56Cx4CgJDeKQeqfz8wAYiSb8JOprWhHH5p77PbmYCvvUuXQ==} - engines: {node: '>=8'} - hasBin: true - is-extglob@2.1.1: resolution: {integrity: sha512-SbKbANkN603Vi4jEZv49LeVJMn4yGwsbzZworEoyEiutsN3nJYdbO36zfhGJ6QEDpOZIFkDtnq5JRxmvl3jsoQ==} engines: {node: '>=0.10.0'} @@ -2902,10 +2351,6 @@ packages: resolution: {integrity: sha512-xelSayHH36ZgE7ZWhli7pW34hNbNl8Ojv5KVmkJD4hBdD3th8Tfk9vYasLM+mXWOZhFkgZfxhLSnrwRr4elSSg==} engines: {node: '>=0.10.0'} - is-interactive@1.0.0: - resolution: {integrity: sha512-2HvIEKRoqS62guEC+qBjpvRubdX910WCMuJTZ+I9yvqKU2/12eSL549HMwtabb4oupdj2sMP50k+XJfB/8JE6w==} - engines: {node: '>=8'} - is-number@7.0.0: resolution: {integrity: sha512-41Cifkg6e8TylSpdtTpeLVMqvSBEVzTttHvERD741+pnZ8ANv0004MRL43QKPDlK9cGvNp6NZWZUBlbGXYxxng==} engines: {node: '>=0.12.0'} @@ -2922,14 +2367,6 @@ packages: resolution: {integrity: sha512-hFoiJiTl63nn+kstHGBtewWSKnQLpyb155KHheA1l39uvtO9nWIop1p3udqPcUd/xbF1VLMO4n7OI6p7RbngDg==} engines: {node: '>=8'} - is-unicode-supported@0.1.0: - resolution: {integrity: sha512-knxG2q4UC3u8stRGyAVJCOdxFmv5DZiRcdlIaAQXAbSfJya+OhopNotLQrstBhququ4ZpuKbDc/8S6mgXgPFPw==} - engines: {node: '>=10'} - - is-wsl@2.2.0: - resolution: {integrity: sha512-fKzAra0rGJUUBwGBgNkHZuToZcn+TtXHpeCgmkMJMMYx1sQDYaCSyjJBSCa2nH1DGm7s3n1oBnohoVTBaN7Lww==} - engines: {node: '>=8'} - isarray@1.0.0: resolution: {integrity: sha512-VLghIWNM6ELQzo7zwmcg0NmTVyWKYjvIeM83yjp0wRDTmUnrM678fQbcKBo6n2CJEF0szoG//ytg+TKla89ALQ==} @@ -3138,9 +2575,6 @@ packages: engines: {node: '>=4'} hasBin: true - json-bigint@1.0.0: - resolution: {integrity: sha512-SiPv/8VpZuWbvLSMtTDU8hEfrZWg/mH/nV/b4o0CYbSxu1UIQPLdwKOCIyLQX+VIPO5vrLX3i8qtqFyhdPSUSQ==} - json-parse-even-better-errors@2.3.1: resolution: {integrity: sha512-xyFwyhro/JEof6Ghe2iz2NcXoj2sloNsWr/XsERDK/oiPCfaNhl5ONfp+jQdAZRQQ0IJWNzH9zIZF7li91kh2w==} @@ -3151,9 +2585,6 @@ packages: json-schema-traverse@1.0.0: resolution: {integrity: sha512-NM8/P9n3XjXhIZn1lLhkFaACTOURQXjWhV4BA/RnOv8xvgqtqpAX9IO4mRQxSx1Rlo4tqzeqb0sOlruaOy3dug==} - json-stringify-safe@5.0.1: - resolution: {integrity: sha512-ZClg6AaYvamvYEE82d3Iyd3vSSIjQ+odgjaTzRuO3s7toCdFKczob2i0zCh7JE8kWn17yvAWhUVxvqGwUalsRA==} - json5@2.2.3: resolution: {integrity: sha512-XmOWe7eyHYH14cLdVPoyg+GOH3rYX++KpzrylJwSW98t3Nk+U8XOl8FWKOgwtzdb8lXGf6zYwDUzeHMWfxasyg==} engines: {node: '>=6'} @@ -3184,6 +2615,9 @@ packages: koffi@2.9.0: resolution: {integrity: sha512-KCsuJ2gM58n6bNdR2Z7gqsh/3TchxxQFbVgax2/UvAjRTgwNSYAJDx9E3jrkBP4jEDHWRCfE47Y2OG+/fiSvEw==} + kuler@2.0.0: + resolution: {integrity: sha512-Xq9nH7KlWZmXAtodXDDRE7vs6DU1gTU8zYDHDiWLSip45Egwq3plLHzPn27NgvzL2r1LMPC1vdqh98sQxtqj4A==} + langchain@0.2.8: resolution: {integrity: sha512-kb2IOMA71xH8e6EXFg0l4S+QSMC/c796pj1+7mPBkR91HHwoyHZhFRrBaZv4tV+Td+Ba91J2uEDBmySklZLpNQ==} engines: {node: '>=18'} @@ -3378,24 +2812,12 @@ packages: resolution: {integrity: sha512-t7hw9pI+WvuwNJXwk5zVHpyhIqzg2qTlklJOf0mVxGSbe3Fp2VieZcduNYjaLDoy6p9uGpQEGWG87WpMKlNq8g==} engines: {node: '>=8'} - lodash.camelcase@4.3.0: - resolution: {integrity: sha512-TwuEnCnxbc3rAvhf/LbG7tJUDzhqXyFnv3dtzLOPgCG/hODL7WFnsbwktkD7yUV0RrreP/l1PALq/YSg6VvjlA==} - lodash.defaults@4.2.0: resolution: {integrity: sha512-qjxPLHd3r5DnsdGacqOMU6pb/avJzdh9tFX2ymgoZE27BmjXrNy/y4LoaiTeAb+O3gL8AfpJGtqfX/ae2leYYQ==} lodash.isarguments@3.1.0: resolution: {integrity: sha512-chi4NHZlZqZD18a0imDHnZPrDeBbTtVN7GXMwuGdRH9qotxAjYs3aVLKc7zNOG9eddR5Ksd8rvFEBc9SsggPpg==} - lodash.isobject@3.0.2: - resolution: {integrity: sha512-3/Qptq2vr7WeJbB4KHUSKlq8Pl7ASXi3UG6CMbBm8WRtXi8+GHm7mKaU3urfpSEzWe2wCIChs6/sdocUsTKJiA==} - - lodash.isplainobject@4.0.6: - resolution: {integrity: sha512-oSXzaWypCMHkPC3NvBEaPHf0KsA5mvPrOPgQWDsbg8n7orZ290M0BmC/jgRZ4vcJ6DTAhjrsSYgdsW/F+MFOBA==} - - lodash.isstring@4.0.1: - resolution: {integrity: sha512-0wJxfxH1wgO3GrbuP+dTTk7op+6L41QCXbGINEmD+ny/G/eCqGzxyCsh7159S+mgDDcoarnBw6PC1PS5+wUGgw==} - lodash.memoize@4.1.2: resolution: {integrity: sha512-t7j+NzmgnQzTAYXcsHYLgimltOV1MXHtlOWf6GjL9Kj8GK5FInw5JotxvbOs+IvV1/Dzo04/fCGfLVs7aXb4Ag==} @@ -3408,14 +2830,14 @@ packages: lodash@4.17.21: resolution: {integrity: sha512-v2kDEe57lecTulaDIuNTPy3Ry4gLGJ6Z1O3vE1krgXZNrsQ+LFTGHVxVjcXPs17LhbZVGedAJv8XZ1tvj5FvSg==} - log-symbols@4.1.0: - resolution: {integrity: sha512-8XPvpAA8uyhfteu8pIvQxpJZ7SYYdpUivZpGy6sFsBuKRY/7rQGavedeB8aK+Zkyq6upMFVL/9AW6vOYzfRyLg==} - engines: {node: '>=10'} - logform@2.6.0: resolution: {integrity: sha512-1ulHeNPp6k/LD8H91o7VYFBng5i1BDE7HoKxVbZiGFidS1Rj65qcywLxX+pVfAPoQJEjRdvKcusKwOupHCVOVQ==} engines: {node: '>= 12.0.0'} + logform@2.6.1: + resolution: {integrity: sha512-CdaO738xRapbKIMVn2m4F6KTj4j7ooJ8POVnebSgKo3KBz5axNXRAL7ZdRjIV6NOr2Uf4vjtRkxrFETOioCqSA==} + engines: {node: '>= 12.0.0'} + loglevel@1.9.1: resolution: {integrity: sha512-hP3I3kCrDIMuRwAwHltphhDM1r8i55H33GgqjXbrisuJhF4kRhW1dNuxsRklp4bXl8DSdLaNLuiL4A/LWRfxvg==} engines: {node: '>= 0.6.0'} @@ -3423,9 +2845,6 @@ packages: logsnag@1.0.0: resolution: {integrity: sha512-HMzjh75OR5EVY7Be4Rw8TcDTIY5UPsrXF1HvQ6EzDi21x5cQcDzi4Ts0Y/ruPCbxKY2KG17YjeeTzErXFewFBg==} - long@5.2.3: - resolution: {integrity: sha512-lcHwpNoggQTObv5apGNCTdJrO69eHOZMi4BNC+rTLER8iHAqGrUVeLh/irVIM7zTw2bOXA8T6uNPeujwOLg/2Q==} - loose-envify@1.4.0: resolution: {integrity: sha512-lyuxPGr/Wfhrlem2CL/UcnUc1zcqKAImBDzukY7Y5F/yQiNdko6+fRLevlw1HgMySw7f611UIY408EtxRSoK3Q==} hasBin: true @@ -3467,6 +2886,11 @@ packages: engines: {node: '>=12.0.0'} hasBin: true + marked@14.1.2: + resolution: {integrity: sha512-f3r0yqpz31VXiDB/wj9GaOB0a2PRLQl6vJmXiFrniNwjkKdvakqJRULhjFKJpxOchlCRiG5fcacoUZY5Xa6PEQ==} + engines: {node: '>= 18'} + hasBin: true + md5@2.3.0: resolution: {integrity: sha512-T1GITYmFaKuO91vxyoQMFETst+O71VUPEU3ze5GNzDm0OWdP8v1ziTaAEPUr/3kLsY3Sftgz242A1SetQiDL7g==} @@ -3736,14 +3160,13 @@ packages: once@1.4.0: resolution: {integrity: sha512-lNaJgI+2Q5URQBkccEKHTQOPaXdUxnZZElQTZY0MFUAuaEqe1E+Nyvgdz/aIyNi6Z9MzO5dv1H8n58/GELp3+w==} + one-time@1.0.0: + resolution: {integrity: sha512-5DXOiRKwuSEcQ/l0kGCF6Q3jcADFv5tSmRaJck/OqkVFcOzutB134KRSfF0xDrL39MNnqxbHBbUUcjZIhTgb2g==} + onetime@5.1.2: resolution: {integrity: sha512-kbpaSSGJTWdAY5KPVeMOKXSrPtr8C8C7wodJbcsd51jRnmD+GZu8Y0VoU6Dm5Z4vWr0Ig/1NKuWRKf7j5aaYSg==} engines: {node: '>=6'} - open@8.4.2: - resolution: {integrity: sha512-7x81NCL719oNbsq/3mh+hVrAWmFuEYUqrq/Iw3kUzH8ReypT9QQ0BLoJS7/G9k6N81XjW4qHWtjWwe/9eLy1EQ==} - engines: {node: '>=12'} - openai@3.3.0: resolution: {integrity: sha512-uqxI/Au+aPRnsaQRe8CojU0eCR7I0mBiKjD3sNMzY6DaC1ZVrc85u98mtJW6voDug8fgGN+DIZmTDxTthxb7dQ==} @@ -3772,10 +3195,6 @@ packages: option@0.2.4: resolution: {integrity: sha512-pkEqbDyl8ou5cpq+VsnQbe/WlEy5qS7xPzMS1U55OCG9KPvwFD46zDbxQIj3egJSFc3D+XhYOPUzz49zQAVy7A==} - ora@5.4.1: - resolution: {integrity: sha512-5b6Y85tPxZZ7QytO+BQzysW31HJku27cRIlkbAXaNx+BdcVi+LlRFmVXzeF6a7JCwJpyw5c4b+YSVImQIrBpuQ==} - engines: {node: '>=10'} - p-finally@1.0.0: resolution: {integrity: sha512-LICb2p9CB7FS+0eR1oqWnHhp0FljGLZCWBE9aix0Uye9W8LTQPwMTYVGWQWIw9RdQiDg4+epXQODwIYJtSJaow==} engines: {node: '>=4'} @@ -3919,9 +3338,6 @@ packages: resolution: {integrity: sha512-JU3teHTNjmE2VCGFzuY8EXzCDVwEqB2a8fsIvwaStHhAWJEeVd1o1QD80CU6+ZdEXXSLbSsuLwJjkCBWqRQUVA==} engines: {node: '>=8.6'} - pino-abstract-transport@1.2.0: - resolution: {integrity: sha512-Guhh8EZfPCfH+PMXAb6rKOjGQEoy0xlAIn+irODG5kgfYV+BQ0rGYYWTIel3P5mmyXqkYkPmdIkywsn6QKUR1Q==} - pirates@4.0.6: resolution: {integrity: sha512-saLsH7WeYYPiD25LDuLRRY/i+6HaPYr6G1OUlN39otzkSTxKnubR9RTxS3/Kk50s1g2JTgFwWQDQyplC5/SHZg==} engines: {node: '>= 6'} @@ -3986,10 +3402,6 @@ packages: proto-list@1.2.4: resolution: {integrity: sha512-vtK/94akxsTMhe0/cbfpR+syPuszcuwhqVjJq26CuNDgFGj682oRBXOP5MJpv2r7JtE8MsiepGIqvvOTBwn2vA==} - protobufjs@7.3.2: - resolution: {integrity: sha512-RXyHaACeqXeqAKGLDl68rQKbmObRsTIn4TYVUUug1KfS47YWCo5MacGITEryugIgZqORCvJWEk4l449POg5Txg==} - engines: {node: '>=12.0.0'} - proxy-addr@2.0.7: resolution: {integrity: sha512-llQsMLSUDUPT44jdrU/O37qlnifitDP+ZwrmmZcoSKyLKvtZxpyV0n2/bD/N4tBAAZ/gJEdZU7KMraoK1+XYAg==} engines: {node: '>= 0.10'} @@ -4139,10 +3551,6 @@ packages: resolution: {integrity: sha512-oKWePCxqpd6FlLvGV1VU0x7bkPmmCNolxzjMf4NczoDnQcIWrAF+cPtZn5i6n+RfD2d9i0tzpKnG6Yk168yIyw==} hasBin: true - restore-cursor@3.1.0: - resolution: {integrity: sha512-l+sSefzHpj5qimhFSE5a8nufZYAM3sBSVMAPtYkmC+4EH2anSGaEMXSD0izRQbu9nfyQ9y5JrVmp7E8oZrUjvA==} - engines: {node: '>=8'} - retry@0.13.1: resolution: {integrity: sha512-XQBQ3I8W1Cge0Seh+6gjj03LbmRFWuoszgK9ooCpwYIrhhoO80pfq4cUkU5DkknwfOfFteRwlZ56PYOGYyFWdg==} engines: {node: '>= 4'} @@ -4213,10 +3621,6 @@ packages: resolution: {integrity: sha512-qqWzuOjSFOuqPjFe4NOsMLafToQQwBSOEpS+FwEt3A2V3vKubTquT3vmLTQpFgMXp8AlFWFuP1qKaJZOtPpVXg==} engines: {node: '>= 0.8.0'} - serialize-error@8.1.0: - resolution: {integrity: sha512-3NnuWfM6vBYoy5gZFvHiYsVbafvI9vZv/+jlIigFn4oP4zjNPK3LhcY0xSCgeb1a5L8jO71Mit9LlNoi2UfDDQ==} - engines: {node: '>=10'} - serve-static@1.15.0: resolution: {integrity: sha512-XGuRDNjXUijsUL0vl6nSD7cwURuzEgglbOaFuZM9g3kwDXOWVTck0jLzjPzGD+TazWbboZYu52/9/XPdUgne9g==} engines: {node: '>= 0.8.0'} @@ -4259,6 +3663,9 @@ packages: resolution: {integrity: sha512-bzyZ1e88w9O1iNJbKnOlvYTrWPDl46O1bG0D3XInv+9tkPrxrN8jUUTiFlDkkmKWgn1M6CfIA13SuGqOa9Korw==} engines: {node: '>=14'} + simple-swizzle@0.2.2: + resolution: {integrity: sha512-JA//kQgZtbuY83m+xT+tXJkmJncGMTFT+C+g2h2R9uxkYIrE2yy9sgmcLhCnw57/WSD+Eh3J97FPEDFnbXnDUg==} + simple-update-notifier@1.1.0: resolution: {integrity: sha512-VpsrsJSUcJEseSbMHkrsrAVSdvVS5I96Qo1QAQ4FxQ9wXFcB+pjj7FB7/us9+GcgfW4ziHtYMc1J0PLczb55mg==} engines: {node: '>=8.10.0'} @@ -4422,6 +3829,9 @@ packages: text-decoder@1.1.0: resolution: {integrity: sha512-TmLJNj6UgX8xcUZo4UDStGQtDiTzF7BzWlzn9g7UWrjkpHr5uJTK1ld16wZ3LXb2vb6jH8qU89dW5whuMdXYdw==} + text-hex@1.0.0: + resolution: {integrity: sha512-uuVGNWzgJ4yhRaNSiubPY7OjISw4sw4E5Uv0wbjp+OzcbmVU/rsT8ujgcXJhn9ypzsgr5vlzpPqP+MBBKcGvbg==} + through@2.3.8: resolution: {integrity: sha512-w89qg7PI8wAdvX60bMDP+bFoD5Dvhm9oLheFp5O4a2QF0cSBGsBX4qZmadPMvVqlLJBBci+WqGGOAPvcDeNSVg==} @@ -4509,10 +3919,6 @@ packages: resolution: {integrity: sha512-0fr/mIH1dlO+x7TlcMy+bIDqKPsw/70tVyeHW787goQjhmqaZe10uwLujubK9q9Lg6Fiho1KUKDYz0Z7k7g5/g==} engines: {node: '>=4'} - type-fest@0.20.2: - resolution: {integrity: sha512-Ne+eE4r0/iWnpAxD852z3A+N0Bt5RN//NjJwRd2VFHEmrywxf5vsZlh4R6lixl6B+wz/8d+maTSAkN1FIkI3LQ==} - engines: {node: '>=10'} - type-fest@0.21.3: resolution: {integrity: sha512-t0rzBq87m3fVcduHDUFhKmyyX+9eo6WQjZvf51Ea/M0Q7+T374Jp1aUiyUl0GKxp8M/OETVHSDvmkyPgvX+X2w==} engines: {node: '>=10'} @@ -4615,9 +4021,6 @@ packages: walker@1.0.8: resolution: {integrity: sha512-ts/8E8l5b7kY0vlWLewOkDXMmPdLcVV4GmOQLyxuSswIJsweeFZtAsMF7k1Nszz+TYBQrlYRmzOnr398y1JemQ==} - wcwidth@1.0.1: - resolution: {integrity: sha512-XHPEwS0q6TaxcvG85+8EYkbiCux2XtWG2mkc47Ng2A77BQu9+DqIOJldST4HgPkuea7dvKSj5VgX3P1d4rW8Tg==} - web-streams-polyfill@3.3.3: resolution: {integrity: sha512-d2JWLCivmZYTSIoge9MsgFCZrt571BikcWGYkjC1khllbTeDlGqZ2D8vD8E/lJa8WGWbb7Plm8/XJYV7IJHZZw==} engines: {node: '>= 8'} @@ -4652,8 +4055,12 @@ packages: engines: {node: '>= 8'} hasBin: true - winston-transport@4.7.0: - resolution: {integrity: sha512-ajBj65K5I7denzer2IYW6+2bNIVqLGDHqDw3Ow8Ohh+vdW+rv4MZ6eiDvHoKhfJFZ2auyN8byXieDDJ96ViONg==} + winston-transport@4.8.0: + resolution: {integrity: sha512-qxSTKswC6llEMZKgCQdaWgDuMJQnhuvF5f2Nk3SNXc4byfQ+voo2mX1Px9dkNOuR8p0KAjfPG29PuYUSIb+vSA==} + engines: {node: '>= 12.0.0'} + + winston@3.14.2: + resolution: {integrity: sha512-CO8cdpBB2yqzEf8v895L+GNKYJiEq8eKlHU38af3snQBQ+sdAIUepjMSguOIJC7ICbzm0ZI+Af2If4vIJrtmOg==} engines: {node: '>= 12.0.0'} wordnet-db@3.1.14: @@ -5038,9 +4445,15 @@ snapshots: dependencies: '@jridgewell/trace-mapping': 0.3.9 + '@dabh/diagnostics@2.0.3': + dependencies: + colorspace: 1.1.4 + enabled: 2.0.0 + kuler: 2.0.0 + '@devil7softwares/pos@1.0.2': {} - '@dqbd/tiktoken@1.0.15': {} + '@dqbd/tiktoken@1.0.16': {} '@flydotio/dockerfile@0.4.11': dependencies: @@ -5050,83 +4463,6 @@ snapshots: shell-quote: 1.8.1 yargs: 17.7.2 - '@grpc/grpc-js@1.10.10': - dependencies: - '@grpc/proto-loader': 0.7.13 - '@js-sdsl/ordered-map': 4.4.2 - - '@grpc/proto-loader@0.7.13': - dependencies: - lodash.camelcase: 4.3.0 - long: 5.2.3 - protobufjs: 7.3.2 - yargs: 17.7.2 - - '@hyperdx/instrumentation-exception@0.1.0(@opentelemetry/api@1.9.0)': - dependencies: - '@hyperdx/instrumentation-sentry-node': 0.1.0(@opentelemetry/api@1.9.0) - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - '@sentry/core': 8.13.0 - '@sentry/types': 8.13.0 - '@sentry/utils': 8.13.0 - json-stringify-safe: 5.0.1 - shimmer: 1.2.1 - tslib: 2.6.3 - transitivePeerDependencies: - - supports-color - - '@hyperdx/instrumentation-sentry-node@0.1.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - json-stringify-safe: 5.0.1 - shimmer: 1.2.1 - tslib: 2.6.3 - transitivePeerDependencies: - - supports-color - - '@hyperdx/node-opentelemetry@0.8.1': - dependencies: - '@hyperdx/instrumentation-exception': 0.1.0(@opentelemetry/api@1.9.0) - '@hyperdx/instrumentation-sentry-node': 0.1.0(@opentelemetry/api@1.9.0) - '@opentelemetry/api': 1.9.0 - '@opentelemetry/api-logs': 0.51.1 - '@opentelemetry/auto-instrumentations-node': 0.46.1(@opentelemetry/api@1.9.0) - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/exporter-logs-otlp-http': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/exporter-metrics-otlp-proto': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/exporter-trace-otlp-proto': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-http': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-runtime-node': 0.4.0(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-logs': 0.51.1(@opentelemetry/api-logs@0.51.1)(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-metrics': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-node': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-trace-base': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - cli-spinners: 2.9.2 - json-stringify-safe: 5.0.1 - lodash.isobject: 3.0.2 - lodash.isplainobject: 4.0.6 - lodash.isstring: 4.0.1 - node-fetch: 2.7.0 - open: 8.4.2 - ora: 5.4.1 - pino-abstract-transport: 1.2.0 - semver: 7.6.2 - shimmer: 1.2.1 - tslib: 2.6.3 - winston-transport: 4.7.0 - transitivePeerDependencies: - - encoding - - supports-color - '@ioredis/commands@1.2.0': {} '@isaacs/cliui@8.0.2': @@ -5336,8 +4672,6 @@ snapshots: '@jridgewell/resolve-uri': 3.1.2 '@jridgewell/sourcemap-codec': 1.4.15 - '@js-sdsl/ordered-map@4.4.2': {} - '@langchain/core@0.2.12(langchain@0.2.8(@supabase/supabase-js@2.44.2)(axios@1.7.2)(cheerio@1.0.0-rc.12)(handlebars@4.7.8)(html-to-text@9.0.5)(ioredis@5.4.1)(mammoth@1.7.2)(mongodb@6.6.2(socks@2.8.3))(openai@4.57.0(zod@3.23.8))(pdf-parse@1.1.1)(puppeteer@22.12.1(typescript@5.4.5))(redis@4.6.14)(ws@8.18.0))(openai@4.57.0(zod@3.23.8))': dependencies: ansi-styles: 5.2.0 @@ -5375,39 +4709,12 @@ snapshots: - langchain - openai - '@logtail/core@0.4.21': - dependencies: - '@logtail/tools': 0.4.21 - '@logtail/types': 0.4.20 - serialize-error: 8.1.0 - - '@logtail/node@0.4.21': - dependencies: - '@logtail/core': 0.4.21 - '@logtail/types': 0.4.20 - '@msgpack/msgpack': 2.8.0 - '@types/stack-trace': 0.0.29 - cross-fetch: 3.1.8 - minimatch: 3.1.2 - serialize-error: 8.1.0 - stack-trace: 0.0.10 - transitivePeerDependencies: - - encoding - - '@logtail/tools@0.4.21': - dependencies: - '@logtail/types': 0.4.20 - - '@logtail/types@0.4.20': {} - '@mixmark-io/domino@2.2.0': {} '@mongodb-js/saslprep@1.1.7': dependencies: sparse-bitfield: 3.0.3 - '@msgpack/msgpack@2.8.0': {} - '@msgpackr-extract/msgpackr-extract-darwin-arm64@3.0.3': optional: true @@ -5434,210 +4741,21 @@ snapshots: '@one-ini/wasm@0.1.1': {} - '@opentelemetry/api-logs@0.51.1': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/api-logs@0.52.1': dependencies: '@opentelemetry/api': 1.9.0 '@opentelemetry/api@1.9.0': {} - '@opentelemetry/auto-instrumentations-node@0.46.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-amqplib': 0.37.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-aws-lambda': 0.41.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-aws-sdk': 0.41.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-bunyan': 0.38.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-cassandra-driver': 0.38.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-connect': 0.36.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-cucumber': 0.6.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-dataloader': 0.9.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-dns': 0.36.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-express': 0.39.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-fastify': 0.36.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-fs': 0.12.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-generic-pool': 0.36.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-graphql': 0.40.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-grpc': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-hapi': 0.38.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-http': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-ioredis': 0.40.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-knex': 0.36.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-koa': 0.40.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-lru-memoizer': 0.37.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-memcached': 0.36.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-mongodb': 0.43.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-mongoose': 0.38.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-mysql': 0.38.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-mysql2': 0.38.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-nestjs-core': 0.37.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-net': 0.36.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-pg': 0.41.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-pino': 0.39.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-redis': 0.39.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-redis-4': 0.39.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-restify': 0.38.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-router': 0.37.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-socket.io': 0.39.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-tedious': 0.10.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-undici': 0.2.0(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation-winston': 0.37.0(@opentelemetry/api@1.9.0) - '@opentelemetry/resource-detector-alibaba-cloud': 0.28.10(@opentelemetry/api@1.9.0) - '@opentelemetry/resource-detector-aws': 1.5.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resource-detector-azure': 0.2.9(@opentelemetry/api@1.9.0) - '@opentelemetry/resource-detector-container': 0.3.11(@opentelemetry/api@1.9.0) - '@opentelemetry/resource-detector-gcp': 0.29.10(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-node': 0.51.1(@opentelemetry/api@1.9.0) - transitivePeerDependencies: - - encoding - - supports-color - - '@opentelemetry/context-async-hooks@1.24.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/context-async-hooks@1.25.1(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 - '@opentelemetry/core@1.24.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/semantic-conventions': 1.24.1 - '@opentelemetry/core@1.25.1(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 '@opentelemetry/semantic-conventions': 1.25.1 - '@opentelemetry/exporter-logs-otlp-http@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/api-logs': 0.51.1 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-exporter-base': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-transformer': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-logs': 0.51.1(@opentelemetry/api-logs@0.51.1)(@opentelemetry/api@1.9.0) - - '@opentelemetry/exporter-metrics-otlp-http@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-exporter-base': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-transformer': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-metrics': 1.24.1(@opentelemetry/api@1.9.0) - - '@opentelemetry/exporter-metrics-otlp-proto@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/exporter-metrics-otlp-http': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-exporter-base': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-proto-exporter-base': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-transformer': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-metrics': 1.24.1(@opentelemetry/api@1.9.0) - - '@opentelemetry/exporter-trace-otlp-grpc@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@grpc/grpc-js': 1.10.10 - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-grpc-exporter-base': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-transformer': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-trace-base': 1.24.1(@opentelemetry/api@1.9.0) - - '@opentelemetry/exporter-trace-otlp-http@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-exporter-base': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-transformer': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-trace-base': 1.24.1(@opentelemetry/api@1.9.0) - - '@opentelemetry/exporter-trace-otlp-proto@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-exporter-base': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-proto-exporter-base': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-transformer': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-trace-base': 1.24.1(@opentelemetry/api@1.9.0) - - '@opentelemetry/exporter-zipkin@1.24.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-trace-base': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.24.1 - - '@opentelemetry/instrumentation-amqplib@0.37.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-aws-lambda@0.41.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/propagator-aws-xray': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - '@types/aws-lambda': 8.10.122 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-aws-sdk@0.41.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/propagation-utils': 0.30.10(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-bunyan@0.38.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/api-logs': 0.51.1 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@types/bunyan': 1.8.9 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-cassandra-driver@0.38.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-connect@0.36.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - '@types/connect': 3.4.36 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-connect@0.38.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5648,39 +4766,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-cucumber@0.6.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-dataloader@0.9.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-dns@0.36.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - semver: 7.6.2 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-express@0.39.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-express@0.41.1(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5690,15 +4775,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-fastify@0.36.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-fastify@0.38.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5708,14 +4784,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-fs@0.12.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-fs@0.14.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5724,21 +4792,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-generic-pool@0.36.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-graphql@0.40.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-graphql@0.42.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5746,23 +4799,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-grpc@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.24.1 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-hapi@0.38.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-hapi@0.40.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5772,16 +4808,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-http@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.24.1 - semver: 7.6.2 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-http@0.52.1(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5792,15 +4818,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-ioredis@0.40.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/redis-common': 0.36.2 - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-ioredis@0.42.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5810,25 +4827,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-knex@0.36.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-koa@0.40.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - '@types/koa': 2.14.0 - '@types/koa__router': 12.0.3 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-koa@0.42.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5838,31 +4836,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-lru-memoizer@0.37.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-memcached@0.36.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - '@types/memcached': 2.2.10 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-mongodb@0.43.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-metrics': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-mongodb@0.46.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5872,15 +4845,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-mongoose@0.38.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-mongoose@0.40.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5890,15 +4854,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-mysql2@0.38.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - '@opentelemetry/sql-common': 0.40.1(@opentelemetry/api@1.9.0) - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-mysql2@0.40.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5908,15 +4863,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-mysql@0.38.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - '@types/mysql': 2.15.22 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-mysql@0.40.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5926,14 +4872,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-nestjs-core@0.37.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-nestjs-core@0.39.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5942,25 +4880,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-net@0.36.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-pg@0.41.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - '@opentelemetry/sql-common': 0.40.1(@opentelemetry/api@1.9.0) - '@types/pg': 8.6.1 - '@types/pg-pool': 2.0.4 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-pg@0.43.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5972,22 +4891,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-pino@0.39.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-redis-4@0.39.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/redis-common': 0.36.2 - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation-redis-4@0.41.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -5997,72 +4900,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/instrumentation-redis@0.39.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/redis-common': 0.36.2 - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-restify@0.38.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-router@0.37.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-runtime-node@0.4.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-socket.io@0.39.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-tedious@0.10.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - '@types/tedious': 4.0.14 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-undici@0.2.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - transitivePeerDependencies: - - supports-color - - '@opentelemetry/instrumentation-winston@0.37.0(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/api-logs': 0.51.1 - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation@0.46.0(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -6075,18 +4912,6 @@ snapshots: - supports-color optional: true - '@opentelemetry/instrumentation@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/api-logs': 0.51.1 - '@types/shimmer': 1.0.5 - import-in-the-middle: 1.7.4 - require-in-the-middle: 7.3.0 - semver: 7.6.2 - shimmer: 1.2.1 - transitivePeerDependencies: - - supports-color - '@opentelemetry/instrumentation@0.52.1(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -6099,119 +4924,14 @@ snapshots: transitivePeerDependencies: - supports-color - '@opentelemetry/otlp-exporter-base@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - - '@opentelemetry/otlp-grpc-exporter-base@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@grpc/grpc-js': 1.10.10 - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-exporter-base': 0.51.1(@opentelemetry/api@1.9.0) - protobufjs: 7.3.2 - - '@opentelemetry/otlp-proto-exporter-base@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/otlp-exporter-base': 0.51.1(@opentelemetry/api@1.9.0) - protobufjs: 7.3.2 - - '@opentelemetry/otlp-transformer@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/api-logs': 0.51.1 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-logs': 0.51.1(@opentelemetry/api-logs@0.51.1)(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-metrics': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-trace-base': 1.24.1(@opentelemetry/api@1.9.0) - - '@opentelemetry/propagation-utils@0.30.10(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - - '@opentelemetry/propagator-aws-xray@1.25.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - - '@opentelemetry/propagator-b3@1.24.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - - '@opentelemetry/propagator-jaeger@1.24.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/redis-common@0.36.2': {} - '@opentelemetry/resource-detector-alibaba-cloud@0.28.10(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/resources': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - - '@opentelemetry/resource-detector-aws@1.5.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - - '@opentelemetry/resource-detector-azure@0.2.9(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/resources': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - - '@opentelemetry/resource-detector-container@0.3.11(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/resources': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - - '@opentelemetry/resource-detector-gcp@0.29.10(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.25.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.25.1 - gcp-metadata: 6.1.0 - transitivePeerDependencies: - - encoding - - supports-color - - '@opentelemetry/resources@1.24.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.24.1 - '@opentelemetry/resources@1.25.1(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 '@opentelemetry/core': 1.25.1(@opentelemetry/api@1.9.0) '@opentelemetry/semantic-conventions': 1.25.1 - '@opentelemetry/sdk-logs@0.51.1(@opentelemetry/api-logs@0.51.1)(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/api-logs': 0.51.1 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.24.1(@opentelemetry/api@1.9.0) - - '@opentelemetry/sdk-metrics@1.24.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.24.1(@opentelemetry/api@1.9.0) - lodash.merge: 4.6.2 - '@opentelemetry/sdk-metrics@1.25.1(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -6219,32 +4939,6 @@ snapshots: '@opentelemetry/resources': 1.25.1(@opentelemetry/api@1.9.0) lodash.merge: 4.6.2 - '@opentelemetry/sdk-node@0.51.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/api-logs': 0.51.1 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/exporter-trace-otlp-grpc': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/exporter-trace-otlp-http': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/exporter-trace-otlp-proto': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/exporter-zipkin': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/instrumentation': 0.51.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-logs': 0.51.1(@opentelemetry/api-logs@0.51.1)(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-metrics': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-trace-base': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-trace-node': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.24.1 - transitivePeerDependencies: - - supports-color - - '@opentelemetry/sdk-trace-base@1.24.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/resources': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/semantic-conventions': 1.24.1 - '@opentelemetry/sdk-trace-base@1.25.1(@opentelemetry/api@1.9.0)': dependencies: '@opentelemetry/api': 1.9.0 @@ -6252,18 +4946,6 @@ snapshots: '@opentelemetry/resources': 1.25.1(@opentelemetry/api@1.9.0) '@opentelemetry/semantic-conventions': 1.25.1 - '@opentelemetry/sdk-trace-node@1.24.1(@opentelemetry/api@1.9.0)': - dependencies: - '@opentelemetry/api': 1.9.0 - '@opentelemetry/context-async-hooks': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/core': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/propagator-b3': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/propagator-jaeger': 1.24.1(@opentelemetry/api@1.9.0) - '@opentelemetry/sdk-trace-base': 1.24.1(@opentelemetry/api@1.9.0) - semver: 7.6.2 - - '@opentelemetry/semantic-conventions@1.24.1': {} - '@opentelemetry/semantic-conventions@1.25.1': {} '@opentelemetry/sql-common@0.40.1(@opentelemetry/api@1.9.0)': @@ -6290,29 +4972,6 @@ snapshots: transitivePeerDependencies: - supports-color - '@protobufjs/aspromise@1.1.2': {} - - '@protobufjs/base64@1.1.2': {} - - '@protobufjs/codegen@2.0.4': {} - - '@protobufjs/eventemitter@1.1.0': {} - - '@protobufjs/fetch@1.1.0': - dependencies: - '@protobufjs/aspromise': 1.1.2 - '@protobufjs/inquire': 1.1.0 - - '@protobufjs/float@1.0.2': {} - - '@protobufjs/inquire@1.1.0': {} - - '@protobufjs/path@1.1.2': {} - - '@protobufjs/pool@1.1.0': {} - - '@protobufjs/utf8@1.1.0': {} - '@puppeteer/browsers@2.2.3': dependencies: debug: 4.3.4 @@ -6405,11 +5064,6 @@ snapshots: - encoding - supports-color - '@sentry/core@8.13.0': - dependencies: - '@sentry/types': 8.13.0 - '@sentry/utils': 8.13.0 - '@sentry/core@8.26.0': dependencies: '@sentry/types': 8.26.0 @@ -6473,14 +5127,8 @@ snapshots: transitivePeerDependencies: - supports-color - '@sentry/types@8.13.0': {} - '@sentry/types@8.26.0': {} - '@sentry/utils@8.13.0': - dependencies: - '@sentry/types': 8.13.0 - '@sentry/utils@8.26.0': dependencies: '@sentry/types': 8.26.0 @@ -6631,12 +5279,6 @@ snapshots: '@tsconfig/recommended@1.0.6': {} - '@types/accepts@1.3.7': - dependencies: - '@types/node': 20.14.1 - - '@types/aws-lambda@8.10.122': {} - '@types/babel__core@7.20.5': dependencies: '@babel/parser': 7.24.6 @@ -6663,10 +5305,6 @@ snapshots: '@types/connect': 3.4.38 '@types/node': 20.14.1 - '@types/bunyan@1.8.9': - dependencies: - '@types/node': 20.14.1 - '@types/connect@3.4.36': dependencies: '@types/node': 20.14.1 @@ -6675,19 +5313,12 @@ snapshots: dependencies: '@types/node': 20.14.1 - '@types/content-disposition@0.5.8': {} - - '@types/cookies@0.9.0': - dependencies: - '@types/connect': 3.4.38 - '@types/express': 4.17.21 - '@types/keygrip': 1.0.6 - '@types/node': 20.14.1 - '@types/cors@2.8.17': dependencies: '@types/node': 20.14.1 + '@types/escape-html@1.0.4': {} + '@types/express-serve-static-core@4.19.3': dependencies: '@types/node': 20.14.1 @@ -6712,8 +5343,6 @@ snapshots: dependencies: '@types/node': 20.14.1 - '@types/http-assert@1.5.5': {} - '@types/http-errors@2.0.4': {} '@types/istanbul-lib-coverage@2.0.6': {} @@ -6731,31 +5360,6 @@ snapshots: expect: 29.7.0 pretty-format: 29.7.0 - '@types/keygrip@1.0.6': {} - - '@types/koa-compose@3.2.8': - dependencies: - '@types/koa': 2.14.0 - - '@types/koa@2.14.0': - dependencies: - '@types/accepts': 1.3.7 - '@types/content-disposition': 0.5.8 - '@types/cookies': 0.9.0 - '@types/http-assert': 1.5.5 - '@types/http-errors': 2.0.4 - '@types/keygrip': 1.0.6 - '@types/koa-compose': 3.2.8 - '@types/node': 20.14.1 - - '@types/koa__router@12.0.3': - dependencies: - '@types/koa': 2.14.0 - - '@types/memcached@2.2.10': - dependencies: - '@types/node': 20.14.1 - '@types/mime@1.3.5': {} '@types/mysql@2.15.22': @@ -6775,6 +5379,8 @@ snapshots: dependencies: undici-types: 5.26.5 + '@types/pdf-parse@1.1.4': {} + '@types/pg-pool@2.0.4': dependencies: '@types/pg': 8.6.1 @@ -6806,14 +5412,8 @@ snapshots: '@types/shimmer@1.0.5': {} - '@types/stack-trace@0.0.29': {} - '@types/stack-utils@2.0.3': {} - '@types/tedious@4.0.14': - dependencies: - '@types/node': 20.14.1 - '@types/triple-beam@1.3.5': {} '@types/uuid@9.0.8': {} @@ -7064,8 +5664,6 @@ snapshots: basic-ftp@5.0.5: {} - bignumber.js@9.1.2: {} - bin-links@4.0.4: dependencies: cmd-shim: 6.0.3 @@ -7077,12 +5675,6 @@ snapshots: binary-search@1.3.6: {} - bl@4.1.0: - dependencies: - buffer: 5.7.1 - inherits: 2.0.4 - readable-stream: 3.6.2 - bluebird@3.4.7: {} body-parser@1.20.2: @@ -7243,20 +5835,12 @@ snapshots: cjs-module-lexer@1.3.1: {} - cli-cursor@3.1.0: - dependencies: - restore-cursor: 3.1.0 - - cli-spinners@2.9.2: {} - cliui@8.0.1: dependencies: string-width: 4.2.3 strip-ansi: 6.0.1 wrap-ansi: 7.0.0 - clone@1.0.4: {} - cluster-key-slot@1.1.2: {} cmd-shim@6.0.3: {} @@ -7279,6 +5863,21 @@ snapshots: color-name@1.1.4: {} + color-string@1.9.1: + dependencies: + color-name: 1.1.4 + simple-swizzle: 0.2.2 + + color@3.2.1: + dependencies: + color-convert: 1.9.3 + color-string: 1.9.1 + + colorspace@1.1.4: + dependencies: + color: 3.2.1 + text-hex: 1.0.0 + combined-stream@1.0.8: dependencies: delayed-stream: 1.0.0 @@ -7407,18 +6006,12 @@ snapshots: deepmerge@4.3.1: {} - defaults@1.0.4: - dependencies: - clone: 1.0.4 - define-data-property@1.1.4: dependencies: es-define-property: 1.0.0 es-errors: 1.3.0 gopd: 1.0.1 - define-lazy-prop@2.0.0: {} - degenerator@5.0.1: dependencies: ast-types: 0.13.4 @@ -7508,6 +6101,8 @@ snapshots: emoji-regex@9.2.2: {} + enabled@2.0.0: {} + encodeurl@1.0.2: {} end-of-stream@1.4.4: @@ -7628,8 +6223,6 @@ snapshots: transitivePeerDependencies: - supports-color - extend@3.0.2: {} - extract-zip@2.0.1: dependencies: debug: 4.3.4 @@ -7692,6 +6285,8 @@ snapshots: flat@5.0.2: {} + fn.name@1.1.0: {} + follow-redirects@1.15.6: {} foreground-child@3.2.1: @@ -7740,25 +6335,6 @@ snapshots: function-bind@1.1.2: {} - gaxios@6.7.0: - dependencies: - extend: 3.0.2 - https-proxy-agent: 7.0.5 - is-stream: 2.0.1 - node-fetch: 2.7.0 - uuid: 10.0.0 - transitivePeerDependencies: - - encoding - - supports-color - - gcp-metadata@6.1.0: - dependencies: - gaxios: 6.7.0 - json-bigint: 1.0.0 - transitivePeerDependencies: - - encoding - - supports-color - generic-pool@3.9.0: {} gensync@1.0.0-beta.2: {} @@ -7978,13 +6554,6 @@ snapshots: module-details-from-path: 1.0.3 optional: true - import-in-the-middle@1.7.4: - dependencies: - acorn: 8.12.0 - acorn-import-attributes: 1.9.5(acorn@8.12.0) - cjs-module-lexer: 1.3.1 - module-details-from-path: 1.0.3 - import-local@3.1.0: dependencies: pkg-dir: 4.2.0 @@ -8026,6 +6595,8 @@ snapshots: is-arrayish@0.2.1: {} + is-arrayish@0.3.2: {} + is-binary-path@2.1.0: dependencies: binary-extensions: 2.3.0 @@ -8036,8 +6607,6 @@ snapshots: dependencies: hasown: 2.0.2 - is-docker@2.2.1: {} - is-extglob@2.1.1: {} is-fullwidth-code-point@3.0.0: {} @@ -8048,8 +6617,6 @@ snapshots: dependencies: is-extglob: 2.1.1 - is-interactive@1.0.0: {} - is-number@7.0.0: {} is-plain-obj@2.1.0: {} @@ -8058,12 +6625,6 @@ snapshots: is-stream@2.0.1: {} - is-unicode-supported@0.1.0: {} - - is-wsl@2.2.0: - dependencies: - is-docker: 2.2.1 - isarray@1.0.0: {} isexe@2.0.0: {} @@ -8476,18 +7037,12 @@ snapshots: jsesc@2.5.2: {} - json-bigint@1.0.0: - dependencies: - bignumber.js: 9.1.2 - json-parse-even-better-errors@2.3.1: {} json-schema-to-zod@2.3.0: {} json-schema-traverse@1.0.0: {} - json-stringify-safe@5.0.1: {} - json5@2.2.3: {} jsonfile@6.1.0: @@ -8513,6 +7068,8 @@ snapshots: koffi@2.9.0: {} + kuler@2.0.0: {} + langchain@0.2.8(@supabase/supabase-js@2.44.2)(axios@1.7.2)(cheerio@1.0.0-rc.12)(handlebars@4.7.8)(html-to-text@9.0.5)(ioredis@5.4.1)(mammoth@1.7.2)(mongodb@6.6.2(socks@2.8.3))(openai@4.57.0(zod@3.23.8))(pdf-parse@1.1.1)(puppeteer@22.12.1(typescript@5.4.5))(redis@4.6.14)(ws@8.18.0): dependencies: '@langchain/core': 0.2.12(langchain@0.2.8(@supabase/supabase-js@2.44.2)(axios@1.7.2)(cheerio@1.0.0-rc.12)(handlebars@4.7.8)(html-to-text@9.0.5)(ioredis@5.4.1)(mammoth@1.7.2)(mongodb@6.6.2(socks@2.8.3))(openai@4.57.0(zod@3.23.8))(pdf-parse@1.1.1)(puppeteer@22.12.1(typescript@5.4.5))(redis@4.6.14)(ws@8.18.0))(openai@4.57.0(zod@3.23.8)) @@ -8579,18 +7136,10 @@ snapshots: dependencies: p-locate: 4.1.0 - lodash.camelcase@4.3.0: {} - lodash.defaults@4.2.0: {} lodash.isarguments@3.1.0: {} - lodash.isobject@3.0.2: {} - - lodash.isplainobject@4.0.6: {} - - lodash.isstring@4.0.1: {} - lodash.memoize@4.1.2: {} lodash.merge@4.6.2: {} @@ -8599,11 +7148,6 @@ snapshots: lodash@4.17.21: {} - log-symbols@4.1.0: - dependencies: - chalk: 4.1.2 - is-unicode-supported: 0.1.0 - logform@2.6.0: dependencies: '@colors/colors': 1.6.0 @@ -8613,6 +7157,15 @@ snapshots: safe-stable-stringify: 2.4.3 triple-beam: 1.4.1 + logform@2.6.1: + dependencies: + '@colors/colors': 1.6.0 + '@types/triple-beam': 1.3.5 + fecha: 4.2.3 + ms: 2.1.3 + safe-stable-stringify: 2.4.3 + triple-beam: 1.4.1 + loglevel@1.9.1: {} logsnag@1.0.0: @@ -8621,8 +7174,6 @@ snapshots: transitivePeerDependencies: - encoding - long@5.2.3: {} - loose-envify@1.4.0: dependencies: js-tokens: 4.0.0 @@ -8670,6 +7221,8 @@ snapshots: underscore: 1.13.6 xmlbuilder: 10.1.1 + marked@14.1.2: {} + md5@2.3.0: dependencies: charenc: 0.0.2 @@ -8937,16 +7490,14 @@ snapshots: dependencies: wrappy: 1.0.2 + one-time@1.0.0: + dependencies: + fn.name: 1.1.0 + onetime@5.1.2: dependencies: mimic-fn: 2.1.0 - open@8.4.2: - dependencies: - define-lazy-prop: 2.0.0 - is-docker: 2.2.1 - is-wsl: 2.2.0 - openai@3.3.0: dependencies: axios: 0.26.1 @@ -8985,18 +7536,6 @@ snapshots: option@0.2.4: {} - ora@5.4.1: - dependencies: - bl: 4.1.0 - chalk: 4.1.2 - cli-cursor: 3.1.0 - cli-spinners: 2.9.2 - is-interactive: 1.0.0 - is-unicode-supported: 0.1.0 - log-symbols: 4.1.0 - strip-ansi: 6.0.1 - wcwidth: 1.0.1 - p-finally@1.0.0: {} p-limit@2.3.0: @@ -9148,11 +7687,6 @@ snapshots: picomatch@2.3.1: {} - pino-abstract-transport@1.2.0: - dependencies: - readable-stream: 4.5.2 - split2: 4.2.0 - pirates@4.0.6: {} pkg-dir@4.2.0: @@ -9221,21 +7755,6 @@ snapshots: proto-list@1.2.4: {} - protobufjs@7.3.2: - dependencies: - '@protobufjs/aspromise': 1.1.2 - '@protobufjs/base64': 1.1.2 - '@protobufjs/codegen': 2.0.4 - '@protobufjs/eventemitter': 1.1.0 - '@protobufjs/fetch': 1.1.0 - '@protobufjs/float': 1.0.2 - '@protobufjs/inquire': 1.1.0 - '@protobufjs/path': 1.1.2 - '@protobufjs/pool': 1.1.0 - '@protobufjs/utf8': 1.1.0 - '@types/node': 20.14.1 - long: 5.2.3 - proxy-addr@2.0.7: dependencies: forwarded: 0.2.0 @@ -9417,11 +7936,6 @@ snapshots: path-parse: 1.0.7 supports-preserve-symlinks-flag: 1.0.0 - restore-cursor@3.1.0: - dependencies: - onetime: 5.1.2 - signal-exit: 3.0.7 - retry@0.13.1: {} rimraf@5.0.7: @@ -9490,10 +8004,6 @@ snapshots: transitivePeerDependencies: - supports-color - serialize-error@8.1.0: - dependencies: - type-fest: 0.20.2 - serve-static@1.15.0: dependencies: encodeurl: 1.0.2 @@ -9539,6 +8049,10 @@ snapshots: signal-exit@4.1.0: {} + simple-swizzle@0.2.2: + dependencies: + is-arrayish: 0.3.2 + simple-update-notifier@1.1.0: dependencies: semver: 7.0.0 @@ -9725,6 +8239,8 @@ snapshots: dependencies: b4a: 1.6.6 + text-hex@1.0.0: {} + through@2.3.8: {} tmpl@1.0.5: {} @@ -9795,8 +8311,6 @@ snapshots: type-detect@4.0.8: {} - type-fest@0.20.2: {} - type-fest@0.21.3: {} type-is@1.6.18: @@ -9880,10 +8394,6 @@ snapshots: dependencies: makeerror: 1.0.12 - wcwidth@1.0.1: - dependencies: - defaults: 1.0.4 - web-streams-polyfill@3.3.3: {} web-streams-polyfill@4.0.0-beta.3: {} @@ -9912,12 +8422,26 @@ snapshots: dependencies: isexe: 2.0.0 - winston-transport@4.7.0: + winston-transport@4.8.0: dependencies: - logform: 2.6.0 - readable-stream: 3.6.2 + logform: 2.6.1 + readable-stream: 4.5.2 triple-beam: 1.4.1 + winston@3.14.2: + dependencies: + '@colors/colors': 1.6.0 + '@dabh/diagnostics': 2.0.3 + async: 3.2.5 + is-stream: 2.0.1 + logform: 2.6.0 + one-time: 1.0.0 + readable-stream: 3.6.2 + safe-stable-stringify: 2.4.3 + stack-trace: 0.0.10 + triple-beam: 1.4.1 + winston-transport: 4.8.0 + wordnet-db@3.1.14: {} wordpos@2.1.0: diff --git a/apps/api/requests.http b/apps/api/requests.http index 3e7bd2b7..809bae7b 100644 --- a/apps/api/requests.http +++ b/apps/api/requests.http @@ -1,15 +1,15 @@ ### Crawl Website POST http://localhost:3002/v0/scrape HTTP/1.1 -Authorization: Bearer fc- +Authorization: Bearer {{$dotenv TEST_API_KEY}} content-type: application/json { - "url":"corterix.com" + "url":"firecrawl.dev" } ### Check Job Status GET http://localhost:3002/v1/crawl/1dd0f924-a36f-4b96-94ea-32ed954dac67 HTTP/1.1 -Authorization: Bearer fc- +Authorization: Bearer {{$dotenv TEST_API_KEY}} ### Check Job Status @@ -18,7 +18,7 @@ GET http://localhost:3002/v0/jobs/active HTTP/1.1 ### Scrape Website POST http://localhost:3002/v0/crawl HTTP/1.1 -Authorization: Bearer fc- +Authorization: Bearer {{$dotenv TEST_API_KEY}} content-type: application/json { @@ -45,7 +45,7 @@ content-type: application/json ### Scrape Website POST http://localhost:3002/v0/scrape HTTP/1.1 -Authorization: Bearer +Authorization: Bearer {{$dotenv TEST_API_KEY}} content-type: application/json { @@ -56,12 +56,12 @@ content-type: application/json ### Check Job Status GET http://localhost:3002/v0/crawl/status/a6053912-d602-4709-841f-3d2cb46fea0a HTTP/1.1 -Authorization: Bearer +Authorization: Bearer {{$dotenv TEST_API_KEY}} ### Get Job Result POST https://api.firecrawl.dev/v0/crawl HTTP/1.1 -Authorization: Bearer +Authorization: Bearer {{$dotenv TEST_API_KEY}} content-type: application/json { @@ -70,7 +70,7 @@ content-type: application/json ### Check Job Status GET https://api.firecrawl.dev/v0/crawl/status/cfcb71ac-23a3-4da5-bd85-d4e58b871d66 -Authorization: Bearer +Authorization: Bearer {{$dotenv TEST_API_KEY}} ### Get Active Jobs Count GET http://localhost:3002/serverHealthCheck diff --git a/apps/api/sharedLibs/go-html-to-md/.gitignore b/apps/api/sharedLibs/go-html-to-md/.gitignore new file mode 100644 index 00000000..bdab47c6 --- /dev/null +++ b/apps/api/sharedLibs/go-html-to-md/.gitignore @@ -0,0 +1,2 @@ +html-to-markdown.so +html-to-markdown.h \ No newline at end of file diff --git a/apps/api/src/__tests__/e2e_full_withAuth/index.test.ts b/apps/api/src/__tests__/e2e_full_withAuth/index.test.ts index b1708abc..dec77131 100644 --- a/apps/api/src/__tests__/e2e_full_withAuth/index.test.ts +++ b/apps/api/src/__tests__/e2e_full_withAuth/index.test.ts @@ -844,7 +844,7 @@ describe("E2E Tests for API Routes", () => { expect(crawlInitResponse.statusCode).toBe(200); expect(crawlInitResponse.body).toHaveProperty("jobId"); - let crawlStatus: string; + let crawlStatus: string = "scraping"; let crawlData = []; while (crawlStatus !== "completed") { const statusResponse = await request(TEST_URL) diff --git a/apps/api/src/__tests__/e2e_noAuth/index.test.ts b/apps/api/src/__tests__/e2e_noAuth/index.test.ts index acb22780..83f676b8 100644 --- a/apps/api/src/__tests__/e2e_noAuth/index.test.ts +++ b/apps/api/src/__tests__/e2e_noAuth/index.test.ts @@ -20,7 +20,6 @@ describe("E2E Tests for API Routes with No Authentication", () => { process.env.SCRAPING_BEE_API_KEY = ""; process.env.OPENAI_API_KEY = ""; process.env.BULL_AUTH_KEY = ""; - process.env.LOGTAIL_KEY = ""; process.env.PLAYWRIGHT_MICROSERVICE_URL = ""; process.env.LLAMAPARSE_API_KEY = ""; process.env.TEST_API_KEY = ""; diff --git a/apps/api/src/__tests__/e2e_v1_withAuth/index.test.ts b/apps/api/src/__tests__/e2e_v1_withAuth/index.test.ts index a4163472..8c3d1731 100644 --- a/apps/api/src/__tests__/e2e_v1_withAuth/index.test.ts +++ b/apps/api/src/__tests__/e2e_v1_withAuth/index.test.ts @@ -1,7 +1,7 @@ import request from "supertest"; import { configDotenv } from "dotenv"; import { - ScrapeRequest, + ScrapeRequestInput, ScrapeResponseRequestTest, } from "../../controllers/v1/types"; @@ -44,7 +44,7 @@ describe("E2E Tests for v1 API Routes", () => { }); it.concurrent("should throw error for blocklisted URL", async () => { - const scrapeRequest: ScrapeRequest = { + const scrapeRequest: ScrapeRequestInput = { url: "https://facebook.com/fake-test", }; @@ -73,7 +73,7 @@ describe("E2E Tests for v1 API Routes", () => { it.concurrent( "should return a successful response with a valid API key", async () => { - const scrapeRequest: ScrapeRequest = { + const scrapeRequest: ScrapeRequestInput = { url: "https://roastmywebsite.ai", }; @@ -125,7 +125,7 @@ describe("E2E Tests for v1 API Routes", () => { it.concurrent( "should return a successful response with a valid API key", async () => { - const scrapeRequest: ScrapeRequest = { + const scrapeRequest: ScrapeRequestInput = { url: "https://arxiv.org/abs/2410.04840", }; @@ -167,7 +167,7 @@ describe("E2E Tests for v1 API Routes", () => { it.concurrent( "should return a successful response with a valid API key and includeHtml set to true", async () => { - const scrapeRequest: ScrapeRequest = { + const scrapeRequest: ScrapeRequestInput = { url: "https://roastmywebsite.ai", formats: ["markdown", "html"], }; @@ -194,7 +194,7 @@ describe("E2E Tests for v1 API Routes", () => { 30000 ); it.concurrent('should return a successful response for a valid scrape with PDF file', async () => { - const scrapeRequest: ScrapeRequest = { + const scrapeRequest: ScrapeRequestInput = { url: "https://arxiv.org/pdf/astro-ph/9301001.pdf" // formats: ["markdown", "html"], }; @@ -217,7 +217,7 @@ describe("E2E Tests for v1 API Routes", () => { }, 60000); it.concurrent('should return a successful response for a valid scrape with PDF file without explicit .pdf extension', async () => { - const scrapeRequest: ScrapeRequest = { + const scrapeRequest: ScrapeRequestInput = { url: "https://arxiv.org/pdf/astro-ph/9301001" }; const response: ScrapeResponseRequestTest = await request(TEST_URL) @@ -240,7 +240,7 @@ describe("E2E Tests for v1 API Routes", () => { }, 60000); it.concurrent("should return a successful response with a valid API key with removeTags option", async () => { - const scrapeRequest: ScrapeRequest = { + const scrapeRequest: ScrapeRequestInput = { url: "https://www.scrapethissite.com/", onlyMainContent: false // default is true }; @@ -261,7 +261,7 @@ describe("E2E Tests for v1 API Routes", () => { expect(responseWithoutRemoveTags.body.data.markdown).toContain("[FAQ](/faq/)"); // .nav expect(responseWithoutRemoveTags.body.data.markdown).toContain("Hartley Brody 2023"); // #footer - const scrapeRequestWithRemoveTags: ScrapeRequest = { + const scrapeRequestWithRemoveTags: ScrapeRequestInput = { url: "https://www.scrapethissite.com/", excludeTags: ['.nav', '#footer', 'strong'], onlyMainContent: false // default is true @@ -407,7 +407,7 @@ describe("E2E Tests for v1 API Routes", () => { it.concurrent( "should return a successful response with a valid API key and includeHtml set to true", async () => { - const scrapeRequest: ScrapeRequest = { + const scrapeRequest: ScrapeRequestInput = { url: "https://roastmywebsite.ai", formats: ["html","rawHtml"], }; @@ -438,7 +438,7 @@ describe("E2E Tests for v1 API Routes", () => { it.concurrent( "should return a successful response with waitFor", async () => { - const scrapeRequest: ScrapeRequest = { + const scrapeRequest: ScrapeRequestInput = { url: "https://ycombinator.com/companies", formats: ["markdown"], waitFor: 8000 @@ -471,7 +471,7 @@ describe("E2E Tests for v1 API Routes", () => { it.concurrent( "should return a successful response with a valid links on page", async () => { - const scrapeRequest: ScrapeRequest = { + const scrapeRequest: ScrapeRequestInput = { url: "https://roastmywebsite.ai", formats: ["links"], }; @@ -672,7 +672,7 @@ describe("POST /v1/crawl", () => { }); it.concurrent("should throw error for blocklisted URL", async () => { - const scrapeRequest: ScrapeRequest = { + const scrapeRequest: ScrapeRequestInput = { url: "https://facebook.com/fake-test", }; diff --git a/apps/api/src/__tests__/e2e_v1_withAuth_all_params/index.test.ts b/apps/api/src/__tests__/e2e_v1_withAuth_all_params/index.test.ts new file mode 100644 index 00000000..5c7feb1f --- /dev/null +++ b/apps/api/src/__tests__/e2e_v1_withAuth_all_params/index.test.ts @@ -0,0 +1,603 @@ +import request from "supertest"; +import { configDotenv } from "dotenv"; +import { + ScrapeRequest, + ScrapeResponseRequestTest, +} from "../../controllers/v1/types"; + +configDotenv(); +const FIRECRAWL_API_URL = "http://127.0.0.1:3002"; +const E2E_TEST_SERVER_URL = "http://firecrawl-e2e-test.vercel.app"; // @rafaelsideguide/firecrawl-e2e-test + +describe("E2E Tests for v1 API Routes", () => { + + it.concurrent('should return a successful response for a scrape with 403 page', async () => { + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post('/v1/scrape') + .set('Authorization', `Bearer ${process.env.TEST_API_KEY}`) + .set('Content-Type', 'application/json') + .send({ url: 'https://httpstat.us/403' }); + + expect(response.statusCode).toBe(200); + expect(response.body).toHaveProperty('data'); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + expect(response.body.data).toHaveProperty('markdown'); + expect(response.body.data).toHaveProperty('metadata'); + expect(response.body.data.metadata.statusCode).toBe(403); + }, 30000); + + it.concurrent("should handle 'formats:markdown (default)' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + expect(response.body).toHaveProperty("data"); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + + expect(response.body.data).toHaveProperty("markdown"); + + expect(response.body.data.markdown).toContain("This page is used for end-to-end (e2e) testing with Firecrawl."); + expect(response.body.data.markdown).toContain("Content with id #content-1"); + // expect(response.body.data.markdown).toContain("Loading..."); + expect(response.body.data.markdown).toContain("Click me!"); + expect(response.body.data.markdown).toContain("Power your AI apps with clean data crawled from any website. It's also open-source."); // firecrawl.dev inside an iframe + expect(response.body.data.markdown).toContain("This content loads only when you see it. Don't blink! 👼"); // the browser always scroll to the bottom + expect(response.body.data.markdown).not.toContain("Header"); // Only main content is returned by default + expect(response.body.data.markdown).not.toContain("footer"); // Only main content is returned by default + expect(response.body.data.markdown).not.toContain("This content is only visible on mobile"); + }, + 30000); + + it.concurrent("should handle 'formats:html' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + formats: ["html"] + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + expect(response.body).toHaveProperty("data"); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + + + expect(response.body.data).not.toHaveProperty("markdown"); + expect(response.body.data).toHaveProperty("html"); + + expect(response.body.data.html).not.toContain("
Header
"); + expect(response.body.data.html).toContain("

This page is used for end-to-end (e2e) testing with Firecrawl.

"); + }, + 30000); + + it.concurrent("should handle 'rawHtml' in 'formats' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + formats: ["rawHtml"] + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + expect(response.body).toHaveProperty("data"); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + + expect(response.body.data).not.toHaveProperty("markdown"); + expect(response.body.data).toHaveProperty("rawHtml"); + + expect(response.body.data.rawHtml).toContain(">This page is used for end-to-end (e2e) testing with Firecrawl.

"); + expect(response.body.data.rawHtml).toContain(">Header"); + }, + 30000); + + // - TODO: tests for links + // - TODO: tests for screenshot + // - TODO: tests for screenshot@fullPage + + it.concurrent("should handle 'headers' parameter correctly", async () => { + // @ts-ignore + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + headers: { "e2e-header-test": "firecrawl" } + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + expect(response.body).toHaveProperty("data"); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + + expect(response.body.data.markdown).toContain("e2e-header-test: firecrawl"); + }, 30000); + + it.concurrent("should handle 'includeTags' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + includeTags: ['#content-1'] + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + expect(response.body).toHaveProperty("data"); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + + expect(response.body.data.markdown).not.toContain("

This page is used for end-to-end (e2e) testing with Firecrawl.

"); + expect(response.body.data.markdown).toContain("Content with id #content-1"); + }, + 30000); + + it.concurrent("should handle 'excludeTags' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + excludeTags: ['#content-1'] + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + expect(response.body).toHaveProperty("data"); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + + expect(response.body.data.markdown).toContain("This page is used for end-to-end (e2e) testing with Firecrawl."); + expect(response.body.data.markdown).not.toContain("Content with id #content-1"); + }, + 30000); + + it.concurrent("should handle 'onlyMainContent' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + formats: ["html", "markdown"], + onlyMainContent: false + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + expect(response.body).toHaveProperty("data"); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + + expect(response.body.data.markdown).toContain("This page is used for end-to-end (e2e) testing with Firecrawl."); + expect(response.body.data.html).toContain("
Header
"); + }, + 30000); + + it.concurrent("should handle 'timeout' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + timeout: 500 + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(408); + + if (!("error" in response.body)) { + throw new Error("Expected response body to have 'error' property"); + } + expect(response.body.error).toBe("Request timed out"); + expect(response.body.success).toBe(false); + }, 30000); + + + it.concurrent("should handle 'mobile' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + mobile: true + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + expect(response.body.data.markdown).toContain("This content is only visible on mobile"); + }, + 30000); + + it.concurrent("should handle 'parsePDF' parameter correctly", + async () => { + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send({ url: 'https://arxiv.org/pdf/astro-ph/9301001.pdf'}); + await new Promise((r) => setTimeout(r, 6000)); + + expect(response.statusCode).toBe(200); + expect(response.body).toHaveProperty('data'); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + + expect(response.body.data.markdown).toContain('arXiv:astro-ph/9301001v1 7 Jan 1993'); + expect(response.body.data.markdown).not.toContain('h7uKu14adDL6yGfnGf2qycY5uq8kC3OKCWkPxm'); + + const responseNoParsePDF: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send({ url: 'https://arxiv.org/pdf/astro-ph/9301001.pdf', parsePDF: false }); + await new Promise((r) => setTimeout(r, 6000)); + + expect(responseNoParsePDF.statusCode).toBe(200); + expect(responseNoParsePDF.body).toHaveProperty('data'); + if (!("data" in responseNoParsePDF.body)) { + throw new Error("Expected response body to have 'data' property"); + } + expect(responseNoParsePDF.body.data.markdown).toContain('h7uKu14adDL6yGfnGf2qycY5uq8kC3OKCWkPxm'); + }, + 30000); + + // it.concurrent("should handle 'location' parameter correctly", + // async () => { + // const scrapeRequest: ScrapeRequest = { + // url: "https://roastmywebsite.ai", + // location: { + // country: "US", + // languages: ["en"] + // } + // }; + + // const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + // .post("/v1/scrape") + // .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + // .set("Content-Type", "application/json") + // .send(scrapeRequest); + + // expect(response.statusCode).toBe(200); + // // Add assertions to verify location is handled correctly + // }, + // 30000); + + it.concurrent("should handle 'skipTlsVerification' parameter correctly", + async () => { + const scrapeRequest = { + url: "https://expired.badssl.com/", + timeout: 120000 + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + console.log("Error1a") + // console.log(response.body) + expect(response.statusCode).toBe(200); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + expect(response.body.data.metadata.pageStatusCode).toBe(500); + console.log("Error?") + + const scrapeRequestWithSkipTlsVerification = { + url: "https://expired.badssl.com/", + skipTlsVerification: true, + timeout: 120000 + + } as ScrapeRequest; + + const responseWithSkipTlsVerification: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequestWithSkipTlsVerification); + + console.log("Error1b") + // console.log(responseWithSkipTlsVerification.body) + expect(responseWithSkipTlsVerification.statusCode).toBe(200); + if (!("data" in responseWithSkipTlsVerification.body)) { + throw new Error("Expected response body to have 'data' property"); + } + // console.log(responseWithSkipTlsVerification.body.data) + expect(responseWithSkipTlsVerification.body.data.markdown).toContain("badssl.com"); + }, + 60000); + + it.concurrent("should handle 'removeBase64Images' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + removeBase64Images: true + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + // console.log(response.body.data.markdown) + // - TODO: not working for every image + // expect(response.body.data.markdown).toContain("Image-Removed"); + }, + 30000); + + it.concurrent("should handle 'action wait' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + actions: [{ + type: "wait", + milliseconds: 10000 + }] + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + expect(response.body.data.markdown).not.toContain("Loading..."); + expect(response.body.data.markdown).toContain("Content loaded after 5 seconds!"); + }, + 30000); + + // screenshot + it.concurrent("should handle 'action screenshot' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + actions: [{ + type: "screenshot" + }] + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + if (!response.body.data.actions?.screenshots) { + throw new Error("Expected response body to have screenshots array"); + } + expect(response.body.data.actions.screenshots[0].length).toBeGreaterThan(0); + expect(response.body.data.actions.screenshots[0]).toContain("https://service.firecrawl.dev/storage/v1/object/public/media/screenshot-"); + + // TODO compare screenshot with expected screenshot + }, + 30000); + + it.concurrent("should handle 'action screenshot@fullPage' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + actions: [{ + type: "screenshot", + fullPage: true + }, + { + type:"scrape" + }] + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + // console.log(response.body.data.actions?.screenshots[0]) + if (!response.body.data.actions?.screenshots) { + throw new Error("Expected response body to have screenshots array"); + } + expect(response.body.data.actions.screenshots[0].length).toBeGreaterThan(0); + expect(response.body.data.actions.screenshots[0]).toContain("https://service.firecrawl.dev/storage/v1/object/public/media/screenshot-"); + + if (!response.body.data.actions?.scrapes) { + throw new Error("Expected response body to have scrapes array"); + } + expect(response.body.data.actions.scrapes[0].url).toBe("https://firecrawl-e2e-test.vercel.app/"); + expect(response.body.data.actions.scrapes[0].html).toContain("This page is used for end-to-end (e2e) testing with Firecrawl.

"); + // TODO compare screenshot with expected full page screenshot + }, + 30000); + + it.concurrent("should handle 'action click' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + actions: [{ + type: "click", + selector: "#click-me" + }] + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + expect(response.body.data.markdown).not.toContain("Click me!"); + expect(response.body.data.markdown).toContain("Text changed after click!"); + }, + 30000); + + it.concurrent("should handle 'action write' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + formats: ["html"], + actions: [{ + type: "click", + selector: "#input-1" + }, + { + type: "write", + text: "Hello, world!" + } + ]} as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + expect(response.statusCode).toBe(200); + if (!("data" in response.body)) { + throw new Error("Expected response body to have 'data' property"); + } + + // TODO: fix this test (need to fix fire-engine first) + // uncomment the following line: + // expect(response.body.data.html).toContain(""); + }, + 30000); + + // TODO: fix this test (need to fix fire-engine first) + it.concurrent("should handle 'action pressKey' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + formats: ["markdown"], + actions: [ + { + type: "press", + key: "ArrowDown" + } + ] + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + // // TODO: fix this test (need to fix fire-engine first) + // // right now response.body is: { success: false, error: '(Internal server error) - null' } + // expect(response.statusCode).toBe(200); + // if (!("data" in response.body)) { + // throw new Error("Expected response body to have 'data' property"); + // } + // expect(response.body.data.markdown).toContain("Last Key Clicked: ArrowDown") + }, + 30000); + + // TODO: fix this test (need to fix fire-engine first) + it.concurrent("should handle 'action scroll' parameter correctly", + async () => { + const scrapeRequest = { + url: E2E_TEST_SERVER_URL, + formats: ["markdown"], + actions: [ + { + type: "click", + selector: "#scroll-bottom-loader" + }, + { + type: "scroll", + direction: "down", + amount: 2000 + } + ] + } as ScrapeRequest; + + const response: ScrapeResponseRequestTest = await request(FIRECRAWL_API_URL) + .post("/v1/scrape") + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .set("Content-Type", "application/json") + .send(scrapeRequest); + + // TODO: uncomment this tests + // expect(response.statusCode).toBe(200); + // if (!("data" in response.body)) { + // throw new Error("Expected response body to have 'data' property"); + // } + // + // expect(response.body.data.markdown).toContain("You have reached the bottom!") + }, + 30000); + + // TODO: test scrape action + +}); \ No newline at end of file diff --git a/apps/api/src/__tests__/e2e_withAuth/index.test.ts b/apps/api/src/__tests__/e2e_withAuth/index.test.ts index 26caf63e..1646843f 100644 --- a/apps/api/src/__tests__/e2e_withAuth/index.test.ts +++ b/apps/api/src/__tests__/e2e_withAuth/index.test.ts @@ -776,7 +776,8 @@ describe("E2E Tests for v0 API Routes", () => { await new Promise((r) => setTimeout(r, 10000)); const completedResponse = await request(TEST_URL) .get(`/v0/crawl/status/${crawlResponse.body.jobId}`) - .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`); + .set("Authorization", `Bearer ${process.env.TEST_API_KEY}`) + .maxResponseSize(4000000000); expect(completedResponse.statusCode).toBe(200); expect(completedResponse.body).toHaveProperty("status"); diff --git a/apps/api/src/controllers/auth.ts b/apps/api/src/controllers/auth.ts index c9f693c5..de74fed0 100644 --- a/apps/api/src/controllers/auth.ts +++ b/apps/api/src/controllers/auth.ts @@ -9,9 +9,8 @@ import { import { supabase_service } from "../services/supabase"; import { withAuth } from "../lib/withAuth"; import { RateLimiterRedis } from "rate-limiter-flexible"; -import { setTraceAttributes } from "@hyperdx/node-opentelemetry"; import { sendNotification } from "../services/notification/email_notification"; -import { Logger } from "../lib/logger"; +import { logger } from "../lib/logger"; import { redlock } from "../services/redlock"; import { deleteKey, getValue } from "../services/redis"; import { setValue } from "../services/redis"; @@ -40,8 +39,8 @@ function normalizedApiIsUuid(potentialUuid: string): boolean { export async function setCachedACUC( api_key: string, acuc: - | AuthCreditUsageChunk - | ((acuc: AuthCreditUsageChunk) => AuthCreditUsageChunk) + | AuthCreditUsageChunk | null + | ((acuc: AuthCreditUsageChunk) => AuthCreditUsageChunk | null) ) { const cacheKeyACUC = `acuc_${api_key}`; const redLockKey = `lock_${cacheKeyACUC}`; @@ -49,7 +48,7 @@ export async function setCachedACUC( try { await redlock.using([redLockKey], 10000, {}, async (signal) => { if (typeof acuc === "function") { - acuc = acuc(JSON.parse(await getValue(cacheKeyACUC))); + acuc = acuc(JSON.parse(await getValue(cacheKeyACUC) ?? "null")); if (acuc === null) { if (signal.aborted) { @@ -69,7 +68,7 @@ export async function setCachedACUC( await setValue(cacheKeyACUC, JSON.stringify(acuc), 600, true); }); } catch (error) { - Logger.error(`Error updating cached ACUC ${cacheKeyACUC}: ${error}`); + logger.error(`Error updating cached ACUC ${cacheKeyACUC}: ${error}`); } } @@ -103,7 +102,7 @@ export async function getACUC( break; } - Logger.warn( + logger.warn( `Failed to retrieve authentication and credit usage data after ${retries}, trying again...` ); retries++; @@ -146,33 +145,14 @@ export async function authenticateUser( res, mode?: RateLimiterMode ): Promise { - return withAuth(supaAuthenticateUser)(req, res, mode); -} - -function setTrace(team_id: string, api_key: string) { - try { - setTraceAttributes({ - team_id, - api_key, - }); - } catch (error) { - Sentry.captureException(error); - Logger.error(`Error setting trace attributes: ${error.message}`); - } + return withAuth(supaAuthenticateUser, { success: true, chunk: null, team_id: "bypass" })(req, res, mode); } export async function supaAuthenticateUser( req, res, mode?: RateLimiterMode -): Promise<{ - success: boolean; - team_id?: string; - error?: string; - status?: number; - plan?: PlanType; - chunk?: AuthCreditUsageChunk; -}> { +): Promise { const authHeader = req.headers.authorization ?? (req.headers["sec-websocket-protocol"] @@ -200,7 +180,7 @@ export async function supaAuthenticateUser( let teamId: string | null = null; let priceId: string | null = null; - let chunk: AuthCreditUsageChunk; + let chunk: AuthCreditUsageChunk | null = null; if (token == "this_is_just_a_preview_token") { if (mode == RateLimiterMode.CrawlStatus) { @@ -233,8 +213,6 @@ export async function supaAuthenticateUser( priceId = chunk.price_id; const plan = getPlanByPriceId(priceId); - // HyperDX Logging - setTrace(teamId, normalizedApi); subscriptionData = { team_id: teamId, plan, @@ -291,7 +269,7 @@ export async function supaAuthenticateUser( try { await rateLimiter.consume(team_endpoint_token); } catch (rateLimiterRes) { - Logger.error(`Rate limit exceeded: ${rateLimiterRes}`); + logger.error(`Rate limit exceeded: ${rateLimiterRes}`); const secs = Math.round(rateLimiterRes.msBeforeNext / 1000) || 1; const retryDate = new Date(Date.now() + rateLimiterRes.msBeforeNext); @@ -318,7 +296,7 @@ export async function supaAuthenticateUser( mode === RateLimiterMode.CrawlStatus || mode === RateLimiterMode.Search) ) { - return { success: true, team_id: "preview" }; + return { success: true, team_id: "preview", chunk: null }; // check the origin of the request and make sure its from firecrawl.dev // const origin = req.headers.origin; // if (origin && origin.includes("firecrawl.dev")){ @@ -333,12 +311,12 @@ export async function supaAuthenticateUser( return { success: true, - team_id: subscriptionData.team_id, - plan: (subscriptionData.plan ?? "") as PlanType, + team_id: teamId ?? undefined, + plan: (subscriptionData?.plan ?? "") as PlanType, chunk, }; } -function getPlanByPriceId(price_id: string): PlanType { +function getPlanByPriceId(price_id: string | null): PlanType { switch (price_id) { case process.env.STRIPE_PRICE_ID_STARTER: return "starter"; diff --git a/apps/api/src/controllers/v0/admin/acuc-cache-clear.ts b/apps/api/src/controllers/v0/admin/acuc-cache-clear.ts index 876ca98a..75acd60a 100644 --- a/apps/api/src/controllers/v0/admin/acuc-cache-clear.ts +++ b/apps/api/src/controllers/v0/admin/acuc-cache-clear.ts @@ -1,7 +1,7 @@ import { Request, Response } from "express"; import { supabase_service } from "../../../services/supabase"; import { clearACUC } from "../../auth"; -import { Logger } from "../../../lib/logger"; +import { logger } from "../../../lib/logger"; export async function acucCacheClearController(req: Request, res: Response) { try { @@ -12,11 +12,11 @@ export async function acucCacheClearController(req: Request, res: Response) { .select("*") .eq("team_id", team_id); - await Promise.all(keys.data.map((x) => clearACUC(x.key))); + await Promise.all((keys.data ?? []).map((x) => clearACUC(x.key))); res.json({ ok: true }); } catch (error) { - Logger.error(`Error clearing ACUC cache via API route: ${error}`); + logger.error(`Error clearing ACUC cache via API route: ${error}`); res.status(500).json({ error: "Internal server error" }); } } diff --git a/apps/api/src/controllers/v0/admin/queue.ts b/apps/api/src/controllers/v0/admin/queue.ts index 71748002..6ef8a992 100644 --- a/apps/api/src/controllers/v0/admin/queue.ts +++ b/apps/api/src/controllers/v0/admin/queue.ts @@ -1,7 +1,7 @@ import { Request, Response } from "express"; import { Job } from "bullmq"; -import { Logger } from "../../../lib/logger"; +import { logger } from "../../../lib/logger"; import { getScrapeQueue } from "../../../services/queue-service"; import { checkAlerts } from "../../../services/alerts"; import { sendSlackWebhook } from "../../../services/alerts/slack"; @@ -10,7 +10,7 @@ export async function cleanBefore24hCompleteJobsController( req: Request, res: Response ) { - Logger.info("🐂 Cleaning jobs older than 24h"); + logger.info("🐂 Cleaning jobs older than 24h"); try { const scrapeQueue = getScrapeQueue(); const batchSize = 10; @@ -31,7 +31,7 @@ export async function cleanBefore24hCompleteJobsController( ).flat(); const before24hJobs = completedJobs.filter( - (job) => job.finishedOn < Date.now() - 24 * 60 * 60 * 1000 + (job) => job.finishedOn !== undefined && job.finishedOn < Date.now() - 24 * 60 * 60 * 1000 ) || []; let count = 0; @@ -45,12 +45,12 @@ export async function cleanBefore24hCompleteJobsController( await job.remove(); count++; } catch (jobError) { - Logger.error(`🐂 Failed to remove job with ID ${job.id}: ${jobError}`); + logger.error(`🐂 Failed to remove job with ID ${job.id}: ${jobError}`); } } return res.status(200).send(`Removed ${count} completed jobs.`); } catch (error) { - Logger.error(`🐂 Failed to clean last 24h complete jobs: ${error}`); + logger.error(`🐂 Failed to clean last 24h complete jobs: ${error}`); return res.status(500).send("Failed to clean jobs"); } } @@ -60,7 +60,7 @@ export async function checkQueuesController(req: Request, res: Response) { await checkAlerts(); return res.status(200).send("Alerts initialized"); } catch (error) { - Logger.debug(`Failed to initialize alerts: ${error}`); + logger.debug(`Failed to initialize alerts: ${error}`); return res.status(500).send("Failed to initialize alerts"); } } @@ -81,7 +81,7 @@ export async function queuesController(req: Request, res: Response) { noActiveJobs, }); } catch (error) { - Logger.error(error); + logger.error(error); return res.status(500).json({ error: error.message }); } } @@ -165,7 +165,7 @@ export async function autoscalerController(req: Request, res: Response) { } if (targetMachineCount !== activeMachines) { - Logger.info( + logger.info( `🐂 Scaling from ${activeMachines} to ${targetMachineCount} - ${webScraperActive} active, ${webScraperWaiting} waiting` ); @@ -193,7 +193,7 @@ export async function autoscalerController(req: Request, res: Response) { count: activeMachines, }); } catch (error) { - Logger.error(error); + logger.error(error); return res.status(500).send("Failed to initialize autoscaler"); } } diff --git a/apps/api/src/controllers/v0/admin/redis-health.ts b/apps/api/src/controllers/v0/admin/redis-health.ts index dc58d745..dc587606 100644 --- a/apps/api/src/controllers/v0/admin/redis-health.ts +++ b/apps/api/src/controllers/v0/admin/redis-health.ts @@ -1,6 +1,6 @@ import { Request, Response } from "express"; import Redis from "ioredis"; -import { Logger } from "../../../lib/logger"; +import { logger } from "../../../lib/logger"; import { redisRateLimitClient } from "../../../services/rate-limiter"; export async function redisHealthController(req: Request, res: Response) { @@ -10,14 +10,14 @@ export async function redisHealthController(req: Request, res: Response) { return await operation(); } catch (error) { if (attempt === retries) throw error; - Logger.warn(`Attempt ${attempt} failed: ${error.message}. Retrying...`); + logger.warn(`Attempt ${attempt} failed: ${error.message}. Retrying...`); await new Promise((resolve) => setTimeout(resolve, 2000)); // Wait 2 seconds before retrying } } }; try { - const queueRedis = new Redis(process.env.REDIS_URL); + const queueRedis = new Redis(process.env.REDIS_URL!); const testKey = "test"; const testValue = "test"; @@ -29,7 +29,7 @@ export async function redisHealthController(req: Request, res: Response) { queueRedisHealth = await retryOperation(() => queueRedis.get(testKey)); await retryOperation(() => queueRedis.del(testKey)); } catch (error) { - Logger.error(`queueRedis health check failed: ${error}`); + logger.error(`queueRedis health check failed: ${error}`); queueRedisHealth = null; } @@ -42,7 +42,7 @@ export async function redisHealthController(req: Request, res: Response) { ); await retryOperation(() => redisRateLimitClient.del(testKey)); } catch (error) { - Logger.error(`redisRateLimitClient health check failed: ${error}`); + logger.error(`redisRateLimitClient health check failed: ${error}`); redisRateLimitHealth = null; } @@ -56,10 +56,10 @@ export async function redisHealthController(req: Request, res: Response) { healthStatus.queueRedis === "healthy" && healthStatus.redisRateLimitClient === "healthy" ) { - Logger.info("Both Redis instances are healthy"); + logger.info("Both Redis instances are healthy"); return res.status(200).json({ status: "healthy", details: healthStatus }); } else { - Logger.info( + logger.info( `Redis instances health check: ${JSON.stringify(healthStatus)}` ); // await sendSlackWebhook( @@ -73,7 +73,7 @@ export async function redisHealthController(req: Request, res: Response) { .json({ status: "unhealthy", details: healthStatus }); } } catch (error) { - Logger.error(`Redis health check failed: ${error}`); + logger.error(`Redis health check failed: ${error}`); // await sendSlackWebhook( // `[REDIS DOWN] Redis instances health check: ${error.message}`, // true diff --git a/apps/api/src/controllers/v0/crawl-cancel.ts b/apps/api/src/controllers/v0/crawl-cancel.ts index efcd454a..e81064f2 100644 --- a/apps/api/src/controllers/v0/crawl-cancel.ts +++ b/apps/api/src/controllers/v0/crawl-cancel.ts @@ -2,7 +2,7 @@ import { Request, Response } from "express"; import { authenticateUser } from "../auth"; import { RateLimiterMode } from "../../../src/types"; import { supabase_service } from "../../../src/services/supabase"; -import { Logger } from "../../../src/lib/logger"; +import { logger } from "../../../src/lib/logger"; import { getCrawl, saveCrawl } from "../../../src/lib/crawl-redis"; import * as Sentry from "@sentry/node"; import { configDotenv } from "dotenv"; @@ -12,15 +12,17 @@ export async function crawlCancelController(req: Request, res: Response) { try { const useDbAuthentication = process.env.USE_DB_AUTHENTICATION === 'true'; - const { success, team_id, error, status } = await authenticateUser( + const auth = await authenticateUser( req, res, RateLimiterMode.CrawlStatus ); - if (!success) { - return res.status(status).json({ error }); + if (!auth.success) { + return res.status(auth.status).json({ error: auth.error }); } + const { team_id } = auth; + const sc = await getCrawl(req.params.jobId); if (!sc) { return res.status(404).json({ error: "Job not found" }); @@ -46,7 +48,7 @@ export async function crawlCancelController(req: Request, res: Response) { sc.cancelled = true; await saveCrawl(req.params.jobId, sc); } catch (error) { - Logger.error(error); + logger.error(error); } res.json({ @@ -54,7 +56,7 @@ export async function crawlCancelController(req: Request, res: Response) { }); } catch (error) { Sentry.captureException(error); - Logger.error(error); + logger.error(error); return res.status(500).json({ error: error.message }); } } diff --git a/apps/api/src/controllers/v0/crawl-status.ts b/apps/api/src/controllers/v0/crawl-status.ts index 66522bcf..7b6e610a 100644 --- a/apps/api/src/controllers/v0/crawl-status.ts +++ b/apps/api/src/controllers/v0/crawl-status.ts @@ -2,15 +2,17 @@ import { Request, Response } from "express"; import { authenticateUser } from "../auth"; import { RateLimiterMode } from "../../../src/types"; import { getScrapeQueue } from "../../../src/services/queue-service"; -import { Logger } from "../../../src/lib/logger"; +import { logger } from "../../../src/lib/logger"; import { getCrawl, getCrawlJobs } from "../../../src/lib/crawl-redis"; import { supabaseGetJobsByCrawlId } from "../../../src/lib/supabase-jobs"; import * as Sentry from "@sentry/node"; import { configDotenv } from "dotenv"; +import { Job } from "bullmq"; +import { toLegacyDocument } from "../v1/types"; configDotenv(); export async function getJobs(crawlId: string, ids: string[]) { - const jobs = (await Promise.all(ids.map(x => getScrapeQueue().getJob(x)))).filter(x => x); + const jobs = (await Promise.all(ids.map(x => getScrapeQueue().getJob(x)))).filter(x => x) as Job[]; if (process.env.USE_DB_AUTHENTICATION === "true") { const supabaseData = await supabaseGetJobsByCrawlId(crawlId); @@ -32,15 +34,17 @@ export async function getJobs(crawlId: string, ids: string[]) { export async function crawlStatusController(req: Request, res: Response) { try { - const { success, team_id, error, status } = await authenticateUser( + const auth = await authenticateUser( req, res, RateLimiterMode.CrawlStatus ); - if (!success) { - return res.status(status).json({ error }); + if (!auth.success) { + return res.status(auth.status).json({ error: auth.error }); } + const { team_id } = auth; + const sc = await getCrawl(req.params.jobId); if (!sc) { return res.status(404).json({ error: "Job not found" }); @@ -90,12 +94,12 @@ export async function crawlStatusController(req: Request, res: Response) { status: jobStatus, current: jobStatuses.filter(x => x === "completed" || x === "failed").length, total: jobs.length, - data: jobStatus === "completed" ? data : null, - partial_data: jobStatus === "completed" ? [] : data.filter(x => x !== null), + data: jobStatus === "completed" ? data.map(x => toLegacyDocument(x, sc.internalOptions)) : null, + partial_data: jobStatus === "completed" ? [] : data.filter(x => x !== null).map(x => toLegacyDocument(x, sc.internalOptions)), }); } catch (error) { Sentry.captureException(error); - Logger.error(error); + logger.error(error); return res.status(500).json({ error: error.message }); } } diff --git a/apps/api/src/controllers/v0/crawl.ts b/apps/api/src/controllers/v0/crawl.ts index fb412147..d502d142 100644 --- a/apps/api/src/controllers/v0/crawl.ts +++ b/apps/api/src/controllers/v0/crawl.ts @@ -9,24 +9,28 @@ import { validateIdempotencyKey } from "../../../src/services/idempotency/valida import { createIdempotencyKey } from "../../../src/services/idempotency/create"; import { defaultCrawlPageOptions, defaultCrawlerOptions, defaultOrigin } from "../../../src/lib/default-values"; import { v4 as uuidv4 } from "uuid"; -import { Logger } from "../../../src/lib/logger"; +import { logger } from "../../../src/lib/logger"; import { addCrawlJob, addCrawlJobs, crawlToCrawler, lockURL, lockURLs, saveCrawl, StoredCrawl } from "../../../src/lib/crawl-redis"; import { getScrapeQueue } from "../../../src/services/queue-service"; import { checkAndUpdateURL } from "../../../src/lib/validateUrl"; import * as Sentry from "@sentry/node"; import { getJobPriority } from "../../lib/job-priority"; +import { fromLegacyScrapeOptions, url as urlSchema } from "../v1/types"; +import { ZodError } from "zod"; export async function crawlController(req: Request, res: Response) { try { - const { success, team_id, error, status, plan, chunk } = await authenticateUser( + const auth = await authenticateUser( req, res, RateLimiterMode.Crawl ); - if (!success) { - return res.status(status).json({ error }); + if (!auth.success) { + return res.status(auth.status).json({ error: auth.error }); } + const { team_id, plan, chunk } = auth; + if (req.headers["x-idempotency-key"]) { const isIdempotencyValid = await validateIdempotencyKey(req); if (!isIdempotencyValid) { @@ -35,7 +39,7 @@ export async function crawlController(req: Request, res: Response) { try { createIdempotencyKey(req); } catch (error) { - Logger.error(error); + logger.error(error); return res.status(500).json({ error: error.message }); } } @@ -77,7 +81,7 @@ export async function crawlController(req: Request, res: Response) { // TODO: need to do this to v1 crawlerOptions.limit = Math.min(remainingCredits, crawlerOptions.limit); - let url = req.body.url; + let url = urlSchema.parse(req.body.url); if (!url) { return res.status(400).json({ error: "Url is required" }); } @@ -123,7 +127,7 @@ export async function crawlController(req: Request, res: Response) { // documents: docs, // }); // } catch (error) { - // Logger.error(error); + // logger.error(error); // return res.status(500).json({ error: error.message }); // } // } @@ -132,10 +136,13 @@ export async function crawlController(req: Request, res: Response) { await logCrawl(id, team_id); + const { scrapeOptions, internalOptions } = fromLegacyScrapeOptions(pageOptions, undefined, undefined); + const sc: StoredCrawl = { originUrl: url, crawlerOptions, - pageOptions, + scrapeOptions, + internalOptions, team_id, plan, createdAt: Date.now(), @@ -170,10 +177,11 @@ export async function crawlController(req: Request, res: Response) { data: { url, mode: "single_urls", - crawlerOptions: crawlerOptions, + crawlerOptions, + scrapeOptions, + internalOptions, team_id, plan, - pageOptions: pageOptions, origin: req.body.origin ?? defaultOrigin, crawl_id: id, sitemapped: true, @@ -208,10 +216,11 @@ export async function crawlController(req: Request, res: Response) { { url, mode: "single_urls", - crawlerOptions: crawlerOptions, + crawlerOptions, + scrapeOptions, + internalOptions, team_id, - plan, - pageOptions: pageOptions, + plan: plan!, origin: req.body.origin ?? defaultOrigin, crawl_id: id, }, @@ -226,7 +235,9 @@ export async function crawlController(req: Request, res: Response) { res.json({ jobId: id }); } catch (error) { Sentry.captureException(error); - Logger.error(error); - return res.status(500).json({ error: error.message }); + logger.error(error); + return res.status(500).json({ error: error instanceof ZodError + ? "Invalid URL" + : error.message }); } } diff --git a/apps/api/src/controllers/v0/crawlPreview.ts b/apps/api/src/controllers/v0/crawlPreview.ts index e9f6e806..8b82bef8 100644 --- a/apps/api/src/controllers/v0/crawlPreview.ts +++ b/apps/api/src/controllers/v0/crawlPreview.ts @@ -3,15 +3,16 @@ import { authenticateUser } from "../auth"; import { RateLimiterMode } from "../../../src/types"; import { isUrlBlocked } from "../../../src/scraper/WebScraper/utils/blocklist"; import { v4 as uuidv4 } from "uuid"; -import { Logger } from "../../../src/lib/logger"; +import { logger } from "../../../src/lib/logger"; import { addCrawlJob, crawlToCrawler, lockURL, saveCrawl, StoredCrawl } from "../../../src/lib/crawl-redis"; import { addScrapeJob } from "../../../src/services/queue-jobs"; import { checkAndUpdateURL } from "../../../src/lib/validateUrl"; import * as Sentry from "@sentry/node"; +import { fromLegacyScrapeOptions } from "../v1/types"; export async function crawlPreviewController(req: Request, res: Response) { try { - const { success, error, status, team_id:a, plan } = await authenticateUser( + const auth = await authenticateUser( req, res, RateLimiterMode.Preview @@ -19,10 +20,12 @@ export async function crawlPreviewController(req: Request, res: Response) { const team_id = "preview"; - if (!success) { - return res.status(status).json({ error }); + if (!auth.success) { + return res.status(auth.status).json({ error: auth.error }); } + const { plan } = auth; + let url = req.body.url; if (!url) { return res.status(400).json({ error: "Url is required" }); @@ -71,7 +74,7 @@ export async function crawlPreviewController(req: Request, res: Response) { // documents: docs, // }); // } catch (error) { - // Logger.error(error); + // logger.error(error); // return res.status(500).json({ error: error.message }); // } // } @@ -84,10 +87,13 @@ export async function crawlPreviewController(req: Request, res: Response) { robots = await this.getRobotsTxt(); } catch (_) {} + const { scrapeOptions, internalOptions } = fromLegacyScrapeOptions(pageOptions, undefined, undefined); + const sc: StoredCrawl = { originUrl: url, crawlerOptions, - pageOptions, + scrapeOptions, + internalOptions, team_id, plan, robots, @@ -107,10 +113,11 @@ export async function crawlPreviewController(req: Request, res: Response) { await addScrapeJob({ url, mode: "single_urls", - crawlerOptions: crawlerOptions, team_id, - plan, - pageOptions: pageOptions, + plan: plan!, + crawlerOptions, + scrapeOptions, + internalOptions, origin: "website-preview", crawl_id: id, sitemapped: true, @@ -123,10 +130,11 @@ export async function crawlPreviewController(req: Request, res: Response) { await addScrapeJob({ url, mode: "single_urls", - crawlerOptions: crawlerOptions, team_id, - plan, - pageOptions: pageOptions, + plan: plan!, + crawlerOptions, + scrapeOptions, + internalOptions, origin: "website-preview", crawl_id: id, }, {}, jobId); @@ -136,7 +144,7 @@ export async function crawlPreviewController(req: Request, res: Response) { res.json({ jobId: id }); } catch (error) { Sentry.captureException(error); - Logger.error(error); + logger.error(error); return res.status(500).json({ error: error.message }); } } diff --git a/apps/api/src/controllers/v0/keyAuth.ts b/apps/api/src/controllers/v0/keyAuth.ts index b70d672a..63915302 100644 --- a/apps/api/src/controllers/v0/keyAuth.ts +++ b/apps/api/src/controllers/v0/keyAuth.ts @@ -8,13 +8,14 @@ import { authenticateUser } from "../auth"; export const keyAuthController = async (req: Request, res: Response) => { try { // make sure to authenticate user first, Bearer - const { success, team_id, error, status } = await authenticateUser( + const auth = await authenticateUser( req, res ); - if (!success) { - return res.status(status).json({ error }); + if (!auth.success) { + return res.status(auth.status).json({ error: auth.error }); } + // if success, return success: true return res.status(200).json({ success: true }); } catch (error) { diff --git a/apps/api/src/controllers/v0/scrape.ts b/apps/api/src/controllers/v0/scrape.ts index 5e6b7c6f..05fb3c41 100644 --- a/apps/api/src/controllers/v0/scrape.ts +++ b/apps/api/src/controllers/v0/scrape.ts @@ -7,7 +7,7 @@ import { import { authenticateUser } from "../auth"; import { PlanType, RateLimiterMode } from "../../types"; import { logJob } from "../../services/logging/log_job"; -import { Document } from "../../lib/entities"; +import { Document, fromLegacyCombo, toLegacyDocument, url as urlSchema } from "../v1/types"; import { isUrlBlocked } from "../../scraper/WebScraper/utils/blocklist"; // Import the isUrlBlocked function import { numTokensFromString } from "../../lib/LLM-extraction/helpers"; import { @@ -19,9 +19,11 @@ import { import { addScrapeJob, waitForJob } from "../../services/queue-jobs"; import { getScrapeQueue } from "../../services/queue-service"; import { v4 as uuidv4 } from "uuid"; -import { Logger } from "../../lib/logger"; +import { logger } from "../../lib/logger"; import * as Sentry from "@sentry/node"; import { getJobPriority } from "../../lib/job-priority"; +import { fromLegacyScrapeOptions } from "../v1/types"; +import { ZodError } from "zod"; export async function scrapeHelper( jobId: string, @@ -35,10 +37,10 @@ export async function scrapeHelper( ): Promise<{ success: boolean; error?: string; - data?: Document; + data?: Document | { url: string }; returnCode: number; }> { - const url = req.body.url; + const url = urlSchema.parse(req.body.url); if (typeof url !== "string") { return { success: false, error: "Url is required", returnCode: 400 }; } @@ -54,15 +56,16 @@ export async function scrapeHelper( const jobPriority = await getJobPriority({ plan, team_id, basePriority: 10 }); + const { scrapeOptions, internalOptions } = fromLegacyCombo(pageOptions, extractorOptions, timeout, crawlerOptions); + await addScrapeJob( { url, mode: "single_urls", - crawlerOptions, team_id, - pageOptions, - plan, - extractorOptions, + scrapeOptions, + internalOptions, + plan: plan!, origin: req.body.origin ?? defaultOrigin, is_scrape: true, }, @@ -81,7 +84,7 @@ export async function scrapeHelper( }, async (span) => { try { - doc = (await waitForJob(jobId, timeout))[0]; + doc = (await waitForJob(jobId, timeout)); } catch (e) { if (e instanceof Error && e.message.startsWith("Job wait")) { span.setAttribute("timedOut", true); @@ -149,7 +152,7 @@ export async function scrapeHelper( return { success: true, - data: doc, + data: toLegacyDocument(doc, internalOptions), returnCode: 200, }; } @@ -158,15 +161,17 @@ export async function scrapeController(req: Request, res: Response) { try { let earlyReturn = false; // make sure to authenticate user first, Bearer - const { success, team_id, error, status, plan, chunk } = await authenticateUser( + const auth = await authenticateUser( req, res, RateLimiterMode.Scrape ); - if (!success) { - return res.status(status).json({ error }); + if (!auth.success) { + return res.status(auth.status).json({ error: auth.error }); } + const { team_id, plan, chunk } = auth; + const crawlerOptions = req.body.crawlerOptions ?? {}; const pageOptions = { ...defaultPageOptions, ...req.body.pageOptions }; const extractorOptions = { @@ -200,7 +205,7 @@ export async function scrapeController(req: Request, res: Response) { return res.status(402).json({ error: "Insufficient credits. For more credits, you can upgrade your plan at https://firecrawl.dev/pricing" }); } } catch (error) { - Logger.error(error); + logger.error(error); earlyReturn = true; return res.status(500).json({ error: @@ -224,8 +229,8 @@ export async function scrapeController(req: Request, res: Response) { const endTime = new Date().getTime(); const timeTakenInSeconds = (endTime - startTime) / 1000; const numTokens = - result.data && result.data.markdown - ? numTokensFromString(result.data.markdown, "gpt-3.5-turbo") + result.data && (result.data as Document).markdown + ? numTokensFromString((result.data as Document).markdown!, "gpt-3.5-turbo") : 0; if (result.success) { @@ -246,7 +251,7 @@ export async function scrapeController(req: Request, res: Response) { if (creditsToBeBilled > 0) { // billing for doc done on queue end, bill only for llm extraction billTeam(team_id, chunk?.sub_id, creditsToBeBilled).catch(error => { - Logger.error(`Failed to bill team ${team_id} for ${creditsToBeBilled} credits: ${error}`); + logger.error(`Failed to bill team ${team_id} for ${creditsToBeBilled} credits: ${error}`); // Optionally, you could notify an admin or add to a retry queue here }); } @@ -254,17 +259,19 @@ export async function scrapeController(req: Request, res: Response) { let doc = result.data; if (!pageOptions || !pageOptions.includeRawHtml) { - if (doc && doc.rawHtml) { - delete doc.rawHtml; + if (doc && (doc as Document).rawHtml) { + delete (doc as Document).rawHtml; } } if(pageOptions && pageOptions.includeExtract) { - if(!pageOptions.includeMarkdown && doc && doc.markdown) { - delete doc.markdown; + if(!pageOptions.includeMarkdown && doc && (doc as Document).markdown) { + delete (doc as Document).markdown; } } + const { scrapeOptions } = fromLegacyScrapeOptions(pageOptions, extractorOptions, timeout); + logJob({ job_id: jobId, success: result.success, @@ -276,21 +283,22 @@ export async function scrapeController(req: Request, res: Response) { mode: "scrape", url: req.body.url, crawlerOptions: crawlerOptions, - pageOptions: pageOptions, + scrapeOptions, origin: origin, - extractor_options: extractorOptions, num_tokens: numTokens, }); return res.status(result.returnCode).json(result); } catch (error) { Sentry.captureException(error); - Logger.error(error); + logger.error(error); return res.status(500).json({ error: - typeof error === "string" - ? error - : error?.message ?? "Internal Server Error", + error instanceof ZodError + ? "Invalid URL" + : typeof error === "string" + ? error + : error?.message ?? "Internal Server Error", }); } } diff --git a/apps/api/src/controllers/v0/search.ts b/apps/api/src/controllers/v0/search.ts index 67cff8eb..e0102406 100644 --- a/apps/api/src/controllers/v0/search.ts +++ b/apps/api/src/controllers/v0/search.ts @@ -1,5 +1,4 @@ import { Request, Response } from "express"; -import { WebScraperDataProvider } from "../../scraper/WebScraper"; import { billTeam, checkTeamCredits } from "../../services/billing/credit_billing"; import { authenticateUser } from "../auth"; import { PlanType, RateLimiterMode } from "../../types"; @@ -8,21 +7,23 @@ import { PageOptions, SearchOptions } from "../../lib/entities"; import { search } from "../../search"; import { isUrlBlocked } from "../../scraper/WebScraper/utils/blocklist"; import { v4 as uuidv4 } from "uuid"; -import { Logger } from "../../lib/logger"; +import { logger } from "../../lib/logger"; import { getScrapeQueue } from "../../services/queue-service"; import { addScrapeJob, waitForJob } from "../../services/queue-jobs"; import * as Sentry from "@sentry/node"; import { getJobPriority } from "../../lib/job-priority"; +import { Job } from "bullmq"; +import { Document, fromLegacyCombo, fromLegacyScrapeOptions, toLegacyDocument } from "../v1/types"; export async function searchHelper( jobId: string, req: Request, team_id: string, - subscription_id: string, + subscription_id: string | null | undefined, crawlerOptions: any, pageOptions: PageOptions, searchOptions: SearchOptions, - plan: PlanType + plan: PlanType | undefined ): Promise<{ success: boolean; error?: string; @@ -35,8 +36,8 @@ export async function searchHelper( return { success: false, error: "Query is required", returnCode: 400 }; } - const tbs = searchOptions.tbs ?? null; - const filter = searchOptions.filter ?? null; + const tbs = searchOptions.tbs ?? undefined; + const filter = searchOptions.filter ?? undefined; let num_results = Math.min(searchOptions.limit ?? 7, 10); if (team_id === "d97c4ceb-290b-4957-8432-2b2a02727d95") { @@ -57,11 +58,12 @@ export async function searchHelper( }); let justSearch = pageOptions.fetchPageContent === false; - + + const { scrapeOptions, internalOptions } = fromLegacyCombo(pageOptions, undefined, 60000, crawlerOptions); if (justSearch) { billTeam(team_id, subscription_id, res.length).catch(error => { - Logger.error(`Failed to bill team ${team_id} for ${res.length} credits: ${error}`); + logger.error(`Failed to bill team ${team_id} for ${res.length} credits: ${error}`); // Optionally, you could notify an admin or add to a retry queue here }); return { success: true, data: res, returnCode: 200 }; @@ -88,9 +90,9 @@ export async function searchHelper( data: { url, mode: "single_urls", - crawlerOptions: crawlerOptions, team_id: team_id, - pageOptions: pageOptions, + scrapeOptions, + internalOptions, }, opts: { jobId: uuid, @@ -104,7 +106,7 @@ export async function searchHelper( await addScrapeJob(job.data as any, {}, job.opts.jobId, job.opts.priority) } - const docs = (await Promise.all(jobDatas.map(x => waitForJob(x.opts.jobId, 60000)))).map(x => x[0]); + const docs = (await Promise.all(jobDatas.map(x => waitForJob(x.opts.jobId, 60000)))).map(x => toLegacyDocument(x, internalOptions)); if (docs.length === 0) { return { success: true, error: "No search results found", returnCode: 200 }; @@ -115,7 +117,7 @@ export async function searchHelper( // make sure doc.content is not empty const filteredDocs = docs.filter( - (doc: { content?: string }) => doc && doc.content && doc.content.trim().length > 0 + (doc: any) => doc && doc.content && doc.content.trim().length > 0 ); if (filteredDocs.length === 0) { @@ -132,14 +134,15 @@ export async function searchHelper( export async function searchController(req: Request, res: Response) { try { // make sure to authenticate user first, Bearer - const { success, team_id, error, status, plan, chunk } = await authenticateUser( + const auth = await authenticateUser( req, res, RateLimiterMode.Search ); - if (!success) { - return res.status(status).json({ error }); + if (!auth.success) { + return res.status(auth.status).json({ error: auth.error }); } + const { team_id, plan, chunk } = auth; const crawlerOptions = req.body.crawlerOptions ?? {}; const pageOptions = req.body.pageOptions ?? { includeHtml: req.body.pageOptions?.includeHtml ?? false, @@ -162,7 +165,7 @@ export async function searchController(req: Request, res: Response) { } } catch (error) { Sentry.captureException(error); - Logger.error(error); + logger.error(error); return res.status(500).json({ error: "Internal server error" }); } const startTime = new Date().getTime(); @@ -189,7 +192,6 @@ export async function searchController(req: Request, res: Response) { mode: "search", url: req.body.query, crawlerOptions: crawlerOptions, - pageOptions: pageOptions, origin: origin, }); return res.status(result.returnCode).json(result); @@ -199,7 +201,7 @@ export async function searchController(req: Request, res: Response) { } Sentry.captureException(error); - Logger.error(error); + logger.error("Unhandled error occurred in search", { error }); return res.status(500).json({ error: error.message }); } } diff --git a/apps/api/src/controllers/v0/status.ts b/apps/api/src/controllers/v0/status.ts index bf8d2834..c5eafc2d 100644 --- a/apps/api/src/controllers/v0/status.ts +++ b/apps/api/src/controllers/v0/status.ts @@ -1,5 +1,5 @@ import { Request, Response } from "express"; -import { Logger } from "../../../src/lib/logger"; +import { logger } from "../../../src/lib/logger"; import { getCrawl, getCrawlJobs } from "../../../src/lib/crawl-redis"; import { getJobs } from "./crawl-status"; import * as Sentry from "@sentry/node"; @@ -37,7 +37,7 @@ export async function crawlJobStatusPreviewController(req: Request, res: Respons }); } catch (error) { Sentry.captureException(error); - Logger.error(error); + logger.error(error); return res.status(500).json({ error: error.message }); } } diff --git a/apps/api/src/controllers/v1/batch-scrape.ts b/apps/api/src/controllers/v1/batch-scrape.ts index 286163df..9c6a288c 100644 --- a/apps/api/src/controllers/v1/batch-scrape.ts +++ b/apps/api/src/controllers/v1/batch-scrape.ts @@ -4,8 +4,6 @@ import { BatchScrapeRequest, batchScrapeRequestSchema, CrawlResponse, - legacyExtractorOptions, - legacyScrapeOptions, RequestWithAuth, } from "./types"; import { @@ -29,19 +27,16 @@ export async function batchScrapeController( await logCrawl(id, req.auth.team_id); - let { remainingCredits } = req.account; + let { remainingCredits } = req.account!; const useDbAuthentication = process.env.USE_DB_AUTHENTICATION === 'true'; if(!useDbAuthentication){ remainingCredits = Infinity; } - const pageOptions = legacyScrapeOptions(req.body); - const extractorOptions = req.body.extract ? legacyExtractorOptions(req.body.extract) : undefined; - - const sc: StoredCrawl = { crawlerOptions: null, - pageOptions, + scrapeOptions: req.body, + internalOptions: {}, team_id: req.auth.team_id, createdAt: Date.now(), plan: req.auth.plan, @@ -64,10 +59,9 @@ export async function batchScrapeController( url: x, mode: "single_urls" as const, team_id: req.auth.team_id, - plan: req.auth.plan, + plan: req.auth.plan!, crawlerOptions: null, - pageOptions, - extractorOptions, + scrapeOptions: req.body, origin: "api", crawl_id: id, sitemapped: true, diff --git a/apps/api/src/controllers/v1/crawl-cancel.ts b/apps/api/src/controllers/v1/crawl-cancel.ts index f8fba824..958318b5 100644 --- a/apps/api/src/controllers/v1/crawl-cancel.ts +++ b/apps/api/src/controllers/v1/crawl-cancel.ts @@ -1,6 +1,6 @@ import { Response } from "express"; import { supabase_service } from "../../services/supabase"; -import { Logger } from "../../lib/logger"; +import { logger } from "../../lib/logger"; import { getCrawl, saveCrawl } from "../../lib/crawl-redis"; import * as Sentry from "@sentry/node"; import { configDotenv } from "dotenv"; @@ -36,7 +36,7 @@ export async function crawlCancelController(req: RequestWithAuth<{ jobId: string sc.cancelled = true; await saveCrawl(req.params.jobId, sc); } catch (error) { - Logger.error(error); + logger.error(error); } res.json({ @@ -44,7 +44,7 @@ export async function crawlCancelController(req: RequestWithAuth<{ jobId: string }); } catch (error) { Sentry.captureException(error); - Logger.error(error); + logger.error(error); return res.status(500).json({ error: error.message }); } } diff --git a/apps/api/src/controllers/v1/crawl-status-ws.ts b/apps/api/src/controllers/v1/crawl-status-ws.ts index 3738e3a2..18222edc 100644 --- a/apps/api/src/controllers/v1/crawl-status-ws.ts +++ b/apps/api/src/controllers/v1/crawl-status-ws.ts @@ -1,14 +1,15 @@ import { authMiddleware } from "../../routes/v1"; import { RateLimiterMode } from "../../types"; import { authenticateUser } from "../auth"; -import { CrawlStatusParams, CrawlStatusResponse, Document, ErrorResponse, legacyDocumentConverter, RequestWithAuth } from "./types"; +import { CrawlStatusParams, CrawlStatusResponse, Document, ErrorResponse, RequestWithAuth } from "./types"; import { WebSocket } from "ws"; import { v4 as uuidv4 } from "uuid"; -import { Logger } from "../../lib/logger"; +import { logger } from "../../lib/logger"; import { getCrawl, getCrawlExpiry, getCrawlJobs, getDoneJobsOrdered, getDoneJobsOrderedLength, getThrottledJobs, isCrawlFinished, isCrawlFinishedLocked } from "../../lib/crawl-redis"; import { getScrapeQueue } from "../../services/queue-service"; import { getJob, getJobs } from "./crawl-status"; import * as Sentry from "@sentry/node"; +import { Job, JobState } from "bullmq"; type ErrorMessage = { type: "error", @@ -56,7 +57,7 @@ async function crawlStatusWS(ws: WebSocket, req: RequestWithAuth { @@ -70,15 +71,14 @@ async function crawlStatusWS(ws: WebSocket, req: RequestWithAuth !doneJobIDs.includes(x)); const jobStatuses = await Promise.all(notDoneJobIDs.map(async x => [x, await getScrapeQueue().getJobState(x)])); - const newlyDoneJobIDs = jobStatuses.filter(x => x[1] === "completed" || x[1] === "failed").map(x => x[0]); - - for (const jobID of newlyDoneJobIDs) { - const job = await getJob(jobID); + const newlyDoneJobIDs: string[] = jobStatuses.filter(x => x[1] === "completed" || x[1] === "failed").map(x => x[0]); + const newlyDoneJobs: Job[] = (await Promise.all(newlyDoneJobIDs.map(x => getJob(x)))).filter(x => x !== undefined) as Job[] + for (const job of newlyDoneJobs) { if (job.returnvalue) { send(ws, { type: "document", - data: legacyDocumentConverter(job.returnvalue), + data: job.returnvalue, }) } else { return close(ws, 3000, { type: "error", error: job.failedReason }); @@ -100,8 +100,8 @@ async function crawlStatusWS(ws: WebSocket, req: RequestWithAuth legacyDocumentConverter(x)), + data: data, } }); @@ -139,19 +139,21 @@ async function crawlStatusWS(ws: WebSocket, req: RequestWithAuth) { try { - const { success, team_id, error, status, plan } = await authenticateUser( + const auth = await authenticateUser( req, null, RateLimiterMode.CrawlStatus, ); - if (!success) { + if (!auth.success) { return close(ws, 3000, { type: "error", - error, + error: auth.error, }); } + const { team_id, plan } = auth; + req.auth = { team_id, plan }; await crawlStatusWS(ws, req); @@ -170,7 +172,7 @@ export async function crawlStatusWSController(ws: WebSocket, req: RequestWithAut } } - Logger.error("Error occurred in WebSocket! (" + req.path + ") -- ID " + id + " -- " + verbose); + logger.error("Error occurred in WebSocket! (" + req.path + ") -- ID " + id + " -- " + verbose); return close(ws, 1011, { type: "error", error: "An unexpected error occurred. Please contact hello@firecrawl.com for help. Your exception ID is " + id diff --git a/apps/api/src/controllers/v1/crawl-status.ts b/apps/api/src/controllers/v1/crawl-status.ts index a8d78293..f150ddc4 100644 --- a/apps/api/src/controllers/v1/crawl-status.ts +++ b/apps/api/src/controllers/v1/crawl-status.ts @@ -1,9 +1,10 @@ import { Response } from "express"; -import { CrawlStatusParams, CrawlStatusResponse, ErrorResponse, legacyDocumentConverter, RequestWithAuth } from "./types"; +import { CrawlStatusParams, CrawlStatusResponse, ErrorResponse, RequestWithAuth } from "./types"; import { getCrawl, getCrawlExpiry, getCrawlJobs, getDoneJobsOrdered, getDoneJobsOrderedLength, getThrottledJobs } from "../../lib/crawl-redis"; import { getScrapeQueue } from "../../services/queue-service"; import { supabaseGetJobById, supabaseGetJobsById } from "../../lib/supabase-jobs"; import { configDotenv } from "dotenv"; +import { Job, JobState } from "bullmq"; configDotenv(); export async function getJob(id: string) { @@ -24,7 +25,7 @@ export async function getJob(id: string) { } export async function getJobs(ids: string[]) { - const jobs = (await Promise.all(ids.map(x => getScrapeQueue().getJob(x)))).filter(x => x); + const jobs: (Job & { id: string })[] = (await Promise.all(ids.map(x => getScrapeQueue().getJob(x)))).filter(x => x) as (Job & {id: string})[]; if (process.env.USE_DB_AUTHENTICATION === "true") { const supabaseData = await supabaseGetJobsById(ids); @@ -63,8 +64,8 @@ export async function crawlStatusController(req: RequestWithAuth 0) { - if (!doneJobs[0].data.pageOptions.includeRawHtml) { + if (!doneJobs[0].data.scrapeOptions.formats.includes("rawHtml")) { for (let ii = 0; ii < doneJobs.length; ii++) { if (data[ii]) { delete data[ii].rawHtml; @@ -142,7 +143,7 @@ export async function crawlStatusController(req: RequestWithAuth legacyDocumentConverter(x)), + data: data, }); } diff --git a/apps/api/src/controllers/v1/crawl.ts b/apps/api/src/controllers/v1/crawl.ts index 53cc04e8..f9f60e71 100644 --- a/apps/api/src/controllers/v1/crawl.ts +++ b/apps/api/src/controllers/v1/crawl.ts @@ -4,9 +4,8 @@ import { CrawlRequest, crawlRequestSchema, CrawlResponse, - legacyCrawlerOptions, - legacyScrapeOptions, RequestWithAuth, + toLegacyCrawlerOptions, } from "./types"; import { addCrawlJob, @@ -20,9 +19,10 @@ import { import { logCrawl } from "../../services/logging/crawl_log"; import { getScrapeQueue } from "../../services/queue-service"; import { addScrapeJob } from "../../services/queue-jobs"; -import { Logger } from "../../lib/logger"; +import { logger } from "../../lib/logger"; import { getJobPriority } from "../../lib/job-priority"; import { callWebhook } from "../../services/webhook"; +import { scrapeOptions as scrapeOptionsSchema } from "./types"; export async function crawlController( req: RequestWithAuth<{}, CrawlResponse, CrawlRequest>, @@ -34,18 +34,22 @@ export async function crawlController( await logCrawl(id, req.auth.team_id); - let { remainingCredits } = req.account; + let { remainingCredits } = req.account!; const useDbAuthentication = process.env.USE_DB_AUTHENTICATION === 'true'; if(!useDbAuthentication){ remainingCredits = Infinity; } - const crawlerOptions = legacyCrawlerOptions(req.body); - const pageOptions = legacyScrapeOptions(req.body.scrapeOptions); + const crawlerOptions = { + ...req.body, + url: undefined, + scrapeOptions: undefined, + }; + const scrapeOptions = req.body.scrapeOptions; // TODO: @rafa, is this right? copied from v0 - if (Array.isArray(crawlerOptions.includes)) { - for (const x of crawlerOptions.includes) { + if (Array.isArray(crawlerOptions.includePaths)) { + for (const x of crawlerOptions.includePaths) { try { new RegExp(x); } catch (e) { @@ -54,8 +58,8 @@ export async function crawlController( } } - if (Array.isArray(crawlerOptions.excludes)) { - for (const x of crawlerOptions.excludes) { + if (Array.isArray(crawlerOptions.excludePaths)) { + for (const x of crawlerOptions.excludePaths) { try { new RegExp(x); } catch (e) { @@ -68,8 +72,9 @@ export async function crawlController( const sc: StoredCrawl = { originUrl: req.body.url, - crawlerOptions, - pageOptions, + crawlerOptions: toLegacyCrawlerOptions(crawlerOptions), + scrapeOptions, + internalOptions: {}, team_id: req.auth.team_id, createdAt: Date.now(), plan: req.auth.plan, @@ -78,9 +83,9 @@ export async function crawlController( const crawler = crawlToCrawler(id, sc); try { - sc.robots = await crawler.getRobotsTxt(pageOptions.skipTlsVerification); + sc.robots = await crawler.getRobotsTxt(scrapeOptions.skipTlsVerification); } catch (e) { - Logger.debug( + logger.debug( `[Crawl] Failed to get robots.txt (this is probably fine!): ${JSON.stringify( e )}` @@ -112,7 +117,7 @@ export async function crawlController( team_id: req.auth.team_id, plan: req.auth.plan, crawlerOptions, - pageOptions, + scrapeOptions, origin: "api", crawl_id: id, sitemapped: true, @@ -142,10 +147,10 @@ export async function crawlController( { url: req.body.url, mode: "single_urls", - crawlerOptions: crawlerOptions, team_id: req.auth.team_id, - plan: req.auth.plan, - pageOptions: pageOptions, + crawlerOptions, + scrapeOptions: scrapeOptionsSchema.parse(scrapeOptions), + plan: req.auth.plan!, origin: "api", crawl_id: id, webhook: req.body.webhook, diff --git a/apps/api/src/controllers/v1/map.ts b/apps/api/src/controllers/v1/map.ts index 5ed3dd51..45856543 100644 --- a/apps/api/src/controllers/v1/map.ts +++ b/apps/api/src/controllers/v1/map.ts @@ -1,9 +1,9 @@ import { Response } from "express"; import { v4 as uuidv4 } from "uuid"; import { - legacyCrawlerOptions, mapRequestSchema, RequestWithAuth, + scrapeOptions, } from "./types"; import { crawlToCrawler, StoredCrawl } from "../../lib/crawl-redis"; import { MapResponse, MapRequest } from "./types"; @@ -18,11 +18,11 @@ import { fireEngineMap } from "../../search/fireEngine"; import { billTeam } from "../../services/billing/credit_billing"; import { logJob } from "../../services/logging/log_job"; import { performCosineSimilarity } from "../../lib/map-cosine"; -import { Logger } from "../../lib/logger"; +import { logger } from "../../lib/logger"; import Redis from "ioredis"; configDotenv(); -const redis = new Redis(process.env.REDIS_URL); +const redis = new Redis(process.env.REDIS_URL!); // Max Links that /map can return const MAX_MAP_LIMIT = 5000; @@ -44,8 +44,12 @@ export async function mapController( const sc: StoredCrawl = { originUrl: req.body.url, - crawlerOptions: legacyCrawlerOptions(req.body), - pageOptions: {}, + crawlerOptions: { + ...req.body, + scrapeOptions: undefined, + }, + scrapeOptions: scrapeOptions.parse({}), + internalOptions: {}, team_id: req.auth.team_id, createdAt: Date.now(), plan: req.auth.plan, @@ -65,8 +69,8 @@ export async function mapController( const cacheKey = `fireEngineMap:${mapUrl}`; const cachedResult = null; - let allResults: any[]; - let pagePromises: Promise[]; + let allResults: any[] = []; + let pagePromises: Promise[] = []; if (cachedResult) { allResults = JSON.parse(cachedResult); @@ -139,7 +143,7 @@ export async function mapController( return null; } }) - .filter((x) => x !== null); + .filter((x) => x !== null) as string[]; // allows for subdomains to be included links = links.filter((x) => isSameDomain(x, req.body.url)); @@ -153,7 +157,7 @@ export async function mapController( links = removeDuplicateUrls(links); billTeam(req.auth.team_id, req.acuc?.sub_id, 1).catch((error) => { - Logger.error( + logger.error( `Failed to bill team ${req.auth.team_id} for 1 credit: ${error}` ); // Optionally, you could notify an admin or add to a retry queue here @@ -175,9 +179,8 @@ export async function mapController( mode: "map", url: req.body.url, crawlerOptions: {}, - pageOptions: {}, + scrapeOptions: {}, origin: req.body.origin, - extractor_options: { mode: "markdown" }, num_tokens: 0, }); diff --git a/apps/api/src/controllers/v1/scrape-status.ts b/apps/api/src/controllers/v1/scrape-status.ts index 5e0aecb6..db50f7d3 100644 --- a/apps/api/src/controllers/v1/scrape-status.ts +++ b/apps/api/src/controllers/v1/scrape-status.ts @@ -12,7 +12,7 @@ export async function scrapeStatusController(req: any, res: any) { const job = await supabaseGetJobByIdOnlyData(req.params.jobId); - if(job.team_id !== "41bdbfe1-0579-4d9b-b6d5-809f16be12f5"){ + if(job?.team_id !== "41bdbfe1-0579-4d9b-b6d5-809f16be12f5"){ return res.status(403).json({ success: false, error: "You are not allowed to access this resource.", diff --git a/apps/api/src/controllers/v1/scrape.ts b/apps/api/src/controllers/v1/scrape.ts index 6fa855f7..3f6609af 100644 --- a/apps/api/src/controllers/v1/scrape.ts +++ b/apps/api/src/controllers/v1/scrape.ts @@ -1,10 +1,7 @@ -import { Request, Response } from "express"; -import { Logger } from "../../lib/logger"; +import { Response } from "express"; +import { logger } from "../../lib/logger"; import { Document, - legacyDocumentConverter, - legacyExtractorOptions, - legacyScrapeOptions, RequestWithAuth, ScrapeRequest, scrapeRequestSchema, @@ -12,7 +9,6 @@ import { } from "./types"; import { billTeam } from "../../services/billing/credit_billing"; import { v4 as uuidv4 } from "uuid"; -import { numTokensFromString } from "../../lib/LLM-extraction/helpers"; import { addScrapeJob, waitForJob } from "../../services/queue-jobs"; import { logJob } from "../../services/logging/log_job"; import { getJobPriority } from "../../lib/job-priority"; @@ -28,8 +24,6 @@ export async function scrapeController( const origin = req.body.origin; const timeout = req.body.timeout; - const pageOptions = legacyScrapeOptions(req.body); - const extractorOptions = req.body.extract ? legacyExtractorOptions(req.body.extract) : undefined; const jobId = uuidv4(); const startTime = new Date().getTime(); @@ -43,11 +37,10 @@ export async function scrapeController( { url: req.body.url, mode: "single_urls", - crawlerOptions: {}, team_id: req.auth.team_id, - plan: req.auth.plan, - pageOptions, - extractorOptions, + scrapeOptions: req.body, + internalOptions: {}, + plan: req.auth.plan!, origin: req.body.origin, is_scrape: true, }, @@ -56,13 +49,13 @@ export async function scrapeController( jobPriority ); - const totalWait = (req.body.waitFor ?? 0) + (req.body.actions ?? []).reduce((a,x) => (x.type === "wait" ? x.milliseconds : 0) + a, 0); + const totalWait = (req.body.waitFor ?? 0) + (req.body.actions ?? []).reduce((a,x) => (x.type === "wait" ? x.milliseconds ?? 0 : 0) + a, 0); - let doc: any | undefined; + let doc: Document; try { - doc = (await waitForJob(jobId, timeout + totalWait))[0]; + doc = await waitForJob(jobId, timeout + totalWait); // TODO: better types for this } catch (e) { - Logger.error(`Error in scrapeController: ${e}`); + logger.error(`Error in scrapeController: ${e}`); if (e instanceof Error && e.message.startsWith("Job wait")) { return res.status(408).json({ success: false, @@ -71,34 +64,19 @@ export async function scrapeController( } else { return res.status(500).json({ success: false, - error: `(Internal server error) - ${e && e?.message ? e.message : e} ${ - extractorOptions && extractorOptions.mode !== "markdown" - ? " - Could be due to LLM parsing issues" - : "" - }`, + error: `(Internal server error) - ${e && e?.message ? e.message : e}`, }); } } await getScrapeQueue().remove(jobId); - if (!doc) { - console.error("!!! PANIC DOC IS", doc); - return res.status(200).json({ - success: true, - warning: "No page found", - data: doc, - }); - } - - delete doc.index; - delete doc.provider; - const endTime = new Date().getTime(); const timeTakenInSeconds = (endTime - startTime) / 1000; const numTokens = - doc && doc.markdown - ? numTokensFromString(doc.markdown, "gpt-3.5-turbo") + doc && doc.extract + // ? numTokensFromString(doc.markdown, "gpt-3.5-turbo") + ? 0 // TODO: fix : 0; let creditsToBeBilled = 1; // Assuming 1 credit per document @@ -111,22 +89,16 @@ export async function scrapeController( } billTeam(req.auth.team_id, req.acuc?.sub_id, creditsToBeBilled).catch(error => { - Logger.error(`Failed to bill team ${req.auth.team_id} for ${creditsToBeBilled} credits: ${error}`); + logger.error(`Failed to bill team ${req.auth.team_id} for ${creditsToBeBilled} credits: ${error}`); // Optionally, you could notify an admin or add to a retry queue here }); - if (!pageOptions || !pageOptions.includeRawHtml) { + if (!req.body.formats.includes("rawHtml")) { if (doc && doc.rawHtml) { delete doc.rawHtml; } } - if(pageOptions && pageOptions.includeExtract) { - if(!pageOptions.includeMarkdown && doc && doc.markdown) { - delete doc.markdown; - } - } - logJob({ job_id: jobId, success: true, @@ -137,16 +109,14 @@ export async function scrapeController( team_id: req.auth.team_id, mode: "scrape", url: req.body.url, - crawlerOptions: {}, - pageOptions: pageOptions, + scrapeOptions: req.body, origin: origin, - extractor_options: extractorOptions, num_tokens: numTokens, }); return res.status(200).json({ success: true, - data: legacyDocumentConverter(doc), + data: doc, scrape_id: origin?.includes("website") ? jobId : undefined, }); } diff --git a/apps/api/src/controllers/v1/types.ts b/apps/api/src/controllers/v1/types.ts index 9585175e..82b24c22 100644 --- a/apps/api/src/controllers/v1/types.ts +++ b/apps/api/src/controllers/v1/types.ts @@ -1,10 +1,11 @@ import { Request, Response } from "express"; import { z } from "zod"; import { isUrlBlocked } from "../../scraper/WebScraper/utils/blocklist"; -import { Action, ExtractorOptions, PageOptions } from "../../lib/entities"; import { protocolIncluded, checkUrl } from "../../lib/validateUrl"; import { PlanType } from "../../types"; import { countries } from "../../lib/validate-country"; +import { ExtractorOptions, PageOptions, ScrapeActionContent, Document as V0Document } from "../../lib/entities"; +import { InternalOptions } from "../../scraper/scrapeURL"; export type Format = | "markdown" @@ -167,6 +168,7 @@ export const scrapeRequestSchema = scrapeOptions.extend({ }); export type ScrapeRequest = z.infer; +export type ScrapeRequestInput = z.input; export const batchScrapeRequestSchema = scrapeOptions.extend({ urls: url.array(), @@ -240,7 +242,7 @@ export const mapRequestSchema = crawlerOptions.extend({ includeSubdomains: z.boolean().default(true), search: z.string().optional(), ignoreSitemap: z.boolean().default(false), - limit: z.number().min(1).max(5000).default(5000).optional(), + limit: z.number().min(1).max(5000).default(5000), }).strict(strictMessage); // export type MapRequest = { @@ -252,13 +254,14 @@ export type MapRequest = z.infer; export type Document = { markdown?: string; - extract?: string; + extract?: any; html?: string; rawHtml?: string; links?: string[]; screenshot?: string; actions?: { - screenshots: string[]; + screenshots?: string[]; + scrapes?: ScrapeActionContent[]; }; warning?: string; metadata: { @@ -291,11 +294,11 @@ export type Document = { publishedTime?: string; articleTag?: string; articleSection?: string; + url?: string; sourceURL?: string; statusCode?: number; error?: string; [key: string]: string | string[] | number | undefined; - }; }; @@ -366,7 +369,7 @@ export type CrawlStatusResponse = type AuthObject = { team_id: string; - plan: PlanType; + plan: PlanType | undefined; }; type Account = { @@ -439,7 +442,7 @@ export interface ResponseWithSentry< sentry?: string, } -export function legacyCrawlerOptions(x: CrawlerOptions) { +export function toLegacyCrawlerOptions(x: CrawlerOptions) { return { includes: x.includePaths, excludes: x.excludePaths, @@ -453,68 +456,90 @@ export function legacyCrawlerOptions(x: CrawlerOptions) { }; } -export function legacyScrapeOptions(x: ScrapeOptions): PageOptions { +export function fromLegacyCrawlerOptions(x: any): { crawlOptions: CrawlerOptions; internalOptions: InternalOptions } { return { - includeMarkdown: x.formats.includes("markdown"), - includeHtml: x.formats.includes("html"), - includeRawHtml: x.formats.includes("rawHtml"), - includeExtract: x.formats.includes("extract"), - onlyIncludeTags: x.includeTags, - removeTags: x.excludeTags, - onlyMainContent: x.onlyMainContent, - waitFor: x.waitFor, - headers: x.headers, - includeLinks: x.formats.includes("links"), - screenshot: x.formats.includes("screenshot"), - fullPageScreenshot: x.formats.includes("screenshot@fullPage"), - parsePDF: x.parsePDF, - actions: x.actions as Action[], // no strict null checking grrrr - mogery - geolocation: x.location ?? x.geolocation, - skipTlsVerification: x.skipTlsVerification, - removeBase64Images: x.removeBase64Images, - mobile: x.mobile, - }; -} - -export function legacyExtractorOptions(x: ExtractOptions): ExtractorOptions { - return { - mode: x.mode ? "llm-extraction" : "markdown", - extractionPrompt: x.prompt ?? "Based on the information on the page, extract the information from the schema.", - extractionSchema: x.schema, - userPrompt: x.prompt ?? "", - }; -} - -export function legacyDocumentConverter(doc: any): Document { - if (doc === null || doc === undefined) return null; - - if (doc.metadata) { - if (doc.metadata.screenshot) { - doc.screenshot = doc.metadata.screenshot; - delete doc.metadata.screenshot; - } - - if (doc.metadata.fullPageScreenshot) { - doc.fullPageScreenshot = doc.metadata.fullPageScreenshot; - delete doc.metadata.fullPageScreenshot; - } - } - - return { - markdown: doc.markdown, - links: doc.linksOnPage, - rawHtml: doc.rawHtml, - html: doc.html, - extract: doc.llm_extraction, - screenshot: doc.screenshot ?? doc.fullPageScreenshot, - actions: doc.actions ?? undefined, - warning: doc.warning ?? undefined, - metadata: { - ...doc.metadata, - pageError: undefined, - pageStatusCode: undefined, - error: doc.metadata?.pageError, - statusCode: doc.metadata?.pageStatusCode, + crawlOptions: crawlerOptions.parse({ + includePaths: x.includes, + excludePaths: x.excludes, + limit: x.maxCrawledLinks ?? x.limit, + maxDepth: x.maxDepth, + allowBackwardLinks: x.allowBackwardCrawling, + allowExternalLinks: x.allowExternalContentLinks, + ignoreSitemap: x.ignoreSitemap, + // TODO: returnOnlyUrls support + }), + internalOptions: { + v0CrawlOnlyUrls: x.returnOnlyUrls, }, }; } + +export function fromLegacyScrapeOptions(pageOptions: PageOptions, extractorOptions: ExtractorOptions | undefined, timeout: number | undefined): { scrapeOptions: ScrapeOptions, internalOptions: InternalOptions } { + return { + scrapeOptions: scrapeOptions.parse({ + formats: [ + (pageOptions.includeMarkdown ?? true) ? "markdown" as const : null, + (pageOptions.includeHtml ?? false) ? "html" as const : null, + (pageOptions.includeRawHtml ?? false) ? "rawHtml" as const : null, + (pageOptions.screenshot ?? false) ? "screenshot" as const : null, + (pageOptions.fullPageScreenshot ?? false) ? "screenshot@fullPage" as const : null, + (extractorOptions !== undefined && extractorOptions.mode.includes("llm-extraction")) ? "extract" as const : null, + "links" + ].filter(x => x !== null), + waitFor: pageOptions.waitFor, + headers: pageOptions.headers, + includeTags: (typeof pageOptions.onlyIncludeTags === "string" ? [pageOptions.onlyIncludeTags] : pageOptions.onlyIncludeTags), + excludeTags: (typeof pageOptions.removeTags === "string" ? [pageOptions.removeTags] : pageOptions.removeTags), + onlyMainContent: pageOptions.onlyMainContent ?? false, + timeout: timeout, + parsePDF: pageOptions.parsePDF, + actions: pageOptions.actions, + location: pageOptions.geolocation, + skipTlsVerification: pageOptions.skipTlsVerification, + removeBase64Images: pageOptions.removeBase64Images, + extract: extractorOptions !== undefined && extractorOptions.mode.includes("llm-extraction") ? { + systemPrompt: extractorOptions.extractionPrompt, + prompt: extractorOptions.userPrompt, + schema: extractorOptions.extractionSchema, + } : undefined, + mobile: pageOptions.mobile, + }), + internalOptions: { + atsv: pageOptions.atsv, + v0DisableJsDom: pageOptions.disableJsDom, + v0UseFastMode: pageOptions.useFastMode, + }, + // TODO: fallback, fetchPageContent, replaceAllPathsWithAbsolutePaths, includeLinks + } +} + +export function fromLegacyCombo(pageOptions: PageOptions, extractorOptions: ExtractorOptions | undefined, timeout: number | undefined, crawlerOptions: any): { scrapeOptions: ScrapeOptions, internalOptions: InternalOptions} { + const { scrapeOptions, internalOptions: i1 } = fromLegacyScrapeOptions(pageOptions, extractorOptions, timeout); + const { internalOptions: i2 } = fromLegacyCrawlerOptions(crawlerOptions); + return { scrapeOptions, internalOptions: Object.assign(i1, i2) }; +} + +export function toLegacyDocument(document: Document, internalOptions: InternalOptions): V0Document | { url: string; } { + if (internalOptions.v0CrawlOnlyUrls) { + return { url: document.metadata.sourceURL! }; + } + + return { + content: document.markdown!, + markdown: document.markdown!, + html: document.html, + rawHtml: document.rawHtml, + linksOnPage: document.links, + llm_extraction: document.extract, + metadata: { + ...document.metadata, + error: undefined, + statusCode: undefined, + pageError: document.metadata.error, + pageStatusCode: document.metadata.statusCode, + screenshot: document.screenshot, + }, + actions: document.actions , + warning: document.warning, + } +} diff --git a/apps/api/src/example.ts b/apps/api/src/example.ts deleted file mode 100644 index edf0faef..00000000 --- a/apps/api/src/example.ts +++ /dev/null @@ -1,19 +0,0 @@ -import { WebScraperDataProvider } from "./scraper/WebScraper"; - -async function example() { - const example = new WebScraperDataProvider(); - - await example.setOptions({ - jobId: "TEST", - mode: "crawl", - urls: ["https://mendable.ai"], - crawlerOptions: {}, - }); - const docs = await example.getDocuments(false); - docs.map((doc) => { - console.log(doc.metadata.sourceURL); - }); - console.log(docs.length); -} - -// example(); diff --git a/apps/api/src/index.ts b/apps/api/src/index.ts index 5ccbb9cc..7f7ec036 100644 --- a/apps/api/src/index.ts +++ b/apps/api/src/index.ts @@ -6,28 +6,24 @@ import bodyParser from "body-parser"; import cors from "cors"; import { getScrapeQueue } from "./services/queue-service"; import { v0Router } from "./routes/v0"; -import { initSDK } from "@hyperdx/node-opentelemetry"; import os from "os"; -import { Logger } from "./lib/logger"; +import { logger } from "./lib/logger"; import { adminRouter } from "./routes/admin"; -import { ScrapeEvents } from "./lib/scrape-events"; import http from 'node:http'; import https from 'node:https'; import CacheableLookup from 'cacheable-lookup'; import { v1Router } from "./routes/v1"; import expressWs from "express-ws"; -import { crawlStatusWSController } from "./controllers/v1/crawl-status-ws"; import { ErrorResponse, ResponseWithSentry } from "./controllers/v1/types"; import { ZodError } from "zod"; import { v4 as uuidv4 } from "uuid"; -import dns from 'node:dns'; const { createBullBoard } = require("@bull-board/api"); const { BullAdapter } = require("@bull-board/api/bullAdapter"); const { ExpressAdapter } = require("@bull-board/express"); const numCPUs = process.env.ENV === "local" ? 2 : os.cpus().length; -Logger.info(`Number of CPUs: ${numCPUs} available`); +logger.info(`Number of CPUs: ${numCPUs} available`); const cacheable = new CacheableLookup() @@ -55,7 +51,6 @@ const { addQueue, removeQueue, setQueues, replaceQueues } = createBullBoard({ serverAdapter: serverAdapter, }); - app.use( `/admin/${process.env.BULL_AUTH_KEY}/queues`, serverAdapter.getRouter() @@ -78,15 +73,10 @@ app.use(adminRouter); const DEFAULT_PORT = process.env.PORT ?? 3002; const HOST = process.env.HOST ?? "localhost"; -// HyperDX OpenTelemetry -if (process.env.ENV === "production") { - initSDK({ consoleCapture: true, additionalInstrumentations: [] }); -} - function startServer(port = DEFAULT_PORT) { const server = app.listen(Number(port), HOST, () => { - Logger.info(`Worker ${process.pid} listening on port ${port}`); - Logger.info( + logger.info(`Worker ${process.pid} listening on port ${port}`); + logger.info( `For the Queue UI, open: http://${HOST}:${port}/admin/${process.env.BULL_AUTH_KEY}/queues` ); }); @@ -103,7 +93,6 @@ app.get(`/serverHealthCheck`, async (req, res) => { const [waitingJobs] = await Promise.all([ scrapeQueue.getWaitingCount(), ]); - const noWaitingJobs = waitingJobs === 0; // 200 if no active jobs, 503 if there are active jobs return res.status(noWaitingJobs ? 200 : 500).json({ @@ -111,7 +100,7 @@ app.get(`/serverHealthCheck`, async (req, res) => { }); } catch (error) { Sentry.captureException(error); - Logger.error(error); + logger.error(error); return res.status(500).json({ error: error.message }); } }); @@ -140,7 +129,7 @@ app.get("/serverHealthCheck/notify", async (req, res) => { // Re-check the waiting jobs count after the timeout waitingJobsCount = await getWaitingJobsCount(); if (waitingJobsCount >= treshold) { - const slackWebhookUrl = process.env.SLACK_WEBHOOK_URL; + const slackWebhookUrl = process.env.SLACK_WEBHOOK_URL!; const message = { text: `⚠️ Warning: The number of active jobs (${waitingJobsCount}) has exceeded the threshold (${treshold}) for more than ${ timeout / 60000 @@ -156,14 +145,14 @@ app.get("/serverHealthCheck/notify", async (req, res) => { }); if (!response.ok) { - Logger.error("Failed to send Slack notification"); + logger.error("Failed to send Slack notification"); } } }, timeout); } } catch (error) { Sentry.captureException(error); - Logger.debug(error); + logger.debug(error); } }; @@ -178,7 +167,7 @@ app.get("/is-production", (req, res) => { app.use((err: unknown, req: Request<{}, ErrorResponse, undefined>, res: Response, next: NextFunction) => { if (err instanceof ZodError) { if (Array.isArray(err.errors) && err.errors.find(x => x.message === "URL uses unsupported protocol")) { - Logger.warn("Unsupported protocol error: " + JSON.stringify(req.body)); + logger.warn("Unsupported protocol error: " + JSON.stringify(req.body)); } res.status(400).json({ success: false, error: "Bad Request", details: err.errors }); @@ -206,11 +195,11 @@ app.use((err: unknown, req: Request<{}, ErrorResponse, undefined>, res: Response } } - Logger.error("Error occurred in request! (" + req.path + ") -- ID " + id + " -- " + verbose); + logger.error("Error occurred in request! (" + req.path + ") -- ID " + id + " -- " + verbose); res.status(500).json({ success: false, error: "An unexpected error occurred. Please contact hello@firecrawl.com for help. Your exception ID is " + id }); }); -Logger.info(`Worker ${process.pid} started`); +logger.info(`Worker ${process.pid} started`); // const sq = getScrapeQueue(); diff --git a/apps/api/src/lib/LLM-extraction/index.ts b/apps/api/src/lib/LLM-extraction/index.ts index d05f9bd7..430dc1d4 100644 --- a/apps/api/src/lib/LLM-extraction/index.ts +++ b/apps/api/src/lib/LLM-extraction/index.ts @@ -4,19 +4,19 @@ const ajv = new Ajv(); // Initialize AJV for JSON schema validation import { generateOpenAICompletions } from "./models"; import { Document, ExtractorOptions } from "../entities"; -import { Logger } from "../logger"; +import { logger } from "../logger"; // Generate completion using OpenAI export async function generateCompletions( documents: Document[], - extractionOptions: ExtractorOptions, + extractionOptions: ExtractorOptions | undefined, mode: "markdown" | "raw-html" ): Promise { // const schema = zodToJsonSchema(options.schema) - const schema = extractionOptions.extractionSchema; - const systemPrompt = extractionOptions.extractionPrompt; - const prompt = extractionOptions.userPrompt; + const schema = extractionOptions?.extractionSchema; + const systemPrompt = extractionOptions?.extractionPrompt; + const prompt = extractionOptions?.userPrompt; const switchVariable = "openAI"; // Placholder, want to think more about how we abstract the model provider @@ -51,7 +51,7 @@ export async function generateCompletions( return completionResult; } catch (error) { - Logger.error(`Error generating completions: ${error}`); + logger.error(`Error generating completions: ${error}`); throw error; } default: diff --git a/apps/api/src/lib/LLM-extraction/models.ts b/apps/api/src/lib/LLM-extraction/models.ts index 23147b12..f777dce9 100644 --- a/apps/api/src/lib/LLM-extraction/models.ts +++ b/apps/api/src/lib/LLM-extraction/models.ts @@ -95,7 +95,7 @@ export async function generateOpenAICompletions({ try { llmExtraction = JSON.parse( - jsonCompletion.choices[0].message.content.trim() + (jsonCompletion.choices[0].message.content ?? "").trim() ); } catch (e) { throw new Error("Invalid JSON"); diff --git a/apps/api/src/lib/batch-process.ts b/apps/api/src/lib/batch-process.ts index 30289fd0..802d1eb1 100644 --- a/apps/api/src/lib/batch-process.ts +++ b/apps/api/src/lib/batch-process.ts @@ -3,7 +3,7 @@ export async function batchProcess( batchSize: number, asyncFunction: (item: T, index: number) => Promise ): Promise { - const batches = []; + const batches: T[][] = []; for (let i = 0; i < array.length; i += batchSize) { const batch = array.slice(i, i + batchSize); batches.push(batch); diff --git a/apps/api/src/lib/crawl-redis.ts b/apps/api/src/lib/crawl-redis.ts index 379bc179..41cbb07c 100644 --- a/apps/api/src/lib/crawl-redis.ts +++ b/apps/api/src/lib/crawl-redis.ts @@ -1,13 +1,16 @@ +import { InternalOptions } from "../scraper/scrapeURL"; +import { ScrapeOptions } from "../controllers/v1/types"; import { WebCrawler } from "../scraper/WebScraper/crawler"; import { redisConnection } from "../services/queue-service"; -import { Logger } from "./logger"; +import { logger } from "./logger"; export type StoredCrawl = { originUrl?: string; crawlerOptions: any; - pageOptions: any; + scrapeOptions: Omit; + internalOptions: InternalOptions; team_id: string; - plan: string; + plan?: string; robots?: string; cancelled?: boolean; createdAt: number; @@ -100,7 +103,7 @@ export async function lockURL(id: string, sc: StoredCrawl, url: string): Promise urlO.hash = ""; url = urlO.href; } catch (error) { - Logger.warn("Failed to normalize URL " + JSON.stringify(url) + ": " + error); + logger.warn("Failed to normalize URL " + JSON.stringify(url) + ": " + error); } const res = (await redisConnection.sadd("crawl:" + id + ":visited", url)) !== 0 @@ -117,7 +120,7 @@ export async function lockURLs(id: string, urls: string[]): Promise { urlO.hash = ""; return urlO.href; } catch (error) { - Logger.warn("Failed to normalize URL " + JSON.stringify(url) + ": " + error); + logger.warn("Failed to normalize URL " + JSON.stringify(url) + ": " + error); } return url; @@ -131,7 +134,7 @@ export async function lockURLs(id: string, urls: string[]): Promise { export function crawlToCrawler(id: string, sc: StoredCrawl): WebCrawler { const crawler = new WebCrawler({ jobId: id, - initialUrl: sc.originUrl, + initialUrl: sc.originUrl!, includes: sc.crawlerOptions?.includes ?? [], excludes: sc.crawlerOptions?.excludes ?? [], maxCrawledLinks: sc.crawlerOptions?.maxCrawledLinks ?? 1000, diff --git a/apps/api/src/lib/entities.ts b/apps/api/src/lib/entities.ts index 2c8f376d..9b68f425 100644 --- a/apps/api/src/lib/entities.ts +++ b/apps/api/src/lib/entities.ts @@ -1,3 +1,5 @@ +import type { Document as V1Document } from "../controllers/v1/types"; + export interface Progress { current: number; total: number; @@ -129,7 +131,8 @@ export class Document { provider?: string; warning?: string; actions?: { - screenshots: string[]; + screenshots?: string[]; + scrapes?: ScrapeActionContent[]; } index?: number; diff --git a/apps/api/src/lib/html-to-markdown.ts b/apps/api/src/lib/html-to-markdown.ts index 396b7fe7..8800d916 100644 --- a/apps/api/src/lib/html-to-markdown.ts +++ b/apps/api/src/lib/html-to-markdown.ts @@ -5,7 +5,7 @@ import "../services/sentry" import * as Sentry from "@sentry/node"; import dotenv from 'dotenv'; -import { Logger } from './logger'; +import { logger } from './logger'; dotenv.config(); // TODO: add a timeout to the Go parser @@ -40,7 +40,7 @@ class GoMarkdownConverter { } } -export async function parseMarkdown(html: string): Promise { +export async function parseMarkdown(html: string | null | undefined): Promise { if (!html) { return ''; } @@ -52,12 +52,12 @@ export async function parseMarkdown(html: string): Promise { markdownContent = processMultiLineLinks(markdownContent); markdownContent = removeSkipToContentLinks(markdownContent); - Logger.info(`HTML to Markdown conversion using Go parser successful`); + logger.info(`HTML to Markdown conversion using Go parser successful`); return markdownContent; } } catch (error) { Sentry.captureException(error); - Logger.error(`Error converting HTML to Markdown with Go parser: ${error}`); + logger.error(`Error converting HTML to Markdown with Go parser: ${error}`); } // Fallback to TurndownService if Go parser fails or is not enabled diff --git a/apps/api/src/lib/job-priority.ts b/apps/api/src/lib/job-priority.ts index 9d046052..27e45230 100644 --- a/apps/api/src/lib/job-priority.ts +++ b/apps/api/src/lib/job-priority.ts @@ -1,6 +1,6 @@ import { redisConnection } from "../../src/services/queue-service"; import { PlanType } from "../../src/types"; -import { Logger } from "./logger"; +import { logger } from "./logger"; const SET_KEY_PREFIX = "limit_team_id:"; export async function addJobPriority(team_id, job_id) { @@ -13,7 +13,7 @@ export async function addJobPriority(team_id, job_id) { // This approach will reset the expiration time to 60 seconds every time a new job is added to the set. await redisConnection.expire(setKey, 60); } catch (e) { - Logger.error(`Add job priority (sadd) failed: ${team_id}, ${job_id}`); + logger.error(`Add job priority (sadd) failed: ${team_id}, ${job_id}`); } } @@ -24,7 +24,7 @@ export async function deleteJobPriority(team_id, job_id) { // remove job_id from the set await redisConnection.srem(setKey, job_id); } catch (e) { - Logger.error(`Delete job priority (srem) failed: ${team_id}, ${job_id}`); + logger.error(`Delete job priority (srem) failed: ${team_id}, ${job_id}`); } } @@ -33,7 +33,7 @@ export async function getJobPriority({ team_id, basePriority = 10, }: { - plan: PlanType; + plan: PlanType | undefined; team_id: string; basePriority?: number; }): Promise { @@ -95,7 +95,7 @@ export async function getJobPriority({ ); } } catch (e) { - Logger.error( + logger.error( `Get job priority failed: ${team_id}, ${plan}, ${basePriority}` ); return basePriority; diff --git a/apps/api/src/lib/load-testing-example.ts b/apps/api/src/lib/load-testing-example.ts deleted file mode 100644 index 01b61db9..00000000 --- a/apps/api/src/lib/load-testing-example.ts +++ /dev/null @@ -1,42 +0,0 @@ -// import { scrapWithFireEngine } from "../../src/scraper/WebScraper/single_url"; - -// const delay = (ms: number) => new Promise((resolve) => setTimeout(resolve, ms)); - -// const scrapInBatches = async ( -// urls: string[], -// batchSize: number, -// delayMs: number -// ) => { -// let successCount = 0; -// let errorCount = 0; - -// for (let i = 0; i < urls.length; i += batchSize) { -// const batch = urls -// .slice(i, i + batchSize) -// .map((url) => scrapWithFireEngine(url)); -// try { -// const results = await Promise.all(batch); -// results.forEach((data, index) => { -// if (data.trim() === "") { -// errorCount++; -// } else { -// successCount++; -// console.log( -// `Scraping result ${i + index + 1}:`, -// data.trim().substring(0, 20) + "..." -// ); -// } -// }); -// } catch (error) { -// console.error("Error during scraping:", error); -// } -// await delay(delayMs); -// } - -// console.log(`Total successful scrapes: ${successCount}`); -// console.log(`Total errored scrapes: ${errorCount}`); -// }; -// function run() { -// const urls = Array.from({ length: 200 }, () => "https://scrapethissite.com"); -// scrapInBatches(urls, 10, 1000); -// } diff --git a/apps/api/src/lib/logger.ts b/apps/api/src/lib/logger.ts index 7eca1ef0..33aa949b 100644 --- a/apps/api/src/lib/logger.ts +++ b/apps/api/src/lib/logger.ts @@ -1,57 +1,82 @@ +import * as winston from "winston"; + import { configDotenv } from "dotenv"; +import Transport from "winston-transport"; configDotenv(); -enum LogLevel { - NONE = 'NONE', // No logs will be output. - ERROR = 'ERROR', // For logging error messages that indicate a failure in a specific operation. - WARN = 'WARN', // For logging potentially harmful situations that are not necessarily errors. - INFO = 'INFO', // For logging informational messages that highlight the progress of the application. - DEBUG = 'DEBUG', // For logging detailed information on the flow through the system, primarily used for debugging. - TRACE = 'TRACE' // For logging more detailed information than the DEBUG level. -} -export class Logger { - static colors = { - ERROR: '\x1b[31m%s\x1b[0m', // Red - WARN: '\x1b[33m%s\x1b[0m', // Yellow - INFO: '\x1b[34m%s\x1b[0m', // Blue - DEBUG: '\x1b[36m%s\x1b[0m', // Cyan - TRACE: '\x1b[35m%s\x1b[0m' // Magenta - }; - - static log (message: string, level: LogLevel) { - const logLevel: LogLevel = LogLevel[process.env.LOGGING_LEVEL as keyof typeof LogLevel] || LogLevel.TRACE; - const levels = [LogLevel.NONE, LogLevel.ERROR, LogLevel.WARN, LogLevel.INFO, LogLevel.DEBUG, LogLevel.TRACE]; - const currentLevelIndex = levels.indexOf(logLevel); - const messageLevelIndex = levels.indexOf(level); - - if (currentLevelIndex >= messageLevelIndex) { - const color = Logger.colors[level]; - console[level.toLowerCase()](color, `[${new Date().toISOString()}]${level} - ${message}`); - - // const useDbAuthentication = process.env.USE_DB_AUTHENTICATION === 'true'; - // if (useDbAuthentication) { - // save to supabase? another place? - // supabase.from('logs').insert({ level: level, message: message, timestamp: new Date().toISOString(), success: boolean }); - // } +const logFormat = winston.format.printf(info => + `${info.timestamp} ${info.level} [${info.metadata.module ?? ""}:${info.metadata.method ?? ""}]: ${info.message} ${info.level.includes("error") || info.level.includes("warn") ? JSON.stringify( + info.metadata, + (_, value) => { + if (value instanceof Error) { + return { + ...value, + name: value.name, + message: value.message, + stack: value.stack, + cause: value.cause, + } + } else { + return value; + } } - } - static error(message: string | any) { - Logger.log(message, LogLevel.ERROR); + ) : ""}` +) + +export const logger = winston.createLogger({ + level: process.env.LOGGING_LEVEL?.toLowerCase() ?? "debug", + format: winston.format.json({ + replacer(key, value) { + if (value instanceof Error) { + return { + ...value, + name: value.name, + message: value.message, + stack: value.stack, + cause: value.cause, + } + } else { + return value; + } + } + }), + transports: [ + new winston.transports.Console({ + format: winston.format.combine( + winston.format.timestamp({ format: "YYYY-MM-DD HH:mm:ss" }), + winston.format.metadata({ fillExcept: ["message", "level", "timestamp"] }), + ...(((process.env.ENV === "production" && process.env.SENTRY_ENVIRONMENT === "dev") || (process.env.ENV !== "production")) ? [winston.format.colorize(), logFormat] : []), + ), + }), + ], +}); + +export type ArrayTransportOptions = Transport.TransportStreamOptions & { + array: any[]; + scrapeId?: string; +}; + +export class ArrayTransport extends Transport { + private array: any[]; + private scrapeId?: string; + + constructor(opts: ArrayTransportOptions) { + super(opts); + this.array = opts.array; + this.scrapeId = opts.scrapeId; } - static warn(message: string) { - Logger.log(message, LogLevel.WARN); - } + log(info, next) { + setImmediate(() => { + this.emit("logged", info); + }); - static info(message: string) { - Logger.log(message, LogLevel.INFO); - } + if (this.scrapeId !== undefined && info.scrapeId !== this.scrapeId) { + return next(); + } - static debug(message: string) { - Logger.log(message, LogLevel.DEBUG); - } + this.array.push(info); - static trace(message: string) { - Logger.log(message, LogLevel.TRACE); + next(); } -} +} \ No newline at end of file diff --git a/apps/api/src/lib/map-cosine.ts b/apps/api/src/lib/map-cosine.ts index db2491a9..2a089548 100644 --- a/apps/api/src/lib/map-cosine.ts +++ b/apps/api/src/lib/map-cosine.ts @@ -1,4 +1,4 @@ -import { Logger } from "./logger"; +import { logger } from "./logger"; export function performCosineSimilarity(links: string[], searchQuery: string) { try { @@ -40,7 +40,7 @@ export function performCosineSimilarity(links: string[], searchQuery: string) { links = a.map((item) => item.link); return links; } catch (error) { - Logger.error(`Error performing cosine similarity: ${error}`); + logger.error(`Error performing cosine similarity: ${error}`); return links; } } diff --git a/apps/api/src/lib/scrape-events.ts b/apps/api/src/lib/scrape-events.ts index ad70dfef..83873a58 100644 --- a/apps/api/src/lib/scrape-events.ts +++ b/apps/api/src/lib/scrape-events.ts @@ -1,8 +1,8 @@ import { Job } from "bullmq"; -import type { baseScrapers } from "../scraper/WebScraper/single_url"; import { supabase_service as supabase } from "../services/supabase"; -import { Logger } from "./logger"; +import { logger } from "./logger"; import { configDotenv } from "dotenv"; +import { Engine } from "../scraper/scrapeURL/engines"; configDotenv(); export type ScrapeErrorEvent = { @@ -15,7 +15,7 @@ export type ScrapeScrapeEvent = { type: "scrape", url: string, worker?: string, - method: (typeof baseScrapers)[number], + method: Engine, result: null | { success: boolean, response_code?: number, @@ -49,7 +49,7 @@ export class ScrapeEvents { }).select().single(); return (result.data as any).id; } catch (error) { - // Logger.error(`Error inserting scrape event: ${error}`); + // logger.error(`Error inserting scrape event: ${error}`); return null; } } @@ -69,7 +69,7 @@ export class ScrapeEvents { } }).eq("id", logId); } catch (error) { - Logger.error(`Error updating scrape result: ${error}`); + logger.error(`Error updating scrape result: ${error}`); } } @@ -81,7 +81,7 @@ export class ScrapeEvents { worker: process.env.FLY_MACHINE_ID, }); } catch (error) { - Logger.error(`Error logging job event: ${error}`); + logger.error(`Error logging job event: ${error}`); } } } diff --git a/apps/api/src/lib/supabase-jobs.ts b/apps/api/src/lib/supabase-jobs.ts index c418a6e0..c9be72a3 100644 --- a/apps/api/src/lib/supabase-jobs.ts +++ b/apps/api/src/lib/supabase-jobs.ts @@ -1,5 +1,5 @@ import { supabase_service } from "../services/supabase"; -import { Logger } from "./logger"; +import { logger } from "./logger"; import * as Sentry from "@sentry/node"; /** @@ -37,7 +37,7 @@ export const supabaseGetJobsById = async (jobIds: string[]) => { .in("job_id", jobIds); if (error) { - Logger.error(`Error in supabaseGetJobsById: ${error}`); + logger.error(`Error in supabaseGetJobsById: ${error}`); Sentry.captureException(error); return []; } @@ -61,7 +61,7 @@ export const supabaseGetJobsByCrawlId = async (crawlId: string) => { .eq("crawl_id", crawlId) if (error) { - Logger.error(`Error in supabaseGetJobsByCrawlId: ${error}`); + logger.error(`Error in supabaseGetJobsByCrawlId: ${error}`); Sentry.captureException(error); return []; } diff --git a/apps/api/src/lib/withAuth.ts b/apps/api/src/lib/withAuth.ts index b45b8973..a6cd539d 100644 --- a/apps/api/src/lib/withAuth.ts +++ b/apps/api/src/lib/withAuth.ts @@ -1,30 +1,25 @@ import { AuthResponse } from "../../src/types"; -import { Logger } from "./logger"; +import { logger } from "./logger"; import * as Sentry from "@sentry/node"; import { configDotenv } from "dotenv"; configDotenv(); let warningCount = 0; -export function withAuth( - originalFunction: (...args: U) => Promise +export function withAuth( + originalFunction: (...args: U) => Promise, + mockSuccess: T, ) { return async function (...args: U): Promise { const useDbAuthentication = process.env.USE_DB_AUTHENTICATION === 'true'; if (!useDbAuthentication) { if (warningCount < 5) { - Logger.warn("You're bypassing authentication"); + logger.warn("You're bypassing authentication"); warningCount++; } return { success: true } as T; } else { - try { - return await originalFunction(...args); - } catch (error) { - Sentry.captureException(error); - Logger.error(`Error in withAuth function: ${error}`); - return { success: false, error: error.message } as T; - } + return await originalFunction(...args); } }; } diff --git a/apps/api/src/main/runWebScraper.ts b/apps/api/src/main/runWebScraper.ts index 8bd0c12c..90d4a47f 100644 --- a/apps/api/src/main/runWebScraper.ts +++ b/apps/api/src/main/runWebScraper.ts @@ -1,151 +1,127 @@ import { Job } from "bullmq"; import { - CrawlResult, WebScraperOptions, RunWebScraperParams, RunWebScraperResult, } from "../types"; -import { WebScraperDataProvider } from "../scraper/WebScraper"; -import { DocumentUrl, Progress } from "../lib/entities"; import { billTeam } from "../services/billing/credit_billing"; -import { Document } from "../lib/entities"; +import { Document } from "../controllers/v1/types"; import { supabase_service } from "../services/supabase"; -import { Logger } from "../lib/logger"; +import { logger } from "../lib/logger"; import { ScrapeEvents } from "../lib/scrape-events"; import { configDotenv } from "dotenv"; +import { EngineResultsTracker, scrapeURL, ScrapeUrlResponse } from "../scraper/scrapeURL"; +import { Engine } from "../scraper/scrapeURL/engines"; configDotenv(); export async function startWebScraperPipeline({ job, token, }: { - job: Job; + job: Job & { id: string }; token: string; }) { - let partialDocs: Document[] = []; return (await runWebScraper({ url: job.data.url, mode: job.data.mode, - crawlerOptions: job.data.crawlerOptions, - extractorOptions: job.data.extractorOptions, - pageOptions: { - ...job.data.pageOptions, + scrapeOptions: { + ...job.data.scrapeOptions, ...(job.data.crawl_id ? ({ - includeRawHtml: true, + formats: job.data.scrapeOptions.formats.concat(["rawHtml"]), }): {}), }, - inProgress: (progress) => { - Logger.debug(`🐂 Job in progress ${job.id}`); - if (progress.currentDocument) { - partialDocs.push(progress.currentDocument); - if (partialDocs.length > 50) { - partialDocs = partialDocs.slice(-50); - } - // job.updateProgress({ ...progress, partialDocs: partialDocs }); - } - }, - onSuccess: (result, mode) => { - Logger.debug(`🐂 Job completed ${job.id}`); - saveJob(job, result, token, mode); - }, - onError: (error) => { - Logger.error(`🐂 Job failed ${job.id}`); - ScrapeEvents.logJobEvent(job, "failed"); - job.moveToFailed(error, token, false); - }, + internalOptions: job.data.internalOptions, + // onSuccess: (result, mode) => { + // logger.debug(`🐂 Job completed ${job.id}`); + // saveJob(job, result, token, mode); + // }, + // onError: (error) => { + // logger.error(`🐂 Job failed ${job.id}`); + // ScrapeEvents.logJobEvent(job, "failed"); + // }, team_id: job.data.team_id, bull_job_id: job.id.toString(), priority: job.opts.priority, is_scrape: job.data.is_scrape ?? false, - })) as { success: boolean; message: string; docs: Document[] }; + })); } export async function runWebScraper({ url, mode, - crawlerOptions, - pageOptions, - extractorOptions, - inProgress, - onSuccess, - onError, + scrapeOptions, + internalOptions, + // onSuccess, + // onError, team_id, bull_job_id, priority, is_scrape=false, -}: RunWebScraperParams): Promise { +}: RunWebScraperParams): Promise { + let response: ScrapeUrlResponse | undefined = undefined; + let engines: EngineResultsTracker = {}; try { - const provider = new WebScraperDataProvider(); - if (mode === "crawl") { - await provider.setOptions({ - jobId: bull_job_id, - mode: mode, - urls: [url], - extractorOptions, - crawlerOptions: crawlerOptions, - pageOptions: pageOptions, - bullJobId: bull_job_id, - priority, - }); - } else { - await provider.setOptions({ - jobId: bull_job_id, - mode: mode, - urls: url.split(","), - extractorOptions, - crawlerOptions: crawlerOptions, - pageOptions: pageOptions, - priority, - teamId: team_id - }); + response = await scrapeURL(bull_job_id, url, scrapeOptions, { priority, ...internalOptions }); + if (!response.success) { + if (response.error instanceof Error) { + throw response.error; + } else { + throw new Error("scrapeURL error: " + (Array.isArray(response.error) ? JSON.stringify(response.error) : typeof response.error === "object" ? JSON.stringify({ ...response.error }) : response.error)); + } } - const docs = (await provider.getDocuments(false, (progress: Progress) => { - inProgress(progress); - })) as Document[]; - - if (docs.length === 0) { - return { - success: true, - message: "No pages found", - docs: [], - }; - } - - // remove docs with empty content - const filteredDocs = crawlerOptions?.returnOnlyUrls - ? docs.map((doc) => { - if (doc.metadata.sourceURL) { - return { url: doc.metadata.sourceURL }; - } - }) - : docs; if(is_scrape === false) { let creditsToBeBilled = 1; // Assuming 1 credit per document - if (extractorOptions && (extractorOptions.mode === "llm-extraction" || extractorOptions.mode === "extract")) { + if (scrapeOptions.extract) { creditsToBeBilled = 5; } - billTeam(team_id, undefined, creditsToBeBilled * filteredDocs.length).catch(error => { - Logger.error(`Failed to bill team ${team_id} for ${creditsToBeBilled * filteredDocs.length} credits: ${error}`); + billTeam(team_id, undefined, creditsToBeBilled).catch(error => { + logger.error(`Failed to bill team ${team_id} for ${creditsToBeBilled} credits: ${error}`); // Optionally, you could notify an admin or add to a retry queue here }); } - - // This is where the returnvalue from the job is set - onSuccess(filteredDocs, mode); + // onSuccess(response.document, mode); - // this return doesn't matter too much for the job completion result - return { success: true, message: "", docs: filteredDocs }; + engines = response.engines; + return response; } catch (error) { - onError(error); - return { success: false, message: error.message, docs: [] }; + engines = response !== undefined ? response.engines : ((typeof error === "object" && error !== null ? (error as any).results ?? {} : {})); + + if (response !== undefined) { + return { + ...response, + success: false, + error, + } + } else { + return { success: false, error, logs: ["no logs -- error coming from runWebScraper"], engines }; + } + // onError(error); + } finally { + const engineOrder = Object.entries(engines).sort((a, b) => a[1].startedAt - b[1].startedAt).map(x => x[0]) as Engine[]; + + for (const engine of engineOrder) { + const result = engines[engine] as Exclude; + ScrapeEvents.insert(bull_job_id, { + type: "scrape", + url, + method: engine, + result: { + success: result.state === "success", + response_code: (result.state === "success" ? result.result.statusCode : undefined), + response_size: (result.state === "success" ? result.result.html.length : undefined), + error: (result.state === "error" ? result.error : result.state === "timeout" ? "Timed out" : undefined), + time_taken: result.finishedAt - result.startedAt, + }, + }); + } } } -const saveJob = async (job: Job, result: any, token: string, mode: string) => { +const saveJob = async (job: Job, result: any, token: string, mode: string, engines?: EngineResultsTracker) => { try { const useDbAuthentication = process.env.USE_DB_AUTHENTICATION === 'true'; if (useDbAuthentication) { @@ -173,6 +149,6 @@ const saveJob = async (job: Job, result: any, token: string, mode: string) => { } ScrapeEvents.logJobEvent(job, "completed"); } catch (error) { - Logger.error(`🐂 Failed to update job status: ${error}`); + logger.error(`🐂 Failed to update job status: ${error}`); } }; diff --git a/apps/api/src/routes/admin.ts b/apps/api/src/routes/admin.ts index 88159060..ac61519a 100644 --- a/apps/api/src/routes/admin.ts +++ b/apps/api/src/routes/admin.ts @@ -6,8 +6,8 @@ import { cleanBefore24hCompleteJobsController, queuesController, } from "../controllers/v0/admin/queue"; -import { acucCacheClearController } from "../controllers/v0/admin/acuc-cache-clear"; import { wrap } from "./v1"; +import { acucCacheClearController } from "../controllers/v0/admin/acuc-cache-clear"; export const adminRouter = express.Router(); diff --git a/apps/api/src/routes/v1.ts b/apps/api/src/routes/v1.ts index 4e4b6052..3eaace3b 100644 --- a/apps/api/src/routes/v1.ts +++ b/apps/api/src/routes/v1.ts @@ -14,7 +14,7 @@ import expressWs from "express-ws"; import { crawlStatusWSController } from "../controllers/v1/crawl-status-ws"; import { isUrlBlocked } from "../scraper/WebScraper/utils/blocklist"; import { crawlCancelController } from "../controllers/v1/crawl-cancel"; -import { Logger } from "../lib/logger"; +import { logger } from "../lib/logger"; import { scrapeStatusController } from "../controllers/v1/scrape-status"; import { concurrencyCheckController } from "../controllers/v1/concurrency-check"; import { batchScrapeController } from "../controllers/v1/batch-scrape"; @@ -32,10 +32,12 @@ function checkCreditsMiddleware(minimum?: number): (req: RequestWithAuth, res: R if (!minimum && req.body) { minimum = (req.body as any)?.limit ?? (req.body as any)?.urls?.length ?? 1; } - const { success, remainingCredits, chunk } = await checkTeamCredits(req.acuc, req.auth.team_id, minimum); - req.acuc = chunk; + const { success, remainingCredits, chunk } = await checkTeamCredits(req.acuc, req.auth.team_id, minimum ?? 1); + if (chunk) { + req.acuc = chunk; + } if (!success) { - Logger.error(`Insufficient credits: ${JSON.stringify({ team_id: req.auth.team_id, minimum, remainingCredits })}`); + logger.error(`Insufficient credits: ${JSON.stringify({ team_id: req.auth.team_id, minimum, remainingCredits })}`); if (!res.headersSent) { return res.status(402).json({ success: false, error: "Insufficient credits to perform this request. For more credits, you can upgrade your plan at https://firecrawl.dev/pricing or try changing the request limit to a lower value." }); } @@ -50,20 +52,27 @@ function checkCreditsMiddleware(minimum?: number): (req: RequestWithAuth, res: R export function authMiddleware(rateLimiterMode: RateLimiterMode): (req: RequestWithMaybeAuth, res: Response, next: NextFunction) => void { return (req, res, next) => { (async () => { - const { success, team_id, error, status, plan, chunk } = await authenticateUser( + const auth = await authenticateUser( req, res, rateLimiterMode, ); - if (!success) { + if (!auth.success) { if (!res.headersSent) { - return res.status(status).json({ success: false, error }); + return res.status(auth.status).json({ success: false, error: auth.error }); + } else { + return; } } + const { team_id, plan, chunk } = auth; + req.auth = { team_id, plan }; - req.acuc = chunk; + req.acuc = chunk ?? undefined; + if (chunk) { + req.account = { remainingCredits: chunk.remaining_credits }; + } next(); })() .catch(err => next(err)); diff --git a/apps/api/src/scraper/WebScraper/__tests__/crawler.test.ts b/apps/api/src/scraper/WebScraper/__tests__/crawler.test.ts index 20419ffa..eba0ddb4 100644 --- a/apps/api/src/scraper/WebScraper/__tests__/crawler.test.ts +++ b/apps/api/src/scraper/WebScraper/__tests__/crawler.test.ts @@ -2,7 +2,6 @@ import { WebCrawler } from '../crawler'; import axios from 'axios'; import robotsParser from 'robots-parser'; -import { getAdjustedMaxDepth } from '../utils/maxDepthUtils'; jest.mock('axios'); jest.mock('robots-parser'); @@ -35,165 +34,6 @@ describe('WebCrawler', () => { }); }); - it('should filter out links that exceed maxDepth param of 2 based on enterURL depth of 0 ', async () => { - const initialUrl = 'http://example.com'; // Set initial URL for this test - const enteredMaxCrawledDepth = 2; - maxCrawledDepth = getAdjustedMaxDepth(initialUrl, enteredMaxCrawledDepth); - - - crawler = new WebCrawler({ - jobId: "TEST", - initialUrl: initialUrl, - includes: [], - excludes: [], - limit: 100, - maxCrawledDepth: maxCrawledDepth, // Set maxDepth for testing - }); - - // Mock sitemap fetching function to return controlled links - crawler['tryFetchSitemapLinks'] = jest.fn().mockResolvedValue([ - initialUrl, // depth 0 - initialUrl + '/page1', // depth 1 - initialUrl + '/page1/page2', // depth 2 - initialUrl + '/page1/page2/page3' // depth 3, should be filtered out - ]); - - const results = await crawler.start(undefined, undefined, undefined, undefined, undefined, maxCrawledDepth); - expect(results).toEqual([ - { url: initialUrl, html: '' }, - { url: initialUrl + '/page1', html: '' }, - { url: initialUrl + '/page1/page2', html: '' } - ]); - - - // Ensure that the link with depth 3 is not included - expect(results.some(r => r.url === initialUrl + '/page1/page2/page3')).toBe(false); - }); - - it('should filter out links that exceed maxDepth param of 0 based on enterURL depth of 0 ', async () => { - const initialUrl = 'http://example.com'; // Set initial URL for this test - const enteredMaxCrawledDepth = 0; - maxCrawledDepth = getAdjustedMaxDepth(initialUrl, enteredMaxCrawledDepth); - - - crawler = new WebCrawler({ - jobId: "TEST", - initialUrl: initialUrl, - includes: [], - excludes: [], - limit: 100, - maxCrawledDepth: maxCrawledDepth, // Set maxDepth for testing - }); - - // Mock sitemap fetching function to return controlled links - crawler['tryFetchSitemapLinks'] = jest.fn().mockResolvedValue([ - initialUrl, // depth 0 - initialUrl + '/page1', // depth 1 - initialUrl + '/page1/page2', // depth 2 - initialUrl + '/page1/page2/page3' // depth 3, should be filtered out - ]); - - const results = await crawler.start(undefined, undefined, undefined, undefined, undefined, maxCrawledDepth); - expect(results).toEqual([ - { url: initialUrl, html: '' }, - ]); - }); - - it('should filter out links that exceed maxDepth param of 1 based on enterURL depth of 1 ', async () => { - const initialUrl = 'http://example.com/page1'; // Set initial URL for this test - const enteredMaxCrawledDepth = 1; - maxCrawledDepth = getAdjustedMaxDepth(initialUrl, enteredMaxCrawledDepth); - - - crawler = new WebCrawler({ - jobId: "TEST", - initialUrl: initialUrl, - includes: [], - excludes: [], - limit: 100, - maxCrawledDepth: maxCrawledDepth, // Set maxDepth for testing - }); - - // Mock sitemap fetching function to return controlled links - crawler['tryFetchSitemapLinks'] = jest.fn().mockResolvedValue([ - initialUrl, // depth 0 - initialUrl + '/page2', // depth 1 - initialUrl + '/page2/page3', // depth 2 - initialUrl + '/page2/page3/page4' // depth 3, should be filtered out - ]); - - const results = await crawler.start(undefined, undefined, undefined, undefined, undefined, maxCrawledDepth); - expect(results).toEqual([ - { url: initialUrl, html: '' }, - { url: initialUrl + '/page2', html: '' } - ]); - }); - - it('should filter out links that exceed maxDepth param of 1 based on enterURL depth of 2 ', async () => { - const initialUrl = 'http://example.com/page1'; // Set initial URL for this test - const enteredMaxCrawledDepth = 2; - maxCrawledDepth = getAdjustedMaxDepth(initialUrl, enteredMaxCrawledDepth); - - - crawler = new WebCrawler({ - jobId: "TEST", - initialUrl: initialUrl, - includes: [], - excludes: [], - limit: 100, - maxCrawledDepth: maxCrawledDepth, // Set maxDepth for testing - }); - - // Mock sitemap fetching function to return controlled links - crawler['tryFetchSitemapLinks'] = jest.fn().mockResolvedValue([ - initialUrl, // depth 0 - initialUrl + '/page2', // depth 1 - initialUrl + '/page2/page3', // depth 2 - initialUrl + '/page2/page3/page4' // depth 3, should be filtered out - ]); - - const results = await crawler.start(undefined, undefined, undefined, undefined, undefined, maxCrawledDepth); - expect(results).toEqual([ - { url: initialUrl, html: '' }, - { url: initialUrl + '/page2', html: '' }, - { url: initialUrl + '/page2/page3', html: '' } - ]); - }); - - it('should handle allowBackwardCrawling option correctly', async () => { - const initialUrl = 'https://mendable.ai/blog'; - - // Setup the crawler with the specific test case options - const crawler = new WebCrawler({ - jobId: "TEST", - initialUrl: initialUrl, - includes: [], - excludes: [], - limit: 100, - maxCrawledDepth: 3, // Example depth - allowBackwardCrawling: true - }); - - // Mock the sitemap fetching function to simulate backward crawling - crawler['tryFetchSitemapLinks'] = jest.fn().mockResolvedValue([ - initialUrl, - 'https://mendable.ai', // backward link - initialUrl + '/page1', - initialUrl + '/page1/page2' - ]); - - const results = await crawler.start(); - expect(results).toEqual([ - { url: initialUrl, html: '' }, - { url: 'https://mendable.ai', html: '' }, // Expect the backward link to be included - { url: initialUrl + '/page1', html: '' }, - { url: initialUrl + '/page1/page2', html: '' } - ]); - - // Check that the backward link is included if allowBackwardCrawling is true - expect(results.some(r => r.url === 'https://mendable.ai')).toBe(true); - }); - it('should respect the limit parameter by not returning more links than specified', async () => { const initialUrl = 'http://example.com'; const limit = 2; // Set a limit for the number of links diff --git a/apps/api/src/scraper/WebScraper/__tests__/single_url.test.ts b/apps/api/src/scraper/WebScraper/__tests__/single_url.test.ts deleted file mode 100644 index 02c8a7e0..00000000 --- a/apps/api/src/scraper/WebScraper/__tests__/single_url.test.ts +++ /dev/null @@ -1,37 +0,0 @@ -import { scrapSingleUrl } from '../single_url'; -import { PageOptions } from '../../../lib/entities'; - - -jest.mock('../single_url', () => { - const originalModule = jest.requireActual('../single_url'); - originalModule.fetchHtmlContent = jest.fn().mockResolvedValue('Test

Roast

'); - - return originalModule; -}); - -describe('scrapSingleUrl', () => { - it('should handle includeHtml option correctly', async () => { - const url = 'https://roastmywebsite.ai'; - const pageOptionsWithHtml: PageOptions = { includeHtml: true }; - const pageOptionsWithoutHtml: PageOptions = { includeHtml: false }; - - const resultWithHtml = await scrapSingleUrl("TEST", url, pageOptionsWithHtml); - const resultWithoutHtml = await scrapSingleUrl("TEST", url, pageOptionsWithoutHtml); - - expect(resultWithHtml.html).toBeDefined(); - expect(resultWithoutHtml.html).toBeUndefined(); - }, 10000); -}); - -it('should return a list of links on the firecrawl.ai page', async () => { - const url = 'https://flutterbricks.com'; - const pageOptions: PageOptions = { includeHtml: true }; - - const result = await scrapSingleUrl("TEST", url, pageOptions); - - // Check if the result contains a list of links - expect(result.linksOnPage).toBeDefined(); - expect(Array.isArray(result.linksOnPage)).toBe(true); - expect(result.linksOnPage.length).toBeGreaterThan(0); - expect(result.linksOnPage).toContain('https://flutterbricks.com/features') -}, 15000); diff --git a/apps/api/src/scraper/WebScraper/crawler.ts b/apps/api/src/scraper/WebScraper/crawler.ts index 1ae166b4..9e3f7cd2 100644 --- a/apps/api/src/scraper/WebScraper/crawler.ts +++ b/apps/api/src/scraper/WebScraper/crawler.ts @@ -2,13 +2,10 @@ import axios, { AxiosError } from "axios"; import cheerio, { load } from "cheerio"; import { URL } from "url"; import { getLinksFromSitemap } from "./sitemap"; -import async from "async"; -import { CrawlerOptions, PageOptions, Progress } from "../../lib/entities"; -import { scrapSingleUrl } from "./single_url"; import robotsParser from "robots-parser"; import { getURLDepth } from "./utils/maxDepthUtils"; import { axiosTimeout } from "../../../src/lib/timeout"; -import { Logger } from "../../../src/lib/logger"; +import { logger } from "../../../src/lib/logger"; import https from "https"; export class WebCrawler { private jobId: string; @@ -73,7 +70,7 @@ export class WebCrawler { try { url = new URL(link.trim(), this.baseUrl); } catch (error) { - Logger.debug(`Error processing link: ${link} | Error: ${error.message}`); + logger.debug(`Error processing link: ${link} | Error: ${error.message}`); return false; } const path = url.pathname; @@ -132,7 +129,7 @@ export class WebCrawler { const isAllowed = this.robots.isAllowed(link, "FireCrawlAgent") ?? true; // Check if the link is disallowed by robots.txt if (!isAllowed) { - Logger.debug(`Link disallowed by robots.txt: ${link}`); + logger.debug(`Link disallowed by robots.txt: ${link}`); return false; } @@ -161,7 +158,7 @@ export class WebCrawler { } public async tryGetSitemap(): Promise<{ url: string; html: string; }[] | null> { - Logger.debug(`Fetching sitemap links from ${this.initialUrl}`); + logger.debug(`Fetching sitemap links from ${this.initialUrl}`); const sitemapLinks = await this.tryFetchSitemapLinks(this.initialUrl); if (sitemapLinks.length > 0) { let filteredLinks = this.filterLinks(sitemapLinks, this.limit, this.maxCrawledDepth); @@ -170,115 +167,6 @@ export class WebCrawler { return null; } - public async start( - inProgress?: (progress: Progress) => void, - pageOptions?: PageOptions, - crawlerOptions?: CrawlerOptions, - concurrencyLimit: number = 5, - limit: number = 10000, - maxDepth: number = 10 - ): Promise<{ url: string, html: string }[]> { - - Logger.debug(`Crawler starting with ${this.initialUrl}`); - // Fetch and parse robots.txt - try { - const txt = await this.getRobotsTxt(); - this.importRobotsTxt(txt); - Logger.debug(`Crawler robots.txt fetched with ${this.robotsTxtUrl}`); - } catch (error) { - Logger.debug(`Failed to fetch robots.txt from ${this.robotsTxtUrl}`); - } - - if (!crawlerOptions?.ignoreSitemap){ - const sm = await this.tryGetSitemap(); - if (sm !== null) { - return sm; - } - } - - const urls = await this.crawlUrls( - [this.initialUrl], - pageOptions, - concurrencyLimit, - inProgress - ); - - if ( - urls.length === 0 && - this.filterLinks([this.initialUrl], limit, this.maxCrawledDepth).length > 0 - ) { - return [{ url: this.initialUrl, html: "" }]; - } - - // make sure to run include exclude here again - const filteredUrls = this.filterLinks(urls.map(urlObj => urlObj.url), limit, this.maxCrawledDepth); - return filteredUrls.map(url => ({ url, html: urls.find(urlObj => urlObj.url === url)?.html || "" })); - } - - private async crawlUrls( - urls: string[], - pageOptions: PageOptions, - concurrencyLimit: number, - inProgress?: (progress: Progress) => void, - ): Promise<{ url: string, html: string }[]> { - const queue = async.queue(async (task: string, callback) => { - Logger.debug(`Crawling ${task}`); - if (this.crawledUrls.size >= Math.min(this.maxCrawledLinks, this.limit)) { - if (callback && typeof callback === "function") { - callback(); - } - return; - } - const newUrls = await this.crawl(task, pageOptions); - // add the initial url if not already added - // if (this.visited.size === 1) { - // let normalizedInitial = this.initialUrl; - // if (!normalizedInitial.endsWith("/")) { - // normalizedInitial = normalizedInitial + "/"; - // } - // if (!newUrls.some(page => page.url === this.initialUrl)) { - // newUrls.push({ url: this.initialUrl, html: "" }); - // } - // } - - newUrls.forEach((page) => this.crawledUrls.set(page.url, page.html)); - - if (inProgress && newUrls.length > 0) { - inProgress({ - current: this.crawledUrls.size, - total: Math.min(this.maxCrawledLinks, this.limit), - status: "SCRAPING", - currentDocumentUrl: newUrls[newUrls.length - 1].url, - }); - } else if (inProgress) { - inProgress({ - current: this.crawledUrls.size, - total: Math.min(this.maxCrawledLinks, this.limit), - status: "SCRAPING", - currentDocumentUrl: task, - }); - } - await this.crawlUrls(newUrls.map((p) => p.url), pageOptions, concurrencyLimit, inProgress); - if (callback && typeof callback === "function") { - callback(); - } - }, concurrencyLimit); - - Logger.debug(`🐂 Pushing ${urls.length} URLs to the queue`); - queue.push( - urls.filter( - (url) => - !this.visited.has(url) && this.robots.isAllowed(url, "FireCrawlAgent") - ), - (err) => { - if (err) Logger.error(`🐂 Error pushing URLs to the queue: ${err}`); - } - ); - await queue.drain(); - Logger.debug(`🐂 Crawled ${this.crawledUrls.size} URLs, Queue drained.`); - return Array.from(this.crawledUrls.entries()).map(([url, html]) => ({ url, html })); - } - public filterURL(href: string, url: string): string | null { let fullUrl = href; if (!href.startsWith("http")) { @@ -346,79 +234,9 @@ export class WebCrawler { return links; } - async crawl(url: string, pageOptions: PageOptions): Promise<{url: string, html: string, pageStatusCode?: number, pageError?: string}[]> { - if (this.visited.has(url) || !this.robots.isAllowed(url, "FireCrawlAgent")) { - return []; - } - this.visited.add(url); - - if (!url.startsWith("http")) { - url = "https://" + url; - } - if (url.endsWith("/")) { - url = url.slice(0, -1); - } - - if (this.isFile(url) || this.isSocialMediaOrEmail(url)) { - return []; - } - - try { - let content: string = ""; - let pageStatusCode: number; - let pageError: string | undefined = undefined; - - // If it is the first link, fetch with single url - if (this.visited.size === 1) { - const page = await scrapSingleUrl(this.jobId, url, { ...pageOptions, includeHtml: true }); - content = page.html ?? ""; - pageStatusCode = page.metadata?.pageStatusCode; - pageError = page.metadata?.pageError || undefined; - } else { - const response = await axios.get(url, { timeout: axiosTimeout }); - content = response.data ?? ""; - pageStatusCode = response.status; - pageError = response.statusText != "OK" ? response.statusText : undefined; - } - - const $ = load(content); - let links: { url: string, html: string, pageStatusCode?: number, pageError?: string }[] = []; - - // Add the initial URL to the list of links - if (this.visited.size === 1) { - links.push({ url, html: content, pageStatusCode, pageError }); - } - - links.push(...this.extractLinksFromHTML(content, url).map(url => ({ url, html: content, pageStatusCode, pageError }))); - - if (this.visited.size === 1) { - return links; - } - - // Create a new list to return to avoid modifying the visited list - return links.filter((link) => !this.visited.has(link.url)); - } catch (error) { - return []; - } - } - private isRobotsAllowed(url: string): boolean { return (this.robots ? (this.robots.isAllowed(url, "FireCrawlAgent") ?? true) : true) } - private normalizeCrawlUrl(url: string): string { - try{ - const urlObj = new URL(url); - urlObj.searchParams.sort(); // Sort query parameters to normalize - return urlObj.toString(); - } catch (error) { - return url; - } - } - - private matchesIncludes(url: string): boolean { - if (this.includes.length === 0 || this.includes[0] == "") return true; - return this.includes.some((pattern) => new RegExp(pattern).test(url)); - } private matchesExcludes(url: string, onlyDomains: boolean = false): boolean { return this.excludes.some((pattern) => { @@ -503,7 +321,7 @@ export class WebCrawler { const urlWithoutQuery = url.split('?')[0].toLowerCase(); return fileExtensions.some((ext) => urlWithoutQuery.endsWith(ext)); } catch (error) { - Logger.error(`Error processing URL in isFile: ${error}`); + logger.error(`Error processing URL in isFile: ${error}`); return false; } } @@ -524,7 +342,6 @@ export class WebCrawler { return socialMediaOrEmail.some((ext) => url.includes(ext)); } - // private async tryFetchSitemapLinks(url: string): Promise { const normalizeUrl = (url: string) => { url = url.replace(/^https?:\/\//, "").replace(/^www\./, ""); @@ -546,7 +363,7 @@ export class WebCrawler { sitemapLinks = await getLinksFromSitemap({ sitemapUrl }); } } catch (error) { - Logger.debug(`Failed to fetch sitemap with axios from ${sitemapUrl}: ${error}`); + logger.debug(`Failed to fetch sitemap with axios from ${sitemapUrl}: ${error}`); if (error instanceof AxiosError && error.response?.status === 404) { // ignore 404 } else { @@ -565,7 +382,7 @@ export class WebCrawler { sitemapLinks = await getLinksFromSitemap({ sitemapUrl: baseUrlSitemap, mode: 'fire-engine' }); } } catch (error) { - Logger.debug(`Failed to fetch sitemap from ${baseUrlSitemap}: ${error}`); + logger.debug(`Failed to fetch sitemap from ${baseUrlSitemap}: ${error}`); if (error instanceof AxiosError && error.response?.status === 404) { // ignore 404 } else { diff --git a/apps/api/src/scraper/WebScraper/custom/handleCustomScraping.ts b/apps/api/src/scraper/WebScraper/custom/handleCustomScraping.ts index e5841978..48aa2ffd 100644 --- a/apps/api/src/scraper/WebScraper/custom/handleCustomScraping.ts +++ b/apps/api/src/scraper/WebScraper/custom/handleCustomScraping.ts @@ -1,4 +1,4 @@ -import { Logger } from "../../../lib/logger"; +import { logger } from "../../../lib/logger"; export async function handleCustomScraping( text: string, @@ -6,7 +6,7 @@ export async function handleCustomScraping( ): Promise<{ scraper: string; url: string; waitAfterLoad?: number, pageOptions?: { scrollXPaths?: string[] } } | null> { // Check for Readme Docs special case if (text.includes(' { - throw new Error("Method not implemented."); - } - - private async convertUrlsToDocuments( - urls: string[], - inProgress?: (progress: Progress) => void, - allHtmls?: string[] - ): Promise { - const totalUrls = urls.length; - let processedUrls = 0; - - const results: (Document | null)[] = new Array(urls.length).fill(null); - for (let i = 0; i < urls.length; i += this.concurrentRequests) { - const batchUrls = urls.slice(i, i + this.concurrentRequests); - await Promise.all( - batchUrls.map(async (url, index) => { - const existingHTML = allHtmls ? allHtmls[i + index] : ""; - const result = await scrapSingleUrl( - this.jobId, - url, - this.pageOptions, - this.extractorOptions, - existingHTML, - this.priority, - this.teamId, - ); - processedUrls++; - if (inProgress) { - inProgress({ - current: processedUrls, - total: totalUrls, - status: "SCRAPING", - currentDocumentUrl: url, - currentDocument: { ...result, index: processedUrls }, - }); - } - - results[i + index] = result; - }) - ); - } - return results.filter((result) => result !== null) as Document[]; - } - - async getDocuments( - useCaching: boolean = false, - inProgress?: (progress: Progress) => void - ): Promise { - this.validateInitialUrl(); - if (!useCaching) { - return this.processDocumentsWithoutCache(inProgress); - } - - return this.processDocumentsWithCache(inProgress); - } - - private validateInitialUrl(): void { - if (this.urls[0].trim() === "") { - throw new Error("Url is required"); - } - } - - /** - * Process documents without cache handling each mode - * @param inProgress inProgress - * @returns documents - */ - private async processDocumentsWithoutCache( - inProgress?: (progress: Progress) => void - ): Promise { - switch (this.mode) { - case "crawl": - return this.handleCrawlMode(inProgress); - case "single_urls": - return this.handleSingleUrlsMode(inProgress); - case "sitemap": - return this.handleSitemapMode(inProgress); - default: - return []; - } - } - - private async cleanIrrelevantPath(links: string[]) { - return links.filter((link) => { - const normalizedInitialUrl = new URL(this.urls[0]); - const normalizedLink = new URL(link); - - // Normalize the hostname to account for www and non-www versions - const initialHostname = normalizedInitialUrl.hostname.replace( - /^www\./, - "" - ); - const linkHostname = normalizedLink.hostname.replace(/^www\./, ""); - - // Ensure the protocol and hostname match, and the path starts with the initial URL's path - return ( - linkHostname === initialHostname && - normalizedLink.pathname.startsWith(normalizedInitialUrl.pathname) - ); - }); - } - - private async handleCrawlMode( - inProgress?: (progress: Progress) => void - ): Promise { - let includes: string[]; - if (Array.isArray(this.includes)) { - if (this.includes[0] != "") { - includes = this.includes; - } - } else { - includes = this.includes.split(','); - } - - let excludes: string[]; - if (Array.isArray(this.excludes)) { - if (this.excludes[0] != "") { - excludes = this.excludes; - } - } else { - excludes = this.excludes.split(','); - } - - const crawler = new WebCrawler({ - jobId: this.jobId, - initialUrl: this.urls[0], - includes, - excludes, - maxCrawledLinks: this.maxCrawledLinks, - maxCrawledDepth: getAdjustedMaxDepth(this.urls[0], this.maxCrawledDepth), - limit: this.limit, - generateImgAltText: this.generateImgAltText, - allowBackwardCrawling: this.allowBackwardCrawling, - allowExternalContentLinks: this.allowExternalContentLinks, - }); - - let links = await crawler.start( - inProgress, - this.pageOptions, - { - ignoreSitemap: this.ignoreSitemap, - }, - 5, - this.limit, - this.maxCrawledDepth - ); - - let allLinks = links.map((e) => e.url); - const allHtmls = links.map((e) => e.html); - - if (this.returnOnlyUrls) { - return this.returnOnlyUrlsResponse(allLinks, inProgress); - } - - let documents = []; - // check if fast mode is enabled and there is html inside the links - if (this.crawlerMode === "fast" && links.some((link) => link.html)) { - documents = await this.processLinks(allLinks, inProgress, allHtmls); - } else { - documents = await this.processLinks(allLinks, inProgress); - } - - return this.cacheAndFinalizeDocuments(documents, allLinks); - } - - private async handleSingleUrlsMode( - inProgress?: (progress: Progress) => void - ): Promise { - const links = this.urls; - - let documents = await this.processLinks(links, inProgress); - return documents; - } - - private async handleSitemapMode( - inProgress?: (progress: Progress) => void - ): Promise { - let links = await getLinksFromSitemap({ sitemapUrl: this.urls[0] }); - links = await this.cleanIrrelevantPath(links); - - if (this.returnOnlyUrls) { - return this.returnOnlyUrlsResponse(links, inProgress); - } - - let documents = await this.processLinks(links, inProgress); - return this.cacheAndFinalizeDocuments(documents, links); - } - - private async returnOnlyUrlsResponse( - links: string[], - inProgress?: (progress: Progress) => void - ): Promise { - inProgress?.({ - current: links.length, - total: links.length, - status: "COMPLETED", - currentDocumentUrl: this.urls[0], - }); - return links.map((url) => ({ - content: "", - html: this.pageOptions?.includeHtml ? "" : undefined, - markdown: "", - metadata: { sourceURL: url, pageStatusCode: 200 }, - })); - } - - private async processLinks( - links: string[], - inProgress?: (progress: Progress) => void, - allHtmls?: string[] - ): Promise { - const pdfLinks = links.filter((link) => link.endsWith(".pdf")); - const docLinks = links.filter( - (link) => link.endsWith(".doc") || link.endsWith(".docx") - ); - - const [pdfDocuments, docxDocuments] = await Promise.all([ - this.fetchPdfDocuments(pdfLinks), - this.fetchDocxDocuments(docLinks), - ]); - - links = links.filter( - (link) => !pdfLinks.includes(link) && !docLinks.includes(link) - ); - - let [documents, sitemapData] = await Promise.all([ - this.convertUrlsToDocuments(links, inProgress, allHtmls), - this.mode === "single_urls" && links.length > 0 - ? this.getSitemapDataForSingleUrl(this.urls[0], links[0], 1500).catch( - (error) => { - Logger.debug(`Failed to fetch sitemap data: ${error}`); - return null; - } - ) - : Promise.resolve(null), - ]); - - if (this.mode === "single_urls" && documents.length > 0) { - documents[0].metadata.sitemap = sitemapData ?? undefined; - } else { - documents = await this.getSitemapData(this.urls[0], documents); - } - - if (this.pageOptions.includeMarkdown) { - documents = this.applyPathReplacements(documents); - } - - if (!this.pageOptions.includeHtml) { - for (let document of documents) { - delete document.html; - } - } - - // documents = await this.applyImgAltText(documents); - if (this.mode === "single_urls" && this.pageOptions.includeExtract) { - const extractionMode = this.extractorOptions?.mode ?? "markdown"; - const completionMode = extractionMode === "llm-extraction-from-raw-html" ? "raw-html" : "markdown"; - - if ( - extractionMode === "llm-extraction" || - extractionMode === "llm-extraction-from-markdown" || - extractionMode === "llm-extraction-from-raw-html" - ) { - documents = await generateCompletions( - documents, - this.extractorOptions, - completionMode - ); - } - } - return documents.concat(pdfDocuments).concat(docxDocuments); - } - - private async fetchPdfDocuments(pdfLinks: string[]): Promise { - return Promise.all( - pdfLinks.map(async (pdfLink) => { - const timer = Date.now(); - const logInsertPromise = ScrapeEvents.insert(this.jobId, { - type: "scrape", - url: pdfLink, - worker: process.env.FLY_MACHINE_ID, - method: "pdf-scrape", - result: null, - }); - - const { content, pageStatusCode, pageError } = await fetchAndProcessPdf( - pdfLink, - this.pageOptions.parsePDF - ); - - const insertedLogId = await logInsertPromise; - ScrapeEvents.updateScrapeResult(insertedLogId, { - response_size: content.length, - success: !(pageStatusCode && pageStatusCode >= 400) && !!content && (content.trim().length >= 100), - error: pageError, - response_code: pageStatusCode, - time_taken: Date.now() - timer, - }); - return { - content: content, - markdown: content, - metadata: { sourceURL: pdfLink, pageStatusCode, pageError }, - provider: "web-scraper", - }; - }) - ); - } - private async fetchDocxDocuments(docxLinks: string[]): Promise { - return Promise.all( - docxLinks.map(async (docxLink) => { - const timer = Date.now(); - const logInsertPromise = ScrapeEvents.insert(this.jobId, { - type: "scrape", - url: docxLink, - worker: process.env.FLY_MACHINE_ID, - method: "docx-scrape", - result: null, - }); - - const { content, pageStatusCode, pageError } = await fetchAndProcessDocx( - docxLink - ); - - const insertedLogId = await logInsertPromise; - ScrapeEvents.updateScrapeResult(insertedLogId, { - response_size: content.length, - success: !(pageStatusCode && pageStatusCode >= 400) && !!content && (content.trim().length >= 100), - error: pageError, - response_code: pageStatusCode, - time_taken: Date.now() - timer, - }); - - return { - content, - metadata: { sourceURL: docxLink, pageStatusCode, pageError }, - provider: "web-scraper", - }; - }) - ); - } - - private applyPathReplacements(documents: Document[]): Document[] { - if (this.replaceAllPathsWithAbsolutePaths) { - documents = replacePathsWithAbsolutePaths(documents); - } - return replaceImgPathsWithAbsolutePaths(documents); - } - - private async applyImgAltText(documents: Document[]): Promise { - return this.generateImgAltText - ? this.generatesImgAltText(documents) - : documents; - } - - private async cacheAndFinalizeDocuments( - documents: Document[], - links: string[] - ): Promise { - // await this.setCachedDocuments(documents, links); - documents = this.removeChildLinks(documents); - return documents.splice(0, this.limit); - } - - private async processDocumentsWithCache( - inProgress?: (progress: Progress) => void - ): Promise { - let documents = await this.getCachedDocuments( - this.urls.slice(0, this.limit) - ); - if (documents.length < this.limit) { - const newDocuments: Document[] = await this.getDocuments( - false, - inProgress - ); - documents = this.mergeNewDocuments(documents, newDocuments); - } - documents = this.filterDocsExcludeInclude(documents); - documents = this.filterDepth(documents); - documents = this.removeChildLinks(documents); - return documents.splice(0, this.limit); - } - - private mergeNewDocuments( - existingDocuments: Document[], - newDocuments: Document[] - ): Document[] { - newDocuments.forEach((doc) => { - if ( - !existingDocuments.some( - (d) => - this.normalizeUrl(d.metadata.sourceURL) === - this.normalizeUrl(doc.metadata?.sourceURL) - ) - ) { - existingDocuments.push(doc); - } - }); - return existingDocuments; - } - - private filterDocsExcludeInclude(documents: Document[]): Document[] { - return documents.filter((document) => { - const url = new URL(document.metadata.sourceURL); - const path = url.pathname; - - if (!Array.isArray(this.excludes)) { - this.excludes = this.excludes.split(','); - } - - if (this.excludes.length > 0 && this.excludes[0] !== "") { - // Check if the link should be excluded - if ( - this.excludes.some((excludePattern) => - new RegExp(excludePattern).test(path) - ) - ) { - return false; - } - } - - if (!Array.isArray(this.includes)) { - this.includes = this.includes.split(','); - } - - if (this.includes.length > 0 && this.includes[0] !== "") { - // Check if the link matches the include patterns, if any are specified - if (this.includes.length > 0) { - return this.includes.some((includePattern) => - new RegExp(includePattern).test(path) - ); - } - } - return true; - }); - } - - private normalizeUrl(url: string): string { - if (url.includes("//www.")) { - return url.replace("//www.", "//"); - } - return url; - } - - private removeChildLinks(documents: Document[]): Document[] { - for (let document of documents) { - if (document?.childrenLinks) delete document.childrenLinks; - } - return documents; - } - - async setCachedDocuments(documents: Document[], childrenLinks?: string[]) { - for (const document of documents) { - if (document.content.trim().length === 0) { - continue; - } - const normalizedUrl = this.normalizeUrl(document.metadata.sourceURL); - await setValue( - "web-scraper-cache:" + normalizedUrl, - JSON.stringify({ - ...document, - childrenLinks: childrenLinks || [], - }), - 60 * 60 - ); // 10 days - } - } - - async getCachedDocuments(urls: string[]): Promise { - let documents: Document[] = []; - for (const url of urls) { - const normalizedUrl = this.normalizeUrl(url); - Logger.debug( - "Getting cached document for web-scraper-cache:" + normalizedUrl - ); - const cachedDocumentString = await getValue( - "web-scraper-cache:" + normalizedUrl - ); - if (cachedDocumentString) { - const cachedDocument = JSON.parse(cachedDocumentString); - documents.push(cachedDocument); - - // get children documents - for (const childUrl of cachedDocument.childrenLinks || []) { - const normalizedChildUrl = this.normalizeUrl(childUrl); - const childCachedDocumentString = await getValue( - "web-scraper-cache:" + normalizedChildUrl - ); - if (childCachedDocumentString) { - const childCachedDocument = JSON.parse(childCachedDocumentString); - if ( - !documents.find( - (doc) => - doc.metadata.sourceURL === - childCachedDocument.metadata.sourceURL - ) - ) { - documents.push(childCachedDocument); - } - } - } - } - } - return documents; - } - - setOptions(options: WebScraperOptions): void { - if (!options.urls) { - throw new Error("Urls are required"); - } - - this.jobId = options.jobId; - this.bullJobId = options.bullJobId; - this.urls = options.urls; - this.mode = options.mode; - this.concurrentRequests = options.concurrentRequests ?? 20; - this.includes = options.crawlerOptions?.includes ?? []; - this.excludes = options.crawlerOptions?.excludes ?? []; - this.maxCrawledLinks = options.crawlerOptions?.maxCrawledLinks ?? 1000; - this.maxCrawledDepth = options.crawlerOptions?.maxDepth ?? 10; - this.returnOnlyUrls = options.crawlerOptions?.returnOnlyUrls ?? false; - this.limit = options.crawlerOptions?.limit ?? 10000; - this.generateImgAltText = - options.crawlerOptions?.generateImgAltText ?? false; - this.pageOptions = { - onlyMainContent: options.pageOptions?.onlyMainContent ?? false, - includeHtml: options.pageOptions?.includeHtml ?? false, - replaceAllPathsWithAbsolutePaths: options.pageOptions?.replaceAllPathsWithAbsolutePaths ?? true, - parsePDF: options.pageOptions?.parsePDF ?? true, - onlyIncludeTags: options.pageOptions?.onlyIncludeTags ?? [], - removeTags: options.pageOptions?.removeTags ?? [], - includeMarkdown: options.pageOptions?.includeMarkdown ?? true, - includeRawHtml: options.pageOptions?.includeRawHtml ?? false, - includeExtract: options.pageOptions?.includeExtract ?? (options.extractorOptions?.mode && options.extractorOptions?.mode !== "markdown") ?? false, - waitFor: options.pageOptions?.waitFor ?? undefined, - headers: options.pageOptions?.headers ?? undefined, - includeLinks: options.pageOptions?.includeLinks ?? true, - fullPageScreenshot: options.pageOptions?.fullPageScreenshot ?? false, - screenshot: options.pageOptions?.screenshot ?? false, - useFastMode: options.pageOptions?.useFastMode ?? false, - disableJsDom: options.pageOptions?.disableJsDom ?? false, - atsv: options.pageOptions?.atsv ?? false, - actions: options.pageOptions?.actions ?? undefined, - geolocation: options.pageOptions?.geolocation ?? undefined, - skipTlsVerification: options.pageOptions?.skipTlsVerification ?? false, - removeBase64Images: options.pageOptions?.removeBase64Images ?? true, - mobile: options.pageOptions?.mobile ?? false, - }; - this.extractorOptions = options.extractorOptions ?? { mode: "markdown" }; - this.replaceAllPathsWithAbsolutePaths = - options.crawlerOptions?.replaceAllPathsWithAbsolutePaths ?? - options.pageOptions?.replaceAllPathsWithAbsolutePaths ?? - false; - - if (typeof options.crawlerOptions?.excludes === 'string') { - this.excludes = options.crawlerOptions?.excludes.split(',').filter((item) => item.trim() !== ""); - } - - if (typeof options.crawlerOptions?.includes === 'string') { - this.includes = options.crawlerOptions?.includes.split(',').filter((item) => item.trim() !== ""); - } - - this.crawlerMode = options.crawlerOptions?.mode ?? "default"; - this.ignoreSitemap = options.crawlerOptions?.ignoreSitemap ?? false; - this.allowBackwardCrawling = - options.crawlerOptions?.allowBackwardCrawling ?? false; - this.allowExternalContentLinks = - options.crawlerOptions?.allowExternalContentLinks ?? false; - this.priority = options.priority; - this.teamId = options.teamId ?? null; - - - - // make sure all urls start with https:// - this.urls = this.urls.map((url) => { - if (!url.trim().startsWith("http")) { - return `https://${url}`; - } - return url; - }); - } - - private async getSitemapData(baseUrl: string, documents: Document[]) { - const sitemapData = await fetchSitemapData(baseUrl); - if (sitemapData) { - for (let i = 0; i < documents.length; i++) { - const docInSitemapData = sitemapData.find( - (data) => - this.normalizeUrl(data.loc) === - this.normalizeUrl(documents[i].metadata.sourceURL) - ); - if (docInSitemapData) { - let sitemapDocData: Partial = {}; - if (docInSitemapData.changefreq) { - sitemapDocData.changefreq = docInSitemapData.changefreq; - } - if (docInSitemapData.priority) { - sitemapDocData.priority = Number(docInSitemapData.priority); - } - if (docInSitemapData.lastmod) { - sitemapDocData.lastmod = docInSitemapData.lastmod; - } - if (Object.keys(sitemapDocData).length !== 0) { - documents[i].metadata.sitemap = sitemapDocData; - } - } - } - } - return documents; - } - private async getSitemapDataForSingleUrl( - baseUrl: string, - url: string, - timeout?: number - ) { - const sitemapData = await fetchSitemapData(baseUrl, timeout); - if (sitemapData) { - const docInSitemapData = sitemapData.find( - (data) => this.normalizeUrl(data.loc) === this.normalizeUrl(url) - ); - if (docInSitemapData) { - let sitemapDocData: Partial = {}; - if (docInSitemapData.changefreq) { - sitemapDocData.changefreq = docInSitemapData.changefreq; - } - if (docInSitemapData.priority) { - sitemapDocData.priority = Number(docInSitemapData.priority); - } - if (docInSitemapData.lastmod) { - sitemapDocData.lastmod = docInSitemapData.lastmod; - } - if (Object.keys(sitemapDocData).length !== 0) { - return sitemapDocData; - } - } - } - return null; - } - generatesImgAltText = async (documents: Document[]): Promise => { - await Promise.all( - documents.map(async (document) => { - const images = document.content.match(/!\[.*?\]\((.*?)\)/g) || []; - - await Promise.all( - images.map(async (image: string) => { - let imageUrl = image.match(/\(([^)]+)\)/)[1]; - let altText = image.match(/\[(.*?)\]/)[1]; - - if ( - !altText && - !imageUrl.startsWith("data:image") && - /\.(png|jpeg|gif|webp)$/.test(imageUrl) - ) { - const imageIndex = document.content.indexOf(image); - const contentLength = document.content.length; - let backText = document.content.substring( - imageIndex + image.length, - Math.min(imageIndex + image.length + 1000, contentLength) - ); - let frontTextStartIndex = Math.max(imageIndex - 1000, 0); - let frontText = document.content.substring( - frontTextStartIndex, - imageIndex - ); - altText = await getImageDescription( - imageUrl, - backText, - frontText, - this.generateImgAltTextModel - ); - } - - document.content = document.content.replace( - image, - `![${altText}](${imageUrl})` - ); - }) - ); - }) - ); - - return documents; - }; - - filterDepth(documents: Document[]): Document[] { - return documents.filter((document) => { - const url = new URL(document.metadata.sourceURL); - return getURLDepth(url.toString()) <= this.maxCrawledDepth; - }); - } -} diff --git a/apps/api/src/scraper/WebScraper/scrapers/fetch.ts b/apps/api/src/scraper/WebScraper/scrapers/fetch.ts deleted file mode 100644 index 0df3be72..00000000 --- a/apps/api/src/scraper/WebScraper/scrapers/fetch.ts +++ /dev/null @@ -1,89 +0,0 @@ -import axios from "axios"; -import { logScrape } from "../../../services/logging/scrape_log"; -import { fetchAndProcessPdf } from "../utils/pdfProcessor"; -import { universalTimeout } from "../global"; -import { Logger } from "../../../lib/logger"; - -/** - * Scrapes a URL with Axios - * @param url The URL to scrape - * @param pageOptions The options for the page - * @returns The scraped content - */ -export async function scrapWithFetch( - url: string, - pageOptions: { parsePDF?: boolean } = { parsePDF: true } -): Promise<{ content: string; pageStatusCode?: number; pageError?: string }> { - const logParams = { - url, - scraper: "fetch", - success: false, - response_code: null, - time_taken_seconds: null, - error_message: null, - html: "", - startTime: Date.now(), - }; - - try { - const response = await axios.get(url, { - headers: { - "Content-Type": "application/json", - }, - timeout: universalTimeout, - transformResponse: [(data) => data], // Prevent axios from parsing JSON automatically - }); - - if (response.status !== 200) { - Logger.debug( - `⛏️ Axios: Failed to fetch url: ${url} with status: ${response.status}` - ); - logParams.error_message = response.statusText; - logParams.response_code = response.status; - return { - content: "", - pageStatusCode: response.status, - pageError: response.statusText, - }; - } - - const contentType = response.headers["content-type"]; - if (contentType && contentType.includes("application/pdf")) { - logParams.success = true; - const { content, pageStatusCode, pageError } = await fetchAndProcessPdf( - url, - pageOptions?.parsePDF - ); - logParams.response_code = pageStatusCode; - logParams.error_message = pageError; - return { content, pageStatusCode: response.status, pageError }; - } else { - const text = response.data; - logParams.success = true; - logParams.html = text; - logParams.response_code = response.status; - return { - content: text, - pageStatusCode: response.status, - pageError: null, - }; - } - } catch (error) { - if (error.code === "ECONNABORTED") { - logParams.error_message = "Request timed out"; - Logger.debug(`⛏️ Axios: Request timed out for ${url}`); - } else { - logParams.error_message = error.message || error; - Logger.debug(`⛏️ Axios: Failed to fetch url: ${url} | Error: ${error}`); - } - return { - content: "", - pageStatusCode: error.response?.status ?? null, - pageError: logParams.error_message, - }; - } finally { - const endTime = Date.now(); - logParams.time_taken_seconds = (endTime - logParams.startTime) / 1000; - await logScrape(logParams); - } -} diff --git a/apps/api/src/scraper/WebScraper/scrapers/fireEngine.ts b/apps/api/src/scraper/WebScraper/scrapers/fireEngine.ts deleted file mode 100644 index 7616774a..00000000 --- a/apps/api/src/scraper/WebScraper/scrapers/fireEngine.ts +++ /dev/null @@ -1,230 +0,0 @@ -import axios from "axios"; -import { Action, FireEngineOptions, FireEngineResponse } from "../../../lib/entities"; -import { logScrape } from "../../../services/logging/scrape_log"; -import { generateRequestParams } from "../single_url"; -import { fetchAndProcessPdf } from "../utils/pdfProcessor"; -import { universalTimeout } from "../global"; -import { Logger } from "../../../lib/logger"; -import * as Sentry from "@sentry/node"; -import axiosRetry from 'axios-retry'; - -axiosRetry(axios, { retries: 3 , onRetry:()=>{ - console.log("Retrying (fire-engine)..."); -}, retryDelay: axiosRetry.exponentialDelay}); -/** - * Scrapes a URL with Fire-Engine - * @param url The URL to scrape - * @param waitFor The time to wait for the page to load - * @param screenshot Whether to take a screenshot - * @param fullPageScreenshot Whether to take a full page screenshot - * @param pageOptions The options for the page - * @param headers The headers to send with the request - * @param options The options for the request - * @returns The scraped content - */ -export async function scrapWithFireEngine({ - url, - actions, - waitFor = 0, - screenshot = false, - fullPageScreenshot = false, - pageOptions = { parsePDF: true, atsv: false, useFastMode: false, disableJsDom: false, geolocation: { country: "US" }, skipTlsVerification: false, removeBase64Images: true, mobile: false }, - fireEngineOptions = {}, - headers, - options, - priority, - teamId, -}: { - url: string; - actions?: Action[]; - waitFor?: number; - screenshot?: boolean; - fullPageScreenshot?: boolean; - pageOptions?: { scrollXPaths?: string[]; parsePDF?: boolean, atsv?: boolean, useFastMode?: boolean, disableJsDom?: boolean, geolocation?: { country?: string }, skipTlsVerification?: boolean, removeBase64Images?: boolean, mobile?: boolean }; - fireEngineOptions?: FireEngineOptions; - headers?: Record; - options?: any; - priority?: number; - teamId?: string; -}): Promise { - const logParams = { - url, - scraper: "fire-engine", - success: false, - response_code: null, - time_taken_seconds: null, - error_message: null, - html: "", - startTime: Date.now(), - }; - - try { - const reqParams = await generateRequestParams(url); - let waitParam = reqParams["params"]?.wait ?? waitFor; - let engineParam = reqParams["params"]?.engine ?? reqParams["params"]?.fireEngineOptions?.engine ?? fireEngineOptions?.engine ?? "chrome-cdp"; - let screenshotParam = reqParams["params"]?.screenshot ?? screenshot; - let fullPageScreenshotParam = reqParams["params"]?.fullPageScreenshot ?? fullPageScreenshot; - let fireEngineOptionsParam : FireEngineOptions = reqParams["params"]?.fireEngineOptions ?? fireEngineOptions; - - - let endpoint = "/scrape"; - - if(options?.endpoint === "request") { - endpoint = "/request"; - } - - let engine = engineParam; // do we want fireEngineOptions as first choice? - - if (pageOptions?.useFastMode) { - fireEngineOptionsParam.engine = "tlsclient"; - engine = "tlsclient"; - } - - Logger.info( - `⛏️ Fire-Engine (${engine}): Scraping ${url} | params: { actions: ${JSON.stringify((actions ?? []).map(x => x.type))}, method: ${fireEngineOptionsParam?.method ?? "null"} }` - ); - - // atsv is only available for beta customers - const betaCustomersString = process.env.BETA_CUSTOMERS; - const betaCustomers = betaCustomersString ? betaCustomersString.split(",") : []; - - if (pageOptions?.atsv && betaCustomers.includes(teamId)) { - fireEngineOptionsParam.atsv = true; - } else { - pageOptions.atsv = false; - } - - const axiosInstance = axios.create({ - headers: { "Content-Type": "application/json" } - }); - - const startTime = Date.now(); - const _response = await Sentry.startSpan({ - name: "Call to fire-engine" - }, async span => { - - return await axiosInstance.post( - process.env.FIRE_ENGINE_BETA_URL + endpoint, - { - url: url, - headers: headers, - wait: waitParam, - screenshot: screenshotParam, - fullPageScreenshot: fullPageScreenshotParam, - disableJsDom: pageOptions?.disableJsDom ?? false, - priority, - engine, - instantReturn: true, - mobile: pageOptions?.mobile ?? false, - ...fireEngineOptionsParam, - atsv: pageOptions?.atsv ?? false, - scrollXPaths: pageOptions?.scrollXPaths ?? [], - geolocation: pageOptions?.geolocation, - skipTlsVerification: pageOptions?.skipTlsVerification ?? false, - removeBase64Images: pageOptions?.removeBase64Images ?? true, - actions: actions, - }, - { - headers: { - "Content-Type": "application/json", - ...(Sentry.isInitialized() ? ({ - "sentry-trace": Sentry.spanToTraceHeader(span), - "baggage": Sentry.spanToBaggageHeader(span), - }) : {}), - } - } - ); - }); - - const waitTotal = (actions ?? []).filter(x => x.type === "wait").reduce((a, x) => (x as { type: "wait"; milliseconds: number; }).milliseconds + a, 0); - - let checkStatusResponse = await axiosInstance.get(`${process.env.FIRE_ENGINE_BETA_URL}/scrape/${_response.data.jobId}`); - - // added 5 seconds to the timeout to account for 'smart wait' - while (checkStatusResponse.data.processing && Date.now() - startTime < universalTimeout + waitTotal + 5000) { - await new Promise(resolve => setTimeout(resolve, 250)); // wait 0.25 seconds - checkStatusResponse = await axiosInstance.get(`${process.env.FIRE_ENGINE_BETA_URL}/scrape/${_response.data.jobId}`); - } - - if (checkStatusResponse.data.processing) { - Logger.debug(`⛏️ Fire-Engine (${engine}): deleting request - jobId: ${_response.data.jobId}`); - axiosInstance.delete( - process.env.FIRE_ENGINE_BETA_URL + `/scrape/${_response.data.jobId}`, { - validateStatus: (status) => true - } - ).catch((error) => { - Logger.debug(`⛏️ Fire-Engine (${engine}): Failed to delete request - jobId: ${_response.data.jobId} | error: ${error}`); - }); - - Logger.debug(`⛏️ Fire-Engine (${engine}): Request timed out for ${url}`); - logParams.error_message = "Request timed out"; - return { html: "", pageStatusCode: null, pageError: "" }; - } - - if (checkStatusResponse.status !== 200 || checkStatusResponse.data.error) { - Logger.debug( - `⛏️ Fire-Engine (${engine}): Failed to fetch url: ${url} \t status: ${checkStatusResponse.status}\t ${checkStatusResponse.data.error}` - ); - - logParams.error_message = checkStatusResponse.data?.pageError ?? checkStatusResponse.data?.error; - logParams.response_code = checkStatusResponse.data?.pageStatusCode; - - if(checkStatusResponse.data && checkStatusResponse.data?.pageStatusCode !== 200) { - Logger.debug(`⛏️ Fire-Engine (${engine}): Failed to fetch url: ${url} \t status: ${checkStatusResponse.data?.pageStatusCode}`); - } - - const pageStatusCode = checkStatusResponse.data?.pageStatusCode ? checkStatusResponse.data?.pageStatusCode : checkStatusResponse.data?.error && checkStatusResponse.data?.error.includes("Dns resolution error for hostname") ? 404 : undefined; - - return { - html: "", - pageStatusCode, - pageError: checkStatusResponse.data?.pageError ?? checkStatusResponse.data?.error, - }; - } - - const contentType = checkStatusResponse.data.responseHeaders?.["content-type"]; - - if (contentType && contentType.includes("application/pdf")) { - const { content, pageStatusCode, pageError } = await fetchAndProcessPdf( - url, - pageOptions?.parsePDF - ); - logParams.success = true; - logParams.response_code = pageStatusCode; - logParams.error_message = pageError; - return { html: content, pageStatusCode, pageError }; - } else { - const data = checkStatusResponse.data; - - logParams.success = - (data.pageStatusCode >= 200 && data.pageStatusCode < 300) || - data.pageStatusCode === 404; - logParams.html = data.content ?? ""; - logParams.response_code = data.pageStatusCode; - logParams.error_message = data.pageError ?? data.error; - - return { - html: data.content ?? "", - screenshots: data.screenshots ?? [data.screenshot] ?? [], - pageStatusCode: data.pageStatusCode, - pageError: data.pageError ?? data.error, - scrapeActionContent: data?.actionContent ?? [], - }; - } - } catch (error) { - if (error.code === "ECONNABORTED") { - Logger.debug(`⛏️ Fire-Engine (catch block): Request timed out for ${url}`); - logParams.error_message = "Request timed out"; - } else { - Logger.debug(`⛏️ Fire-Engine(catch block): Failed to fetch url: ${url} | Error: ${error}`); - logParams.error_message = error.message || error; - } - return { html: "", pageStatusCode: null, pageError: logParams.error_message }; - } finally { - const endTime = Date.now(); - logParams.time_taken_seconds = (endTime - logParams.startTime) / 1000; - await logScrape(logParams, pageOptions); - } -} - - diff --git a/apps/api/src/scraper/WebScraper/scrapers/playwright.ts b/apps/api/src/scraper/WebScraper/scrapers/playwright.ts deleted file mode 100644 index 09c7353b..00000000 --- a/apps/api/src/scraper/WebScraper/scrapers/playwright.ts +++ /dev/null @@ -1,111 +0,0 @@ -import axios from "axios"; -import { logScrape } from "../../../services/logging/scrape_log"; -import { generateRequestParams } from "../single_url"; -import { fetchAndProcessPdf } from "../utils/pdfProcessor"; -import { universalTimeout } from "../global"; -import { Logger } from "../../../lib/logger"; - -/** - * Scrapes a URL with Playwright - * @param url The URL to scrape - * @param waitFor The time to wait for the page to load - * @param headers The headers to send with the request - * @param pageOptions The options for the page - * @returns The scraped content - */ -export async function scrapWithPlaywright( - url: string, - waitFor: number = 0, - headers?: Record, - pageOptions: { parsePDF?: boolean } = { parsePDF: true } -): Promise<{ content: string; pageStatusCode?: number; pageError?: string }> { - const logParams = { - url, - scraper: "playwright", - success: false, - response_code: null, - time_taken_seconds: null, - error_message: null, - html: "", - startTime: Date.now(), - }; - - try { - const reqParams = await generateRequestParams(url); - // If the user has passed a wait parameter in the request, use that - const waitParam = reqParams["params"]?.wait ?? waitFor; - - const response = await axios.post( - process.env.PLAYWRIGHT_MICROSERVICE_URL, - { - url: url, - wait_after_load: waitParam, - timeout: universalTimeout + waitParam, - headers: headers, - }, - { - headers: { - "Content-Type": "application/json", - }, - timeout: universalTimeout + waitParam, // Add waitParam to timeout to account for the wait time - transformResponse: [(data) => data], // Prevent axios from parsing JSON automatically - } - ); - - if (response.status !== 200) { - Logger.debug( - `⛏️ Playwright: Failed to fetch url: ${url} | status: ${response.status}, error: ${response.data?.pageError}` - ); - logParams.error_message = response.data?.pageError; - logParams.response_code = response.data?.pageStatusCode; - return { - content: "", - pageStatusCode: response.data?.pageStatusCode, - pageError: response.data?.pageError, - }; - } - - const contentType = response.headers["content-type"]; - if (contentType && contentType.includes("application/pdf")) { - logParams.success = true; - const { content, pageStatusCode, pageError } = await fetchAndProcessPdf(url, pageOptions?.parsePDF); - logParams.response_code = pageStatusCode; - logParams.error_message = pageError; - return { content, pageStatusCode, pageError }; - } else { - const textData = response.data; - try { - const data = JSON.parse(textData); - const html = data.content; - logParams.success = true; - logParams.html = html; - logParams.response_code = data.pageStatusCode; - logParams.error_message = data.pageError; - return { - content: html ?? "", - pageStatusCode: data.pageStatusCode, - pageError: data.pageError, - }; - } catch (jsonError) { - logParams.error_message = jsonError.message || jsonError; - Logger.debug( - `⛏️ Playwright: Error parsing JSON response for url: ${url} | Error: ${jsonError}` - ); - return { content: "", pageStatusCode: null, pageError: logParams.error_message }; - } - } - } catch (error) { - if (error.code === "ECONNABORTED") { - logParams.error_message = "Request timed out"; - Logger.debug(`⛏️ Playwright: Request timed out for ${url}`); - } else { - logParams.error_message = error.message || error; - Logger.debug(`⛏️ Playwright: Failed to fetch url: ${url} | Error: ${error}`); - } - return { content: "", pageStatusCode: null, pageError: logParams.error_message }; - } finally { - const endTime = Date.now(); - logParams.time_taken_seconds = (endTime - logParams.startTime) / 1000; - await logScrape(logParams); - } -} diff --git a/apps/api/src/scraper/WebScraper/scrapers/scrapingBee.ts b/apps/api/src/scraper/WebScraper/scrapers/scrapingBee.ts deleted file mode 100644 index b72fa8b2..00000000 --- a/apps/api/src/scraper/WebScraper/scrapers/scrapingBee.ts +++ /dev/null @@ -1,92 +0,0 @@ -import { logScrape } from "../../../services/logging/scrape_log"; -import { generateRequestParams } from "../single_url"; -import { fetchAndProcessPdf } from "../utils/pdfProcessor"; -import { universalTimeout } from "../global"; -import { ScrapingBeeClient } from "scrapingbee"; -import { Logger } from "../../../lib/logger"; - -/** - * Scrapes a URL with ScrapingBee - * @param url The URL to scrape - * @param wait_browser The browser event to wait for - * @param timeout The timeout for the scrape - * @param pageOptions The options for the page - * @returns The scraped content - */ -export async function scrapWithScrapingBee( - url: string, - wait_browser: string = "domcontentloaded", - timeout: number = universalTimeout, - pageOptions: { parsePDF?: boolean } = { parsePDF: true } - ): Promise<{ content: string; pageStatusCode?: number; pageError?: string }> { - const logParams = { - url, - scraper: wait_browser === "networkidle2" ? "scrapingBeeLoad" : "scrapingBee", - success: false, - response_code: null, - time_taken_seconds: null, - error_message: null, - html: "", - startTime: Date.now(), - }; - try { - const client = new ScrapingBeeClient(process.env.SCRAPING_BEE_API_KEY); - const clientParams = await generateRequestParams( - url, - wait_browser, - timeout - ); - const response = await client.get({ - ...clientParams, - params: { - ...clientParams.params, - transparent_status_code: "True", - }, - }); - Logger.info( - `⛏️ ScrapingBee: Scraping ${url}` - ); - const contentType = response.headers["content-type"]; - if (contentType && contentType.includes("application/pdf")) { - logParams.success = true; - const { content, pageStatusCode, pageError } = await fetchAndProcessPdf(url, pageOptions?.parsePDF); - logParams.response_code = pageStatusCode; - logParams.error_message = pageError; - return { content, pageStatusCode, pageError }; - } else { - let text = ""; - try { - const decoder = new TextDecoder(); - text = decoder.decode(response.data); - logParams.success = true; - } catch (decodeError) { - Logger.debug( - `⛏️ ScrapingBee: Error decoding response data for url: ${url} | Error: ${decodeError}` - ); - logParams.error_message = decodeError.message || decodeError; - } - logParams.response_code = response.status; - logParams.html = text; - logParams.success = response.status >= 200 && response.status < 300 || response.status === 404; - logParams.error_message = response.statusText !== "OK" ? response.statusText : undefined; - return { - content: text, - pageStatusCode: response.status, - pageError: response.statusText !== "OK" ? response.statusText : undefined, - }; - } - } catch (error) { - Logger.debug(`⛏️ ScrapingBee: Error fetching url: ${url} | Error: ${error}`); - logParams.error_message = error.message || error; - logParams.response_code = error.response?.status; - return { - content: "", - pageStatusCode: error.response?.status, - pageError: error.response?.statusText, - }; - } finally { - const endTime = Date.now(); - logParams.time_taken_seconds = (endTime - logParams.startTime) / 1000; - await logScrape(logParams); - } - } \ No newline at end of file diff --git a/apps/api/src/scraper/WebScraper/single_url.ts b/apps/api/src/scraper/WebScraper/single_url.ts deleted file mode 100644 index 38ec74f0..00000000 --- a/apps/api/src/scraper/WebScraper/single_url.ts +++ /dev/null @@ -1,506 +0,0 @@ -import * as cheerio from "cheerio"; -import { extractMetadata } from "./utils/metadata"; -import dotenv from "dotenv"; -import { - Document, - PageOptions, - FireEngineResponse, - ExtractorOptions, - Action, -} from "../../lib/entities"; -import { parseMarkdown } from "../../lib/html-to-markdown"; -import { urlSpecificParams } from "./utils/custom/website_params"; -import { fetchAndProcessPdf } from "./utils/pdfProcessor"; -import { handleCustomScraping } from "./custom/handleCustomScraping"; -import { removeUnwantedElements } from "./utils/removeUnwantedElements"; -import { scrapWithFetch } from "./scrapers/fetch"; -import { scrapWithFireEngine } from "./scrapers/fireEngine"; -import { scrapWithPlaywright } from "./scrapers/playwright"; -import { scrapWithScrapingBee } from "./scrapers/scrapingBee"; -import { extractLinks } from "./utils/utils"; -import { Logger } from "../../lib/logger"; -import { ScrapeEvents } from "../../lib/scrape-events"; -import { clientSideError } from "../../strings"; -import { ScrapeActionContent } from "../../lib/entities"; -import { removeBase64Images } from "./utils/removeBase64Images"; - -dotenv.config(); - -const useScrapingBee = process.env.SCRAPING_BEE_API_KEY !== '' && process.env.SCRAPING_BEE_API_KEY !== undefined; -const useFireEngine = process.env.FIRE_ENGINE_BETA_URL !== '' && process.env.FIRE_ENGINE_BETA_URL !== undefined; - -export const baseScrapers = [ - useFireEngine ? "fire-engine;chrome-cdp" : undefined, - useFireEngine ? "fire-engine" : undefined, - useScrapingBee ? "scrapingBee" : undefined, - useFireEngine ? undefined : "playwright", - useScrapingBee ? "scrapingBeeLoad" : undefined, - "fetch", -].filter(Boolean); - -export async function generateRequestParams( - url: string, - wait_browser: string = "domcontentloaded", - timeout: number = 15000 -): Promise { - const defaultParams = { - url: url, - params: { timeout: timeout, wait_browser: wait_browser }, - headers: { "ScrapingService-Request": "TRUE" }, - }; - - try { - const urlKey = new URL(url).hostname.replace(/^www\./, ""); - if (urlSpecificParams.hasOwnProperty(urlKey)) { - return { ...defaultParams, ...urlSpecificParams[urlKey] }; - } else { - return defaultParams; - } - } catch (error) { - Logger.error(`Error generating URL key: ${error}`); - return defaultParams; - } -} - -/** - * Get the order of scrapers to be used for scraping a URL - * If the user doesn't have envs set for a specific scraper, it will be removed from the order. - * @param defaultScraper The default scraper to use if the URL does not have a specific scraper order defined - * @returns The order of scrapers to be used for scraping a URL - */ -function getScrapingFallbackOrder( - defaultScraper?: string, - isWaitPresent: boolean = false, - isScreenshotPresent: boolean = false, - isHeadersPresent: boolean = false, - isActionsPresent: boolean = false, -) { - if (isActionsPresent) { - return useFireEngine ? ["fire-engine;chrome-cdp"] : []; - } - - const availableScrapers = baseScrapers.filter((scraper) => { - switch (scraper) { - case "scrapingBee": - case "scrapingBeeLoad": - return !!process.env.SCRAPING_BEE_API_KEY; - case "fire-engine": - return !!process.env.FIRE_ENGINE_BETA_URL; - case "fire-engine;chrome-cdp": - return !!process.env.FIRE_ENGINE_BETA_URL; - case "playwright": - return !!process.env.PLAYWRIGHT_MICROSERVICE_URL; - default: - return true; - } - }); - - let defaultOrder = [ - useFireEngine ? "fire-engine;chrome-cdp" : undefined, - useFireEngine ? "fire-engine" : undefined, - useScrapingBee ? "scrapingBee" : undefined, - useScrapingBee ? "scrapingBeeLoad" : undefined, - useFireEngine ? undefined : "playwright", - "fetch", - ].filter(Boolean); - - // if (isWaitPresent || isScreenshotPresent || isHeadersPresent) { - // defaultOrder = [ - // "fire-engine", - // useFireEngine ? undefined : "playwright", - // ...defaultOrder.filter( - // (scraper) => scraper !== "fire-engine" && scraper !== "playwright" - // ), - // ].filter(Boolean); - // } - - const filteredDefaultOrder = defaultOrder.filter( - (scraper: (typeof baseScrapers)[number]) => - availableScrapers.includes(scraper) - ); - const uniqueScrapers = new Set( - defaultScraper - ? [defaultScraper, ...filteredDefaultOrder, ...availableScrapers] - : [...filteredDefaultOrder, ...availableScrapers] - ); - - const scrapersInOrder = Array.from(uniqueScrapers); - return scrapersInOrder as (typeof baseScrapers)[number][]; -} - - - -export async function scrapSingleUrl( - jobId: string, - urlToScrap: string, - pageOptions: PageOptions, - extractorOptions?: ExtractorOptions, - existingHtml?: string, - priority?: number, - teamId?: string -): Promise { - pageOptions = { - includeMarkdown: pageOptions.includeMarkdown ?? true, - includeExtract: pageOptions.includeExtract ?? false, - onlyMainContent: pageOptions.onlyMainContent ?? false, - includeHtml: pageOptions.includeHtml ?? false, - includeRawHtml: pageOptions.includeRawHtml ?? false, - waitFor: pageOptions.waitFor ?? undefined, - screenshot: pageOptions.screenshot ?? false, - fullPageScreenshot: pageOptions.fullPageScreenshot ?? false, - headers: pageOptions.headers ?? undefined, - includeLinks: pageOptions.includeLinks ?? true, - replaceAllPathsWithAbsolutePaths: pageOptions.replaceAllPathsWithAbsolutePaths ?? true, - parsePDF: pageOptions.parsePDF ?? true, - removeTags: pageOptions.removeTags ?? [], - onlyIncludeTags: pageOptions.onlyIncludeTags ?? [], - useFastMode: pageOptions.useFastMode ?? false, - disableJsDom: pageOptions.disableJsDom ?? false, - atsv: pageOptions.atsv ?? false, - actions: pageOptions.actions ?? undefined, - geolocation: pageOptions.geolocation ?? undefined, - skipTlsVerification: pageOptions.skipTlsVerification ?? false, - removeBase64Images: pageOptions.removeBase64Images ?? true, - mobile: pageOptions.mobile ?? false, - } - - if (extractorOptions) { - extractorOptions = { - mode: extractorOptions?.mode ?? "llm-extraction-from-markdown", - } - } - - if (!existingHtml) { - existingHtml = ""; - } - - urlToScrap = urlToScrap.trim(); - - const attemptScraping = async ( - url: string, - method: (typeof baseScrapers)[number] - ) => { - let scraperResponse: { - text: string; - screenshot: string; - actions?: { - screenshots?: string[]; - scrapes?: ScrapeActionContent[]; - }; - metadata: { pageStatusCode?: number; pageError?: string | null }; - } = { text: "", screenshot: "", metadata: {} }; - let screenshot = ""; - - const timer = Date.now(); - const logInsertPromise = ScrapeEvents.insert(jobId, { - type: "scrape", - url, - worker: process.env.FLY_MACHINE_ID, - method, - result: null, - }); - - switch (method) { - case "fire-engine": - case "fire-engine;chrome-cdp": - - let engine: "playwright" | "chrome-cdp" | "tlsclient" = "playwright"; - if (method === "fire-engine;chrome-cdp") { - engine = "chrome-cdp"; - } - - if (process.env.FIRE_ENGINE_BETA_URL) { - const processedActions: Action[] = pageOptions.actions?.flatMap((action: Action, index: number, array: Action[]) => { - if (action.type === "click" || action.type === "write" || action.type === "press") { - const result: Action[] = []; - // Don't add a wait if the previous action is a wait - // if (index === 0 || array[index - 1].type !== "wait") { - // result.push({ type: "wait", milliseconds: 1200 } as Action); - // } - // Fire-engine now handles wait times automatically, leaving the code here for now - result.push(action); - // Don't add a wait if the next action is a wait - // if (index === array.length - 1 || array[index + 1].type !== "wait") { - // result.push({ type: "wait", milliseconds: 1200 } as Action); - // } - return result; - } - return [action as Action]; - }) ?? [] as Action[]; - - const response = await scrapWithFireEngine({ - url, - ...(engine === "chrome-cdp" ? ({ - actions: [ - ...(pageOptions.waitFor ? [{ - type: "wait" as const, - milliseconds: pageOptions.waitFor, - }] : []), - ...((pageOptions.screenshot || pageOptions.fullPageScreenshot) ? [{ - type: "screenshot" as const, - fullPage: !!pageOptions.fullPageScreenshot, - }] : []), - ...processedActions, - ], - }) : ({ - waitFor: pageOptions.waitFor, - screenshot: pageOptions.screenshot, - fullPageScreenshot: pageOptions.fullPageScreenshot, - })), - pageOptions: pageOptions, - headers: pageOptions.headers, - fireEngineOptions: { - engine: engine, - atsv: pageOptions.atsv, - disableJsDom: pageOptions.disableJsDom, - }, - priority, - teamId, - }); - scraperResponse.text = response.html; - if (pageOptions.screenshot || pageOptions.fullPageScreenshot) { - scraperResponse.screenshot = (response.screenshots ?? []).splice(0, 1)[0] ?? ""; - } - if (pageOptions.actions) { - scraperResponse.actions = { - screenshots: response.screenshots ?? [], - scrapes: response.scrapeActionContent ?? [], - }; - } - scraperResponse.metadata.pageStatusCode = response.pageStatusCode; - scraperResponse.metadata.pageError = response.pageError; - } - break; - case "scrapingBee": - if (process.env.SCRAPING_BEE_API_KEY) { - const response = await scrapWithScrapingBee( - url, - "domcontentloaded", - pageOptions.fallback === false ? 7000 : 15000 - ); - scraperResponse.text = response.content; - scraperResponse.metadata.pageStatusCode = response.pageStatusCode; - scraperResponse.metadata.pageError = response.pageError; - } - break; - case "playwright": - if (process.env.PLAYWRIGHT_MICROSERVICE_URL) { - const response = await scrapWithPlaywright( - url, - pageOptions.waitFor, - pageOptions.headers - ); - scraperResponse.text = response.content; - scraperResponse.metadata.pageStatusCode = response.pageStatusCode; - scraperResponse.metadata.pageError = response.pageError; - } - break; - case "scrapingBeeLoad": - if (process.env.SCRAPING_BEE_API_KEY) { - const response = await scrapWithScrapingBee(url, "networkidle2"); - scraperResponse.text = response.content; - scraperResponse.metadata.pageStatusCode = response.pageStatusCode; - scraperResponse.metadata.pageError = response.pageError; - } - break; - case "fetch": - const response = await scrapWithFetch(url); - scraperResponse.text = response.content; - scraperResponse.metadata.pageStatusCode = response.pageStatusCode; - scraperResponse.metadata.pageError = response.pageError; - break; - } - - let customScrapedContent: FireEngineResponse | null = null; - - // Check for custom scraping conditions - const customScraperResult = await handleCustomScraping( - scraperResponse.text, - url - ); - - if (customScraperResult) { - switch (customScraperResult.scraper) { - case "fire-engine": - customScrapedContent = await scrapWithFireEngine({ - url: customScraperResult.url, - actions: customScraperResult.waitAfterLoad ? ([ - { - type: "wait", - milliseconds: customScraperResult.waitAfterLoad, - } - ]) : ([]), - pageOptions: customScraperResult.pageOptions, - }); - break; - case "pdf": - const { content, pageStatusCode, pageError } = - await fetchAndProcessPdf( - customScraperResult.url, - pageOptions?.parsePDF - ); - customScrapedContent = { - html: content, - pageStatusCode, - pageError, - }; - break; - } - } - - if (customScrapedContent) { - scraperResponse.text = customScrapedContent.html; - } - //* TODO: add an optional to return markdown or structured/extracted content - let cleanedHtml = removeUnwantedElements(scraperResponse.text, pageOptions); - let text = await parseMarkdown(cleanedHtml); - if (pageOptions.removeBase64Images) { - text = await removeBase64Images(text); - } - - const insertedLogId = await logInsertPromise; - ScrapeEvents.updateScrapeResult(insertedLogId, { - response_size: scraperResponse.text.length, - success: !(scraperResponse.metadata.pageStatusCode && scraperResponse.metadata.pageStatusCode >= 400) && !!text && (text.trim().length >= 100), - error: scraperResponse.metadata.pageError, - response_code: scraperResponse.metadata.pageStatusCode, - time_taken: Date.now() - timer, - }); - - return { - text, - html: cleanedHtml, - rawHtml: scraperResponse.text, - screenshot: scraperResponse.screenshot, - actions: scraperResponse.actions, - pageStatusCode: scraperResponse.metadata.pageStatusCode, - pageError: scraperResponse.metadata.pageError || undefined, - }; - }; - - let { text, html, rawHtml, screenshot, actions, pageStatusCode, pageError } = { - text: "", - html: "", - rawHtml: "", - screenshot: "", - actions: undefined, - pageStatusCode: 200, - pageError: undefined, - }; - try { - let urlKey = urlToScrap; - try { - urlKey = new URL(urlToScrap).hostname.replace(/^www\./, ""); - } catch (error) { - Logger.error(`Invalid URL key, trying: ${urlToScrap}`); - } - const defaultScraper = urlSpecificParams[urlKey]?.defaultScraper ?? ""; - const scrapersInOrder = getScrapingFallbackOrder( - defaultScraper, - pageOptions && pageOptions.waitFor && pageOptions.waitFor > 0, - pageOptions && (pageOptions.screenshot || pageOptions.fullPageScreenshot) && (pageOptions.screenshot === true || pageOptions.fullPageScreenshot === true), - pageOptions && pageOptions.headers && pageOptions.headers !== undefined, - pageOptions && Array.isArray(pageOptions.actions) && pageOptions.actions.length > 0, - ); - - for (const scraper of scrapersInOrder) { - // If exists text coming from crawler, use it - if (existingHtml && existingHtml.trim().length >= 100 && !existingHtml.includes(clientSideError)) { - let cleanedHtml = removeUnwantedElements(existingHtml, pageOptions); - text = await parseMarkdown(cleanedHtml); - html = cleanedHtml; - break; - } - - const attempt = await attemptScraping(urlToScrap, scraper); - text = attempt.text ?? ""; - html = attempt.html ?? ""; - rawHtml = attempt.rawHtml ?? ""; - screenshot = attempt.screenshot ?? ""; - actions = attempt.actions ?? undefined; - - if (attempt.pageStatusCode) { - pageStatusCode = attempt.pageStatusCode; - } - - if (attempt.pageError && (attempt.pageStatusCode >= 400 || scrapersInOrder.indexOf(scraper) === scrapersInOrder.length - 1)) { // force pageError if it's the last scraper and it failed too - pageError = attempt.pageError; - - if (attempt.pageStatusCode < 400 || !attempt.pageStatusCode) { - pageStatusCode = 500; - } - } else if (attempt && attempt.pageStatusCode && attempt.pageStatusCode < 400) { - pageError = undefined; - } - - if ((text && text.trim().length >= 100) || (typeof screenshot === "string" && screenshot.length > 0)) { - Logger.debug(`⛏️ ${scraper}: Successfully scraped ${urlToScrap} with text length >= 100 or screenshot, breaking`); - break; - } - if (pageStatusCode && (pageStatusCode == 404 || pageStatusCode == 400)) { - Logger.debug(`⛏️ ${scraper}: Successfully scraped ${urlToScrap} with status code ${pageStatusCode}, breaking`); - break; - } - // const nextScraperIndex = scrapersInOrder.indexOf(scraper) + 1; - // if (nextScraperIndex < scrapersInOrder.length) { - // Logger.debug(`⛏️ ${scraper} Failed to fetch URL: ${urlToScrap} with status: ${pageStatusCode}, error: ${pageError} | Falling back to ${scrapersInOrder[nextScraperIndex]}`); - // } - } - - if (!text) { - throw new Error(`All scraping methods failed for URL: ${urlToScrap}`); - } - - const soup = cheerio.load(rawHtml); - const metadata = extractMetadata(soup, urlToScrap); - - let linksOnPage: string[] | undefined; - - if (pageOptions.includeLinks) { - linksOnPage = extractLinks(rawHtml, urlToScrap); - } - - let document: Document = { - content: text, - markdown: pageOptions.includeMarkdown || pageOptions.includeExtract ? text : undefined, - html: pageOptions.includeHtml ? html : undefined, - rawHtml: - pageOptions.includeRawHtml || - (extractorOptions?.mode === "llm-extraction-from-raw-html" && pageOptions.includeExtract) - ? rawHtml - : undefined, - linksOnPage: pageOptions.includeLinks ? linksOnPage : undefined, - actions, - metadata: { - ...metadata, - ...(screenshot && screenshot.length > 0 ? ({ - screenshot, - }) : {}), - sourceURL: urlToScrap, - pageStatusCode: pageStatusCode, - pageError: pageError, - }, - }; - - return document; - } catch (error) { - Logger.debug(`⛏️ Error: ${error.message} - Failed to fetch URL: ${urlToScrap}`); - ScrapeEvents.insert(jobId, { - type: "error", - message: typeof error === "string" ? error : typeof error.message === "string" ? error.message : JSON.stringify(error), - stack: error.stack, - }); - - return { - content: "", - markdown: pageOptions.includeMarkdown || pageOptions.includeExtract ? "" : undefined, - html: "", - linksOnPage: pageOptions.includeLinks ? [] : undefined, - metadata: { - sourceURL: urlToScrap, - pageStatusCode: pageStatusCode, - pageError: pageError, - }, - } as Document; - } -} diff --git a/apps/api/src/scraper/WebScraper/sitemap.ts b/apps/api/src/scraper/WebScraper/sitemap.ts index 756cd765..05b3d00d 100644 --- a/apps/api/src/scraper/WebScraper/sitemap.ts +++ b/apps/api/src/scraper/WebScraper/sitemap.ts @@ -1,9 +1,10 @@ import axios from "axios"; import { axiosTimeout } from "../../lib/timeout"; import { parseStringPromise } from "xml2js"; -import { scrapWithFireEngine } from "./scrapers/fireEngine"; import { WebCrawler } from "./crawler"; -import { Logger } from "../../lib/logger"; +import { logger } from "../../lib/logger"; +import { scrapeURL } from "../scrapeURL"; +import { scrapeOptions } from "../../controllers/v1/types"; export async function getLinksFromSitemap( { @@ -17,17 +18,20 @@ export async function getLinksFromSitemap( } ): Promise { try { - let content: string; + let content: string = ""; try { if (mode === 'axios' || process.env.FIRE_ENGINE_BETA_URL === '') { const response = await axios.get(sitemapUrl, { timeout: axiosTimeout }); content = response.data; } else if (mode === 'fire-engine') { - const response = await scrapWithFireEngine({ url: sitemapUrl, fireEngineOptions: { engine:"playwright" } }); - content = response.html; + const response = await scrapeURL("sitemap", sitemapUrl, scrapeOptions.parse({ formats: ["rawHtml"] }), { forceEngine: "fire-engine;playwright" });; + if (!response.success) { + throw response.error; + } + content = response.document.rawHtml!; } } catch (error) { - Logger.error(`Request failed for ${sitemapUrl}: ${error.message}`); + logger.error(`Request failed for ${sitemapUrl}: ${error.message}`); return allUrls; } @@ -47,7 +51,7 @@ export async function getLinksFromSitemap( allUrls.push(...validUrls); } } catch (error) { - Logger.debug(`Error processing sitemapUrl: ${sitemapUrl} | Error: ${error.message}`); + logger.debug(`Error processing sitemapUrl: ${sitemapUrl} | Error: ${error.message}`); } return allUrls; diff --git a/apps/api/src/scraper/WebScraper/utils/__tests__/docxProcessor.test.ts b/apps/api/src/scraper/WebScraper/utils/__tests__/docxProcessor.test.ts deleted file mode 100644 index 53237ef8..00000000 --- a/apps/api/src/scraper/WebScraper/utils/__tests__/docxProcessor.test.ts +++ /dev/null @@ -1,15 +0,0 @@ -import * as docxProcessor from "../docxProcessor"; - -describe("DOCX Processing Module - Integration Test", () => { - it("should correctly process a simple DOCX file without the LLAMAPARSE_API_KEY", async () => { - delete process.env.LLAMAPARSE_API_KEY; - const { content, pageStatusCode, pageError } = await docxProcessor.fetchAndProcessDocx( - "https://nvca.org/wp-content/uploads/2019/06/NVCA-Model-Document-Stock-Purchase-Agreement.docx" - ); - expect(content.trim()).toContain( - "SERIES A PREFERRED STOCK PURCHASE AGREEMENT" - ); - expect(pageStatusCode).toBe(200); - expect(pageError).toBeUndefined(); - }); -}); diff --git a/apps/api/src/scraper/WebScraper/utils/__tests__/parseTable.test.ts b/apps/api/src/scraper/WebScraper/utils/__tests__/parseTable.test.ts deleted file mode 100644 index 8d644c7b..00000000 --- a/apps/api/src/scraper/WebScraper/utils/__tests__/parseTable.test.ts +++ /dev/null @@ -1,128 +0,0 @@ -import { parseTablesToMarkdown, convertTableElementToMarkdown, convertTableRowElementToMarkdown, createMarkdownDividerRow } from '../parseTable'; -import cheerio from 'cheerio'; - -describe('parseTablesToMarkdown', () => { - it('converts a simple HTML table to Markdown', async () => { - const html = ` - - - - -
Header 1Header 2
Row 1 Col 1Row 1 Col 2
Row 2 Col 1Row 2 Col 2
- `; - const expectedMarkdown = `
| Header 1 | Header 2 |\n| --- | --- |\n| Row 1 Col 1 | Row 1 Col 2 |\n| Row 2 Col 1 | Row 2 Col 2 |
`; - const markdown = await parseTablesToMarkdown(html); - expect(markdown).toBe(expectedMarkdown); - }); - - it('converts a table with a single row to Markdown', async () => { - const html = ` - - - -
Header 1Header 2
Row 1 Col 1Row 1 Col 2
- `; - const expectedMarkdown = `
| Header 1 | Header 2 |\n| --- | --- |\n| Row 1 Col 1 | Row 1 Col 2 |
`; - const markdown = await parseTablesToMarkdown(html); - expect(markdown).toBe(expectedMarkdown); - }); - - it('converts a table with a single column to Markdown', async () => { - const html = ` - - - - -
Header 1
Row 1 Col 1
Row 2 Col 1
- `; - const expectedMarkdown = `
| Header 1 |\n| --- |\n| Row 1 Col 1 |\n| Row 2 Col 1 |
`; - const markdown = await parseTablesToMarkdown(html); - expect(markdown).toBe(expectedMarkdown); - }); - - it('converts a table with a single cell to Markdown', async () => { - const html = ` - - - -
Header 1
Row 1 Col 1
- `; - const expectedMarkdown = `
| Header 1 |\n| --- |\n| Row 1 Col 1 |
`; - const markdown = await parseTablesToMarkdown(html); - expect(markdown).toBe(expectedMarkdown); - }); - - it('converts a table with no header to Markdown', async () => { - const html = ` - - - -
Row 1 Col 1Row 1 Col 2
Row 2 Col 1Row 2 Col 2
- `; - const expectedMarkdown = `
| Row 1 Col 1 | Row 1 Col 2 |\n| Row 2 Col 1 | Row 2 Col 2 |
`; - const markdown = await parseTablesToMarkdown(html); - expect(markdown).toBe(expectedMarkdown); - }); - - it('converts a table with no rows to Markdown', async () => { - const html = ` - -
- `; - const expectedMarkdown = `
`; - const markdown = await parseTablesToMarkdown(html); - expect(markdown).toBe(expectedMarkdown); - }); - - it('converts a table with no cells to Markdown', async () => { - const html = ` - - -
- `; - const expectedMarkdown = `
`; - const markdown = await parseTablesToMarkdown(html); - expect(markdown).toBe(expectedMarkdown); - }); - - it('converts a table with no columns to Markdown', async () => { - const html = ` - - -
- `; - const expectedMarkdown = `
`; - const markdown = await parseTablesToMarkdown(html); - expect(markdown).toBe(expectedMarkdown); - }); - - it('converts a table with no table to Markdown', async () => { - const html = ``; - const expectedMarkdown = ``; - const markdown = await parseTablesToMarkdown(html); - expect(markdown).toBe(expectedMarkdown); - }); - -it('converts a table inside of a bunch of html noise', async () => { - const html = ` -
-

Some text before

- - - -
Row 1 Col 1Row 1 Col 2
Row 2 Col 1Row 2 Col 2
-

Some text after

-
- `; - const expectedMarkdown = `
-

Some text before

-
| Row 1 Col 1 | Row 1 Col 2 | -| Row 2 Col 1 | Row 2 Col 2 |
-

Some text after

-
`; - - const markdown = await parseTablesToMarkdown(html); - expect(markdown).toBe(expectedMarkdown); -}); - -}); diff --git a/apps/api/src/scraper/WebScraper/utils/__tests__/pdfProcessor.test.ts b/apps/api/src/scraper/WebScraper/utils/__tests__/pdfProcessor.test.ts deleted file mode 100644 index 18302654..00000000 --- a/apps/api/src/scraper/WebScraper/utils/__tests__/pdfProcessor.test.ts +++ /dev/null @@ -1,19 +0,0 @@ -import * as pdfProcessor from '../pdfProcessor'; - -describe('PDF Processing Module - Integration Test', () => { - it('should correctly process a simple PDF file without the LLAMAPARSE_API_KEY', async () => { - delete process.env.LLAMAPARSE_API_KEY; - const { content, pageStatusCode, pageError } = await pdfProcessor.fetchAndProcessPdf('https://s3.us-east-1.amazonaws.com/storage.mendable.ai/rafa-testing/test%20%281%29.pdf', true); - expect(content.trim()).toEqual("Dummy PDF file"); - expect(pageStatusCode).toEqual(200); - expect(pageError).toBeUndefined(); - }); - - it('should return a successful response for a valid scrape with PDF file and parsePDF set to false', async () => { - const { content, pageStatusCode, pageError } = await pdfProcessor.fetchAndProcessPdf('https://arxiv.org/pdf/astro-ph/9301001.pdf', false); - expect(pageStatusCode).toBe(200); - expect(pageError).toBeUndefined(); - expect(content).toContain('/Title(arXiv:astro-ph/9301001v1 7 Jan 1993)>>endobj'); - }, 60000); // 60 seconds - -}); diff --git a/apps/api/src/scraper/WebScraper/utils/__tests__/removeUnwantedElements.test.ts b/apps/api/src/scraper/WebScraper/utils/__tests__/removeUnwantedElements.test.ts deleted file mode 100644 index b3d4a244..00000000 --- a/apps/api/src/scraper/WebScraper/utils/__tests__/removeUnwantedElements.test.ts +++ /dev/null @@ -1,192 +0,0 @@ -import { removeUnwantedElements } from "../removeUnwantedElements"; -import { PageOptions } from "../../../../lib/entities"; - -describe('removeUnwantedElements', () => { - it('should remove script, style, iframe, noscript, meta, and head tags', () => { - const html = `Test
Content
`; - const options: PageOptions = {}; - const result = removeUnwantedElements(html, options); - expect(result).not.toContain('