Compare commits
2 Commits
main
...
7793588e0b
| Author | SHA1 | Date | |
|---|---|---|---|
| 7793588e0b | |||
| 23b14536cc |
16
.dockerignore
Normal file
16
.dockerignore
Normal file
@@ -0,0 +1,16 @@
|
||||
node_modules
|
||||
Dockerfile*
|
||||
docker-compose*
|
||||
.dockerignore
|
||||
.git
|
||||
.gitignore
|
||||
README.md
|
||||
LICENSE
|
||||
.vscode
|
||||
Makefile
|
||||
helm-charts
|
||||
.env
|
||||
.dev.vars
|
||||
.editorconfig
|
||||
.idea
|
||||
coverage*
|
||||
@@ -1 +1 @@
|
||||
nlx lint-staged
|
||||
bunx lint-staged
|
||||
|
||||
41
Dockerfile
Normal file
41
Dockerfile
Normal file
@@ -0,0 +1,41 @@
|
||||
# use the official Bun image
|
||||
# see all versions at https://hub.docker.com/r/oven/bun/tags
|
||||
FROM oven/bun:1 as base
|
||||
WORKDIR /usr/app
|
||||
|
||||
# install dependencies into temp directory
|
||||
# this will cache them and speed up future builds
|
||||
FROM base AS install
|
||||
RUN mkdir -p /tmp/dev
|
||||
COPY package.json bun.lockb /tmp/dev/
|
||||
RUN cd /tmp/dev && bun install --frozen-lockfile
|
||||
|
||||
# install with --production (exclude devDependencies)
|
||||
RUN mkdir -p /tmp/prod
|
||||
COPY package.json bun.lockb /tmp/prod/
|
||||
RUN cd /tmp/prod && bun install --frozen-lockfile --production
|
||||
|
||||
# copy node_modules from temp directory
|
||||
# then copy all (non-ignored) project files into the image
|
||||
FROM base AS prerelease
|
||||
COPY --from=install /tmp/dev/node_modules node_modules
|
||||
COPY . .
|
||||
|
||||
# [optional] tests & build
|
||||
ENV NODE_ENV=production
|
||||
RUN bun test
|
||||
RUN bun build --compile src/index.ts --outfile=aniplay
|
||||
|
||||
# copy production dependencies and source code into final image
|
||||
FROM base AS release
|
||||
COPY --from=install /tmp/prod/node_modules node_modules
|
||||
COPY --from=prerelease /usr/app/src ./src
|
||||
COPY --from=prerelease /usr/app/package.json .
|
||||
COPY --from=prerelease /usr/app/tsconfig.json .
|
||||
# TODO: uncomment once v2 is ready
|
||||
# COPY --from=prerelease /usr/app/drizzle.config.ts .
|
||||
|
||||
# run the app
|
||||
USER bun
|
||||
EXPOSE 3000
|
||||
ENTRYPOINT [ "bun", "run", "prod:server" ]
|
||||
76
README.md
76
README.md
@@ -1,72 +1,12 @@
|
||||
# Aniplay API
|
||||
```
|
||||
npm install
|
||||
npm run dev
|
||||
```
|
||||
|
||||
API for [Aniplay](https://github.com/silverAndroid/aniplay), built with Cloudflare Workers, Hono, and Drizzle ORM.
|
||||
|
||||
## Tech Stack
|
||||
|
||||
- **Cloudflare Workers**: Serverless execution environment.
|
||||
- **Hono**: Ultrafast web framework (OpenAPI).
|
||||
- **GraphQL**: Used internally for communicating with the [AniList](https://anilist.co) API.
|
||||
- **Drizzle ORM**: TypeScript ORM for D1 (Cloudflare's serverless SQL database).
|
||||
- **Vitest**: Testing framework.
|
||||
|
||||
## Prerequisites
|
||||
|
||||
- **Node.js**
|
||||
- **pnpm**: Package manager.
|
||||
|
||||
## Getting Started
|
||||
|
||||
1. **Installation**
|
||||
|
||||
```bash
|
||||
pnpm install
|
||||
```
|
||||
|
||||
2. **Environment Setup**
|
||||
Generate the environment types:
|
||||
|
||||
```bash
|
||||
pnpm exec wrangler types
|
||||
```
|
||||
|
||||
3. **Database Setup**
|
||||
Apply migrations to the local D1 database:
|
||||
```bash
|
||||
pnpm exec wrangler d1 migrations apply aniplay
|
||||
```
|
||||
```
|
||||
npm run deploy
|
||||
```
|
||||
|
||||
## Development
|
||||
|
||||
### Running Locally
|
||||
|
||||
Start the development server:
|
||||
|
||||
```bash
|
||||
pnpm run dev
|
||||
```
|
||||
|
||||
### Testing
|
||||
|
||||
Run the tests using Vitest:
|
||||
|
||||
```bash
|
||||
pnpm test
|
||||
```
|
||||
|
||||
## Deployment
|
||||
|
||||
Deploy to Cloudflare Workers:
|
||||
|
||||
```bash
|
||||
pnpm run deploy
|
||||
```
|
||||
|
||||
## Project Structure
|
||||
|
||||
- `src/controllers`: API route handlers (titles, episodes, search, etc.)
|
||||
- `src/libs`: Shared utilities and logic (AniList integration, background tasks)
|
||||
- `src/middleware`: Middleware handlers (authentication, authorization, etc.)
|
||||
- `src/models`: Database schema and models
|
||||
- `src/scripts`: Utility scripts for maintenance and setup
|
||||
- `src/types`: TypeScript type definitions
|
||||
If a route is internal-only or doesn't need to appear on the OpenAPI spec (that's autogenerated by Hono), use the `Hono` class. Otherwise, use the `OpenAPIHono` class from `@hono/zod-openapi`.
|
||||
|
||||
7
bunfig.toml
Normal file
7
bunfig.toml
Normal file
@@ -0,0 +1,7 @@
|
||||
[test]
|
||||
preload = [
|
||||
"./testSetup.ts",
|
||||
"./src/mocks/consumet.ts",
|
||||
"./src/mocks/getGoogleAuthToken.ts",
|
||||
"./src/mocks/cloudflare.ts",
|
||||
]
|
||||
@@ -6,7 +6,8 @@
|
||||
"type": "module",
|
||||
"scripts": {
|
||||
"dev": "wrangler dev src/index.ts --port 8080",
|
||||
"deploy": "wrangler deploy --minify src/index.ts",
|
||||
"env:generate": "tsx src/scripts/generateEnv.ts",
|
||||
"env:verify": "tsx src/scripts/verifyEnv.ts",
|
||||
"db:generate": "drizzle-kit generate",
|
||||
"db:migrate": "drizzle-kit migrate",
|
||||
"test": "vitest",
|
||||
@@ -16,6 +17,7 @@
|
||||
"tsx": "tsx"
|
||||
},
|
||||
"dependencies": {
|
||||
"@consumet/extensions": "github:consumet/consumet.ts#3dd0ccb",
|
||||
"@hono/swagger-ui": "^0.5.1",
|
||||
"@hono/zod-openapi": "^0.19.5",
|
||||
"@hono/zod-validator": "^0.2.2",
|
||||
@@ -25,14 +27,15 @@
|
||||
"graphql-request": "^7.4.0",
|
||||
"hono": "^4.7.7",
|
||||
"jose": "^5.10.0",
|
||||
"lodash.isequal": "^4.5.0",
|
||||
"lodash.mapkeys": "^4.6.0",
|
||||
"luxon": "^3.6.1",
|
||||
"zod": "^3.24.3"
|
||||
},
|
||||
"devDependencies": {
|
||||
"@cloudflare/vitest-pool-workers": "^0.10.15",
|
||||
"@graphql-typed-document-node/core": "^3.2.0",
|
||||
"@trivago/prettier-plugin-sort-imports": "^4.3.0",
|
||||
"@types/lodash.isequal": "^4.5.8",
|
||||
"@types/lodash.mapkeys": "^4.6.9",
|
||||
"@types/luxon": "^3.6.2",
|
||||
"@types/node": "^25.0.1",
|
||||
@@ -51,6 +54,7 @@
|
||||
"ts-morph": "^22.0.0",
|
||||
"tsx": "^4.20.6",
|
||||
"typescript": "^5.8.3",
|
||||
"util": "^0.12.5",
|
||||
"vitest": "~3.2.4",
|
||||
"wrangler": "^4.51.0",
|
||||
"zx": "8.1.5"
|
||||
|
||||
554
pnpm-lock.yaml
generated
554
pnpm-lock.yaml
generated
@@ -7,6 +7,9 @@ settings:
|
||||
importers:
|
||||
.:
|
||||
dependencies:
|
||||
"@consumet/extensions":
|
||||
specifier: github:consumet/consumet.ts#3dd0ccb
|
||||
version: https://codeload.github.com/consumet/consumet.ts/tar.gz/3dd0ccb
|
||||
"@hono/swagger-ui":
|
||||
specifier: ^0.5.1
|
||||
version: 0.5.2(hono@4.10.8)
|
||||
@@ -34,6 +37,9 @@ importers:
|
||||
jose:
|
||||
specifier: ^5.10.0
|
||||
version: 5.10.0
|
||||
lodash.isequal:
|
||||
specifier: ^4.5.0
|
||||
version: 4.5.0
|
||||
lodash.mapkeys:
|
||||
specifier: ^4.6.0
|
||||
version: 4.6.0
|
||||
@@ -47,12 +53,12 @@ importers:
|
||||
"@cloudflare/vitest-pool-workers":
|
||||
specifier: ^0.10.15
|
||||
version: 0.10.15(@vitest/runner@3.2.4)(@vitest/snapshot@3.2.4)(vitest@3.2.4)
|
||||
"@graphql-typed-document-node/core":
|
||||
specifier: ^3.2.0
|
||||
version: 3.2.0(graphql@16.12.0)
|
||||
"@trivago/prettier-plugin-sort-imports":
|
||||
specifier: ^4.3.0
|
||||
version: 4.3.0(prettier@3.7.4)
|
||||
"@types/lodash.isequal":
|
||||
specifier: ^4.5.8
|
||||
version: 4.5.8
|
||||
"@types/lodash.mapkeys":
|
||||
specifier: ^4.6.9
|
||||
version: 4.6.9
|
||||
@@ -107,6 +113,9 @@ importers:
|
||||
typescript:
|
||||
specifier: ^5.8.3
|
||||
version: 5.9.3
|
||||
util:
|
||||
specifier: ^0.12.5
|
||||
version: 0.12.5
|
||||
vitest:
|
||||
specifier: ~3.2.4
|
||||
version: 3.2.4(@types/node@25.0.1)(@vitest/ui@3.2.4)(tsx@4.21.0)(yaml@2.8.2)
|
||||
@@ -429,6 +438,13 @@ packages:
|
||||
cpu: [x64]
|
||||
os: [win32]
|
||||
|
||||
"@consumet/extensions@https://codeload.github.com/consumet/consumet.ts/tar.gz/3dd0ccb":
|
||||
resolution:
|
||||
{
|
||||
tarball: https://codeload.github.com/consumet/consumet.ts/tar.gz/3dd0ccb,
|
||||
}
|
||||
version: 1.7.0
|
||||
|
||||
"@cspotcode/source-map-support@0.8.1":
|
||||
resolution:
|
||||
{
|
||||
@@ -1953,6 +1969,12 @@ packages:
|
||||
integrity: sha512-D5qGUYwjvnNNextdU59/+fI+spnwtTFmyQP0h+PfIOSkNfpU6AOICUOkm4i0OnSk+NyjdPJrxCDro0sJsWlRpQ==,
|
||||
}
|
||||
|
||||
"@types/lodash.isequal@4.5.8":
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-uput6pg4E/tj2LGxCZo9+y27JNyB2OZuuI/T5F+ylVDYuqICLG2/ktjxx0v6GvVntAf8TvEzeQLcV0ffRirXuA==,
|
||||
}
|
||||
|
||||
"@types/lodash.mapkeys@4.6.9":
|
||||
resolution:
|
||||
{
|
||||
@@ -2132,6 +2154,12 @@ packages:
|
||||
integrity: sha512-xvsWESUJn0JN421Xb9MQw6AsMHRCUknCe0Wjlxvjud80mU4E6hQf1A6NzQKcYNmYw62MfzEtXc+badstZP3JpQ==,
|
||||
}
|
||||
|
||||
ascii-url-encoder@1.2.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-jRQMrz1ZqL7gnK60Xbu1S8kDi01UN+lHBnhvXZZqjd7nms7t4jOv4o9nE8bbfDh1Ts/KEtW2nXGR5CmydDBmlw==,
|
||||
}
|
||||
|
||||
assertion-error@2.0.1:
|
||||
resolution:
|
||||
{
|
||||
@@ -2145,6 +2173,19 @@ packages:
|
||||
integrity: sha512-Oei9OH4tRh0YqU3GxhX79dM/mwVgvbZJaSNaRk+bshkj0S5cfHcgYakreBjrHwatXKbz+IoIdYLxrKim2MjW0Q==,
|
||||
}
|
||||
|
||||
available-typed-arrays@1.0.7:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-wvUjBtSGN7+7SjNpq/9M2Tg350UZD3q62IFZLbRAR1bSMlCo1ZaeW+BJ+D090e4hIIZLBcTDWe4Mh4jvUDajzQ==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
axios@0.27.2:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-t+yRIyySRTp/wua5xEr+z1q60QmLq8ABsS5O9Me1AsE5dfKqgnCFzwiCZZ/cGNd1lq4/7akDWMxdhVlucjmnOQ==,
|
||||
}
|
||||
|
||||
balanced-match@1.0.2:
|
||||
resolution:
|
||||
{
|
||||
@@ -2170,6 +2211,12 @@ packages:
|
||||
integrity: sha512-F1+K8EbfOZE49dtoPtmxUQrpXaBIl3ICvasLh+nJta0xkz+9kF/7uet9fLnwKqhDrmj6g+6K3Tw9yQPUg2ka5g==,
|
||||
}
|
||||
|
||||
boolbase@1.0.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-JZOSA7Mo9sNGB8+UjSgzdLtokWAky1zbztM3WRLCbZ70/3cTANmQmOdR7y2g+J0e2WXywy1yS468tY+IruqEww==,
|
||||
}
|
||||
|
||||
brace-expansion@2.0.2:
|
||||
resolution:
|
||||
{
|
||||
@@ -2217,6 +2264,20 @@ packages:
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
call-bind@1.0.8:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-oKlSFMcMwpUg2ednkhQ454wfWiU/ul3CkJe/PEHcTKuiX6RpbehUiFMXu13HalGZxfUwCQzZG747YXBn1im9ww==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
call-bound@1.0.4:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-+ys997U96po4Kx/ABpBCqhA9EuxJaQWDQg7295H4hBphv3IZg0boBKuwYpt4YXp6MZ5AmZQnU/tyMTlRpaSejg==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
caniuse-lite@1.0.30001760:
|
||||
resolution:
|
||||
{
|
||||
@@ -2244,6 +2305,19 @@ packages:
|
||||
}
|
||||
engines: { node: ">= 16" }
|
||||
|
||||
cheerio-select@2.1.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-9v9kG0LvzrlcungtnJtpGNxY+fzECQKhK4EGJX2vByejiMX84MFNQw4UxPJl3bFbTMw+Dfs37XaIkCwTZfLh4g==,
|
||||
}
|
||||
|
||||
cheerio@1.1.2:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-IkxPpb5rS/d1IiLbHMgfPuS0FgiWTtFIm/Nj+2woXDLTZ7fOT2eqzgYbdMlLweqlHbsZjxEChoVK+7iph7jyQg==,
|
||||
}
|
||||
engines: { node: ">=20.18.1" }
|
||||
|
||||
cjs-module-lexer@1.4.3:
|
||||
resolution:
|
||||
{
|
||||
@@ -2349,6 +2423,25 @@ packages:
|
||||
}
|
||||
engines: { node: ">= 8" }
|
||||
|
||||
crypto-js@4.2.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-KALDyEYgpY+Rlob/iriUtjV6d5Eq+Y191A5g4UqLAi8CyGP9N1+FdVbkc1SxKc2r4YAYqG8JzO2KGL+AizD70Q==,
|
||||
}
|
||||
|
||||
css-select@5.2.2:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-TizTzUddG/xYLA3NXodFM0fSbNizXjOKhqiQQwvhlspadZokn1KDy0NZFS0wuEubIYAV5/c1/lAr0TaaFXEXzw==,
|
||||
}
|
||||
|
||||
css-what@6.2.2:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-u/O3vwbptzhMs3L1fQE82ZSLHQQfto5gyZzwteVIEyeaY5Fc7R4dapF/BvRoSYFeqfBk4m0V1Vafq5Pjv25wvA==,
|
||||
}
|
||||
engines: { node: ">= 6" }
|
||||
|
||||
data-uri-to-buffer@2.0.2:
|
||||
resolution:
|
||||
{
|
||||
@@ -2374,6 +2467,13 @@ packages:
|
||||
}
|
||||
engines: { node: ">=6" }
|
||||
|
||||
define-data-property@1.1.4:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-rBMvIzlpA8v6E+SJZoo++HAYqsLrkg7MSfIinMPFhmkorw7X+dOXVJQs+QT69zGkzMyfDnIMN2Wid1+NbL3T+A==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
delayed-stream@1.0.0:
|
||||
resolution:
|
||||
{
|
||||
@@ -2394,6 +2494,31 @@ packages:
|
||||
integrity: sha512-jDwizj+IlEZBunHcOuuFVBnIMPAEHvTsJj0BcIp94xYguLRVBcXO853px/MyIJvbVzWdsGvrRweIUWJw8hBP7A==,
|
||||
}
|
||||
|
||||
dom-serializer@2.0.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-wIkAryiqt/nV5EQKqQpo3SToSOV9J0DnbJqwK7Wv/Trc92zIAYZ4FlMu+JPFW1DfGFt81ZTCGgDEabffXeLyJg==,
|
||||
}
|
||||
|
||||
domelementtype@2.3.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-OLETBj6w0OsagBwdXnPdN0cnMfF9opN69co+7ZrbfPGrdpPVNBUj02spi6B1N7wChLQiPn4CSH/zJvXw56gmHw==,
|
||||
}
|
||||
|
||||
domhandler@5.0.3:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-cgwlv/1iFQiFnU96XXgROh8xTeetsnJiDsTc7TYCLFd9+/WNkIqPTxiM/8pSd8VIrhXGTf1Ny1q1hquVqDJB5w==,
|
||||
}
|
||||
engines: { node: ">= 4" }
|
||||
|
||||
domutils@3.2.2:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-6kZKyUajlDuqlHKVX1w7gyslj9MPIXzIFiz/rGu35uC1wMi+kMhQwGhl4lt9unC9Vb9INnY9Z3/ZA3+FhASLaw==,
|
||||
}
|
||||
|
||||
dotenv@17.2.3:
|
||||
resolution:
|
||||
{
|
||||
@@ -2546,6 +2671,26 @@ packages:
|
||||
integrity: sha512-L18DaJsXSUk2+42pv8mLs5jJT2hqFkFE4j21wOmgbUqsZ2hL72NsUU785g9RXgo3s0ZNgVl42TiHp3ZtOv/Vyg==,
|
||||
}
|
||||
|
||||
encoding-sniffer@0.2.1:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-5gvq20T6vfpekVtqrYQsSCFZ1wEg5+wW0/QaZMWkFr6BqD3NfKs0rLCx4rrVlSWJeZb5NBJgVLswK/w2MWU+Gw==,
|
||||
}
|
||||
|
||||
entities@4.5.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-V0hjH4dGPh9Ao5p0MoRY6BVqtwCjhz6vI5LT8AJ55H+4g9/4vbHx1I54fS0XuclLhDHArPQCiMjDxjaL8fPxhw==,
|
||||
}
|
||||
engines: { node: ">=0.12" }
|
||||
|
||||
entities@6.0.1:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-aN97NXWF6AWBTahfVOIrB/NShkzi5H7F9r1s9mD3cDj4Ko5f2qhhVoYMibXF7GlLveb/D2ioWay8lxI97Ven3g==,
|
||||
}
|
||||
engines: { node: ">=0.12" }
|
||||
|
||||
environment@1.1.0:
|
||||
resolution:
|
||||
{
|
||||
@@ -2730,6 +2875,25 @@ packages:
|
||||
integrity: sha512-GX+ysw4PBCz0PzosHDepZGANEuFCMLrnRTiEy9McGjmkCQYwRq4A/X786G/fjM/+OjsWSU1ZrY5qyARZmO/uwg==,
|
||||
}
|
||||
|
||||
follow-redirects@1.15.11:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-deG2P0JfjrTxl50XGCDyfI97ZGVCxIpfKYmfyrQ54n5FO/0gfIES8C/Psl6kWVDolizcaaxZJnTS0QSMxvnsBQ==,
|
||||
}
|
||||
engines: { node: ">=4.0" }
|
||||
peerDependencies:
|
||||
debug: "*"
|
||||
peerDependenciesMeta:
|
||||
debug:
|
||||
optional: true
|
||||
|
||||
for-each@0.3.5:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-dKx12eRCVIzqCxFGplyFKJMPvLEWgmNtUrpTiJIR5u97zEhRG8ySrtboPHZXx7daLxQVrl643cTzbab2tkQjxg==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
foreground-child@3.3.1:
|
||||
resolution:
|
||||
{
|
||||
@@ -2778,6 +2942,13 @@ packages:
|
||||
}
|
||||
engines: { node: ">=14" }
|
||||
|
||||
generator-function@2.0.1:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-SFdFmIJi+ybC0vjlHN0ZGVGHc3lgE0DxPAT0djjVg+kjOnSqclqmj0KQ7ykTOLP6YxoqOvuAODGdcHJn+43q3g==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
gensync@1.0.0-beta.2:
|
||||
resolution:
|
||||
{
|
||||
@@ -2897,6 +3068,12 @@ packages:
|
||||
}
|
||||
engines: { node: ">=8" }
|
||||
|
||||
has-property-descriptors@1.0.2:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-55JNKuIW+vq4Ke1BjOTjM2YctQIvCT7GFzHwmfZPGo5wnrgkid0YQtnAleFSqumZm4az3n2BS+erby5ipJdgrg==,
|
||||
}
|
||||
|
||||
has-symbols@1.1.0:
|
||||
resolution:
|
||||
{
|
||||
@@ -2931,6 +3108,12 @@ packages:
|
||||
integrity: sha512-H2iMtd0I4Mt5eYiapRdIDjp+XzelXQ0tFE4JS7YFwFevXXMmOp9myNrUvCg0D6ws8iqkRPBfKHgbwig1SmlLfg==,
|
||||
}
|
||||
|
||||
htmlparser2@10.0.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-TwAZM+zE5Tq3lrEHvOlvwgj1XLWQCtaaibSN11Q+gGBAS7Y1uZSWwXXRe4iF6OXnaq1riyQAPFOBtYc77Mxq0g==,
|
||||
}
|
||||
|
||||
https-proxy-agent@7.0.6:
|
||||
resolution:
|
||||
{
|
||||
@@ -2959,12 +3142,39 @@ packages:
|
||||
engines: { node: ">=18" }
|
||||
hasBin: true
|
||||
|
||||
iconv-lite@0.6.3:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-4fCk79wshMdzMp2rH06qWrJE4iolqLhCUH+OiuIgU++RB0+94NlDL81atO7GX55uUKueo0txHNtvEyI6D7WdMw==,
|
||||
}
|
||||
engines: { node: ">=0.10.0" }
|
||||
|
||||
inherits@2.0.4:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-k/vGaX4/Yla3WzyMCvTQOXYeIHvqOKtnqBduzTHpzpQZzAskKMhZ2K+EnBiSM9zGSoIFeMpXKxa4dYeZIQqewQ==,
|
||||
}
|
||||
|
||||
is-arguments@1.2.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-7bVbi0huj/wrIAOzb8U1aszg9kdi3KN/CyU19CTI7tAoZYEZoL9yCDXpbXN+uPsuWnP02cyug1gleqq+TU+YCA==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
is-arrayish@0.3.4:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-m6UrgzFVUYawGBh1dUsWR5M2Clqic9RVXC/9f8ceNlv2IcO9j9J/z8UoCLPqtsPBFNzEpfR3xftohbfqDx8EQA==,
|
||||
}
|
||||
|
||||
is-callable@1.2.7:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-1BC0BVFhS/p0qtw6enp8e+8OD0UrK0oFLztSjNzhcKA3WDuJxxAPXzPuPtKkjEY9UUoEWlX/8fgKeu2S8i9JTA==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
is-extglob@2.1.1:
|
||||
resolution:
|
||||
{
|
||||
@@ -2993,6 +3203,13 @@ packages:
|
||||
}
|
||||
engines: { node: ">=18" }
|
||||
|
||||
is-generator-function@1.1.2:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-upqt1SkGkODW9tsGNG5mtXTXtECizwtS2kA161M+gJPc1xdb/Ax629af6YrTwcOeQHbewrPNlE5Dx7kzvXTizA==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
is-glob@4.0.3:
|
||||
resolution:
|
||||
{
|
||||
@@ -3007,6 +3224,13 @@ packages:
|
||||
}
|
||||
engines: { node: ">=0.12.0" }
|
||||
|
||||
is-regex@1.2.1:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-MjYsKHO5O7mCsmRGxWcLWheFqN9DJ/2TmngvjKXihe6efViPqc274+Fx/4fYj/r03+ESvBdTXK0V6tA3rgez1g==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
is-stream@2.0.1:
|
||||
resolution:
|
||||
{
|
||||
@@ -3021,6 +3245,13 @@ packages:
|
||||
}
|
||||
engines: { node: ^12.20.0 || ^14.13.1 || >=16.0.0 }
|
||||
|
||||
is-typed-array@1.1.15:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-p3EcsicXjit7SaskXHs1hA91QxgTw46Fv6EFKKGS5DRFLD8yKnohjF3hxoju94b/OcMZoQukzpPpBE9uLVKzgQ==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
isexe@2.0.0:
|
||||
resolution:
|
||||
{
|
||||
@@ -3157,6 +3388,13 @@ packages:
|
||||
}
|
||||
engines: { node: ">=18.0.0" }
|
||||
|
||||
lodash.isequal@4.5.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-pDo3lu8Jhfjqls6GkMgpahsF9kCyayhgykjyLMNFTKWrpVdAQtYyB4muAMWozBB4ig/dtWAmsMxLEI8wuz+DYQ==,
|
||||
}
|
||||
deprecated: This package is deprecated. Use require('node:util').isDeepStrictEqual instead.
|
||||
|
||||
lodash.mapkeys@4.6.0:
|
||||
resolution:
|
||||
{
|
||||
@@ -3382,6 +3620,12 @@ packages:
|
||||
}
|
||||
engines: { node: ^12.20.0 || ^14.13.1 || >=16.0.0 }
|
||||
|
||||
nth-check@2.1.1:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-lqjrjmaOoAnWfMmBPL+XNnynZh2+swxiX3WUE0s4yEHI6m+AwrK2UZOimIRl3X/4QctVqS8AiZjFqyOGrMXb/w==,
|
||||
}
|
||||
|
||||
onetime@6.0.0:
|
||||
resolution:
|
||||
{
|
||||
@@ -3408,6 +3652,24 @@ packages:
|
||||
integrity: sha512-UEZIS3/by4OC8vL3P2dTXRETpebLI2NiI5vIrjaD/5UtrkFX/tNbwjTSRAGC/+7CAo2pIcBaRgWmcBBHcsaCIw==,
|
||||
}
|
||||
|
||||
parse5-htmlparser2-tree-adapter@7.1.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-ruw5xyKs6lrpo9x9rCZqZZnIUntICjQAd0Wsmp396Ul9lN/h+ifgVV1x1gZHi8euej6wTfpqX8j+BFQxF0NS/g==,
|
||||
}
|
||||
|
||||
parse5-parser-stream@7.1.2:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-JyeQc9iwFLn5TbvvqACIF/VXG6abODeB3Fwmv/TGdLk2LfbWkaySGY72at4+Ty7EkPZj854u4CrICqNk2qIbow==,
|
||||
}
|
||||
|
||||
parse5@7.3.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-IInvU7fabl34qmi9gY8XOVxhYyMyuH2xUNpb2q8/Y+7552KlejkRvqvD19nMoUW/uQGGbqNpA6Tufu5FL5BZgw==,
|
||||
}
|
||||
|
||||
path-browserify@1.0.1:
|
||||
resolution:
|
||||
{
|
||||
@@ -3482,6 +3744,13 @@ packages:
|
||||
engines: { node: ">=0.10" }
|
||||
hasBin: true
|
||||
|
||||
possible-typed-array-names@1.1.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-/+5VFTchJDoVj3bhoqi6UeymcD00DAwb1nJwamzPvHEszJ4FpF6SNNbUbOS8yI56qHzdV8eK0qEfOSiodkTdxg==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
postcss@8.5.6:
|
||||
resolution:
|
||||
{
|
||||
@@ -3564,6 +3833,19 @@ packages:
|
||||
integrity: sha512-rp3So07KcdmmKbGvgaNxQSJr7bGVSVk5S9Eq1F+ppbRo70+YeaDxkw5Dd8NPN+GD6bjnYm2VuPuCXmpuYvmCXQ==,
|
||||
}
|
||||
|
||||
safe-regex-test@1.1.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-x/+Cz4YrimQxQccJf5mKEbIa1NzeCRNI5Ecl/ekmlYaampdNLPalVyIcCZNNH3MvmqBugV5TMYZXv0ljslUlaw==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
safer-buffer@2.1.2:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-YZo3K82SD7Riyi0E1EQPojLz7kpepnSQI9IyPbHHg1XXXevb5dJI7tpyN2ADxGcQbHG7vcyRHk0cbwqcQriUtg==,
|
||||
}
|
||||
|
||||
semver@6.3.1:
|
||||
resolution:
|
||||
{
|
||||
@@ -3579,6 +3861,13 @@ packages:
|
||||
engines: { node: ">=10" }
|
||||
hasBin: true
|
||||
|
||||
set-function-length@1.2.2:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-pgRc4hJ4/sNjWCSS9AmnS40x3bNMDTknHgL5UaMBTMyJnU90EgWh1Rz+MC9eFu4BuN/UwZjKQuY/1v3rM7HMfg==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
sharp@0.33.5:
|
||||
resolution:
|
||||
{
|
||||
@@ -3889,6 +4178,13 @@ packages:
|
||||
}
|
||||
engines: { node: ">=20.18.1" }
|
||||
|
||||
undici@7.16.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-QEg3HPMll0o3t2ourKwOeUAZ159Kn9mx5pnzHRQO8+Wixmh88YdZRiIwat0iNzNNXn0yoEtXJqFpyW7eM8BV7g==,
|
||||
}
|
||||
engines: { node: ">=20.18.1" }
|
||||
|
||||
unenv@2.0.0-rc.24:
|
||||
resolution:
|
||||
{
|
||||
@@ -3904,6 +4200,12 @@ packages:
|
||||
peerDependencies:
|
||||
browserslist: ">= 4.21.0"
|
||||
|
||||
util@0.12.5:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-kZf/K6hEIrWHI6XqOFUiiMa+79wE/D8Q+NCNAWclkyg3b4d2k7s0QGepNjiABc+aR3N1PAyHL7p6UcLY6LmrnA==,
|
||||
}
|
||||
|
||||
uuid@9.0.1:
|
||||
resolution:
|
||||
{
|
||||
@@ -4006,12 +4308,33 @@ packages:
|
||||
integrity: sha512-2JAn3z8AR6rjK8Sm8orRC0h/bcl/DqL7tRPdGZ4I1CjdF+EaMLmYxBHyXuKL849eucPFhvBoxMsflfOb8kxaeQ==,
|
||||
}
|
||||
|
||||
whatwg-encoding@3.1.1:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-6qN4hJdMwfYBtE3YBTTHhoeuUrDBPZmbQaxWAqSALV/MeEnR5z1xd8UKud2RAkFoPkmB+hli1TZSnyi84xz1vQ==,
|
||||
}
|
||||
engines: { node: ">=18" }
|
||||
|
||||
whatwg-mimetype@4.0.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-QaKxh0eNIi2mE9p2vEdzfagOKHCcj1pJ56EEHGQOVxp8r9/iszLUUV7v89x9O1p/T+NlTM5W7jW6+cz4Fq1YVg==,
|
||||
}
|
||||
engines: { node: ">=18" }
|
||||
|
||||
whatwg-url@5.0.0:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-saE57nupxk6v3HY35+jzBwYa0rKSy0XR8JSxZPwgLr7ys0IBzhGviA1/TUGJLmSVqs8pb9AnvICXEuOHLprYTw==,
|
||||
}
|
||||
|
||||
which-typed-array@1.1.19:
|
||||
resolution:
|
||||
{
|
||||
integrity: sha512-rEvr90Bck4WZt9HHFC4DJMsjvu7x+r6bImz0/BrbWb7A2djJ8hnZMrWnHo9F8ssv0OMErasDhftrfROTyqSDrw==,
|
||||
}
|
||||
engines: { node: ">= 0.4" }
|
||||
|
||||
which@2.0.2:
|
||||
resolution:
|
||||
{
|
||||
@@ -4361,6 +4684,17 @@ snapshots:
|
||||
"@cloudflare/workerd-windows-64@1.20251210.0":
|
||||
optional: true
|
||||
|
||||
"@consumet/extensions@https://codeload.github.com/consumet/consumet.ts/tar.gz/3dd0ccb":
|
||||
dependencies:
|
||||
ascii-url-encoder: 1.2.0
|
||||
axios: 0.27.2
|
||||
cheerio: 1.1.2
|
||||
crypto-js: 4.2.0
|
||||
form-data: 4.0.5
|
||||
husky: 9.1.7
|
||||
transitivePeerDependencies:
|
||||
- debug
|
||||
|
||||
"@cspotcode/source-map-support@0.8.1":
|
||||
dependencies:
|
||||
"@jridgewell/trace-mapping": 0.3.9
|
||||
@@ -4977,6 +5311,10 @@ snapshots:
|
||||
"@types/node": 25.0.1
|
||||
optional: true
|
||||
|
||||
"@types/lodash.isequal@4.5.8":
|
||||
dependencies:
|
||||
"@types/lodash": 4.17.21
|
||||
|
||||
"@types/lodash.mapkeys@4.6.9":
|
||||
dependencies:
|
||||
"@types/lodash": 4.17.21
|
||||
@@ -5099,10 +5437,23 @@ snapshots:
|
||||
dependencies:
|
||||
printable-characters: 1.0.42
|
||||
|
||||
ascii-url-encoder@1.2.0: {}
|
||||
|
||||
assertion-error@2.0.1: {}
|
||||
|
||||
asynckit@0.4.0: {}
|
||||
|
||||
available-typed-arrays@1.0.7:
|
||||
dependencies:
|
||||
possible-typed-array-names: 1.1.0
|
||||
|
||||
axios@0.27.2:
|
||||
dependencies:
|
||||
follow-redirects: 1.15.11
|
||||
form-data: 4.0.5
|
||||
transitivePeerDependencies:
|
||||
- debug
|
||||
|
||||
balanced-match@1.0.2: {}
|
||||
|
||||
baseline-browser-mapping@2.9.7: {}
|
||||
@@ -5111,6 +5462,8 @@ snapshots:
|
||||
|
||||
blake3-wasm@2.1.5: {}
|
||||
|
||||
boolbase@1.0.0: {}
|
||||
|
||||
brace-expansion@2.0.2:
|
||||
dependencies:
|
||||
balanced-match: 1.0.2
|
||||
@@ -5138,6 +5491,18 @@ snapshots:
|
||||
es-errors: 1.3.0
|
||||
function-bind: 1.1.2
|
||||
|
||||
call-bind@1.0.8:
|
||||
dependencies:
|
||||
call-bind-apply-helpers: 1.0.2
|
||||
es-define-property: 1.0.1
|
||||
get-intrinsic: 1.3.0
|
||||
set-function-length: 1.2.2
|
||||
|
||||
call-bound@1.0.4:
|
||||
dependencies:
|
||||
call-bind-apply-helpers: 1.0.2
|
||||
get-intrinsic: 1.3.0
|
||||
|
||||
caniuse-lite@1.0.30001760: {}
|
||||
|
||||
chai@5.3.3:
|
||||
@@ -5152,6 +5517,29 @@ snapshots:
|
||||
|
||||
check-error@2.1.1: {}
|
||||
|
||||
cheerio-select@2.1.0:
|
||||
dependencies:
|
||||
boolbase: 1.0.0
|
||||
css-select: 5.2.2
|
||||
css-what: 6.2.2
|
||||
domelementtype: 2.3.0
|
||||
domhandler: 5.0.3
|
||||
domutils: 3.2.2
|
||||
|
||||
cheerio@1.1.2:
|
||||
dependencies:
|
||||
cheerio-select: 2.1.0
|
||||
dom-serializer: 2.0.0
|
||||
domhandler: 5.0.3
|
||||
domutils: 3.2.2
|
||||
encoding-sniffer: 0.2.1
|
||||
htmlparser2: 10.0.0
|
||||
parse5: 7.3.0
|
||||
parse5-htmlparser2-tree-adapter: 7.1.0
|
||||
parse5-parser-stream: 7.1.2
|
||||
undici: 7.16.0
|
||||
whatwg-mimetype: 4.0.0
|
||||
|
||||
cjs-module-lexer@1.4.3: {}
|
||||
|
||||
cli-cursor@5.0.0:
|
||||
@@ -5213,6 +5601,18 @@ snapshots:
|
||||
shebang-command: 2.0.0
|
||||
which: 2.0.2
|
||||
|
||||
crypto-js@4.2.0: {}
|
||||
|
||||
css-select@5.2.2:
|
||||
dependencies:
|
||||
boolbase: 1.0.0
|
||||
css-what: 6.2.2
|
||||
domhandler: 5.0.3
|
||||
domutils: 3.2.2
|
||||
nth-check: 2.1.1
|
||||
|
||||
css-what@6.2.2: {}
|
||||
|
||||
data-uri-to-buffer@2.0.2: {}
|
||||
|
||||
debug@4.4.3:
|
||||
@@ -5221,12 +5621,36 @@ snapshots:
|
||||
|
||||
deep-eql@5.0.2: {}
|
||||
|
||||
define-data-property@1.1.4:
|
||||
dependencies:
|
||||
es-define-property: 1.0.1
|
||||
es-errors: 1.3.0
|
||||
gopd: 1.2.0
|
||||
|
||||
delayed-stream@1.0.0: {}
|
||||
|
||||
detect-libc@2.1.2: {}
|
||||
|
||||
devalue@5.6.1: {}
|
||||
|
||||
dom-serializer@2.0.0:
|
||||
dependencies:
|
||||
domelementtype: 2.3.0
|
||||
domhandler: 5.0.3
|
||||
entities: 4.5.0
|
||||
|
||||
domelementtype@2.3.0: {}
|
||||
|
||||
domhandler@5.0.3:
|
||||
dependencies:
|
||||
domelementtype: 2.3.0
|
||||
|
||||
domutils@3.2.2:
|
||||
dependencies:
|
||||
dom-serializer: 2.0.0
|
||||
domelementtype: 2.3.0
|
||||
domhandler: 5.0.3
|
||||
|
||||
dotenv@17.2.3: {}
|
||||
|
||||
drizzle-kit@0.31.8:
|
||||
@@ -5260,6 +5684,15 @@ snapshots:
|
||||
|
||||
emoji-regex@9.2.2: {}
|
||||
|
||||
encoding-sniffer@0.2.1:
|
||||
dependencies:
|
||||
iconv-lite: 0.6.3
|
||||
whatwg-encoding: 3.1.1
|
||||
|
||||
entities@4.5.0: {}
|
||||
|
||||
entities@6.0.1: {}
|
||||
|
||||
environment@1.1.0: {}
|
||||
|
||||
error-stack-parser-es@1.0.5: {}
|
||||
@@ -5452,6 +5885,12 @@ snapshots:
|
||||
|
||||
flatted@3.3.3: {}
|
||||
|
||||
follow-redirects@1.15.11: {}
|
||||
|
||||
for-each@0.3.5:
|
||||
dependencies:
|
||||
is-callable: 1.2.7
|
||||
|
||||
foreground-child@3.3.1:
|
||||
dependencies:
|
||||
cross-spawn: 7.0.6
|
||||
@@ -5488,6 +5927,8 @@ snapshots:
|
||||
- encoding
|
||||
- supports-color
|
||||
|
||||
generator-function@2.0.1: {}
|
||||
|
||||
gensync@1.0.0-beta.2: {}
|
||||
|
||||
get-east-asian-width@1.4.0: {}
|
||||
@@ -5569,6 +6010,10 @@ snapshots:
|
||||
|
||||
has-flag@4.0.0: {}
|
||||
|
||||
has-property-descriptors@1.0.2:
|
||||
dependencies:
|
||||
es-define-property: 1.0.1
|
||||
|
||||
has-symbols@1.1.0: {}
|
||||
|
||||
has-tostringtag@1.0.2:
|
||||
@@ -5583,6 +6028,13 @@ snapshots:
|
||||
|
||||
html-escaper@2.0.2: {}
|
||||
|
||||
htmlparser2@10.0.0:
|
||||
dependencies:
|
||||
domelementtype: 2.3.0
|
||||
domhandler: 5.0.3
|
||||
domutils: 3.2.2
|
||||
entities: 6.0.1
|
||||
|
||||
https-proxy-agent@7.0.6:
|
||||
dependencies:
|
||||
agent-base: 7.1.4
|
||||
@@ -5598,8 +6050,21 @@ snapshots:
|
||||
|
||||
husky@9.1.7: {}
|
||||
|
||||
iconv-lite@0.6.3:
|
||||
dependencies:
|
||||
safer-buffer: 2.1.2
|
||||
|
||||
inherits@2.0.4: {}
|
||||
|
||||
is-arguments@1.2.0:
|
||||
dependencies:
|
||||
call-bound: 1.0.4
|
||||
has-tostringtag: 1.0.2
|
||||
|
||||
is-arrayish@0.3.4: {}
|
||||
|
||||
is-callable@1.2.7: {}
|
||||
|
||||
is-extglob@2.1.1: {}
|
||||
|
||||
is-fullwidth-code-point@3.0.0: {}
|
||||
@@ -5610,16 +6075,35 @@ snapshots:
|
||||
dependencies:
|
||||
get-east-asian-width: 1.4.0
|
||||
|
||||
is-generator-function@1.1.2:
|
||||
dependencies:
|
||||
call-bound: 1.0.4
|
||||
generator-function: 2.0.1
|
||||
get-proto: 1.0.1
|
||||
has-tostringtag: 1.0.2
|
||||
safe-regex-test: 1.1.0
|
||||
|
||||
is-glob@4.0.3:
|
||||
dependencies:
|
||||
is-extglob: 2.1.1
|
||||
|
||||
is-number@7.0.0: {}
|
||||
|
||||
is-regex@1.2.1:
|
||||
dependencies:
|
||||
call-bound: 1.0.4
|
||||
gopd: 1.2.0
|
||||
has-tostringtag: 1.0.2
|
||||
hasown: 2.0.2
|
||||
|
||||
is-stream@2.0.1: {}
|
||||
|
||||
is-stream@3.0.0: {}
|
||||
|
||||
is-typed-array@1.1.15:
|
||||
dependencies:
|
||||
which-typed-array: 1.1.19
|
||||
|
||||
isexe@2.0.0: {}
|
||||
|
||||
istanbul-lib-coverage@3.2.2: {}
|
||||
@@ -5712,6 +6196,8 @@ snapshots:
|
||||
rfdc: 1.4.1
|
||||
wrap-ansi: 9.0.2
|
||||
|
||||
lodash.isequal@4.5.0: {}
|
||||
|
||||
lodash.mapkeys@4.6.0: {}
|
||||
|
||||
lodash@4.17.21: {}
|
||||
@@ -5834,6 +6320,10 @@ snapshots:
|
||||
dependencies:
|
||||
path-key: 4.0.0
|
||||
|
||||
nth-check@2.1.1:
|
||||
dependencies:
|
||||
boolbase: 1.0.0
|
||||
|
||||
onetime@6.0.0:
|
||||
dependencies:
|
||||
mimic-fn: 4.0.0
|
||||
@@ -5848,6 +6338,19 @@ snapshots:
|
||||
|
||||
package-json-from-dist@1.0.1: {}
|
||||
|
||||
parse5-htmlparser2-tree-adapter@7.1.0:
|
||||
dependencies:
|
||||
domhandler: 5.0.3
|
||||
parse5: 7.3.0
|
||||
|
||||
parse5-parser-stream@7.1.2:
|
||||
dependencies:
|
||||
parse5: 7.3.0
|
||||
|
||||
parse5@7.3.0:
|
||||
dependencies:
|
||||
entities: 6.0.1
|
||||
|
||||
path-browserify@1.0.1: {}
|
||||
|
||||
path-key@3.1.1: {}
|
||||
@@ -5873,6 +6376,8 @@ snapshots:
|
||||
|
||||
pidtree@0.6.0: {}
|
||||
|
||||
possible-typed-array-names@1.1.0: {}
|
||||
|
||||
postcss@8.5.6:
|
||||
dependencies:
|
||||
nanoid: 3.3.11
|
||||
@@ -5935,10 +6440,27 @@ snapshots:
|
||||
|
||||
safe-buffer@5.2.1: {}
|
||||
|
||||
safe-regex-test@1.1.0:
|
||||
dependencies:
|
||||
call-bound: 1.0.4
|
||||
es-errors: 1.3.0
|
||||
is-regex: 1.2.1
|
||||
|
||||
safer-buffer@2.1.2: {}
|
||||
|
||||
semver@6.3.1: {}
|
||||
|
||||
semver@7.7.3: {}
|
||||
|
||||
set-function-length@1.2.2:
|
||||
dependencies:
|
||||
define-data-property: 1.1.4
|
||||
es-errors: 1.3.0
|
||||
function-bind: 1.1.2
|
||||
get-intrinsic: 1.3.0
|
||||
gopd: 1.2.0
|
||||
has-property-descriptors: 1.0.2
|
||||
|
||||
sharp@0.33.5:
|
||||
dependencies:
|
||||
color: 4.2.3
|
||||
@@ -6115,6 +6637,8 @@ snapshots:
|
||||
|
||||
undici@7.14.0: {}
|
||||
|
||||
undici@7.16.0: {}
|
||||
|
||||
unenv@2.0.0-rc.24:
|
||||
dependencies:
|
||||
pathe: 2.0.3
|
||||
@@ -6125,6 +6649,14 @@ snapshots:
|
||||
escalade: 3.2.0
|
||||
picocolors: 1.1.1
|
||||
|
||||
util@0.12.5:
|
||||
dependencies:
|
||||
inherits: 2.0.4
|
||||
is-arguments: 1.2.0
|
||||
is-generator-function: 1.1.2
|
||||
is-typed-array: 1.1.15
|
||||
which-typed-array: 1.1.19
|
||||
|
||||
uuid@9.0.1: {}
|
||||
|
||||
vite-node@3.2.4(@types/node@25.0.1)(tsx@4.21.0)(yaml@2.8.2):
|
||||
@@ -6208,11 +6740,27 @@ snapshots:
|
||||
|
||||
webidl-conversions@3.0.1: {}
|
||||
|
||||
whatwg-encoding@3.1.1:
|
||||
dependencies:
|
||||
iconv-lite: 0.6.3
|
||||
|
||||
whatwg-mimetype@4.0.0: {}
|
||||
|
||||
whatwg-url@5.0.0:
|
||||
dependencies:
|
||||
tr46: 0.0.3
|
||||
webidl-conversions: 3.0.1
|
||||
|
||||
which-typed-array@1.1.19:
|
||||
dependencies:
|
||||
available-typed-arrays: 1.0.7
|
||||
call-bind: 1.0.8
|
||||
call-bound: 1.0.4
|
||||
for-each: 0.3.5
|
||||
get-proto: 1.0.1
|
||||
gopd: 1.2.0
|
||||
has-tostringtag: 1.0.2
|
||||
|
||||
which@2.0.2:
|
||||
dependencies:
|
||||
isexe: 2.0.0
|
||||
|
||||
5
src/consumet.ts
Normal file
5
src/consumet.ts
Normal file
@@ -0,0 +1,5 @@
|
||||
import { ANIME, META } from "@consumet/extensions";
|
||||
import fetchAdapter from "@haverstack/axios-fetch-adapter";
|
||||
|
||||
const gogoAnime = new ANIME.Gogoanime(undefined, undefined, fetchAdapter);
|
||||
export const aniList = new META.Anilist(gogoAnime, undefined, fetchAdapter);
|
||||
@@ -52,7 +52,7 @@ app.openapi(route, async (c) => {
|
||||
// Check if we should use mock data
|
||||
const { useMockData } = await import("~/libs/useMockData");
|
||||
if (useMockData()) {
|
||||
const { mockEpisodes } = await import("~/mocks");
|
||||
const { mockEpisodes } = await import("~/mocks/mockData");
|
||||
|
||||
return c.json({
|
||||
success: true,
|
||||
|
||||
64
src/controllers/episodes/getEpisodeUrl/anify.ts
Normal file
64
src/controllers/episodes/getEpisodeUrl/anify.ts
Normal file
@@ -0,0 +1,64 @@
|
||||
import { sortByProperty } from "~/libs/sortByProperty";
|
||||
import type { FetchUrlResponse } from "~/types/episode/fetch-url-response";
|
||||
|
||||
import { type SkipTime, convertSkipTime } from "./convertSkipTime";
|
||||
import {
|
||||
audioPriority,
|
||||
qualityPriority,
|
||||
subtitlesPriority,
|
||||
} from "./priorities";
|
||||
|
||||
export async function getSourcesFromAnify(
|
||||
provider: string,
|
||||
watchId: string,
|
||||
aniListId: number,
|
||||
): Promise<FetchUrlResponse | null> {
|
||||
const response = await fetch("https://anify.eltik.cc/sources", {
|
||||
body: JSON.stringify({
|
||||
watchId,
|
||||
providerId: provider,
|
||||
episodeNumber: "1",
|
||||
id: aniListId.toString(),
|
||||
subType: "sub",
|
||||
}),
|
||||
method: "POST",
|
||||
}).then((res) => res.json() as Promise<AnifySourcesResponse>);
|
||||
const { sources, subtitles, audio, intro, outro, headers } = response;
|
||||
|
||||
if (!sources || sources.length === 0) {
|
||||
return null;
|
||||
}
|
||||
|
||||
const source = sources.sort(sortByProperty(qualityPriority, "quality"))[0]
|
||||
?.url;
|
||||
subtitles?.sort(sortByProperty(subtitlesPriority, "lang"));
|
||||
audio?.sort(sortByProperty(audioPriority, "lang"));
|
||||
|
||||
return {
|
||||
source,
|
||||
audio,
|
||||
subtitles,
|
||||
intro: convertSkipTime(intro),
|
||||
outro: convertSkipTime(outro),
|
||||
headers: Object.keys(headers ?? {}).length > 0 ? headers : undefined,
|
||||
};
|
||||
}
|
||||
|
||||
interface AnifySourcesResponse {
|
||||
sources: VideoSource[];
|
||||
subtitles: LanguageSource[];
|
||||
audio: LanguageSource[];
|
||||
intro: SkipTime;
|
||||
outro: SkipTime;
|
||||
headers?: Record<string, string>;
|
||||
}
|
||||
|
||||
interface VideoSource {
|
||||
url: string;
|
||||
quality: string;
|
||||
}
|
||||
|
||||
interface LanguageSource {
|
||||
url: string;
|
||||
lang: string;
|
||||
}
|
||||
@@ -123,7 +123,7 @@ app.openapi(route, async (c) => {
|
||||
// Check if we should use mock data
|
||||
const { useMockData } = await import("~/libs/useMockData");
|
||||
if (useMockData()) {
|
||||
const { mockEpisodeUrl } = await import("~/mocks");
|
||||
const { mockEpisodeUrl } = await import("~/mocks/mockData");
|
||||
|
||||
return c.json({ success: true, result: mockEpisodeUrl });
|
||||
}
|
||||
|
||||
@@ -84,7 +84,7 @@ app.openapi(route, async (c) => {
|
||||
isComplete,
|
||||
);
|
||||
if (isComplete) {
|
||||
await updateWatchStatus(deviceId, aniListId, "COMPLETED");
|
||||
await updateWatchStatus(c.req, deviceId, aniListId, "COMPLETED");
|
||||
}
|
||||
|
||||
if (!user) {
|
||||
|
||||
@@ -15,7 +15,7 @@ type AiringSchedule = {
|
||||
id: number;
|
||||
};
|
||||
|
||||
export async function getUpcomingTitlesFromAnilist() {
|
||||
export async function getUpcomingTitlesFromAnilist(req: HonoRequest) {
|
||||
const durableObjectId = env.ANILIST_DO.idFromName("GLOBAL");
|
||||
const stub = env.ANILIST_DO.get(durableObjectId);
|
||||
|
||||
|
||||
@@ -9,8 +9,8 @@ import { getUpcomingTitlesFromAnilist } from "./anilist";
|
||||
|
||||
const app = new Hono();
|
||||
|
||||
export async function checkUpcomingTitles() {
|
||||
const titles = await getUpcomingTitlesFromAnilist();
|
||||
app.post("/", async (c) => {
|
||||
const titles = await getUpcomingTitlesFromAnilist(c.req);
|
||||
|
||||
await Promise.allSettled(
|
||||
titles.map(async (title) => {
|
||||
@@ -44,10 +44,6 @@ export async function checkUpcomingTitles() {
|
||||
});
|
||||
}),
|
||||
);
|
||||
}
|
||||
|
||||
app.post("/", async (c) => {
|
||||
await checkUpcomingTitles();
|
||||
|
||||
return c.json(SuccessResponse, 200);
|
||||
});
|
||||
|
||||
@@ -30,7 +30,7 @@ export async function fetchPopularTitlesFromAnilist(
|
||||
);
|
||||
break;
|
||||
case "upcoming":
|
||||
data = await stub.nextSeasonPopular(next.season, next.year, page, limit);
|
||||
data = await stub.nextSeasonPopular(next.season, next.year, limit);
|
||||
break;
|
||||
default:
|
||||
throw new Error(`Unknown category: ${category}`);
|
||||
|
||||
@@ -41,7 +41,7 @@ app.openapi(route, async (c) => {
|
||||
// Check if we should use mock data
|
||||
const { useMockData } = await import("~/libs/useMockData");
|
||||
if (useMockData()) {
|
||||
const { mockSearchResults } = await import("~/mocks");
|
||||
const { mockSearchResults } = await import("~/mocks/mockData");
|
||||
|
||||
// Paginate mock results
|
||||
const startIndex = (page - 1) * limit;
|
||||
|
||||
@@ -51,7 +51,7 @@ describe('requests the "/title" route', () => {
|
||||
headers: new Headers({ "x-anilist-token": "asd" }),
|
||||
});
|
||||
|
||||
await expect(response.json()).resolves.toMatchSnapshot();
|
||||
expect(await response.json()).toMatchSnapshot();
|
||||
expect(response.status).toBe(200);
|
||||
});
|
||||
|
||||
@@ -63,7 +63,7 @@ describe('requests the "/title" route', () => {
|
||||
|
||||
const response = await app.request("/title?id=10");
|
||||
|
||||
await expect(response.json()).resolves.toMatchSnapshot();
|
||||
expect(await response.json()).toMatchSnapshot();
|
||||
expect(response.status).toBe(200);
|
||||
});
|
||||
|
||||
@@ -75,7 +75,7 @@ describe('requests the "/title" route', () => {
|
||||
|
||||
const response = await app.request("/title?id=-1");
|
||||
|
||||
await expect(response.json()).resolves.toEqual({ success: false });
|
||||
expect(await response.json()).toEqual({ success: false });
|
||||
expect(response.status).toBe(404);
|
||||
});
|
||||
});
|
||||
|
||||
@@ -2,7 +2,6 @@ import { OpenAPIHono, createRoute, z } from "@hono/zod-openapi";
|
||||
|
||||
import { fetchTitleFromAnilist } from "~/libs/anilist/getTitle";
|
||||
import { fetchFromMultipleSources } from "~/libs/fetchFromMultipleSources";
|
||||
import { userProfileMiddleware } from "~/middleware/userProfile";
|
||||
import {
|
||||
AniListIdQuerySchema,
|
||||
ErrorResponse,
|
||||
@@ -10,7 +9,6 @@ import {
|
||||
SuccessResponseSchema,
|
||||
} from "~/types/schema";
|
||||
import { Title } from "~/types/title";
|
||||
import type { User } from "~/types/user";
|
||||
|
||||
const app = new OpenAPIHono();
|
||||
|
||||
@@ -42,7 +40,6 @@ const route = createRoute({
|
||||
description: "Title could not be found",
|
||||
},
|
||||
},
|
||||
middleware: [userProfileMiddleware],
|
||||
});
|
||||
|
||||
app.openapi(route, async (c) => {
|
||||
@@ -52,18 +49,13 @@ app.openapi(route, async (c) => {
|
||||
// Check if we should use mock data
|
||||
const { useMockData } = await import("~/libs/useMockData");
|
||||
if (useMockData()) {
|
||||
const { mockTitleDetails } = await import("~/mocks");
|
||||
const { mockTitleDetails } = await import("~/mocks/mockData");
|
||||
|
||||
return c.json({ success: true, result: mockTitleDetails() }, 200);
|
||||
}
|
||||
|
||||
const { result: title, errorOccurred } = await fetchFromMultipleSources([
|
||||
() =>
|
||||
fetchTitleFromAnilist(
|
||||
aniListId,
|
||||
(c.get("user") as User)?.id,
|
||||
aniListToken ?? undefined,
|
||||
),
|
||||
() => fetchTitleFromAnilist(aniListId, aniListToken ?? undefined),
|
||||
]);
|
||||
|
||||
if (errorOccurred) {
|
||||
|
||||
@@ -30,11 +30,19 @@ const DeleteMediaListEntryMutation = graphql(`
|
||||
}
|
||||
`);
|
||||
|
||||
export async function updateWatchStatusOnAnilist(
|
||||
/** Updates the watch status for a title on Anilist. If the token is null, the watch status will not be updated.
|
||||
*
|
||||
* @returns true if the watch status was updated or if the token was null, false if it was not
|
||||
*/
|
||||
export async function maybeUpdateWatchStatusOnAnilist(
|
||||
titleId: number,
|
||||
watchStatus: WatchStatus | null,
|
||||
aniListToken: string,
|
||||
aniListToken: string | undefined,
|
||||
) {
|
||||
if (!aniListToken) {
|
||||
return true;
|
||||
}
|
||||
|
||||
const client = new GraphQLClient("https://graphql.anilist.co/");
|
||||
const headers = new Headers({ Authorization: `Bearer ${aniListToken}` });
|
||||
|
||||
|
||||
@@ -22,6 +22,8 @@ vi.mock("~/mocks", () => ({
|
||||
describe("requests the /watch-status route", () => {
|
||||
const db = getTestDb(env);
|
||||
let app: typeof import("../../../src/index").app;
|
||||
let maybeUpdateWatchStatusOnAnilist: any;
|
||||
let queueTask: any;
|
||||
let maybeScheduleNextAiringEpisode: any;
|
||||
let removeTask: any;
|
||||
|
||||
@@ -29,6 +31,10 @@ describe("requests the /watch-status route", () => {
|
||||
await resetTestDb(db);
|
||||
vi.resetModules();
|
||||
|
||||
vi.doMock("./anilist", () => ({
|
||||
maybeUpdateWatchStatusOnAnilist: vi.fn().mockResolvedValue(undefined),
|
||||
}));
|
||||
|
||||
vi.doMock("~/libs/tasks/queueTask", () => ({
|
||||
queueTask: vi.fn().mockResolvedValue(undefined),
|
||||
}));
|
||||
@@ -46,6 +52,10 @@ describe("requests the /watch-status route", () => {
|
||||
}));
|
||||
|
||||
app = (await import("~/index")).app;
|
||||
maybeUpdateWatchStatusOnAnilist = (
|
||||
await import("~/controllers/watch-status/anilist")
|
||||
).maybeUpdateWatchStatusOnAnilist;
|
||||
queueTask = (await import("~/libs/tasks/queueTask")).queueTask;
|
||||
removeTask = (await import("~/libs/tasks/removeTask")).removeTask;
|
||||
maybeScheduleNextAiringEpisode = (
|
||||
await import("~/libs/maybeScheduleNextAiringEpisode")
|
||||
@@ -109,6 +119,34 @@ describe("requests the /watch-status route", () => {
|
||||
expect(res.status).toBe(500);
|
||||
});
|
||||
|
||||
it("saving title, Anilist request fails, should succeed", async () => {
|
||||
vi.mocked(maybeUpdateWatchStatusOnAnilist).mockRejectedValue(
|
||||
new Error("Anilist failed"),
|
||||
);
|
||||
|
||||
const res = await app.request(
|
||||
"/watch-status",
|
||||
{
|
||||
method: "POST",
|
||||
headers: new Headers({
|
||||
"x-anilist-token": "asd",
|
||||
"Content-Type": "application/json",
|
||||
}),
|
||||
body: JSON.stringify({
|
||||
deviceId: "123",
|
||||
watchStatus: "CURRENT",
|
||||
titleId: -1,
|
||||
}),
|
||||
},
|
||||
env,
|
||||
);
|
||||
|
||||
await expect(res.json()).resolves.toEqual({ success: true });
|
||||
expect(res.status).toBe(200);
|
||||
// Should queue task if direct update fails
|
||||
expect(queueTask).toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it("watch status is null, should succeed", async () => {
|
||||
const res = await app.request(
|
||||
"/watch-status",
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
import { OpenAPIHono, createRoute, z } from "@hono/zod-openapi";
|
||||
import type { HonoRequest } from "hono";
|
||||
|
||||
import { AnilistUpdateType } from "~/libs/anilist/updateType.ts";
|
||||
import { maybeScheduleNextAiringEpisode } from "~/libs/maybeScheduleNextAiringEpisode";
|
||||
@@ -15,12 +16,15 @@ import {
|
||||
} from "~/types/schema";
|
||||
import { WatchStatus } from "~/types/title/watchStatus";
|
||||
|
||||
import { maybeUpdateWatchStatusOnAnilist } from "./anilist";
|
||||
|
||||
const app = new OpenAPIHono<Cloudflare.Env>();
|
||||
|
||||
const UpdateWatchStatusRequest = z.object({
|
||||
deviceId: z.string(),
|
||||
watchStatus: WatchStatus.nullable(),
|
||||
titleId: AniListIdSchema,
|
||||
isRetrying: z.boolean().optional().default(false),
|
||||
});
|
||||
|
||||
const route = createRoute({
|
||||
@@ -62,6 +66,7 @@ const route = createRoute({
|
||||
});
|
||||
|
||||
export async function updateWatchStatus(
|
||||
req: HonoRequest,
|
||||
deviceId: string,
|
||||
titleId: number,
|
||||
watchStatus: WatchStatus | null,
|
||||
@@ -79,8 +84,14 @@ export async function updateWatchStatus(
|
||||
}
|
||||
|
||||
app.openapi(route, async (c) => {
|
||||
const { deviceId, watchStatus, titleId } =
|
||||
await c.req.json<typeof UpdateWatchStatusRequest._type>();
|
||||
const {
|
||||
deviceId,
|
||||
watchStatus,
|
||||
titleId,
|
||||
isRetrying = false,
|
||||
} = await c.req.json<typeof UpdateWatchStatusRequest._type>();
|
||||
const aniListToken = c.req.header("X-AniList-Token");
|
||||
|
||||
// Check if we should use mock data
|
||||
const { useMockData } = await import("~/libs/useMockData");
|
||||
if (useMockData()) {
|
||||
@@ -88,24 +99,35 @@ app.openapi(route, async (c) => {
|
||||
return c.json(SuccessResponse, { status: 200 });
|
||||
}
|
||||
|
||||
try {
|
||||
await updateWatchStatus(deviceId, titleId, watchStatus);
|
||||
} catch (error) {
|
||||
console.error("Error setting watch status");
|
||||
console.error(error);
|
||||
return c.json(ErrorResponse, { status: 500 });
|
||||
if (!isRetrying) {
|
||||
try {
|
||||
await updateWatchStatus(c.req, deviceId, titleId, watchStatus);
|
||||
} catch (error) {
|
||||
console.error("Error setting watch status");
|
||||
console.error(error);
|
||||
return c.json(ErrorResponse, { status: 500 });
|
||||
}
|
||||
}
|
||||
|
||||
const aniListToken = c.req.header("X-AniList-Token");
|
||||
if (aniListToken) {
|
||||
try {
|
||||
await maybeUpdateWatchStatusOnAnilist(
|
||||
Number(titleId),
|
||||
watchStatus,
|
||||
aniListToken,
|
||||
);
|
||||
} catch (error) {
|
||||
console.error("Failed to update watch status on Anilist");
|
||||
console.error(error);
|
||||
if (isRetrying) {
|
||||
return c.json(ErrorResponse, { status: 500 });
|
||||
}
|
||||
|
||||
await queueTask(
|
||||
"ANILIST_UPDATES",
|
||||
{
|
||||
[AnilistUpdateType.UpdateWatchStatus]: {
|
||||
aniListToken,
|
||||
titleId,
|
||||
watchStatus,
|
||||
},
|
||||
deviceId,
|
||||
watchStatus,
|
||||
titleId,
|
||||
updateType: AnilistUpdateType.UpdateWatchStatus,
|
||||
},
|
||||
{ req: c.req, scheduleConfig: { delay: { minute: 1 } } },
|
||||
|
||||
120
src/index.ts
120
src/index.ts
@@ -1,18 +1,11 @@
|
||||
import { swaggerUI } from "@hono/swagger-ui";
|
||||
import { OpenAPIHono } from "@hono/zod-openapi";
|
||||
import { Duration, type DurationLike } from "luxon";
|
||||
|
||||
import { onNewEpisode } from "~/controllers/internal/new-episode";
|
||||
import { AnilistUpdateType } from "~/libs/anilist/updateType";
|
||||
import { calculateExponentialBackoff } from "~/libs/calculateExponentialBackoff";
|
||||
import { maybeUpdateLastConnectedAt } from "~/controllers/maybeUpdateLastConnectedAt";
|
||||
import type { QueueName } from "~/libs/tasks/queueName.ts";
|
||||
import {
|
||||
MAX_QUEUE_DELAY_SECONDS,
|
||||
type QueueBody,
|
||||
} from "~/libs/tasks/queueTask";
|
||||
import { maybeUpdateLastConnectedAt } from "~/middleware/maybeUpdateLastConnectedAt";
|
||||
|
||||
import { checkUpcomingTitles } from "./controllers/internal/upcoming-titles";
|
||||
import { onNewEpisode } from "./controllers/internal/new-episode";
|
||||
import type { QueueBody } from "./libs/tasks/queueTask";
|
||||
|
||||
export const app = new OpenAPIHono<{ Bindings: Env }>();
|
||||
|
||||
@@ -79,101 +72,28 @@ app.get("/docs", swaggerUI({ url: "/openapi.json" }));
|
||||
export default {
|
||||
fetch: app.fetch,
|
||||
async queue(batch) {
|
||||
onMessageQueue(batch, async (message, queueName) => {
|
||||
switch (queueName) {
|
||||
case "ANILIST_UPDATES":
|
||||
const anilistUpdateBody =
|
||||
message.body as QueueBody["ANILIST_UPDATES"];
|
||||
console.log("queue run", message.body);
|
||||
switch (anilistUpdateBody.updateType) {
|
||||
case AnilistUpdateType.UpdateWatchStatus:
|
||||
if (!anilistUpdateBody[AnilistUpdateType.UpdateWatchStatus]) {
|
||||
console.error(
|
||||
`Discarding update, unknown body ${JSON.stringify(message.body)}`,
|
||||
);
|
||||
return;
|
||||
}
|
||||
|
||||
const { updateWatchStatusOnAnilist } =
|
||||
await import("~/controllers/watch-status/anilist");
|
||||
const payload =
|
||||
anilistUpdateBody[AnilistUpdateType.UpdateWatchStatus];
|
||||
await updateWatchStatusOnAnilist(
|
||||
payload.titleId,
|
||||
payload.watchStatus,
|
||||
payload.aniListToken,
|
||||
);
|
||||
break;
|
||||
default:
|
||||
throw new Error(
|
||||
`Unhandled update type: ${anilistUpdateBody.updateType}`,
|
||||
);
|
||||
}
|
||||
break;
|
||||
case "NEW_EPISODE":
|
||||
const newEpisodeBody = message.body as QueueBody["NEW_EPISODE"];
|
||||
switch (batch.queue as QueueName) {
|
||||
case "ANILIST_UPDATES":
|
||||
batch.retryAll();
|
||||
break;
|
||||
case "NEW_EPISODE":
|
||||
for (const message of (batch as MessageBatch<QueueBody["NEW_EPISODE"]>)
|
||||
.messages) {
|
||||
await onNewEpisode(
|
||||
newEpisodeBody.aniListId,
|
||||
newEpisodeBody.episodeNumber,
|
||||
message.body.aniListId,
|
||||
message.body.episodeNumber,
|
||||
);
|
||||
break;
|
||||
default:
|
||||
throw new Error(`Unhandled queue name: ${queueName}`);
|
||||
}
|
||||
});
|
||||
message.ack();
|
||||
}
|
||||
break;
|
||||
}
|
||||
},
|
||||
async scheduled(event, env, ctx) {
|
||||
switch (event.cron) {
|
||||
case "0 */12 * * *":
|
||||
const { processDelayedTasks } =
|
||||
await import("~/libs/tasks/processDelayedTasks");
|
||||
await processDelayedTasks(env);
|
||||
break;
|
||||
case "0 18 * * *":
|
||||
const { checkUpcomingTitles } =
|
||||
await import("~/controllers/internal/upcoming-titles");
|
||||
await checkUpcomingTitles();
|
||||
break;
|
||||
default:
|
||||
throw new Error(`Unhandled cron: ${event.cron}`);
|
||||
}
|
||||
const { processDelayedTasks } = await import(
|
||||
"~/libs/tasks/processDelayedTasks"
|
||||
);
|
||||
await processDelayedTasks(env, ctx);
|
||||
},
|
||||
} satisfies ExportedHandler<Env>;
|
||||
|
||||
const retryDelayConfig: Partial<
|
||||
Record<QueueName, { min: DurationLike; max: DurationLike }>
|
||||
> = {
|
||||
NEW_EPISODE: {
|
||||
min: Duration.fromObject({ hours: 1 }),
|
||||
max: Duration.fromObject({ hours: 12 }),
|
||||
},
|
||||
};
|
||||
|
||||
function onMessageQueue<QN extends QueueName>(
|
||||
messageBatch: MessageBatch<unknown>,
|
||||
callback: (message: Message<QueueBody[QN]>, queueName: QN) => void,
|
||||
) {
|
||||
for (const message of messageBatch.messages) {
|
||||
try {
|
||||
callback(message as Message<QueueBody[QN]>, messageBatch.queue as QN);
|
||||
message.ack();
|
||||
} catch (error) {
|
||||
console.error(
|
||||
`Failed to process message ${message.id} for queue ${messageBatch.queue} with body ${JSON.stringify(message.body)}`,
|
||||
);
|
||||
console.error(error);
|
||||
message.retry({
|
||||
delaySeconds: Math.min(
|
||||
calculateExponentialBackoff({
|
||||
attempt: message.attempts,
|
||||
baseMin: retryDelayConfig[messageBatch.queue as QN]?.min,
|
||||
absCap: retryDelayConfig[messageBatch.queue as QN]?.max,
|
||||
}),
|
||||
MAX_QUEUE_DELAY_SECONDS,
|
||||
),
|
||||
});
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
export { AnilistDurableObject as AnilistDo } from "~/libs/anilist/anilist-do.ts";
|
||||
|
||||
@@ -1,7 +1,6 @@
|
||||
import type { TypedDocumentNode } from "@graphql-typed-document-node/core";
|
||||
import { DurableObject } from "cloudflare:workers";
|
||||
import { print } from "graphql";
|
||||
import { DateTime } from "luxon";
|
||||
import { z } from "zod";
|
||||
|
||||
import {
|
||||
@@ -9,7 +8,6 @@ import {
|
||||
GetNextEpisodeAiringAtQuery,
|
||||
GetPopularTitlesQuery,
|
||||
GetTitleQuery,
|
||||
GetTitleUserDataQuery,
|
||||
GetTrendingTitlesQuery,
|
||||
GetUpcomingTitlesQuery,
|
||||
GetUserProfileQuery,
|
||||
@@ -20,7 +18,6 @@ import {
|
||||
SearchQuery,
|
||||
} from "~/libs/anilist/queries";
|
||||
import { sleep } from "~/libs/sleep.ts";
|
||||
import type { Title } from "~/types/title";
|
||||
|
||||
const nextAiringEpisodeSchema = z.nullable(
|
||||
z.object({
|
||||
@@ -41,54 +38,30 @@ export class AnilistDurableObject extends DurableObject {
|
||||
return new Response("Not found", { status: 404 });
|
||||
}
|
||||
|
||||
async getTitle(
|
||||
id: number,
|
||||
userId?: number,
|
||||
token?: string,
|
||||
): Promise<Title | null> {
|
||||
const promises: Promise<any>[] = [
|
||||
this.handleCachedRequest(
|
||||
`title:${id}`,
|
||||
async () => {
|
||||
const anilistResponse = await this.fetchFromAnilist(GetTitleQuery, {
|
||||
id,
|
||||
});
|
||||
return anilistResponse?.Media ?? null;
|
||||
},
|
||||
(media) => {
|
||||
if (!media) return undefined;
|
||||
|
||||
// Cast to any to access fragment fields without unmasking
|
||||
const nextAiringEpisode = nextAiringEpisodeSchema.parse(
|
||||
(media as any)?.nextAiringEpisode,
|
||||
);
|
||||
return nextAiringEpisode?.airingAt
|
||||
? DateTime.fromMillis(nextAiringEpisode?.airingAt)
|
||||
: undefined;
|
||||
},
|
||||
),
|
||||
];
|
||||
promises.push(
|
||||
userId
|
||||
? this.handleCachedRequest(
|
||||
`title:${id}:${userId}`,
|
||||
async () => {
|
||||
const anilistResponse = await this.fetchFromAnilist(
|
||||
GetTitleUserDataQuery,
|
||||
{ id },
|
||||
{ token },
|
||||
);
|
||||
return anilistResponse?.Media ?? null;
|
||||
},
|
||||
DateTime.now().plus({ days: 1 }),
|
||||
)
|
||||
: Promise.resolve({ mediaListEntry: null }),
|
||||
async getTitle(id: number, token?: string) {
|
||||
return this.handleCachedRequest(
|
||||
`title:${id}`,
|
||||
async () => {
|
||||
const anilistResponse = await this.fetchFromAnilist(
|
||||
GetTitleQuery,
|
||||
{ id },
|
||||
token,
|
||||
);
|
||||
return anilistResponse?.Media ?? null;
|
||||
},
|
||||
(media) => {
|
||||
if (!media) return undefined;
|
||||
// Cast to any to access fragment fields without unmasking
|
||||
const nextAiringEpisode = nextAiringEpisodeSchema.parse(
|
||||
(media as any)?.nextAiringEpisode,
|
||||
);
|
||||
const airingAt = (nextAiringEpisode?.airingAt ?? 0) * 1000;
|
||||
if (airingAt) {
|
||||
return airingAt - Date.now();
|
||||
}
|
||||
return undefined;
|
||||
},
|
||||
);
|
||||
|
||||
return Promise.all(promises).then(([title, userTitle]) => ({
|
||||
...title,
|
||||
...userTitle,
|
||||
}));
|
||||
}
|
||||
|
||||
async getNextEpisodeAiringAt(id: number) {
|
||||
@@ -100,7 +73,7 @@ export class AnilistDurableObject extends DurableObject {
|
||||
});
|
||||
return data?.Media;
|
||||
},
|
||||
DateTime.now().plus({ hours: 1 }),
|
||||
60 * 60 * 1000,
|
||||
);
|
||||
}
|
||||
|
||||
@@ -115,7 +88,7 @@ export class AnilistDurableObject extends DurableObject {
|
||||
});
|
||||
return data?.Page;
|
||||
},
|
||||
DateTime.now().plus({ hours: 1 }),
|
||||
60 * 60 * 1000,
|
||||
);
|
||||
}
|
||||
|
||||
@@ -128,7 +101,8 @@ export class AnilistDurableObject extends DurableObject {
|
||||
) {
|
||||
return this.handleCachedRequest(
|
||||
`popular:${JSON.stringify({ season, seasonYear, nextSeason, nextYear, limit })}`,
|
||||
() => {
|
||||
async () => {
|
||||
console.log(nextSeason, nextYear, print(BrowsePopularQuery));
|
||||
return this.fetchFromAnilist(BrowsePopularQuery, {
|
||||
season,
|
||||
seasonYear,
|
||||
@@ -137,27 +111,21 @@ export class AnilistDurableObject extends DurableObject {
|
||||
limit,
|
||||
});
|
||||
},
|
||||
DateTime.now().plus({ days: 1 }),
|
||||
24 * 60 * 60 * 1000,
|
||||
);
|
||||
}
|
||||
|
||||
async nextSeasonPopular(
|
||||
nextSeason: any,
|
||||
nextYear: number,
|
||||
page: number,
|
||||
limit: number,
|
||||
) {
|
||||
async nextSeasonPopular(nextSeason: any, nextYear: number, limit: number) {
|
||||
return this.handleCachedRequest(
|
||||
`next_season:${JSON.stringify({ nextSeason, nextYear, page, limit })}`,
|
||||
`next_season:${JSON.stringify({ nextSeason, nextYear, limit })}`,
|
||||
async () => {
|
||||
return this.fetchFromAnilist(NextSeasonPopularQuery, {
|
||||
nextSeason,
|
||||
nextYear,
|
||||
limit,
|
||||
page,
|
||||
}).then((data) => data?.Page);
|
||||
});
|
||||
},
|
||||
DateTime.now().plus({ days: 1 }),
|
||||
24 * 60 * 60 * 1000,
|
||||
);
|
||||
}
|
||||
|
||||
@@ -170,14 +138,15 @@ export class AnilistDurableObject extends DurableObject {
|
||||
return this.handleCachedRequest(
|
||||
`popular:${JSON.stringify({ page, limit, season, seasonYear })}`,
|
||||
async () => {
|
||||
return this.fetchFromAnilist(GetPopularTitlesQuery, {
|
||||
const data = await this.fetchFromAnilist(GetPopularTitlesQuery, {
|
||||
page,
|
||||
limit,
|
||||
season,
|
||||
seasonYear,
|
||||
}).then((data) => data?.Page);
|
||||
});
|
||||
return data?.Page;
|
||||
},
|
||||
DateTime.now().plus({ days: 1 }),
|
||||
24 * 60 * 60 * 1000,
|
||||
);
|
||||
}
|
||||
|
||||
@@ -191,7 +160,7 @@ export class AnilistDurableObject extends DurableObject {
|
||||
});
|
||||
return data?.Page;
|
||||
},
|
||||
DateTime.now().plus({ days: 1 }),
|
||||
24 * 60 * 60 * 1000,
|
||||
);
|
||||
}
|
||||
|
||||
@@ -210,7 +179,7 @@ export class AnilistDurableObject extends DurableObject {
|
||||
});
|
||||
return data?.Page;
|
||||
},
|
||||
DateTime.now().plus({ days: 1 }),
|
||||
24 * 60 * 60 * 1000,
|
||||
);
|
||||
}
|
||||
|
||||
@@ -218,10 +187,10 @@ export class AnilistDurableObject extends DurableObject {
|
||||
return this.handleCachedRequest(
|
||||
`user:${token}`,
|
||||
async () => {
|
||||
const data = await this.fetchFromAnilist(GetUserQuery, {}, { token });
|
||||
const data = await this.fetchFromAnilist(GetUserQuery, {}, token);
|
||||
return data?.Viewer;
|
||||
},
|
||||
DateTime.now().plus({ days: 30 }),
|
||||
60 * 60 * 24 * 30 * 1000,
|
||||
);
|
||||
}
|
||||
|
||||
@@ -232,11 +201,11 @@ export class AnilistDurableObject extends DurableObject {
|
||||
const data = await this.fetchFromAnilist(
|
||||
GetUserProfileQuery,
|
||||
{ token },
|
||||
{ token },
|
||||
token,
|
||||
);
|
||||
return data?.Viewer;
|
||||
},
|
||||
DateTime.now().plus({ days: 30 }),
|
||||
60 * 60 * 24 * 30 * 1000,
|
||||
);
|
||||
}
|
||||
|
||||
@@ -248,7 +217,7 @@ export class AnilistDurableObject extends DurableObject {
|
||||
const data = await this.fetchFromAnilist(
|
||||
MarkEpisodeAsWatchedMutation,
|
||||
{ titleId, episodeNumber },
|
||||
{ token },
|
||||
token,
|
||||
);
|
||||
return data?.SaveMediaListEntry;
|
||||
}
|
||||
@@ -257,7 +226,7 @@ export class AnilistDurableObject extends DurableObject {
|
||||
const data = await this.fetchFromAnilist(
|
||||
MarkTitleAsWatchedMutation,
|
||||
{ titleId },
|
||||
{ token },
|
||||
token,
|
||||
);
|
||||
return data?.SaveMediaListEntry;
|
||||
}
|
||||
@@ -266,7 +235,7 @@ export class AnilistDurableObject extends DurableObject {
|
||||
async handleCachedRequest<T>(
|
||||
key: string,
|
||||
fetcher: () => Promise<T>,
|
||||
ttl?: DateTime | ((data: T) => DateTime | undefined),
|
||||
ttl?: number | ((data: T) => number | undefined),
|
||||
) {
|
||||
const cache = await this.state.storage.get(key);
|
||||
console.debug(`Retrieving request ${key} from cache:`, cache != null);
|
||||
@@ -278,8 +247,9 @@ export class AnilistDurableObject extends DurableObject {
|
||||
await this.state.storage.put(key, result);
|
||||
|
||||
const calculatedTtl = typeof ttl === "function" ? ttl(result) : ttl;
|
||||
if (calculatedTtl) {
|
||||
const alarmTime = calculatedTtl.toMillis();
|
||||
|
||||
if (calculatedTtl && calculatedTtl > 0) {
|
||||
const alarmTime = Date.now() + calculatedTtl;
|
||||
await this.state.storage.setAlarm(alarmTime);
|
||||
await this.state.storage.put(`alarm:${key}`, alarmTime);
|
||||
}
|
||||
@@ -290,13 +260,11 @@ export class AnilistDurableObject extends DurableObject {
|
||||
async alarm() {
|
||||
const now = Date.now();
|
||||
const alarms = await this.state.storage.list({ prefix: "alarm:" });
|
||||
console.debug(`Retrieved alarms from cache:`, Object.entries(alarms));
|
||||
for (const [key, ttl] of Object.entries(alarms)) {
|
||||
if (now >= ttl) {
|
||||
// The key in alarms is `alarm:${storageKey}`
|
||||
// We want to delete the storageKey
|
||||
const storageKey = key.replace("alarm:", "");
|
||||
console.debug(`Deleting storage key ${storageKey} & alarm ${key}`);
|
||||
await this.state.storage.delete(storageKey);
|
||||
await this.state.storage.delete(key);
|
||||
}
|
||||
@@ -304,13 +272,10 @@ export class AnilistDurableObject extends DurableObject {
|
||||
}
|
||||
|
||||
async fetchFromAnilist<Result = any, Variables = any>(
|
||||
query: TypedDocumentNode<Result, Variables>,
|
||||
queryString: string,
|
||||
variables: Variables,
|
||||
{
|
||||
token,
|
||||
shouldRetryOnRateLimit = true,
|
||||
}: { token?: string | undefined; shouldRetryOnRateLimit?: boolean } = {},
|
||||
): Promise<Result | undefined> {
|
||||
token?: string | undefined,
|
||||
): Promise<Result> {
|
||||
const headers: any = {
|
||||
"Content-Type": "application/json",
|
||||
};
|
||||
@@ -321,7 +286,7 @@ export class AnilistDurableObject extends DurableObject {
|
||||
|
||||
// Use the query passed in, or fallback if needed (though we expect it to be passed)
|
||||
// We print the query to string
|
||||
const queryString = print(query);
|
||||
// const queryString = print(query);
|
||||
|
||||
const response = await fetch(`${this.env.PROXY_URL}/proxy`, {
|
||||
method: "POST",
|
||||
@@ -340,17 +305,14 @@ export class AnilistDurableObject extends DurableObject {
|
||||
});
|
||||
|
||||
// 1. Handle Rate Limiting (429)
|
||||
if (shouldRetryOnRateLimit && response.status === 429) {
|
||||
if (response.status === 429) {
|
||||
const retryAfter = await response
|
||||
.json<{ headers: Record<string, string> }>()
|
||||
.json()
|
||||
.then(({ headers }) => new Headers(headers).get("Retry-After"));
|
||||
console.log("429, retrying in", retryAfter);
|
||||
|
||||
await sleep(Number(retryAfter || 1) * 1000); // specific fallback or ensure logic
|
||||
return this.fetchFromAnilist(query, variables, {
|
||||
token,
|
||||
shouldRetryOnRateLimit: false,
|
||||
});
|
||||
return this.fetchFromAnilist(query, variables, token);
|
||||
}
|
||||
|
||||
// 2. Handle HTTP Errors (like 404 or 500)
|
||||
|
||||
@@ -5,11 +5,10 @@ import type { Title } from "~/types/title";
|
||||
|
||||
export async function fetchTitleFromAnilist(
|
||||
id: number,
|
||||
userId?: number | undefined,
|
||||
token?: string | undefined,
|
||||
): Promise<Title | undefined> {
|
||||
if (useMockData()) {
|
||||
const { mockTitleDetails } = await import("~/mocks");
|
||||
const { mockTitleDetails } = await import("~/mocks/mockData");
|
||||
return mockTitleDetails();
|
||||
}
|
||||
|
||||
@@ -18,7 +17,8 @@ export async function fetchTitleFromAnilist(
|
||||
);
|
||||
const stub = env.ANILIST_DO.get(durableObjectId);
|
||||
|
||||
const data = await stub.getTitle(id, userId, token);
|
||||
const data = await stub.getTitle(id, token);
|
||||
|
||||
if (!data) {
|
||||
return undefined;
|
||||
}
|
||||
|
||||
@@ -14,18 +14,6 @@ export const GetTitleQuery = graphql(
|
||||
[MediaFragment],
|
||||
);
|
||||
|
||||
export const GetTitleUserDataQuery = graphql(`
|
||||
query GetTitleUserData($id: Int!) {
|
||||
Media(id: $id) {
|
||||
mediaListEntry {
|
||||
id
|
||||
progress
|
||||
status
|
||||
}
|
||||
}
|
||||
}
|
||||
`);
|
||||
|
||||
export const SearchQuery = graphql(
|
||||
`
|
||||
query Search($query: String!, $page: Int!, $limit: Int!) {
|
||||
@@ -259,9 +247,8 @@ export const NextSeasonPopularQuery = graphql(
|
||||
$nextSeason: MediaSeason
|
||||
$nextYear: Int
|
||||
$limit: Int!
|
||||
$page: Int!
|
||||
) {
|
||||
Page(page: $page, perPage: $limit) {
|
||||
Page(page: 1, perPage: $limit) {
|
||||
media(
|
||||
season: $nextSeason
|
||||
seasonYear: $nextYear
|
||||
|
||||
@@ -1,53 +0,0 @@
|
||||
import { Duration, type DurationLike } from "luxon";
|
||||
|
||||
interface CalculateExponentialBackoffOptions {
|
||||
attempt: number;
|
||||
baseMin?: DurationLike;
|
||||
absCap?: DurationLike;
|
||||
fuzzFactor?: number;
|
||||
}
|
||||
|
||||
/**
|
||||
* Generates a backoff time where both the Minimum floor and Maximum ceiling
|
||||
* are "fuzzed" with jitter to prevent clustering at the edges.
|
||||
*
|
||||
* @param attempt - The current retry attempt (0-indexed).
|
||||
* @param baseMin - The nominal minimum wait time (default: 1s).
|
||||
* @param absCap - The absolute maximum wait time (default: 60s).
|
||||
* @param fuzzFactor - How much to wobble the edges (0.1 = +/- 10%).
|
||||
*
|
||||
* @returns A random duration between the nominal minimum and maximum, in seconds.
|
||||
*/
|
||||
export function calculateExponentialBackoff({
|
||||
attempt,
|
||||
baseMin: baseMinDuration = Duration.fromObject({ minutes: 1 }),
|
||||
absCap: absCapDuration = Duration.fromObject({ hours: 1 }),
|
||||
fuzzFactor = 0.2,
|
||||
}: CalculateExponentialBackoffOptions): number {
|
||||
const baseMin = Duration.fromDurationLike(baseMinDuration).as("seconds");
|
||||
const absCap = Duration.fromDurationLike(absCapDuration).as("seconds");
|
||||
|
||||
// 1. Calculate nominal boundaries
|
||||
// Example: If baseMin is 1s, the nominal boundaries are 1s, 2s, 4s, 8s... (The 'ceiling' grows exponentially)
|
||||
const nominalMin = baseMin;
|
||||
const nominalCeiling = Math.min(baseMin * Math.pow(2, attempt), absCap);
|
||||
|
||||
// 2. Fuzz the Min (The Floor)
|
||||
// Example: If min is 1s and fuzz is 0.2, the floor becomes random between 0.8s and 1.2s
|
||||
const minFuzz = nominalMin * fuzzFactor;
|
||||
const fuzzedMin = nominalMin + (Math.random() * 2 * minFuzz - minFuzz);
|
||||
|
||||
// 3. Fuzz the Max (The Ceiling)
|
||||
// Example: If ceiling is 4s (and fuzz is 0.2), it becomes random between 3.2s and 4.8s
|
||||
const maxFuzz = nominalCeiling * fuzzFactor;
|
||||
const fuzzedCeiling =
|
||||
nominalCeiling + (Math.random() * 2 * maxFuzz - maxFuzz);
|
||||
|
||||
// Safety: Ensure we don't return a negative number or cross boundaries weirdly
|
||||
// (e.g. if fuzz makes min > max, we swap or clamp)
|
||||
const safeMin = Math.max(0, fuzzedMin);
|
||||
const safeMax = Math.max(safeMin, fuzzedCeiling);
|
||||
|
||||
// 4. Return random value in the new fuzzy range
|
||||
return safeMin + Math.random() * (safeMax - safeMin);
|
||||
}
|
||||
@@ -3,13 +3,11 @@ import mapKeys from "lodash.mapkeys";
|
||||
|
||||
import { Case, changeStringCase } from "../changeStringCase";
|
||||
|
||||
export function getAdminSdkCredentials(
|
||||
env: Cloudflare.Env = cloudflareEnv,
|
||||
): AdminSdkCredentials {
|
||||
export function getAdminSdkCredentials(env: Cloudflare.Env = cloudflareEnv) {
|
||||
return mapKeys(
|
||||
JSON.parse(env.ADMIN_SDK_JSON) as AdminSdkCredentials,
|
||||
(_, key) => changeStringCase(key, Case.snake_case, Case.camelCase),
|
||||
) satisfies AdminSdkCredentials;
|
||||
);
|
||||
}
|
||||
|
||||
export interface AdminSdkCredentials {
|
||||
|
||||
@@ -1,158 +1,204 @@
|
||||
import { env } from "cloudflare:test";
|
||||
import { DateTime } from "luxon";
|
||||
import { beforeEach, describe, expect, it, vi } from "vitest";
|
||||
|
||||
import { getTestEnv } from "../test/getTestEnv";
|
||||
import { processDelayedTasks } from "./processDelayedTasks";
|
||||
|
||||
describe("processDelayedTasks", () => {
|
||||
beforeEach(async () => {
|
||||
const tasksToDelete = await env.DELAYED_TASKS.list({
|
||||
prefix: "delayed-task:",
|
||||
});
|
||||
console.log(`Found ${tasksToDelete.keys.length} tasks to delete`);
|
||||
for (const task of tasksToDelete.keys) {
|
||||
await env.DELAYED_TASKS.delete(task.name);
|
||||
}
|
||||
let mockEnv: Cloudflare.Env;
|
||||
let mockCtx: ExecutionContext;
|
||||
let kvGetSpy: ReturnType<typeof vi.fn>;
|
||||
let kvDeleteSpy: ReturnType<typeof vi.fn>;
|
||||
let kvPutSpy: ReturnType<typeof vi.fn>;
|
||||
let queueSendSpy: ReturnType<typeof vi.fn>;
|
||||
|
||||
beforeEach(() => {
|
||||
kvGetSpy = vi.fn(() => Promise.resolve(null));
|
||||
kvDeleteSpy = vi.fn(() => Promise.resolve());
|
||||
kvPutSpy = vi.fn(() => Promise.resolve());
|
||||
queueSendSpy = vi.fn(() => Promise.resolve());
|
||||
|
||||
mockEnv = {
|
||||
DELAYED_TASKS: {
|
||||
get: kvGetSpy,
|
||||
delete: kvDeleteSpy,
|
||||
put: kvPutSpy,
|
||||
list: vi.fn(() =>
|
||||
Promise.resolve({
|
||||
keys: [],
|
||||
list_complete: true as const,
|
||||
cacheStatus: null,
|
||||
}),
|
||||
),
|
||||
getWithMetadata: vi.fn(() =>
|
||||
Promise.resolve({ value: null, metadata: null }),
|
||||
),
|
||||
} as any,
|
||||
NEW_EPISODE: {
|
||||
send: queueSendSpy,
|
||||
} as any,
|
||||
ANILIST_UPDATES: {
|
||||
send: vi.fn(() => Promise.resolve()),
|
||||
} as any,
|
||||
} as any;
|
||||
|
||||
mockCtx = {
|
||||
waitUntil: vi.fn(() => {}),
|
||||
passThroughOnException: vi.fn(() => {}),
|
||||
} as any;
|
||||
});
|
||||
|
||||
it("handles empty KV namespace", async () => {
|
||||
await processDelayedTasks(env);
|
||||
await processDelayedTasks(mockEnv, mockCtx);
|
||||
|
||||
await expect(
|
||||
env.DELAYED_TASKS.list({ prefix: "delayed-task:" }).then(
|
||||
(result) => result.keys,
|
||||
),
|
||||
).resolves.toHaveLength(0);
|
||||
expect(kvDeleteSpy).not.toHaveBeenCalled();
|
||||
expect(queueSendSpy).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it("queues tasks within 9 hours of scheduled time", async () => {
|
||||
const now = DateTime.now();
|
||||
const scheduledTime = now.plus({ hours: 6 }).toSeconds();
|
||||
it("queues tasks within 12 hours of scheduled time", async () => {
|
||||
const now = Math.floor(Date.now() / 1000);
|
||||
const scheduledTime = now + 6 * 3600; // 6 hours from now
|
||||
|
||||
const taskMetadata = {
|
||||
queueName: "NEW_EPISODE",
|
||||
body: { aniListId: 123, episodeNumber: 1 },
|
||||
headers: { "Content-Type": "application/json" },
|
||||
scheduledEpochTime: scheduledTime,
|
||||
taskId: "task-1",
|
||||
createdAt: now.minus({ hours: 18 }).toSeconds(),
|
||||
createdAt: now - 18 * 3600,
|
||||
retryCount: 0,
|
||||
};
|
||||
await env.DELAYED_TASKS.put(
|
||||
`delayed-task:${scheduledTime}:task-1`,
|
||||
JSON.stringify(taskMetadata),
|
||||
|
||||
mockEnv.DELAYED_TASKS.list = vi.fn(() =>
|
||||
Promise.resolve({
|
||||
keys: [{ name: `delayed-task:${scheduledTime}:task-1` }],
|
||||
list_complete: true as const,
|
||||
cacheStatus: null,
|
||||
}),
|
||||
);
|
||||
|
||||
await processDelayedTasks(env);
|
||||
kvGetSpy.mockReturnValue(Promise.resolve(JSON.stringify(taskMetadata)));
|
||||
|
||||
await expect(
|
||||
env.DELAYED_TASKS.get(`delayed-task:${scheduledTime}:task-1`),
|
||||
).resolves.toBeNull();
|
||||
await processDelayedTasks(mockEnv, mockCtx);
|
||||
|
||||
expect(queueSendSpy).toHaveBeenCalledTimes(1);
|
||||
expect(kvDeleteSpy).toHaveBeenCalledTimes(1);
|
||||
expect(kvDeleteSpy).toHaveBeenCalledWith(
|
||||
`delayed-task:${scheduledTime}:task-1`,
|
||||
);
|
||||
});
|
||||
|
||||
it("does not queue tasks beyond 9 hours", async () => {
|
||||
const now = DateTime.now();
|
||||
const scheduledTime = now.plus({ hours: 24 }).toSeconds();
|
||||
it("does not queue tasks beyond 12 hours", async () => {
|
||||
const now = Math.floor(Date.now() / 1000);
|
||||
const scheduledTime = now + 24 * 3600; // 24 hours from now
|
||||
|
||||
const taskMetadata = {
|
||||
queueName: "NEW_EPISODE",
|
||||
body: { aniListId: 456, episodeNumber: 2 },
|
||||
headers: { "Content-Type": "application/json" },
|
||||
scheduledEpochTime: scheduledTime,
|
||||
taskId: "task-2",
|
||||
createdAt: now.toSeconds(),
|
||||
createdAt: now,
|
||||
retryCount: 0,
|
||||
};
|
||||
await env.DELAYED_TASKS.put(
|
||||
`delayed-task:${scheduledTime}:task-2`,
|
||||
JSON.stringify(taskMetadata),
|
||||
|
||||
mockEnv.DELAYED_TASKS.list = vi.fn(() =>
|
||||
Promise.resolve({
|
||||
keys: [{ name: `delayed-task:${scheduledTime}:task-2` }],
|
||||
list_complete: true as const,
|
||||
cacheStatus: null,
|
||||
}),
|
||||
);
|
||||
|
||||
await processDelayedTasks(env);
|
||||
kvGetSpy.mockReturnValue(Promise.resolve(JSON.stringify(taskMetadata)));
|
||||
|
||||
await expect(
|
||||
env.DELAYED_TASKS.get(`delayed-task:${scheduledTime}:task-2`),
|
||||
).resolves.toBeTruthy();
|
||||
await processDelayedTasks(mockEnv, mockCtx);
|
||||
|
||||
expect(queueSendSpy).not.toHaveBeenCalled();
|
||||
expect(kvDeleteSpy).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it("increments retry count on queue failure", async () => {
|
||||
const now = DateTime.now();
|
||||
const scheduledTime = now.plus({ hours: 1 }).toSeconds();
|
||||
const now = Math.floor(Date.now() / 1000);
|
||||
const scheduledTime = now + 1 * 3600; // 1 hour from now
|
||||
|
||||
const taskMetadata = {
|
||||
queueName: "NEW_EPISODE",
|
||||
body: { aniListId: 789, episodeNumber: 3 },
|
||||
headers: { "Content-Type": "application/json" },
|
||||
scheduledEpochTime: scheduledTime,
|
||||
taskId: "task-3",
|
||||
createdAt: now.minus({ hours: 23 }).toSeconds(),
|
||||
createdAt: now - 23 * 3600,
|
||||
retryCount: 0,
|
||||
};
|
||||
const mockEnv = getTestEnv({
|
||||
NEW_EPISODE: {
|
||||
send: vi.fn().mockRejectedValue(new Error("Queue error")),
|
||||
sendBatch: vi.fn().mockRejectedValue(new Error("Queue error")),
|
||||
},
|
||||
});
|
||||
await mockEnv.DELAYED_TASKS.put(
|
||||
`delayed-task:${scheduledTime}:task-3`,
|
||||
JSON.stringify(taskMetadata),
|
||||
|
||||
mockEnv.DELAYED_TASKS.list = vi.fn(() =>
|
||||
Promise.resolve({
|
||||
keys: [{ name: `delayed-task:${scheduledTime}:task-3` }],
|
||||
list_complete: true as const,
|
||||
cacheStatus: null,
|
||||
}),
|
||||
);
|
||||
|
||||
await processDelayedTasks(mockEnv);
|
||||
kvGetSpy.mockReturnValue(Promise.resolve(JSON.stringify(taskMetadata)));
|
||||
queueSendSpy.mockRejectedValue(new Error("Queue error"));
|
||||
|
||||
const updatedMetadata = JSON.parse(
|
||||
(await mockEnv.DELAYED_TASKS.get(
|
||||
`delayed-task:${scheduledTime}:task-3`,
|
||||
))!,
|
||||
);
|
||||
await processDelayedTasks(mockEnv, mockCtx);
|
||||
|
||||
expect(kvPutSpy).toHaveBeenCalledTimes(1);
|
||||
const updatedMetadata = JSON.parse(kvPutSpy.mock.calls[0][1]);
|
||||
expect(updatedMetadata.retryCount).toBe(1);
|
||||
expect(kvDeleteSpy).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it("logs alert after 3 failed attempts", async () => {
|
||||
const consoleErrorSpy = vi.fn(() => {});
|
||||
const originalConsoleError = console.error;
|
||||
console.error = consoleErrorSpy as any;
|
||||
const now = DateTime.now();
|
||||
const scheduledTime = now.plus({ hours: 1 }).toSeconds();
|
||||
|
||||
const now = Math.floor(Date.now() / 1000);
|
||||
const scheduledTime = now + 1 * 3600;
|
||||
|
||||
const taskMetadata = {
|
||||
queueName: "NEW_EPISODE",
|
||||
body: { aniListId: 789, episodeNumber: 4 },
|
||||
body: { aniListId: 999, episodeNumber: 4 },
|
||||
headers: { "Content-Type": "application/json" },
|
||||
scheduledEpochTime: scheduledTime,
|
||||
taskId: "task-4",
|
||||
createdAt: now.minus({ hours: 23 }).toSeconds(),
|
||||
retryCount: 2,
|
||||
createdAt: now - 23 * 3600,
|
||||
retryCount: 2, // Will become 3 after this failure
|
||||
};
|
||||
const mockEnv = getTestEnv({
|
||||
NEW_EPISODE: {
|
||||
send: vi.fn().mockRejectedValue(new Error("Queue error")),
|
||||
sendBatch: vi.fn().mockRejectedValue(new Error("Queue error")),
|
||||
},
|
||||
});
|
||||
await mockEnv.DELAYED_TASKS.put(
|
||||
`delayed-task:${scheduledTime}:task-4`,
|
||||
JSON.stringify(taskMetadata),
|
||||
|
||||
mockEnv.DELAYED_TASKS.list = vi.fn(() =>
|
||||
Promise.resolve({
|
||||
keys: [{ name: `delayed-task:${scheduledTime}:task-4` }],
|
||||
list_complete: true as const,
|
||||
cacheStatus: null,
|
||||
}),
|
||||
);
|
||||
|
||||
await processDelayedTasks(mockEnv);
|
||||
kvGetSpy.mockReturnValue(Promise.resolve(JSON.stringify(taskMetadata)));
|
||||
queueSendSpy.mockRejectedValue(new Error("Queue error"));
|
||||
|
||||
await processDelayedTasks(mockEnv, mockCtx);
|
||||
|
||||
// Check that alert was logged
|
||||
const alertCalls = consoleErrorSpy.mock.calls.filter((call: any) =>
|
||||
call[0]?.includes("🚨 ALERT"),
|
||||
);
|
||||
expect(alertCalls.length).toBeGreaterThan(0);
|
||||
|
||||
console.error = originalConsoleError;
|
||||
});
|
||||
|
||||
it("handles multiple tasks in single cron run", async () => {
|
||||
const now = DateTime.now();
|
||||
const now = Math.floor(Date.now() / 1000);
|
||||
|
||||
const task1Metadata = {
|
||||
queueName: "NEW_EPISODE",
|
||||
body: { aniListId: 100, episodeNumber: 1 },
|
||||
headers: { "Content-Type": "application/json" },
|
||||
scheduledEpochTime: now.plus({ hours: 2 }).toSeconds(),
|
||||
scheduledEpochTime: now + 2 * 3600,
|
||||
taskId: "task-1",
|
||||
createdAt: now.minus({ hours: 20 }).toSeconds(),
|
||||
createdAt: now - 20 * 3600,
|
||||
retryCount: 0,
|
||||
};
|
||||
|
||||
@@ -160,53 +206,47 @@ describe("processDelayedTasks", () => {
|
||||
queueName: "NEW_EPISODE",
|
||||
body: { aniListId: 200, episodeNumber: 2 },
|
||||
headers: { "Content-Type": "application/json" },
|
||||
scheduledEpochTime: now.plus({ hours: 5 }).toSeconds(),
|
||||
scheduledEpochTime: now + 5 * 3600,
|
||||
taskId: "task-2",
|
||||
createdAt: now.minus({ hours: 19 }).toSeconds(),
|
||||
createdAt: now - 19 * 3600,
|
||||
retryCount: 0,
|
||||
};
|
||||
await env.DELAYED_TASKS.put(
|
||||
`delayed-task:${task1Metadata.scheduledEpochTime}:task-1`,
|
||||
JSON.stringify(task1Metadata),
|
||||
);
|
||||
await env.DELAYED_TASKS.put(
|
||||
`delayed-task:${task2Metadata.scheduledEpochTime}:task-2`,
|
||||
JSON.stringify(task2Metadata),
|
||||
|
||||
mockEnv.DELAYED_TASKS.list = vi.fn(() =>
|
||||
Promise.resolve({
|
||||
keys: [
|
||||
{ name: `delayed-task:${task1Metadata.scheduledEpochTime}:task-1` },
|
||||
{ name: `delayed-task:${task2Metadata.scheduledEpochTime}:task-2` },
|
||||
],
|
||||
list_complete: true as const,
|
||||
cacheStatus: null,
|
||||
}),
|
||||
);
|
||||
|
||||
await processDelayedTasks(env);
|
||||
kvGetSpy
|
||||
.mockReturnValueOnce(Promise.resolve(JSON.stringify(task1Metadata)))
|
||||
.mockReturnValueOnce(Promise.resolve(JSON.stringify(task2Metadata)));
|
||||
|
||||
await expect(
|
||||
env.DELAYED_TASKS.get(
|
||||
`delayed-task:${task1Metadata.scheduledEpochTime}:task-1`,
|
||||
),
|
||||
).resolves.toBeNull();
|
||||
await expect(
|
||||
env.DELAYED_TASKS.get(
|
||||
`delayed-task:${task2Metadata.scheduledEpochTime}:task-2`,
|
||||
),
|
||||
).resolves.toBeNull();
|
||||
await processDelayedTasks(mockEnv, mockCtx);
|
||||
|
||||
expect(queueSendSpy).toHaveBeenCalledTimes(2);
|
||||
expect(kvDeleteSpy).toHaveBeenCalledTimes(2);
|
||||
});
|
||||
|
||||
it("skips tasks with null values in KV", async () => {
|
||||
const queueSendSpy = vi.fn().mockResolvedValue(undefined);
|
||||
const mockEnv = getTestEnv({
|
||||
NEW_EPISODE: {
|
||||
send: queueSendSpy,
|
||||
sendBatch: queueSendSpy,
|
||||
},
|
||||
ANILIST_UPDATES: {
|
||||
send: queueSendSpy,
|
||||
sendBatch: queueSendSpy,
|
||||
},
|
||||
});
|
||||
await mockEnv.DELAYED_TASKS.put(`delayed-task:123:invalid`, null);
|
||||
mockEnv.DELAYED_TASKS.list = vi.fn(() =>
|
||||
Promise.resolve({
|
||||
keys: [{ name: "delayed-task:123:invalid" }],
|
||||
list_complete: true as const,
|
||||
cacheStatus: null,
|
||||
}),
|
||||
);
|
||||
|
||||
await processDelayedTasks(mockEnv);
|
||||
kvGetSpy.mockReturnValue(Promise.resolve(null));
|
||||
|
||||
await processDelayedTasks(mockEnv, mockCtx);
|
||||
|
||||
expect(queueSendSpy).not.toHaveBeenCalled();
|
||||
await expect(
|
||||
mockEnv.DELAYED_TASKS.get(`delayed-task:123:invalid`),
|
||||
).resolves.toBeNull();
|
||||
expect(kvDeleteSpy).not.toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
|
||||
@@ -2,11 +2,15 @@ import { DateTime } from "luxon";
|
||||
|
||||
import type { DelayedTaskMetadata } from "./delayedTask";
|
||||
import { deserializeDelayedTask } from "./delayedTask";
|
||||
import { MAX_QUEUE_DELAY_SECONDS, queueTask } from "./queueTask";
|
||||
import { queueTask } from "./queueTask";
|
||||
|
||||
const MAX_DELAY_SECONDS = 12 * 60 * 60; // 43,200 seconds (12 hours)
|
||||
const RETRY_ALERT_THRESHOLD = 3;
|
||||
|
||||
export async function processDelayedTasks(env: Cloudflare.Env): Promise<void> {
|
||||
export async function processDelayedTasks(
|
||||
env: Cloudflare.Env,
|
||||
ctx: ExecutionContext,
|
||||
): Promise<void> {
|
||||
console.log("Starting delayed task processing cron job");
|
||||
|
||||
const kvNamespace = env.DELAYED_TASKS;
|
||||
@@ -27,7 +31,7 @@ export async function processDelayedTasks(env: Cloudflare.Env): Promise<void> {
|
||||
console.log(`Found ${keys.length} delayed tasks to check`);
|
||||
|
||||
const currentTime = Math.floor(Date.now() / 1000);
|
||||
const maxQueueTime = currentTime + MAX_QUEUE_DELAY_SECONDS;
|
||||
const twelveHoursFromNow = currentTime + MAX_DELAY_SECONDS;
|
||||
|
||||
let processedCount = 0;
|
||||
let queuedCount = 0;
|
||||
@@ -36,17 +40,16 @@ export async function processDelayedTasks(env: Cloudflare.Env): Promise<void> {
|
||||
for (const key of keys) {
|
||||
try {
|
||||
const value = await kvNamespace.get(key.name);
|
||||
if (!value || value == "null") {
|
||||
console.warn(`Task key ${key.name} has no value, removing`);
|
||||
await kvNamespace.delete(key.name);
|
||||
if (!value) {
|
||||
console.warn(`Task key ${key.name} has no value, skipping`);
|
||||
continue;
|
||||
}
|
||||
|
||||
const metadata: DelayedTaskMetadata = deserializeDelayedTask(value);
|
||||
processedCount++;
|
||||
|
||||
// Check if task is ready to be queued (within 9 hours of scheduled time)
|
||||
if (metadata.scheduledEpochTime <= maxQueueTime) {
|
||||
// Check if task is ready to be queued (within 12 hours of scheduled time)
|
||||
if (metadata.scheduledEpochTime <= twelveHoursFromNow) {
|
||||
const remainingDelay = Math.max(
|
||||
0,
|
||||
metadata.scheduledEpochTime - currentTime,
|
||||
@@ -97,7 +100,7 @@ export async function processDelayedTasks(env: Cloudflare.Env): Promise<void> {
|
||||
}
|
||||
} else {
|
||||
const hoursUntilReady =
|
||||
(metadata.scheduledEpochTime - maxQueueTime) / 3600;
|
||||
(metadata.scheduledEpochTime - twelveHoursFromNow) / 3600;
|
||||
console.log(
|
||||
`Task ${metadata.taskId} not ready yet (${hoursUntilReady.toFixed(1)} hours until queueable)`,
|
||||
);
|
||||
|
||||
@@ -35,8 +35,8 @@ describe("queueTask - delayed task handling", () => {
|
||||
(globalThis as any).crypto = { randomUUID: vi.fn(() => "test-uuid-123") };
|
||||
});
|
||||
|
||||
describe("tasks with delay <= 9 hours", () => {
|
||||
it("queues task directly when delay is less than 9 hours", async () => {
|
||||
describe("tasks with delay <= 12 hours", () => {
|
||||
it("queues task directly when delay is less than 12 hours", async () => {
|
||||
await queueTask(
|
||||
"NEW_EPISODE",
|
||||
{ aniListId: 123, episodeNumber: 1 },
|
||||
@@ -52,12 +52,12 @@ describe("queueTask - delayed task handling", () => {
|
||||
expect(kvPutSpy).not.toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it("queues task directly when delay is exactly 9 hours", async () => {
|
||||
it("queues task directly when delay is exactly 12 hours", async () => {
|
||||
await queueTask(
|
||||
"NEW_EPISODE",
|
||||
{ aniListId: 456, episodeNumber: 2 },
|
||||
{
|
||||
scheduleConfig: { delay: { hours: 9 } },
|
||||
scheduleConfig: { delay: { hours: 12 } },
|
||||
env: mockEnv,
|
||||
},
|
||||
);
|
||||
|
||||
@@ -9,11 +9,9 @@ import type { QueueName } from "./queueName";
|
||||
|
||||
export type QueueBody = {
|
||||
ANILIST_UPDATES: {
|
||||
[AnilistUpdateType.UpdateWatchStatus]: {
|
||||
titleId: number;
|
||||
watchStatus: WatchStatus | null;
|
||||
aniListToken: string;
|
||||
};
|
||||
deviceId: string;
|
||||
watchStatus: WatchStatus | null;
|
||||
titleId: number;
|
||||
updateType: AnilistUpdateType;
|
||||
};
|
||||
NEW_EPISODE: { aniListId: number; episodeNumber: number };
|
||||
@@ -30,10 +28,6 @@ interface QueueTaskOptionalArgs {
|
||||
env?: Cloudflare.Env;
|
||||
}
|
||||
|
||||
export const MAX_QUEUE_DELAY_SECONDS = Duration.fromObject({ hours: 12 }).as(
|
||||
"seconds",
|
||||
);
|
||||
|
||||
export async function queueTask(
|
||||
queueName: QueueName,
|
||||
body: QueueBody[QueueName],
|
||||
@@ -46,14 +40,17 @@ export async function queueTask(
|
||||
req?.header(),
|
||||
);
|
||||
|
||||
const MAX_DELAY_SECONDS = 12 * 60 * 60; // 43,200 seconds (12 hours)
|
||||
|
||||
// If delay exceeds 12 hours, store in KV for later processing
|
||||
if (scheduleTime > MAX_QUEUE_DELAY_SECONDS) {
|
||||
if (scheduleTime > MAX_DELAY_SECONDS) {
|
||||
if (!env || !env.DELAYED_TASKS) {
|
||||
throw new Error("DELAYED_TASKS KV namespace not available");
|
||||
}
|
||||
|
||||
const { generateTaskKey, serializeDelayedTask } =
|
||||
await import("./delayedTask");
|
||||
const { generateTaskKey, serializeDelayedTask } = await import(
|
||||
"./delayedTask"
|
||||
);
|
||||
const taskId = crypto.randomUUID();
|
||||
const scheduledEpochTime = Math.floor(Date.now() / 1000) + scheduleTime;
|
||||
|
||||
@@ -132,9 +129,6 @@ function buildTask(
|
||||
scheduleTime = Duration.fromDurationLike(delay).as("second");
|
||||
}
|
||||
}
|
||||
const authorizationHeader = headers?.["X-Anilist-Token"]
|
||||
? { Authorization: `Bearer ${headers["X-Anilist-Token"]}` }
|
||||
: {};
|
||||
|
||||
switch (queueName) {
|
||||
case "ANILIST_UPDATES":
|
||||
@@ -143,8 +137,8 @@ function buildTask(
|
||||
body,
|
||||
scheduleTime,
|
||||
headers: {
|
||||
...authorizationHeader,
|
||||
"Content-Type": "application/json",
|
||||
"X-Anilist-Token": headers?.["X-Anilist-Token"],
|
||||
},
|
||||
};
|
||||
default:
|
||||
|
||||
@@ -8,12 +8,10 @@ export function getTestEnvVariables(): Cloudflare.Env {
|
||||
export function getTestEnv({
|
||||
ADMIN_SDK_JSON = '{"client_email": "test@test.com", "project_id": "test-26g38"}',
|
||||
LOG_DB_QUERIES = "false",
|
||||
...mockEnv
|
||||
}: Partial<Cloudflare.Env> = {}): Cloudflare.Env {
|
||||
return {
|
||||
...env,
|
||||
ADMIN_SDK_JSON,
|
||||
LOG_DB_QUERIES,
|
||||
...mockEnv,
|
||||
};
|
||||
}
|
||||
|
||||
@@ -1,25 +0,0 @@
|
||||
import { createMiddleware } from "hono/factory";
|
||||
|
||||
import type { User } from "~/types/user";
|
||||
|
||||
export const userProfileMiddleware = createMiddleware<
|
||||
Cloudflare.Env & {
|
||||
Variables: {
|
||||
user: User;
|
||||
};
|
||||
Bindings: Env;
|
||||
}
|
||||
>(async (c, next) => {
|
||||
const aniListToken = await c.req.header("X-AniList-Token");
|
||||
if (!aniListToken) {
|
||||
return next();
|
||||
}
|
||||
|
||||
const user = await c.env.ANILIST_DO.getByName("GLOBAL").getUser(aniListToken);
|
||||
if (!user) {
|
||||
return c.json({ error: "User not found" }, 401);
|
||||
}
|
||||
|
||||
c.set("user", user);
|
||||
return next();
|
||||
});
|
||||
47
src/scripts/generateEnv.ts
Normal file
47
src/scripts/generateEnv.ts
Normal file
@@ -0,0 +1,47 @@
|
||||
import { Project } from "ts-morph";
|
||||
import { $ } from "zx";
|
||||
|
||||
import { logStep } from "~/libs/logStep";
|
||||
|
||||
await logStep(
|
||||
'Re-generating "env.d.ts"',
|
||||
() => $`wrangler types src/types/env.d.ts`.quiet(),
|
||||
"Generated env.d.ts",
|
||||
);
|
||||
|
||||
const secretNames = await logStep(
|
||||
"Fetching secrets from Cloudflare",
|
||||
async (): Promise<string[]> => {
|
||||
const { stdout } = await $`wrangler secret list`.quiet();
|
||||
return JSON.parse(stdout.toString()).map(
|
||||
(secret: { name: string; type: "secret_text" }) => secret.name,
|
||||
);
|
||||
},
|
||||
"Fetched secrets",
|
||||
);
|
||||
|
||||
const project = new Project({});
|
||||
|
||||
const envSourceFile = project.addSourceFileAtPath("src/types/env.d.ts");
|
||||
envSourceFile.insertImportDeclaration(2, {
|
||||
isTypeOnly: true,
|
||||
moduleSpecifier: "hono",
|
||||
namedImports: ["Env as HonoEnv"],
|
||||
});
|
||||
envSourceFile
|
||||
.getInterfaceOrThrow("Env")
|
||||
.addExtends(["HonoEnv", "Record<string, unknown>"]);
|
||||
envSourceFile.getInterfaceOrThrow("Env").addProperties(
|
||||
secretNames.map((name) => ({
|
||||
name,
|
||||
type: `string`,
|
||||
})),
|
||||
);
|
||||
|
||||
await project.save();
|
||||
|
||||
await logStep(
|
||||
"Formatting env.d.ts",
|
||||
() => $`prettier --write src/types/env.d.ts`.quiet(),
|
||||
"Formatted env.d.ts",
|
||||
);
|
||||
29
src/scripts/ipCheck.ts
Normal file
29
src/scripts/ipCheck.ts
Normal file
@@ -0,0 +1,29 @@
|
||||
// import { GraphQLClient } from "graphql-request";
|
||||
import { HttpsProxyAgent } from "https-proxy-agent";
|
||||
import nodeFetch from "node-fetch";
|
||||
|
||||
// import { GetTitleQuery } from "../libs/anilist/getTitle.ts";
|
||||
|
||||
const agent = new HttpsProxyAgent(
|
||||
"http://ruru:pdh!CQB@kpc3vyb3cwc@45.56.108.251:3128",
|
||||
);
|
||||
const response = await nodeFetch("https://httpbin.org/ip", { agent });
|
||||
console.log(await response.text());
|
||||
console.log(response.status);
|
||||
console.log(nodeFetch);
|
||||
|
||||
// const client = new GraphQLClient("https://graphql.anilist.co/", {
|
||||
// fetch: (input, init) => {
|
||||
// console.log("custom fetch");
|
||||
// const agent = new HttpsProxyAgent(
|
||||
// "http://ruru:pdh!CQB@kpc3vyb3cwc@45.56.108.251:3128",
|
||||
// );
|
||||
// return nodeFetch(input, { ...init, agent });
|
||||
// },
|
||||
// });
|
||||
|
||||
// console.log(
|
||||
// await client
|
||||
// .request(GetTitleQuery, { id: 186794 })
|
||||
// .then((data) => data?.Media ?? undefined),
|
||||
// );
|
||||
40
src/scripts/verifyEnv.ts
Normal file
40
src/scripts/verifyEnv.ts
Normal file
@@ -0,0 +1,40 @@
|
||||
import { readFile } from "fs/promises";
|
||||
import { $, sleep } from "zx";
|
||||
|
||||
import { logStep } from "~/libs/logStep";
|
||||
|
||||
await $`cp src/types/env.d.ts /tmp/env.d.ts`.quiet();
|
||||
|
||||
await logStep(
|
||||
'Generating "env.d.ts"',
|
||||
// @ts-ignore
|
||||
() => import("./generateEnv"),
|
||||
"Generated env.d.ts",
|
||||
);
|
||||
|
||||
await logStep("Comparing env.d.ts", async () => {
|
||||
function filterComments(content: Buffer) {
|
||||
return content
|
||||
.toString()
|
||||
.split("\n")
|
||||
.filter((line) => !line.trim().startsWith("//"))
|
||||
.join("\n");
|
||||
}
|
||||
|
||||
const currentFileContent = filterComments(await readFile("/tmp/env.d.ts"));
|
||||
const generatedFileContent = filterComments(
|
||||
await readFile("src/types/env.d.ts"),
|
||||
);
|
||||
|
||||
if (currentFileContent === generatedFileContent) {
|
||||
console.log("env.d.ts is up to date");
|
||||
return;
|
||||
}
|
||||
|
||||
const isCI = process.env["IS_CI"] === "true";
|
||||
const vcsCommand = isCI ? "git" : "sl";
|
||||
await $`${vcsCommand} diff src/types/env.d.ts`.stdio("inherit");
|
||||
// add 1 second to make sure spawn completes
|
||||
await sleep(1000);
|
||||
throw new Error("env.d.ts is out of date");
|
||||
});
|
||||
@@ -21,6 +21,11 @@ export const MediaFragment = graphql(`
|
||||
medium
|
||||
}
|
||||
countryOfOrigin
|
||||
mediaListEntry {
|
||||
id
|
||||
progress
|
||||
status
|
||||
}
|
||||
nextAiringEpisode {
|
||||
timeUntilAiring
|
||||
airingAt
|
||||
|
||||
@@ -3,24 +3,20 @@ import { z } from "zod";
|
||||
export type User = z.infer<typeof User>;
|
||||
export const User = z
|
||||
.object({
|
||||
statistics: z.object({
|
||||
minutesWatched: z.number().openapi({ type: "integer", format: "int64" }),
|
||||
episodesWatched: z.number().openapi({ type: "integer", format: "int64" }),
|
||||
count: z
|
||||
.number()
|
||||
.int() /* .openapi({ type: "integer", format: "int64" }) */,
|
||||
meanScore: z.number().openapi({ type: "number", format: "float" }),
|
||||
}),
|
||||
id: z.number().openapi({ type: "integer", format: "int64" }),
|
||||
name: z.string(),
|
||||
avatar: z.object({
|
||||
medium: z.string(),
|
||||
large: z.string(),
|
||||
}),
|
||||
})
|
||||
.optional()
|
||||
.nullable();
|
||||
|
||||
export type UserProfile = z.infer<typeof UserProfile>;
|
||||
export const UserProfile = z.object({
|
||||
statistics: z.object({
|
||||
minutesWatched: z.number().openapi({ type: "integer", format: "int64" }),
|
||||
episodesWatched: z.number().openapi({ type: "integer", format: "int64" }),
|
||||
count: z.number().int(),
|
||||
meanScore: z.number().openapi({ type: "number", format: "float" }),
|
||||
}),
|
||||
id: z.number().openapi({ type: "integer", format: "int64" }),
|
||||
name: z.string(),
|
||||
avatar: z.object({
|
||||
medium: z.string(),
|
||||
large: z.string(),
|
||||
}),
|
||||
});
|
||||
|
||||
145
worker-configuration.d.ts
vendored
145
worker-configuration.d.ts
vendored
@@ -2,32 +2,32 @@
|
||||
// Generated by Wrangler by running `wrangler types` (hash: df24977940a31745cb42d562b6645de2)
|
||||
// Runtime types generated with workerd@1.20251210.0 2025-11-28 nodejs_compat
|
||||
declare namespace Cloudflare {
|
||||
interface GlobalProps {
|
||||
mainModule: typeof import("./src/index");
|
||||
durableNamespaces: "AnilistDo";
|
||||
}
|
||||
interface Env {
|
||||
DELAYED_TASKS: KVNamespace;
|
||||
ADMIN_SDK_JSON: string;
|
||||
CLOUDFLARE_TOKEN: string;
|
||||
CLOUDFLARE_D1_TOKEN: string;
|
||||
CLOUDFLARE_ACCOUNT_ID: string;
|
||||
CLOUDFLARE_DATABASE_ID: string;
|
||||
PROXY_URL: string;
|
||||
USE_MOCK_DATA: string;
|
||||
LOG_DB_QUERIES: string;
|
||||
ANILIST_DO: DurableObjectNamespace<import("./src/index").AnilistDo>;
|
||||
DB: D1Database;
|
||||
ANILIST_UPDATES: Queue;
|
||||
NEW_EPISODE: Queue;
|
||||
}
|
||||
interface GlobalProps {
|
||||
mainModule: typeof import("./src/index");
|
||||
durableNamespaces: "AnilistDo";
|
||||
}
|
||||
interface Env {
|
||||
DELAYED_TASKS: KVNamespace;
|
||||
ADMIN_SDK_JSON: string;
|
||||
CLOUDFLARE_TOKEN: string;
|
||||
CLOUDFLARE_D1_TOKEN: string;
|
||||
CLOUDFLARE_ACCOUNT_ID: string;
|
||||
CLOUDFLARE_DATABASE_ID: string;
|
||||
PROXY_URL: string;
|
||||
USE_MOCK_DATA: string;
|
||||
LOG_DB_QUERIES: string;
|
||||
ANILIST_DO: DurableObjectNamespace<import("./src/index").AnilistDo>;
|
||||
DB: D1Database;
|
||||
ANILIST_UPDATES: Queue;
|
||||
NEW_EPISODE: Queue;
|
||||
}
|
||||
}
|
||||
interface Env extends Cloudflare.Env { }
|
||||
interface Env extends Cloudflare.Env {}
|
||||
type StringifyValues<EnvType extends Record<string, unknown>> = {
|
||||
[Binding in keyof EnvType]: EnvType[Binding] extends string ? EnvType[Binding] : string;
|
||||
[Binding in keyof EnvType]: EnvType[Binding] extends string ? EnvType[Binding] : string;
|
||||
};
|
||||
declare namespace NodeJS {
|
||||
interface ProcessEnv extends StringifyValues<Pick<Cloudflare.Env, "ADMIN_SDK_JSON" | "CLOUDFLARE_TOKEN" | "CLOUDFLARE_D1_TOKEN" | "CLOUDFLARE_ACCOUNT_ID" | "CLOUDFLARE_DATABASE_ID" | "PROXY_URL" | "USE_MOCK_DATA" | "LOG_DB_QUERIES">> { }
|
||||
interface ProcessEnv extends StringifyValues<Pick<Cloudflare.Env, "ADMIN_SDK_JSON" | "CLOUDFLARE_TOKEN" | "CLOUDFLARE_D1_TOKEN" | "CLOUDFLARE_ACCOUNT_ID" | "CLOUDFLARE_DATABASE_ID" | "PROXY_URL" | "USE_MOCK_DATA" | "LOG_DB_QUERIES">> {}
|
||||
}
|
||||
|
||||
// Begin runtime types
|
||||
@@ -1644,7 +1644,7 @@ declare abstract class Body {
|
||||
*/
|
||||
declare var Response: {
|
||||
prototype: Response;
|
||||
new(body?: BodyInit | null, init?: ResponseInit): Response;
|
||||
new (body?: BodyInit | null, init?: ResponseInit): Response;
|
||||
error(): Response;
|
||||
redirect(url: string, status?: number): Response;
|
||||
json(any: any, maybeInit?: (ResponseInit | Response)): Response;
|
||||
@@ -2192,7 +2192,7 @@ interface ReadableStream<R = any> {
|
||||
*/
|
||||
declare const ReadableStream: {
|
||||
prototype: ReadableStream;
|
||||
new(underlyingSource: UnderlyingByteSource, strategy?: QueuingStrategy<Uint8Array>): ReadableStream<Uint8Array>;
|
||||
new (underlyingSource: UnderlyingByteSource, strategy?: QueuingStrategy<Uint8Array>): ReadableStream<Uint8Array>;
|
||||
new <R = any>(underlyingSource?: UnderlyingSource<R>, strategy?: QueuingStrategy<R>): ReadableStream<R>;
|
||||
};
|
||||
/**
|
||||
@@ -3034,7 +3034,7 @@ type WebSocketEventMap = {
|
||||
*/
|
||||
declare var WebSocket: {
|
||||
prototype: WebSocket;
|
||||
new(url: string, protocols?: (string[] | string)): WebSocket;
|
||||
new (url: string, protocols?: (string[] | string)): WebSocket;
|
||||
readonly READY_STATE_CONNECTING: number;
|
||||
readonly CONNECTING: number;
|
||||
readonly READY_STATE_OPEN: number;
|
||||
@@ -3091,7 +3091,7 @@ interface WebSocket extends EventTarget<WebSocketEventMap> {
|
||||
extensions: string | null;
|
||||
}
|
||||
declare const WebSocketPair: {
|
||||
new(): {
|
||||
new (): {
|
||||
0: WebSocket;
|
||||
1: WebSocket;
|
||||
};
|
||||
@@ -9413,21 +9413,21 @@ interface IncomingRequestCfPropertiesTLSClientAuthPlaceholder {
|
||||
certNotAfter: "";
|
||||
}
|
||||
/** Possible outcomes of TLS verification */
|
||||
declare type CertVerificationStatus =
|
||||
/** Authentication succeeded */
|
||||
"SUCCESS"
|
||||
/** No certificate was presented */
|
||||
| "NONE"
|
||||
/** Failed because the certificate was self-signed */
|
||||
| "FAILED:self signed certificate"
|
||||
/** Failed because the certificate failed a trust chain check */
|
||||
| "FAILED:unable to verify the first certificate"
|
||||
/** Failed because the certificate not yet valid */
|
||||
| "FAILED:certificate is not yet valid"
|
||||
/** Failed because the certificate is expired */
|
||||
| "FAILED:certificate has expired"
|
||||
/** Failed for another unspecified reason */
|
||||
| "FAILED";
|
||||
declare type CertVerificationStatus =
|
||||
/** Authentication succeeded */
|
||||
"SUCCESS"
|
||||
/** No certificate was presented */
|
||||
| "NONE"
|
||||
/** Failed because the certificate was self-signed */
|
||||
| "FAILED:self signed certificate"
|
||||
/** Failed because the certificate failed a trust chain check */
|
||||
| "FAILED:unable to verify the first certificate"
|
||||
/** Failed because the certificate not yet valid */
|
||||
| "FAILED:certificate is not yet valid"
|
||||
/** Failed because the certificate is expired */
|
||||
| "FAILED:certificate has expired"
|
||||
/** Failed for another unspecified reason */
|
||||
| "FAILED";
|
||||
/**
|
||||
* An upstream endpoint's response to a TCP `keepalive` message from Cloudflare.
|
||||
*/
|
||||
@@ -9477,15 +9477,15 @@ interface D1ExecResult {
|
||||
count: number;
|
||||
duration: number;
|
||||
}
|
||||
type D1SessionConstraint =
|
||||
// Indicates that the first query should go to the primary, and the rest queries
|
||||
// using the same D1DatabaseSession will go to any replica that is consistent with
|
||||
// the bookmark maintained by the session (returned by the first query).
|
||||
'first-primary'
|
||||
// Indicates that the first query can go anywhere (primary or replica), and the rest queries
|
||||
// using the same D1DatabaseSession will go to any replica that is consistent with
|
||||
// the bookmark maintained by the session (returned by the first query).
|
||||
| 'first-unconstrained';
|
||||
type D1SessionConstraint =
|
||||
// Indicates that the first query should go to the primary, and the rest queries
|
||||
// using the same D1DatabaseSession will go to any replica that is consistent with
|
||||
// the bookmark maintained by the session (returned by the first query).
|
||||
'first-primary'
|
||||
// Indicates that the first query can go anywhere (primary or replica), and the rest queries
|
||||
// using the same D1DatabaseSession will go to any replica that is consistent with
|
||||
// the bookmark maintained by the session (returned by the first query).
|
||||
| 'first-unconstrained';
|
||||
type D1SessionBookmark = string;
|
||||
declare abstract class D1Database {
|
||||
prepare(query: string): D1PreparedStatement;
|
||||
@@ -9599,7 +9599,7 @@ declare type EmailExportedHandler<Env = unknown> = (message: ForwardableEmailMes
|
||||
declare module "cloudflare:email" {
|
||||
let _EmailMessage: {
|
||||
prototype: EmailMessage;
|
||||
new(from: string, to: string, raw: ReadableStream | string): EmailMessage;
|
||||
new (from: string, to: string, raw: ReadableStream | string): EmailMessage;
|
||||
};
|
||||
export { _EmailMessage as EmailMessage };
|
||||
}
|
||||
@@ -10058,17 +10058,17 @@ declare namespace Rpc {
|
||||
// The reason for using a generic type here is to build a serializable subset of structured
|
||||
// cloneable composite types. This allows types defined with the "interface" keyword to pass the
|
||||
// serializable check as well. Otherwise, only types defined with the "type" keyword would pass.
|
||||
type Serializable<T> =
|
||||
// Structured cloneables
|
||||
BaseType
|
||||
// Structured cloneable composites
|
||||
| Map<T extends Map<infer U, unknown> ? Serializable<U> : never, T extends Map<unknown, infer U> ? Serializable<U> : never> | Set<T extends Set<infer U> ? Serializable<U> : never> | ReadonlyArray<T extends ReadonlyArray<infer U> ? Serializable<U> : never> | {
|
||||
[K in keyof T]: K extends number | string ? Serializable<T[K]> : never;
|
||||
}
|
||||
// Special types
|
||||
| Stub<Stubable>
|
||||
// Serialized as stubs, see `Stubify`
|
||||
| Stubable;
|
||||
type Serializable<T> =
|
||||
// Structured cloneables
|
||||
BaseType
|
||||
// Structured cloneable composites
|
||||
| Map<T extends Map<infer U, unknown> ? Serializable<U> : never, T extends Map<unknown, infer U> ? Serializable<U> : never> | Set<T extends Set<infer U> ? Serializable<U> : never> | ReadonlyArray<T extends ReadonlyArray<infer U> ? Serializable<U> : never> | {
|
||||
[K in keyof T]: K extends number | string ? Serializable<T[K]> : never;
|
||||
}
|
||||
// Special types
|
||||
| Stub<Stubable>
|
||||
// Serialized as stubs, see `Stubify`
|
||||
| Stubable;
|
||||
// Base type for all RPC stubs, including common memory management methods.
|
||||
// `T` is used as a marker type for unwrapping `Stub`s later.
|
||||
interface StubBase<T extends Stubable> extends Disposable {
|
||||
@@ -10083,8 +10083,8 @@ declare namespace Rpc {
|
||||
type Stubify<T> = T extends Stubable ? Stub<T> : T extends Map<infer K, infer V> ? Map<Stubify<K>, Stubify<V>> : T extends Set<infer V> ? Set<Stubify<V>> : T extends Array<infer V> ? Array<Stubify<V>> : T extends ReadonlyArray<infer V> ? ReadonlyArray<Stubify<V>> : T extends BaseType ? T : T extends {
|
||||
[key: string | number]: any;
|
||||
} ? {
|
||||
[K in keyof T]: Stubify<T[K]>;
|
||||
} : T;
|
||||
[K in keyof T]: Stubify<T[K]>;
|
||||
} : T;
|
||||
// Recursively rewrite all `Stub<T>`s with the corresponding `T`s.
|
||||
// Note we use `StubBase` instead of `Stub` here to avoid circular dependencies:
|
||||
// `Stub` depends on `Provider`, which depends on `Unstubify`, which would depend on `Stub`.
|
||||
@@ -10092,8 +10092,8 @@ declare namespace Rpc {
|
||||
type Unstubify<T> = T extends StubBase<infer V> ? V : T extends Map<infer K, infer V> ? Map<Unstubify<K>, Unstubify<V>> : T extends Set<infer V> ? Set<Unstubify<V>> : T extends Array<infer V> ? Array<Unstubify<V>> : T extends ReadonlyArray<infer V> ? ReadonlyArray<Unstubify<V>> : T extends BaseType ? T : T extends {
|
||||
[key: string | number]: unknown;
|
||||
} ? {
|
||||
[K in keyof T]: Unstubify<T[K]>;
|
||||
} : T;
|
||||
[K in keyof T]: Unstubify<T[K]>;
|
||||
} : T;
|
||||
type UnstubifyAll<A extends any[]> = {
|
||||
[I in keyof A]: Unstubify<A[I]>;
|
||||
};
|
||||
@@ -10166,7 +10166,7 @@ declare namespace Cloudflare {
|
||||
[K in keyof MainModule]: LoopbackForExport<MainModule[K]>
|
||||
// If the export is listed in `durableNamespaces`, then it is also a
|
||||
// DurableObjectNamespace.
|
||||
& (K extends GlobalProp<"durableNamespaces", never> ? MainModule[K] extends new (...args: any[]) => infer DoInstance ? DoInstance extends Rpc.DurableObjectBranded ? DurableObjectNamespace<DoInstance> : DurableObjectNamespace<undefined> : DurableObjectNamespace<undefined> : {});
|
||||
& (K extends GlobalProp<"durableNamespaces", never> ? MainModule[K] extends new (...args: any[]) => infer DoInstance ? DoInstance extends Rpc.DurableObjectBranded ? DurableObjectNamespace<DoInstance> : DurableObjectNamespace<undefined> : DurableObjectNamespace<undefined> : {});
|
||||
};
|
||||
}
|
||||
declare namespace CloudflareWorkersModule {
|
||||
@@ -10251,9 +10251,6 @@ declare namespace CloudflareWorkersModule {
|
||||
export const env: Cloudflare.Env;
|
||||
export const exports: Cloudflare.Exports;
|
||||
}
|
||||
declare module 'cloudflare:test' {
|
||||
export = CloudflareWorkersModule;
|
||||
}
|
||||
declare module 'cloudflare:workers' {
|
||||
export = CloudflareWorkersModule;
|
||||
}
|
||||
@@ -10825,10 +10822,10 @@ interface WorkflowInstanceCreateOptions<PARAMS = unknown> {
|
||||
}
|
||||
type InstanceStatus = {
|
||||
status: 'queued' // means that instance is waiting to be started (see concurrency limits)
|
||||
| 'running' | 'paused' | 'errored' | 'terminated' // user terminated the instance while it was running
|
||||
| 'complete' | 'waiting' // instance is hibernating and waiting for sleep or event to finish
|
||||
| 'waitingForPause' // instance is finishing the current work to pause
|
||||
| 'unknown';
|
||||
| 'running' | 'paused' | 'errored' | 'terminated' // user terminated the instance while it was running
|
||||
| 'complete' | 'waiting' // instance is hibernating and waiting for sleep or event to finish
|
||||
| 'waitingForPause' // instance is finishing the current work to pause
|
||||
| 'unknown';
|
||||
error?: {
|
||||
name: string;
|
||||
message: string;
|
||||
|
||||
@@ -39,14 +39,6 @@ deleted_classes = ["AnilistDo"]
|
||||
tag = "v4"
|
||||
new_sqlite_classes = ["AnilistDo"]
|
||||
|
||||
[[migrations]]
|
||||
tag = "v5"
|
||||
deleted_classes = ["AnilistDo"]
|
||||
|
||||
[[migrations]]
|
||||
tag = "v6"
|
||||
new_sqlite_classes = ["AnilistDo"]
|
||||
|
||||
[[queues.producers]]
|
||||
queue = "anilist-updates"
|
||||
binding = "ANILIST_UPDATES"
|
||||
@@ -67,7 +59,7 @@ id = "c8db249d8ee7462b91f9c374321776e4"
|
||||
preview_id = "ff38240eb2aa4b1388c705f4974f5aec"
|
||||
|
||||
[triggers]
|
||||
crons = ["0 */12 * * *", "0 18 * * *"]
|
||||
crons = ["0 */12 * * *"]
|
||||
|
||||
[[d1_databases]]
|
||||
binding = "DB"
|
||||
|
||||
Reference in New Issue
Block a user