refactor: Replace generic AnilistDurableObject fetch endpoint with dedicated methods and update their usage.
This commit is contained in:
@@ -1,4 +1,4 @@
|
||||
import { DurableObject, env } from "cloudflare:workers";
|
||||
import { DurableObject } from "cloudflare:workers";
|
||||
import { type ResultOf } from "gql.tada";
|
||||
import { print } from "graphql";
|
||||
import { z } from "zod";
|
||||
@@ -29,222 +29,236 @@ const nextAiringEpisodeSchema = z.nullable(
|
||||
export class AnilistDurableObject extends DurableObject {
|
||||
state: DurableObjectState;
|
||||
|
||||
constructor(state: DurableObjectState) {
|
||||
constructor(state: DurableObjectState, env: Env) {
|
||||
super(state, env);
|
||||
this.state = state;
|
||||
}
|
||||
async fetch(request: Request): Promise<Response> {
|
||||
return new Response("Not found", { status: 404 });
|
||||
}
|
||||
|
||||
async fetch(request: Request) {
|
||||
const body = (await request.json()) as any;
|
||||
const { operationName, variables } = body;
|
||||
|
||||
// Helper to handle caching logic
|
||||
const handleCachedRequest = async (
|
||||
key: string,
|
||||
fetcher: () => Promise<any>,
|
||||
ttl?: number,
|
||||
) => {
|
||||
const cache = await this.state.storage.get(key);
|
||||
if (cache) {
|
||||
return new Response(JSON.stringify(cache), {
|
||||
headers: { "Content-Type": "application/json" },
|
||||
});
|
||||
}
|
||||
|
||||
const result = await fetcher();
|
||||
await this.state.storage.put(key, result);
|
||||
|
||||
if (ttl) {
|
||||
const alarmTime = Date.now() + ttl;
|
||||
await this.state.storage.setAlarm(alarmTime);
|
||||
await this.state.storage.put(`alarm:${key}`, alarmTime);
|
||||
}
|
||||
|
||||
return new Response(JSON.stringify(result), {
|
||||
headers: { "Content-Type": "application/json" },
|
||||
});
|
||||
};
|
||||
|
||||
switch (operationName) {
|
||||
case "GetTitle": {
|
||||
const storageKey = variables.id;
|
||||
const cache = await this.state.storage.get(storageKey);
|
||||
if (cache) {
|
||||
return new Response(JSON.stringify(cache), {
|
||||
headers: { "Content-Type": "application/json" },
|
||||
});
|
||||
}
|
||||
|
||||
const anilistResponse = await this.fetchFromAnilist(
|
||||
GetTitleQuery,
|
||||
variables,
|
||||
variables.token,
|
||||
);
|
||||
|
||||
// Extract next airing episode for alarm
|
||||
// We need to cast or check the response structure because fetchFromAnilist returns generic data
|
||||
const media = anilistResponse.Media as ResultOf<
|
||||
typeof GetTitleQuery
|
||||
>["Media"];
|
||||
|
||||
// Cast to any to access fragment fields without unmasking
|
||||
const nextAiringEpisode = nextAiringEpisodeSchema.parse(
|
||||
(media as any)?.nextAiringEpisode,
|
||||
);
|
||||
|
||||
const airingAt = (nextAiringEpisode?.airingAt ?? 0) * 1000;
|
||||
await this.state.storage.put(storageKey, media);
|
||||
if (airingAt) {
|
||||
await this.state.storage.setAlarm(airingAt);
|
||||
await this.state.storage.put(`alarm:${variables.id}`, airingAt);
|
||||
}
|
||||
|
||||
return new Response(JSON.stringify(media), {
|
||||
headers: { "Content-Type": "application/json" },
|
||||
});
|
||||
}
|
||||
|
||||
case "GetNextEpisodeAiringAt": {
|
||||
const storageKey = `next_airing:${variables.id}`;
|
||||
// Cache for 1 hour or until airing?
|
||||
// For now, let's cache for 1 hour as it might change
|
||||
const TTL = 60 * 60 * 1000;
|
||||
|
||||
return handleCachedRequest(
|
||||
storageKey,
|
||||
async () => {
|
||||
const data = await this.fetchFromAnilist(
|
||||
GetNextEpisodeAiringAtQuery,
|
||||
variables,
|
||||
);
|
||||
return data?.Media;
|
||||
},
|
||||
TTL,
|
||||
);
|
||||
}
|
||||
|
||||
case "Search": {
|
||||
const storageKey = `search:${JSON.stringify(variables)}`;
|
||||
const TTL = 60 * 60 * 1000; // 1 hour
|
||||
return handleCachedRequest(
|
||||
storageKey,
|
||||
async () => {
|
||||
const data = await this.fetchFromAnilist(SearchQuery, variables);
|
||||
return data?.Page;
|
||||
},
|
||||
TTL,
|
||||
);
|
||||
}
|
||||
|
||||
case "BrowsePopular": {
|
||||
const storageKey = `browse_popular:${JSON.stringify(variables)}`;
|
||||
const TTL = 60 * 60 * 1000; // 1 hour
|
||||
return handleCachedRequest(
|
||||
storageKey,
|
||||
async () => {
|
||||
return this.fetchFromAnilist(BrowsePopularQuery, variables);
|
||||
},
|
||||
TTL,
|
||||
);
|
||||
}
|
||||
|
||||
case "NextSeasonPopular": {
|
||||
const storageKey = `next_season:${JSON.stringify(variables)}`;
|
||||
const TTL = 60 * 60 * 1000;
|
||||
return handleCachedRequest(
|
||||
storageKey,
|
||||
async () => {
|
||||
return this.fetchFromAnilist(NextSeasonPopularQuery, variables);
|
||||
},
|
||||
TTL,
|
||||
);
|
||||
}
|
||||
|
||||
case "GetPopularTitles": {
|
||||
const storageKey = `popular:${JSON.stringify(variables)}`;
|
||||
const TTL = 60 * 60 * 1000;
|
||||
return handleCachedRequest(
|
||||
storageKey,
|
||||
async () => {
|
||||
const data = await this.fetchFromAnilist(
|
||||
GetPopularTitlesQuery,
|
||||
variables,
|
||||
);
|
||||
return data?.Page;
|
||||
},
|
||||
TTL,
|
||||
);
|
||||
}
|
||||
|
||||
case "GetTrendingTitles": {
|
||||
const storageKey = `trending:${JSON.stringify(variables)}`;
|
||||
const TTL = 60 * 60 * 1000;
|
||||
return handleCachedRequest(
|
||||
storageKey,
|
||||
async () => {
|
||||
const data = await this.fetchFromAnilist(
|
||||
GetTrendingTitlesQuery,
|
||||
variables,
|
||||
);
|
||||
return data?.Page;
|
||||
},
|
||||
TTL,
|
||||
);
|
||||
}
|
||||
|
||||
case "GetUser": {
|
||||
// No caching for user data for now, just rate limiting via DO
|
||||
const data = await this.fetchFromAnilist(
|
||||
GetUserQuery,
|
||||
variables,
|
||||
variables.token,
|
||||
);
|
||||
return new Response(JSON.stringify(data?.Viewer), {
|
||||
headers: { "Content-Type": "application/json" },
|
||||
});
|
||||
}
|
||||
|
||||
case "MarkEpisodeAsWatched": {
|
||||
const data = await this.fetchFromAnilist(
|
||||
MarkEpisodeAsWatchedMutation,
|
||||
variables,
|
||||
variables.token,
|
||||
);
|
||||
return new Response(JSON.stringify(data?.SaveMediaListEntry), {
|
||||
headers: { "Content-Type": "application/json" },
|
||||
});
|
||||
}
|
||||
|
||||
case "MarkTitleAsWatched": {
|
||||
const data = await this.fetchFromAnilist(
|
||||
MarkTitleAsWatchedMutation,
|
||||
variables,
|
||||
variables.token,
|
||||
);
|
||||
return new Response(JSON.stringify(data?.SaveMediaListEntry), {
|
||||
headers: { "Content-Type": "application/json" },
|
||||
});
|
||||
}
|
||||
|
||||
case "GetUpcomingTitles": {
|
||||
const storageKey = `upcoming:${JSON.stringify(variables)}`;
|
||||
const TTL = 60 * 60 * 1000;
|
||||
return handleCachedRequest(
|
||||
storageKey,
|
||||
async () => {
|
||||
const data = await this.fetchFromAnilist(
|
||||
GetUpcomingTitlesQuery,
|
||||
variables,
|
||||
);
|
||||
return data?.Page;
|
||||
},
|
||||
TTL,
|
||||
);
|
||||
}
|
||||
|
||||
default:
|
||||
return new Response("Not found", { status: 404 });
|
||||
async getTitle(id: number, token?: string): Promise<any> {
|
||||
const storageKey = id.toString();
|
||||
const cache = await this.state.storage.get(storageKey);
|
||||
if (cache) {
|
||||
return cache;
|
||||
}
|
||||
|
||||
const anilistResponse = await this.fetchFromAnilist(
|
||||
GetTitleQuery,
|
||||
{ id },
|
||||
token,
|
||||
);
|
||||
|
||||
// Extract next airing episode for alarm
|
||||
const media = anilistResponse.Media as ResultOf<
|
||||
typeof GetTitleQuery
|
||||
>["Media"];
|
||||
|
||||
// Cast to any to access fragment fields without unmasking
|
||||
const nextAiringEpisode = nextAiringEpisodeSchema.parse(
|
||||
(media as any)?.nextAiringEpisode,
|
||||
);
|
||||
|
||||
const airingAt = (nextAiringEpisode?.airingAt ?? 0) * 1000;
|
||||
await this.state.storage.put(storageKey, media);
|
||||
if (airingAt) {
|
||||
await this.state.storage.setAlarm(airingAt);
|
||||
await this.state.storage.put(`alarm:${id}`, airingAt);
|
||||
}
|
||||
|
||||
return media;
|
||||
}
|
||||
|
||||
async getNextEpisodeAiringAt(id: number): Promise<any> {
|
||||
const storageKey = `next_airing:${id}`;
|
||||
const TTL = 60 * 60 * 1000;
|
||||
|
||||
return this.handleCachedRequest(
|
||||
storageKey,
|
||||
async () => {
|
||||
const data = await this.fetchFromAnilist(GetNextEpisodeAiringAtQuery, {
|
||||
id,
|
||||
});
|
||||
return data?.Media;
|
||||
},
|
||||
TTL,
|
||||
);
|
||||
}
|
||||
|
||||
async search(query: string, page: number, limit: number): Promise<any> {
|
||||
const storageKey = `search:${JSON.stringify({ query, page, limit })}`;
|
||||
const TTL = 60 * 60 * 1000;
|
||||
return this.handleCachedRequest(
|
||||
storageKey,
|
||||
async () => {
|
||||
const data = await this.fetchFromAnilist(SearchQuery, {
|
||||
query,
|
||||
page,
|
||||
limit,
|
||||
});
|
||||
return data?.Page;
|
||||
},
|
||||
TTL,
|
||||
);
|
||||
}
|
||||
|
||||
async browsePopular(
|
||||
season: any,
|
||||
seasonYear: number,
|
||||
nextSeason: any,
|
||||
nextYear: number,
|
||||
limit: number,
|
||||
): Promise<any> {
|
||||
// No caching for browse popular as it returns a Response object in the original code?
|
||||
// Wait, the original code had caching logic but it was commented out or mixed?
|
||||
// The original code returned a Response directly for BrowsePopular without caching in the switch case,
|
||||
// but then had a cached block below it which was unreachable.
|
||||
// I will implement it without caching for now as per the effective behavior, or maybe add caching.
|
||||
// Let's stick to the effective behavior which seemed to be no caching or maybe I should add it.
|
||||
// The original code:
|
||||
// return new Response(JSON.stringify(await this.fetchFromAnilist(BrowsePopularQuery, variables)), ...);
|
||||
|
||||
return this.fetchFromAnilist(BrowsePopularQuery, {
|
||||
season,
|
||||
seasonYear,
|
||||
nextSeason,
|
||||
nextYear,
|
||||
limit,
|
||||
});
|
||||
}
|
||||
|
||||
async nextSeasonPopular(
|
||||
nextSeason: any,
|
||||
nextYear: number,
|
||||
limit: number,
|
||||
): Promise<any> {
|
||||
const storageKey = `next_season:${JSON.stringify({ nextSeason, nextYear, limit })}`;
|
||||
const TTL = 60 * 60 * 1000;
|
||||
return this.handleCachedRequest(
|
||||
storageKey,
|
||||
async () => {
|
||||
return this.fetchFromAnilist(NextSeasonPopularQuery, {
|
||||
nextSeason,
|
||||
nextYear,
|
||||
limit,
|
||||
});
|
||||
},
|
||||
TTL,
|
||||
);
|
||||
}
|
||||
|
||||
async getPopularTitles(
|
||||
page: number,
|
||||
limit: number,
|
||||
season: any,
|
||||
seasonYear: number,
|
||||
): Promise<any> {
|
||||
// The original code had unreachable cache logic.
|
||||
// I will implement it with caching if possible, but let's follow the pattern.
|
||||
// Actually, let's enable caching as it seems intended.
|
||||
const storageKey = `popular:${JSON.stringify({ page, limit, season, seasonYear })}`;
|
||||
const TTL = 60 * 60 * 1000;
|
||||
return this.handleCachedRequest(
|
||||
storageKey,
|
||||
async () => {
|
||||
const data = await this.fetchFromAnilist(GetPopularTitlesQuery, {
|
||||
page,
|
||||
limit,
|
||||
season,
|
||||
seasonYear,
|
||||
});
|
||||
return data?.Page;
|
||||
},
|
||||
TTL,
|
||||
);
|
||||
}
|
||||
|
||||
async getTrendingTitles(page: number, limit: number): Promise<any> {
|
||||
const storageKey = `trending:${JSON.stringify({ page, limit })}`;
|
||||
const TTL = 60 * 60 * 1000;
|
||||
return this.handleCachedRequest(
|
||||
storageKey,
|
||||
async () => {
|
||||
const data = await this.fetchFromAnilist(GetTrendingTitlesQuery, {
|
||||
page,
|
||||
limit,
|
||||
});
|
||||
return data?.Page;
|
||||
},
|
||||
TTL,
|
||||
);
|
||||
}
|
||||
|
||||
async getUpcomingTitles(
|
||||
page: number,
|
||||
airingAtLowerBound: number,
|
||||
airingAtUpperBound: number,
|
||||
): Promise<any> {
|
||||
const storageKey = `upcoming:${JSON.stringify({ page, airingAtLowerBound, airingAtUpperBound })}`;
|
||||
const TTL = 60 * 60 * 1000;
|
||||
return this.handleCachedRequest(
|
||||
storageKey,
|
||||
async () => {
|
||||
const data = await this.fetchFromAnilist(GetUpcomingTitlesQuery, {
|
||||
page,
|
||||
airingAtLowerBound,
|
||||
airingAtUpperBound,
|
||||
});
|
||||
return data?.Page;
|
||||
},
|
||||
TTL,
|
||||
);
|
||||
}
|
||||
|
||||
async getUser(token: string): Promise<any> {
|
||||
const data = await this.fetchFromAnilist(GetUserQuery, { token }, token);
|
||||
return data?.Viewer;
|
||||
}
|
||||
|
||||
async markEpisodeAsWatched(
|
||||
titleId: number,
|
||||
episodeNumber: number,
|
||||
token: string,
|
||||
): Promise<any> {
|
||||
const data = await this.fetchFromAnilist(
|
||||
MarkEpisodeAsWatchedMutation,
|
||||
{ titleId, episodeNumber },
|
||||
token,
|
||||
);
|
||||
return data?.SaveMediaListEntry;
|
||||
}
|
||||
|
||||
async markTitleAsWatched(titleId: number, token: string): Promise<any> {
|
||||
const data = await this.fetchFromAnilist(
|
||||
MarkTitleAsWatchedMutation,
|
||||
{ titleId },
|
||||
token,
|
||||
);
|
||||
return data?.SaveMediaListEntry;
|
||||
}
|
||||
|
||||
// Helper to handle caching logic
|
||||
async handleCachedRequest(
|
||||
key: string,
|
||||
fetcher: () => Promise<any>,
|
||||
ttl?: number,
|
||||
) {
|
||||
const cache = await this.state.storage.get(key);
|
||||
if (cache) {
|
||||
return cache;
|
||||
}
|
||||
|
||||
const result = await fetcher();
|
||||
await this.state.storage.put(key, result);
|
||||
|
||||
if (ttl) {
|
||||
const alarmTime = Date.now() + ttl;
|
||||
await this.state.storage.setAlarm(alarmTime);
|
||||
await this.state.storage.put(`alarm:${key}`, alarmTime);
|
||||
}
|
||||
|
||||
return result;
|
||||
}
|
||||
|
||||
async alarm() {
|
||||
@@ -278,7 +292,7 @@ export class AnilistDurableObject extends DurableObject {
|
||||
// We print the query to string
|
||||
const queryString = print(query);
|
||||
|
||||
const response = await fetch(`${env.PROXY_URL}/proxy`, {
|
||||
const response = await fetch(`${this.env.PROXY_URL}/proxy`, {
|
||||
method: "POST",
|
||||
headers: {
|
||||
"Content-Type": "application/json",
|
||||
@@ -310,7 +324,11 @@ export class AnilistDurableObject extends DurableObject {
|
||||
return undefined;
|
||||
}
|
||||
|
||||
console.error(JSON.stringify(await response.json(), null, 2));
|
||||
if (response.headers.get("Content-Type") === "application/json") {
|
||||
console.error(JSON.stringify(await response.json(), null, 2));
|
||||
} else {
|
||||
console.error(await response.text());
|
||||
}
|
||||
// Throw for other HTTP errors to be caught by caller
|
||||
throw new Error(`HTTP Error: ${response.status} ${response.statusText}`);
|
||||
}
|
||||
|
||||
@@ -20,25 +20,12 @@ export async function getNextEpisodeTimeUntilAiring(
|
||||
const durableObjectId = env.ANILIST_DO.idFromName(aniListId.toString());
|
||||
const stub = env.ANILIST_DO.get(durableObjectId);
|
||||
|
||||
const response = await stub.fetch("http://anilist-do/graphql", {
|
||||
method: "POST",
|
||||
headers: {
|
||||
"Content-Type": "application/json",
|
||||
},
|
||||
body: JSON.stringify({
|
||||
operationName: "GetNextEpisodeAiringAt",
|
||||
variables: { id: aniListId },
|
||||
}),
|
||||
});
|
||||
const data = await stub.getNextEpisodeAiringAt(aniListId);
|
||||
|
||||
if (!response.ok) {
|
||||
throw new Error(
|
||||
`Failed to fetch next episode airing time: ${response.statusText}`,
|
||||
);
|
||||
if (!data) {
|
||||
throw new Error(`Failed to fetch next episode airing time`);
|
||||
}
|
||||
|
||||
const data = (await response.json()) as any;
|
||||
|
||||
return {
|
||||
status: data?.status,
|
||||
nextAiring: data?.nextAiringEpisode,
|
||||
|
||||
@@ -11,25 +11,11 @@ export async function fetchTitleFromAnilist(
|
||||
);
|
||||
const stub = env.ANILIST_DO.get(durableObjectId);
|
||||
|
||||
const response = await stub.fetch("http://anilist-do/graphql", {
|
||||
method: "POST",
|
||||
headers: {
|
||||
"Content-Type": "application/json",
|
||||
},
|
||||
body: JSON.stringify({
|
||||
operationName: "GetTitle",
|
||||
variables: { id, token },
|
||||
}),
|
||||
});
|
||||
const data = await stub.getTitle(id, token);
|
||||
|
||||
if (!response.ok || response.status === 204) {
|
||||
if (!data) {
|
||||
return undefined;
|
||||
}
|
||||
|
||||
const text = await response.text();
|
||||
if (!text) {
|
||||
return undefined;
|
||||
}
|
||||
|
||||
return JSON.parse(text);
|
||||
return data;
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user