Implement Cloudflare guest token caching

This commit is contained in:
dangered wolf 2022-08-18 03:09:41 -04:00
parent 4ae10034db
commit 7707c0d2d0
No known key found for this signature in database
GPG key ID: 41E4D37680ED8B58

View file

@ -1,8 +1,11 @@
import { Constants } from './constants'; import { Constants } from './constants';
export const fetchUsingGuest = async (status: string, event: FetchEvent): Promise<TimelineBlobPartial> => { export const fetchUsingGuest = async (
status: string,
event: FetchEvent
): Promise<TimelineBlobPartial> => {
let apiAttempts = 0; let apiAttempts = 0;
let cachedTokenFailed = false; let newTokenGenerated = false;
const tokenHeaders: { [header: string]: string } = { const tokenHeaders: { [header: string]: string } = {
Authorization: Constants.GUEST_BEARER_TOKEN, Authorization: Constants.GUEST_BEARER_TOKEN,
@ -14,11 +17,21 @@ export const fetchUsingGuest = async (status: string, event: FetchEvent): Promis
{ {
method: 'POST', method: 'POST',
headers: tokenHeaders, headers: tokenHeaders,
body: ''
}
);
/* A dummy version of the request only used for Cloudflare caching purposes.
The reason it exists at all is because Cloudflare won't cache POST requests. */
const guestTokenRequestCacheDummy = new Request(
`${Constants.TWITTER_API_ROOT}/1.1/guest/activate.json`,
{
method: 'GET',
headers: tokenHeaders,
cf: { cf: {
cacheEverything: true, cacheEverything: true,
cacheTtl: 600 cacheTtl: 300
}, }
body: ''
} }
); );
@ -41,19 +54,19 @@ export const fetchUsingGuest = async (status: string, event: FetchEvent): Promis
let activate: Response | null = null; let activate: Response | null = null;
if (!cachedTokenFailed) { if (!newTokenGenerated) {
const cachedResponse = await cache.match(guestTokenRequest); const cachedResponse = await cache.match(guestTokenRequestCacheDummy);
if (cachedResponse) { if (cachedResponse) {
console.log('Token cache hit'); console.log('Token cache hit');
activate = cachedResponse; activate = cachedResponse;
} } else {
console.log('Token cache miss'); console.log('Token cache miss');
cachedTokenFailed = true; newTokenGenerated = true;
}
} }
if (cachedTokenFailed || activate === null) { if (newTokenGenerated || activate === null) {
/* If all goes according to plan, we have a guest token we can use to call API /* If all goes according to plan, we have a guest token we can use to call API
AFAIK there is no limit to how many guest tokens you can request. AFAIK there is no limit to how many guest tokens you can request.
@ -66,14 +79,14 @@ export const fetchUsingGuest = async (status: string, event: FetchEvent): Promis
let activateJson: { guest_token: string }; let activateJson: { guest_token: string };
try { try {
activateJson = (await activate.json()) as { guest_token: string }; activateJson = (await activate.clone().json()) as { guest_token: string };
} catch (e: unknown) { } catch (e: unknown) {
continue; continue;
} }
const guestToken = activateJson.guest_token; const guestToken = activateJson.guest_token;
console.log('Activated guest:', activateJson); console.log(newTokenGenerated ? 'Activated guest:' : 'Using guest:', activateJson);
console.log('Guest token:', guestToken); console.log('Guest token:', guestToken);
/* Just some cookies to mimick what the Twitter Web App would send */ /* Just some cookies to mimick what the Twitter Web App would send */
@ -107,10 +120,21 @@ export const fetchUsingGuest = async (status: string, event: FetchEvent): Promis
/* We'll usually only hit this if we get an invalid response from Twitter. /* We'll usually only hit this if we get an invalid response from Twitter.
It's uncommon, but it happens */ It's uncommon, but it happens */
console.error('Unknown error while fetching conversation from API'); console.error('Unknown error while fetching conversation from API');
cachedTokenFailed = true; event && event.waitUntil(cache.delete(guestTokenRequestCacheDummy));
newTokenGenerated = true;
continue; continue;
} }
const remainingRateLimit = parseInt(
apiRequest.headers.get('x-rate-limit-remaining') || '0'
);
console.log(`Remaining rate limit: ${remainingRateLimit} requests`);
/* Running out of requests within our rate limit, let's purge the cache */
if (remainingRateLimit < 20) {
console.log(`Purging token on this edge due to low rate limit remaining`);
event && event.waitUntil(cache.delete(guestTokenRequestCacheDummy));
}
if ( if (
typeof conversation.globalObjects === 'undefined' && typeof conversation.globalObjects === 'undefined' &&
(typeof conversation.errors === 'undefined' || (typeof conversation.errors === 'undefined' ||
@ -118,11 +142,14 @@ export const fetchUsingGuest = async (status: string, event: FetchEvent): Promis
239) /* TODO: i forgot what code 239 actually is lol */ 239) /* TODO: i forgot what code 239 actually is lol */
) { ) {
console.log('Failed to fetch conversation, got', conversation); console.log('Failed to fetch conversation, got', conversation);
cachedTokenFailed = true; newTokenGenerated = true;
continue; continue;
} }
/* Once we've confirmed we have a working guest token, let's cache it! */ console.log('Caching guest token');
event.waitUntil(cache.put(guestTokenRequest, activate.clone())); /* If we've generated a new token, we'll cache it */
if (newTokenGenerated) {
event && event.waitUntil(cache.put(guestTokenRequestCacheDummy, activate.clone()));
}
conversation.guestToken = guestToken; conversation.guestToken = guestToken;
return conversation; return conversation;
} }