1
0
Fork 0
peertube/server/lib/activitypub/crawl.ts

63 lines
1.7 KiB
TypeScript
Raw Normal View History

import * as Bluebird from 'bluebird'
2020-01-31 10:56:52 -05:00
import { URL } from 'url'
2021-03-08 08:24:11 -05:00
import { ActivityPubOrderedCollection } from '../../../shared/models/activitypub'
import { logger } from '../../helpers/logger'
import { doJSONRequest } from '../../helpers/requests'
import { ACTIVITY_PUB, REQUEST_TIMEOUT, WEBSERVER } from '../../initializers/constants'
2019-03-19 11:23:02 -04:00
type HandlerFunction<T> = (items: T[]) => (Promise<any> | Bluebird<any>)
type CleanerFunction = (startedDate: Date) => (Promise<any> | Bluebird<any>)
2021-03-08 08:24:11 -05:00
async function crawlCollectionPage <T> (argUrl: string, handler: HandlerFunction<T>, cleaner?: CleanerFunction) {
let url = argUrl
logger.info('Crawling ActivityPub data on %s.', url)
const options = {
activityPub: true,
2021-03-03 04:10:55 -05:00
timeout: REQUEST_TIMEOUT
}
2019-03-19 11:23:02 -04:00
const startDate = new Date()
2021-03-08 08:24:11 -05:00
const response = await doJSONRequest<ActivityPubOrderedCollection<T>>(url, options)
const firstBody = response.body
2020-01-31 10:56:52 -05:00
const limit = ACTIVITY_PUB.FETCH_PAGE_LIMIT
let i = 0
let nextLink = firstBody.first
while (nextLink && i < limit) {
let body: any
2019-04-25 08:23:15 -04:00
if (typeof nextLink === 'string') {
// Don't crawl ourselves
2020-01-31 10:56:52 -05:00
const remoteHost = new URL(nextLink).host
if (remoteHost === WEBSERVER.HOST) continue
2021-03-08 08:24:11 -05:00
url = nextLink
2021-03-08 08:24:11 -05:00
const res = await doJSONRequest<ActivityPubOrderedCollection<T>>(url, options)
body = res.body
} else {
// nextLink is already the object we want
body = nextLink
}
nextLink = body.next
i++
if (Array.isArray(body.orderedItems)) {
const items = body.orderedItems
2021-03-08 08:24:11 -05:00
logger.info('Processing %i ActivityPub items for %s.', items.length, url)
await handler(items)
}
}
2019-03-19 11:23:02 -04:00
if (cleaner) await cleaner(startDate)
}
export {
crawlCollectionPage
}