0305db28c9
* Add support for saving video files to object storage * Add support for custom url generation on s3 stored files Uses two config keys to support url generation that doesn't directly go to (compatible s3). Can be used to generate urls to any cache server or CDN. * Upload files to s3 concurrently and delete originals afterwards * Only publish after move to object storage is complete * Use base url instead of url template * Fix mistyped config field * Add rudenmentary way to download before transcode * Implement Chocobozzz suggestions https://github.com/Chocobozzz/PeerTube/pull/4290#issuecomment-891670478 The remarks in question: Try to use objectStorage prefix instead of s3 prefix for your function/variables/config names Prefer to use a tree for the config: s3.streaming_playlists_bucket -> object_storage.streaming_playlists.bucket Use uppercase for config: S3.STREAMING_PLAYLISTS_BUCKETINFO.bucket -> OBJECT_STORAGE.STREAMING_PLAYLISTS.BUCKET (maybe BUCKET_NAME instead of BUCKET) I suggest to rename moveJobsRunning to pendingMovingJobs (or better, create a dedicated videoJobInfo table with a pendingMove & videoId columns so we could also use this table to track pending transcoding jobs) https://github.com/Chocobozzz/PeerTube/pull/4290/files#diff-3e26d41ca4bda1de8e1747af70ca2af642abcc1e9e0bfb94239ff2165acfbde5R19 uses a string instead of an integer I think we should store the origin object storage URL in fileUrl, without base_url injection. Instead, inject the base_url at "runtime" so admins can easily change this configuration without running a script to update DB URLs * Import correct function * Support multipart upload * Remove import of node 15.0 module stream/promises * Extend maximum upload job length Using the same value as for redundancy downloading seems logical * Use dynamic part size for really large uploads Also adds very small part size for local testing * Fix decreasePendingMove query * Resolve various PR comments * Move to object storage after optimize * Make upload size configurable and increase default * Prune webtorrent files that are stored in object storage * Move files after transcoding jobs * Fix federation * Add video path manager * Support move to external storage job in client * Fix live object storage tests Co-authored-by: Chocobozzz <me@florianbigard.com>
440 lines
13 KiB
TypeScript
440 lines
13 KiB
TypeScript
|
|
import { createServer, Server } from 'net'
|
|
import { isTestInstance } from '@server/helpers/core-utils'
|
|
import {
|
|
computeResolutionsToTranscode,
|
|
ffprobePromise,
|
|
getVideoFileBitrate,
|
|
getVideoFileFPS,
|
|
getVideoFileResolution
|
|
} from '@server/helpers/ffprobe-utils'
|
|
import { logger, loggerTagsFactory } from '@server/helpers/logger'
|
|
import { CONFIG, registerConfigChangedHandler } from '@server/initializers/config'
|
|
import { P2P_MEDIA_LOADER_PEER_VERSION, VIDEO_LIVE, VIEW_LIFETIME } from '@server/initializers/constants'
|
|
import { UserModel } from '@server/models/user/user'
|
|
import { VideoModel } from '@server/models/video/video'
|
|
import { VideoLiveModel } from '@server/models/video/video-live'
|
|
import { VideoStreamingPlaylistModel } from '@server/models/video/video-streaming-playlist'
|
|
import { MStreamingPlaylistVideo, MVideo, MVideoLiveVideo } from '@server/types/models'
|
|
import { VideoState, VideoStreamingPlaylistType } from '@shared/models'
|
|
import { federateVideoIfNeeded } from '../activitypub/videos'
|
|
import { JobQueue } from '../job-queue'
|
|
import { PeerTubeSocket } from '../peertube-socket'
|
|
import { generateHLSMasterPlaylistFilename, generateHlsSha256SegmentsFilename } from '../paths'
|
|
import { LiveQuotaStore } from './live-quota-store'
|
|
import { LiveSegmentShaStore } from './live-segment-sha-store'
|
|
import { cleanupLive } from './live-utils'
|
|
import { MuxingSession } from './shared'
|
|
|
|
const NodeRtmpSession = require('node-media-server/src/node_rtmp_session')
|
|
const context = require('node-media-server/src/node_core_ctx')
|
|
const nodeMediaServerLogger = require('node-media-server/src/node_core_logger')
|
|
|
|
// Disable node media server logs
|
|
nodeMediaServerLogger.setLogType(0)
|
|
|
|
const config = {
|
|
rtmp: {
|
|
port: CONFIG.LIVE.RTMP.PORT,
|
|
chunk_size: VIDEO_LIVE.RTMP.CHUNK_SIZE,
|
|
gop_cache: VIDEO_LIVE.RTMP.GOP_CACHE,
|
|
ping: VIDEO_LIVE.RTMP.PING,
|
|
ping_timeout: VIDEO_LIVE.RTMP.PING_TIMEOUT
|
|
},
|
|
transcoding: {
|
|
ffmpeg: 'ffmpeg'
|
|
}
|
|
}
|
|
|
|
const lTags = loggerTagsFactory('live')
|
|
|
|
class LiveManager {
|
|
|
|
private static instance: LiveManager
|
|
|
|
private readonly muxingSessions = new Map<string, MuxingSession>()
|
|
private readonly videoSessions = new Map<number, string>()
|
|
// Values are Date().getTime()
|
|
private readonly watchersPerVideo = new Map<number, number[]>()
|
|
|
|
private rtmpServer: Server
|
|
|
|
private constructor () {
|
|
}
|
|
|
|
init () {
|
|
const events = this.getContext().nodeEvent
|
|
events.on('postPublish', (sessionId: string, streamPath: string) => {
|
|
logger.debug('RTMP received stream', { id: sessionId, streamPath, ...lTags(sessionId) })
|
|
|
|
const splittedPath = streamPath.split('/')
|
|
if (splittedPath.length !== 3 || splittedPath[1] !== VIDEO_LIVE.RTMP.BASE_PATH) {
|
|
logger.warn('Live path is incorrect.', { streamPath, ...lTags(sessionId) })
|
|
return this.abortSession(sessionId)
|
|
}
|
|
|
|
this.handleSession(sessionId, streamPath, splittedPath[2])
|
|
.catch(err => logger.error('Cannot handle sessions.', { err, ...lTags(sessionId) }))
|
|
})
|
|
|
|
events.on('donePublish', sessionId => {
|
|
logger.info('Live session ended.', { sessionId, ...lTags(sessionId) })
|
|
})
|
|
|
|
registerConfigChangedHandler(() => {
|
|
if (!this.rtmpServer && CONFIG.LIVE.ENABLED === true) {
|
|
this.run()
|
|
return
|
|
}
|
|
|
|
if (this.rtmpServer && CONFIG.LIVE.ENABLED === false) {
|
|
this.stop()
|
|
}
|
|
})
|
|
|
|
// Cleanup broken lives, that were terminated by a server restart for example
|
|
this.handleBrokenLives()
|
|
.catch(err => logger.error('Cannot handle broken lives.', { err, ...lTags() }))
|
|
|
|
setInterval(() => this.updateLiveViews(), VIEW_LIFETIME.LIVE)
|
|
}
|
|
|
|
run () {
|
|
logger.info('Running RTMP server on port %d', config.rtmp.port, lTags())
|
|
|
|
this.rtmpServer = createServer(socket => {
|
|
const session = new NodeRtmpSession(config, socket)
|
|
|
|
session.run()
|
|
})
|
|
|
|
this.rtmpServer.on('error', err => {
|
|
logger.error('Cannot run RTMP server.', { err, ...lTags() })
|
|
})
|
|
|
|
this.rtmpServer.listen(CONFIG.LIVE.RTMP.PORT)
|
|
}
|
|
|
|
stop () {
|
|
logger.info('Stopping RTMP server.', lTags())
|
|
|
|
this.rtmpServer.close()
|
|
this.rtmpServer = undefined
|
|
|
|
// Sessions is an object
|
|
this.getContext().sessions.forEach((session: any) => {
|
|
if (session instanceof NodeRtmpSession) {
|
|
session.stop()
|
|
}
|
|
})
|
|
}
|
|
|
|
isRunning () {
|
|
return !!this.rtmpServer
|
|
}
|
|
|
|
stopSessionOf (videoId: number) {
|
|
const sessionId = this.videoSessions.get(videoId)
|
|
if (!sessionId) return
|
|
|
|
this.videoSessions.delete(videoId)
|
|
this.abortSession(sessionId)
|
|
}
|
|
|
|
addViewTo (videoId: number) {
|
|
if (this.videoSessions.has(videoId) === false) return
|
|
|
|
let watchers = this.watchersPerVideo.get(videoId)
|
|
|
|
if (!watchers) {
|
|
watchers = []
|
|
this.watchersPerVideo.set(videoId, watchers)
|
|
}
|
|
|
|
watchers.push(new Date().getTime())
|
|
}
|
|
|
|
private getContext () {
|
|
return context
|
|
}
|
|
|
|
private abortSession (sessionId: string) {
|
|
const session = this.getContext().sessions.get(sessionId)
|
|
if (session) {
|
|
session.stop()
|
|
this.getContext().sessions.delete(sessionId)
|
|
}
|
|
|
|
const muxingSession = this.muxingSessions.get(sessionId)
|
|
if (muxingSession) {
|
|
// Muxing session will fire and event so we correctly cleanup the session
|
|
muxingSession.abort()
|
|
|
|
this.muxingSessions.delete(sessionId)
|
|
}
|
|
}
|
|
|
|
private async handleSession (sessionId: string, streamPath: string, streamKey: string) {
|
|
const videoLive = await VideoLiveModel.loadByStreamKey(streamKey)
|
|
if (!videoLive) {
|
|
logger.warn('Unknown live video with stream key %s.', streamKey, lTags(sessionId))
|
|
return this.abortSession(sessionId)
|
|
}
|
|
|
|
const video = videoLive.Video
|
|
if (video.isBlacklisted()) {
|
|
logger.warn('Video is blacklisted. Refusing stream %s.', streamKey, lTags(sessionId, video.uuid))
|
|
return this.abortSession(sessionId)
|
|
}
|
|
|
|
// Cleanup old potential live files (could happen with a permanent live)
|
|
LiveSegmentShaStore.Instance.cleanupShaSegments(video.uuid)
|
|
|
|
const oldStreamingPlaylist = await VideoStreamingPlaylistModel.loadHLSPlaylistByVideo(video.id)
|
|
if (oldStreamingPlaylist) {
|
|
await cleanupLive(video, oldStreamingPlaylist)
|
|
}
|
|
|
|
this.videoSessions.set(video.id, sessionId)
|
|
|
|
const rtmpUrl = 'rtmp://127.0.0.1:' + config.rtmp.port + streamPath
|
|
|
|
const now = Date.now()
|
|
const probe = await ffprobePromise(rtmpUrl)
|
|
|
|
const [ { resolution, ratio }, fps, bitrate ] = await Promise.all([
|
|
getVideoFileResolution(rtmpUrl, probe),
|
|
getVideoFileFPS(rtmpUrl, probe),
|
|
getVideoFileBitrate(rtmpUrl, probe)
|
|
])
|
|
|
|
logger.info(
|
|
'%s probing took %d ms (bitrate: %d, fps: %d, resolution: %d)',
|
|
rtmpUrl, Date.now() - now, bitrate, fps, resolution, lTags(sessionId, video.uuid)
|
|
)
|
|
|
|
const allResolutions = this.buildAllResolutionsToTranscode(resolution)
|
|
|
|
logger.info(
|
|
'Will mux/transcode live video of original resolution %d.', resolution,
|
|
{ allResolutions, ...lTags(sessionId, video.uuid) }
|
|
)
|
|
|
|
const streamingPlaylist = await this.createLivePlaylist(video, allResolutions)
|
|
|
|
return this.runMuxingSession({
|
|
sessionId,
|
|
videoLive,
|
|
streamingPlaylist,
|
|
rtmpUrl,
|
|
fps,
|
|
bitrate,
|
|
ratio,
|
|
allResolutions
|
|
})
|
|
}
|
|
|
|
private async runMuxingSession (options: {
|
|
sessionId: string
|
|
videoLive: MVideoLiveVideo
|
|
streamingPlaylist: MStreamingPlaylistVideo
|
|
rtmpUrl: string
|
|
fps: number
|
|
bitrate: number
|
|
ratio: number
|
|
allResolutions: number[]
|
|
}) {
|
|
const { sessionId, videoLive, streamingPlaylist, allResolutions, fps, bitrate, ratio, rtmpUrl } = options
|
|
const videoUUID = videoLive.Video.uuid
|
|
const localLTags = lTags(sessionId, videoUUID)
|
|
|
|
const user = await UserModel.loadByLiveId(videoLive.id)
|
|
LiveQuotaStore.Instance.addNewLive(user.id, videoLive.id)
|
|
|
|
const muxingSession = new MuxingSession({
|
|
context: this.getContext(),
|
|
user,
|
|
sessionId,
|
|
videoLive,
|
|
streamingPlaylist,
|
|
rtmpUrl,
|
|
bitrate,
|
|
ratio,
|
|
fps,
|
|
allResolutions
|
|
})
|
|
|
|
muxingSession.on('master-playlist-created', () => this.publishAndFederateLive(videoLive, localLTags))
|
|
|
|
muxingSession.on('bad-socket-health', ({ videoId }) => {
|
|
logger.error(
|
|
'Too much data in client socket stream (ffmpeg is too slow to transcode the video).' +
|
|
' Stopping session of video %s.', videoUUID,
|
|
localLTags
|
|
)
|
|
|
|
this.stopSessionOf(videoId)
|
|
})
|
|
|
|
muxingSession.on('duration-exceeded', ({ videoId }) => {
|
|
logger.info('Stopping session of %s: max duration exceeded.', videoUUID, localLTags)
|
|
|
|
this.stopSessionOf(videoId)
|
|
})
|
|
|
|
muxingSession.on('quota-exceeded', ({ videoId }) => {
|
|
logger.info('Stopping session of %s: user quota exceeded.', videoUUID, localLTags)
|
|
|
|
this.stopSessionOf(videoId)
|
|
})
|
|
|
|
muxingSession.on('ffmpeg-error', ({ sessionId }) => this.abortSession(sessionId))
|
|
muxingSession.on('ffmpeg-end', ({ videoId }) => {
|
|
this.onMuxingFFmpegEnd(videoId)
|
|
})
|
|
|
|
muxingSession.on('after-cleanup', ({ videoId }) => {
|
|
this.muxingSessions.delete(sessionId)
|
|
|
|
muxingSession.destroy()
|
|
|
|
return this.onAfterMuxingCleanup(videoId)
|
|
.catch(err => logger.error('Error in end transmuxing.', { err, ...localLTags }))
|
|
})
|
|
|
|
this.muxingSessions.set(sessionId, muxingSession)
|
|
|
|
muxingSession.runMuxing()
|
|
.catch(err => {
|
|
logger.error('Cannot run muxing.', { err, ...localLTags })
|
|
this.abortSession(sessionId)
|
|
})
|
|
}
|
|
|
|
private async publishAndFederateLive (live: MVideoLiveVideo, localLTags: { tags: string[] }) {
|
|
const videoId = live.videoId
|
|
|
|
try {
|
|
const video = await VideoModel.loadAndPopulateAccountAndServerAndTags(videoId)
|
|
|
|
logger.info('Will publish and federate live %s.', video.url, localLTags)
|
|
|
|
video.state = VideoState.PUBLISHED
|
|
await video.save()
|
|
|
|
live.Video = video
|
|
|
|
setTimeout(() => {
|
|
federateVideoIfNeeded(video, false)
|
|
.catch(err => logger.error('Cannot federate live video %s.', video.url, { err, ...localLTags }))
|
|
|
|
PeerTubeSocket.Instance.sendVideoLiveNewState(video)
|
|
}, VIDEO_LIVE.SEGMENT_TIME_SECONDS * 1000 * VIDEO_LIVE.EDGE_LIVE_DELAY_SEGMENTS_NOTIFICATION)
|
|
} catch (err) {
|
|
logger.error('Cannot save/federate live video %d.', videoId, { err, ...localLTags })
|
|
}
|
|
}
|
|
|
|
private onMuxingFFmpegEnd (videoId: number) {
|
|
this.watchersPerVideo.delete(videoId)
|
|
this.videoSessions.delete(videoId)
|
|
}
|
|
|
|
private async onAfterMuxingCleanup (videoUUID: string, cleanupNow = false) {
|
|
try {
|
|
const fullVideo = await VideoModel.loadAndPopulateAccountAndServerAndTags(videoUUID)
|
|
if (!fullVideo) return
|
|
|
|
const live = await VideoLiveModel.loadByVideoId(fullVideo.id)
|
|
|
|
if (!live.permanentLive) {
|
|
JobQueue.Instance.createJob({
|
|
type: 'video-live-ending',
|
|
payload: {
|
|
videoId: fullVideo.id
|
|
}
|
|
}, { delay: cleanupNow ? 0 : VIDEO_LIVE.CLEANUP_DELAY })
|
|
|
|
fullVideo.state = VideoState.LIVE_ENDED
|
|
} else {
|
|
fullVideo.state = VideoState.WAITING_FOR_LIVE
|
|
}
|
|
|
|
await fullVideo.save()
|
|
|
|
PeerTubeSocket.Instance.sendVideoLiveNewState(fullVideo)
|
|
|
|
await federateVideoIfNeeded(fullVideo, false)
|
|
} catch (err) {
|
|
logger.error('Cannot save/federate new video state of live streaming of video %d.', videoUUID, { err, ...lTags(videoUUID) })
|
|
}
|
|
}
|
|
|
|
private async updateLiveViews () {
|
|
if (!this.isRunning()) return
|
|
|
|
if (!isTestInstance()) logger.info('Updating live video views.', lTags())
|
|
|
|
for (const videoId of this.watchersPerVideo.keys()) {
|
|
const notBefore = new Date().getTime() - VIEW_LIFETIME.LIVE
|
|
|
|
const watchers = this.watchersPerVideo.get(videoId)
|
|
|
|
const numWatchers = watchers.length
|
|
|
|
const video = await VideoModel.loadAndPopulateAccountAndServerAndTags(videoId)
|
|
video.views = numWatchers
|
|
await video.save()
|
|
|
|
await federateVideoIfNeeded(video, false)
|
|
|
|
PeerTubeSocket.Instance.sendVideoViewsUpdate(video)
|
|
|
|
// Only keep not expired watchers
|
|
const newWatchers = watchers.filter(w => w > notBefore)
|
|
this.watchersPerVideo.set(videoId, newWatchers)
|
|
|
|
logger.debug('New live video views for %s is %d.', video.url, numWatchers, lTags())
|
|
}
|
|
}
|
|
|
|
private async handleBrokenLives () {
|
|
const videoUUIDs = await VideoModel.listPublishedLiveUUIDs()
|
|
|
|
for (const uuid of videoUUIDs) {
|
|
await this.onAfterMuxingCleanup(uuid, true)
|
|
}
|
|
}
|
|
|
|
private buildAllResolutionsToTranscode (originResolution: number) {
|
|
const resolutionsEnabled = CONFIG.LIVE.TRANSCODING.ENABLED
|
|
? computeResolutionsToTranscode(originResolution, 'live')
|
|
: []
|
|
|
|
return resolutionsEnabled.concat([ originResolution ])
|
|
}
|
|
|
|
private async createLivePlaylist (video: MVideo, allResolutions: number[]): Promise<MStreamingPlaylistVideo> {
|
|
const playlist = await VideoStreamingPlaylistModel.loadOrGenerate(video)
|
|
|
|
playlist.playlistFilename = generateHLSMasterPlaylistFilename(true)
|
|
playlist.segmentsSha256Filename = generateHlsSha256SegmentsFilename(true)
|
|
|
|
playlist.p2pMediaLoaderPeerVersion = P2P_MEDIA_LOADER_PEER_VERSION
|
|
playlist.type = VideoStreamingPlaylistType.HLS
|
|
|
|
playlist.assignP2PMediaLoaderInfoHashes(video, allResolutions)
|
|
|
|
return playlist.save()
|
|
}
|
|
|
|
static get Instance () {
|
|
return this.instance || (this.instance = new this())
|
|
}
|
|
}
|
|
|
|
// ---------------------------------------------------------------------------
|
|
|
|
export {
|
|
LiveManager
|
|
}
|