Project

General

Profile

Actions

Video Release Process » History » Revision 5

« Previous | Revision 5/9 (diff) | Next »
tnt, 02/27/2022 01:32 PM


Video Release Process

Unless the presenter has an objection, talks are recorded and published afterward. This page details the current process.

Note that the process uses Davinci Resolve for the editing. It's not OSS but this is what I'm used to and have already setup for other jobs, so this is what I use.

Source material

The BBB instance raw stream dump can be downloaded just by looking up the link in the page source. Or simply using the script below and feeding it the meeting ID which is in the URL when viewing the raw stream dump through the BBB interface.

Once downloaded, ffmpeg is used to repackage the video streams in `mp4` container and transcode the audio from OPUS into FLAC to bring them into containers/formats that the linux version of resolve can load. (Note newer version of resolve should support MKV now, but that hasn't been tried. Sticking with the "tried and working" process for now).

#!/bin/bash
set -e

ODC_SLUG="$1" 
MEETING_ID="$2" 

curl -o "tmp_webcams.webm"   "https://meeting5.franken.de/presentation/${MEETING_ID}/video/webcams.webm" 
curl -o "tmp_deskshare.webm" "https://meeting5.franken.de/presentation/${MEETING_ID}/deskshare/deskshare.webm" 

ffmpeg -i "tmp_webcams.webm"   -vn -c:a flac -sample_fmt s16 "${ODC_SLUG}-audio.flac" 
ffmpeg -i "tmp_webcams.webm"   -an -c:v copy                 "${ODC_SLUG}-webcam-vp9.mp4" 
ffmpeg -i "tmp_deskshare.webm" -an -c:v copy                 "${ODC_SLUG}-screen-vp9.mp4" 

rm "tmp_webcams.webm" "tmp_deskshare.webm" 

Edit

Just do a normal edit, trimming start / stop and anything that shouldn't be in there (like technical issues or whatever).
Some notes / guidelines :

  • Project Settings
    • 1280x720 24fps
    • Davinci YRGB (not color managed)
    • Fairlight -14 LUFS target loudness
  • Intro:
    • Use a Fusion composition (two examples attached), tweak as needed (font size, ...)
    • About 10 second long
    • Start speaker audio intro ~ 1-2 second in
    • Transition to slide: Dip-To-Color-Dissolve, 1s, Ease In-Out
  • Q&A:
    • If there is no such slide, use a Fusion composition for video
    • Cut out any long blanks
  • Outro:
    • Check there is a "thanks for watching" or similar
    • Fade to black ~ 3 second
  • Floating Head:
    • Magic mask with expand and feather edge. Add alpha out on color page from the magic mask node.
    • Add power window to control the crop
  • Audio:
    • FX: Dialogue processor filter with default for "Male VO"
    • EQ: Lo cut at 100 Hz, Hi cut at 10 kHz
    • Adjust audio level in Q/A section for the individual speaker
  • Master Render
    • Video: Grassvalley HQX 720p
    • Audio: Linear PCM 16 bits

Local encode

Once the master render is done, the file is transcoded into a few formats more suitable for online viewing using ffmpeg using the script below :
Theses are relatively low bitrates but perfectly fine for "slides" type of content with voice-over in 720p.
(Note the script is design to work on a machine with a NVidia card and driver to have hw acceleration of h264 and h265 encoding)

#!/bin/bash
set -e

ODC_SLUG="$1" 
ODC_RENDER_PATH="$2" 
ODC_RENDER_MASTER="${ODC_RENDER_PATH}/osmodevcall-${ODC_SLUG}_master.mov" 

ffmpeg \
    -hwaccel cuda -hwaccel_output_format cuda \
    -i "${ODC_RENDER_MASTER}" \
    -c:v h264_nvenc -b:v 1M -pix_fmt yuv420p \
    -c:a aac -b:a 96k \
    "${ODC_RENDER_PATH}/osmodevcall-${ODC_SLUG}_h264_420.mp4" 

ffmpeg \
    -hwaccel cuda -hwaccel_output_format cuda \
    -i "${ODC_RENDER_MASTER}" \
    -c:v hevc_nvenc -b:v 512k -pix_fmt yuv420p \
    -c:a aac -b:a 96k \
    "${ODC_RENDER_PATH}/osmodevcall-${ODC_SLUG}_h265_420.mp4" 

ffmpeg \
    -i "${ODC_RENDER_MASTER}" \
    -c:v libvpx-vp9 -b:v 400k \
    -c:a libopus -b:a 80k \
    "${ODC_RENDER_PATH}/osmodevcall-${ODC_SLUG}_vp9.webm" 

VOC upload

The final step is to feed the master render to the VOC rendering pipeline so the talk can be published on https://media.ccc.de .

  • Generate schedule file
    • Using a fork of the VOC tool : https://github.com/smunaut/voctosched in the osmodevcall branch
    • Edit the osmodevcall.tsv to add the new talk
    • Run ./schedule.py -c osmodevcall.ini to regenerate schedule-extended.xml
    • Upload that file on some public facing HTTP server
  • Upload master
  • In the VOC tracker
    • Use the "Import" button to import the updated XML from its source. Only import new tickets in the confirmation screen.
    • Once imported, edit the top level ticket for the talk and change state from "staging" to "staged".
    • The "Recording" subticket will then trigger and go to "Recording" state while its downloading the master file.
    • When done, it'll go to 'Recorded' state at which point, you need to click the 'cut' button. No actual cutting is needed, just select language and confirm with the "I finished cutting" button.
    • And it's done. All the rest should proceed automatically until all the version are in the "Released" state
Files (4)
qa-fusion.comp qa-fusion.comp 11.8 KB Fusion composition for Q&A tnt, 01/16/2022 08:26 AM
OsmoDevCall.png View OsmoDevCall.png 41.6 KB OsmoDevCall logo tnt, 01/16/2022 08:26 AM
title-sub-fusion.comp title-sub-fusion.comp 16.9 KB Fusion composition for Title screen with sub-title tnt, 01/16/2022 08:26 AM
title-fusion.comp title-fusion.comp 14.9 KB Fusion composition for Title screen tnt, 01/16/2022 08:26 AM

Updated by tnt about 2 years ago · 5 revisions

Add picture from clipboard (Maximum size: 48.8 MB)