fix: Gstreamer pipeline quitting too soon.

TODO: Fix error with of an overflowing ``GST:Buffer``
This commit is contained in:
Wanjohi
2024-01-03 06:09:59 +03:00
parent 60cae1e7b8
commit 4ec4dc27b5
2 changed files with 416 additions and 268 deletions

View File

@@ -75,7 +75,7 @@ async fn main() -> anyhow::Result<()> {
quinn::Endpoint::client(config.bind)?; quinn::Endpoint::client(config.bind)?;
endpoint.set_default_client_config(quinn_client_config); endpoint.set_default_client_config(quinn_client_config);
log::info!("connecting to relay: url={}", config.url); println!("connecting to relay: url={}", config.url);
let session = webtransport_quinn::connect(&endpoint, &config.url) let session = webtransport_quinn::connect(&endpoint, &config.url)
.await .await

View File

@@ -1,4 +1,4 @@
use anyhow::{self,Context}; use anyhow::{self, Context};
use gst::prelude::*; use gst::prelude::*;
use gst::ClockTime; use gst::ClockTime;
@@ -6,10 +6,10 @@ use std::sync::{Arc, Mutex};
use moq_transport::cache::{broadcast, fragment, segment, track}; use moq_transport::cache::{broadcast, fragment, segment, track};
use moq_transport::VarInt; use moq_transport::VarInt;
use std::collections::HashMap;
use std::io::Cursor;
use serde_json::json; use serde_json::json;
use std::cmp::max; use std::cmp::max;
use std::collections::HashMap;
use std::io::Cursor;
use std::time; use std::time;
use mp4::{self, ReadBox}; use mp4::{self, ReadBox};
@@ -64,6 +64,9 @@ impl Mp4Parser {
let mut atom_bytes = Vec::with_capacity(atom_size); let mut atom_bytes = Vec::with_capacity(atom_size);
// TODO: Swap vectors? // TODO: Swap vectors?
atom_bytes.extend_from_slice(&self.buf[0..atom_size]); atom_bytes.extend_from_slice(&self.buf[0..atom_size]);
println!("self.buffer length in Mp4 parser {:?}", self.buf.len());
println!("atom size {:?}", atom_size);
assert_eq!(self.buf.len(), atom_size); assert_eq!(self.buf.len(), atom_size);
self.buf.clear(); self.buf.clear();
Some(Mp4Atom { Some(Mp4Atom {
@@ -105,17 +108,19 @@ struct State {
// We hold on to publisher so we don't close then while media is still being published. // We hold on to publisher so we don't close then while media is still being published.
broadcast: broadcast::Publisher, broadcast: broadcast::Publisher,
catalog: Option<track::Publisher>, catalog: Option<track::Publisher>,
init: Option<track::Publisher>, init: Option<track::Publisher>,
// Tracks based on their track ID. // Tracks based on their track ID.
tracks: Option<HashMap<u32, Track>>, tracks: Option<HashMap<u32, Track>>,
current: Option<u32>, current: Option<u32>,
} }
pub struct GST {} pub struct GST {}
impl GST { impl GST {
pub async fn run(mut broadcast: broadcast::Publisher) -> anyhow::Result<()> { pub async fn run(mut broadcast: broadcast::Publisher) -> anyhow::Result<()> {
println!("starting up GST.run");
gst::init()?; gst::init()?;
//FIXME: add audio pipeline //FIXME: add audio pipeline
@@ -140,10 +145,10 @@ impl GST {
mp4_parser: Mp4Parser::new(), mp4_parser: Mp4Parser::new(),
// wave: "sine".to_string(), // wave: "sine".to_string(),
catalog: None, catalog: None,
init: None, init: None,
// Tracks based on their track ID. // Tracks based on their track ID.
tracks: None, tracks: None,
current: None, current: None,
})); }));
@@ -227,27 +232,49 @@ impl GST {
appsink.set_callbacks( appsink.set_callbacks(
gst_app::AppSinkCallbacks::builder() gst_app::AppSinkCallbacks::builder()
.new_sample(move |sink| { .new_sample(move |sink| {
let sample = sink.pull_sample().map_err(|_| gst::FlowError::Eos)?; let sample = sink
.pull_sample()
.with_context(|| "Error pulling sample")
.map_err(|e| {
eprintln!("{:?}", e);
gst::FlowError::Eos
})?;
// The muxer only outputs non-empty buffer lists // The muxer only outputs non-empty buffer lists
let buffer_list = sample.buffer_list_owned().expect("no buffer list"); let buffer_list = sample.buffer_list_owned().expect("no buffer list");
println!("buffer is empty {:?}", buffer_list.is_empty());
assert!(!buffer_list.is_empty()); assert!(!buffer_list.is_empty());
println!("bufferlist is this long {:?}", buffer_list.len());
let first = buffer_list.get(0).unwrap(); let first = buffer_list.get(0).unwrap();
// Each list contains a full segment, i.e. does not start with a DELTA_UNIT // Each list contains a full segment, i.e. does not start with a DELTA_UNIT
println!("first buffer has a delta unit {:?}", first.flags().contains(gst::BufferFlags::DELTA_UNIT));
assert!(!first.flags().contains(gst::BufferFlags::DELTA_UNIT)); assert!(!first.flags().contains(gst::BufferFlags::DELTA_UNIT));
for buffer in &*buffer_list { for buffer in &*buffer_list {
let map = buffer.map_readable().map_err(|_| gst::FlowError::Error)?; let map = buffer
.map_readable()
.with_context(|| "Error mapping buffer to readable")
.map_err(|e| {
eprintln!("{:?}", e);
gst::FlowError::Error
})?;
let input_buf = map.as_ref(); let input_buf = map.as_ref();
let mut state = state.lock().unwrap(); let mut state = state.lock().unwrap();
//FIXME: The buffer in the mp4parser is "overflowing"
//TODO: Find another way to "slice" a buffer at a time, probably use this method https://github.com/sdroege/gst-plugin-rs/blob/d9397ef1743ac92e84784d00b93dc0877d44f966/mux/fmp4/examples/hls_live.rs#L256C17-L280C18 or this
state.mp4_parser.add(input_buf); state.mp4_parser.add(input_buf);
// Update cummulative fragment variables. // Update cummulative fragment variables.
// Buffer PTS, etc. are only valid if this buffer contains MDAT data. // Buffer PTS, etc. are only valid if this buffer contains MDAT data.
if state.mp4_parser.have_mdat() { if state.mp4_parser.have_mdat() {
println!("buffer has pts {:?}", buffer.pts().is_some());
assert!(buffer.pts().is_some()); assert!(buffer.pts().is_some());
if state.fragment_pts.is_none() || state.fragment_pts > buffer.pts() { if state.fragment_pts.is_none() || state.fragment_pts > buffer.pts() {
state.fragment_pts = buffer.pts(); state.fragment_pts = buffer.pts();
@@ -302,108 +329,169 @@ impl GST {
let mut state = state.lock().unwrap(); let mut state = state.lock().unwrap();
match state.mp4_parser.pop_atom() { match state.mp4_parser.pop_atom() {
Some(atom) => { Some(atom) => {
log::info!("atom_size={}, atom_type={}", atom.len(), atom.atom_type); log::info!(
"atom_size={}, atom_type={}",
atom.len(),
atom.atom_type
);
match atom.atom_type { match atom.atom_type {
ATOM_TYPE_FTYPE => { ATOM_TYPE_FTYPE => {
state.ftype_atom = Some(atom); state.ftype_atom = Some(atom);
log::info!("ftype_atom={:?}", state.ftype_atom); log::info!("ftype_atom={:?}", state.ftype_atom);
}, }
ATOM_TYPE_MOOV => { ATOM_TYPE_MOOV => {
state.moov_atom = Some(atom); state.moov_atom = Some(atom);
log::info!("moov_atom={:?}", state.moov_atom); log::info!("moov_atom={:?}", state.moov_atom);
match (state.ftype_atom.as_ref(), state.moov_atom.as_ref()) { match (state.ftype_atom.as_ref(), state.moov_atom.as_ref())
{
(Some(ftype_atom), Some(moov_atom)) => { (Some(ftype_atom), Some(moov_atom)) => {
let output_buf_len = ftype_atom.len() + moov_atom.len(); let output_buf_len =
let mut gst_buffer = gst::Buffer::with_size(output_buf_len).unwrap(); ftype_atom.len() + moov_atom.len();
let mut gst_buffer =
gst::Buffer::with_size(output_buf_len).unwrap();
{ {
let buffer_ref = gst_buffer.get_mut().unwrap(); let buffer_ref = gst_buffer.get_mut().unwrap();
buffer_ref.set_pts(state.fragment_pts); buffer_ref.set_pts(state.fragment_pts);
buffer_ref.set_dts(state.fragment_dts); buffer_ref.set_dts(state.fragment_dts);
let pts_plus_duration =state.fragment_max_pts_plus_duration.clone(); let pts_plus_duration = state
.fragment_max_pts_plus_duration
.clone();
let fragment_pts = state.fragment_pts.clone(); let fragment_pts = state.fragment_pts.clone();
let duration = match (pts_plus_duration, fragment_pts) { let duration =
(Some(pts_plus_duration), Some(fragment_pts)) => { match (pts_plus_duration, fragment_pts) {
Some(pts_plus_duration - fragment_pts) (
} Some(pts_plus_duration),
// Handle the case where one or both values are `None` Some(fragment_pts),
_ => None, ) => Some(
}; pts_plus_duration - fragment_pts,
),
// Handle the case where one or both values are `None`
_ => None,
};
buffer_ref.set_duration(duration); buffer_ref.set_duration(duration);
buffer_ref.set_offset(state.fragment_offset.unwrap_or(gst_sys::GST_BUFFER_OFFSET_NONE)); buffer_ref.set_offset(
buffer_ref.set_offset_end(state.fragment_offset_end.unwrap_or(gst_sys::GST_BUFFER_OFFSET_NONE)); state.fragment_offset.unwrap_or(
gst_sys::GST_BUFFER_OFFSET_NONE,
),
);
buffer_ref.set_offset_end(
state.fragment_offset_end.unwrap_or(
gst_sys::GST_BUFFER_OFFSET_NONE,
),
);
//this is a header //this is a header
buffer_ref.set_flags(gst::BufferFlags::HEADER); buffer_ref.set_flags(gst::BufferFlags::HEADER);
let mut buffer_map = buffer_ref.map_writable().unwrap(); let mut buffer_map =
buffer_ref.map_writable().unwrap();
let slice = buffer_map.as_mut_slice(); let slice = buffer_map.as_mut_slice();
let mut pos = 0; let mut pos = 0;
slice[pos..pos+ftype_atom.len()].copy_from_slice(&ftype_atom.atom_bytes); slice[pos..pos + ftype_atom.len()]
.copy_from_slice(&ftype_atom.atom_bytes);
pos += ftype_atom.len(); pos += ftype_atom.len();
slice[pos..pos+moov_atom.len()].copy_from_slice(&moov_atom.atom_bytes); slice[pos..pos + moov_atom.len()]
.copy_from_slice(&moov_atom.atom_bytes);
pos += moov_atom.len(); pos += moov_atom.len();
log::info!("real size of the header atom={:?}", pos); println!(
log::info!("expected size of the header atom={:?}", output_buf_len); "real size of the header atom={:?}",
pos
);
println!(
"expected size of the header atom={:?}",
output_buf_len
);
assert_eq!(pos, output_buf_len); assert_eq!(pos, output_buf_len);
}; };
log::info!("pushing the header atom={:?}", gst_buffer); log::info!(
"pushing the header atom={:?}",
gst_buffer
);
// Create the catalog track with a single segment. // Create the catalog track with a single segment.
let mut init_track = state.broadcast.clone().create_track("0.mp4").map_err(|_| gst::FlowError::Error)?; let mut init_track = state
let init_segment = init_track.create_segment(segment::Info { .broadcast
sequence: VarInt::ZERO, .clone()
priority: 0, .create_track("0.mp4")
expires: None, .map_err(|_| gst::FlowError::Error)?;
}).map_err(|_| gst::FlowError::Error)?; let init_segment = init_track
.create_segment(segment::Info {
sequence: VarInt::ZERO,
priority: 0,
expires: None,
})
.map_err(|_| gst::FlowError::Error)?;
// Create a single fragment, optionally setting the size // Create a single fragment, optionally setting the size
let mut init_fragment = init_segment.final_fragment(VarInt::ZERO).map_err(|_| gst::FlowError::Error)?; let mut init_fragment = init_segment
.final_fragment(VarInt::ZERO)
.map_err(|_| gst::FlowError::Error)?;
let buffer_map = gst_buffer.map_readable().map_err(|_| gst::FlowError::Error)?; let buffer_map = gst_buffer
.map_readable()
.map_err(|_| gst::FlowError::Error)?;
// Create a a Vec<u8> object from the data slice // Create a a Vec<u8> object from the data slice
let bytes = buffer_map.as_slice().to_vec(); let bytes = buffer_map.as_slice().to_vec();
init_fragment.chunk(bytes.into()).map_err(|_| gst::FlowError::Error)?; init_fragment
.chunk(bytes.into())
.map_err(|_| gst::FlowError::Error)?;
// We're going to parse the moov box. // We're going to parse the moov box.
// We have to read the moov box header to correctly advance the cursor for the mp4 crate. // We have to read the moov box header to correctly advance the cursor for the mp4 crate.
let mut moov_reader = Cursor::new(moov_atom.atom_bytes.clone()); let mut moov_reader =
let moov_header = mp4::BoxHeader::read(&mut moov_reader).map_err(|_| gst::FlowError::Error)?; Cursor::new(moov_atom.atom_bytes.clone());
let moov_header =
mp4::BoxHeader::read(&mut moov_reader)
.map_err(|_| gst::FlowError::Error)?;
// Parse the moov box so we can detect the timescales for each track. // Parse the moov box so we can detect the timescales for each track.
let moov = mp4::MoovBox::read_box(&mut moov_reader, moov_header.size).map_err(|_| gst::FlowError::Error)?; let moov = mp4::MoovBox::read_box(
&mut moov_reader,
moov_header.size,
)
.map_err(|_| gst::FlowError::Error)?;
let mut tracks = HashMap::new(); let mut tracks = HashMap::new();
for trak in &moov.traks { for trak in &moov.traks {
let id = trak.tkhd.track_id; let id = trak.tkhd.track_id;
let name = format!("{}.m4s", id); let name = format!("{}.m4s", id);
let timescale = track_timescale(&moov, id); let timescale = track_timescale(&moov, id);
// Store the track publisher in a map so we can update it later. // Store the track publisher in a map so we can update it later.
let track = broadcast.create_track(&name).map_err(|_| gst::FlowError::Error)?; let track = broadcast
.create_track(&name)
.map_err(|_| gst::FlowError::Error)?;
let track = Track::new(track, timescale); let track = Track::new(track, timescale);
tracks.insert(id, track); tracks.insert(id, track);
} }
let mut catalog = broadcast.create_track(".catalog").map_err(|_| gst::FlowError::Error)?; let mut catalog = broadcast
.create_track(".catalog")
.map_err(|_| gst::FlowError::Error)?;
// Create the catalog track // Create the catalog track
Self::serve_catalog(&mut catalog, &init_track.name, &moov).map_err(|_| gst::FlowError::Error)?; Self::serve_catalog(
&mut catalog,
&init_track.name,
&moov,
)
.map_err(|_| gst::FlowError::Error)?;
state.tracks = Some(tracks); state.tracks = Some(tracks);
state.init = Some(init_track); state.init = Some(init_track);
state.catalog = Some(catalog); state.catalog = Some(catalog);
} }
_ => { _ => {
log::warn!("Received moov without ftype"); log::warn!("Received moov without ftype");
}, }
} }
}, }
ATOM_TYPE_MOOF => { ATOM_TYPE_MOOF => {
log::info!("pushing the moof_atom={:?}", atom); log::info!("pushing the moof_atom={:?}", atom);
@@ -417,17 +505,23 @@ impl GST {
}; };
let mut reader = Cursor::new(atom.atom_bytes.clone()); let mut reader = Cursor::new(atom.atom_bytes.clone());
let header = mp4::BoxHeader::read(&mut reader).map_err(|_| gst::FlowError::Error)?; let header = mp4::BoxHeader::read(&mut reader)
let moof = mp4::MoofBox::read_box(&mut reader, header.size).map_err(|_| gst::FlowError::Error)?; .map_err(|_| gst::FlowError::Error)?;
let moof = mp4::MoofBox::read_box(&mut reader, header.size)
.map_err(|_| gst::FlowError::Error)?;
// Process the moof. // Process the moof.
let fragment = Fragment::new(moof).map_err(|_| gst::FlowError::Error)?; let fragment = Fragment::new(moof)
.map_err(|_| gst::FlowError::Error)?;
// Get the track for this moof. // Get the track for this moof.
let track = tracks.get_mut(&fragment.track).context("failed to find track").map_err(|_| gst::FlowError::Error)?; let track = tracks
.get_mut(&fragment.track)
.context("failed to find track")
.map_err(|_| gst::FlowError::Error)?;
// Save the track ID for the next iteration, which must be a mdat. // Save the track ID for the next iteration, which must be a mdat.
if current.clone().is_none(){ if current.clone().is_none() {
log::error!("multiple moof atoms"); log::error!("multiple moof atoms");
return Err(gst::FlowError::Error); return Err(gst::FlowError::Error);
} }
@@ -435,15 +529,21 @@ impl GST {
current.replace(fragment.track); current.replace(fragment.track);
// Publish the moof header, creating a new segment if it's a keyframe. // Publish the moof header, creating a new segment if it's a keyframe.
track.header(atom.atom_bytes.clone(), fragment).context("failed to publish moof").map_err(|_| gst::FlowError::Error)?; track
.header(atom.atom_bytes.clone(), fragment)
}, .context("failed to publish moof")
.map_err(|_| gst::FlowError::Error)?;
}
ATOM_TYPE_MDAT => { ATOM_TYPE_MDAT => {
log::info!("pushing the mdat_atom={:?}", atom); log::info!("pushing the mdat_atom={:?}", atom);
// Get the track ID from the previous moof. // Get the track ID from the previous moof.
let track = state.current.take().context("missing moof").map_err(|_| gst::FlowError::Error)?; let track = state
.current
.take()
.context("missing moof")
.map_err(|_| gst::FlowError::Error)?;
let tracks = if let Some(tracks) = &mut state.tracks { let tracks = if let Some(tracks) = &mut state.tracks {
tracks tracks
} else { } else {
@@ -451,19 +551,25 @@ impl GST {
return Err(gst::FlowError::Error); return Err(gst::FlowError::Error);
}; };
let track = tracks.get_mut(&track).context("failed to find track").map_err(|_| gst::FlowError::Error)?; let track = tracks
.get_mut(&track)
.context("failed to find track")
.map_err(|_| gst::FlowError::Error)?;
// Publish the mdat atom. // Publish the mdat atom.
track.data(atom.atom_bytes.clone()).context("failed to publish mdat").map_err(|_| gst::FlowError::Error)?; track
}, .data(atom.atom_bytes.clone())
.context("failed to publish mdat")
.map_err(|_| gst::FlowError::Error)?;
}
_ => { _ => {
log::warn!("Unknown atom type {:?}", atom); log::warn!("Unknown atom type {:?}", atom);
}, }
} }
}, }
None => break, None => break,
} }
}; }
Ok(gst::FlowSuccess::Ok) Ok(gst::FlowSuccess::Ok)
}) })
@@ -473,266 +579,308 @@ impl GST {
.build(), .build(),
); );
pipeline.set_state(gst::State::Playing)?;
let bus = pipeline
.bus()
.expect("Pipeline without bus. Shouldn't happen!");
for msg in bus.iter_timed(gst::ClockTime::NONE) {
use gst::MessageView;
match msg.view() {
MessageView::Eos(..) => {
println!("EOS");
break;
}
MessageView::Error(err) => {
pipeline.set_state(gst::State::Null)?;
eprintln!(
"Got error from {}: {} ({})",
msg.src()
.map(|s| String::from(s.path_string()))
.unwrap_or_else(|| "None".into()),
err.error(),
err.debug().unwrap_or_else(|| "".into()),
);
break;
}
_ => (),
}
}
pipeline.set_state(gst::State::Null)?;
Ok(()) Ok(())
} }
//Copied from https://github.com/kixelated/moq-rs/blob/f64c2e894e4f09873aeef4620dbf20aaabe5d12e/moq-pub/src/media.rs#L127 //Copied from https://github.com/kixelated/moq-rs/blob/f64c2e894e4f09873aeef4620dbf20aaabe5d12e/moq-pub/src/media.rs#L127
fn serve_catalog( fn serve_catalog(
track: &mut track::Publisher, track: &mut track::Publisher,
init_track_name: &str, init_track_name: &str,
moov: &mp4::MoovBox, moov: &mp4::MoovBox,
) -> Result<(), anyhow::Error> { ) -> Result<(), anyhow::Error> {
let segment = track.create_segment(segment::Info { println!("serving the catalog");
sequence: VarInt::ZERO,
priority: 0,
expires: None,
})?;
let mut tracks = Vec::new(); let segment = track.create_segment(segment::Info {
sequence: VarInt::ZERO,
priority: 0,
expires: None,
})?;
for trak in &moov.traks { let mut tracks = Vec::new();
let mut track = json!({
"container": "mp4",
"init_track": init_track_name,
"data_track": format!("{}.m4s", trak.tkhd.track_id),
});
let stsd = &trak.mdia.minf.stbl.stsd; for trak in &moov.traks {
if let Some(avc1) = &stsd.avc1 { let mut track = json!({
// avc1[.PPCCLL] "container": "mp4",
// "init_track": init_track_name,
// let profile = 0x64; "data_track": format!("{}.m4s", trak.tkhd.track_id),
// let constraints = 0x00; });
// let level = 0x1f;
let profile = avc1.avcc.avc_profile_indication;
let constraints = avc1.avcc.profile_compatibility; // Not 100% certain here, but it's 0x00 on my current test video
let level = avc1.avcc.avc_level_indication;
let width = avc1.width; let stsd = &trak.mdia.minf.stbl.stsd;
let height = avc1.height; if let Some(avc1) = &stsd.avc1 {
// avc1[.PPCCLL]
//
// let profile = 0x64;
// let constraints = 0x00;
// let level = 0x1f;
let profile = avc1.avcc.avc_profile_indication;
let constraints = avc1.avcc.profile_compatibility; // Not 100% certain here, but it's 0x00 on my current test video
let level = avc1.avcc.avc_level_indication;
let codec = rfc6381_codec::Codec::avc1(profile, constraints, level); let width = avc1.width;
let codec_str = codec.to_string(); let height = avc1.height;
track["kind"] = json!("video"); let codec = rfc6381_codec::Codec::avc1(profile, constraints, level);
track["codec"] = json!(codec_str); let codec_str = codec.to_string();
track["width"] = json!(width);
track["height"] = json!(height);
} else if let Some(_hev1) = &stsd.hev1 {
// TODO https://github.com/gpac/mp4box.js/blob/325741b592d910297bf609bc7c400fc76101077b/src/box-codecs.js#L106
anyhow::bail!("HEVC not yet supported")
} else if let Some(mp4a) = &stsd.mp4a {
let desc = &mp4a
.esds
.as_ref()
.context("missing esds box for MP4a")?
.es_desc
.dec_config;
let codec_str = format!("mp4a.{:02x}.{}", desc.object_type_indication, desc.dec_specific.profile); track["kind"] = json!("video");
track["codec"] = json!(codec_str);
track["width"] = json!(width);
track["height"] = json!(height);
} else if let Some(_hev1) = &stsd.hev1 {
// TODO https://github.com/gpac/mp4box.js/blob/325741b592d910297bf609bc7c400fc76101077b/src/box-codecs.js#L106
anyhow::bail!("HEVC not yet supported")
} else if let Some(mp4a) = &stsd.mp4a {
let desc = &mp4a
.esds
.as_ref()
.context("missing esds box for MP4a")?
.es_desc
.dec_config;
track["kind"] = json!("audio"); let codec_str = format!(
track["codec"] = json!(codec_str); "mp4a.{:02x}.{}",
track["channel_count"] = json!(mp4a.channelcount); desc.object_type_indication, desc.dec_specific.profile
track["sample_rate"] = json!(mp4a.samplerate.value()); );
track["sample_size"] = json!(mp4a.samplesize);
let bitrate = max(desc.max_bitrate, desc.avg_bitrate); track["kind"] = json!("audio");
if bitrate > 0 { track["codec"] = json!(codec_str);
track["bit_rate"] = json!(bitrate); track["channel_count"] = json!(mp4a.channelcount);
} track["sample_rate"] = json!(mp4a.samplerate.value());
} else if let Some(vp09) = &stsd.vp09 { track["sample_size"] = json!(mp4a.samplesize);
// https://github.com/gpac/mp4box.js/blob/325741b592d910297bf609bc7c400fc76101077b/src/box-codecs.js#L238
let vpcc = &vp09.vpcc;
let codec_str = format!("vp09.0.{:02x}.{:02x}.{:02x}", vpcc.profile, vpcc.level, vpcc.bit_depth);
track["kind"] = json!("video"); let bitrate = max(desc.max_bitrate, desc.avg_bitrate);
track["codec"] = json!(codec_str); if bitrate > 0 {
track["width"] = json!(vp09.width); // no idea if this needs to be multiplied track["bit_rate"] = json!(bitrate);
track["height"] = json!(vp09.height); // no idea if this needs to be multiplied }
} else if let Some(vp09) = &stsd.vp09 {
// https://github.com/gpac/mp4box.js/blob/325741b592d910297bf609bc7c400fc76101077b/src/box-codecs.js#L238
let vpcc = &vp09.vpcc;
let codec_str = format!(
"vp09.0.{:02x}.{:02x}.{:02x}",
vpcc.profile, vpcc.level, vpcc.bit_depth
);
// TODO Test if this actually works; I'm just guessing based on mp4box.js track["kind"] = json!("video");
anyhow::bail!("VP9 not yet supported") track["codec"] = json!(codec_str);
} else { track["width"] = json!(vp09.width); // no idea if this needs to be multiplied
// TODO add av01 support: https://github.com/gpac/mp4box.js/blob/325741b592d910297bf609bc7c400fc76101077b/src/box-codecs.js#L251 track["height"] = json!(vp09.height); // no idea if this needs to be multiplied
anyhow::bail!("unknown codec for track: {}", trak.tkhd.track_id);
}
tracks.push(track); // TODO Test if this actually works; I'm just guessing based on mp4box.js
} anyhow::bail!("VP9 not yet supported")
} else {
// TODO add av01 support: https://github.com/gpac/mp4box.js/blob/325741b592d910297bf609bc7c400fc76101077b/src/box-codecs.js#L251
anyhow::bail!("unknown codec for track: {}", trak.tkhd.track_id);
}
let catalog = json!({ tracks.push(track);
"tracks": tracks }
});
let catalog_str = serde_json::to_string_pretty(&catalog)?; let catalog = json!({
log::info!("catalog: {}", catalog_str); "tracks": tracks
});
// Create a single fragment for the segment. let catalog_str = serde_json::to_string_pretty(&catalog)?;
let mut fragment = segment.final_fragment(VarInt::ZERO)?; log::info!("catalog: {}", catalog_str);
// Add the segment and add the fragment. // Create a single fragment for the segment.
fragment.chunk(catalog_str.into())?; let mut fragment = segment.final_fragment(VarInt::ZERO)?;
Ok(()) // Add the segment and add the fragment.
} fragment.chunk(catalog_str.into())?;
Ok(())
}
} }
struct Track { struct Track {
// The track we're producing // The track we're producing
track: track::Publisher, track: track::Publisher,
// The current segment // The current segment
current: Option<fragment::Publisher>, current: Option<fragment::Publisher>,
// The number of units per second. // The number of units per second.
timescale: u64, timescale: u64,
// The number of segments produced. // The number of segments produced.
sequence: u64, sequence: u64,
} }
impl Track { impl Track {
fn new(track: track::Publisher, timescale: u64) -> Self { fn new(track: track::Publisher, timescale: u64) -> Self {
Self { Self {
track, track,
sequence: 0, sequence: 0,
current: None, current: None,
timescale, timescale,
} }
} }
pub fn header(&mut self, raw: Vec<u8>, fragment: Fragment) -> anyhow::Result<()> { pub fn header(&mut self, raw: Vec<u8>, fragment: Fragment) -> anyhow::Result<()> {
if let Some(current) = self.current.as_mut() { if let Some(current) = self.current.as_mut() {
if !fragment.keyframe { if !fragment.keyframe {
// Use the existing segment // Use the existing segment
current.chunk(raw.into())?; current.chunk(raw.into())?;
return Ok(()); return Ok(());
} }
} }
// Otherwise make a new segment // Otherwise make a new segment
// Compute the timestamp in milliseconds. // Compute the timestamp in milliseconds.
// Overflows after 583 million years, so we're fine. // Overflows after 583 million years, so we're fine.
let timestamp: u32 = fragment let timestamp: u32 = fragment
.timestamp(self.timescale) .timestamp(self.timescale)
.as_millis() .as_millis()
.try_into() .try_into()
.context("timestamp too large")?; .context("timestamp too large")?;
// Create a new segment. // Create a new segment.
let segment = self.track.create_segment(segment::Info { let segment = self.track.create_segment(segment::Info {
sequence: VarInt::try_from(self.sequence).context("sequence too large")?, sequence: VarInt::try_from(self.sequence).context("sequence too large")?,
// Newer segments are higher priority // Newer segments are higher priority
priority: u32::MAX.checked_sub(timestamp).context("priority too large")?, priority: u32::MAX
.checked_sub(timestamp)
.context("priority too large")?,
// Delete segments after 10s. // Delete segments after 10s.
expires: Some(time::Duration::from_secs(10)), expires: Some(time::Duration::from_secs(10)),
})?; })?;
// Create a single fragment for the segment that we will keep appending. // Create a single fragment for the segment that we will keep appending.
let mut fragment = segment.final_fragment(VarInt::ZERO)?; let mut fragment = segment.final_fragment(VarInt::ZERO)?;
self.sequence += 1; self.sequence += 1;
// Insert the raw atom into the segment. // Insert the raw atom into the segment.
fragment.chunk(raw.into())?; fragment.chunk(raw.into())?;
// Save for the next iteration // Save for the next iteration
self.current = Some(fragment); self.current = Some(fragment);
Ok(()) Ok(())
} }
pub fn data(&mut self, raw: Vec<u8>) -> anyhow::Result<()> { pub fn data(&mut self, raw: Vec<u8>) -> anyhow::Result<()> {
let fragment = self.current.as_mut().context("missing current fragment")?; let fragment = self.current.as_mut().context("missing current fragment")?;
fragment.chunk(raw.into())?; fragment.chunk(raw.into())?;
Ok(()) Ok(())
} }
} }
struct Fragment { struct Fragment {
// The track for this fragment. // The track for this fragment.
track: u32, track: u32,
// The timestamp of the first sample in this fragment, in timescale units. // The timestamp of the first sample in this fragment, in timescale units.
timestamp: u64, timestamp: u64,
// True if this fragment is a keyframe. // True if this fragment is a keyframe.
keyframe: bool, keyframe: bool,
} }
impl Fragment { impl Fragment {
fn new(moof: mp4::MoofBox) -> anyhow::Result<Self> { fn new(moof: mp4::MoofBox) -> anyhow::Result<Self> {
// We can't split the mdat atom, so this is impossible to support // We can't split the mdat atom, so this is impossible to support
anyhow::ensure!(moof.trafs.len() == 1, "multiple tracks per moof atom"); anyhow::ensure!(moof.trafs.len() == 1, "multiple tracks per moof atom");
let track = moof.trafs[0].tfhd.track_id; let track = moof.trafs[0].tfhd.track_id;
// Parse the moof to get some timing information to sleep. // Parse the moof to get some timing information to sleep.
let timestamp = sample_timestamp(&moof).expect("couldn't find timestamp"); let timestamp = sample_timestamp(&moof).expect("couldn't find timestamp");
// Detect if we should start a new segment. // Detect if we should start a new segment.
let keyframe = sample_keyframe(&moof); let keyframe = sample_keyframe(&moof);
Ok(Self { Ok(Self {
track, track,
timestamp, timestamp,
keyframe, keyframe,
}) })
} }
// Convert from timescale units to a duration. // Convert from timescale units to a duration.
fn timestamp(&self, timescale: u64) -> time::Duration { fn timestamp(&self, timescale: u64) -> time::Duration {
time::Duration::from_millis(1000 * self.timestamp / timescale) time::Duration::from_millis(1000 * self.timestamp / timescale)
} }
} }
fn sample_timestamp(moof: &mp4::MoofBox) -> Option<u64> { fn sample_timestamp(moof: &mp4::MoofBox) -> Option<u64> {
Some(moof.trafs.first()?.tfdt.as_ref()?.base_media_decode_time) Some(moof.trafs.first()?.tfdt.as_ref()?.base_media_decode_time)
} }
fn sample_keyframe(moof: &mp4::MoofBox) -> bool { fn sample_keyframe(moof: &mp4::MoofBox) -> bool {
for traf in &moof.trafs { for traf in &moof.trafs {
// TODO trak default flags if this is None // TODO trak default flags if this is None
let default_flags = traf.tfhd.default_sample_flags.unwrap_or_default(); let default_flags = traf.tfhd.default_sample_flags.unwrap_or_default();
let trun = match &traf.trun { let trun = match &traf.trun {
Some(t) => t, Some(t) => t,
None => return false, None => return false,
}; };
for i in 0..trun.sample_count { for i in 0..trun.sample_count {
let mut flags = match trun.sample_flags.get(i as usize) { let mut flags = match trun.sample_flags.get(i as usize) {
Some(f) => *f, Some(f) => *f,
None => default_flags, None => default_flags,
}; };
if i == 0 && trun.first_sample_flags.is_some() { if i == 0 && trun.first_sample_flags.is_some() {
flags = trun.first_sample_flags.unwrap(); flags = trun.first_sample_flags.unwrap();
} }
// https://chromium.googlesource.com/chromium/src/media/+/master/formats/mp4/track_run_iterator.cc#177 // https://chromium.googlesource.com/chromium/src/media/+/master/formats/mp4/track_run_iterator.cc#177
let keyframe = (flags >> 24) & 0x3 == 0x2; // kSampleDependsOnNoOther let keyframe = (flags >> 24) & 0x3 == 0x2; // kSampleDependsOnNoOther
let non_sync = (flags >> 16) & 0x1 == 0x1; // kSampleIsNonSyncSample let non_sync = (flags >> 16) & 0x1 == 0x1; // kSampleIsNonSyncSample
if keyframe && !non_sync { if keyframe && !non_sync {
return true; return true;
} }
} }
} }
false false
} }
// Find the timescale for the given track. // Find the timescale for the given track.
fn track_timescale(moov: &mp4::MoovBox, track_id: u32) -> u64 { fn track_timescale(moov: &mp4::MoovBox, track_id: u32) -> u64 {
let trak = moov let trak = moov
.traks .traks
.iter() .iter()
.find(|trak| trak.tkhd.track_id == track_id) .find(|trak| trak.tkhd.track_id == track_id)
.expect("failed to find trak"); .expect("failed to find trak");
trak.mdia.mdhd.timescale as u64 trak.mdia.mdhd.timescale as u64
} }