123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378 |
- var Stream = require('../utils/stream.js');
- var m2ts = require('../m2ts/m2ts.js');
- var codecs = require('../codecs/index.js');
- var AudioSegmentStream = require('./audio-segment-stream.js');
- var VideoSegmentStream = require('./video-segment-stream.js');
- var trackInfo = require('../mp4/track-decode-info.js');
- var isLikelyAacData = require('../aac/utils').isLikelyAacData;
- var AdtsStream = require('../codecs/adts');
- var AacStream = require('../aac/index');
- var clock = require('../utils/clock');
- var createPipeline = function(object) {
- object.prototype = new Stream();
- object.prototype.init.call(object);
- return object;
- };
- var tsPipeline = function(options) {
- var
- pipeline = {
- type: 'ts',
- tracks: {
- audio: null,
- video: null
- },
- packet: new m2ts.TransportPacketStream(),
- parse: new m2ts.TransportParseStream(),
- elementary: new m2ts.ElementaryStream(),
- timestampRollover: new m2ts.TimestampRolloverStream(),
- adts: new codecs.Adts(),
- h264: new codecs.h264.H264Stream(),
- captionStream: new m2ts.CaptionStream(options),
- metadataStream: new m2ts.MetadataStream()
- };
- pipeline.headOfPipeline = pipeline.packet;
- // Transport Stream
- pipeline.packet
- .pipe(pipeline.parse)
- .pipe(pipeline.elementary)
- .pipe(pipeline.timestampRollover);
- // H264
- pipeline.timestampRollover
- .pipe(pipeline.h264);
- // Hook up CEA-608/708 caption stream
- pipeline.h264
- .pipe(pipeline.captionStream);
- pipeline.timestampRollover
- .pipe(pipeline.metadataStream);
- // ADTS
- pipeline.timestampRollover
- .pipe(pipeline.adts);
- pipeline.elementary.on('data', function(data) {
- if (data.type !== 'metadata') {
- return;
- }
- for (var i = 0; i < data.tracks.length; i++) {
- if (!pipeline.tracks[data.tracks[i].type]) {
- pipeline.tracks[data.tracks[i].type] = data.tracks[i];
- pipeline.tracks[data.tracks[i].type].timelineStartInfo.baseMediaDecodeTime = options.baseMediaDecodeTime;
- }
- }
- if (pipeline.tracks.video && !pipeline.videoSegmentStream) {
- pipeline.videoSegmentStream = new VideoSegmentStream(pipeline.tracks.video, options);
- pipeline.videoSegmentStream.on('timelineStartInfo', function(timelineStartInfo) {
- if (pipeline.tracks.audio && !options.keepOriginalTimestamps) {
- pipeline.audioSegmentStream.setEarliestDts(timelineStartInfo.dts - options.baseMediaDecodeTime);
- }
- });
- pipeline.videoSegmentStream.on('timingInfo',
- pipeline.trigger.bind(pipeline, 'videoTimingInfo'));
- pipeline.videoSegmentStream.on('data', function(data) {
- pipeline.trigger('data', {
- type: 'video',
- data: data
- });
- });
- pipeline.videoSegmentStream.on('done',
- pipeline.trigger.bind(pipeline, 'done'));
- pipeline.videoSegmentStream.on('partialdone',
- pipeline.trigger.bind(pipeline, 'partialdone'));
- pipeline.videoSegmentStream.on('endedtimeline',
- pipeline.trigger.bind(pipeline, 'endedtimeline'));
- pipeline.h264
- .pipe(pipeline.videoSegmentStream);
- }
- if (pipeline.tracks.audio && !pipeline.audioSegmentStream) {
- pipeline.audioSegmentStream = new AudioSegmentStream(pipeline.tracks.audio, options);
- pipeline.audioSegmentStream.on('data', function(data) {
- pipeline.trigger('data', {
- type: 'audio',
- data: data
- });
- });
- pipeline.audioSegmentStream.on('done',
- pipeline.trigger.bind(pipeline, 'done'));
- pipeline.audioSegmentStream.on('partialdone',
- pipeline.trigger.bind(pipeline, 'partialdone'));
- pipeline.audioSegmentStream.on('endedtimeline',
- pipeline.trigger.bind(pipeline, 'endedtimeline'));
- pipeline.audioSegmentStream.on('timingInfo',
- pipeline.trigger.bind(pipeline, 'audioTimingInfo'));
- pipeline.adts
- .pipe(pipeline.audioSegmentStream);
- }
- // emit pmt info
- pipeline.trigger('trackinfo', {
- hasAudio: !!pipeline.tracks.audio,
- hasVideo: !!pipeline.tracks.video
- });
- });
- pipeline.captionStream.on('data', function(caption) {
- var timelineStartPts;
- if (pipeline.tracks.video) {
- timelineStartPts = pipeline.tracks.video.timelineStartInfo.pts || 0;
- } else {
- // This will only happen if we encounter caption packets before
- // video data in a segment. This is an unusual/unlikely scenario,
- // so we assume the timeline starts at zero for now.
- timelineStartPts = 0;
- }
- // Translate caption PTS times into second offsets into the
- // video timeline for the segment
- caption.startTime = clock.metadataTsToSeconds(caption.startPts, timelineStartPts, options.keepOriginalTimestamps);
- caption.endTime = clock.metadataTsToSeconds(caption.endPts, timelineStartPts, options.keepOriginalTimestamps);
- pipeline.trigger('caption', caption);
- });
- pipeline = createPipeline(pipeline);
- pipeline.metadataStream.on('data', pipeline.trigger.bind(pipeline, 'id3Frame'));
- return pipeline;
- };
- var aacPipeline = function(options) {
- var
- pipeline = {
- type: 'aac',
- tracks: {
- audio: null
- },
- metadataStream: new m2ts.MetadataStream(),
- aacStream: new AacStream(),
- audioRollover: new m2ts.TimestampRolloverStream('audio'),
- timedMetadataRollover: new m2ts.TimestampRolloverStream('timed-metadata'),
- adtsStream: new AdtsStream(true)
- };
- // set up the parsing pipeline
- pipeline.headOfPipeline = pipeline.aacStream;
- pipeline.aacStream
- .pipe(pipeline.audioRollover)
- .pipe(pipeline.adtsStream);
- pipeline.aacStream
- .pipe(pipeline.timedMetadataRollover)
- .pipe(pipeline.metadataStream);
- pipeline.metadataStream.on('timestamp', function(frame) {
- pipeline.aacStream.setTimestamp(frame.timeStamp);
- });
- pipeline.aacStream.on('data', function(data) {
- if ((data.type !== 'timed-metadata' && data.type !== 'audio') || pipeline.audioSegmentStream) {
- return;
- }
- pipeline.tracks.audio = pipeline.tracks.audio || {
- timelineStartInfo: {
- baseMediaDecodeTime: options.baseMediaDecodeTime
- },
- codec: 'adts',
- type: 'audio'
- };
- // hook up the audio segment stream to the first track with aac data
- pipeline.audioSegmentStream = new AudioSegmentStream(pipeline.tracks.audio, options);
- pipeline.audioSegmentStream.on('data', function(data) {
- pipeline.trigger('data', {
- type: 'audio',
- data: data
- });
- });
- pipeline.audioSegmentStream.on('partialdone',
- pipeline.trigger.bind(pipeline, 'partialdone'));
- pipeline.audioSegmentStream.on('done', pipeline.trigger.bind(pipeline, 'done'));
- pipeline.audioSegmentStream.on('endedtimeline',
- pipeline.trigger.bind(pipeline, 'endedtimeline'));
- pipeline.audioSegmentStream.on('timingInfo',
- pipeline.trigger.bind(pipeline, 'audioTimingInfo'));
- // Set up the final part of the audio pipeline
- pipeline.adtsStream
- .pipe(pipeline.audioSegmentStream);
- pipeline.trigger('trackinfo', {
- hasAudio: !!pipeline.tracks.audio,
- hasVideo: !!pipeline.tracks.video
- });
- });
- // set the pipeline up as a stream before binding to get access to the trigger function
- pipeline = createPipeline(pipeline);
- pipeline.metadataStream.on('data', pipeline.trigger.bind(pipeline, 'id3Frame'));
- return pipeline;
- };
- var setupPipelineListeners = function(pipeline, transmuxer) {
- pipeline.on('data', transmuxer.trigger.bind(transmuxer, 'data'));
- pipeline.on('done', transmuxer.trigger.bind(transmuxer, 'done'));
- pipeline.on('partialdone', transmuxer.trigger.bind(transmuxer, 'partialdone'));
- pipeline.on('endedtimeline', transmuxer.trigger.bind(transmuxer, 'endedtimeline'));
- pipeline.on('audioTimingInfo', transmuxer.trigger.bind(transmuxer, 'audioTimingInfo'));
- pipeline.on('videoTimingInfo', transmuxer.trigger.bind(transmuxer, 'videoTimingInfo'));
- pipeline.on('trackinfo', transmuxer.trigger.bind(transmuxer, 'trackinfo'));
- pipeline.on('id3Frame', function(event) {
- // add this to every single emitted segment even though it's only needed for the first
- event.dispatchType = pipeline.metadataStream.dispatchType;
- // keep original time, can be adjusted if needed at a higher level
- event.cueTime = clock.videoTsToSeconds(event.pts);
- transmuxer.trigger('id3Frame', event);
- });
- pipeline.on('caption', function(event) {
- transmuxer.trigger('caption', event);
- });
- };
- var Transmuxer = function(options) {
- var
- pipeline = null,
- hasFlushed = true;
- options = options || {};
- Transmuxer.prototype.init.call(this);
- options.baseMediaDecodeTime = options.baseMediaDecodeTime || 0;
- this.push = function(bytes) {
- if (hasFlushed) {
- var isAac = isLikelyAacData(bytes);
- if (isAac && (!pipeline || pipeline.type !== 'aac')) {
- pipeline = aacPipeline(options);
- setupPipelineListeners(pipeline, this);
- } else if (!isAac && (!pipeline || pipeline.type !== 'ts')) {
- pipeline = tsPipeline(options);
- setupPipelineListeners(pipeline, this);
- }
- hasFlushed = false;
- }
- pipeline.headOfPipeline.push(bytes);
- };
- this.flush = function() {
- if (!pipeline) {
- return;
- }
- hasFlushed = true;
- pipeline.headOfPipeline.flush();
- };
- this.partialFlush = function() {
- if (!pipeline) {
- return;
- }
- pipeline.headOfPipeline.partialFlush();
- };
- this.endTimeline = function() {
- if (!pipeline) {
- return;
- }
- pipeline.headOfPipeline.endTimeline();
- };
- this.reset = function() {
- if (!pipeline) {
- return;
- }
- pipeline.headOfPipeline.reset();
- };
- this.setBaseMediaDecodeTime = function(baseMediaDecodeTime) {
- if (!options.keepOriginalTimestamps) {
- options.baseMediaDecodeTime = baseMediaDecodeTime;
- }
- if (!pipeline) {
- return;
- }
- if (pipeline.tracks.audio) {
- pipeline.tracks.audio.timelineStartInfo.dts = undefined;
- pipeline.tracks.audio.timelineStartInfo.pts = undefined;
- trackInfo.clearDtsInfo(pipeline.tracks.audio);
- if (pipeline.audioRollover) {
- pipeline.audioRollover.discontinuity();
- }
- }
- if (pipeline.tracks.video) {
- if (pipeline.videoSegmentStream) {
- pipeline.videoSegmentStream.gopCache_ = [];
- }
- pipeline.tracks.video.timelineStartInfo.dts = undefined;
- pipeline.tracks.video.timelineStartInfo.pts = undefined;
- trackInfo.clearDtsInfo(pipeline.tracks.video);
- // pipeline.captionStream.reset();
- }
- if (pipeline.timestampRollover) {
- pipeline.timestampRollover.discontinuity();
- }
- };
- this.setRemux = function(val) {
- options.remux = val;
- if (pipeline && pipeline.coalesceStream) {
- pipeline.coalesceStream.setRemux(val);
- }
- };
- this.setAudioAppendStart = function(audioAppendStart) {
- if (!pipeline || !pipeline.tracks.audio || !pipeline.audioSegmentStream) {
- return;
- }
- pipeline.audioSegmentStream.setAudioAppendStart(audioAppendStart);
- };
- // TODO GOP alignment support
- // Support may be a bit trickier than with full segment appends, as GOPs may be split
- // and processed in a more granular fashion
- this.alignGopsWith = function(gopsToAlignWith) {
- return;
- };
- };
- Transmuxer.prototype = new Stream();
- module.exports = Transmuxer;
|