diff options
Diffstat (limited to 'priv/static/vendored/video.js/video.es.js')
-rw-r--r-- | priv/static/vendored/video.js/video.es.js | 52506 |
1 files changed, 52506 insertions, 0 deletions
diff --git a/priv/static/vendored/video.js/video.es.js b/priv/static/vendored/video.js/video.es.js new file mode 100644 index 0000000..2cc0305 --- /dev/null +++ b/priv/static/vendored/video.js/video.es.js @@ -0,0 +1,52506 @@ +/** + * @license + * Video.js 8.12.0 <http://videojs.com/> + * Copyright Brightcove, Inc. <https://www.brightcove.com/> + * Available under Apache License Version 2.0 + * <https://github.com/videojs/video.js/blob/main/LICENSE> + * + * Includes vtt.js <https://github.com/mozilla/vtt.js> + * Available under Apache License Version 2.0 + * <https://github.com/mozilla/vtt.js/blob/main/LICENSE> + */ + +import window$1 from 'global/window'; +import document from 'global/document'; +import keycode from 'keycode'; +import safeParseTuple from 'safe-json-parse/tuple'; +import XHR from '@videojs/xhr'; +import vtt from 'videojs-vtt.js'; +import _extends from '@babel/runtime/helpers/extends'; +import _resolveUrl from '@videojs/vhs-utils/es/resolve-url.js'; +import { Parser } from 'm3u8-parser'; +import { DEFAULT_VIDEO_CODEC, DEFAULT_AUDIO_CODEC, parseCodecs, muxerSupportsCodec, browserSupportsCodec, translateLegacyCodec, codecsFromDefault, isAudioCodec, getMimeForCodec } from '@videojs/vhs-utils/es/codecs.js'; +import { simpleTypeFromSourceType } from '@videojs/vhs-utils/es/media-types.js'; +import { isArrayBufferView, concatTypedArrays, stringToBytes, toUint8 } from '@videojs/vhs-utils/es/byte-helpers'; +import { generateSidxKey, parseUTCTiming, parse, addSidxSegmentsToPlaylist } from 'mpd-parser'; +import parseSidx from 'mux.js/lib/tools/parse-sidx'; +import { getId3Offset } from '@videojs/vhs-utils/es/id3-helpers'; +import { detectContainerForBytes, isLikelyFmp4MediaSegment } from '@videojs/vhs-utils/es/containers'; +import { ONE_SECOND_IN_TS } from 'mux.js/lib/utils/clock'; + +var version$6 = "8.12.0"; + +/** + * An Object that contains lifecycle hooks as keys which point to an array + * of functions that are run when a lifecycle is triggered + * + * @private + */ +const hooks_ = {}; + +/** + * Get a list of hooks for a specific lifecycle + * + * @param {string} type + * the lifecycle to get hooks from + * + * @param {Function|Function[]} [fn] + * Optionally add a hook (or hooks) to the lifecycle that your are getting. + * + * @return {Array} + * an array of hooks, or an empty array if there are none. + */ +const hooks = function (type, fn) { + hooks_[type] = hooks_[type] || []; + if (fn) { + hooks_[type] = hooks_[type].concat(fn); + } + return hooks_[type]; +}; + +/** + * Add a function hook to a specific videojs lifecycle. + * + * @param {string} type + * the lifecycle to hook the function to. + * + * @param {Function|Function[]} + * The function or array of functions to attach. + */ +const hook = function (type, fn) { + hooks(type, fn); +}; + +/** + * Remove a hook from a specific videojs lifecycle. + * + * @param {string} type + * the lifecycle that the function hooked to + * + * @param {Function} fn + * The hooked function to remove + * + * @return {boolean} + * The function that was removed or undef + */ +const removeHook = function (type, fn) { + const index = hooks(type).indexOf(fn); + if (index <= -1) { + return false; + } + hooks_[type] = hooks_[type].slice(); + hooks_[type].splice(index, 1); + return true; +}; + +/** + * Add a function hook that will only run once to a specific videojs lifecycle. + * + * @param {string} type + * the lifecycle to hook the function to. + * + * @param {Function|Function[]} + * The function or array of functions to attach. + */ +const hookOnce = function (type, fn) { + hooks(type, [].concat(fn).map(original => { + const wrapper = (...args) => { + removeHook(type, wrapper); + return original(...args); + }; + return wrapper; + })); +}; + +/** + * @file fullscreen-api.js + * @module fullscreen-api + */ + +/** + * Store the browser-specific methods for the fullscreen API. + * + * @type {Object} + * @see [Specification]{@link https://fullscreen.spec.whatwg.org} + * @see [Map Approach From Screenfull.js]{@link https://github.com/sindresorhus/screenfull.js} + */ +const FullscreenApi = { + prefixed: true +}; + +// browser API methods +const apiMap = [['requestFullscreen', 'exitFullscreen', 'fullscreenElement', 'fullscreenEnabled', 'fullscreenchange', 'fullscreenerror', 'fullscreen'], +// WebKit +['webkitRequestFullscreen', 'webkitExitFullscreen', 'webkitFullscreenElement', 'webkitFullscreenEnabled', 'webkitfullscreenchange', 'webkitfullscreenerror', '-webkit-full-screen']]; +const specApi = apiMap[0]; +let browserApi; + +// determine the supported set of functions +for (let i = 0; i < apiMap.length; i++) { + // check for exitFullscreen function + if (apiMap[i][1] in document) { + browserApi = apiMap[i]; + break; + } +} + +// map the browser API names to the spec API names +if (browserApi) { + for (let i = 0; i < browserApi.length; i++) { + FullscreenApi[specApi[i]] = browserApi[i]; + } + FullscreenApi.prefixed = browserApi[0] !== specApi[0]; +} + +/** + * @file create-logger.js + * @module create-logger + */ + +// This is the private tracking variable for the logging history. +let history = []; + +/** + * Log messages to the console and history based on the type of message + * + * @private + * @param {string} name + * The name of the console method to use. + * + * @param {Object} log + * The arguments to be passed to the matching console method. + * + * @param {string} [styles] + * styles for name + */ +const LogByTypeFactory = (name, log, styles) => (type, level, args) => { + const lvl = log.levels[level]; + const lvlRegExp = new RegExp(`^(${lvl})$`); + let resultName = name; + if (type !== 'log') { + // Add the type to the front of the message when it's not "log". + args.unshift(type.toUpperCase() + ':'); + } + if (styles) { + resultName = `%c${name}`; + args.unshift(styles); + } + + // Add console prefix after adding to history. + args.unshift(resultName + ':'); + + // Add a clone of the args at this point to history. + if (history) { + history.push([].concat(args)); + + // only store 1000 history entries + const splice = history.length - 1000; + history.splice(0, splice > 0 ? splice : 0); + } + + // If there's no console then don't try to output messages, but they will + // still be stored in history. + if (!window$1.console) { + return; + } + + // Was setting these once outside of this function, but containing them + // in the function makes it easier to test cases where console doesn't exist + // when the module is executed. + let fn = window$1.console[type]; + if (!fn && type === 'debug') { + // Certain browsers don't have support for console.debug. For those, we + // should default to the closest comparable log. + fn = window$1.console.info || window$1.console.log; + } + + // Bail out if there's no console or if this type is not allowed by the + // current logging level. + if (!fn || !lvl || !lvlRegExp.test(type)) { + return; + } + fn[Array.isArray(args) ? 'apply' : 'call'](window$1.console, args); +}; +function createLogger$1(name, delimiter = ':', styles = '') { + // This is the private tracking variable for logging level. + let level = 'info'; + + // the curried logByType bound to the specific log and history + let logByType; + + /** + * Logs plain debug messages. Similar to `console.log`. + * + * Due to [limitations](https://github.com/jsdoc3/jsdoc/issues/955#issuecomment-313829149) + * of our JSDoc template, we cannot properly document this as both a function + * and a namespace, so its function signature is documented here. + * + * #### Arguments + * ##### *args + * *[] + * + * Any combination of values that could be passed to `console.log()`. + * + * #### Return Value + * + * `undefined` + * + * @namespace + * @param {...*} args + * One or more messages or objects that should be logged. + */ + const log = function (...args) { + logByType('log', level, args); + }; + + // This is the logByType helper that the logging methods below use + logByType = LogByTypeFactory(name, log, styles); + + /** + * Create a new subLogger which chains the old name to the new name. + * + * For example, doing `videojs.log.createLogger('player')` and then using that logger will log the following: + * ```js + * mylogger('foo'); + * // > VIDEOJS: player: foo + * ``` + * + * @param {string} subName + * The name to add call the new logger + * @param {string} [subDelimiter] + * Optional delimiter + * @param {string} [subStyles] + * Optional styles + * @return {Object} + */ + log.createLogger = (subName, subDelimiter, subStyles) => { + const resultDelimiter = subDelimiter !== undefined ? subDelimiter : delimiter; + const resultStyles = subStyles !== undefined ? subStyles : styles; + const resultName = `${name} ${resultDelimiter} ${subName}`; + return createLogger$1(resultName, resultDelimiter, resultStyles); + }; + + /** + * Create a new logger. + * + * @param {string} newName + * The name for the new logger + * @param {string} [newDelimiter] + * Optional delimiter + * @param {string} [newStyles] + * Optional styles + * @return {Object} + */ + log.createNewLogger = (newName, newDelimiter, newStyles) => { + return createLogger$1(newName, newDelimiter, newStyles); + }; + + /** + * Enumeration of available logging levels, where the keys are the level names + * and the values are `|`-separated strings containing logging methods allowed + * in that logging level. These strings are used to create a regular expression + * matching the function name being called. + * + * Levels provided by Video.js are: + * + * - `off`: Matches no calls. Any value that can be cast to `false` will have + * this effect. The most restrictive. + * - `all`: Matches only Video.js-provided functions (`debug`, `log`, + * `log.warn`, and `log.error`). + * - `debug`: Matches `log.debug`, `log`, `log.warn`, and `log.error` calls. + * - `info` (default): Matches `log`, `log.warn`, and `log.error` calls. + * - `warn`: Matches `log.warn` and `log.error` calls. + * - `error`: Matches only `log.error` calls. + * + * @type {Object} + */ + log.levels = { + all: 'debug|log|warn|error', + off: '', + debug: 'debug|log|warn|error', + info: 'log|warn|error', + warn: 'warn|error', + error: 'error', + DEFAULT: level + }; + + /** + * Get or set the current logging level. + * + * If a string matching a key from {@link module:log.levels} is provided, acts + * as a setter. + * + * @param {'all'|'debug'|'info'|'warn'|'error'|'off'} [lvl] + * Pass a valid level to set a new logging level. + * + * @return {string} + * The current logging level. + */ + log.level = lvl => { + if (typeof lvl === 'string') { + if (!log.levels.hasOwnProperty(lvl)) { + throw new Error(`"${lvl}" in not a valid log level`); + } + level = lvl; + } + return level; + }; + + /** + * Returns an array containing everything that has been logged to the history. + * + * This array is a shallow clone of the internal history record. However, its + * contents are _not_ cloned; so, mutating objects inside this array will + * mutate them in history. + * + * @return {Array} + */ + log.history = () => history ? [].concat(history) : []; + + /** + * Allows you to filter the history by the given logger name + * + * @param {string} fname + * The name to filter by + * + * @return {Array} + * The filtered list to return + */ + log.history.filter = fname => { + return (history || []).filter(historyItem => { + // if the first item in each historyItem includes `fname`, then it's a match + return new RegExp(`.*${fname}.*`).test(historyItem[0]); + }); + }; + + /** + * Clears the internal history tracking, but does not prevent further history + * tracking. + */ + log.history.clear = () => { + if (history) { + history.length = 0; + } + }; + + /** + * Disable history tracking if it is currently enabled. + */ + log.history.disable = () => { + if (history !== null) { + history.length = 0; + history = null; + } + }; + + /** + * Enable history tracking if it is currently disabled. + */ + log.history.enable = () => { + if (history === null) { + history = []; + } + }; + + /** + * Logs error messages. Similar to `console.error`. + * + * @param {...*} args + * One or more messages or objects that should be logged as an error + */ + log.error = (...args) => logByType('error', level, args); + + /** + * Logs warning messages. Similar to `console.warn`. + * + * @param {...*} args + * One or more messages or objects that should be logged as a warning. + */ + log.warn = (...args) => logByType('warn', level, args); + + /** + * Logs debug messages. Similar to `console.debug`, but may also act as a comparable + * log if `console.debug` is not available + * + * @param {...*} args + * One or more messages or objects that should be logged as debug. + */ + log.debug = (...args) => logByType('debug', level, args); + return log; +} + +/** + * @file log.js + * @module log + */ +const log$1 = createLogger$1('VIDEOJS'); +const createLogger = log$1.createLogger; + +/** + * @file obj.js + * @module obj + */ + +/** + * @callback obj:EachCallback + * + * @param {*} value + * The current key for the object that is being iterated over. + * + * @param {string} key + * The current key-value for object that is being iterated over + */ + +/** + * @callback obj:ReduceCallback + * + * @param {*} accum + * The value that is accumulating over the reduce loop. + * + * @param {*} value + * The current key for the object that is being iterated over. + * + * @param {string} key + * The current key-value for object that is being iterated over + * + * @return {*} + * The new accumulated value. + */ +const toString = Object.prototype.toString; + +/** + * Get the keys of an Object + * + * @param {Object} + * The Object to get the keys from + * + * @return {string[]} + * An array of the keys from the object. Returns an empty array if the + * object passed in was invalid or had no keys. + * + * @private + */ +const keys = function (object) { + return isObject(object) ? Object.keys(object) : []; +}; + +/** + * Array-like iteration for objects. + * + * @param {Object} object + * The object to iterate over + * + * @param {obj:EachCallback} fn + * The callback function which is called for each key in the object. + */ +function each(object, fn) { + keys(object).forEach(key => fn(object[key], key)); +} + +/** + * Array-like reduce for objects. + * + * @param {Object} object + * The Object that you want to reduce. + * + * @param {Function} fn + * A callback function which is called for each key in the object. It + * receives the accumulated value and the per-iteration value and key + * as arguments. + * + * @param {*} [initial = 0] + * Starting value + * + * @return {*} + * The final accumulated value. + */ +function reduce(object, fn, initial = 0) { + return keys(object).reduce((accum, key) => fn(accum, object[key], key), initial); +} + +/** + * Returns whether a value is an object of any kind - including DOM nodes, + * arrays, regular expressions, etc. Not functions, though. + * + * This avoids the gotcha where using `typeof` on a `null` value + * results in `'object'`. + * + * @param {Object} value + * @return {boolean} + */ +function isObject(value) { + return !!value && typeof value === 'object'; +} + +/** + * Returns whether an object appears to be a "plain" object - that is, a + * direct instance of `Object`. + * + * @param {Object} value + * @return {boolean} + */ +function isPlain(value) { + return isObject(value) && toString.call(value) === '[object Object]' && value.constructor === Object; +} + +/** + * Merge two objects recursively. + * + * Performs a deep merge like + * {@link https://lodash.com/docs/4.17.10#merge|lodash.merge}, but only merges + * plain objects (not arrays, elements, or anything else). + * + * Non-plain object values will be copied directly from the right-most + * argument. + * + * @param {Object[]} sources + * One or more objects to merge into a new object. + * + * @return {Object} + * A new object that is the merged result of all sources. + */ +function merge$1(...sources) { + const result = {}; + sources.forEach(source => { + if (!source) { + return; + } + each(source, (value, key) => { + if (!isPlain(value)) { + result[key] = value; + return; + } + if (!isPlain(result[key])) { + result[key] = {}; + } + result[key] = merge$1(result[key], value); + }); + }); + return result; +} + +/** + * Returns an array of values for a given object + * + * @param {Object} source - target object + * @return {Array<unknown>} - object values + */ +function values(source = {}) { + const result = []; + for (const key in source) { + if (source.hasOwnProperty(key)) { + const value = source[key]; + result.push(value); + } + } + return result; +} + +/** + * Object.defineProperty but "lazy", which means that the value is only set after + * it is retrieved the first time, rather than being set right away. + * + * @param {Object} obj the object to set the property on + * @param {string} key the key for the property to set + * @param {Function} getValue the function used to get the value when it is needed. + * @param {boolean} setter whether a setter should be allowed or not + */ +function defineLazyProperty(obj, key, getValue, setter = true) { + const set = value => Object.defineProperty(obj, key, { + value, + enumerable: true, + writable: true + }); + const options = { + configurable: true, + enumerable: true, + get() { + const value = getValue(); + set(value); + return value; + } + }; + if (setter) { + options.set = set; + } + return Object.defineProperty(obj, key, options); +} + +var Obj = /*#__PURE__*/Object.freeze({ + __proto__: null, + each: each, + reduce: reduce, + isObject: isObject, + isPlain: isPlain, + merge: merge$1, + values: values, + defineLazyProperty: defineLazyProperty +}); + +/** + * @file browser.js + * @module browser + */ + +/** + * Whether or not this device is an iPod. + * + * @static + * @type {Boolean} + */ +let IS_IPOD = false; + +/** + * The detected iOS version - or `null`. + * + * @static + * @type {string|null} + */ +let IOS_VERSION = null; + +/** + * Whether or not this is an Android device. + * + * @static + * @type {Boolean} + */ +let IS_ANDROID = false; + +/** + * The detected Android version - or `null` if not Android or indeterminable. + * + * @static + * @type {number|string|null} + */ +let ANDROID_VERSION; + +/** + * Whether or not this is Mozilla Firefox. + * + * @static + * @type {Boolean} + */ +let IS_FIREFOX = false; + +/** + * Whether or not this is Microsoft Edge. + * + * @static + * @type {Boolean} + */ +let IS_EDGE = false; + +/** + * Whether or not this is any Chromium Browser + * + * @static + * @type {Boolean} + */ +let IS_CHROMIUM = false; + +/** + * Whether or not this is any Chromium browser that is not Edge. + * + * This will also be `true` for Chrome on iOS, which will have different support + * as it is actually Safari under the hood. + * + * Deprecated, as the behaviour to not match Edge was to prevent Legacy Edge's UA matching. + * IS_CHROMIUM should be used instead. + * "Chromium but not Edge" could be explicitly tested with IS_CHROMIUM && !IS_EDGE + * + * @static + * @deprecated + * @type {Boolean} + */ +let IS_CHROME = false; + +/** + * The detected Chromium version - or `null`. + * + * @static + * @type {number|null} + */ +let CHROMIUM_VERSION = null; + +/** + * The detected Google Chrome version - or `null`. + * This has always been the _Chromium_ version, i.e. would return on Chromium Edge. + * Deprecated, use CHROMIUM_VERSION instead. + * + * @static + * @deprecated + * @type {number|null} + */ +let CHROME_VERSION = null; + +/** + * The detected Internet Explorer version - or `null`. + * + * @static + * @deprecated + * @type {number|null} + */ +let IE_VERSION = null; + +/** + * Whether or not this is desktop Safari. + * + * @static + * @type {Boolean} + */ +let IS_SAFARI = false; + +/** + * Whether or not this is a Windows machine. + * + * @static + * @type {Boolean} + */ +let IS_WINDOWS = false; + +/** + * Whether or not this device is an iPad. + * + * @static + * @type {Boolean} + */ +let IS_IPAD = false; + +/** + * Whether or not this device is an iPhone. + * + * @static + * @type {Boolean} + */ +// The Facebook app's UIWebView identifies as both an iPhone and iPad, so +// to identify iPhones, we need to exclude iPads. +// http://artsy.github.io/blog/2012/10/18/the-perils-of-ios-user-agent-sniffing/ +let IS_IPHONE = false; + +/** + * Whether or not this is a Tizen device. + * + * @static + * @type {Boolean} + */ +let IS_TIZEN = false; + +/** + * Whether or not this is a WebOS device. + * + * @static + * @type {Boolean} + */ +let IS_WEBOS = false; + +/** + * Whether or not this is a Smart TV (Tizen or WebOS) device. + * + * @static + * @type {Boolean} + */ +let IS_SMART_TV = false; + +/** + * Whether or not this device is touch-enabled. + * + * @static + * @const + * @type {Boolean} + */ +const TOUCH_ENABLED = Boolean(isReal() && ('ontouchstart' in window$1 || window$1.navigator.maxTouchPoints || window$1.DocumentTouch && window$1.document instanceof window$1.DocumentTouch)); +const UAD = window$1.navigator && window$1.navigator.userAgentData; +if (UAD && UAD.platform && UAD.brands) { + // If userAgentData is present, use it instead of userAgent to avoid warnings + // Currently only implemented on Chromium + // userAgentData does not expose Android version, so ANDROID_VERSION remains `null` + + IS_ANDROID = UAD.platform === 'Android'; + IS_EDGE = Boolean(UAD.brands.find(b => b.brand === 'Microsoft Edge')); + IS_CHROMIUM = Boolean(UAD.brands.find(b => b.brand === 'Chromium')); + IS_CHROME = !IS_EDGE && IS_CHROMIUM; + CHROMIUM_VERSION = CHROME_VERSION = (UAD.brands.find(b => b.brand === 'Chromium') || {}).version || null; + IS_WINDOWS = UAD.platform === 'Windows'; +} + +// If the browser is not Chromium, either userAgentData is not present which could be an old Chromium browser, +// or it's a browser that has added userAgentData since that we don't have tests for yet. In either case, +// the checks need to be made agiainst the regular userAgent string. +if (!IS_CHROMIUM) { + const USER_AGENT = window$1.navigator && window$1.navigator.userAgent || ''; + IS_IPOD = /iPod/i.test(USER_AGENT); + IOS_VERSION = function () { + const match = USER_AGENT.match(/OS (\d+)_/i); + if (match && match[1]) { + return match[1]; + } + return null; + }(); + IS_ANDROID = /Android/i.test(USER_AGENT); + ANDROID_VERSION = function () { + // This matches Android Major.Minor.Patch versions + // ANDROID_VERSION is Major.Minor as a Number, if Minor isn't available, then only Major is returned + const match = USER_AGENT.match(/Android (\d+)(?:\.(\d+))?(?:\.(\d+))*/i); + if (!match) { + return null; + } + const major = match[1] && parseFloat(match[1]); + const minor = match[2] && parseFloat(match[2]); + if (major && minor) { + return parseFloat(match[1] + '.' + match[2]); + } else if (major) { + return major; + } + return null; + }(); + IS_FIREFOX = /Firefox/i.test(USER_AGENT); + IS_EDGE = /Edg/i.test(USER_AGENT); + IS_CHROMIUM = /Chrome/i.test(USER_AGENT) || /CriOS/i.test(USER_AGENT); + IS_CHROME = !IS_EDGE && IS_CHROMIUM; + CHROMIUM_VERSION = CHROME_VERSION = function () { + const match = USER_AGENT.match(/(Chrome|CriOS)\/(\d+)/); + if (match && match[2]) { + return parseFloat(match[2]); + } + return null; + }(); + IE_VERSION = function () { + const result = /MSIE\s(\d+)\.\d/.exec(USER_AGENT); + let version = result && parseFloat(result[1]); + if (!version && /Trident\/7.0/i.test(USER_AGENT) && /rv:11.0/.test(USER_AGENT)) { + // IE 11 has a different user agent string than other IE versions + version = 11.0; + } + return version; + }(); + IS_TIZEN = /Tizen/i.test(USER_AGENT); + IS_WEBOS = /Web0S/i.test(USER_AGENT); + IS_SMART_TV = IS_TIZEN || IS_WEBOS; + IS_SAFARI = /Safari/i.test(USER_AGENT) && !IS_CHROME && !IS_ANDROID && !IS_EDGE && !IS_SMART_TV; + IS_WINDOWS = /Windows/i.test(USER_AGENT); + IS_IPAD = /iPad/i.test(USER_AGENT) || IS_SAFARI && TOUCH_ENABLED && !/iPhone/i.test(USER_AGENT); + IS_IPHONE = /iPhone/i.test(USER_AGENT) && !IS_IPAD; +} + +/** + * Whether or not this is an iOS device. + * + * @static + * @const + * @type {Boolean} + */ +const IS_IOS = IS_IPHONE || IS_IPAD || IS_IPOD; + +/** + * Whether or not this is any flavor of Safari - including iOS. + * + * @static + * @const + * @type {Boolean} + */ +const IS_ANY_SAFARI = (IS_SAFARI || IS_IOS) && !IS_CHROME; + +var browser = /*#__PURE__*/Object.freeze({ + __proto__: null, + get IS_IPOD () { return IS_IPOD; }, + get IOS_VERSION () { return IOS_VERSION; }, + get IS_ANDROID () { return IS_ANDROID; }, + get ANDROID_VERSION () { return ANDROID_VERSION; }, + get IS_FIREFOX () { return IS_FIREFOX; }, + get IS_EDGE () { return IS_EDGE; }, + get IS_CHROMIUM () { return IS_CHROMIUM; }, + get IS_CHROME () { return IS_CHROME; }, + get CHROMIUM_VERSION () { return CHROMIUM_VERSION; }, + get CHROME_VERSION () { return CHROME_VERSION; }, + get IE_VERSION () { return IE_VERSION; }, + get IS_SAFARI () { return IS_SAFARI; }, + get IS_WINDOWS () { return IS_WINDOWS; }, + get IS_IPAD () { return IS_IPAD; }, + get IS_IPHONE () { return IS_IPHONE; }, + get IS_TIZEN () { return IS_TIZEN; }, + get IS_WEBOS () { return IS_WEBOS; }, + get IS_SMART_TV () { return IS_SMART_TV; }, + TOUCH_ENABLED: TOUCH_ENABLED, + IS_IOS: IS_IOS, + IS_ANY_SAFARI: IS_ANY_SAFARI +}); + +/** + * @file dom.js + * @module dom + */ + +/** + * Detect if a value is a string with any non-whitespace characters. + * + * @private + * @param {string} str + * The string to check + * + * @return {boolean} + * Will be `true` if the string is non-blank, `false` otherwise. + * + */ +function isNonBlankString(str) { + // we use str.trim as it will trim any whitespace characters + // from the front or back of non-whitespace characters. aka + // Any string that contains non-whitespace characters will + // still contain them after `trim` but whitespace only strings + // will have a length of 0, failing this check. + return typeof str === 'string' && Boolean(str.trim()); +} + +/** + * Throws an error if the passed string has whitespace. This is used by + * class methods to be relatively consistent with the classList API. + * + * @private + * @param {string} str + * The string to check for whitespace. + * + * @throws {Error} + * Throws an error if there is whitespace in the string. + */ +function throwIfWhitespace(str) { + // str.indexOf instead of regex because str.indexOf is faster performance wise. + if (str.indexOf(' ') >= 0) { + throw new Error('class has illegal whitespace characters'); + } +} + +/** + * Whether the current DOM interface appears to be real (i.e. not simulated). + * + * @return {boolean} + * Will be `true` if the DOM appears to be real, `false` otherwise. + */ +function isReal() { + // Both document and window will never be undefined thanks to `global`. + return document === window$1.document; +} + +/** + * Determines, via duck typing, whether or not a value is a DOM element. + * + * @param {*} value + * The value to check. + * + * @return {boolean} + * Will be `true` if the value is a DOM element, `false` otherwise. + */ +function isEl(value) { + return isObject(value) && value.nodeType === 1; +} + +/** + * Determines if the current DOM is embedded in an iframe. + * + * @return {boolean} + * Will be `true` if the DOM is embedded in an iframe, `false` + * otherwise. + */ +function isInFrame() { + // We need a try/catch here because Safari will throw errors when attempting + // to get either `parent` or `self` + try { + return window$1.parent !== window$1.self; + } catch (x) { + return true; + } +} + +/** + * Creates functions to query the DOM using a given method. + * + * @private + * @param {string} method + * The method to create the query with. + * + * @return {Function} + * The query method + */ +function createQuerier(method) { + return function (selector, context) { + if (!isNonBlankString(selector)) { + return document[method](null); + } + if (isNonBlankString(context)) { + context = document.querySelector(context); + } + const ctx = isEl(context) ? context : document; + return ctx[method] && ctx[method](selector); + }; +} + +/** + * Creates an element and applies properties, attributes, and inserts content. + * + * @param {string} [tagName='div'] + * Name of tag to be created. + * + * @param {Object} [properties={}] + * Element properties to be applied. + * + * @param {Object} [attributes={}] + * Element attributes to be applied. + * + * @param {ContentDescriptor} [content] + * A content descriptor object. + * + * @return {Element} + * The element that was created. + */ +function createEl(tagName = 'div', properties = {}, attributes = {}, content) { + const el = document.createElement(tagName); + Object.getOwnPropertyNames(properties).forEach(function (propName) { + const val = properties[propName]; + + // Handle textContent since it's not supported everywhere and we have a + // method for it. + if (propName === 'textContent') { + textContent(el, val); + } else if (el[propName] !== val || propName === 'tabIndex') { + el[propName] = val; + } + }); + Object.getOwnPropertyNames(attributes).forEach(function (attrName) { + el.setAttribute(attrName, attributes[attrName]); + }); + if (content) { + appendContent(el, content); + } + return el; +} + +/** + * Injects text into an element, replacing any existing contents entirely. + * + * @param {HTMLElement} el + * The element to add text content into + * + * @param {string} text + * The text content to add. + * + * @return {Element} + * The element with added text content. + */ +function textContent(el, text) { + if (typeof el.textContent === 'undefined') { + el.innerText = text; + } else { + el.textContent = text; + } + return el; +} + +/** + * Insert an element as the first child node of another + * + * @param {Element} child + * Element to insert + * + * @param {Element} parent + * Element to insert child into + */ +function prependTo(child, parent) { + if (parent.firstChild) { + parent.insertBefore(child, parent.firstChild); + } else { + parent.appendChild(child); + } +} + +/** + * Check if an element has a class name. + * + * @param {Element} element + * Element to check + * + * @param {string} classToCheck + * Class name to check for + * + * @return {boolean} + * Will be `true` if the element has a class, `false` otherwise. + * + * @throws {Error} + * Throws an error if `classToCheck` has white space. + */ +function hasClass(element, classToCheck) { + throwIfWhitespace(classToCheck); + return element.classList.contains(classToCheck); +} + +/** + * Add a class name to an element. + * + * @param {Element} element + * Element to add class name to. + * + * @param {...string} classesToAdd + * One or more class name to add. + * + * @return {Element} + * The DOM element with the added class name. + */ +function addClass(element, ...classesToAdd) { + element.classList.add(...classesToAdd.reduce((prev, current) => prev.concat(current.split(/\s+/)), [])); + return element; +} + +/** + * Remove a class name from an element. + * + * @param {Element} element + * Element to remove a class name from. + * + * @param {...string} classesToRemove + * One or more class name to remove. + * + * @return {Element} + * The DOM element with class name removed. + */ +function removeClass(element, ...classesToRemove) { + // Protect in case the player gets disposed + if (!element) { + log$1.warn("removeClass was called with an element that doesn't exist"); + return null; + } + element.classList.remove(...classesToRemove.reduce((prev, current) => prev.concat(current.split(/\s+/)), [])); + return element; +} + +/** + * The callback definition for toggleClass. + * + * @callback module:dom~PredicateCallback + * @param {Element} element + * The DOM element of the Component. + * + * @param {string} classToToggle + * The `className` that wants to be toggled + * + * @return {boolean|undefined} + * If `true` is returned, the `classToToggle` will be added to the + * `element`. If `false`, the `classToToggle` will be removed from + * the `element`. If `undefined`, the callback will be ignored. + */ + +/** + * Adds or removes a class name to/from an element depending on an optional + * condition or the presence/absence of the class name. + * + * @param {Element} element + * The element to toggle a class name on. + * + * @param {string} classToToggle + * The class that should be toggled. + * + * @param {boolean|module:dom~PredicateCallback} [predicate] + * See the return value for {@link module:dom~PredicateCallback} + * + * @return {Element} + * The element with a class that has been toggled. + */ +function toggleClass(element, classToToggle, predicate) { + if (typeof predicate === 'function') { + predicate = predicate(element, classToToggle); + } + if (typeof predicate !== 'boolean') { + predicate = undefined; + } + classToToggle.split(/\s+/).forEach(className => element.classList.toggle(className, predicate)); + return element; +} + +/** + * Apply attributes to an HTML element. + * + * @param {Element} el + * Element to add attributes to. + * + * @param {Object} [attributes] + * Attributes to be applied. + */ +function setAttributes(el, attributes) { + Object.getOwnPropertyNames(attributes).forEach(function (attrName) { + const attrValue = attributes[attrName]; + if (attrValue === null || typeof attrValue === 'undefined' || attrValue === false) { + el.removeAttribute(attrName); + } else { + el.setAttribute(attrName, attrValue === true ? '' : attrValue); + } + }); +} + +/** + * Get an element's attribute values, as defined on the HTML tag. + * + * Attributes are not the same as properties. They're defined on the tag + * or with setAttribute. + * + * @param {Element} tag + * Element from which to get tag attributes. + * + * @return {Object} + * All attributes of the element. Boolean attributes will be `true` or + * `false`, others will be strings. + */ +function getAttributes(tag) { + const obj = {}; + + // known boolean attributes + // we can check for matching boolean properties, but not all browsers + // and not all tags know about these attributes, so, we still want to check them manually + const knownBooleans = ['autoplay', 'controls', 'playsinline', 'loop', 'muted', 'default', 'defaultMuted']; + if (tag && tag.attributes && tag.attributes.length > 0) { + const attrs = tag.attributes; + for (let i = attrs.length - 1; i >= 0; i--) { + const attrName = attrs[i].name; + /** @type {boolean|string} */ + let attrVal = attrs[i].value; + + // check for known booleans + // the matching element property will return a value for typeof + if (knownBooleans.includes(attrName)) { + // the value of an included boolean attribute is typically an empty + // string ('') which would equal false if we just check for a false value. + // we also don't want support bad code like autoplay='false' + attrVal = attrVal !== null ? true : false; + } + obj[attrName] = attrVal; + } + } + return obj; +} + +/** + * Get the value of an element's attribute. + * + * @param {Element} el + * A DOM element. + * + * @param {string} attribute + * Attribute to get the value of. + * + * @return {string} + * The value of the attribute. + */ +function getAttribute(el, attribute) { + return el.getAttribute(attribute); +} + +/** + * Set the value of an element's attribute. + * + * @param {Element} el + * A DOM element. + * + * @param {string} attribute + * Attribute to set. + * + * @param {string} value + * Value to set the attribute to. + */ +function setAttribute(el, attribute, value) { + el.setAttribute(attribute, value); +} + +/** + * Remove an element's attribute. + * + * @param {Element} el + * A DOM element. + * + * @param {string} attribute + * Attribute to remove. + */ +function removeAttribute(el, attribute) { + el.removeAttribute(attribute); +} + +/** + * Attempt to block the ability to select text. + */ +function blockTextSelection() { + document.body.focus(); + document.onselectstart = function () { + return false; + }; +} + +/** + * Turn off text selection blocking. + */ +function unblockTextSelection() { + document.onselectstart = function () { + return true; + }; +} + +/** + * Identical to the native `getBoundingClientRect` function, but ensures that + * the method is supported at all (it is in all browsers we claim to support) + * and that the element is in the DOM before continuing. + * + * This wrapper function also shims properties which are not provided by some + * older browsers (namely, IE8). + * + * Additionally, some browsers do not support adding properties to a + * `ClientRect`/`DOMRect` object; so, we shallow-copy it with the standard + * properties (except `x` and `y` which are not widely supported). This helps + * avoid implementations where keys are non-enumerable. + * + * @param {Element} el + * Element whose `ClientRect` we want to calculate. + * + * @return {Object|undefined} + * Always returns a plain object - or `undefined` if it cannot. + */ +function getBoundingClientRect(el) { + if (el && el.getBoundingClientRect && el.parentNode) { + const rect = el.getBoundingClientRect(); + const result = {}; + ['bottom', 'height', 'left', 'right', 'top', 'width'].forEach(k => { + if (rect[k] !== undefined) { + result[k] = rect[k]; + } + }); + if (!result.height) { + result.height = parseFloat(computedStyle(el, 'height')); + } + if (!result.width) { + result.width = parseFloat(computedStyle(el, 'width')); + } + return result; + } +} + +/** + * Represents the position of a DOM element on the page. + * + * @typedef {Object} module:dom~Position + * + * @property {number} left + * Pixels to the left. + * + * @property {number} top + * Pixels from the top. + */ + +/** + * Get the position of an element in the DOM. + * + * Uses `getBoundingClientRect` technique from John Resig. + * + * @see http://ejohn.org/blog/getboundingclientrect-is-awesome/ + * + * @param {Element} el + * Element from which to get offset. + * + * @return {module:dom~Position} + * The position of the element that was passed in. + */ +function findPosition(el) { + if (!el || el && !el.offsetParent) { + return { + left: 0, + top: 0, + width: 0, + height: 0 + }; + } + const width = el.offsetWidth; + const height = el.offsetHeight; + let left = 0; + let top = 0; + while (el.offsetParent && el !== document[FullscreenApi.fullscreenElement]) { + left += el.offsetLeft; + top += el.offsetTop; + el = el.offsetParent; + } + return { + left, + top, + width, + height + }; +} + +/** + * Represents x and y coordinates for a DOM element or mouse pointer. + * + * @typedef {Object} module:dom~Coordinates + * + * @property {number} x + * x coordinate in pixels + * + * @property {number} y + * y coordinate in pixels + */ + +/** + * Get the pointer position within an element. + * + * The base on the coordinates are the bottom left of the element. + * + * @param {Element} el + * Element on which to get the pointer position on. + * + * @param {Event} event + * Event object. + * + * @return {module:dom~Coordinates} + * A coordinates object corresponding to the mouse position. + * + */ +function getPointerPosition(el, event) { + const translated = { + x: 0, + y: 0 + }; + if (IS_IOS) { + let item = el; + while (item && item.nodeName.toLowerCase() !== 'html') { + const transform = computedStyle(item, 'transform'); + if (/^matrix/.test(transform)) { + const values = transform.slice(7, -1).split(/,\s/).map(Number); + translated.x += values[4]; + translated.y += values[5]; + } else if (/^matrix3d/.test(transform)) { + const values = transform.slice(9, -1).split(/,\s/).map(Number); + translated.x += values[12]; + translated.y += values[13]; + } + item = item.parentNode; + } + } + const position = {}; + const boxTarget = findPosition(event.target); + const box = findPosition(el); + const boxW = box.width; + const boxH = box.height; + let offsetY = event.offsetY - (box.top - boxTarget.top); + let offsetX = event.offsetX - (box.left - boxTarget.left); + if (event.changedTouches) { + offsetX = event.changedTouches[0].pageX - box.left; + offsetY = event.changedTouches[0].pageY + box.top; + if (IS_IOS) { + offsetX -= translated.x; + offsetY -= translated.y; + } + } + position.y = 1 - Math.max(0, Math.min(1, offsetY / boxH)); + position.x = Math.max(0, Math.min(1, offsetX / boxW)); + return position; +} + +/** + * Determines, via duck typing, whether or not a value is a text node. + * + * @param {*} value + * Check if this value is a text node. + * + * @return {boolean} + * Will be `true` if the value is a text node, `false` otherwise. + */ +function isTextNode(value) { + return isObject(value) && value.nodeType === 3; +} + +/** + * Empties the contents of an element. + * + * @param {Element} el + * The element to empty children from + * + * @return {Element} + * The element with no children + */ +function emptyEl(el) { + while (el.firstChild) { + el.removeChild(el.firstChild); + } + return el; +} + +/** + * This is a mixed value that describes content to be injected into the DOM + * via some method. It can be of the following types: + * + * Type | Description + * -----------|------------- + * `string` | The value will be normalized into a text node. + * `Element` | The value will be accepted as-is. + * `Text` | A TextNode. The value will be accepted as-is. + * `Array` | A one-dimensional array of strings, elements, text nodes, or functions. These functions should return a string, element, or text node (any other return value, like an array, will be ignored). + * `Function` | A function, which is expected to return a string, element, text node, or array - any of the other possible values described above. This means that a content descriptor could be a function that returns an array of functions, but those second-level functions must return strings, elements, or text nodes. + * + * @typedef {string|Element|Text|Array|Function} ContentDescriptor + */ + +/** + * Normalizes content for eventual insertion into the DOM. + * + * This allows a wide range of content definition methods, but helps protect + * from falling into the trap of simply writing to `innerHTML`, which could + * be an XSS concern. + * + * The content for an element can be passed in multiple types and + * combinations, whose behavior is as follows: + * + * @param {ContentDescriptor} content + * A content descriptor value. + * + * @return {Array} + * All of the content that was passed in, normalized to an array of + * elements or text nodes. + */ +function normalizeContent(content) { + // First, invoke content if it is a function. If it produces an array, + // that needs to happen before normalization. + if (typeof content === 'function') { + content = content(); + } + + // Next up, normalize to an array, so one or many items can be normalized, + // filtered, and returned. + return (Array.isArray(content) ? content : [content]).map(value => { + // First, invoke value if it is a function to produce a new value, + // which will be subsequently normalized to a Node of some kind. + if (typeof value === 'function') { + value = value(); + } + if (isEl(value) || isTextNode(value)) { + return value; + } + if (typeof value === 'string' && /\S/.test(value)) { + return document.createTextNode(value); + } + }).filter(value => value); +} + +/** + * Normalizes and appends content to an element. + * + * @param {Element} el + * Element to append normalized content to. + * + * @param {ContentDescriptor} content + * A content descriptor value. + * + * @return {Element} + * The element with appended normalized content. + */ +function appendContent(el, content) { + normalizeContent(content).forEach(node => el.appendChild(node)); + return el; +} + +/** + * Normalizes and inserts content into an element; this is identical to + * `appendContent()`, except it empties the element first. + * + * @param {Element} el + * Element to insert normalized content into. + * + * @param {ContentDescriptor} content + * A content descriptor value. + * + * @return {Element} + * The element with inserted normalized content. + */ +function insertContent(el, content) { + return appendContent(emptyEl(el), content); +} + +/** + * Check if an event was a single left click. + * + * @param {MouseEvent} event + * Event object. + * + * @return {boolean} + * Will be `true` if a single left click, `false` otherwise. + */ +function isSingleLeftClick(event) { + // Note: if you create something draggable, be sure to + // call it on both `mousedown` and `mousemove` event, + // otherwise `mousedown` should be enough for a button + + if (event.button === undefined && event.buttons === undefined) { + // Why do we need `buttons` ? + // Because, middle mouse sometimes have this: + // e.button === 0 and e.buttons === 4 + // Furthermore, we want to prevent combination click, something like + // HOLD middlemouse then left click, that would be + // e.button === 0, e.buttons === 5 + // just `button` is not gonna work + + // Alright, then what this block does ? + // this is for chrome `simulate mobile devices` + // I want to support this as well + + return true; + } + if (event.button === 0 && event.buttons === undefined) { + // Touch screen, sometimes on some specific device, `buttons` + // doesn't have anything (safari on ios, blackberry...) + + return true; + } + + // `mouseup` event on a single left click has + // `button` and `buttons` equal to 0 + if (event.type === 'mouseup' && event.button === 0 && event.buttons === 0) { + return true; + } + if (event.button !== 0 || event.buttons !== 1) { + // This is the reason we have those if else block above + // if any special case we can catch and let it slide + // we do it above, when get to here, this definitely + // is-not-left-click + + return false; + } + return true; +} + +/** + * Finds a single DOM element matching `selector` within the optional + * `context` of another DOM element (defaulting to `document`). + * + * @param {string} selector + * A valid CSS selector, which will be passed to `querySelector`. + * + * @param {Element|String} [context=document] + * A DOM element within which to query. Can also be a selector + * string in which case the first matching element will be used + * as context. If missing (or no element matches selector), falls + * back to `document`. + * + * @return {Element|null} + * The element that was found or null. + */ +const $ = createQuerier('querySelector'); + +/** + * Finds a all DOM elements matching `selector` within the optional + * `context` of another DOM element (defaulting to `document`). + * + * @param {string} selector + * A valid CSS selector, which will be passed to `querySelectorAll`. + * + * @param {Element|String} [context=document] + * A DOM element within which to query. Can also be a selector + * string in which case the first matching element will be used + * as context. If missing (or no element matches selector), falls + * back to `document`. + * + * @return {NodeList} + * A element list of elements that were found. Will be empty if none + * were found. + * + */ +const $$ = createQuerier('querySelectorAll'); + +/** + * A safe getComputedStyle. + * + * This is needed because in Firefox, if the player is loaded in an iframe with + * `display:none`, then `getComputedStyle` returns `null`, so, we do a + * null-check to make sure that the player doesn't break in these cases. + * + * @param {Element} el + * The element you want the computed style of + * + * @param {string} prop + * The property name you want + * + * @see https://bugzilla.mozilla.org/show_bug.cgi?id=548397 + */ +function computedStyle(el, prop) { + if (!el || !prop) { + return ''; + } + if (typeof window$1.getComputedStyle === 'function') { + let computedStyleValue; + try { + computedStyleValue = window$1.getComputedStyle(el); + } catch (e) { + return ''; + } + return computedStyleValue ? computedStyleValue.getPropertyValue(prop) || computedStyleValue[prop] : ''; + } + return ''; +} + +/** + * Copy document style sheets to another window. + * + * @param {Window} win + * The window element you want to copy the document style sheets to. + * + */ +function copyStyleSheetsToWindow(win) { + [...document.styleSheets].forEach(styleSheet => { + try { + const cssRules = [...styleSheet.cssRules].map(rule => rule.cssText).join(''); + const style = document.createElement('style'); + style.textContent = cssRules; + win.document.head.appendChild(style); + } catch (e) { + const link = document.createElement('link'); + link.rel = 'stylesheet'; + link.type = styleSheet.type; + // For older Safari this has to be the string; on other browsers setting the MediaList works + link.media = styleSheet.media.mediaText; + link.href = styleSheet.href; + win.document.head.appendChild(link); + } + }); +} + +var Dom = /*#__PURE__*/Object.freeze({ + __proto__: null, + isReal: isReal, + isEl: isEl, + isInFrame: isInFrame, + createEl: createEl, + textContent: textContent, + prependTo: prependTo, + hasClass: hasClass, + addClass: addClass, + removeClass: removeClass, + toggleClass: toggleClass, + setAttributes: setAttributes, + getAttributes: getAttributes, + getAttribute: getAttribute, + setAttribute: setAttribute, + removeAttribute: removeAttribute, + blockTextSelection: blockTextSelection, + unblockTextSelection: unblockTextSelection, + getBoundingClientRect: getBoundingClientRect, + findPosition: findPosition, + getPointerPosition: getPointerPosition, + isTextNode: isTextNode, + emptyEl: emptyEl, + normalizeContent: normalizeContent, + appendContent: appendContent, + insertContent: insertContent, + isSingleLeftClick: isSingleLeftClick, + $: $, + $$: $$, + computedStyle: computedStyle, + copyStyleSheetsToWindow: copyStyleSheetsToWindow +}); + +/** + * @file setup.js - Functions for setting up a player without + * user interaction based on the data-setup `attribute` of the video tag. + * + * @module setup + */ +let _windowLoaded = false; +let videojs$1; + +/** + * Set up any tags that have a data-setup `attribute` when the player is started. + */ +const autoSetup = function () { + if (videojs$1.options.autoSetup === false) { + return; + } + const vids = Array.prototype.slice.call(document.getElementsByTagName('video')); + const audios = Array.prototype.slice.call(document.getElementsByTagName('audio')); + const divs = Array.prototype.slice.call(document.getElementsByTagName('video-js')); + const mediaEls = vids.concat(audios, divs); + + // Check if any media elements exist + if (mediaEls && mediaEls.length > 0) { + for (let i = 0, e = mediaEls.length; i < e; i++) { + const mediaEl = mediaEls[i]; + + // Check if element exists, has getAttribute func. + if (mediaEl && mediaEl.getAttribute) { + // Make sure this player hasn't already been set up. + if (mediaEl.player === undefined) { + const options = mediaEl.getAttribute('data-setup'); + + // Check if data-setup attr exists. + // We only auto-setup if they've added the data-setup attr. + if (options !== null) { + // Create new video.js instance. + videojs$1(mediaEl); + } + } + + // If getAttribute isn't defined, we need to wait for the DOM. + } else { + autoSetupTimeout(1); + break; + } + } + + // No videos were found, so keep looping unless page is finished loading. + } else if (!_windowLoaded) { + autoSetupTimeout(1); + } +}; + +/** + * Wait until the page is loaded before running autoSetup. This will be called in + * autoSetup if `hasLoaded` returns false. + * + * @param {number} wait + * How long to wait in ms + * + * @param {module:videojs} [vjs] + * The videojs library function + */ +function autoSetupTimeout(wait, vjs) { + // Protect against breakage in non-browser environments + if (!isReal()) { + return; + } + if (vjs) { + videojs$1 = vjs; + } + window$1.setTimeout(autoSetup, wait); +} + +/** + * Used to set the internal tracking of window loaded state to true. + * + * @private + */ +function setWindowLoaded() { + _windowLoaded = true; + window$1.removeEventListener('load', setWindowLoaded); +} +if (isReal()) { + if (document.readyState === 'complete') { + setWindowLoaded(); + } else { + /** + * Listen for the load event on window, and set _windowLoaded to true. + * + * We use a standard event listener here to avoid incrementing the GUID + * before any players are created. + * + * @listens load + */ + window$1.addEventListener('load', setWindowLoaded); + } +} + +/** + * @file stylesheet.js + * @module stylesheet + */ + +/** + * Create a DOM style element given a className for it. + * + * @param {string} className + * The className to add to the created style element. + * + * @return {Element} + * The element that was created. + */ +const createStyleElement = function (className) { + const style = document.createElement('style'); + style.className = className; + return style; +}; + +/** + * Add text to a DOM element. + * + * @param {Element} el + * The Element to add text content to. + * + * @param {string} content + * The text to add to the element. + */ +const setTextContent = function (el, content) { + if (el.styleSheet) { + el.styleSheet.cssText = content; + } else { + el.textContent = content; + } +}; + +/** + * @file dom-data.js + * @module dom-data + */ + +/** + * Element Data Store. + * + * Allows for binding data to an element without putting it directly on the + * element. Ex. Event listeners are stored here. + * (also from jsninja.com, slightly modified and updated for closure compiler) + * + * @type {Object} + * @private + */ +var DomData = new WeakMap(); + +/** + * @file guid.js + * @module guid + */ + +// Default value for GUIDs. This allows us to reset the GUID counter in tests. +// +// The initial GUID is 3 because some users have come to rely on the first +// default player ID ending up as `vjs_video_3`. +// +// See: https://github.com/videojs/video.js/pull/6216 +const _initialGuid = 3; + +/** + * Unique ID for an element or function + * + * @type {Number} + */ +let _guid = _initialGuid; + +/** + * Get a unique auto-incrementing ID by number that has not been returned before. + * + * @return {number} + * A new unique ID. + */ +function newGUID() { + return _guid++; +} + +/** + * @file events.js. An Event System (John Resig - Secrets of a JS Ninja http://jsninja.com/) + * (Original book version wasn't completely usable, so fixed some things and made Closure Compiler compatible) + * This should work very similarly to jQuery's events, however it's based off the book version which isn't as + * robust as jquery's, so there's probably some differences. + * + * @file events.js + * @module events + */ + +/** + * Clean up the listener cache and dispatchers + * + * @param {Element|Object} elem + * Element to clean up + * + * @param {string} type + * Type of event to clean up + */ +function _cleanUpEvents(elem, type) { + if (!DomData.has(elem)) { + return; + } + const data = DomData.get(elem); + + // Remove the events of a particular type if there are none left + if (data.handlers[type].length === 0) { + delete data.handlers[type]; + // data.handlers[type] = null; + // Setting to null was causing an error with data.handlers + + // Remove the meta-handler from the element + if (elem.removeEventListener) { + elem.removeEventListener(type, data.dispatcher, false); + } else if (elem.detachEvent) { + elem.detachEvent('on' + type, data.dispatcher); + } + } + + // Remove the events object if there are no types left + if (Object.getOwnPropertyNames(data.handlers).length <= 0) { + delete data.handlers; + delete data.dispatcher; + delete data.disabled; + } + + // Finally remove the element data if there is no data left + if (Object.getOwnPropertyNames(data).length === 0) { + DomData.delete(elem); + } +} + +/** + * Loops through an array of event types and calls the requested method for each type. + * + * @param {Function} fn + * The event method we want to use. + * + * @param {Element|Object} elem + * Element or object to bind listeners to + * + * @param {string[]} types + * Type of event to bind to. + * + * @param {Function} callback + * Event listener. + */ +function _handleMultipleEvents(fn, elem, types, callback) { + types.forEach(function (type) { + // Call the event method for each one of the types + fn(elem, type, callback); + }); +} + +/** + * Fix a native event to have standard property values + * + * @param {Object} event + * Event object to fix. + * + * @return {Object} + * Fixed event object. + */ +function fixEvent(event) { + if (event.fixed_) { + return event; + } + function returnTrue() { + return true; + } + function returnFalse() { + return false; + } + + // Test if fixing up is needed + // Used to check if !event.stopPropagation instead of isPropagationStopped + // But native events return true for stopPropagation, but don't have + // other expected methods like isPropagationStopped. Seems to be a problem + // with the Javascript Ninja code. So we're just overriding all events now. + if (!event || !event.isPropagationStopped || !event.isImmediatePropagationStopped) { + const old = event || window$1.event; + event = {}; + // Clone the old object so that we can modify the values event = {}; + // IE8 Doesn't like when you mess with native event properties + // Firefox returns false for event.hasOwnProperty('type') and other props + // which makes copying more difficult. + // TODO: Probably best to create a whitelist of event props + for (const key in old) { + // Safari 6.0.3 warns you if you try to copy deprecated layerX/Y + // Chrome warns you if you try to copy deprecated keyboardEvent.keyLocation + // and webkitMovementX/Y + // Lighthouse complains if Event.path is copied + if (key !== 'layerX' && key !== 'layerY' && key !== 'keyLocation' && key !== 'webkitMovementX' && key !== 'webkitMovementY' && key !== 'path') { + // Chrome 32+ warns if you try to copy deprecated returnValue, but + // we still want to if preventDefault isn't supported (IE8). + if (!(key === 'returnValue' && old.preventDefault)) { + event[key] = old[key]; + } + } + } + + // The event occurred on this element + if (!event.target) { + event.target = event.srcElement || document; + } + + // Handle which other element the event is related to + if (!event.relatedTarget) { + event.relatedTarget = event.fromElement === event.target ? event.toElement : event.fromElement; + } + + // Stop the default browser action + event.preventDefault = function () { + if (old.preventDefault) { + old.preventDefault(); + } + event.returnValue = false; + old.returnValue = false; + event.defaultPrevented = true; + }; + event.defaultPrevented = false; + + // Stop the event from bubbling + event.stopPropagation = function () { + if (old.stopPropagation) { + old.stopPropagation(); + } + event.cancelBubble = true; + old.cancelBubble = true; + event.isPropagationStopped = returnTrue; + }; + event.isPropagationStopped = returnFalse; + + // Stop the event from bubbling and executing other handlers + event.stopImmediatePropagation = function () { + if (old.stopImmediatePropagation) { + old.stopImmediatePropagation(); + } + event.isImmediatePropagationStopped = returnTrue; + event.stopPropagation(); + }; + event.isImmediatePropagationStopped = returnFalse; + + // Handle mouse position + if (event.clientX !== null && event.clientX !== undefined) { + const doc = document.documentElement; + const body = document.body; + event.pageX = event.clientX + (doc && doc.scrollLeft || body && body.scrollLeft || 0) - (doc && doc.clientLeft || body && body.clientLeft || 0); + event.pageY = event.clientY + (doc && doc.scrollTop || body && body.scrollTop || 0) - (doc && doc.clientTop || body && body.clientTop || 0); + } + + // Handle key presses + event.which = event.charCode || event.keyCode; + + // Fix button for mouse clicks: + // 0 == left; 1 == middle; 2 == right + if (event.button !== null && event.button !== undefined) { + // The following is disabled because it does not pass videojs-standard + // and... yikes. + /* eslint-disable */ + event.button = event.button & 1 ? 0 : event.button & 4 ? 1 : event.button & 2 ? 2 : 0; + /* eslint-enable */ + } + } + + event.fixed_ = true; + // Returns fixed-up instance + return event; +} + +/** + * Whether passive event listeners are supported + */ +let _supportsPassive; +const supportsPassive = function () { + if (typeof _supportsPassive !== 'boolean') { + _supportsPassive = false; + try { + const opts = Object.defineProperty({}, 'passive', { + get() { + _supportsPassive = true; + } + }); + window$1.addEventListener('test', null, opts); + window$1.removeEventListener('test', null, opts); + } catch (e) { + // disregard + } + } + return _supportsPassive; +}; + +/** + * Touch events Chrome expects to be passive + */ +const passiveEvents = ['touchstart', 'touchmove']; + +/** + * Add an event listener to element + * It stores the handler function in a separate cache object + * and adds a generic handler to the element's event, + * along with a unique id (guid) to the element. + * + * @param {Element|Object} elem + * Element or object to bind listeners to + * + * @param {string|string[]} type + * Type of event to bind to. + * + * @param {Function} fn + * Event listener. + */ +function on(elem, type, fn) { + if (Array.isArray(type)) { + return _handleMultipleEvents(on, elem, type, fn); + } + if (!DomData.has(elem)) { + DomData.set(elem, {}); + } + const data = DomData.get(elem); + + // We need a place to store all our handler data + if (!data.handlers) { + data.handlers = {}; + } + if (!data.handlers[type]) { + data.handlers[type] = []; + } + if (!fn.guid) { + fn.guid = newGUID(); + } + data.handlers[type].push(fn); + if (!data.dispatcher) { + data.disabled = false; + data.dispatcher = function (event, hash) { + if (data.disabled) { + return; + } + event = fixEvent(event); + const handlers = data.handlers[event.type]; + if (handlers) { + // Copy handlers so if handlers are added/removed during the process it doesn't throw everything off. + const handlersCopy = handlers.slice(0); + for (let m = 0, n = handlersCopy.length; m < n; m++) { + if (event.isImmediatePropagationStopped()) { + break; + } else { + try { + handlersCopy[m].call(elem, event, hash); + } catch (e) { + log$1.error(e); + } + } + } + } + }; + } + if (data.handlers[type].length === 1) { + if (elem.addEventListener) { + let options = false; + if (supportsPassive() && passiveEvents.indexOf(type) > -1) { + options = { + passive: true + }; + } + elem.addEventListener(type, data.dispatcher, options); + } else if (elem.attachEvent) { + elem.attachEvent('on' + type, data.dispatcher); + } + } +} + +/** + * Removes event listeners from an element + * + * @param {Element|Object} elem + * Object to remove listeners from. + * + * @param {string|string[]} [type] + * Type of listener to remove. Don't include to remove all events from element. + * + * @param {Function} [fn] + * Specific listener to remove. Don't include to remove listeners for an event + * type. + */ +function off(elem, type, fn) { + // Don't want to add a cache object through getElData if not needed + if (!DomData.has(elem)) { + return; + } + const data = DomData.get(elem); + + // If no events exist, nothing to unbind + if (!data.handlers) { + return; + } + if (Array.isArray(type)) { + return _handleMultipleEvents(off, elem, type, fn); + } + + // Utility function + const removeType = function (el, t) { + data.handlers[t] = []; + _cleanUpEvents(el, t); + }; + + // Are we removing all bound events? + if (type === undefined) { + for (const t in data.handlers) { + if (Object.prototype.hasOwnProperty.call(data.handlers || {}, t)) { + removeType(elem, t); + } + } + return; + } + const handlers = data.handlers[type]; + + // If no handlers exist, nothing to unbind + if (!handlers) { + return; + } + + // If no listener was provided, remove all listeners for type + if (!fn) { + removeType(elem, type); + return; + } + + // We're only removing a single handler + if (fn.guid) { + for (let n = 0; n < handlers.length; n++) { + if (handlers[n].guid === fn.guid) { + handlers.splice(n--, 1); + } + } + } + _cleanUpEvents(elem, type); +} + +/** + * Trigger an event for an element + * + * @param {Element|Object} elem + * Element to trigger an event on + * + * @param {EventTarget~Event|string} event + * A string (the type) or an event object with a type attribute + * + * @param {Object} [hash] + * data hash to pass along with the event + * + * @return {boolean|undefined} + * Returns the opposite of `defaultPrevented` if default was + * prevented. Otherwise, returns `undefined` + */ +function trigger(elem, event, hash) { + // Fetches element data and a reference to the parent (for bubbling). + // Don't want to add a data object to cache for every parent, + // so checking hasElData first. + const elemData = DomData.has(elem) ? DomData.get(elem) : {}; + const parent = elem.parentNode || elem.ownerDocument; + // type = event.type || event, + // handler; + + // If an event name was passed as a string, creates an event out of it + if (typeof event === 'string') { + event = { + type: event, + target: elem + }; + } else if (!event.target) { + event.target = elem; + } + + // Normalizes the event properties. + event = fixEvent(event); + + // If the passed element has a dispatcher, executes the established handlers. + if (elemData.dispatcher) { + elemData.dispatcher.call(elem, event, hash); + } + + // Unless explicitly stopped or the event does not bubble (e.g. media events) + // recursively calls this function to bubble the event up the DOM. + if (parent && !event.isPropagationStopped() && event.bubbles === true) { + trigger.call(null, parent, event, hash); + + // If at the top of the DOM, triggers the default action unless disabled. + } else if (!parent && !event.defaultPrevented && event.target && event.target[event.type]) { + if (!DomData.has(event.target)) { + DomData.set(event.target, {}); + } + const targetData = DomData.get(event.target); + + // Checks if the target has a default action for this event. + if (event.target[event.type]) { + // Temporarily disables event dispatching on the target as we have already executed the handler. + targetData.disabled = true; + // Executes the default action. + if (typeof event.target[event.type] === 'function') { + event.target[event.type](); + } + // Re-enables event dispatching. + targetData.disabled = false; + } + } + + // Inform the triggerer if the default was prevented by returning false + return !event.defaultPrevented; +} + +/** + * Trigger a listener only once for an event. + * + * @param {Element|Object} elem + * Element or object to bind to. + * + * @param {string|string[]} type + * Name/type of event + * + * @param {Event~EventListener} fn + * Event listener function + */ +function one(elem, type, fn) { + if (Array.isArray(type)) { + return _handleMultipleEvents(one, elem, type, fn); + } + const func = function () { + off(elem, type, func); + fn.apply(this, arguments); + }; + + // copy the guid to the new function so it can removed using the original function's ID + func.guid = fn.guid = fn.guid || newGUID(); + on(elem, type, func); +} + +/** + * Trigger a listener only once and then turn if off for all + * configured events + * + * @param {Element|Object} elem + * Element or object to bind to. + * + * @param {string|string[]} type + * Name/type of event + * + * @param {Event~EventListener} fn + * Event listener function + */ +function any(elem, type, fn) { + const func = function () { + off(elem, type, func); + fn.apply(this, arguments); + }; + + // copy the guid to the new function so it can removed using the original function's ID + func.guid = fn.guid = fn.guid || newGUID(); + + // multiple ons, but one off for everything + on(elem, type, func); +} + +var Events = /*#__PURE__*/Object.freeze({ + __proto__: null, + fixEvent: fixEvent, + on: on, + off: off, + trigger: trigger, + one: one, + any: any +}); + +/** + * @file fn.js + * @module fn + */ +const UPDATE_REFRESH_INTERVAL = 30; + +/** + * A private, internal-only function for changing the context of a function. + * + * It also stores a unique id on the function so it can be easily removed from + * events. + * + * @private + * @function + * @param {*} context + * The object to bind as scope. + * + * @param {Function} fn + * The function to be bound to a scope. + * + * @param {number} [uid] + * An optional unique ID for the function to be set + * + * @return {Function} + * The new function that will be bound into the context given + */ +const bind_ = function (context, fn, uid) { + // Make sure the function has a unique ID + if (!fn.guid) { + fn.guid = newGUID(); + } + + // Create the new function that changes the context + const bound = fn.bind(context); + + // Allow for the ability to individualize this function + // Needed in the case where multiple objects might share the same prototype + // IF both items add an event listener with the same function, then you try to remove just one + // it will remove both because they both have the same guid. + // when using this, you need to use the bind method when you remove the listener as well. + // currently used in text tracks + bound.guid = uid ? uid + '_' + fn.guid : fn.guid; + return bound; +}; + +/** + * Wraps the given function, `fn`, with a new function that only invokes `fn` + * at most once per every `wait` milliseconds. + * + * @function + * @param {Function} fn + * The function to be throttled. + * + * @param {number} wait + * The number of milliseconds by which to throttle. + * + * @return {Function} + */ +const throttle = function (fn, wait) { + let last = window$1.performance.now(); + const throttled = function (...args) { + const now = window$1.performance.now(); + if (now - last >= wait) { + fn(...args); + last = now; + } + }; + return throttled; +}; + +/** + * Creates a debounced function that delays invoking `func` until after `wait` + * milliseconds have elapsed since the last time the debounced function was + * invoked. + * + * Inspired by lodash and underscore implementations. + * + * @function + * @param {Function} func + * The function to wrap with debounce behavior. + * + * @param {number} wait + * The number of milliseconds to wait after the last invocation. + * + * @param {boolean} [immediate] + * Whether or not to invoke the function immediately upon creation. + * + * @param {Object} [context=window] + * The "context" in which the debounced function should debounce. For + * example, if this function should be tied to a Video.js player, + * the player can be passed here. Alternatively, defaults to the + * global `window` object. + * + * @return {Function} + * A debounced function. + */ +const debounce = function (func, wait, immediate, context = window$1) { + let timeout; + const cancel = () => { + context.clearTimeout(timeout); + timeout = null; + }; + + /* eslint-disable consistent-this */ + const debounced = function () { + const self = this; + const args = arguments; + let later = function () { + timeout = null; + later = null; + if (!immediate) { + func.apply(self, args); + } + }; + if (!timeout && immediate) { + func.apply(self, args); + } + context.clearTimeout(timeout); + timeout = context.setTimeout(later, wait); + }; + /* eslint-enable consistent-this */ + + debounced.cancel = cancel; + return debounced; +}; + +var Fn = /*#__PURE__*/Object.freeze({ + __proto__: null, + UPDATE_REFRESH_INTERVAL: UPDATE_REFRESH_INTERVAL, + bind_: bind_, + throttle: throttle, + debounce: debounce +}); + +/** + * @file src/js/event-target.js + */ +let EVENT_MAP; + +/** + * `EventTarget` is a class that can have the same API as the DOM `EventTarget`. It + * adds shorthand functions that wrap around lengthy functions. For example: + * the `on` function is a wrapper around `addEventListener`. + * + * @see [EventTarget Spec]{@link https://www.w3.org/TR/DOM-Level-2-Events/events.html#Events-EventTarget} + * @class EventTarget + */ +class EventTarget$2 { + /** + * Adds an `event listener` to an instance of an `EventTarget`. An `event listener` is a + * function that will get called when an event with a certain name gets triggered. + * + * @param {string|string[]} type + * An event name or an array of event names. + * + * @param {Function} fn + * The function to call with `EventTarget`s + */ + on(type, fn) { + // Remove the addEventListener alias before calling Events.on + // so we don't get into an infinite type loop + const ael = this.addEventListener; + this.addEventListener = () => {}; + on(this, type, fn); + this.addEventListener = ael; + } + /** + * Removes an `event listener` for a specific event from an instance of `EventTarget`. + * This makes it so that the `event listener` will no longer get called when the + * named event happens. + * + * @param {string|string[]} type + * An event name or an array of event names. + * + * @param {Function} fn + * The function to remove. + */ + off(type, fn) { + off(this, type, fn); + } + /** + * This function will add an `event listener` that gets triggered only once. After the + * first trigger it will get removed. This is like adding an `event listener` + * with {@link EventTarget#on} that calls {@link EventTarget#off} on itself. + * + * @param {string|string[]} type + * An event name or an array of event names. + * + * @param {Function} fn + * The function to be called once for each event name. + */ + one(type, fn) { + // Remove the addEventListener aliasing Events.on + // so we don't get into an infinite type loop + const ael = this.addEventListener; + this.addEventListener = () => {}; + one(this, type, fn); + this.addEventListener = ael; + } + /** + * This function will add an `event listener` that gets triggered only once and is + * removed from all events. This is like adding an array of `event listener`s + * with {@link EventTarget#on} that calls {@link EventTarget#off} on all events the + * first time it is triggered. + * + * @param {string|string[]} type + * An event name or an array of event names. + * + * @param {Function} fn + * The function to be called once for each event name. + */ + any(type, fn) { + // Remove the addEventListener aliasing Events.on + // so we don't get into an infinite type loop + const ael = this.addEventListener; + this.addEventListener = () => {}; + any(this, type, fn); + this.addEventListener = ael; + } + /** + * This function causes an event to happen. This will then cause any `event listeners` + * that are waiting for that event, to get called. If there are no `event listeners` + * for an event then nothing will happen. + * + * If the name of the `Event` that is being triggered is in `EventTarget.allowedEvents_`. + * Trigger will also call the `on` + `uppercaseEventName` function. + * + * Example: + * 'click' is in `EventTarget.allowedEvents_`, so, trigger will attempt to call + * `onClick` if it exists. + * + * @param {string|EventTarget~Event|Object} event + * The name of the event, an `Event`, or an object with a key of type set to + * an event name. + */ + trigger(event) { + const type = event.type || event; + + // deprecation + // In a future version we should default target to `this` + // similar to how we default the target to `elem` in + // `Events.trigger`. Right now the default `target` will be + // `document` due to the `Event.fixEvent` call. + if (typeof event === 'string') { + event = { + type + }; + } + event = fixEvent(event); + if (this.allowedEvents_[type] && this['on' + type]) { + this['on' + type](event); + } + trigger(this, event); + } + queueTrigger(event) { + // only set up EVENT_MAP if it'll be used + if (!EVENT_MAP) { + EVENT_MAP = new Map(); + } + const type = event.type || event; + let map = EVENT_MAP.get(this); + if (!map) { + map = new Map(); + EVENT_MAP.set(this, map); + } + const oldTimeout = map.get(type); + map.delete(type); + window$1.clearTimeout(oldTimeout); + const timeout = window$1.setTimeout(() => { + map.delete(type); + // if we cleared out all timeouts for the current target, delete its map + if (map.size === 0) { + map = null; + EVENT_MAP.delete(this); + } + this.trigger(event); + }, 0); + map.set(type, timeout); + } +} + +/** + * A Custom DOM event. + * + * @typedef {CustomEvent} Event + * @see [Properties]{@link https://developer.mozilla.org/en-US/docs/Web/API/CustomEvent} + */ + +/** + * All event listeners should follow the following format. + * + * @callback EventListener + * @this {EventTarget} + * + * @param {Event} event + * the event that triggered this function + * + * @param {Object} [hash] + * hash of data sent during the event + */ + +/** + * An object containing event names as keys and booleans as values. + * + * > NOTE: If an event name is set to a true value here {@link EventTarget#trigger} + * will have extra functionality. See that function for more information. + * + * @property EventTarget.prototype.allowedEvents_ + * @protected + */ +EventTarget$2.prototype.allowedEvents_ = {}; + +/** + * An alias of {@link EventTarget#on}. Allows `EventTarget` to mimic + * the standard DOM API. + * + * @function + * @see {@link EventTarget#on} + */ +EventTarget$2.prototype.addEventListener = EventTarget$2.prototype.on; + +/** + * An alias of {@link EventTarget#off}. Allows `EventTarget` to mimic + * the standard DOM API. + * + * @function + * @see {@link EventTarget#off} + */ +EventTarget$2.prototype.removeEventListener = EventTarget$2.prototype.off; + +/** + * An alias of {@link EventTarget#trigger}. Allows `EventTarget` to mimic + * the standard DOM API. + * + * @function + * @see {@link EventTarget#trigger} + */ +EventTarget$2.prototype.dispatchEvent = EventTarget$2.prototype.trigger; + +/** + * @file mixins/evented.js + * @module evented + */ +const objName = obj => { + if (typeof obj.name === 'function') { + return obj.name(); + } + if (typeof obj.name === 'string') { + return obj.name; + } + if (obj.name_) { + return obj.name_; + } + if (obj.constructor && obj.constructor.name) { + return obj.constructor.name; + } + return typeof obj; +}; + +/** + * Returns whether or not an object has had the evented mixin applied. + * + * @param {Object} object + * An object to test. + * + * @return {boolean} + * Whether or not the object appears to be evented. + */ +const isEvented = object => object instanceof EventTarget$2 || !!object.eventBusEl_ && ['on', 'one', 'off', 'trigger'].every(k => typeof object[k] === 'function'); + +/** + * Adds a callback to run after the evented mixin applied. + * + * @param {Object} target + * An object to Add + * @param {Function} callback + * The callback to run. + */ +const addEventedCallback = (target, callback) => { + if (isEvented(target)) { + callback(); + } else { + if (!target.eventedCallbacks) { + target.eventedCallbacks = []; + } + target.eventedCallbacks.push(callback); + } +}; + +/** + * Whether a value is a valid event type - non-empty string or array. + * + * @private + * @param {string|Array} type + * The type value to test. + * + * @return {boolean} + * Whether or not the type is a valid event type. + */ +const isValidEventType = type => +// The regex here verifies that the `type` contains at least one non- +// whitespace character. +typeof type === 'string' && /\S/.test(type) || Array.isArray(type) && !!type.length; + +/** + * Validates a value to determine if it is a valid event target. Throws if not. + * + * @private + * @throws {Error} + * If the target does not appear to be a valid event target. + * + * @param {Object} target + * The object to test. + * + * @param {Object} obj + * The evented object we are validating for + * + * @param {string} fnName + * The name of the evented mixin function that called this. + */ +const validateTarget = (target, obj, fnName) => { + if (!target || !target.nodeName && !isEvented(target)) { + throw new Error(`Invalid target for ${objName(obj)}#${fnName}; must be a DOM node or evented object.`); + } +}; + +/** + * Validates a value to determine if it is a valid event target. Throws if not. + * + * @private + * @throws {Error} + * If the type does not appear to be a valid event type. + * + * @param {string|Array} type + * The type to test. + * + * @param {Object} obj +* The evented object we are validating for + * + * @param {string} fnName + * The name of the evented mixin function that called this. + */ +const validateEventType = (type, obj, fnName) => { + if (!isValidEventType(type)) { + throw new Error(`Invalid event type for ${objName(obj)}#${fnName}; must be a non-empty string or array.`); + } +}; + +/** + * Validates a value to determine if it is a valid listener. Throws if not. + * + * @private + * @throws {Error} + * If the listener is not a function. + * + * @param {Function} listener + * The listener to test. + * + * @param {Object} obj + * The evented object we are validating for + * + * @param {string} fnName + * The name of the evented mixin function that called this. + */ +const validateListener = (listener, obj, fnName) => { + if (typeof listener !== 'function') { + throw new Error(`Invalid listener for ${objName(obj)}#${fnName}; must be a function.`); + } +}; + +/** + * Takes an array of arguments given to `on()` or `one()`, validates them, and + * normalizes them into an object. + * + * @private + * @param {Object} self + * The evented object on which `on()` or `one()` was called. This + * object will be bound as the `this` value for the listener. + * + * @param {Array} args + * An array of arguments passed to `on()` or `one()`. + * + * @param {string} fnName + * The name of the evented mixin function that called this. + * + * @return {Object} + * An object containing useful values for `on()` or `one()` calls. + */ +const normalizeListenArgs = (self, args, fnName) => { + // If the number of arguments is less than 3, the target is always the + // evented object itself. + const isTargetingSelf = args.length < 3 || args[0] === self || args[0] === self.eventBusEl_; + let target; + let type; + let listener; + if (isTargetingSelf) { + target = self.eventBusEl_; + + // Deal with cases where we got 3 arguments, but we are still listening to + // the evented object itself. + if (args.length >= 3) { + args.shift(); + } + [type, listener] = args; + } else { + [target, type, listener] = args; + } + validateTarget(target, self, fnName); + validateEventType(type, self, fnName); + validateListener(listener, self, fnName); + listener = bind_(self, listener); + return { + isTargetingSelf, + target, + type, + listener + }; +}; + +/** + * Adds the listener to the event type(s) on the target, normalizing for + * the type of target. + * + * @private + * @param {Element|Object} target + * A DOM node or evented object. + * + * @param {string} method + * The event binding method to use ("on" or "one"). + * + * @param {string|Array} type + * One or more event type(s). + * + * @param {Function} listener + * A listener function. + */ +const listen = (target, method, type, listener) => { + validateTarget(target, target, method); + if (target.nodeName) { + Events[method](target, type, listener); + } else { + target[method](type, listener); + } +}; + +/** + * Contains methods that provide event capabilities to an object which is passed + * to {@link module:evented|evented}. + * + * @mixin EventedMixin + */ +const EventedMixin = { + /** + * Add a listener to an event (or events) on this object or another evented + * object. + * + * @param {string|Array|Element|Object} targetOrType + * If this is a string or array, it represents the event type(s) + * that will trigger the listener. + * + * Another evented object can be passed here instead, which will + * cause the listener to listen for events on _that_ object. + * + * In either case, the listener's `this` value will be bound to + * this object. + * + * @param {string|Array|Function} typeOrListener + * If the first argument was a string or array, this should be the + * listener function. Otherwise, this is a string or array of event + * type(s). + * + * @param {Function} [listener] + * If the first argument was another evented object, this will be + * the listener function. + */ + on(...args) { + const { + isTargetingSelf, + target, + type, + listener + } = normalizeListenArgs(this, args, 'on'); + listen(target, 'on', type, listener); + + // If this object is listening to another evented object. + if (!isTargetingSelf) { + // If this object is disposed, remove the listener. + const removeListenerOnDispose = () => this.off(target, type, listener); + + // Use the same function ID as the listener so we can remove it later it + // using the ID of the original listener. + removeListenerOnDispose.guid = listener.guid; + + // Add a listener to the target's dispose event as well. This ensures + // that if the target is disposed BEFORE this object, we remove the + // removal listener that was just added. Otherwise, we create a memory leak. + const removeRemoverOnTargetDispose = () => this.off('dispose', removeListenerOnDispose); + + // Use the same function ID as the listener so we can remove it later + // it using the ID of the original listener. + removeRemoverOnTargetDispose.guid = listener.guid; + listen(this, 'on', 'dispose', removeListenerOnDispose); + listen(target, 'on', 'dispose', removeRemoverOnTargetDispose); + } + }, + /** + * Add a listener to an event (or events) on this object or another evented + * object. The listener will be called once per event and then removed. + * + * @param {string|Array|Element|Object} targetOrType + * If this is a string or array, it represents the event type(s) + * that will trigger the listener. + * + * Another evented object can be passed here instead, which will + * cause the listener to listen for events on _that_ object. + * + * In either case, the listener's `this` value will be bound to + * this object. + * + * @param {string|Array|Function} typeOrListener + * If the first argument was a string or array, this should be the + * listener function. Otherwise, this is a string or array of event + * type(s). + * + * @param {Function} [listener] + * If the first argument was another evented object, this will be + * the listener function. + */ + one(...args) { + const { + isTargetingSelf, + target, + type, + listener + } = normalizeListenArgs(this, args, 'one'); + + // Targeting this evented object. + if (isTargetingSelf) { + listen(target, 'one', type, listener); + + // Targeting another evented object. + } else { + // TODO: This wrapper is incorrect! It should only + // remove the wrapper for the event type that called it. + // Instead all listeners are removed on the first trigger! + // see https://github.com/videojs/video.js/issues/5962 + const wrapper = (...largs) => { + this.off(target, type, wrapper); + listener.apply(null, largs); + }; + + // Use the same function ID as the listener so we can remove it later + // it using the ID of the original listener. + wrapper.guid = listener.guid; + listen(target, 'one', type, wrapper); + } + }, + /** + * Add a listener to an event (or events) on this object or another evented + * object. The listener will only be called once for the first event that is triggered + * then removed. + * + * @param {string|Array|Element|Object} targetOrType + * If this is a string or array, it represents the event type(s) + * that will trigger the listener. + * + * Another evented object can be passed here instead, which will + * cause the listener to listen for events on _that_ object. + * + * In either case, the listener's `this` value will be bound to + * this object. + * + * @param {string|Array|Function} typeOrListener + * If the first argument was a string or array, this should be the + * listener function. Otherwise, this is a string or array of event + * type(s). + * + * @param {Function} [listener] + * If the first argument was another evented object, this will be + * the listener function. + */ + any(...args) { + const { + isTargetingSelf, + target, + type, + listener + } = normalizeListenArgs(this, args, 'any'); + + // Targeting this evented object. + if (isTargetingSelf) { + listen(target, 'any', type, listener); + + // Targeting another evented object. + } else { + const wrapper = (...largs) => { + this.off(target, type, wrapper); + listener.apply(null, largs); + }; + + // Use the same function ID as the listener so we can remove it later + // it using the ID of the original listener. + wrapper.guid = listener.guid; + listen(target, 'any', type, wrapper); + } + }, + /** + * Removes listener(s) from event(s) on an evented object. + * + * @param {string|Array|Element|Object} [targetOrType] + * If this is a string or array, it represents the event type(s). + * + * Another evented object can be passed here instead, in which case + * ALL 3 arguments are _required_. + * + * @param {string|Array|Function} [typeOrListener] + * If the first argument was a string or array, this may be the + * listener function. Otherwise, this is a string or array of event + * type(s). + * + * @param {Function} [listener] + * If the first argument was another evented object, this will be + * the listener function; otherwise, _all_ listeners bound to the + * event type(s) will be removed. + */ + off(targetOrType, typeOrListener, listener) { + // Targeting this evented object. + if (!targetOrType || isValidEventType(targetOrType)) { + off(this.eventBusEl_, targetOrType, typeOrListener); + + // Targeting another evented object. + } else { + const target = targetOrType; + const type = typeOrListener; + + // Fail fast and in a meaningful way! + validateTarget(target, this, 'off'); + validateEventType(type, this, 'off'); + validateListener(listener, this, 'off'); + + // Ensure there's at least a guid, even if the function hasn't been used + listener = bind_(this, listener); + + // Remove the dispose listener on this evented object, which was given + // the same guid as the event listener in on(). + this.off('dispose', listener); + if (target.nodeName) { + off(target, type, listener); + off(target, 'dispose', listener); + } else if (isEvented(target)) { + target.off(type, listener); + target.off('dispose', listener); + } + } + }, + /** + * Fire an event on this evented object, causing its listeners to be called. + * + * @param {string|Object} event + * An event type or an object with a type property. + * + * @param {Object} [hash] + * An additional object to pass along to listeners. + * + * @return {boolean} + * Whether or not the default behavior was prevented. + */ + trigger(event, hash) { + validateTarget(this.eventBusEl_, this, 'trigger'); + const type = event && typeof event !== 'string' ? event.type : event; + if (!isValidEventType(type)) { + throw new Error(`Invalid event type for ${objName(this)}#trigger; ` + 'must be a non-empty string or object with a type key that has a non-empty value.'); + } + return trigger(this.eventBusEl_, event, hash); + } +}; + +/** + * Applies {@link module:evented~EventedMixin|EventedMixin} to a target object. + * + * @param {Object} target + * The object to which to add event methods. + * + * @param {Object} [options={}] + * Options for customizing the mixin behavior. + * + * @param {string} [options.eventBusKey] + * By default, adds a `eventBusEl_` DOM element to the target object, + * which is used as an event bus. If the target object already has a + * DOM element that should be used, pass its key here. + * + * @return {Object} + * The target object. + */ +function evented(target, options = {}) { + const { + eventBusKey + } = options; + + // Set or create the eventBusEl_. + if (eventBusKey) { + if (!target[eventBusKey].nodeName) { + throw new Error(`The eventBusKey "${eventBusKey}" does not refer to an element.`); + } + target.eventBusEl_ = target[eventBusKey]; + } else { + target.eventBusEl_ = createEl('span', { + className: 'vjs-event-bus' + }); + } + Object.assign(target, EventedMixin); + if (target.eventedCallbacks) { + target.eventedCallbacks.forEach(callback => { + callback(); + }); + } + + // When any evented object is disposed, it removes all its listeners. + target.on('dispose', () => { + target.off(); + [target, target.el_, target.eventBusEl_].forEach(function (val) { + if (val && DomData.has(val)) { + DomData.delete(val); + } + }); + window$1.setTimeout(() => { + target.eventBusEl_ = null; + }, 0); + }); + return target; +} + +/** + * @file mixins/stateful.js + * @module stateful + */ + +/** + * Contains methods that provide statefulness to an object which is passed + * to {@link module:stateful}. + * + * @mixin StatefulMixin + */ +const StatefulMixin = { + /** + * A hash containing arbitrary keys and values representing the state of + * the object. + * + * @type {Object} + */ + state: {}, + /** + * Set the state of an object by mutating its + * {@link module:stateful~StatefulMixin.state|state} object in place. + * + * @fires module:stateful~StatefulMixin#statechanged + * @param {Object|Function} stateUpdates + * A new set of properties to shallow-merge into the plugin state. + * Can be a plain object or a function returning a plain object. + * + * @return {Object|undefined} + * An object containing changes that occurred. If no changes + * occurred, returns `undefined`. + */ + setState(stateUpdates) { + // Support providing the `stateUpdates` state as a function. + if (typeof stateUpdates === 'function') { + stateUpdates = stateUpdates(); + } + let changes; + each(stateUpdates, (value, key) => { + // Record the change if the value is different from what's in the + // current state. + if (this.state[key] !== value) { + changes = changes || {}; + changes[key] = { + from: this.state[key], + to: value + }; + } + this.state[key] = value; + }); + + // Only trigger "statechange" if there were changes AND we have a trigger + // function. This allows us to not require that the target object be an + // evented object. + if (changes && isEvented(this)) { + /** + * An event triggered on an object that is both + * {@link module:stateful|stateful} and {@link module:evented|evented} + * indicating that its state has changed. + * + * @event module:stateful~StatefulMixin#statechanged + * @type {Object} + * @property {Object} changes + * A hash containing the properties that were changed and + * the values they were changed `from` and `to`. + */ + this.trigger({ + changes, + type: 'statechanged' + }); + } + return changes; + } +}; + +/** + * Applies {@link module:stateful~StatefulMixin|StatefulMixin} to a target + * object. + * + * If the target object is {@link module:evented|evented} and has a + * `handleStateChanged` method, that method will be automatically bound to the + * `statechanged` event on itself. + * + * @param {Object} target + * The object to be made stateful. + * + * @param {Object} [defaultState] + * A default set of properties to populate the newly-stateful object's + * `state` property. + * + * @return {Object} + * Returns the `target`. + */ +function stateful(target, defaultState) { + Object.assign(target, StatefulMixin); + + // This happens after the mixing-in because we need to replace the `state` + // added in that step. + target.state = Object.assign({}, target.state, defaultState); + + // Auto-bind the `handleStateChanged` method of the target object if it exists. + if (typeof target.handleStateChanged === 'function' && isEvented(target)) { + target.on('statechanged', target.handleStateChanged); + } + return target; +} + +/** + * @file str.js + * @module to-lower-case + */ + +/** + * Lowercase the first letter of a string. + * + * @param {string} string + * String to be lowercased + * + * @return {string} + * The string with a lowercased first letter + */ +const toLowerCase = function (string) { + if (typeof string !== 'string') { + return string; + } + return string.replace(/./, w => w.toLowerCase()); +}; + +/** + * Uppercase the first letter of a string. + * + * @param {string} string + * String to be uppercased + * + * @return {string} + * The string with an uppercased first letter + */ +const toTitleCase$1 = function (string) { + if (typeof string !== 'string') { + return string; + } + return string.replace(/./, w => w.toUpperCase()); +}; + +/** + * Compares the TitleCase versions of the two strings for equality. + * + * @param {string} str1 + * The first string to compare + * + * @param {string} str2 + * The second string to compare + * + * @return {boolean} + * Whether the TitleCase versions of the strings are equal + */ +const titleCaseEquals = function (str1, str2) { + return toTitleCase$1(str1) === toTitleCase$1(str2); +}; + +var Str = /*#__PURE__*/Object.freeze({ + __proto__: null, + toLowerCase: toLowerCase, + toTitleCase: toTitleCase$1, + titleCaseEquals: titleCaseEquals +}); + +/** + * Player Component - Base class for all UI objects + * + * @file component.js + */ + +/** + * Base class for all UI Components. + * Components are UI objects which represent both a javascript object and an element + * in the DOM. They can be children of other components, and can have + * children themselves. + * + * Components can also use methods from {@link EventTarget} + */ +class Component$1 { + /** + * A callback that is called when a component is ready. Does not have any + * parameters and any callback value will be ignored. + * + * @callback ReadyCallback + * @this Component + */ + + /** + * Creates an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of component options. + * + * @param {Object[]} [options.children] + * An array of children objects to initialize this component with. Children objects have + * a name property that will be used if more than one component of the same type needs to be + * added. + * + * @param {string} [options.className] + * A class or space separated list of classes to add the component + * + * @param {ReadyCallback} [ready] + * Function that gets called when the `Component` is ready. + */ + constructor(player, options, ready) { + // The component might be the player itself and we can't pass `this` to super + if (!player && this.play) { + this.player_ = player = this; // eslint-disable-line + } else { + this.player_ = player; + } + this.isDisposed_ = false; + + // Hold the reference to the parent component via `addChild` method + this.parentComponent_ = null; + + // Make a copy of prototype.options_ to protect against overriding defaults + this.options_ = merge$1({}, this.options_); + + // Updated options with supplied options + options = this.options_ = merge$1(this.options_, options); + + // Get ID from options or options element if one is supplied + this.id_ = options.id || options.el && options.el.id; + + // If there was no ID from the options, generate one + if (!this.id_) { + // Don't require the player ID function in the case of mock players + const id = player && player.id && player.id() || 'no_player'; + this.id_ = `${id}_component_${newGUID()}`; + } + this.name_ = options.name || null; + + // Create element if one wasn't provided in options + if (options.el) { + this.el_ = options.el; + } else if (options.createEl !== false) { + this.el_ = this.createEl(); + } + if (options.className && this.el_) { + options.className.split(' ').forEach(c => this.addClass(c)); + } + + // Remove the placeholder event methods. If the component is evented, the + // real methods are added next + ['on', 'off', 'one', 'any', 'trigger'].forEach(fn => { + this[fn] = undefined; + }); + + // if evented is anything except false, we want to mixin in evented + if (options.evented !== false) { + // Make this an evented object and use `el_`, if available, as its event bus + evented(this, { + eventBusKey: this.el_ ? 'el_' : null + }); + this.handleLanguagechange = this.handleLanguagechange.bind(this); + this.on(this.player_, 'languagechange', this.handleLanguagechange); + } + stateful(this, this.constructor.defaultState); + this.children_ = []; + this.childIndex_ = {}; + this.childNameIndex_ = {}; + this.setTimeoutIds_ = new Set(); + this.setIntervalIds_ = new Set(); + this.rafIds_ = new Set(); + this.namedRafs_ = new Map(); + this.clearingTimersOnDispose_ = false; + + // Add any child components in options + if (options.initChildren !== false) { + this.initChildren(); + } + + // Don't want to trigger ready here or it will go before init is actually + // finished for all children that run this constructor + this.ready(ready); + if (options.reportTouchActivity !== false) { + this.enableTouchActivity(); + } + } + + // `on`, `off`, `one`, `any` and `trigger` are here so tsc includes them in definitions. + // They are replaced or removed in the constructor + + /** + * Adds an `event listener` to an instance of an `EventTarget`. An `event listener` is a + * function that will get called when an event with a certain name gets triggered. + * + * @param {string|string[]} type + * An event name or an array of event names. + * + * @param {Function} fn + * The function to call with `EventTarget`s + */ + on(type, fn) {} + + /** + * Removes an `event listener` for a specific event from an instance of `EventTarget`. + * This makes it so that the `event listener` will no longer get called when the + * named event happens. + * + * @param {string|string[]} type + * An event name or an array of event names. + * + * @param {Function} [fn] + * The function to remove. If not specified, all listeners managed by Video.js will be removed. + */ + off(type, fn) {} + + /** + * This function will add an `event listener` that gets triggered only once. After the + * first trigger it will get removed. This is like adding an `event listener` + * with {@link EventTarget#on} that calls {@link EventTarget#off} on itself. + * + * @param {string|string[]} type + * An event name or an array of event names. + * + * @param {Function} fn + * The function to be called once for each event name. + */ + one(type, fn) {} + + /** + * This function will add an `event listener` that gets triggered only once and is + * removed from all events. This is like adding an array of `event listener`s + * with {@link EventTarget#on} that calls {@link EventTarget#off} on all events the + * first time it is triggered. + * + * @param {string|string[]} type + * An event name or an array of event names. + * + * @param {Function} fn + * The function to be called once for each event name. + */ + any(type, fn) {} + + /** + * This function causes an event to happen. This will then cause any `event listeners` + * that are waiting for that event, to get called. If there are no `event listeners` + * for an event then nothing will happen. + * + * If the name of the `Event` that is being triggered is in `EventTarget.allowedEvents_`. + * Trigger will also call the `on` + `uppercaseEventName` function. + * + * Example: + * 'click' is in `EventTarget.allowedEvents_`, so, trigger will attempt to call + * `onClick` if it exists. + * + * @param {string|Event|Object} event + * The name of the event, an `Event`, or an object with a key of type set to + * an event name. + * + * @param {Object} [hash] + * Optionally extra argument to pass through to an event listener + */ + trigger(event, hash) {} + + /** + * Dispose of the `Component` and all child components. + * + * @fires Component#dispose + * + * @param {Object} options + * @param {Element} options.originalEl element with which to replace player element + */ + dispose(options = {}) { + // Bail out if the component has already been disposed. + if (this.isDisposed_) { + return; + } + if (this.readyQueue_) { + this.readyQueue_.length = 0; + } + + /** + * Triggered when a `Component` is disposed. + * + * @event Component#dispose + * @type {Event} + * + * @property {boolean} [bubbles=false] + * set to false so that the dispose event does not + * bubble up + */ + this.trigger({ + type: 'dispose', + bubbles: false + }); + this.isDisposed_ = true; + + // Dispose all children. + if (this.children_) { + for (let i = this.children_.length - 1; i >= 0; i--) { + if (this.children_[i].dispose) { + this.children_[i].dispose(); + } + } + } + + // Delete child references + this.children_ = null; + this.childIndex_ = null; + this.childNameIndex_ = null; + this.parentComponent_ = null; + if (this.el_) { + // Remove element from DOM + if (this.el_.parentNode) { + if (options.restoreEl) { + this.el_.parentNode.replaceChild(options.restoreEl, this.el_); + } else { + this.el_.parentNode.removeChild(this.el_); + } + } + this.el_ = null; + } + + // remove reference to the player after disposing of the element + this.player_ = null; + } + + /** + * Determine whether or not this component has been disposed. + * + * @return {boolean} + * If the component has been disposed, will be `true`. Otherwise, `false`. + */ + isDisposed() { + return Boolean(this.isDisposed_); + } + + /** + * Return the {@link Player} that the `Component` has attached to. + * + * @return { import('./player').default } + * The player that this `Component` has attached to. + */ + player() { + return this.player_; + } + + /** + * Deep merge of options objects with new options. + * > Note: When both `obj` and `options` contain properties whose values are objects. + * The two properties get merged using {@link module:obj.merge} + * + * @param {Object} obj + * The object that contains new options. + * + * @return {Object} + * A new object of `this.options_` and `obj` merged together. + */ + options(obj) { + if (!obj) { + return this.options_; + } + this.options_ = merge$1(this.options_, obj); + return this.options_; + } + + /** + * Get the `Component`s DOM element + * + * @return {Element} + * The DOM element for this `Component`. + */ + el() { + return this.el_; + } + + /** + * Create the `Component`s DOM element. + * + * @param {string} [tagName] + * Element's DOM node type. e.g. 'div' + * + * @param {Object} [properties] + * An object of properties that should be set. + * + * @param {Object} [attributes] + * An object of attributes that should be set. + * + * @return {Element} + * The element that gets created. + */ + createEl(tagName, properties, attributes) { + return createEl(tagName, properties, attributes); + } + + /** + * Localize a string given the string in english. + * + * If tokens are provided, it'll try and run a simple token replacement on the provided string. + * The tokens it looks for look like `{1}` with the index being 1-indexed into the tokens array. + * + * If a `defaultValue` is provided, it'll use that over `string`, + * if a value isn't found in provided language files. + * This is useful if you want to have a descriptive key for token replacement + * but have a succinct localized string and not require `en.json` to be included. + * + * Currently, it is used for the progress bar timing. + * ```js + * { + * "progress bar timing: currentTime={1} duration={2}": "{1} of {2}" + * } + * ``` + * It is then used like so: + * ```js + * this.localize('progress bar timing: currentTime={1} duration{2}', + * [this.player_.currentTime(), this.player_.duration()], + * '{1} of {2}'); + * ``` + * + * Which outputs something like: `01:23 of 24:56`. + * + * + * @param {string} string + * The string to localize and the key to lookup in the language files. + * @param {string[]} [tokens] + * If the current item has token replacements, provide the tokens here. + * @param {string} [defaultValue] + * Defaults to `string`. Can be a default value to use for token replacement + * if the lookup key is needed to be separate. + * + * @return {string} + * The localized string or if no localization exists the english string. + */ + localize(string, tokens, defaultValue = string) { + const code = this.player_.language && this.player_.language(); + const languages = this.player_.languages && this.player_.languages(); + const language = languages && languages[code]; + const primaryCode = code && code.split('-')[0]; + const primaryLang = languages && languages[primaryCode]; + let localizedString = defaultValue; + if (language && language[string]) { + localizedString = language[string]; + } else if (primaryLang && primaryLang[string]) { + localizedString = primaryLang[string]; + } + if (tokens) { + localizedString = localizedString.replace(/\{(\d+)\}/g, function (match, index) { + const value = tokens[index - 1]; + let ret = value; + if (typeof value === 'undefined') { + ret = match; + } + return ret; + }); + } + return localizedString; + } + + /** + * Handles language change for the player in components. Should be overridden by sub-components. + * + * @abstract + */ + handleLanguagechange() {} + + /** + * Return the `Component`s DOM element. This is where children get inserted. + * This will usually be the the same as the element returned in {@link Component#el}. + * + * @return {Element} + * The content element for this `Component`. + */ + contentEl() { + return this.contentEl_ || this.el_; + } + + /** + * Get this `Component`s ID + * + * @return {string} + * The id of this `Component` + */ + id() { + return this.id_; + } + + /** + * Get the `Component`s name. The name gets used to reference the `Component` + * and is set during registration. + * + * @return {string} + * The name of this `Component`. + */ + name() { + return this.name_; + } + + /** + * Get an array of all child components + * + * @return {Array} + * The children + */ + children() { + return this.children_; + } + + /** + * Returns the child `Component` with the given `id`. + * + * @param {string} id + * The id of the child `Component` to get. + * + * @return {Component|undefined} + * The child `Component` with the given `id` or undefined. + */ + getChildById(id) { + return this.childIndex_[id]; + } + + /** + * Returns the child `Component` with the given `name`. + * + * @param {string} name + * The name of the child `Component` to get. + * + * @return {Component|undefined} + * The child `Component` with the given `name` or undefined. + */ + getChild(name) { + if (!name) { + return; + } + return this.childNameIndex_[name]; + } + + /** + * Returns the descendant `Component` following the givent + * descendant `names`. For instance ['foo', 'bar', 'baz'] would + * try to get 'foo' on the current component, 'bar' on the 'foo' + * component and 'baz' on the 'bar' component and return undefined + * if any of those don't exist. + * + * @param {...string[]|...string} names + * The name of the child `Component` to get. + * + * @return {Component|undefined} + * The descendant `Component` following the given descendant + * `names` or undefined. + */ + getDescendant(...names) { + // flatten array argument into the main array + names = names.reduce((acc, n) => acc.concat(n), []); + let currentChild = this; + for (let i = 0; i < names.length; i++) { + currentChild = currentChild.getChild(names[i]); + if (!currentChild || !currentChild.getChild) { + return; + } + } + return currentChild; + } + + /** + * Adds an SVG icon element to another element or component. + * + * @param {string} iconName + * The name of icon. A list of all the icon names can be found at 'sandbox/svg-icons.html' + * + * @param {Element} [el=this.el()] + * Element to set the title on. Defaults to the current Component's element. + * + * @return {Element} + * The newly created icon element. + */ + setIcon(iconName, el = this.el()) { + // TODO: In v9 of video.js, we will want to remove font icons entirely. + // This means this check, as well as the others throughout the code, and + // the unecessary CSS for font icons, will need to be removed. + // See https://github.com/videojs/video.js/pull/8260 as to which components + // need updating. + if (!this.player_.options_.experimentalSvgIcons) { + return; + } + const xmlnsURL = 'http://www.w3.org/2000/svg'; + + // The below creates an element in the format of: + // <span><svg><use>....</use></svg></span> + const iconContainer = createEl('span', { + className: 'vjs-icon-placeholder vjs-svg-icon' + }, { + 'aria-hidden': 'true' + }); + const svgEl = document.createElementNS(xmlnsURL, 'svg'); + svgEl.setAttributeNS(null, 'viewBox', '0 0 512 512'); + const useEl = document.createElementNS(xmlnsURL, 'use'); + svgEl.appendChild(useEl); + useEl.setAttributeNS(null, 'href', `#vjs-icon-${iconName}`); + iconContainer.appendChild(svgEl); + + // Replace a pre-existing icon if one exists. + if (this.iconIsSet_) { + el.replaceChild(iconContainer, el.querySelector('.vjs-icon-placeholder')); + } else { + el.appendChild(iconContainer); + } + this.iconIsSet_ = true; + return iconContainer; + } + + /** + * Add a child `Component` inside the current `Component`. + * + * @param {string|Component} child + * The name or instance of a child to add. + * + * @param {Object} [options={}] + * The key/value store of options that will get passed to children of + * the child. + * + * @param {number} [index=this.children_.length] + * The index to attempt to add a child into. + * + * + * @return {Component} + * The `Component` that gets added as a child. When using a string the + * `Component` will get created by this process. + */ + addChild(child, options = {}, index = this.children_.length) { + let component; + let componentName; + + // If child is a string, create component with options + if (typeof child === 'string') { + componentName = toTitleCase$1(child); + const componentClassName = options.componentClass || componentName; + + // Set name through options + options.name = componentName; + + // Create a new object & element for this controls set + // If there's no .player_, this is a player + const ComponentClass = Component$1.getComponent(componentClassName); + if (!ComponentClass) { + throw new Error(`Component ${componentClassName} does not exist`); + } + + // data stored directly on the videojs object may be + // misidentified as a component to retain + // backwards-compatibility with 4.x. check to make sure the + // component class can be instantiated. + if (typeof ComponentClass !== 'function') { + return null; + } + component = new ComponentClass(this.player_ || this, options); + + // child is a component instance + } else { + component = child; + } + if (component.parentComponent_) { + component.parentComponent_.removeChild(component); + } + this.children_.splice(index, 0, component); + component.parentComponent_ = this; + if (typeof component.id === 'function') { + this.childIndex_[component.id()] = component; + } + + // If a name wasn't used to create the component, check if we can use the + // name function of the component + componentName = componentName || component.name && toTitleCase$1(component.name()); + if (componentName) { + this.childNameIndex_[componentName] = component; + this.childNameIndex_[toLowerCase(componentName)] = component; + } + + // Add the UI object's element to the container div (box) + // Having an element is not required + if (typeof component.el === 'function' && component.el()) { + // If inserting before a component, insert before that component's element + let refNode = null; + if (this.children_[index + 1]) { + // Most children are components, but the video tech is an HTML element + if (this.children_[index + 1].el_) { + refNode = this.children_[index + 1].el_; + } else if (isEl(this.children_[index + 1])) { + refNode = this.children_[index + 1]; + } + } + this.contentEl().insertBefore(component.el(), refNode); + } + + // Return so it can stored on parent object if desired. + return component; + } + + /** + * Remove a child `Component` from this `Component`s list of children. Also removes + * the child `Component`s element from this `Component`s element. + * + * @param {Component} component + * The child `Component` to remove. + */ + removeChild(component) { + if (typeof component === 'string') { + component = this.getChild(component); + } + if (!component || !this.children_) { + return; + } + let childFound = false; + for (let i = this.children_.length - 1; i >= 0; i--) { + if (this.children_[i] === component) { + childFound = true; + this.children_.splice(i, 1); + break; + } + } + if (!childFound) { + return; + } + component.parentComponent_ = null; + this.childIndex_[component.id()] = null; + this.childNameIndex_[toTitleCase$1(component.name())] = null; + this.childNameIndex_[toLowerCase(component.name())] = null; + const compEl = component.el(); + if (compEl && compEl.parentNode === this.contentEl()) { + this.contentEl().removeChild(component.el()); + } + } + + /** + * Add and initialize default child `Component`s based upon options. + */ + initChildren() { + const children = this.options_.children; + if (children) { + // `this` is `parent` + const parentOptions = this.options_; + const handleAdd = child => { + const name = child.name; + let opts = child.opts; + + // Allow options for children to be set at the parent options + // e.g. videojs(id, { controlBar: false }); + // instead of videojs(id, { children: { controlBar: false }); + if (parentOptions[name] !== undefined) { + opts = parentOptions[name]; + } + + // Allow for disabling default components + // e.g. options['children']['posterImage'] = false + if (opts === false) { + return; + } + + // Allow options to be passed as a simple boolean if no configuration + // is necessary. + if (opts === true) { + opts = {}; + } + + // We also want to pass the original player options + // to each component as well so they don't need to + // reach back into the player for options later. + opts.playerOptions = this.options_.playerOptions; + + // Create and add the child component. + // Add a direct reference to the child by name on the parent instance. + // If two of the same component are used, different names should be supplied + // for each + const newChild = this.addChild(name, opts); + if (newChild) { + this[name] = newChild; + } + }; + + // Allow for an array of children details to passed in the options + let workingChildren; + const Tech = Component$1.getComponent('Tech'); + if (Array.isArray(children)) { + workingChildren = children; + } else { + workingChildren = Object.keys(children); + } + workingChildren + // children that are in this.options_ but also in workingChildren would + // give us extra children we do not want. So, we want to filter them out. + .concat(Object.keys(this.options_).filter(function (child) { + return !workingChildren.some(function (wchild) { + if (typeof wchild === 'string') { + return child === wchild; + } + return child === wchild.name; + }); + })).map(child => { + let name; + let opts; + if (typeof child === 'string') { + name = child; + opts = children[name] || this.options_[name] || {}; + } else { + name = child.name; + opts = child; + } + return { + name, + opts + }; + }).filter(child => { + // we have to make sure that child.name isn't in the techOrder since + // techs are registered as Components but can't aren't compatible + // See https://github.com/videojs/video.js/issues/2772 + const c = Component$1.getComponent(child.opts.componentClass || toTitleCase$1(child.name)); + return c && !Tech.isTech(c); + }).forEach(handleAdd); + } + } + + /** + * Builds the default DOM class name. Should be overridden by sub-components. + * + * @return {string} + * The DOM class name for this object. + * + * @abstract + */ + buildCSSClass() { + // Child classes can include a function that does: + // return 'CLASS NAME' + this._super(); + return ''; + } + + /** + * Bind a listener to the component's ready state. + * Different from event listeners in that if the ready event has already happened + * it will trigger the function immediately. + * + * @param {ReadyCallback} fn + * Function that gets called when the `Component` is ready. + * + * @return {Component} + * Returns itself; method can be chained. + */ + ready(fn, sync = false) { + if (!fn) { + return; + } + if (!this.isReady_) { + this.readyQueue_ = this.readyQueue_ || []; + this.readyQueue_.push(fn); + return; + } + if (sync) { + fn.call(this); + } else { + // Call the function asynchronously by default for consistency + this.setTimeout(fn, 1); + } + } + + /** + * Trigger all the ready listeners for this `Component`. + * + * @fires Component#ready + */ + triggerReady() { + this.isReady_ = true; + + // Ensure ready is triggered asynchronously + this.setTimeout(function () { + const readyQueue = this.readyQueue_; + + // Reset Ready Queue + this.readyQueue_ = []; + if (readyQueue && readyQueue.length > 0) { + readyQueue.forEach(function (fn) { + fn.call(this); + }, this); + } + + // Allow for using event listeners also + /** + * Triggered when a `Component` is ready. + * + * @event Component#ready + * @type {Event} + */ + this.trigger('ready'); + }, 1); + } + + /** + * Find a single DOM element matching a `selector`. This can be within the `Component`s + * `contentEl()` or another custom context. + * + * @param {string} selector + * A valid CSS selector, which will be passed to `querySelector`. + * + * @param {Element|string} [context=this.contentEl()] + * A DOM element within which to query. Can also be a selector string in + * which case the first matching element will get used as context. If + * missing `this.contentEl()` gets used. If `this.contentEl()` returns + * nothing it falls back to `document`. + * + * @return {Element|null} + * the dom element that was found, or null + * + * @see [Information on CSS Selectors](https://developer.mozilla.org/en-US/docs/Web/Guide/CSS/Getting_Started/Selectors) + */ + $(selector, context) { + return $(selector, context || this.contentEl()); + } + + /** + * Finds all DOM element matching a `selector`. This can be within the `Component`s + * `contentEl()` or another custom context. + * + * @param {string} selector + * A valid CSS selector, which will be passed to `querySelectorAll`. + * + * @param {Element|string} [context=this.contentEl()] + * A DOM element within which to query. Can also be a selector string in + * which case the first matching element will get used as context. If + * missing `this.contentEl()` gets used. If `this.contentEl()` returns + * nothing it falls back to `document`. + * + * @return {NodeList} + * a list of dom elements that were found + * + * @see [Information on CSS Selectors](https://developer.mozilla.org/en-US/docs/Web/Guide/CSS/Getting_Started/Selectors) + */ + $$(selector, context) { + return $$(selector, context || this.contentEl()); + } + + /** + * Check if a component's element has a CSS class name. + * + * @param {string} classToCheck + * CSS class name to check. + * + * @return {boolean} + * - True if the `Component` has the class. + * - False if the `Component` does not have the class` + */ + hasClass(classToCheck) { + return hasClass(this.el_, classToCheck); + } + + /** + * Add a CSS class name to the `Component`s element. + * + * @param {...string} classesToAdd + * One or more CSS class name to add. + */ + addClass(...classesToAdd) { + addClass(this.el_, ...classesToAdd); + } + + /** + * Remove a CSS class name from the `Component`s element. + * + * @param {...string} classesToRemove + * One or more CSS class name to remove. + */ + removeClass(...classesToRemove) { + removeClass(this.el_, ...classesToRemove); + } + + /** + * Add or remove a CSS class name from the component's element. + * - `classToToggle` gets added when {@link Component#hasClass} would return false. + * - `classToToggle` gets removed when {@link Component#hasClass} would return true. + * + * @param {string} classToToggle + * The class to add or remove based on (@link Component#hasClass} + * + * @param {boolean|Dom~predicate} [predicate] + * An {@link Dom~predicate} function or a boolean + */ + toggleClass(classToToggle, predicate) { + toggleClass(this.el_, classToToggle, predicate); + } + + /** + * Show the `Component`s element if it is hidden by removing the + * 'vjs-hidden' class name from it. + */ + show() { + this.removeClass('vjs-hidden'); + } + + /** + * Hide the `Component`s element if it is currently showing by adding the + * 'vjs-hidden` class name to it. + */ + hide() { + this.addClass('vjs-hidden'); + } + + /** + * Lock a `Component`s element in its visible state by adding the 'vjs-lock-showing' + * class name to it. Used during fadeIn/fadeOut. + * + * @private + */ + lockShowing() { + this.addClass('vjs-lock-showing'); + } + + /** + * Unlock a `Component`s element from its visible state by removing the 'vjs-lock-showing' + * class name from it. Used during fadeIn/fadeOut. + * + * @private + */ + unlockShowing() { + this.removeClass('vjs-lock-showing'); + } + + /** + * Get the value of an attribute on the `Component`s element. + * + * @param {string} attribute + * Name of the attribute to get the value from. + * + * @return {string|null} + * - The value of the attribute that was asked for. + * - Can be an empty string on some browsers if the attribute does not exist + * or has no value + * - Most browsers will return null if the attribute does not exist or has + * no value. + * + * @see [DOM API]{@link https://developer.mozilla.org/en-US/docs/Web/API/Element/getAttribute} + */ + getAttribute(attribute) { + return getAttribute(this.el_, attribute); + } + + /** + * Set the value of an attribute on the `Component`'s element + * + * @param {string} attribute + * Name of the attribute to set. + * + * @param {string} value + * Value to set the attribute to. + * + * @see [DOM API]{@link https://developer.mozilla.org/en-US/docs/Web/API/Element/setAttribute} + */ + setAttribute(attribute, value) { + setAttribute(this.el_, attribute, value); + } + + /** + * Remove an attribute from the `Component`s element. + * + * @param {string} attribute + * Name of the attribute to remove. + * + * @see [DOM API]{@link https://developer.mozilla.org/en-US/docs/Web/API/Element/removeAttribute} + */ + removeAttribute(attribute) { + removeAttribute(this.el_, attribute); + } + + /** + * Get or set the width of the component based upon the CSS styles. + * See {@link Component#dimension} for more detailed information. + * + * @param {number|string} [num] + * The width that you want to set postfixed with '%', 'px' or nothing. + * + * @param {boolean} [skipListeners] + * Skip the componentresize event trigger + * + * @return {number|undefined} + * The width when getting, zero if there is no width + */ + width(num, skipListeners) { + return this.dimension('width', num, skipListeners); + } + + /** + * Get or set the height of the component based upon the CSS styles. + * See {@link Component#dimension} for more detailed information. + * + * @param {number|string} [num] + * The height that you want to set postfixed with '%', 'px' or nothing. + * + * @param {boolean} [skipListeners] + * Skip the componentresize event trigger + * + * @return {number|undefined} + * The height when getting, zero if there is no height + */ + height(num, skipListeners) { + return this.dimension('height', num, skipListeners); + } + + /** + * Set both the width and height of the `Component` element at the same time. + * + * @param {number|string} width + * Width to set the `Component`s element to. + * + * @param {number|string} height + * Height to set the `Component`s element to. + */ + dimensions(width, height) { + // Skip componentresize listeners on width for optimization + this.width(width, true); + this.height(height); + } + + /** + * Get or set width or height of the `Component` element. This is the shared code + * for the {@link Component#width} and {@link Component#height}. + * + * Things to know: + * - If the width or height in an number this will return the number postfixed with 'px'. + * - If the width/height is a percent this will return the percent postfixed with '%' + * - Hidden elements have a width of 0 with `window.getComputedStyle`. This function + * defaults to the `Component`s `style.width` and falls back to `window.getComputedStyle`. + * See [this]{@link http://www.foliotek.com/devblog/getting-the-width-of-a-hidden-element-with-jquery-using-width/} + * for more information + * - If you want the computed style of the component, use {@link Component#currentWidth} + * and {@link {Component#currentHeight} + * + * @fires Component#componentresize + * + * @param {string} widthOrHeight + 8 'width' or 'height' + * + * @param {number|string} [num] + 8 New dimension + * + * @param {boolean} [skipListeners] + * Skip componentresize event trigger + * + * @return {number|undefined} + * The dimension when getting or 0 if unset + */ + dimension(widthOrHeight, num, skipListeners) { + if (num !== undefined) { + // Set to zero if null or literally NaN (NaN !== NaN) + if (num === null || num !== num) { + num = 0; + } + + // Check if using css width/height (% or px) and adjust + if (('' + num).indexOf('%') !== -1 || ('' + num).indexOf('px') !== -1) { + this.el_.style[widthOrHeight] = num; + } else if (num === 'auto') { + this.el_.style[widthOrHeight] = ''; + } else { + this.el_.style[widthOrHeight] = num + 'px'; + } + + // skipListeners allows us to avoid triggering the resize event when setting both width and height + if (!skipListeners) { + /** + * Triggered when a component is resized. + * + * @event Component#componentresize + * @type {Event} + */ + this.trigger('componentresize'); + } + return; + } + + // Not setting a value, so getting it + // Make sure element exists + if (!this.el_) { + return 0; + } + + // Get dimension value from style + const val = this.el_.style[widthOrHeight]; + const pxIndex = val.indexOf('px'); + if (pxIndex !== -1) { + // Return the pixel value with no 'px' + return parseInt(val.slice(0, pxIndex), 10); + } + + // No px so using % or no style was set, so falling back to offsetWidth/height + // If component has display:none, offset will return 0 + // TODO: handle display:none and no dimension style using px + return parseInt(this.el_['offset' + toTitleCase$1(widthOrHeight)], 10); + } + + /** + * Get the computed width or the height of the component's element. + * + * Uses `window.getComputedStyle`. + * + * @param {string} widthOrHeight + * A string containing 'width' or 'height'. Whichever one you want to get. + * + * @return {number} + * The dimension that gets asked for or 0 if nothing was set + * for that dimension. + */ + currentDimension(widthOrHeight) { + let computedWidthOrHeight = 0; + if (widthOrHeight !== 'width' && widthOrHeight !== 'height') { + throw new Error('currentDimension only accepts width or height value'); + } + computedWidthOrHeight = computedStyle(this.el_, widthOrHeight); + + // remove 'px' from variable and parse as integer + computedWidthOrHeight = parseFloat(computedWidthOrHeight); + + // if the computed value is still 0, it's possible that the browser is lying + // and we want to check the offset values. + // This code also runs wherever getComputedStyle doesn't exist. + if (computedWidthOrHeight === 0 || isNaN(computedWidthOrHeight)) { + const rule = `offset${toTitleCase$1(widthOrHeight)}`; + computedWidthOrHeight = this.el_[rule]; + } + return computedWidthOrHeight; + } + + /** + * An object that contains width and height values of the `Component`s + * computed style. Uses `window.getComputedStyle`. + * + * @typedef {Object} Component~DimensionObject + * + * @property {number} width + * The width of the `Component`s computed style. + * + * @property {number} height + * The height of the `Component`s computed style. + */ + + /** + * Get an object that contains computed width and height values of the + * component's element. + * + * Uses `window.getComputedStyle`. + * + * @return {Component~DimensionObject} + * The computed dimensions of the component's element. + */ + currentDimensions() { + return { + width: this.currentDimension('width'), + height: this.currentDimension('height') + }; + } + + /** + * Get the computed width of the component's element. + * + * Uses `window.getComputedStyle`. + * + * @return {number} + * The computed width of the component's element. + */ + currentWidth() { + return this.currentDimension('width'); + } + + /** + * Get the computed height of the component's element. + * + * Uses `window.getComputedStyle`. + * + * @return {number} + * The computed height of the component's element. + */ + currentHeight() { + return this.currentDimension('height'); + } + + /** + * Set the focus to this component + */ + focus() { + this.el_.focus(); + } + + /** + * Remove the focus from this component + */ + blur() { + this.el_.blur(); + } + + /** + * When this Component receives a `keydown` event which it does not process, + * it passes the event to the Player for handling. + * + * @param {KeyboardEvent} event + * The `keydown` event that caused this function to be called. + */ + handleKeyDown(event) { + if (this.player_) { + // We only stop propagation here because we want unhandled events to fall + // back to the browser. Exclude Tab for focus trapping. + if (!keycode.isEventKey(event, 'Tab')) { + event.stopPropagation(); + } + this.player_.handleKeyDown(event); + } + } + + /** + * Many components used to have a `handleKeyPress` method, which was poorly + * named because it listened to a `keydown` event. This method name now + * delegates to `handleKeyDown`. This means anyone calling `handleKeyPress` + * will not see their method calls stop working. + * + * @param {KeyboardEvent} event + * The event that caused this function to be called. + */ + handleKeyPress(event) { + this.handleKeyDown(event); + } + + /** + * Emit a 'tap' events when touch event support gets detected. This gets used to + * support toggling the controls through a tap on the video. They get enabled + * because every sub-component would have extra overhead otherwise. + * + * @protected + * @fires Component#tap + * @listens Component#touchstart + * @listens Component#touchmove + * @listens Component#touchleave + * @listens Component#touchcancel + * @listens Component#touchend + */ + emitTapEvents() { + // Track the start time so we can determine how long the touch lasted + let touchStart = 0; + let firstTouch = null; + + // Maximum movement allowed during a touch event to still be considered a tap + // Other popular libs use anywhere from 2 (hammer.js) to 15, + // so 10 seems like a nice, round number. + const tapMovementThreshold = 10; + + // The maximum length a touch can be while still being considered a tap + const touchTimeThreshold = 200; + let couldBeTap; + this.on('touchstart', function (event) { + // If more than one finger, don't consider treating this as a click + if (event.touches.length === 1) { + // Copy pageX/pageY from the object + firstTouch = { + pageX: event.touches[0].pageX, + pageY: event.touches[0].pageY + }; + // Record start time so we can detect a tap vs. "touch and hold" + touchStart = window$1.performance.now(); + // Reset couldBeTap tracking + couldBeTap = true; + } + }); + this.on('touchmove', function (event) { + // If more than one finger, don't consider treating this as a click + if (event.touches.length > 1) { + couldBeTap = false; + } else if (firstTouch) { + // Some devices will throw touchmoves for all but the slightest of taps. + // So, if we moved only a small distance, this could still be a tap + const xdiff = event.touches[0].pageX - firstTouch.pageX; + const ydiff = event.touches[0].pageY - firstTouch.pageY; + const touchDistance = Math.sqrt(xdiff * xdiff + ydiff * ydiff); + if (touchDistance > tapMovementThreshold) { + couldBeTap = false; + } + } + }); + const noTap = function () { + couldBeTap = false; + }; + + // TODO: Listen to the original target. http://youtu.be/DujfpXOKUp8?t=13m8s + this.on('touchleave', noTap); + this.on('touchcancel', noTap); + + // When the touch ends, measure how long it took and trigger the appropriate + // event + this.on('touchend', function (event) { + firstTouch = null; + // Proceed only if the touchmove/leave/cancel event didn't happen + if (couldBeTap === true) { + // Measure how long the touch lasted + const touchTime = window$1.performance.now() - touchStart; + + // Make sure the touch was less than the threshold to be considered a tap + if (touchTime < touchTimeThreshold) { + // Don't let browser turn this into a click + event.preventDefault(); + /** + * Triggered when a `Component` is tapped. + * + * @event Component#tap + * @type {MouseEvent} + */ + this.trigger('tap'); + // It may be good to copy the touchend event object and change the + // type to tap, if the other event properties aren't exact after + // Events.fixEvent runs (e.g. event.target) + } + } + }); + } + + /** + * This function reports user activity whenever touch events happen. This can get + * turned off by any sub-components that wants touch events to act another way. + * + * Report user touch activity when touch events occur. User activity gets used to + * determine when controls should show/hide. It is simple when it comes to mouse + * events, because any mouse event should show the controls. So we capture mouse + * events that bubble up to the player and report activity when that happens. + * With touch events it isn't as easy as `touchstart` and `touchend` toggle player + * controls. So touch events can't help us at the player level either. + * + * User activity gets checked asynchronously. So what could happen is a tap event + * on the video turns the controls off. Then the `touchend` event bubbles up to + * the player. Which, if it reported user activity, would turn the controls right + * back on. We also don't want to completely block touch events from bubbling up. + * Furthermore a `touchmove` event and anything other than a tap, should not turn + * controls back on. + * + * @listens Component#touchstart + * @listens Component#touchmove + * @listens Component#touchend + * @listens Component#touchcancel + */ + enableTouchActivity() { + // Don't continue if the root player doesn't support reporting user activity + if (!this.player() || !this.player().reportUserActivity) { + return; + } + + // listener for reporting that the user is active + const report = bind_(this.player(), this.player().reportUserActivity); + let touchHolding; + this.on('touchstart', function () { + report(); + // For as long as the they are touching the device or have their mouse down, + // we consider them active even if they're not moving their finger or mouse. + // So we want to continue to update that they are active + this.clearInterval(touchHolding); + // report at the same interval as activityCheck + touchHolding = this.setInterval(report, 250); + }); + const touchEnd = function (event) { + report(); + // stop the interval that maintains activity if the touch is holding + this.clearInterval(touchHolding); + }; + this.on('touchmove', report); + this.on('touchend', touchEnd); + this.on('touchcancel', touchEnd); + } + + /** + * A callback that has no parameters and is bound into `Component`s context. + * + * @callback Component~GenericCallback + * @this Component + */ + + /** + * Creates a function that runs after an `x` millisecond timeout. This function is a + * wrapper around `window.setTimeout`. There are a few reasons to use this one + * instead though: + * 1. It gets cleared via {@link Component#clearTimeout} when + * {@link Component#dispose} gets called. + * 2. The function callback will gets turned into a {@link Component~GenericCallback} + * + * > Note: You can't use `window.clearTimeout` on the id returned by this function. This + * will cause its dispose listener not to get cleaned up! Please use + * {@link Component#clearTimeout} or {@link Component#dispose} instead. + * + * @param {Component~GenericCallback} fn + * The function that will be run after `timeout`. + * + * @param {number} timeout + * Timeout in milliseconds to delay before executing the specified function. + * + * @return {number} + * Returns a timeout ID that gets used to identify the timeout. It can also + * get used in {@link Component#clearTimeout} to clear the timeout that + * was set. + * + * @listens Component#dispose + * @see [Similar to]{@link https://developer.mozilla.org/en-US/docs/Web/API/WindowTimers/setTimeout} + */ + setTimeout(fn, timeout) { + // declare as variables so they are properly available in timeout function + // eslint-disable-next-line + var timeoutId; + fn = bind_(this, fn); + this.clearTimersOnDispose_(); + timeoutId = window$1.setTimeout(() => { + if (this.setTimeoutIds_.has(timeoutId)) { + this.setTimeoutIds_.delete(timeoutId); + } + fn(); + }, timeout); + this.setTimeoutIds_.add(timeoutId); + return timeoutId; + } + + /** + * Clears a timeout that gets created via `window.setTimeout` or + * {@link Component#setTimeout}. If you set a timeout via {@link Component#setTimeout} + * use this function instead of `window.clearTimout`. If you don't your dispose + * listener will not get cleaned up until {@link Component#dispose}! + * + * @param {number} timeoutId + * The id of the timeout to clear. The return value of + * {@link Component#setTimeout} or `window.setTimeout`. + * + * @return {number} + * Returns the timeout id that was cleared. + * + * @see [Similar to]{@link https://developer.mozilla.org/en-US/docs/Web/API/WindowTimers/clearTimeout} + */ + clearTimeout(timeoutId) { + if (this.setTimeoutIds_.has(timeoutId)) { + this.setTimeoutIds_.delete(timeoutId); + window$1.clearTimeout(timeoutId); + } + return timeoutId; + } + + /** + * Creates a function that gets run every `x` milliseconds. This function is a wrapper + * around `window.setInterval`. There are a few reasons to use this one instead though. + * 1. It gets cleared via {@link Component#clearInterval} when + * {@link Component#dispose} gets called. + * 2. The function callback will be a {@link Component~GenericCallback} + * + * @param {Component~GenericCallback} fn + * The function to run every `x` seconds. + * + * @param {number} interval + * Execute the specified function every `x` milliseconds. + * + * @return {number} + * Returns an id that can be used to identify the interval. It can also be be used in + * {@link Component#clearInterval} to clear the interval. + * + * @listens Component#dispose + * @see [Similar to]{@link https://developer.mozilla.org/en-US/docs/Web/API/WindowTimers/setInterval} + */ + setInterval(fn, interval) { + fn = bind_(this, fn); + this.clearTimersOnDispose_(); + const intervalId = window$1.setInterval(fn, interval); + this.setIntervalIds_.add(intervalId); + return intervalId; + } + + /** + * Clears an interval that gets created via `window.setInterval` or + * {@link Component#setInterval}. If you set an interval via {@link Component#setInterval} + * use this function instead of `window.clearInterval`. If you don't your dispose + * listener will not get cleaned up until {@link Component#dispose}! + * + * @param {number} intervalId + * The id of the interval to clear. The return value of + * {@link Component#setInterval} or `window.setInterval`. + * + * @return {number} + * Returns the interval id that was cleared. + * + * @see [Similar to]{@link https://developer.mozilla.org/en-US/docs/Web/API/WindowTimers/clearInterval} + */ + clearInterval(intervalId) { + if (this.setIntervalIds_.has(intervalId)) { + this.setIntervalIds_.delete(intervalId); + window$1.clearInterval(intervalId); + } + return intervalId; + } + + /** + * Queues up a callback to be passed to requestAnimationFrame (rAF), but + * with a few extra bonuses: + * + * - Supports browsers that do not support rAF by falling back to + * {@link Component#setTimeout}. + * + * - The callback is turned into a {@link Component~GenericCallback} (i.e. + * bound to the component). + * + * - Automatic cancellation of the rAF callback is handled if the component + * is disposed before it is called. + * + * @param {Component~GenericCallback} fn + * A function that will be bound to this component and executed just + * before the browser's next repaint. + * + * @return {number} + * Returns an rAF ID that gets used to identify the timeout. It can + * also be used in {@link Component#cancelAnimationFrame} to cancel + * the animation frame callback. + * + * @listens Component#dispose + * @see [Similar to]{@link https://developer.mozilla.org/en-US/docs/Web/API/window/requestAnimationFrame} + */ + requestAnimationFrame(fn) { + this.clearTimersOnDispose_(); + + // declare as variables so they are properly available in rAF function + // eslint-disable-next-line + var id; + fn = bind_(this, fn); + id = window$1.requestAnimationFrame(() => { + if (this.rafIds_.has(id)) { + this.rafIds_.delete(id); + } + fn(); + }); + this.rafIds_.add(id); + return id; + } + + /** + * Request an animation frame, but only one named animation + * frame will be queued. Another will never be added until + * the previous one finishes. + * + * @param {string} name + * The name to give this requestAnimationFrame + * + * @param {Component~GenericCallback} fn + * A function that will be bound to this component and executed just + * before the browser's next repaint. + */ + requestNamedAnimationFrame(name, fn) { + if (this.namedRafs_.has(name)) { + return; + } + this.clearTimersOnDispose_(); + fn = bind_(this, fn); + const id = this.requestAnimationFrame(() => { + fn(); + if (this.namedRafs_.has(name)) { + this.namedRafs_.delete(name); + } + }); + this.namedRafs_.set(name, id); + return name; + } + + /** + * Cancels a current named animation frame if it exists. + * + * @param {string} name + * The name of the requestAnimationFrame to cancel. + */ + cancelNamedAnimationFrame(name) { + if (!this.namedRafs_.has(name)) { + return; + } + this.cancelAnimationFrame(this.namedRafs_.get(name)); + this.namedRafs_.delete(name); + } + + /** + * Cancels a queued callback passed to {@link Component#requestAnimationFrame} + * (rAF). + * + * If you queue an rAF callback via {@link Component#requestAnimationFrame}, + * use this function instead of `window.cancelAnimationFrame`. If you don't, + * your dispose listener will not get cleaned up until {@link Component#dispose}! + * + * @param {number} id + * The rAF ID to clear. The return value of {@link Component#requestAnimationFrame}. + * + * @return {number} + * Returns the rAF ID that was cleared. + * + * @see [Similar to]{@link https://developer.mozilla.org/en-US/docs/Web/API/window/cancelAnimationFrame} + */ + cancelAnimationFrame(id) { + if (this.rafIds_.has(id)) { + this.rafIds_.delete(id); + window$1.cancelAnimationFrame(id); + } + return id; + } + + /** + * A function to setup `requestAnimationFrame`, `setTimeout`, + * and `setInterval`, clearing on dispose. + * + * > Previously each timer added and removed dispose listeners on it's own. + * For better performance it was decided to batch them all, and use `Set`s + * to track outstanding timer ids. + * + * @private + */ + clearTimersOnDispose_() { + if (this.clearingTimersOnDispose_) { + return; + } + this.clearingTimersOnDispose_ = true; + this.one('dispose', () => { + [['namedRafs_', 'cancelNamedAnimationFrame'], ['rafIds_', 'cancelAnimationFrame'], ['setTimeoutIds_', 'clearTimeout'], ['setIntervalIds_', 'clearInterval']].forEach(([idName, cancelName]) => { + // for a `Set` key will actually be the value again + // so forEach((val, val) =>` but for maps we want to use + // the key. + this[idName].forEach((val, key) => this[cancelName](key)); + }); + this.clearingTimersOnDispose_ = false; + }); + } + + /** + * Register a `Component` with `videojs` given the name and the component. + * + * > NOTE: {@link Tech}s should not be registered as a `Component`. {@link Tech}s + * should be registered using {@link Tech.registerTech} or + * {@link videojs:videojs.registerTech}. + * + * > NOTE: This function can also be seen on videojs as + * {@link videojs:videojs.registerComponent}. + * + * @param {string} name + * The name of the `Component` to register. + * + * @param {Component} ComponentToRegister + * The `Component` class to register. + * + * @return {Component} + * The `Component` that was registered. + */ + static registerComponent(name, ComponentToRegister) { + if (typeof name !== 'string' || !name) { + throw new Error(`Illegal component name, "${name}"; must be a non-empty string.`); + } + const Tech = Component$1.getComponent('Tech'); + + // We need to make sure this check is only done if Tech has been registered. + const isTech = Tech && Tech.isTech(ComponentToRegister); + const isComp = Component$1 === ComponentToRegister || Component$1.prototype.isPrototypeOf(ComponentToRegister.prototype); + if (isTech || !isComp) { + let reason; + if (isTech) { + reason = 'techs must be registered using Tech.registerTech()'; + } else { + reason = 'must be a Component subclass'; + } + throw new Error(`Illegal component, "${name}"; ${reason}.`); + } + name = toTitleCase$1(name); + if (!Component$1.components_) { + Component$1.components_ = {}; + } + const Player = Component$1.getComponent('Player'); + if (name === 'Player' && Player && Player.players) { + const players = Player.players; + const playerNames = Object.keys(players); + + // If we have players that were disposed, then their name will still be + // in Players.players. So, we must loop through and verify that the value + // for each item is not null. This allows registration of the Player component + // after all players have been disposed or before any were created. + if (players && playerNames.length > 0 && playerNames.map(pname => players[pname]).every(Boolean)) { + throw new Error('Can not register Player component after player has been created.'); + } + } + Component$1.components_[name] = ComponentToRegister; + Component$1.components_[toLowerCase(name)] = ComponentToRegister; + return ComponentToRegister; + } + + /** + * Get a `Component` based on the name it was registered with. + * + * @param {string} name + * The Name of the component to get. + * + * @return {typeof Component} + * The `Component` that got registered under the given name. + */ + static getComponent(name) { + if (!name || !Component$1.components_) { + return; + } + return Component$1.components_[name]; + } +} +Component$1.registerComponent('Component', Component$1); + +/** + * @file time.js + * @module time + */ + +/** + * Returns the time for the specified index at the start or end + * of a TimeRange object. + * + * @typedef {Function} TimeRangeIndex + * + * @param {number} [index=0] + * The range number to return the time for. + * + * @return {number} + * The time offset at the specified index. + * + * @deprecated The index argument must be provided. + * In the future, leaving it out will throw an error. + */ + +/** + * An object that contains ranges of time, which mimics {@link TimeRanges}. + * + * @typedef {Object} TimeRange + * + * @property {number} length + * The number of time ranges represented by this object. + * + * @property {module:time~TimeRangeIndex} start + * Returns the time offset at which a specified time range begins. + * + * @property {module:time~TimeRangeIndex} end + * Returns the time offset at which a specified time range ends. + * + * @see https://developer.mozilla.org/en-US/docs/Web/API/TimeRanges + */ + +/** + * Check if any of the time ranges are over the maximum index. + * + * @private + * @param {string} fnName + * The function name to use for logging + * + * @param {number} index + * The index to check + * + * @param {number} maxIndex + * The maximum possible index + * + * @throws {Error} if the timeRanges provided are over the maxIndex + */ +function rangeCheck(fnName, index, maxIndex) { + if (typeof index !== 'number' || index < 0 || index > maxIndex) { + throw new Error(`Failed to execute '${fnName}' on 'TimeRanges': The index provided (${index}) is non-numeric or out of bounds (0-${maxIndex}).`); + } +} + +/** + * Get the time for the specified index at the start or end + * of a TimeRange object. + * + * @private + * @param {string} fnName + * The function name to use for logging + * + * @param {string} valueIndex + * The property that should be used to get the time. should be + * 'start' or 'end' + * + * @param {Array} ranges + * An array of time ranges + * + * @param {Array} [rangeIndex=0] + * The index to start the search at + * + * @return {number} + * The time that offset at the specified index. + * + * @deprecated rangeIndex must be set to a value, in the future this will throw an error. + * @throws {Error} if rangeIndex is more than the length of ranges + */ +function getRange(fnName, valueIndex, ranges, rangeIndex) { + rangeCheck(fnName, rangeIndex, ranges.length - 1); + return ranges[rangeIndex][valueIndex]; +} + +/** + * Create a time range object given ranges of time. + * + * @private + * @param {Array} [ranges] + * An array of time ranges. + * + * @return {TimeRange} + */ +function createTimeRangesObj(ranges) { + let timeRangesObj; + if (ranges === undefined || ranges.length === 0) { + timeRangesObj = { + length: 0, + start() { + throw new Error('This TimeRanges object is empty'); + }, + end() { + throw new Error('This TimeRanges object is empty'); + } + }; + } else { + timeRangesObj = { + length: ranges.length, + start: getRange.bind(null, 'start', 0, ranges), + end: getRange.bind(null, 'end', 1, ranges) + }; + } + if (window$1.Symbol && window$1.Symbol.iterator) { + timeRangesObj[window$1.Symbol.iterator] = () => (ranges || []).values(); + } + return timeRangesObj; +} + +/** + * Create a `TimeRange` object which mimics an + * {@link https://developer.mozilla.org/en-US/docs/Web/API/TimeRanges|HTML5 TimeRanges instance}. + * + * @param {number|Array[]} start + * The start of a single range (a number) or an array of ranges (an + * array of arrays of two numbers each). + * + * @param {number} end + * The end of a single range. Cannot be used with the array form of + * the `start` argument. + * + * @return {TimeRange} + */ +function createTimeRanges$1(start, end) { + if (Array.isArray(start)) { + return createTimeRangesObj(start); + } else if (start === undefined || end === undefined) { + return createTimeRangesObj(); + } + return createTimeRangesObj([[start, end]]); +} + +/** + * Format seconds as a time string, H:MM:SS or M:SS. Supplying a guide (in + * seconds) will force a number of leading zeros to cover the length of the + * guide. + * + * @private + * @param {number} seconds + * Number of seconds to be turned into a string + * + * @param {number} guide + * Number (in seconds) to model the string after + * + * @return {string} + * Time formatted as H:MM:SS or M:SS + */ +const defaultImplementation = function (seconds, guide) { + seconds = seconds < 0 ? 0 : seconds; + let s = Math.floor(seconds % 60); + let m = Math.floor(seconds / 60 % 60); + let h = Math.floor(seconds / 3600); + const gm = Math.floor(guide / 60 % 60); + const gh = Math.floor(guide / 3600); + + // handle invalid times + if (isNaN(seconds) || seconds === Infinity) { + // '-' is false for all relational operators (e.g. <, >=) so this setting + // will add the minimum number of fields specified by the guide + h = m = s = '-'; + } + + // Check if we need to show hours + h = h > 0 || gh > 0 ? h + ':' : ''; + + // If hours are showing, we may need to add a leading zero. + // Always show at least one digit of minutes. + m = ((h || gm >= 10) && m < 10 ? '0' + m : m) + ':'; + + // Check if leading zero is need for seconds + s = s < 10 ? '0' + s : s; + return h + m + s; +}; + +// Internal pointer to the current implementation. +let implementation = defaultImplementation; + +/** + * Replaces the default formatTime implementation with a custom implementation. + * + * @param {Function} customImplementation + * A function which will be used in place of the default formatTime + * implementation. Will receive the current time in seconds and the + * guide (in seconds) as arguments. + */ +function setFormatTime(customImplementation) { + implementation = customImplementation; +} + +/** + * Resets formatTime to the default implementation. + */ +function resetFormatTime() { + implementation = defaultImplementation; +} + +/** + * Delegates to either the default time formatting function or a custom + * function supplied via `setFormatTime`. + * + * Formats seconds as a time string (H:MM:SS or M:SS). Supplying a + * guide (in seconds) will force a number of leading zeros to cover the + * length of the guide. + * + * @example formatTime(125, 600) === "02:05" + * @param {number} seconds + * Number of seconds to be turned into a string + * + * @param {number} guide + * Number (in seconds) to model the string after + * + * @return {string} + * Time formatted as H:MM:SS or M:SS + */ +function formatTime(seconds, guide = seconds) { + return implementation(seconds, guide); +} + +var Time = /*#__PURE__*/Object.freeze({ + __proto__: null, + createTimeRanges: createTimeRanges$1, + createTimeRange: createTimeRanges$1, + setFormatTime: setFormatTime, + resetFormatTime: resetFormatTime, + formatTime: formatTime +}); + +/** + * @file buffer.js + * @module buffer + */ + +/** + * Compute the percentage of the media that has been buffered. + * + * @param { import('./time').TimeRange } buffered + * The current `TimeRanges` object representing buffered time ranges + * + * @param {number} duration + * Total duration of the media + * + * @return {number} + * Percent buffered of the total duration in decimal form. + */ +function bufferedPercent(buffered, duration) { + let bufferedDuration = 0; + let start; + let end; + if (!duration) { + return 0; + } + if (!buffered || !buffered.length) { + buffered = createTimeRanges$1(0, 0); + } + for (let i = 0; i < buffered.length; i++) { + start = buffered.start(i); + end = buffered.end(i); + + // buffered end can be bigger than duration by a very small fraction + if (end > duration) { + end = duration; + } + bufferedDuration += end - start; + } + return bufferedDuration / duration; +} + +/** + * @file media-error.js + */ + +/** + * A Custom `MediaError` class which mimics the standard HTML5 `MediaError` class. + * + * @param {number|string|Object|MediaError} value + * This can be of multiple types: + * - number: should be a standard error code + * - string: an error message (the code will be 0) + * - Object: arbitrary properties + * - `MediaError` (native): used to populate a video.js `MediaError` object + * - `MediaError` (video.js): will return itself if it's already a + * video.js `MediaError` object. + * + * @see [MediaError Spec]{@link https://dev.w3.org/html5/spec-author-view/video.html#mediaerror} + * @see [Encrypted MediaError Spec]{@link https://www.w3.org/TR/2013/WD-encrypted-media-20130510/#error-codes} + * + * @class MediaError + */ +function MediaError(value) { + // Allow redundant calls to this constructor to avoid having `instanceof` + // checks peppered around the code. + if (value instanceof MediaError) { + return value; + } + if (typeof value === 'number') { + this.code = value; + } else if (typeof value === 'string') { + // default code is zero, so this is a custom error + this.message = value; + } else if (isObject(value)) { + // We assign the `code` property manually because native `MediaError` objects + // do not expose it as an own/enumerable property of the object. + if (typeof value.code === 'number') { + this.code = value.code; + } + Object.assign(this, value); + } + if (!this.message) { + this.message = MediaError.defaultMessages[this.code] || ''; + } +} + +/** + * The error code that refers two one of the defined `MediaError` types + * + * @type {Number} + */ +MediaError.prototype.code = 0; + +/** + * An optional message that to show with the error. Message is not part of the HTML5 + * video spec but allows for more informative custom errors. + * + * @type {String} + */ +MediaError.prototype.message = ''; + +/** + * An optional status code that can be set by plugins to allow even more detail about + * the error. For example a plugin might provide a specific HTTP status code and an + * error message for that code. Then when the plugin gets that error this class will + * know how to display an error message for it. This allows a custom message to show + * up on the `Player` error overlay. + * + * @type {Array} + */ +MediaError.prototype.status = null; + +/** + * An object containing an error type, as well as other information regarding the error. + * + * @typedef {{errorType: string, [key: string]: any}} ErrorMetadata + */ + +/** + * An optional object to give more detail about the error. This can be used to give + * a higher level of specificity to an error versus the more generic MediaError codes. + * `metadata` expects an `errorType` string that should align with the values from videojs.Error. + * + * @type {ErrorMetadata} + */ +MediaError.prototype.metadata = null; + +/** + * Errors indexed by the W3C standard. The order **CANNOT CHANGE**! See the + * specification listed under {@link MediaError} for more information. + * + * @enum {array} + * @readonly + * @property {string} 0 - MEDIA_ERR_CUSTOM + * @property {string} 1 - MEDIA_ERR_ABORTED + * @property {string} 2 - MEDIA_ERR_NETWORK + * @property {string} 3 - MEDIA_ERR_DECODE + * @property {string} 4 - MEDIA_ERR_SRC_NOT_SUPPORTED + * @property {string} 5 - MEDIA_ERR_ENCRYPTED + */ +MediaError.errorTypes = ['MEDIA_ERR_CUSTOM', 'MEDIA_ERR_ABORTED', 'MEDIA_ERR_NETWORK', 'MEDIA_ERR_DECODE', 'MEDIA_ERR_SRC_NOT_SUPPORTED', 'MEDIA_ERR_ENCRYPTED']; + +/** + * The default `MediaError` messages based on the {@link MediaError.errorTypes}. + * + * @type {Array} + * @constant + */ +MediaError.defaultMessages = { + 1: 'You aborted the media playback', + 2: 'A network error caused the media download to fail part-way.', + 3: 'The media playback was aborted due to a corruption problem or because the media used features your browser did not support.', + 4: 'The media could not be loaded, either because the server or network failed or because the format is not supported.', + 5: 'The media is encrypted and we do not have the keys to decrypt it.' +}; + +/** + * W3C error code for any custom error. + * + * @member MediaError#MEDIA_ERR_CUSTOM + * @constant {number} + * @default 0 + */ +MediaError.MEDIA_ERR_CUSTOM = 0; + +/** + * W3C error code for any custom error. + * + * @member MediaError.MEDIA_ERR_CUSTOM + * @constant {number} + * @default 0 + */ +MediaError.prototype.MEDIA_ERR_CUSTOM = 0; + +/** + * W3C error code for media error aborted. + * + * @member MediaError#MEDIA_ERR_ABORTED + * @constant {number} + * @default 1 + */ +MediaError.MEDIA_ERR_ABORTED = 1; + +/** + * W3C error code for media error aborted. + * + * @member MediaError.MEDIA_ERR_ABORTED + * @constant {number} + * @default 1 + */ +MediaError.prototype.MEDIA_ERR_ABORTED = 1; + +/** + * W3C error code for any network error. + * + * @member MediaError#MEDIA_ERR_NETWORK + * @constant {number} + * @default 2 + */ +MediaError.MEDIA_ERR_NETWORK = 2; + +/** + * W3C error code for any network error. + * + * @member MediaError.MEDIA_ERR_NETWORK + * @constant {number} + * @default 2 + */ +MediaError.prototype.MEDIA_ERR_NETWORK = 2; + +/** + * W3C error code for any decoding error. + * + * @member MediaError#MEDIA_ERR_DECODE + * @constant {number} + * @default 3 + */ +MediaError.MEDIA_ERR_DECODE = 3; + +/** + * W3C error code for any decoding error. + * + * @member MediaError.MEDIA_ERR_DECODE + * @constant {number} + * @default 3 + */ +MediaError.prototype.MEDIA_ERR_DECODE = 3; + +/** + * W3C error code for any time that a source is not supported. + * + * @member MediaError#MEDIA_ERR_SRC_NOT_SUPPORTED + * @constant {number} + * @default 4 + */ +MediaError.MEDIA_ERR_SRC_NOT_SUPPORTED = 4; + +/** + * W3C error code for any time that a source is not supported. + * + * @member MediaError.MEDIA_ERR_SRC_NOT_SUPPORTED + * @constant {number} + * @default 4 + */ +MediaError.prototype.MEDIA_ERR_SRC_NOT_SUPPORTED = 4; + +/** + * W3C error code for any time that a source is encrypted. + * + * @member MediaError#MEDIA_ERR_ENCRYPTED + * @constant {number} + * @default 5 + */ +MediaError.MEDIA_ERR_ENCRYPTED = 5; + +/** + * W3C error code for any time that a source is encrypted. + * + * @member MediaError.MEDIA_ERR_ENCRYPTED + * @constant {number} + * @default 5 + */ +MediaError.prototype.MEDIA_ERR_ENCRYPTED = 5; + +/** + * Returns whether an object is `Promise`-like (i.e. has a `then` method). + * + * @param {Object} value + * An object that may or may not be `Promise`-like. + * + * @return {boolean} + * Whether or not the object is `Promise`-like. + */ +function isPromise(value) { + return value !== undefined && value !== null && typeof value.then === 'function'; +} + +/** + * Silence a Promise-like object. + * + * This is useful for avoiding non-harmful, but potentially confusing "uncaught + * play promise" rejection error messages. + * + * @param {Object} value + * An object that may or may not be `Promise`-like. + */ +function silencePromise(value) { + if (isPromise(value)) { + value.then(null, e => {}); + } +} + +/** + * @file text-track-list-converter.js Utilities for capturing text track state and + * re-creating tracks based on a capture. + * + * @module text-track-list-converter + */ + +/** + * Examine a single {@link TextTrack} and return a JSON-compatible javascript object that + * represents the {@link TextTrack}'s state. + * + * @param {TextTrack} track + * The text track to query. + * + * @return {Object} + * A serializable javascript representation of the TextTrack. + * @private + */ +const trackToJson_ = function (track) { + const ret = ['kind', 'label', 'language', 'id', 'inBandMetadataTrackDispatchType', 'mode', 'src'].reduce((acc, prop, i) => { + if (track[prop]) { + acc[prop] = track[prop]; + } + return acc; + }, { + cues: track.cues && Array.prototype.map.call(track.cues, function (cue) { + return { + startTime: cue.startTime, + endTime: cue.endTime, + text: cue.text, + id: cue.id + }; + }) + }); + return ret; +}; + +/** + * Examine a {@link Tech} and return a JSON-compatible javascript array that represents the + * state of all {@link TextTrack}s currently configured. The return array is compatible with + * {@link text-track-list-converter:jsonToTextTracks}. + * + * @param { import('../tech/tech').default } tech + * The tech object to query + * + * @return {Array} + * A serializable javascript representation of the {@link Tech}s + * {@link TextTrackList}. + */ +const textTracksToJson = function (tech) { + const trackEls = tech.$$('track'); + const trackObjs = Array.prototype.map.call(trackEls, t => t.track); + const tracks = Array.prototype.map.call(trackEls, function (trackEl) { + const json = trackToJson_(trackEl.track); + if (trackEl.src) { + json.src = trackEl.src; + } + return json; + }); + return tracks.concat(Array.prototype.filter.call(tech.textTracks(), function (track) { + return trackObjs.indexOf(track) === -1; + }).map(trackToJson_)); +}; + +/** + * Create a set of remote {@link TextTrack}s on a {@link Tech} based on an array of javascript + * object {@link TextTrack} representations. + * + * @param {Array} json + * An array of `TextTrack` representation objects, like those that would be + * produced by `textTracksToJson`. + * + * @param {Tech} tech + * The `Tech` to create the `TextTrack`s on. + */ +const jsonToTextTracks = function (json, tech) { + json.forEach(function (track) { + const addedTrack = tech.addRemoteTextTrack(track).track; + if (!track.src && track.cues) { + track.cues.forEach(cue => addedTrack.addCue(cue)); + } + }); + return tech.textTracks(); +}; +var textTrackConverter = { + textTracksToJson, + jsonToTextTracks, + trackToJson_ +}; + +/** + * @file modal-dialog.js + */ +const MODAL_CLASS_NAME = 'vjs-modal-dialog'; + +/** + * The `ModalDialog` displays over the video and its controls, which blocks + * interaction with the player until it is closed. + * + * Modal dialogs include a "Close" button and will close when that button + * is activated - or when ESC is pressed anywhere. + * + * @extends Component + */ +class ModalDialog extends Component$1 { + /** + * Create an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + * + * @param { import('./utils/dom').ContentDescriptor} [options.content=undefined] + * Provide customized content for this modal. + * + * @param {string} [options.description] + * A text description for the modal, primarily for accessibility. + * + * @param {boolean} [options.fillAlways=false] + * Normally, modals are automatically filled only the first time + * they open. This tells the modal to refresh its content + * every time it opens. + * + * @param {string} [options.label] + * A text label for the modal, primarily for accessibility. + * + * @param {boolean} [options.pauseOnOpen=true] + * If `true`, playback will will be paused if playing when + * the modal opens, and resumed when it closes. + * + * @param {boolean} [options.temporary=true] + * If `true`, the modal can only be opened once; it will be + * disposed as soon as it's closed. + * + * @param {boolean} [options.uncloseable=false] + * If `true`, the user will not be able to close the modal + * through the UI in the normal ways. Programmatic closing is + * still possible. + */ + constructor(player, options) { + super(player, options); + this.handleKeyDown_ = e => this.handleKeyDown(e); + this.close_ = e => this.close(e); + this.opened_ = this.hasBeenOpened_ = this.hasBeenFilled_ = false; + this.closeable(!this.options_.uncloseable); + this.content(this.options_.content); + + // Make sure the contentEl is defined AFTER any children are initialized + // because we only want the contents of the modal in the contentEl + // (not the UI elements like the close button). + this.contentEl_ = createEl('div', { + className: `${MODAL_CLASS_NAME}-content` + }, { + role: 'document' + }); + this.descEl_ = createEl('p', { + className: `${MODAL_CLASS_NAME}-description vjs-control-text`, + id: this.el().getAttribute('aria-describedby') + }); + textContent(this.descEl_, this.description()); + this.el_.appendChild(this.descEl_); + this.el_.appendChild(this.contentEl_); + } + + /** + * Create the `ModalDialog`'s DOM element + * + * @return {Element} + * The DOM element that gets created. + */ + createEl() { + return super.createEl('div', { + className: this.buildCSSClass(), + tabIndex: -1 + }, { + 'aria-describedby': `${this.id()}_description`, + 'aria-hidden': 'true', + 'aria-label': this.label(), + 'role': 'dialog', + 'aria-live': 'polite' + }); + } + dispose() { + this.contentEl_ = null; + this.descEl_ = null; + this.previouslyActiveEl_ = null; + super.dispose(); + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `${MODAL_CLASS_NAME} vjs-hidden ${super.buildCSSClass()}`; + } + + /** + * Returns the label string for this modal. Primarily used for accessibility. + * + * @return {string} + * the localized or raw label of this modal. + */ + label() { + return this.localize(this.options_.label || 'Modal Window'); + } + + /** + * Returns the description string for this modal. Primarily used for + * accessibility. + * + * @return {string} + * The localized or raw description of this modal. + */ + description() { + let desc = this.options_.description || this.localize('This is a modal window.'); + + // Append a universal closeability message if the modal is closeable. + if (this.closeable()) { + desc += ' ' + this.localize('This modal can be closed by pressing the Escape key or activating the close button.'); + } + return desc; + } + + /** + * Opens the modal. + * + * @fires ModalDialog#beforemodalopen + * @fires ModalDialog#modalopen + */ + open() { + if (this.opened_) { + if (this.options_.fillAlways) { + this.fill(); + } + return; + } + const player = this.player(); + + /** + * Fired just before a `ModalDialog` is opened. + * + * @event ModalDialog#beforemodalopen + * @type {Event} + */ + this.trigger('beforemodalopen'); + this.opened_ = true; + + // Fill content if the modal has never opened before and + // never been filled. + if (this.options_.fillAlways || !this.hasBeenOpened_ && !this.hasBeenFilled_) { + this.fill(); + } + + // If the player was playing, pause it and take note of its previously + // playing state. + this.wasPlaying_ = !player.paused(); + if (this.options_.pauseOnOpen && this.wasPlaying_) { + player.pause(); + } + this.on('keydown', this.handleKeyDown_); + + // Hide controls and note if they were enabled. + this.hadControls_ = player.controls(); + player.controls(false); + this.show(); + this.conditionalFocus_(); + this.el().setAttribute('aria-hidden', 'false'); + + /** + * Fired just after a `ModalDialog` is opened. + * + * @event ModalDialog#modalopen + * @type {Event} + */ + this.trigger('modalopen'); + this.hasBeenOpened_ = true; + } + + /** + * If the `ModalDialog` is currently open or closed. + * + * @param {boolean} [value] + * If given, it will open (`true`) or close (`false`) the modal. + * + * @return {boolean} + * the current open state of the modaldialog + */ + opened(value) { + if (typeof value === 'boolean') { + this[value ? 'open' : 'close'](); + } + return this.opened_; + } + + /** + * Closes the modal, does nothing if the `ModalDialog` is + * not open. + * + * @fires ModalDialog#beforemodalclose + * @fires ModalDialog#modalclose + */ + close() { + if (!this.opened_) { + return; + } + const player = this.player(); + + /** + * Fired just before a `ModalDialog` is closed. + * + * @event ModalDialog#beforemodalclose + * @type {Event} + */ + this.trigger('beforemodalclose'); + this.opened_ = false; + if (this.wasPlaying_ && this.options_.pauseOnOpen) { + player.play(); + } + this.off('keydown', this.handleKeyDown_); + if (this.hadControls_) { + player.controls(true); + } + this.hide(); + this.el().setAttribute('aria-hidden', 'true'); + + /** + * Fired just after a `ModalDialog` is closed. + * + * @event ModalDialog#modalclose + * @type {Event} + */ + this.trigger('modalclose'); + this.conditionalBlur_(); + if (this.options_.temporary) { + this.dispose(); + } + } + + /** + * Check to see if the `ModalDialog` is closeable via the UI. + * + * @param {boolean} [value] + * If given as a boolean, it will set the `closeable` option. + * + * @return {boolean} + * Returns the final value of the closable option. + */ + closeable(value) { + if (typeof value === 'boolean') { + const closeable = this.closeable_ = !!value; + let close = this.getChild('closeButton'); + + // If this is being made closeable and has no close button, add one. + if (closeable && !close) { + // The close button should be a child of the modal - not its + // content element, so temporarily change the content element. + const temp = this.contentEl_; + this.contentEl_ = this.el_; + close = this.addChild('closeButton', { + controlText: 'Close Modal Dialog' + }); + this.contentEl_ = temp; + this.on(close, 'close', this.close_); + } + + // If this is being made uncloseable and has a close button, remove it. + if (!closeable && close) { + this.off(close, 'close', this.close_); + this.removeChild(close); + close.dispose(); + } + } + return this.closeable_; + } + + /** + * Fill the modal's content element with the modal's "content" option. + * The content element will be emptied before this change takes place. + */ + fill() { + this.fillWith(this.content()); + } + + /** + * Fill the modal's content element with arbitrary content. + * The content element will be emptied before this change takes place. + * + * @fires ModalDialog#beforemodalfill + * @fires ModalDialog#modalfill + * + * @param { import('./utils/dom').ContentDescriptor} [content] + * The same rules apply to this as apply to the `content` option. + */ + fillWith(content) { + const contentEl = this.contentEl(); + const parentEl = contentEl.parentNode; + const nextSiblingEl = contentEl.nextSibling; + + /** + * Fired just before a `ModalDialog` is filled with content. + * + * @event ModalDialog#beforemodalfill + * @type {Event} + */ + this.trigger('beforemodalfill'); + this.hasBeenFilled_ = true; + + // Detach the content element from the DOM before performing + // manipulation to avoid modifying the live DOM multiple times. + parentEl.removeChild(contentEl); + this.empty(); + insertContent(contentEl, content); + /** + * Fired just after a `ModalDialog` is filled with content. + * + * @event ModalDialog#modalfill + * @type {Event} + */ + this.trigger('modalfill'); + + // Re-inject the re-filled content element. + if (nextSiblingEl) { + parentEl.insertBefore(contentEl, nextSiblingEl); + } else { + parentEl.appendChild(contentEl); + } + + // make sure that the close button is last in the dialog DOM + const closeButton = this.getChild('closeButton'); + if (closeButton) { + parentEl.appendChild(closeButton.el_); + } + } + + /** + * Empties the content element. This happens anytime the modal is filled. + * + * @fires ModalDialog#beforemodalempty + * @fires ModalDialog#modalempty + */ + empty() { + /** + * Fired just before a `ModalDialog` is emptied. + * + * @event ModalDialog#beforemodalempty + * @type {Event} + */ + this.trigger('beforemodalempty'); + emptyEl(this.contentEl()); + + /** + * Fired just after a `ModalDialog` is emptied. + * + * @event ModalDialog#modalempty + * @type {Event} + */ + this.trigger('modalempty'); + } + + /** + * Gets or sets the modal content, which gets normalized before being + * rendered into the DOM. + * + * This does not update the DOM or fill the modal, but it is called during + * that process. + * + * @param { import('./utils/dom').ContentDescriptor} [value] + * If defined, sets the internal content value to be used on the + * next call(s) to `fill`. This value is normalized before being + * inserted. To "clear" the internal content value, pass `null`. + * + * @return { import('./utils/dom').ContentDescriptor} + * The current content of the modal dialog + */ + content(value) { + if (typeof value !== 'undefined') { + this.content_ = value; + } + return this.content_; + } + + /** + * conditionally focus the modal dialog if focus was previously on the player. + * + * @private + */ + conditionalFocus_() { + const activeEl = document.activeElement; + const playerEl = this.player_.el_; + this.previouslyActiveEl_ = null; + if (playerEl.contains(activeEl) || playerEl === activeEl) { + this.previouslyActiveEl_ = activeEl; + this.focus(); + } + } + + /** + * conditionally blur the element and refocus the last focused element + * + * @private + */ + conditionalBlur_() { + if (this.previouslyActiveEl_) { + this.previouslyActiveEl_.focus(); + this.previouslyActiveEl_ = null; + } + } + + /** + * Keydown handler. Attached when modal is focused. + * + * @listens keydown + */ + handleKeyDown(event) { + // Do not allow keydowns to reach out of the modal dialog. + event.stopPropagation(); + if (keycode.isEventKey(event, 'Escape') && this.closeable()) { + event.preventDefault(); + this.close(); + return; + } + + // exit early if it isn't a tab key + if (!keycode.isEventKey(event, 'Tab')) { + return; + } + const focusableEls = this.focusableEls_(); + const activeEl = this.el_.querySelector(':focus'); + let focusIndex; + for (let i = 0; i < focusableEls.length; i++) { + if (activeEl === focusableEls[i]) { + focusIndex = i; + break; + } + } + if (document.activeElement === this.el_) { + focusIndex = 0; + } + if (event.shiftKey && focusIndex === 0) { + focusableEls[focusableEls.length - 1].focus(); + event.preventDefault(); + } else if (!event.shiftKey && focusIndex === focusableEls.length - 1) { + focusableEls[0].focus(); + event.preventDefault(); + } + } + + /** + * get all focusable elements + * + * @private + */ + focusableEls_() { + const allChildren = this.el_.querySelectorAll('*'); + return Array.prototype.filter.call(allChildren, child => { + return (child instanceof window$1.HTMLAnchorElement || child instanceof window$1.HTMLAreaElement) && child.hasAttribute('href') || (child instanceof window$1.HTMLInputElement || child instanceof window$1.HTMLSelectElement || child instanceof window$1.HTMLTextAreaElement || child instanceof window$1.HTMLButtonElement) && !child.hasAttribute('disabled') || child instanceof window$1.HTMLIFrameElement || child instanceof window$1.HTMLObjectElement || child instanceof window$1.HTMLEmbedElement || child.hasAttribute('tabindex') && child.getAttribute('tabindex') !== -1 || child.hasAttribute('contenteditable'); + }); + } +} + +/** + * Default options for `ModalDialog` default options. + * + * @type {Object} + * @private + */ +ModalDialog.prototype.options_ = { + pauseOnOpen: true, + temporary: true +}; +Component$1.registerComponent('ModalDialog', ModalDialog); + +/** + * @file track-list.js + */ + +/** + * Common functionaliy between {@link TextTrackList}, {@link AudioTrackList}, and + * {@link VideoTrackList} + * + * @extends EventTarget + */ +class TrackList extends EventTarget$2 { + /** + * Create an instance of this class + * + * @param { import('./track').default[] } tracks + * A list of tracks to initialize the list with. + * + * @abstract + */ + constructor(tracks = []) { + super(); + this.tracks_ = []; + + /** + * @memberof TrackList + * @member {number} length + * The current number of `Track`s in the this Trackist. + * @instance + */ + Object.defineProperty(this, 'length', { + get() { + return this.tracks_.length; + } + }); + for (let i = 0; i < tracks.length; i++) { + this.addTrack(tracks[i]); + } + } + + /** + * Add a {@link Track} to the `TrackList` + * + * @param { import('./track').default } track + * The audio, video, or text track to add to the list. + * + * @fires TrackList#addtrack + */ + addTrack(track) { + const index = this.tracks_.length; + if (!('' + index in this)) { + Object.defineProperty(this, index, { + get() { + return this.tracks_[index]; + } + }); + } + + // Do not add duplicate tracks + if (this.tracks_.indexOf(track) === -1) { + this.tracks_.push(track); + /** + * Triggered when a track is added to a track list. + * + * @event TrackList#addtrack + * @type {Event} + * @property {Track} track + * A reference to track that was added. + */ + this.trigger({ + track, + type: 'addtrack', + target: this + }); + } + + /** + * Triggered when a track label is changed. + * + * @event TrackList#addtrack + * @type {Event} + * @property {Track} track + * A reference to track that was added. + */ + track.labelchange_ = () => { + this.trigger({ + track, + type: 'labelchange', + target: this + }); + }; + if (isEvented(track)) { + track.addEventListener('labelchange', track.labelchange_); + } + } + + /** + * Remove a {@link Track} from the `TrackList` + * + * @param { import('./track').default } rtrack + * The audio, video, or text track to remove from the list. + * + * @fires TrackList#removetrack + */ + removeTrack(rtrack) { + let track; + for (let i = 0, l = this.length; i < l; i++) { + if (this[i] === rtrack) { + track = this[i]; + if (track.off) { + track.off(); + } + this.tracks_.splice(i, 1); + break; + } + } + if (!track) { + return; + } + + /** + * Triggered when a track is removed from track list. + * + * @event TrackList#removetrack + * @type {Event} + * @property {Track} track + * A reference to track that was removed. + */ + this.trigger({ + track, + type: 'removetrack', + target: this + }); + } + + /** + * Get a Track from the TrackList by a tracks id + * + * @param {string} id - the id of the track to get + * @method getTrackById + * @return { import('./track').default } + * @private + */ + getTrackById(id) { + let result = null; + for (let i = 0, l = this.length; i < l; i++) { + const track = this[i]; + if (track.id === id) { + result = track; + break; + } + } + return result; + } +} + +/** + * Triggered when a different track is selected/enabled. + * + * @event TrackList#change + * @type {Event} + */ + +/** + * Events that can be called with on + eventName. See {@link EventHandler}. + * + * @property {Object} TrackList#allowedEvents_ + * @protected + */ +TrackList.prototype.allowedEvents_ = { + change: 'change', + addtrack: 'addtrack', + removetrack: 'removetrack', + labelchange: 'labelchange' +}; + +// emulate attribute EventHandler support to allow for feature detection +for (const event in TrackList.prototype.allowedEvents_) { + TrackList.prototype['on' + event] = null; +} + +/** + * @file audio-track-list.js + */ + +/** + * Anywhere we call this function we diverge from the spec + * as we only support one enabled audiotrack at a time + * + * @param {AudioTrackList} list + * list to work on + * + * @param { import('./audio-track').default } track + * The track to skip + * + * @private + */ +const disableOthers$1 = function (list, track) { + for (let i = 0; i < list.length; i++) { + if (!Object.keys(list[i]).length || track.id === list[i].id) { + continue; + } + // another audio track is enabled, disable it + list[i].enabled = false; + } +}; + +/** + * The current list of {@link AudioTrack} for a media file. + * + * @see [Spec]{@link https://html.spec.whatwg.org/multipage/embedded-content.html#audiotracklist} + * @extends TrackList + */ +class AudioTrackList extends TrackList { + /** + * Create an instance of this class. + * + * @param { import('./audio-track').default[] } [tracks=[]] + * A list of `AudioTrack` to instantiate the list with. + */ + constructor(tracks = []) { + // make sure only 1 track is enabled + // sorted from last index to first index + for (let i = tracks.length - 1; i >= 0; i--) { + if (tracks[i].enabled) { + disableOthers$1(tracks, tracks[i]); + break; + } + } + super(tracks); + this.changing_ = false; + } + + /** + * Add an {@link AudioTrack} to the `AudioTrackList`. + * + * @param { import('./audio-track').default } track + * The AudioTrack to add to the list + * + * @fires TrackList#addtrack + */ + addTrack(track) { + if (track.enabled) { + disableOthers$1(this, track); + } + super.addTrack(track); + // native tracks don't have this + if (!track.addEventListener) { + return; + } + track.enabledChange_ = () => { + // when we are disabling other tracks (since we don't support + // more than one track at a time) we will set changing_ + // to true so that we don't trigger additional change events + if (this.changing_) { + return; + } + this.changing_ = true; + disableOthers$1(this, track); + this.changing_ = false; + this.trigger('change'); + }; + + /** + * @listens AudioTrack#enabledchange + * @fires TrackList#change + */ + track.addEventListener('enabledchange', track.enabledChange_); + } + removeTrack(rtrack) { + super.removeTrack(rtrack); + if (rtrack.removeEventListener && rtrack.enabledChange_) { + rtrack.removeEventListener('enabledchange', rtrack.enabledChange_); + rtrack.enabledChange_ = null; + } + } +} + +/** + * @file video-track-list.js + */ + +/** + * Un-select all other {@link VideoTrack}s that are selected. + * + * @param {VideoTrackList} list + * list to work on + * + * @param { import('./video-track').default } track + * The track to skip + * + * @private + */ +const disableOthers = function (list, track) { + for (let i = 0; i < list.length; i++) { + if (!Object.keys(list[i]).length || track.id === list[i].id) { + continue; + } + // another video track is enabled, disable it + list[i].selected = false; + } +}; + +/** + * The current list of {@link VideoTrack} for a video. + * + * @see [Spec]{@link https://html.spec.whatwg.org/multipage/embedded-content.html#videotracklist} + * @extends TrackList + */ +class VideoTrackList extends TrackList { + /** + * Create an instance of this class. + * + * @param {VideoTrack[]} [tracks=[]] + * A list of `VideoTrack` to instantiate the list with. + */ + constructor(tracks = []) { + // make sure only 1 track is enabled + // sorted from last index to first index + for (let i = tracks.length - 1; i >= 0; i--) { + if (tracks[i].selected) { + disableOthers(tracks, tracks[i]); + break; + } + } + super(tracks); + this.changing_ = false; + + /** + * @member {number} VideoTrackList#selectedIndex + * The current index of the selected {@link VideoTrack`}. + */ + Object.defineProperty(this, 'selectedIndex', { + get() { + for (let i = 0; i < this.length; i++) { + if (this[i].selected) { + return i; + } + } + return -1; + }, + set() {} + }); + } + + /** + * Add a {@link VideoTrack} to the `VideoTrackList`. + * + * @param { import('./video-track').default } track + * The VideoTrack to add to the list + * + * @fires TrackList#addtrack + */ + addTrack(track) { + if (track.selected) { + disableOthers(this, track); + } + super.addTrack(track); + // native tracks don't have this + if (!track.addEventListener) { + return; + } + track.selectedChange_ = () => { + if (this.changing_) { + return; + } + this.changing_ = true; + disableOthers(this, track); + this.changing_ = false; + this.trigger('change'); + }; + + /** + * @listens VideoTrack#selectedchange + * @fires TrackList#change + */ + track.addEventListener('selectedchange', track.selectedChange_); + } + removeTrack(rtrack) { + super.removeTrack(rtrack); + if (rtrack.removeEventListener && rtrack.selectedChange_) { + rtrack.removeEventListener('selectedchange', rtrack.selectedChange_); + rtrack.selectedChange_ = null; + } + } +} + +/** + * @file text-track-list.js + */ + +/** + * The current list of {@link TextTrack} for a media file. + * + * @see [Spec]{@link https://html.spec.whatwg.org/multipage/embedded-content.html#texttracklist} + * @extends TrackList + */ +class TextTrackList extends TrackList { + /** + * Add a {@link TextTrack} to the `TextTrackList` + * + * @param { import('./text-track').default } track + * The text track to add to the list. + * + * @fires TrackList#addtrack + */ + addTrack(track) { + super.addTrack(track); + if (!this.queueChange_) { + this.queueChange_ = () => this.queueTrigger('change'); + } + if (!this.triggerSelectedlanguagechange) { + this.triggerSelectedlanguagechange_ = () => this.trigger('selectedlanguagechange'); + } + + /** + * @listens TextTrack#modechange + * @fires TrackList#change + */ + track.addEventListener('modechange', this.queueChange_); + const nonLanguageTextTrackKind = ['metadata', 'chapters']; + if (nonLanguageTextTrackKind.indexOf(track.kind) === -1) { + track.addEventListener('modechange', this.triggerSelectedlanguagechange_); + } + } + removeTrack(rtrack) { + super.removeTrack(rtrack); + + // manually remove the event handlers we added + if (rtrack.removeEventListener) { + if (this.queueChange_) { + rtrack.removeEventListener('modechange', this.queueChange_); + } + if (this.selectedlanguagechange_) { + rtrack.removeEventListener('modechange', this.triggerSelectedlanguagechange_); + } + } + } +} + +/** + * @file html-track-element-list.js + */ + +/** + * The current list of {@link HtmlTrackElement}s. + */ +class HtmlTrackElementList { + /** + * Create an instance of this class. + * + * @param {HtmlTrackElement[]} [tracks=[]] + * A list of `HtmlTrackElement` to instantiate the list with. + */ + constructor(trackElements = []) { + this.trackElements_ = []; + + /** + * @memberof HtmlTrackElementList + * @member {number} length + * The current number of `Track`s in the this Trackist. + * @instance + */ + Object.defineProperty(this, 'length', { + get() { + return this.trackElements_.length; + } + }); + for (let i = 0, length = trackElements.length; i < length; i++) { + this.addTrackElement_(trackElements[i]); + } + } + + /** + * Add an {@link HtmlTrackElement} to the `HtmlTrackElementList` + * + * @param {HtmlTrackElement} trackElement + * The track element to add to the list. + * + * @private + */ + addTrackElement_(trackElement) { + const index = this.trackElements_.length; + if (!('' + index in this)) { + Object.defineProperty(this, index, { + get() { + return this.trackElements_[index]; + } + }); + } + + // Do not add duplicate elements + if (this.trackElements_.indexOf(trackElement) === -1) { + this.trackElements_.push(trackElement); + } + } + + /** + * Get an {@link HtmlTrackElement} from the `HtmlTrackElementList` given an + * {@link TextTrack}. + * + * @param {TextTrack} track + * The track associated with a track element. + * + * @return {HtmlTrackElement|undefined} + * The track element that was found or undefined. + * + * @private + */ + getTrackElementByTrack_(track) { + let trackElement_; + for (let i = 0, length = this.trackElements_.length; i < length; i++) { + if (track === this.trackElements_[i].track) { + trackElement_ = this.trackElements_[i]; + break; + } + } + return trackElement_; + } + + /** + * Remove a {@link HtmlTrackElement} from the `HtmlTrackElementList` + * + * @param {HtmlTrackElement} trackElement + * The track element to remove from the list. + * + * @private + */ + removeTrackElement_(trackElement) { + for (let i = 0, length = this.trackElements_.length; i < length; i++) { + if (trackElement === this.trackElements_[i]) { + if (this.trackElements_[i].track && typeof this.trackElements_[i].track.off === 'function') { + this.trackElements_[i].track.off(); + } + if (typeof this.trackElements_[i].off === 'function') { + this.trackElements_[i].off(); + } + this.trackElements_.splice(i, 1); + break; + } + } + } +} + +/** + * @file text-track-cue-list.js + */ + +/** + * @typedef {Object} TextTrackCueList~TextTrackCue + * + * @property {string} id + * The unique id for this text track cue + * + * @property {number} startTime + * The start time for this text track cue + * + * @property {number} endTime + * The end time for this text track cue + * + * @property {boolean} pauseOnExit + * Pause when the end time is reached if true. + * + * @see [Spec]{@link https://html.spec.whatwg.org/multipage/embedded-content.html#texttrackcue} + */ + +/** + * A List of TextTrackCues. + * + * @see [Spec]{@link https://html.spec.whatwg.org/multipage/embedded-content.html#texttrackcuelist} + */ +class TextTrackCueList { + /** + * Create an instance of this class.. + * + * @param {Array} cues + * A list of cues to be initialized with + */ + constructor(cues) { + TextTrackCueList.prototype.setCues_.call(this, cues); + + /** + * @memberof TextTrackCueList + * @member {number} length + * The current number of `TextTrackCue`s in the TextTrackCueList. + * @instance + */ + Object.defineProperty(this, 'length', { + get() { + return this.length_; + } + }); + } + + /** + * A setter for cues in this list. Creates getters + * an an index for the cues. + * + * @param {Array} cues + * An array of cues to set + * + * @private + */ + setCues_(cues) { + const oldLength = this.length || 0; + let i = 0; + const l = cues.length; + this.cues_ = cues; + this.length_ = cues.length; + const defineProp = function (index) { + if (!('' + index in this)) { + Object.defineProperty(this, '' + index, { + get() { + return this.cues_[index]; + } + }); + } + }; + if (oldLength < l) { + i = oldLength; + for (; i < l; i++) { + defineProp.call(this, i); + } + } + } + + /** + * Get a `TextTrackCue` that is currently in the `TextTrackCueList` by id. + * + * @param {string} id + * The id of the cue that should be searched for. + * + * @return {TextTrackCueList~TextTrackCue|null} + * A single cue or null if none was found. + */ + getCueById(id) { + let result = null; + for (let i = 0, l = this.length; i < l; i++) { + const cue = this[i]; + if (cue.id === id) { + result = cue; + break; + } + } + return result; + } +} + +/** + * @file track-kinds.js + */ + +/** + * All possible `VideoTrackKind`s + * + * @see https://html.spec.whatwg.org/multipage/embedded-content.html#dom-videotrack-kind + * @typedef VideoTrack~Kind + * @enum + */ +const VideoTrackKind = { + alternative: 'alternative', + captions: 'captions', + main: 'main', + sign: 'sign', + subtitles: 'subtitles', + commentary: 'commentary' +}; + +/** + * All possible `AudioTrackKind`s + * + * @see https://html.spec.whatwg.org/multipage/embedded-content.html#dom-audiotrack-kind + * @typedef AudioTrack~Kind + * @enum + */ +const AudioTrackKind = { + 'alternative': 'alternative', + 'descriptions': 'descriptions', + 'main': 'main', + 'main-desc': 'main-desc', + 'translation': 'translation', + 'commentary': 'commentary' +}; + +/** + * All possible `TextTrackKind`s + * + * @see https://html.spec.whatwg.org/multipage/embedded-content.html#dom-texttrack-kind + * @typedef TextTrack~Kind + * @enum + */ +const TextTrackKind = { + subtitles: 'subtitles', + captions: 'captions', + descriptions: 'descriptions', + chapters: 'chapters', + metadata: 'metadata' +}; + +/** + * All possible `TextTrackMode`s + * + * @see https://html.spec.whatwg.org/multipage/embedded-content.html#texttrackmode + * @typedef TextTrack~Mode + * @enum + */ +const TextTrackMode = { + disabled: 'disabled', + hidden: 'hidden', + showing: 'showing' +}; + +/** + * @file track.js + */ + +/** + * A Track class that contains all of the common functionality for {@link AudioTrack}, + * {@link VideoTrack}, and {@link TextTrack}. + * + * > Note: This class should not be used directly + * + * @see {@link https://html.spec.whatwg.org/multipage/embedded-content.html} + * @extends EventTarget + * @abstract + */ +class Track extends EventTarget$2 { + /** + * Create an instance of this class. + * + * @param {Object} [options={}] + * Object of option names and values + * + * @param {string} [options.kind=''] + * A valid kind for the track type you are creating. + * + * @param {string} [options.id='vjs_track_' + Guid.newGUID()] + * A unique id for this AudioTrack. + * + * @param {string} [options.label=''] + * The menu label for this track. + * + * @param {string} [options.language=''] + * A valid two character language code. + * + * @abstract + */ + constructor(options = {}) { + super(); + const trackProps = { + id: options.id || 'vjs_track_' + newGUID(), + kind: options.kind || '', + language: options.language || '' + }; + let label = options.label || ''; + + /** + * @memberof Track + * @member {string} id + * The id of this track. Cannot be changed after creation. + * @instance + * + * @readonly + */ + + /** + * @memberof Track + * @member {string} kind + * The kind of track that this is. Cannot be changed after creation. + * @instance + * + * @readonly + */ + + /** + * @memberof Track + * @member {string} language + * The two letter language code for this track. Cannot be changed after + * creation. + * @instance + * + * @readonly + */ + + for (const key in trackProps) { + Object.defineProperty(this, key, { + get() { + return trackProps[key]; + }, + set() {} + }); + } + + /** + * @memberof Track + * @member {string} label + * The label of this track. Cannot be changed after creation. + * @instance + * + * @fires Track#labelchange + */ + Object.defineProperty(this, 'label', { + get() { + return label; + }, + set(newLabel) { + if (newLabel !== label) { + label = newLabel; + + /** + * An event that fires when label changes on this track. + * + * > Note: This is not part of the spec! + * + * @event Track#labelchange + * @type {Event} + */ + this.trigger('labelchange'); + } + } + }); + } +} + +/** + * @file url.js + * @module url + */ + +/** + * @typedef {Object} url:URLObject + * + * @property {string} protocol + * The protocol of the url that was parsed. + * + * @property {string} hostname + * The hostname of the url that was parsed. + * + * @property {string} port + * The port of the url that was parsed. + * + * @property {string} pathname + * The pathname of the url that was parsed. + * + * @property {string} search + * The search query of the url that was parsed. + * + * @property {string} hash + * The hash of the url that was parsed. + * + * @property {string} host + * The host of the url that was parsed. + */ + +/** + * Resolve and parse the elements of a URL. + * + * @function + * @param {String} url + * The url to parse + * + * @return {url:URLObject} + * An object of url details + */ +const parseUrl = function (url) { + // This entire method can be replace with URL once we are able to drop IE11 + + const props = ['protocol', 'hostname', 'port', 'pathname', 'search', 'hash', 'host']; + + // add the url to an anchor and let the browser parse the URL + const a = document.createElement('a'); + a.href = url; + + // Copy the specific URL properties to a new object + // This is also needed for IE because the anchor loses its + // properties when it's removed from the dom + const details = {}; + for (let i = 0; i < props.length; i++) { + details[props[i]] = a[props[i]]; + } + + // IE adds the port to the host property unlike everyone else. If + // a port identifier is added for standard ports, strip it. + if (details.protocol === 'http:') { + details.host = details.host.replace(/:80$/, ''); + } + if (details.protocol === 'https:') { + details.host = details.host.replace(/:443$/, ''); + } + if (!details.protocol) { + details.protocol = window$1.location.protocol; + } + + /* istanbul ignore if */ + if (!details.host) { + details.host = window$1.location.host; + } + return details; +}; + +/** + * Get absolute version of relative URL. + * + * @function + * @param {string} url + * URL to make absolute + * + * @return {string} + * Absolute URL + * + * @see http://stackoverflow.com/questions/470832/getting-an-absolute-url-from-a-relative-one-ie6-issue + */ +const getAbsoluteURL = function (url) { + // Check if absolute URL + if (!url.match(/^https?:\/\//)) { + // Add the url to an anchor and let the browser parse it to convert to an absolute url + const a = document.createElement('a'); + a.href = url; + url = a.href; + } + return url; +}; + +/** + * Returns the extension of the passed file name. It will return an empty string + * if passed an invalid path. + * + * @function + * @param {string} path + * The fileName path like '/path/to/file.mp4' + * + * @return {string} + * The extension in lower case or an empty string if no + * extension could be found. + */ +const getFileExtension = function (path) { + if (typeof path === 'string') { + const splitPathRe = /^(\/?)([\s\S]*?)((?:\.{1,2}|[^\/]+?)(\.([^\.\/\?]+)))(?:[\/]*|[\?].*)$/; + const pathParts = splitPathRe.exec(path); + if (pathParts) { + return pathParts.pop().toLowerCase(); + } + } + return ''; +}; + +/** + * Returns whether the url passed is a cross domain request or not. + * + * @function + * @param {string} url + * The url to check. + * + * @param {Object} [winLoc] + * the domain to check the url against, defaults to window.location + * + * @param {string} [winLoc.protocol] + * The window location protocol defaults to window.location.protocol + * + * @param {string} [winLoc.host] + * The window location host defaults to window.location.host + * + * @return {boolean} + * Whether it is a cross domain request or not. + */ +const isCrossOrigin = function (url, winLoc = window$1.location) { + const urlInfo = parseUrl(url); + + // IE8 protocol relative urls will return ':' for protocol + const srcProtocol = urlInfo.protocol === ':' ? winLoc.protocol : urlInfo.protocol; + + // Check if url is for another domain/origin + // IE8 doesn't know location.origin, so we won't rely on it here + const crossOrigin = srcProtocol + urlInfo.host !== winLoc.protocol + winLoc.host; + return crossOrigin; +}; + +var Url = /*#__PURE__*/Object.freeze({ + __proto__: null, + parseUrl: parseUrl, + getAbsoluteURL: getAbsoluteURL, + getFileExtension: getFileExtension, + isCrossOrigin: isCrossOrigin +}); + +/** + * @file text-track.js + */ + +/** + * Takes a webvtt file contents and parses it into cues + * + * @param {string} srcContent + * webVTT file contents + * + * @param {TextTrack} track + * TextTrack to add cues to. Cues come from the srcContent. + * + * @private + */ +const parseCues = function (srcContent, track) { + const parser = new window$1.WebVTT.Parser(window$1, window$1.vttjs, window$1.WebVTT.StringDecoder()); + const errors = []; + parser.oncue = function (cue) { + track.addCue(cue); + }; + parser.onparsingerror = function (error) { + errors.push(error); + }; + parser.onflush = function () { + track.trigger({ + type: 'loadeddata', + target: track + }); + }; + parser.parse(srcContent); + if (errors.length > 0) { + if (window$1.console && window$1.console.groupCollapsed) { + window$1.console.groupCollapsed(`Text Track parsing errors for ${track.src}`); + } + errors.forEach(error => log$1.error(error)); + if (window$1.console && window$1.console.groupEnd) { + window$1.console.groupEnd(); + } + } + parser.flush(); +}; + +/** + * Load a `TextTrack` from a specified url. + * + * @param {string} src + * Url to load track from. + * + * @param {TextTrack} track + * Track to add cues to. Comes from the content at the end of `url`. + * + * @private + */ +const loadTrack = function (src, track) { + const opts = { + uri: src + }; + const crossOrigin = isCrossOrigin(src); + if (crossOrigin) { + opts.cors = crossOrigin; + } + const withCredentials = track.tech_.crossOrigin() === 'use-credentials'; + if (withCredentials) { + opts.withCredentials = withCredentials; + } + XHR(opts, bind_(this, function (err, response, responseBody) { + if (err) { + return log$1.error(err, response); + } + track.loaded_ = true; + + // Make sure that vttjs has loaded, otherwise, wait till it finished loading + // NOTE: this is only used for the alt/video.novtt.js build + if (typeof window$1.WebVTT !== 'function') { + if (track.tech_) { + // to prevent use before define eslint error, we define loadHandler + // as a let here + track.tech_.any(['vttjsloaded', 'vttjserror'], event => { + if (event.type === 'vttjserror') { + log$1.error(`vttjs failed to load, stopping trying to process ${track.src}`); + return; + } + return parseCues(responseBody, track); + }); + } + } else { + parseCues(responseBody, track); + } + })); +}; + +/** + * A representation of a single `TextTrack`. + * + * @see [Spec]{@link https://html.spec.whatwg.org/multipage/embedded-content.html#texttrack} + * @extends Track + */ +class TextTrack extends Track { + /** + * Create an instance of this class. + * + * @param {Object} options={} + * Object of option names and values + * + * @param { import('../tech/tech').default } options.tech + * A reference to the tech that owns this TextTrack. + * + * @param {TextTrack~Kind} [options.kind='subtitles'] + * A valid text track kind. + * + * @param {TextTrack~Mode} [options.mode='disabled'] + * A valid text track mode. + * + * @param {string} [options.id='vjs_track_' + Guid.newGUID()] + * A unique id for this TextTrack. + * + * @param {string} [options.label=''] + * The menu label for this track. + * + * @param {string} [options.language=''] + * A valid two character language code. + * + * @param {string} [options.srclang=''] + * A valid two character language code. An alternative, but deprioritized + * version of `options.language` + * + * @param {string} [options.src] + * A url to TextTrack cues. + * + * @param {boolean} [options.default] + * If this track should default to on or off. + */ + constructor(options = {}) { + if (!options.tech) { + throw new Error('A tech was not provided.'); + } + const settings = merge$1(options, { + kind: TextTrackKind[options.kind] || 'subtitles', + language: options.language || options.srclang || '' + }); + let mode = TextTrackMode[settings.mode] || 'disabled'; + const default_ = settings.default; + if (settings.kind === 'metadata' || settings.kind === 'chapters') { + mode = 'hidden'; + } + super(settings); + this.tech_ = settings.tech; + this.cues_ = []; + this.activeCues_ = []; + this.preload_ = this.tech_.preloadTextTracks !== false; + const cues = new TextTrackCueList(this.cues_); + const activeCues = new TextTrackCueList(this.activeCues_); + let changed = false; + this.timeupdateHandler = bind_(this, function (event = {}) { + if (this.tech_.isDisposed()) { + return; + } + if (!this.tech_.isReady_) { + if (event.type !== 'timeupdate') { + this.rvf_ = this.tech_.requestVideoFrameCallback(this.timeupdateHandler); + } + return; + } + + // Accessing this.activeCues for the side-effects of updating itself + // due to its nature as a getter function. Do not remove or cues will + // stop updating! + // Use the setter to prevent deletion from uglify (pure_getters rule) + this.activeCues = this.activeCues; + if (changed) { + this.trigger('cuechange'); + changed = false; + } + if (event.type !== 'timeupdate') { + this.rvf_ = this.tech_.requestVideoFrameCallback(this.timeupdateHandler); + } + }); + const disposeHandler = () => { + this.stopTracking(); + }; + this.tech_.one('dispose', disposeHandler); + if (mode !== 'disabled') { + this.startTracking(); + } + Object.defineProperties(this, { + /** + * @memberof TextTrack + * @member {boolean} default + * If this track was set to be on or off by default. Cannot be changed after + * creation. + * @instance + * + * @readonly + */ + default: { + get() { + return default_; + }, + set() {} + }, + /** + * @memberof TextTrack + * @member {string} mode + * Set the mode of this TextTrack to a valid {@link TextTrack~Mode}. Will + * not be set if setting to an invalid mode. + * @instance + * + * @fires TextTrack#modechange + */ + mode: { + get() { + return mode; + }, + set(newMode) { + if (!TextTrackMode[newMode]) { + return; + } + if (mode === newMode) { + return; + } + mode = newMode; + if (!this.preload_ && mode !== 'disabled' && this.cues.length === 0) { + // On-demand load. + loadTrack(this.src, this); + } + this.stopTracking(); + if (mode !== 'disabled') { + this.startTracking(); + } + /** + * An event that fires when mode changes on this track. This allows + * the TextTrackList that holds this track to act accordingly. + * + * > Note: This is not part of the spec! + * + * @event TextTrack#modechange + * @type {Event} + */ + this.trigger('modechange'); + } + }, + /** + * @memberof TextTrack + * @member {TextTrackCueList} cues + * The text track cue list for this TextTrack. + * @instance + */ + cues: { + get() { + if (!this.loaded_) { + return null; + } + return cues; + }, + set() {} + }, + /** + * @memberof TextTrack + * @member {TextTrackCueList} activeCues + * The list text track cues that are currently active for this TextTrack. + * @instance + */ + activeCues: { + get() { + if (!this.loaded_) { + return null; + } + + // nothing to do + if (this.cues.length === 0) { + return activeCues; + } + const ct = this.tech_.currentTime(); + const active = []; + for (let i = 0, l = this.cues.length; i < l; i++) { + const cue = this.cues[i]; + if (cue.startTime <= ct && cue.endTime >= ct) { + active.push(cue); + } + } + changed = false; + if (active.length !== this.activeCues_.length) { + changed = true; + } else { + for (let i = 0; i < active.length; i++) { + if (this.activeCues_.indexOf(active[i]) === -1) { + changed = true; + } + } + } + this.activeCues_ = active; + activeCues.setCues_(this.activeCues_); + return activeCues; + }, + // /!\ Keep this setter empty (see the timeupdate handler above) + set() {} + } + }); + if (settings.src) { + this.src = settings.src; + if (!this.preload_) { + // Tracks will load on-demand. + // Act like we're loaded for other purposes. + this.loaded_ = true; + } + if (this.preload_ || settings.kind !== 'subtitles' && settings.kind !== 'captions') { + loadTrack(this.src, this); + } + } else { + this.loaded_ = true; + } + } + startTracking() { + // More precise cues based on requestVideoFrameCallback with a requestAnimationFram fallback + this.rvf_ = this.tech_.requestVideoFrameCallback(this.timeupdateHandler); + // Also listen to timeupdate in case rVFC/rAF stops (window in background, audio in video el) + this.tech_.on('timeupdate', this.timeupdateHandler); + } + stopTracking() { + if (this.rvf_) { + this.tech_.cancelVideoFrameCallback(this.rvf_); + this.rvf_ = undefined; + } + this.tech_.off('timeupdate', this.timeupdateHandler); + } + + /** + * Add a cue to the internal list of cues. + * + * @param {TextTrack~Cue} cue + * The cue to add to our internal list + */ + addCue(originalCue) { + let cue = originalCue; + + // Testing if the cue is a VTTCue in a way that survives minification + if (!('getCueAsHTML' in cue)) { + cue = new window$1.vttjs.VTTCue(originalCue.startTime, originalCue.endTime, originalCue.text); + for (const prop in originalCue) { + if (!(prop in cue)) { + cue[prop] = originalCue[prop]; + } + } + + // make sure that `id` is copied over + cue.id = originalCue.id; + cue.originalCue_ = originalCue; + } + const tracks = this.tech_.textTracks(); + for (let i = 0; i < tracks.length; i++) { + if (tracks[i] !== this) { + tracks[i].removeCue(cue); + } + } + this.cues_.push(cue); + this.cues.setCues_(this.cues_); + } + + /** + * Remove a cue from our internal list + * + * @param {TextTrack~Cue} removeCue + * The cue to remove from our internal list + */ + removeCue(removeCue) { + let i = this.cues_.length; + while (i--) { + const cue = this.cues_[i]; + if (cue === removeCue || cue.originalCue_ && cue.originalCue_ === removeCue) { + this.cues_.splice(i, 1); + this.cues.setCues_(this.cues_); + break; + } + } + } +} + +/** + * cuechange - One or more cues in the track have become active or stopped being active. + * @protected + */ +TextTrack.prototype.allowedEvents_ = { + cuechange: 'cuechange' +}; + +/** + * A representation of a single `AudioTrack`. If it is part of an {@link AudioTrackList} + * only one `AudioTrack` in the list will be enabled at a time. + * + * @see [Spec]{@link https://html.spec.whatwg.org/multipage/embedded-content.html#audiotrack} + * @extends Track + */ +class AudioTrack extends Track { + /** + * Create an instance of this class. + * + * @param {Object} [options={}] + * Object of option names and values + * + * @param {AudioTrack~Kind} [options.kind=''] + * A valid audio track kind + * + * @param {string} [options.id='vjs_track_' + Guid.newGUID()] + * A unique id for this AudioTrack. + * + * @param {string} [options.label=''] + * The menu label for this track. + * + * @param {string} [options.language=''] + * A valid two character language code. + * + * @param {boolean} [options.enabled] + * If this track is the one that is currently playing. If this track is part of + * an {@link AudioTrackList}, only one {@link AudioTrack} will be enabled. + */ + constructor(options = {}) { + const settings = merge$1(options, { + kind: AudioTrackKind[options.kind] || '' + }); + super(settings); + let enabled = false; + + /** + * @memberof AudioTrack + * @member {boolean} enabled + * If this `AudioTrack` is enabled or not. When setting this will + * fire {@link AudioTrack#enabledchange} if the state of enabled is changed. + * @instance + * + * @fires VideoTrack#selectedchange + */ + Object.defineProperty(this, 'enabled', { + get() { + return enabled; + }, + set(newEnabled) { + // an invalid or unchanged value + if (typeof newEnabled !== 'boolean' || newEnabled === enabled) { + return; + } + enabled = newEnabled; + + /** + * An event that fires when enabled changes on this track. This allows + * the AudioTrackList that holds this track to act accordingly. + * + * > Note: This is not part of the spec! Native tracks will do + * this internally without an event. + * + * @event AudioTrack#enabledchange + * @type {Event} + */ + this.trigger('enabledchange'); + } + }); + + // if the user sets this track to selected then + // set selected to that true value otherwise + // we keep it false + if (settings.enabled) { + this.enabled = settings.enabled; + } + this.loaded_ = true; + } +} + +/** + * A representation of a single `VideoTrack`. + * + * @see [Spec]{@link https://html.spec.whatwg.org/multipage/embedded-content.html#videotrack} + * @extends Track + */ +class VideoTrack extends Track { + /** + * Create an instance of this class. + * + * @param {Object} [options={}] + * Object of option names and values + * + * @param {string} [options.kind=''] + * A valid {@link VideoTrack~Kind} + * + * @param {string} [options.id='vjs_track_' + Guid.newGUID()] + * A unique id for this AudioTrack. + * + * @param {string} [options.label=''] + * The menu label for this track. + * + * @param {string} [options.language=''] + * A valid two character language code. + * + * @param {boolean} [options.selected] + * If this track is the one that is currently playing. + */ + constructor(options = {}) { + const settings = merge$1(options, { + kind: VideoTrackKind[options.kind] || '' + }); + super(settings); + let selected = false; + + /** + * @memberof VideoTrack + * @member {boolean} selected + * If this `VideoTrack` is selected or not. When setting this will + * fire {@link VideoTrack#selectedchange} if the state of selected changed. + * @instance + * + * @fires VideoTrack#selectedchange + */ + Object.defineProperty(this, 'selected', { + get() { + return selected; + }, + set(newSelected) { + // an invalid or unchanged value + if (typeof newSelected !== 'boolean' || newSelected === selected) { + return; + } + selected = newSelected; + + /** + * An event that fires when selected changes on this track. This allows + * the VideoTrackList that holds this track to act accordingly. + * + * > Note: This is not part of the spec! Native tracks will do + * this internally without an event. + * + * @event VideoTrack#selectedchange + * @type {Event} + */ + this.trigger('selectedchange'); + } + }); + + // if the user sets this track to selected then + // set selected to that true value otherwise + // we keep it false + if (settings.selected) { + this.selected = settings.selected; + } + } +} + +/** + * @file html-track-element.js + */ + +/** + * A single track represented in the DOM. + * + * @see [Spec]{@link https://html.spec.whatwg.org/multipage/embedded-content.html#htmltrackelement} + * @extends EventTarget + */ +class HTMLTrackElement extends EventTarget$2 { + /** + * Create an instance of this class. + * + * @param {Object} options={} + * Object of option names and values + * + * @param { import('../tech/tech').default } options.tech + * A reference to the tech that owns this HTMLTrackElement. + * + * @param {TextTrack~Kind} [options.kind='subtitles'] + * A valid text track kind. + * + * @param {TextTrack~Mode} [options.mode='disabled'] + * A valid text track mode. + * + * @param {string} [options.id='vjs_track_' + Guid.newGUID()] + * A unique id for this TextTrack. + * + * @param {string} [options.label=''] + * The menu label for this track. + * + * @param {string} [options.language=''] + * A valid two character language code. + * + * @param {string} [options.srclang=''] + * A valid two character language code. An alternative, but deprioritized + * version of `options.language` + * + * @param {string} [options.src] + * A url to TextTrack cues. + * + * @param {boolean} [options.default] + * If this track should default to on or off. + */ + constructor(options = {}) { + super(); + let readyState; + const track = new TextTrack(options); + this.kind = track.kind; + this.src = track.src; + this.srclang = track.language; + this.label = track.label; + this.default = track.default; + Object.defineProperties(this, { + /** + * @memberof HTMLTrackElement + * @member {HTMLTrackElement~ReadyState} readyState + * The current ready state of the track element. + * @instance + */ + readyState: { + get() { + return readyState; + } + }, + /** + * @memberof HTMLTrackElement + * @member {TextTrack} track + * The underlying TextTrack object. + * @instance + * + */ + track: { + get() { + return track; + } + } + }); + readyState = HTMLTrackElement.NONE; + + /** + * @listens TextTrack#loadeddata + * @fires HTMLTrackElement#load + */ + track.addEventListener('loadeddata', () => { + readyState = HTMLTrackElement.LOADED; + this.trigger({ + type: 'load', + target: this + }); + }); + } +} + +/** + * @protected + */ +HTMLTrackElement.prototype.allowedEvents_ = { + load: 'load' +}; + +/** + * The text track not loaded state. + * + * @type {number} + * @static + */ +HTMLTrackElement.NONE = 0; + +/** + * The text track loading state. + * + * @type {number} + * @static + */ +HTMLTrackElement.LOADING = 1; + +/** + * The text track loaded state. + * + * @type {number} + * @static + */ +HTMLTrackElement.LOADED = 2; + +/** + * The text track failed to load state. + * + * @type {number} + * @static + */ +HTMLTrackElement.ERROR = 3; + +/* + * This file contains all track properties that are used in + * player.js, tech.js, html5.js and possibly other techs in the future. + */ + +const NORMAL = { + audio: { + ListClass: AudioTrackList, + TrackClass: AudioTrack, + capitalName: 'Audio' + }, + video: { + ListClass: VideoTrackList, + TrackClass: VideoTrack, + capitalName: 'Video' + }, + text: { + ListClass: TextTrackList, + TrackClass: TextTrack, + capitalName: 'Text' + } +}; +Object.keys(NORMAL).forEach(function (type) { + NORMAL[type].getterName = `${type}Tracks`; + NORMAL[type].privateName = `${type}Tracks_`; +}); +const REMOTE = { + remoteText: { + ListClass: TextTrackList, + TrackClass: TextTrack, + capitalName: 'RemoteText', + getterName: 'remoteTextTracks', + privateName: 'remoteTextTracks_' + }, + remoteTextEl: { + ListClass: HtmlTrackElementList, + TrackClass: HTMLTrackElement, + capitalName: 'RemoteTextTrackEls', + getterName: 'remoteTextTrackEls', + privateName: 'remoteTextTrackEls_' + } +}; +const ALL = Object.assign({}, NORMAL, REMOTE); +REMOTE.names = Object.keys(REMOTE); +NORMAL.names = Object.keys(NORMAL); +ALL.names = [].concat(REMOTE.names).concat(NORMAL.names); + +/** + * @file tech.js + */ + +/** + * An Object containing a structure like: `{src: 'url', type: 'mimetype'}` or string + * that just contains the src url alone. + * * `var SourceObject = {src: 'http://ex.com/video.mp4', type: 'video/mp4'};` + * `var SourceString = 'http://example.com/some-video.mp4';` + * + * @typedef {Object|string} SourceObject + * + * @property {string} src + * The url to the source + * + * @property {string} type + * The mime type of the source + */ + +/** + * A function used by {@link Tech} to create a new {@link TextTrack}. + * + * @private + * + * @param {Tech} self + * An instance of the Tech class. + * + * @param {string} kind + * `TextTrack` kind (subtitles, captions, descriptions, chapters, or metadata) + * + * @param {string} [label] + * Label to identify the text track + * + * @param {string} [language] + * Two letter language abbreviation + * + * @param {Object} [options={}] + * An object with additional text track options + * + * @return {TextTrack} + * The text track that was created. + */ +function createTrackHelper(self, kind, label, language, options = {}) { + const tracks = self.textTracks(); + options.kind = kind; + if (label) { + options.label = label; + } + if (language) { + options.language = language; + } + options.tech = self; + const track = new ALL.text.TrackClass(options); + tracks.addTrack(track); + return track; +} + +/** + * This is the base class for media playback technology controllers, such as + * {@link HTML5} + * + * @extends Component + */ +class Tech extends Component$1 { + /** + * Create an instance of this Tech. + * + * @param {Object} [options] + * The key/value store of player options. + * + * @param {Function} [ready] + * Callback function to call when the `HTML5` Tech is ready. + */ + constructor(options = {}, ready = function () {}) { + // we don't want the tech to report user activity automatically. + // This is done manually in addControlsListeners + options.reportTouchActivity = false; + super(null, options, ready); + this.onDurationChange_ = e => this.onDurationChange(e); + this.trackProgress_ = e => this.trackProgress(e); + this.trackCurrentTime_ = e => this.trackCurrentTime(e); + this.stopTrackingCurrentTime_ = e => this.stopTrackingCurrentTime(e); + this.disposeSourceHandler_ = e => this.disposeSourceHandler(e); + this.queuedHanders_ = new Set(); + + // keep track of whether the current source has played at all to + // implement a very limited played() + this.hasStarted_ = false; + this.on('playing', function () { + this.hasStarted_ = true; + }); + this.on('loadstart', function () { + this.hasStarted_ = false; + }); + ALL.names.forEach(name => { + const props = ALL[name]; + if (options && options[props.getterName]) { + this[props.privateName] = options[props.getterName]; + } + }); + + // Manually track progress in cases where the browser/tech doesn't report it. + if (!this.featuresProgressEvents) { + this.manualProgressOn(); + } + + // Manually track timeupdates in cases where the browser/tech doesn't report it. + if (!this.featuresTimeupdateEvents) { + this.manualTimeUpdatesOn(); + } + ['Text', 'Audio', 'Video'].forEach(track => { + if (options[`native${track}Tracks`] === false) { + this[`featuresNative${track}Tracks`] = false; + } + }); + if (options.nativeCaptions === false || options.nativeTextTracks === false) { + this.featuresNativeTextTracks = false; + } else if (options.nativeCaptions === true || options.nativeTextTracks === true) { + this.featuresNativeTextTracks = true; + } + if (!this.featuresNativeTextTracks) { + this.emulateTextTracks(); + } + this.preloadTextTracks = options.preloadTextTracks !== false; + this.autoRemoteTextTracks_ = new ALL.text.ListClass(); + this.initTrackListeners(); + + // Turn on component tap events only if not using native controls + if (!options.nativeControlsForTouch) { + this.emitTapEvents(); + } + if (this.constructor) { + this.name_ = this.constructor.name || 'Unknown Tech'; + } + } + + /** + * A special function to trigger source set in a way that will allow player + * to re-trigger if the player or tech are not ready yet. + * + * @fires Tech#sourceset + * @param {string} src The source string at the time of the source changing. + */ + triggerSourceset(src) { + if (!this.isReady_) { + // on initial ready we have to trigger source set + // 1ms after ready so that player can watch for it. + this.one('ready', () => this.setTimeout(() => this.triggerSourceset(src), 1)); + } + + /** + * Fired when the source is set on the tech causing the media element + * to reload. + * + * @see {@link Player#event:sourceset} + * @event Tech#sourceset + * @type {Event} + */ + this.trigger({ + src, + type: 'sourceset' + }); + } + + /* Fallbacks for unsupported event types + ================================================================================ */ + + /** + * Polyfill the `progress` event for browsers that don't support it natively. + * + * @see {@link Tech#trackProgress} + */ + manualProgressOn() { + this.on('durationchange', this.onDurationChange_); + this.manualProgress = true; + + // Trigger progress watching when a source begins loading + this.one('ready', this.trackProgress_); + } + + /** + * Turn off the polyfill for `progress` events that was created in + * {@link Tech#manualProgressOn} + */ + manualProgressOff() { + this.manualProgress = false; + this.stopTrackingProgress(); + this.off('durationchange', this.onDurationChange_); + } + + /** + * This is used to trigger a `progress` event when the buffered percent changes. It + * sets an interval function that will be called every 500 milliseconds to check if the + * buffer end percent has changed. + * + * > This function is called by {@link Tech#manualProgressOn} + * + * @param {Event} event + * The `ready` event that caused this to run. + * + * @listens Tech#ready + * @fires Tech#progress + */ + trackProgress(event) { + this.stopTrackingProgress(); + this.progressInterval = this.setInterval(bind_(this, function () { + // Don't trigger unless buffered amount is greater than last time + + const numBufferedPercent = this.bufferedPercent(); + if (this.bufferedPercent_ !== numBufferedPercent) { + /** + * See {@link Player#progress} + * + * @event Tech#progress + * @type {Event} + */ + this.trigger('progress'); + } + this.bufferedPercent_ = numBufferedPercent; + if (numBufferedPercent === 1) { + this.stopTrackingProgress(); + } + }), 500); + } + + /** + * Update our internal duration on a `durationchange` event by calling + * {@link Tech#duration}. + * + * @param {Event} event + * The `durationchange` event that caused this to run. + * + * @listens Tech#durationchange + */ + onDurationChange(event) { + this.duration_ = this.duration(); + } + + /** + * Get and create a `TimeRange` object for buffering. + * + * @return { import('../utils/time').TimeRange } + * The time range object that was created. + */ + buffered() { + return createTimeRanges$1(0, 0); + } + + /** + * Get the percentage of the current video that is currently buffered. + * + * @return {number} + * A number from 0 to 1 that represents the decimal percentage of the + * video that is buffered. + * + */ + bufferedPercent() { + return bufferedPercent(this.buffered(), this.duration_); + } + + /** + * Turn off the polyfill for `progress` events that was created in + * {@link Tech#manualProgressOn} + * Stop manually tracking progress events by clearing the interval that was set in + * {@link Tech#trackProgress}. + */ + stopTrackingProgress() { + this.clearInterval(this.progressInterval); + } + + /** + * Polyfill the `timeupdate` event for browsers that don't support it. + * + * @see {@link Tech#trackCurrentTime} + */ + manualTimeUpdatesOn() { + this.manualTimeUpdates = true; + this.on('play', this.trackCurrentTime_); + this.on('pause', this.stopTrackingCurrentTime_); + } + + /** + * Turn off the polyfill for `timeupdate` events that was created in + * {@link Tech#manualTimeUpdatesOn} + */ + manualTimeUpdatesOff() { + this.manualTimeUpdates = false; + this.stopTrackingCurrentTime(); + this.off('play', this.trackCurrentTime_); + this.off('pause', this.stopTrackingCurrentTime_); + } + + /** + * Sets up an interval function to track current time and trigger `timeupdate` every + * 250 milliseconds. + * + * @listens Tech#play + * @triggers Tech#timeupdate + */ + trackCurrentTime() { + if (this.currentTimeInterval) { + this.stopTrackingCurrentTime(); + } + this.currentTimeInterval = this.setInterval(function () { + /** + * Triggered at an interval of 250ms to indicated that time is passing in the video. + * + * @event Tech#timeupdate + * @type {Event} + */ + this.trigger({ + type: 'timeupdate', + target: this, + manuallyTriggered: true + }); + + // 42 = 24 fps // 250 is what Webkit uses // FF uses 15 + }, 250); + } + + /** + * Stop the interval function created in {@link Tech#trackCurrentTime} so that the + * `timeupdate` event is no longer triggered. + * + * @listens {Tech#pause} + */ + stopTrackingCurrentTime() { + this.clearInterval(this.currentTimeInterval); + + // #1002 - if the video ends right before the next timeupdate would happen, + // the progress bar won't make it all the way to the end + this.trigger({ + type: 'timeupdate', + target: this, + manuallyTriggered: true + }); + } + + /** + * Turn off all event polyfills, clear the `Tech`s {@link AudioTrackList}, + * {@link VideoTrackList}, and {@link TextTrackList}, and dispose of this Tech. + * + * @fires Component#dispose + */ + dispose() { + // clear out all tracks because we can't reuse them between techs + this.clearTracks(NORMAL.names); + + // Turn off any manual progress or timeupdate tracking + if (this.manualProgress) { + this.manualProgressOff(); + } + if (this.manualTimeUpdates) { + this.manualTimeUpdatesOff(); + } + super.dispose(); + } + + /** + * Clear out a single `TrackList` or an array of `TrackLists` given their names. + * + * > Note: Techs without source handlers should call this between sources for `video` + * & `audio` tracks. You don't want to use them between tracks! + * + * @param {string[]|string} types + * TrackList names to clear, valid names are `video`, `audio`, and + * `text`. + */ + clearTracks(types) { + types = [].concat(types); + // clear out all tracks because we can't reuse them between techs + types.forEach(type => { + const list = this[`${type}Tracks`]() || []; + let i = list.length; + while (i--) { + const track = list[i]; + if (type === 'text') { + this.removeRemoteTextTrack(track); + } + list.removeTrack(track); + } + }); + } + + /** + * Remove any TextTracks added via addRemoteTextTrack that are + * flagged for automatic garbage collection + */ + cleanupAutoTextTracks() { + const list = this.autoRemoteTextTracks_ || []; + let i = list.length; + while (i--) { + const track = list[i]; + this.removeRemoteTextTrack(track); + } + } + + /** + * Reset the tech, which will removes all sources and reset the internal readyState. + * + * @abstract + */ + reset() {} + + /** + * Get the value of `crossOrigin` from the tech. + * + * @abstract + * + * @see {Html5#crossOrigin} + */ + crossOrigin() {} + + /** + * Set the value of `crossOrigin` on the tech. + * + * @abstract + * + * @param {string} crossOrigin the crossOrigin value + * @see {Html5#setCrossOrigin} + */ + setCrossOrigin() {} + + /** + * Get or set an error on the Tech. + * + * @param {MediaError} [err] + * Error to set on the Tech + * + * @return {MediaError|null} + * The current error object on the tech, or null if there isn't one. + */ + error(err) { + if (err !== undefined) { + this.error_ = new MediaError(err); + this.trigger('error'); + } + return this.error_; + } + + /** + * Returns the `TimeRange`s that have been played through for the current source. + * + * > NOTE: This implementation is incomplete. It does not track the played `TimeRange`. + * It only checks whether the source has played at all or not. + * + * @return { import('../utils/time').TimeRange } + * - A single time range if this video has played + * - An empty set of ranges if not. + */ + played() { + if (this.hasStarted_) { + return createTimeRanges$1(0, 0); + } + return createTimeRanges$1(); + } + + /** + * Start playback + * + * @abstract + * + * @see {Html5#play} + */ + play() {} + + /** + * Set whether we are scrubbing or not + * + * @abstract + * @param {boolean} _isScrubbing + * - true for we are currently scrubbing + * - false for we are no longer scrubbing + * + * @see {Html5#setScrubbing} + */ + setScrubbing(_isScrubbing) {} + + /** + * Get whether we are scrubbing or not + * + * @abstract + * + * @see {Html5#scrubbing} + */ + scrubbing() {} + + /** + * Causes a manual time update to occur if {@link Tech#manualTimeUpdatesOn} was + * previously called. + * + * @param {number} _seconds + * Set the current time of the media to this. + * @fires Tech#timeupdate + */ + setCurrentTime(_seconds) { + // improve the accuracy of manual timeupdates + if (this.manualTimeUpdates) { + /** + * A manual `timeupdate` event. + * + * @event Tech#timeupdate + * @type {Event} + */ + this.trigger({ + type: 'timeupdate', + target: this, + manuallyTriggered: true + }); + } + } + + /** + * Turn on listeners for {@link VideoTrackList}, {@link {AudioTrackList}, and + * {@link TextTrackList} events. + * + * This adds {@link EventTarget~EventListeners} for `addtrack`, and `removetrack`. + * + * @fires Tech#audiotrackchange + * @fires Tech#videotrackchange + * @fires Tech#texttrackchange + */ + initTrackListeners() { + /** + * Triggered when tracks are added or removed on the Tech {@link AudioTrackList} + * + * @event Tech#audiotrackchange + * @type {Event} + */ + + /** + * Triggered when tracks are added or removed on the Tech {@link VideoTrackList} + * + * @event Tech#videotrackchange + * @type {Event} + */ + + /** + * Triggered when tracks are added or removed on the Tech {@link TextTrackList} + * + * @event Tech#texttrackchange + * @type {Event} + */ + NORMAL.names.forEach(name => { + const props = NORMAL[name]; + const trackListChanges = () => { + this.trigger(`${name}trackchange`); + }; + const tracks = this[props.getterName](); + tracks.addEventListener('removetrack', trackListChanges); + tracks.addEventListener('addtrack', trackListChanges); + this.on('dispose', () => { + tracks.removeEventListener('removetrack', trackListChanges); + tracks.removeEventListener('addtrack', trackListChanges); + }); + }); + } + + /** + * Emulate TextTracks using vtt.js if necessary + * + * @fires Tech#vttjsloaded + * @fires Tech#vttjserror + */ + addWebVttScript_() { + if (window$1.WebVTT) { + return; + } + + // Initially, Tech.el_ is a child of a dummy-div wait until the Component system + // signals that the Tech is ready at which point Tech.el_ is part of the DOM + // before inserting the WebVTT script + if (document.body.contains(this.el())) { + // load via require if available and vtt.js script location was not passed in + // as an option. novtt builds will turn the above require call into an empty object + // which will cause this if check to always fail. + if (!this.options_['vtt.js'] && isPlain(vtt) && Object.keys(vtt).length > 0) { + this.trigger('vttjsloaded'); + return; + } + + // load vtt.js via the script location option or the cdn of no location was + // passed in + const script = document.createElement('script'); + script.src = this.options_['vtt.js'] || 'https://vjs.zencdn.net/vttjs/0.14.1/vtt.min.js'; + script.onload = () => { + /** + * Fired when vtt.js is loaded. + * + * @event Tech#vttjsloaded + * @type {Event} + */ + this.trigger('vttjsloaded'); + }; + script.onerror = () => { + /** + * Fired when vtt.js was not loaded due to an error + * + * @event Tech#vttjsloaded + * @type {Event} + */ + this.trigger('vttjserror'); + }; + this.on('dispose', () => { + script.onload = null; + script.onerror = null; + }); + // but have not loaded yet and we set it to true before the inject so that + // we don't overwrite the injected window.WebVTT if it loads right away + window$1.WebVTT = true; + this.el().parentNode.appendChild(script); + } else { + this.ready(this.addWebVttScript_); + } + } + + /** + * Emulate texttracks + * + */ + emulateTextTracks() { + const tracks = this.textTracks(); + const remoteTracks = this.remoteTextTracks(); + const handleAddTrack = e => tracks.addTrack(e.track); + const handleRemoveTrack = e => tracks.removeTrack(e.track); + remoteTracks.on('addtrack', handleAddTrack); + remoteTracks.on('removetrack', handleRemoveTrack); + this.addWebVttScript_(); + const updateDisplay = () => this.trigger('texttrackchange'); + const textTracksChanges = () => { + updateDisplay(); + for (let i = 0; i < tracks.length; i++) { + const track = tracks[i]; + track.removeEventListener('cuechange', updateDisplay); + if (track.mode === 'showing') { + track.addEventListener('cuechange', updateDisplay); + } + } + }; + textTracksChanges(); + tracks.addEventListener('change', textTracksChanges); + tracks.addEventListener('addtrack', textTracksChanges); + tracks.addEventListener('removetrack', textTracksChanges); + this.on('dispose', function () { + remoteTracks.off('addtrack', handleAddTrack); + remoteTracks.off('removetrack', handleRemoveTrack); + tracks.removeEventListener('change', textTracksChanges); + tracks.removeEventListener('addtrack', textTracksChanges); + tracks.removeEventListener('removetrack', textTracksChanges); + for (let i = 0; i < tracks.length; i++) { + const track = tracks[i]; + track.removeEventListener('cuechange', updateDisplay); + } + }); + } + + /** + * Create and returns a remote {@link TextTrack} object. + * + * @param {string} kind + * `TextTrack` kind (subtitles, captions, descriptions, chapters, or metadata) + * + * @param {string} [label] + * Label to identify the text track + * + * @param {string} [language] + * Two letter language abbreviation + * + * @return {TextTrack} + * The TextTrack that gets created. + */ + addTextTrack(kind, label, language) { + if (!kind) { + throw new Error('TextTrack kind is required but was not provided'); + } + return createTrackHelper(this, kind, label, language); + } + + /** + * Create an emulated TextTrack for use by addRemoteTextTrack + * + * This is intended to be overridden by classes that inherit from + * Tech in order to create native or custom TextTracks. + * + * @param {Object} options + * The object should contain the options to initialize the TextTrack with. + * + * @param {string} [options.kind] + * `TextTrack` kind (subtitles, captions, descriptions, chapters, or metadata). + * + * @param {string} [options.label]. + * Label to identify the text track + * + * @param {string} [options.language] + * Two letter language abbreviation. + * + * @return {HTMLTrackElement} + * The track element that gets created. + */ + createRemoteTextTrack(options) { + const track = merge$1(options, { + tech: this + }); + return new REMOTE.remoteTextEl.TrackClass(track); + } + + /** + * Creates a remote text track object and returns an html track element. + * + * > Note: This can be an emulated {@link HTMLTrackElement} or a native one. + * + * @param {Object} options + * See {@link Tech#createRemoteTextTrack} for more detailed properties. + * + * @param {boolean} [manualCleanup=false] + * - When false: the TextTrack will be automatically removed from the video + * element whenever the source changes + * - When True: The TextTrack will have to be cleaned up manually + * + * @return {HTMLTrackElement} + * An Html Track Element. + * + */ + addRemoteTextTrack(options = {}, manualCleanup) { + const htmlTrackElement = this.createRemoteTextTrack(options); + if (typeof manualCleanup !== 'boolean') { + manualCleanup = false; + } + + // store HTMLTrackElement and TextTrack to remote list + this.remoteTextTrackEls().addTrackElement_(htmlTrackElement); + this.remoteTextTracks().addTrack(htmlTrackElement.track); + if (manualCleanup === false) { + // create the TextTrackList if it doesn't exist + this.ready(() => this.autoRemoteTextTracks_.addTrack(htmlTrackElement.track)); + } + return htmlTrackElement; + } + + /** + * Remove a remote text track from the remote `TextTrackList`. + * + * @param {TextTrack} track + * `TextTrack` to remove from the `TextTrackList` + */ + removeRemoteTextTrack(track) { + const trackElement = this.remoteTextTrackEls().getTrackElementByTrack_(track); + + // remove HTMLTrackElement and TextTrack from remote list + this.remoteTextTrackEls().removeTrackElement_(trackElement); + this.remoteTextTracks().removeTrack(track); + this.autoRemoteTextTracks_.removeTrack(track); + } + + /** + * Gets available media playback quality metrics as specified by the W3C's Media + * Playback Quality API. + * + * @see [Spec]{@link https://wicg.github.io/media-playback-quality} + * + * @return {Object} + * An object with supported media playback quality metrics + * + * @abstract + */ + getVideoPlaybackQuality() { + return {}; + } + + /** + * Attempt to create a floating video window always on top of other windows + * so that users may continue consuming media while they interact with other + * content sites, or applications on their device. + * + * @see [Spec]{@link https://wicg.github.io/picture-in-picture} + * + * @return {Promise|undefined} + * A promise with a Picture-in-Picture window if the browser supports + * Promises (or one was passed in as an option). It returns undefined + * otherwise. + * + * @abstract + */ + requestPictureInPicture() { + return Promise.reject(); + } + + /** + * A method to check for the value of the 'disablePictureInPicture' <video> property. + * Defaults to true, as it should be considered disabled if the tech does not support pip + * + * @abstract + */ + disablePictureInPicture() { + return true; + } + + /** + * A method to set or unset the 'disablePictureInPicture' <video> property. + * + * @abstract + */ + setDisablePictureInPicture() {} + + /** + * A fallback implementation of requestVideoFrameCallback using requestAnimationFrame + * + * @param {function} cb + * @return {number} request id + */ + requestVideoFrameCallback(cb) { + const id = newGUID(); + if (!this.isReady_ || this.paused()) { + this.queuedHanders_.add(id); + this.one('playing', () => { + if (this.queuedHanders_.has(id)) { + this.queuedHanders_.delete(id); + cb(); + } + }); + } else { + this.requestNamedAnimationFrame(id, cb); + } + return id; + } + + /** + * A fallback implementation of cancelVideoFrameCallback + * + * @param {number} id id of callback to be cancelled + */ + cancelVideoFrameCallback(id) { + if (this.queuedHanders_.has(id)) { + this.queuedHanders_.delete(id); + } else { + this.cancelNamedAnimationFrame(id); + } + } + + /** + * A method to set a poster from a `Tech`. + * + * @abstract + */ + setPoster() {} + + /** + * A method to check for the presence of the 'playsinline' <video> attribute. + * + * @abstract + */ + playsinline() {} + + /** + * A method to set or unset the 'playsinline' <video> attribute. + * + * @abstract + */ + setPlaysinline() {} + + /** + * Attempt to force override of native audio tracks. + * + * @param {boolean} override - If set to true native audio will be overridden, + * otherwise native audio will potentially be used. + * + * @abstract + */ + overrideNativeAudioTracks(override) {} + + /** + * Attempt to force override of native video tracks. + * + * @param {boolean} override - If set to true native video will be overridden, + * otherwise native video will potentially be used. + * + * @abstract + */ + overrideNativeVideoTracks(override) {} + + /** + * Check if the tech can support the given mime-type. + * + * The base tech does not support any type, but source handlers might + * overwrite this. + * + * @param {string} _type + * The mimetype to check for support + * + * @return {string} + * 'probably', 'maybe', or empty string + * + * @see [Spec]{@link https://developer.mozilla.org/en-US/docs/Web/API/HTMLMediaElement/canPlayType} + * + * @abstract + */ + canPlayType(_type) { + return ''; + } + + /** + * Check if the type is supported by this tech. + * + * The base tech does not support any type, but source handlers might + * overwrite this. + * + * @param {string} _type + * The media type to check + * @return {string} Returns the native video element's response + */ + static canPlayType(_type) { + return ''; + } + + /** + * Check if the tech can support the given source + * + * @param {Object} srcObj + * The source object + * @param {Object} options + * The options passed to the tech + * @return {string} 'probably', 'maybe', or '' (empty string) + */ + static canPlaySource(srcObj, options) { + return Tech.canPlayType(srcObj.type); + } + + /* + * Return whether the argument is a Tech or not. + * Can be passed either a Class like `Html5` or a instance like `player.tech_` + * + * @param {Object} component + * The item to check + * + * @return {boolean} + * Whether it is a tech or not + * - True if it is a tech + * - False if it is not + */ + static isTech(component) { + return component.prototype instanceof Tech || component instanceof Tech || component === Tech; + } + + /** + * Registers a `Tech` into a shared list for videojs. + * + * @param {string} name + * Name of the `Tech` to register. + * + * @param {Object} tech + * The `Tech` class to register. + */ + static registerTech(name, tech) { + if (!Tech.techs_) { + Tech.techs_ = {}; + } + if (!Tech.isTech(tech)) { + throw new Error(`Tech ${name} must be a Tech`); + } + if (!Tech.canPlayType) { + throw new Error('Techs must have a static canPlayType method on them'); + } + if (!Tech.canPlaySource) { + throw new Error('Techs must have a static canPlaySource method on them'); + } + name = toTitleCase$1(name); + Tech.techs_[name] = tech; + Tech.techs_[toLowerCase(name)] = tech; + if (name !== 'Tech') { + // camel case the techName for use in techOrder + Tech.defaultTechOrder_.push(name); + } + return tech; + } + + /** + * Get a `Tech` from the shared list by name. + * + * @param {string} name + * `camelCase` or `TitleCase` name of the Tech to get + * + * @return {Tech|undefined} + * The `Tech` or undefined if there was no tech with the name requested. + */ + static getTech(name) { + if (!name) { + return; + } + if (Tech.techs_ && Tech.techs_[name]) { + return Tech.techs_[name]; + } + name = toTitleCase$1(name); + if (window$1 && window$1.videojs && window$1.videojs[name]) { + log$1.warn(`The ${name} tech was added to the videojs object when it should be registered using videojs.registerTech(name, tech)`); + return window$1.videojs[name]; + } + } +} + +/** + * Get the {@link VideoTrackList} + * + * @returns {VideoTrackList} + * @method Tech.prototype.videoTracks + */ + +/** + * Get the {@link AudioTrackList} + * + * @returns {AudioTrackList} + * @method Tech.prototype.audioTracks + */ + +/** + * Get the {@link TextTrackList} + * + * @returns {TextTrackList} + * @method Tech.prototype.textTracks + */ + +/** + * Get the remote element {@link TextTrackList} + * + * @returns {TextTrackList} + * @method Tech.prototype.remoteTextTracks + */ + +/** + * Get the remote element {@link HtmlTrackElementList} + * + * @returns {HtmlTrackElementList} + * @method Tech.prototype.remoteTextTrackEls + */ + +ALL.names.forEach(function (name) { + const props = ALL[name]; + Tech.prototype[props.getterName] = function () { + this[props.privateName] = this[props.privateName] || new props.ListClass(); + return this[props.privateName]; + }; +}); + +/** + * List of associated text tracks + * + * @type {TextTrackList} + * @private + * @property Tech#textTracks_ + */ + +/** + * List of associated audio tracks. + * + * @type {AudioTrackList} + * @private + * @property Tech#audioTracks_ + */ + +/** + * List of associated video tracks. + * + * @type {VideoTrackList} + * @private + * @property Tech#videoTracks_ + */ + +/** + * Boolean indicating whether the `Tech` supports volume control. + * + * @type {boolean} + * @default + */ +Tech.prototype.featuresVolumeControl = true; + +/** + * Boolean indicating whether the `Tech` supports muting volume. + * + * @type {boolean} + * @default + */ +Tech.prototype.featuresMuteControl = true; + +/** + * Boolean indicating whether the `Tech` supports fullscreen resize control. + * Resizing plugins using request fullscreen reloads the plugin + * + * @type {boolean} + * @default + */ +Tech.prototype.featuresFullscreenResize = false; + +/** + * Boolean indicating whether the `Tech` supports changing the speed at which the video + * plays. Examples: + * - Set player to play 2x (twice) as fast + * - Set player to play 0.5x (half) as fast + * + * @type {boolean} + * @default + */ +Tech.prototype.featuresPlaybackRate = false; + +/** + * Boolean indicating whether the `Tech` supports the `progress` event. + * This will be used to determine if {@link Tech#manualProgressOn} should be called. + * + * @type {boolean} + * @default + */ +Tech.prototype.featuresProgressEvents = false; + +/** + * Boolean indicating whether the `Tech` supports the `sourceset` event. + * + * A tech should set this to `true` and then use {@link Tech#triggerSourceset} + * to trigger a {@link Tech#event:sourceset} at the earliest time after getting + * a new source. + * + * @type {boolean} + * @default + */ +Tech.prototype.featuresSourceset = false; + +/** + * Boolean indicating whether the `Tech` supports the `timeupdate` event. + * This will be used to determine if {@link Tech#manualTimeUpdates} should be called. + * + * @type {boolean} + * @default + */ +Tech.prototype.featuresTimeupdateEvents = false; + +/** + * Boolean indicating whether the `Tech` supports the native `TextTrack`s. + * This will help us integrate with native `TextTrack`s if the browser supports them. + * + * @type {boolean} + * @default + */ +Tech.prototype.featuresNativeTextTracks = false; + +/** + * Boolean indicating whether the `Tech` supports `requestVideoFrameCallback`. + * + * @type {boolean} + * @default + */ +Tech.prototype.featuresVideoFrameCallback = false; + +/** + * A functional mixin for techs that want to use the Source Handler pattern. + * Source handlers are scripts for handling specific formats. + * The source handler pattern is used for adaptive formats (HLS, DASH) that + * manually load video data and feed it into a Source Buffer (Media Source Extensions) + * Example: `Tech.withSourceHandlers.call(MyTech);` + * + * @param {Tech} _Tech + * The tech to add source handler functions to. + * + * @mixes Tech~SourceHandlerAdditions + */ +Tech.withSourceHandlers = function (_Tech) { + /** + * Register a source handler + * + * @param {Function} handler + * The source handler class + * + * @param {number} [index] + * Register it at the following index + */ + _Tech.registerSourceHandler = function (handler, index) { + let handlers = _Tech.sourceHandlers; + if (!handlers) { + handlers = _Tech.sourceHandlers = []; + } + if (index === undefined) { + // add to the end of the list + index = handlers.length; + } + handlers.splice(index, 0, handler); + }; + + /** + * Check if the tech can support the given type. Also checks the + * Techs sourceHandlers. + * + * @param {string} type + * The mimetype to check. + * + * @return {string} + * 'probably', 'maybe', or '' (empty string) + */ + _Tech.canPlayType = function (type) { + const handlers = _Tech.sourceHandlers || []; + let can; + for (let i = 0; i < handlers.length; i++) { + can = handlers[i].canPlayType(type); + if (can) { + return can; + } + } + return ''; + }; + + /** + * Returns the first source handler that supports the source. + * + * TODO: Answer question: should 'probably' be prioritized over 'maybe' + * + * @param {SourceObject} source + * The source object + * + * @param {Object} options + * The options passed to the tech + * + * @return {SourceHandler|null} + * The first source handler that supports the source or null if + * no SourceHandler supports the source + */ + _Tech.selectSourceHandler = function (source, options) { + const handlers = _Tech.sourceHandlers || []; + let can; + for (let i = 0; i < handlers.length; i++) { + can = handlers[i].canHandleSource(source, options); + if (can) { + return handlers[i]; + } + } + return null; + }; + + /** + * Check if the tech can support the given source. + * + * @param {SourceObject} srcObj + * The source object + * + * @param {Object} options + * The options passed to the tech + * + * @return {string} + * 'probably', 'maybe', or '' (empty string) + */ + _Tech.canPlaySource = function (srcObj, options) { + const sh = _Tech.selectSourceHandler(srcObj, options); + if (sh) { + return sh.canHandleSource(srcObj, options); + } + return ''; + }; + + /** + * When using a source handler, prefer its implementation of + * any function normally provided by the tech. + */ + const deferrable = ['seekable', 'seeking', 'duration']; + + /** + * A wrapper around {@link Tech#seekable} that will call a `SourceHandler`s seekable + * function if it exists, with a fallback to the Techs seekable function. + * + * @method _Tech.seekable + */ + + /** + * A wrapper around {@link Tech#duration} that will call a `SourceHandler`s duration + * function if it exists, otherwise it will fallback to the techs duration function. + * + * @method _Tech.duration + */ + + deferrable.forEach(function (fnName) { + const originalFn = this[fnName]; + if (typeof originalFn !== 'function') { + return; + } + this[fnName] = function () { + if (this.sourceHandler_ && this.sourceHandler_[fnName]) { + return this.sourceHandler_[fnName].apply(this.sourceHandler_, arguments); + } + return originalFn.apply(this, arguments); + }; + }, _Tech.prototype); + + /** + * Create a function for setting the source using a source object + * and source handlers. + * Should never be called unless a source handler was found. + * + * @param {SourceObject} source + * A source object with src and type keys + */ + _Tech.prototype.setSource = function (source) { + let sh = _Tech.selectSourceHandler(source, this.options_); + if (!sh) { + // Fall back to a native source handler when unsupported sources are + // deliberately set + if (_Tech.nativeSourceHandler) { + sh = _Tech.nativeSourceHandler; + } else { + log$1.error('No source handler found for the current source.'); + } + } + + // Dispose any existing source handler + this.disposeSourceHandler(); + this.off('dispose', this.disposeSourceHandler_); + if (sh !== _Tech.nativeSourceHandler) { + this.currentSource_ = source; + } + this.sourceHandler_ = sh.handleSource(source, this, this.options_); + this.one('dispose', this.disposeSourceHandler_); + }; + + /** + * Clean up any existing SourceHandlers and listeners when the Tech is disposed. + * + * @listens Tech#dispose + */ + _Tech.prototype.disposeSourceHandler = function () { + // if we have a source and get another one + // then we are loading something new + // than clear all of our current tracks + if (this.currentSource_) { + this.clearTracks(['audio', 'video']); + this.currentSource_ = null; + } + + // always clean up auto-text tracks + this.cleanupAutoTextTracks(); + if (this.sourceHandler_) { + if (this.sourceHandler_.dispose) { + this.sourceHandler_.dispose(); + } + this.sourceHandler_ = null; + } + }; +}; + +// The base Tech class needs to be registered as a Component. It is the only +// Tech that can be registered as a Component. +Component$1.registerComponent('Tech', Tech); +Tech.registerTech('Tech', Tech); + +/** + * A list of techs that should be added to techOrder on Players + * + * @private + */ +Tech.defaultTechOrder_ = []; + +/** + * @file middleware.js + * @module middleware + */ +const middlewares = {}; +const middlewareInstances = {}; +const TERMINATOR = {}; + +/** + * A middleware object is a plain JavaScript object that has methods that + * match the {@link Tech} methods found in the lists of allowed + * {@link module:middleware.allowedGetters|getters}, + * {@link module:middleware.allowedSetters|setters}, and + * {@link module:middleware.allowedMediators|mediators}. + * + * @typedef {Object} MiddlewareObject + */ + +/** + * A middleware factory function that should return a + * {@link module:middleware~MiddlewareObject|MiddlewareObject}. + * + * This factory will be called for each player when needed, with the player + * passed in as an argument. + * + * @callback MiddlewareFactory + * @param { import('../player').default } player + * A Video.js player. + */ + +/** + * Define a middleware that the player should use by way of a factory function + * that returns a middleware object. + * + * @param {string} type + * The MIME type to match or `"*"` for all MIME types. + * + * @param {MiddlewareFactory} middleware + * A middleware factory function that will be executed for + * matching types. + */ +function use(type, middleware) { + middlewares[type] = middlewares[type] || []; + middlewares[type].push(middleware); +} + +/** + * Asynchronously sets a source using middleware by recursing through any + * matching middlewares and calling `setSource` on each, passing along the + * previous returned value each time. + * + * @param { import('../player').default } player + * A {@link Player} instance. + * + * @param {Tech~SourceObject} src + * A source object. + * + * @param {Function} + * The next middleware to run. + */ +function setSource(player, src, next) { + player.setTimeout(() => setSourceHelper(src, middlewares[src.type], next, player), 1); +} + +/** + * When the tech is set, passes the tech to each middleware's `setTech` method. + * + * @param {Object[]} middleware + * An array of middleware instances. + * + * @param { import('../tech/tech').default } tech + * A Video.js tech. + */ +function setTech(middleware, tech) { + middleware.forEach(mw => mw.setTech && mw.setTech(tech)); +} + +/** + * Calls a getter on the tech first, through each middleware + * from right to left to the player. + * + * @param {Object[]} middleware + * An array of middleware instances. + * + * @param { import('../tech/tech').default } tech + * The current tech. + * + * @param {string} method + * A method name. + * + * @return {*} + * The final value from the tech after middleware has intercepted it. + */ +function get(middleware, tech, method) { + return middleware.reduceRight(middlewareIterator(method), tech[method]()); +} + +/** + * Takes the argument given to the player and calls the setter method on each + * middleware from left to right to the tech. + * + * @param {Object[]} middleware + * An array of middleware instances. + * + * @param { import('../tech/tech').default } tech + * The current tech. + * + * @param {string} method + * A method name. + * + * @param {*} arg + * The value to set on the tech. + * + * @return {*} + * The return value of the `method` of the `tech`. + */ +function set(middleware, tech, method, arg) { + return tech[method](middleware.reduce(middlewareIterator(method), arg)); +} + +/** + * Takes the argument given to the player and calls the `call` version of the + * method on each middleware from left to right. + * + * Then, call the passed in method on the tech and return the result unchanged + * back to the player, through middleware, this time from right to left. + * + * @param {Object[]} middleware + * An array of middleware instances. + * + * @param { import('../tech/tech').default } tech + * The current tech. + * + * @param {string} method + * A method name. + * + * @param {*} arg + * The value to set on the tech. + * + * @return {*} + * The return value of the `method` of the `tech`, regardless of the + * return values of middlewares. + */ +function mediate(middleware, tech, method, arg = null) { + const callMethod = 'call' + toTitleCase$1(method); + const middlewareValue = middleware.reduce(middlewareIterator(callMethod), arg); + const terminated = middlewareValue === TERMINATOR; + // deprecated. The `null` return value should instead return TERMINATOR to + // prevent confusion if a techs method actually returns null. + const returnValue = terminated ? null : tech[method](middlewareValue); + executeRight(middleware, method, returnValue, terminated); + return returnValue; +} + +/** + * Enumeration of allowed getters where the keys are method names. + * + * @type {Object} + */ +const allowedGetters = { + buffered: 1, + currentTime: 1, + duration: 1, + muted: 1, + played: 1, + paused: 1, + seekable: 1, + volume: 1, + ended: 1 +}; + +/** + * Enumeration of allowed setters where the keys are method names. + * + * @type {Object} + */ +const allowedSetters = { + setCurrentTime: 1, + setMuted: 1, + setVolume: 1 +}; + +/** + * Enumeration of allowed mediators where the keys are method names. + * + * @type {Object} + */ +const allowedMediators = { + play: 1, + pause: 1 +}; +function middlewareIterator(method) { + return (value, mw) => { + // if the previous middleware terminated, pass along the termination + if (value === TERMINATOR) { + return TERMINATOR; + } + if (mw[method]) { + return mw[method](value); + } + return value; + }; +} +function executeRight(mws, method, value, terminated) { + for (let i = mws.length - 1; i >= 0; i--) { + const mw = mws[i]; + if (mw[method]) { + mw[method](terminated, value); + } + } +} + +/** + * Clear the middleware cache for a player. + * + * @param { import('../player').default } player + * A {@link Player} instance. + */ +function clearCacheForPlayer(player) { + middlewareInstances[player.id()] = null; +} + +/** + * { + * [playerId]: [[mwFactory, mwInstance], ...] + * } + * + * @private + */ +function getOrCreateFactory(player, mwFactory) { + const mws = middlewareInstances[player.id()]; + let mw = null; + if (mws === undefined || mws === null) { + mw = mwFactory(player); + middlewareInstances[player.id()] = [[mwFactory, mw]]; + return mw; + } + for (let i = 0; i < mws.length; i++) { + const [mwf, mwi] = mws[i]; + if (mwf !== mwFactory) { + continue; + } + mw = mwi; + } + if (mw === null) { + mw = mwFactory(player); + mws.push([mwFactory, mw]); + } + return mw; +} +function setSourceHelper(src = {}, middleware = [], next, player, acc = [], lastRun = false) { + const [mwFactory, ...mwrest] = middleware; + + // if mwFactory is a string, then we're at a fork in the road + if (typeof mwFactory === 'string') { + setSourceHelper(src, middlewares[mwFactory], next, player, acc, lastRun); + + // if we have an mwFactory, call it with the player to get the mw, + // then call the mw's setSource method + } else if (mwFactory) { + const mw = getOrCreateFactory(player, mwFactory); + + // if setSource isn't present, implicitly select this middleware + if (!mw.setSource) { + acc.push(mw); + return setSourceHelper(src, mwrest, next, player, acc, lastRun); + } + mw.setSource(Object.assign({}, src), function (err, _src) { + // something happened, try the next middleware on the current level + // make sure to use the old src + if (err) { + return setSourceHelper(src, mwrest, next, player, acc, lastRun); + } + + // we've succeeded, now we need to go deeper + acc.push(mw); + + // if it's the same type, continue down the current chain + // otherwise, we want to go down the new chain + setSourceHelper(_src, src.type === _src.type ? mwrest : middlewares[_src.type], next, player, acc, lastRun); + }); + } else if (mwrest.length) { + setSourceHelper(src, mwrest, next, player, acc, lastRun); + } else if (lastRun) { + next(src, acc); + } else { + setSourceHelper(src, middlewares['*'], next, player, acc, true); + } +} + +/** + * Mimetypes + * + * @see https://www.iana.org/assignments/media-types/media-types.xhtml + * @typedef Mimetypes~Kind + * @enum + */ +const MimetypesKind = { + opus: 'video/ogg', + ogv: 'video/ogg', + mp4: 'video/mp4', + mov: 'video/mp4', + m4v: 'video/mp4', + mkv: 'video/x-matroska', + m4a: 'audio/mp4', + mp3: 'audio/mpeg', + aac: 'audio/aac', + caf: 'audio/x-caf', + flac: 'audio/flac', + oga: 'audio/ogg', + wav: 'audio/wav', + m3u8: 'application/x-mpegURL', + mpd: 'application/dash+xml', + jpg: 'image/jpeg', + jpeg: 'image/jpeg', + gif: 'image/gif', + png: 'image/png', + svg: 'image/svg+xml', + webp: 'image/webp' +}; + +/** + * Get the mimetype of a given src url if possible + * + * @param {string} src + * The url to the src + * + * @return {string} + * return the mimetype if it was known or empty string otherwise + */ +const getMimetype = function (src = '') { + const ext = getFileExtension(src); + const mimetype = MimetypesKind[ext.toLowerCase()]; + return mimetype || ''; +}; + +/** + * Find the mime type of a given source string if possible. Uses the player + * source cache. + * + * @param { import('../player').default } player + * The player object + * + * @param {string} src + * The source string + * + * @return {string} + * The type that was found + */ +const findMimetype = (player, src) => { + if (!src) { + return ''; + } + + // 1. check for the type in the `source` cache + if (player.cache_.source.src === src && player.cache_.source.type) { + return player.cache_.source.type; + } + + // 2. see if we have this source in our `currentSources` cache + const matchingSources = player.cache_.sources.filter(s => s.src === src); + if (matchingSources.length) { + return matchingSources[0].type; + } + + // 3. look for the src url in source elements and use the type there + const sources = player.$$('source'); + for (let i = 0; i < sources.length; i++) { + const s = sources[i]; + if (s.type && s.src && s.src === src) { + return s.type; + } + } + + // 4. finally fallback to our list of mime types based on src url extension + return getMimetype(src); +}; + +/** + * @module filter-source + */ + +/** + * Filter out single bad source objects or multiple source objects in an + * array. Also flattens nested source object arrays into a 1 dimensional + * array of source objects. + * + * @param {Tech~SourceObject|Tech~SourceObject[]} src + * The src object to filter + * + * @return {Tech~SourceObject[]} + * An array of sourceobjects containing only valid sources + * + * @private + */ +const filterSource = function (src) { + // traverse array + if (Array.isArray(src)) { + let newsrc = []; + src.forEach(function (srcobj) { + srcobj = filterSource(srcobj); + if (Array.isArray(srcobj)) { + newsrc = newsrc.concat(srcobj); + } else if (isObject(srcobj)) { + newsrc.push(srcobj); + } + }); + src = newsrc; + } else if (typeof src === 'string' && src.trim()) { + // convert string into object + src = [fixSource({ + src + })]; + } else if (isObject(src) && typeof src.src === 'string' && src.src && src.src.trim()) { + // src is already valid + src = [fixSource(src)]; + } else { + // invalid source, turn it into an empty array + src = []; + } + return src; +}; + +/** + * Checks src mimetype, adding it when possible + * + * @param {Tech~SourceObject} src + * The src object to check + * @return {Tech~SourceObject} + * src Object with known type + */ +function fixSource(src) { + if (!src.type) { + const mimetype = getMimetype(src.src); + if (mimetype) { + src.type = mimetype; + } + } + return src; +} + +var icons = "<svg xmlns=\"http://www.w3.org/2000/svg\">\n <defs>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-play\">\n <path d=\"M16 10v28l22-14z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-pause\">\n <path d=\"M12 38h8V10h-8v28zm16-28v28h8V10h-8z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-audio\">\n <path d=\"M24 2C14.06 2 6 10.06 6 20v14c0 3.31 2.69 6 6 6h6V24h-8v-4c0-7.73 6.27-14 14-14s14 6.27 14 14v4h-8v16h6c3.31 0 6-2.69 6-6V20c0-9.94-8.06-18-18-18z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-captions\">\n <path d=\"M38 8H10c-2.21 0-4 1.79-4 4v24c0 2.21 1.79 4 4 4h28c2.21 0 4-1.79 4-4V12c0-2.21-1.79-4-4-4zM22 22h-3v-1h-4v6h4v-1h3v2a2 2 0 0 1-2 2h-6a2 2 0 0 1-2-2v-8a2 2 0 0 1 2-2h6a2 2 0 0 1 2 2v2zm14 0h-3v-1h-4v6h4v-1h3v2a2 2 0 0 1-2 2h-6a2 2 0 0 1-2-2v-8a2 2 0 0 1 2-2h6a2 2 0 0 1 2 2v2z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-subtitles\">\n <path d=\"M40 8H8c-2.21 0-4 1.79-4 4v24c0 2.21 1.79 4 4 4h32c2.21 0 4-1.79 4-4V12c0-2.21-1.79-4-4-4zM8 24h8v4H8v-4zm20 12H8v-4h20v4zm12 0h-8v-4h8v4zm0-8H20v-4h20v4z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-fullscreen-enter\">\n <path d=\"M14 28h-4v10h10v-4h-6v-6zm-4-8h4v-6h6v-4H10v10zm24 14h-6v4h10V28h-4v6zm-6-24v4h6v6h4V10H28z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-fullscreen-exit\">\n <path d=\"M10 32h6v6h4V28H10v4zm6-16h-6v4h10V10h-4v6zm12 22h4v-6h6v-4H28v10zm4-22v-6h-4v10h10v-4h-6z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-play-circle\">\n <path d=\"M20 33l12-9-12-9v18zm4-29C12.95 4 4 12.95 4 24s8.95 20 20 20 20-8.95 20-20S35.05 4 24 4zm0 36c-8.82 0-16-7.18-16-16S15.18 8 24 8s16 7.18 16 16-7.18 16-16 16z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-volume-mute\">\n <path d=\"M33 24c0-3.53-2.04-6.58-5-8.05v4.42l4.91 4.91c.06-.42.09-.85.09-1.28zm5 0c0 1.88-.41 3.65-1.08 5.28l3.03 3.03C41.25 29.82 42 27 42 24c0-8.56-5.99-15.72-14-17.54v4.13c5.78 1.72 10 7.07 10 13.41zM8.55 6L6 8.55 15.45 18H6v12h8l10 10V26.55l8.51 8.51c-1.34 1.03-2.85 1.86-4.51 2.36v4.13a17.94 17.94 0 0 0 7.37-3.62L39.45 42 42 39.45l-18-18L8.55 6zM24 8l-4.18 4.18L24 16.36V8z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-volume-low\">\n <path d=\"M14 18v12h8l10 10V8L22 18h-8z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-volume-medium\">\n <path d=\"M37 24c0-3.53-2.04-6.58-5-8.05v16.11c2.96-1.48 5-4.53 5-8.06zm-27-6v12h8l10 10V8L18 18h-8z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-volume-high\">\n <path d=\"M6 18v12h8l10 10V8L14 18H6zm27 6c0-3.53-2.04-6.58-5-8.05v16.11c2.96-1.48 5-4.53 5-8.06zM28 6.46v4.13c5.78 1.72 10 7.07 10 13.41s-4.22 11.69-10 13.41v4.13c8.01-1.82 14-8.97 14-17.54S36.01 8.28 28 6.46z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-spinner\">\n <path d=\"M18.8 21l9.53-16.51C26.94 4.18 25.49 4 24 4c-4.8 0-9.19 1.69-12.64 4.51l7.33 12.69.11-.2zm24.28-3c-1.84-5.85-6.3-10.52-11.99-12.68L23.77 18h19.31zm.52 2H28.62l.58 1 9.53 16.5C41.99 33.94 44 29.21 44 24c0-1.37-.14-2.71-.4-4zm-26.53 4l-7.8-13.5C6.01 14.06 4 18.79 4 24c0 1.37.14 2.71.4 4h14.98l-2.31-4zM4.92 30c1.84 5.85 6.3 10.52 11.99 12.68L24.23 30H4.92zm22.54 0l-7.8 13.51c1.4.31 2.85.49 4.34.49 4.8 0 9.19-1.69 12.64-4.51L29.31 26.8 27.46 30z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 24 24\" id=\"vjs-icon-hd\">\n <path d=\"M19 3H5a2 2 0 0 0-2 2v14a2 2 0 0 0 2 2h14c1.1 0 2-.9 2-2V5c0-1.1-.9-2-2-2zm-8 12H9.5v-2h-2v2H6V9h1.5v2.5h2V9H11v6zm2-6h4c.55 0 1 .45 1 1v4c0 .55-.45 1-1 1h-4V9zm1.5 4.5h2v-3h-2v3z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-chapters\">\n <path d=\"M6 26h4v-4H6v4zm0 8h4v-4H6v4zm0-16h4v-4H6v4zm8 8h28v-4H14v4zm0 8h28v-4H14v4zm0-20v4h28v-4H14z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 40 40\" id=\"vjs-icon-downloading\">\n <path d=\"M18.208 36.875q-3.208-.292-5.979-1.729-2.771-1.438-4.812-3.729-2.042-2.292-3.188-5.229-1.146-2.938-1.146-6.23 0-6.583 4.334-11.416 4.333-4.834 10.833-5.5v3.166q-5.167.75-8.583 4.646Q6.25 14.75 6.25 19.958q0 5.209 3.396 9.104 3.396 3.896 8.562 4.646zM20 28.417L11.542 20l2.083-2.083 4.917 4.916v-11.25h2.916v11.25l4.875-4.916L28.417 20zm1.792 8.458v-3.167q1.833-.25 3.541-.958 1.709-.708 3.167-1.875l2.333 2.292q-1.958 1.583-4.25 2.541-2.291.959-4.791 1.167zm6.791-27.792q-1.541-1.125-3.25-1.854-1.708-.729-3.541-1.021V3.042q2.5.25 4.77 1.208 2.271.958 4.271 2.5zm4.584 21.584l-2.25-2.25q1.166-1.5 1.854-3.209.687-1.708.937-3.541h3.209q-.292 2.5-1.229 4.791-.938 2.292-2.521 4.209zm.541-12.417q-.291-1.833-.958-3.562-.667-1.73-1.833-3.188l2.375-2.208q1.541 1.916 2.458 4.208.917 2.292 1.167 4.75z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-file-download\">\n <path d=\"M10.8 40.55q-1.35 0-2.375-1T7.4 37.15v-7.7h3.4v7.7h26.35v-7.7h3.4v7.7q0 1.4-1 2.4t-2.4 1zM24 32.1L13.9 22.05l2.45-2.45 5.95 5.95V7.15h3.4v18.4l5.95-5.95 2.45 2.45z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-file-download-done\">\n <path d=\"M9.8 40.5v-3.45h28.4v3.45zm9.2-9.05L7.4 19.85l2.45-2.35L19 26.65l19.2-19.2 2.4 2.4z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-file-download-off\">\n <path d=\"M4.9 4.75L43.25 43.1 41 45.3l-4.75-4.75q-.05.05-.075.025-.025-.025-.075-.025H10.8q-1.35 0-2.375-1T7.4 37.15v-7.7h3.4v7.7h22.05l-7-7-1.85 1.8L13.9 21.9l1.85-1.85L2.7 7zm26.75 14.7l2.45 2.45-3.75 3.8-2.45-2.5zM25.7 7.15V21.1l-3.4-3.45V7.15z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-share\">\n <path d=\"M36 32.17c-1.52 0-2.89.59-3.93 1.54L17.82 25.4c.11-.45.18-.92.18-1.4s-.07-.95-.18-1.4l14.1-8.23c1.07 1 2.5 1.62 4.08 1.62 3.31 0 6-2.69 6-6s-2.69-6-6-6-6 2.69-6 6c0 .48.07.95.18 1.4l-14.1 8.23c-1.07-1-2.5-1.62-4.08-1.62-3.31 0-6 2.69-6 6s2.69 6 6 6c1.58 0 3.01-.62 4.08-1.62l14.25 8.31c-.1.42-.16.86-.16 1.31A5.83 5.83 0 1 0 36 32.17z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-cog\">\n <path d=\"M38.86 25.95c.08-.64.14-1.29.14-1.95s-.06-1.31-.14-1.95l4.23-3.31c.38-.3.49-.84.24-1.28l-4-6.93c-.25-.43-.77-.61-1.22-.43l-4.98 2.01c-1.03-.79-2.16-1.46-3.38-1.97L29 4.84c-.09-.47-.5-.84-1-.84h-8c-.5 0-.91.37-.99.84l-.75 5.3a14.8 14.8 0 0 0-3.38 1.97L9.9 10.1a1 1 0 0 0-1.22.43l-4 6.93c-.25.43-.14.97.24 1.28l4.22 3.31C9.06 22.69 9 23.34 9 24s.06 1.31.14 1.95l-4.22 3.31c-.38.3-.49.84-.24 1.28l4 6.93c.25.43.77.61 1.22.43l4.98-2.01c1.03.79 2.16 1.46 3.38 1.97l.75 5.3c.08.47.49.84.99.84h8c.5 0 .91-.37.99-.84l.75-5.3a14.8 14.8 0 0 0 3.38-1.97l4.98 2.01a1 1 0 0 0 1.22-.43l4-6.93c.25-.43.14-.97-.24-1.28l-4.22-3.31zM24 31c-3.87 0-7-3.13-7-7s3.13-7 7-7 7 3.13 7 7-3.13 7-7 7z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-square\">\n <path d=\"M36 8H12c-2.21 0-4 1.79-4 4v24c0 2.21 1.79 4 4 4h24c2.21 0 4-1.79 4-4V12c0-2.21-1.79-4-4-4zm0 28H12V12h24v24z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-circle\">\n <circle cx=\"24\" cy=\"24\" r=\"20\"></circle>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-circle-outline\">\n <path d=\"M24 4C12.95 4 4 12.95 4 24s8.95 20 20 20 20-8.95 20-20S35.05 4 24 4zm0 36c-8.82 0-16-7.18-16-16S15.18 8 24 8s16 7.18 16 16-7.18 16-16 16z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-circle-inner-circle\">\n <path d=\"M24 4C12.97 4 4 12.97 4 24s8.97 20 20 20 20-8.97 20-20S35.03 4 24 4zm0 36c-8.82 0-16-7.18-16-16S15.18 8 24 8s16 7.18 16 16-7.18 16-16 16zm6-16c0 3.31-2.69 6-6 6s-6-2.69-6-6 2.69-6 6-6 6 2.69 6 6z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-cancel\">\n <path d=\"M24 4C12.95 4 4 12.95 4 24s8.95 20 20 20 20-8.95 20-20S35.05 4 24 4zm10 27.17L31.17 34 24 26.83 16.83 34 14 31.17 21.17 24 14 16.83 16.83 14 24 21.17 31.17 14 34 16.83 26.83 24 34 31.17z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-replay\">\n <path d=\"M24 10V2L14 12l10 10v-8c6.63 0 12 5.37 12 12s-5.37 12-12 12-12-5.37-12-12H8c0 8.84 7.16 16 16 16s16-7.16 16-16-7.16-16-16-16z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-repeat\">\n <path d=\"M14 14h20v6l8-8-8-8v6H10v12h4v-8zm20 20H14v-6l-8 8 8 8v-6h24V26h-4v8z\"></path>\n </symbol>\n <symbol viewBox=\"0 96 48 48\" id=\"vjs-icon-replay-5\">\n <path d=\"M17.689 98l-8.697 8.696 8.697 8.697 2.486-2.485-4.32-4.319h1.302c4.93 0 9.071 1.722 12.424 5.165 3.352 3.443 5.029 7.638 5.029 12.584h3.55c0-2.958-.553-5.73-1.658-8.313-1.104-2.583-2.622-4.841-4.555-6.774-1.932-1.932-4.19-3.45-6.773-4.555-2.584-1.104-5.355-1.657-8.313-1.657H15.5l4.615-4.615zm-8.08 21.659v13.861h11.357v5.008H9.609V143h12.7c.834 0 1.55-.298 2.146-.894.596-.597.895-1.31.895-2.145v-7.781c0-.835-.299-1.55-.895-2.147a2.929 2.929 0 0 0-2.147-.894h-8.227v-5.096H25.35v-4.384z\"></path>\n </symbol>\n <symbol viewBox=\"0 96 48 48\" id=\"vjs-icon-replay-10\">\n <path d=\"M42.315 125.63c0-4.997-1.694-9.235-5.08-12.713-3.388-3.479-7.571-5.218-12.552-5.218h-1.315l4.363 4.363-2.51 2.51-8.787-8.786L25.221 97l2.45 2.45-4.662 4.663h1.375c2.988 0 5.788.557 8.397 1.673 2.61 1.116 4.892 2.65 6.844 4.602 1.953 1.953 3.487 4.234 4.602 6.844 1.116 2.61 1.674 5.41 1.674 8.398zM8.183 142v-19.657H3.176V117.8h9.643V142zm13.63 0c-1.156 0-2.127-.393-2.912-1.178-.778-.778-1.168-1.746-1.168-2.902v-16.04c0-1.156.393-2.127 1.178-2.912.779-.779 1.746-1.168 2.902-1.168h7.696c1.156 0 2.126.392 2.911 1.177.779.78 1.168 1.747 1.168 2.903v16.04c0 1.156-.392 2.127-1.177 2.912-.779.779-1.746 1.168-2.902 1.168zm.556-4.636h6.583v-15.02H22.37z\"></path>\n </symbol>\n <symbol viewBox=\"0 96 48 48\" id=\"vjs-icon-replay-30\">\n <path d=\"M26.047 97l-8.733 8.732 8.733 8.733 2.496-2.494-4.336-4.338h1.307c4.95 0 9.108 1.73 12.474 5.187 3.367 3.458 5.051 7.668 5.051 12.635h3.565c0-2.97-.556-5.751-1.665-8.346-1.109-2.594-2.633-4.862-4.574-6.802-1.94-1.941-4.208-3.466-6.803-4.575-2.594-1.109-5.375-1.664-8.345-1.664H23.85l4.634-4.634zM2.555 117.531v4.688h10.297v5.25H5.873v4.687h6.979v5.156H2.555V142H13.36c1.061 0 1.95-.395 2.668-1.186.718-.79 1.076-1.772 1.076-2.94v-16.218c0-1.168-.358-2.149-1.076-2.94-.717-.79-1.607-1.185-2.668-1.185zm22.482.14c-1.149 0-2.11.39-2.885 1.165-.78.78-1.172 1.744-1.172 2.893v15.943c0 1.149.388 2.11 1.163 2.885.78.78 1.745 1.172 2.894 1.172h7.649c1.148 0 2.11-.388 2.884-1.163.78-.78 1.17-1.745 1.17-2.894v-15.943c0-1.15-.386-2.111-1.16-2.885-.78-.78-1.746-1.172-2.894-1.172zm.553 4.518h6.545v14.93H25.59z\"></path>\n </symbol>\n <symbol viewBox=\"0 96 48 48\" id=\"vjs-icon-forward-5\">\n <path d=\"M29.508 97l-2.431 2.43 4.625 4.625h-1.364c-2.965 0-5.742.554-8.332 1.66-2.589 1.107-4.851 2.629-6.788 4.566-1.937 1.937-3.458 4.2-4.565 6.788-1.107 2.59-1.66 5.367-1.66 8.331h3.557c0-4.957 1.68-9.16 5.04-12.611 3.36-3.45 7.51-5.177 12.451-5.177h1.304l-4.326 4.33 2.49 2.49 8.715-8.716zm-9.783 21.61v13.89h11.382v5.018H19.725V142h12.727a2.93 2.93 0 0 0 2.15-.896 2.93 2.93 0 0 0 .896-2.15v-7.798c0-.837-.299-1.554-.896-2.152a2.93 2.93 0 0 0-2.15-.896h-8.245V123h11.29v-4.392z\"></path>\n </symbol>\n <symbol viewBox=\"0 96 48 48\" id=\"vjs-icon-forward-10\">\n <path d=\"M23.119 97l-2.386 2.383 4.538 4.538h-1.339c-2.908 0-5.633.543-8.173 1.63-2.54 1.085-4.76 2.577-6.66 4.478-1.9 1.9-3.392 4.12-4.478 6.66-1.085 2.54-1.629 5.264-1.629 8.172h3.49c0-4.863 1.648-8.986 4.944-12.372 3.297-3.385 7.368-5.078 12.216-5.078h1.279l-4.245 4.247 2.443 2.442 8.55-8.55zm-9.52 21.45v4.42h4.871V142h4.513v-23.55zm18.136 0c-1.125 0-2.066.377-2.824 1.135-.764.764-1.148 1.709-1.148 2.834v15.612c0 1.124.38 2.066 1.139 2.824.764.764 1.708 1.145 2.833 1.145h7.489c1.125 0 2.066-.378 2.824-1.136.764-.764 1.145-1.709 1.145-2.833v-15.612c0-1.125-.378-2.067-1.136-2.825-.764-.764-1.708-1.145-2.833-1.145zm.54 4.42h6.408v14.617h-6.407z\"></path>\n </symbol>\n <symbol viewBox=\"0 96 48 48\" id=\"vjs-icon-forward-30\">\n <path d=\"M25.549 97l-2.437 2.434 4.634 4.635H26.38c-2.97 0-5.753.555-8.347 1.664-2.594 1.109-4.861 2.633-6.802 4.574-1.94 1.94-3.465 4.207-4.574 6.802-1.109 2.594-1.664 5.377-1.664 8.347h3.565c0-4.967 1.683-9.178 5.05-12.636 3.366-3.458 7.525-5.187 12.475-5.187h1.307l-4.335 4.338 2.495 2.494 8.732-8.732zm-11.553 20.53v4.689h10.297v5.249h-6.978v4.688h6.978v5.156H13.996V142h10.808c1.06 0 1.948-.395 2.666-1.186.718-.79 1.077-1.771 1.077-2.94v-16.217c0-1.169-.36-2.15-1.077-2.94-.718-.79-1.605-1.186-2.666-1.186zm21.174.168c-1.149 0-2.11.389-2.884 1.163-.78.78-1.172 1.745-1.172 2.894v15.942c0 1.15.388 2.11 1.162 2.885.78.78 1.745 1.17 2.894 1.17h7.649c1.149 0 2.11-.386 2.885-1.16.78-.78 1.17-1.746 1.17-2.895v-15.942c0-1.15-.387-2.11-1.161-2.885-.78-.78-1.745-1.172-2.894-1.172zm.552 4.516h6.542v14.931h-6.542z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 512 512\" id=\"vjs-icon-audio-description\">\n <g fill-rule=\"evenodd\"><path d=\"M227.29 381.351V162.993c50.38-1.017 89.108-3.028 117.631 17.126 27.374 19.342 48.734 56.965 44.89 105.325-4.067 51.155-41.335 94.139-89.776 98.475-24.085 2.155-71.972 0-71.972 0s-.84-1.352-.773-2.568m48.755-54.804c31.43 1.26 53.208-16.633 56.495-45.386 4.403-38.51-21.188-63.552-58.041-60.796v103.612c-.036 1.466.575 2.22 1.546 2.57\"></path><path d=\"M383.78 381.328c13.336 3.71 17.387-11.06 23.215-21.408 12.722-22.571 22.294-51.594 22.445-84.774.221-47.594-18.343-82.517-35.6-106.182h-8.51c-.587 3.874 2.226 7.315 3.865 10.276 13.166 23.762 25.367 56.553 25.54 94.194.2 43.176-14.162 79.278-30.955 107.894\"></path><path d=\"M425.154 381.328c13.336 3.71 17.384-11.061 23.215-21.408 12.721-22.571 22.291-51.594 22.445-84.774.221-47.594-18.343-82.517-35.6-106.182h-8.511c-.586 3.874 2.226 7.315 3.866 10.276 13.166 23.762 25.367 56.553 25.54 94.194.2 43.176-14.162 79.278-30.955 107.894\"></path><path d=\"M466.26 381.328c13.337 3.71 17.385-11.061 23.216-21.408 12.722-22.571 22.292-51.594 22.445-84.774.221-47.594-18.343-82.517-35.6-106.182h-8.51c-.587 3.874 2.225 7.315 3.865 10.276 13.166 23.762 25.367 56.553 25.54 94.194.2 43.176-14.162 79.278-30.955 107.894M4.477 383.005H72.58l18.573-28.484 64.169-.135s.065 19.413.065 28.62h48.756V160.307h-58.816c-5.653 9.537-140.85 222.697-140.85 222.697zm152.667-145.282v71.158l-40.453-.27 40.453-70.888z\"></path></g>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-next-item\">\n <path d=\"M12 36l17-12-17-12v24zm20-24v24h4V12h-4z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-previous-item\">\n <path d=\"M12 12h4v24h-4zm7 12l17 12V12z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-shuffle\">\n <path d=\"M21.17 18.34L10.83 8 8 10.83l10.34 10.34 2.83-2.83zM29 8l4.09 4.09L8 37.17 10.83 40l25.09-25.09L40 19V8H29zm.66 18.83l-2.83 2.83 6.26 6.26L29 40h11V29l-4.09 4.09-6.25-6.26z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-cast\">\n <path d=\"M42 6H6c-2.21 0-4 1.79-4 4v6h4v-6h36v28H28v4h14c2.21 0 4-1.79 4-4V10c0-2.21-1.79-4-4-4zM2 36v6h6c0-3.31-2.69-6-6-6zm0-8v4c5.52 0 10 4.48 10 10h4c0-7.73-6.27-14-14-14zm0-8v4c9.94 0 18 8.06 18 18h4c0-12.15-9.85-22-22-22z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 48 48\" id=\"vjs-icon-picture-in-picture-enter\">\n <path d=\"M38 22H22v11.99h16V22zm8 16V9.96C46 7.76 44.2 6 42 6H6C3.8 6 2 7.76 2 9.96V38c0 2.2 1.8 4 4 4h36c2.2 0 4-1.8 4-4zm-4 .04H6V9.94h36v28.1z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 22 18\" id=\"vjs-icon-picture-in-picture-exit\">\n <path d=\"M18 4H4v10h14V4zm4 12V1.98C22 .88 21.1 0 20 0H2C.9 0 0 .88 0 1.98V16c0 1.1.9 2 2 2h18c1.1 0 2-.9 2-2zm-2 .02H2V1.97h18v14.05z\"></path>\n <path fill=\"none\" d=\"M-1-3h24v24H-1z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 1792 1792\" id=\"vjs-icon-facebook\">\n <path d=\"M1343 12v264h-157q-86 0-116 36t-30 108v189h293l-39 296h-254v759H734V905H479V609h255V391q0-186 104-288.5T1115 0q147 0 228 12z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 1792 1792\" id=\"vjs-icon-linkedin\">\n <path d=\"M477 625v991H147V625h330zm21-306q1 73-50.5 122T312 490h-2q-82 0-132-49t-50-122q0-74 51.5-122.5T314 148t133 48.5T498 319zm1166 729v568h-329v-530q0-105-40.5-164.5T1168 862q-63 0-105.5 34.5T999 982q-11 30-11 81v553H659q2-399 2-647t-1-296l-1-48h329v144h-2q20-32 41-56t56.5-52 87-43.5T1285 602q171 0 275 113.5t104 332.5z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 1792 1792\" id=\"vjs-icon-twitter\">\n <path d=\"M1684 408q-67 98-162 167 1 14 1 42 0 130-38 259.5T1369.5 1125 1185 1335.5t-258 146-323 54.5q-271 0-496-145 35 4 78 4 225 0 401-138-105-2-188-64.5T285 1033q33 5 61 5 43 0 85-11-112-23-185.5-111.5T172 710v-4q68 38 146 41-66-44-105-115t-39-154q0-88 44-163 121 149 294.5 238.5T884 653q-8-38-8-74 0-134 94.5-228.5T1199 256q140 0 236 102 109-21 205-78-37 115-142 178 93-10 186-50z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 1792 1792\" id=\"vjs-icon-tumblr\">\n <path d=\"M1328 1329l80 237q-23 35-111 66t-177 32q-104 2-190.5-26T787 1564t-95-106-55.5-120-16.5-118V676H452V461q72-26 129-69.5t91-90 58-102 34-99T779 12q1-5 4.5-8.5T791 0h244v424h333v252h-334v518q0 30 6.5 56t22.5 52.5 49.5 41.5 81.5 14q78-2 134-29z\"></path>\n </symbol>\n <symbol viewBox=\"0 0 1792 1792\" id=\"vjs-icon-pinterest\">\n <path d=\"M1664 896q0 209-103 385.5T1281.5 1561 896 1664q-111 0-218-32 59-93 78-164 9-34 54-211 20 39 73 67.5t114 28.5q121 0 216-68.5t147-188.5 52-270q0-114-59.5-214T1180 449t-255-63q-105 0-196 29t-154.5 77-109 110.5-67 129.5T377 866q0 104 40 183t117 111q30 12 38-20 2-7 8-31t8-30q6-23-11-43-51-61-51-151 0-151 104.5-259.5T904 517q151 0 235.5 82t84.5 213q0 170-68.5 289T980 1220q-61 0-98-43.5T859 1072q8-35 26.5-93.5t30-103T927 800q0-50-27-83t-77-33q-62 0-105 57t-43 142q0 73 25 122l-99 418q-17 70-13 177-206-91-333-281T128 896q0-209 103-385.5T510.5 231 896 128t385.5 103T1561 510.5 1664 896z\"></path>\n </symbol>\n </defs>\n</svg>"; + +/** + * @file loader.js + */ + +/** + * The `MediaLoader` is the `Component` that decides which playback technology to load + * when a player is initialized. + * + * @extends Component + */ +class MediaLoader extends Component$1 { + /** + * Create an instance of this class. + * + * @param { import('../player').default } player + * The `Player` that this class should attach to. + * + * @param {Object} [options] + * The key/value store of player options. + * + * @param {Function} [ready] + * The function that is run when this component is ready. + */ + constructor(player, options, ready) { + // MediaLoader has no element + const options_ = merge$1({ + createEl: false + }, options); + super(player, options_, ready); + + // If there are no sources when the player is initialized, + // load the first supported playback technology. + + if (!options.playerOptions.sources || options.playerOptions.sources.length === 0) { + for (let i = 0, j = options.playerOptions.techOrder; i < j.length; i++) { + const techName = toTitleCase$1(j[i]); + let tech = Tech.getTech(techName); + + // Support old behavior of techs being registered as components. + // Remove once that deprecated behavior is removed. + if (!techName) { + tech = Component$1.getComponent(techName); + } + + // Check if the browser supports this technology + if (tech && tech.isSupported()) { + player.loadTech_(techName); + break; + } + } + } else { + // Loop through playback technologies (e.g. HTML5) and check for support. + // Then load the best source. + // A few assumptions here: + // All playback technologies respect preload false. + player.src(options.playerOptions.sources); + } + } +} +Component$1.registerComponent('MediaLoader', MediaLoader); + +/** + * @file clickable-component.js + */ + +/** + * Component which is clickable or keyboard actionable, but is not a + * native HTML button. + * + * @extends Component + */ +class ClickableComponent extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of component options. + * + * @param {function} [options.clickHandler] + * The function to call when the button is clicked / activated + * + * @param {string} [options.controlText] + * The text to set on the button + * + * @param {string} [options.className] + * A class or space separated list of classes to add the component + * + */ + constructor(player, options) { + super(player, options); + if (this.options_.controlText) { + this.controlText(this.options_.controlText); + } + this.handleMouseOver_ = e => this.handleMouseOver(e); + this.handleMouseOut_ = e => this.handleMouseOut(e); + this.handleClick_ = e => this.handleClick(e); + this.handleKeyDown_ = e => this.handleKeyDown(e); + this.emitTapEvents(); + this.enable(); + } + + /** + * Create the `ClickableComponent`s DOM element. + * + * @param {string} [tag=div] + * The element's node type. + * + * @param {Object} [props={}] + * An object of properties that should be set on the element. + * + * @param {Object} [attributes={}] + * An object of attributes that should be set on the element. + * + * @return {Element} + * The element that gets created. + */ + createEl(tag = 'div', props = {}, attributes = {}) { + props = Object.assign({ + className: this.buildCSSClass(), + tabIndex: 0 + }, props); + if (tag === 'button') { + log$1.error(`Creating a ClickableComponent with an HTML element of ${tag} is not supported; use a Button instead.`); + } + + // Add ARIA attributes for clickable element which is not a native HTML button + attributes = Object.assign({ + role: 'button' + }, attributes); + this.tabIndex_ = props.tabIndex; + const el = createEl(tag, props, attributes); + if (!this.player_.options_.experimentalSvgIcons) { + el.appendChild(createEl('span', { + className: 'vjs-icon-placeholder' + }, { + 'aria-hidden': true + })); + } + this.createControlTextEl(el); + return el; + } + dispose() { + // remove controlTextEl_ on dispose + this.controlTextEl_ = null; + super.dispose(); + } + + /** + * Create a control text element on this `ClickableComponent` + * + * @param {Element} [el] + * Parent element for the control text. + * + * @return {Element} + * The control text element that gets created. + */ + createControlTextEl(el) { + this.controlTextEl_ = createEl('span', { + className: 'vjs-control-text' + }, { + // let the screen reader user know that the text of the element may change + 'aria-live': 'polite' + }); + if (el) { + el.appendChild(this.controlTextEl_); + } + this.controlText(this.controlText_, el); + return this.controlTextEl_; + } + + /** + * Get or set the localize text to use for the controls on the `ClickableComponent`. + * + * @param {string} [text] + * Control text for element. + * + * @param {Element} [el=this.el()] + * Element to set the title on. + * + * @return {string} + * - The control text when getting + */ + controlText(text, el = this.el()) { + if (text === undefined) { + return this.controlText_ || 'Need Text'; + } + const localizedText = this.localize(text); + + /** @protected */ + this.controlText_ = text; + textContent(this.controlTextEl_, localizedText); + if (!this.nonIconControl && !this.player_.options_.noUITitleAttributes) { + // Set title attribute if only an icon is shown + el.setAttribute('title', localizedText); + } + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-control vjs-button ${super.buildCSSClass()}`; + } + + /** + * Enable this `ClickableComponent` + */ + enable() { + if (!this.enabled_) { + this.enabled_ = true; + this.removeClass('vjs-disabled'); + this.el_.setAttribute('aria-disabled', 'false'); + if (typeof this.tabIndex_ !== 'undefined') { + this.el_.setAttribute('tabIndex', this.tabIndex_); + } + this.on(['tap', 'click'], this.handleClick_); + this.on('keydown', this.handleKeyDown_); + } + } + + /** + * Disable this `ClickableComponent` + */ + disable() { + this.enabled_ = false; + this.addClass('vjs-disabled'); + this.el_.setAttribute('aria-disabled', 'true'); + if (typeof this.tabIndex_ !== 'undefined') { + this.el_.removeAttribute('tabIndex'); + } + this.off('mouseover', this.handleMouseOver_); + this.off('mouseout', this.handleMouseOut_); + this.off(['tap', 'click'], this.handleClick_); + this.off('keydown', this.handleKeyDown_); + } + + /** + * Handles language change in ClickableComponent for the player in components + * + * + */ + handleLanguagechange() { + this.controlText(this.controlText_); + } + + /** + * Event handler that is called when a `ClickableComponent` receives a + * `click` or `tap` event. + * + * @param {Event} event + * The `tap` or `click` event that caused this function to be called. + * + * @listens tap + * @listens click + * @abstract + */ + handleClick(event) { + if (this.options_.clickHandler) { + this.options_.clickHandler.call(this, arguments); + } + } + + /** + * Event handler that is called when a `ClickableComponent` receives a + * `keydown` event. + * + * By default, if the key is Space or Enter, it will trigger a `click` event. + * + * @param {KeyboardEvent} event + * The `keydown` event that caused this function to be called. + * + * @listens keydown + */ + handleKeyDown(event) { + // Support Space or Enter key operation to fire a click event. Also, + // prevent the event from propagating through the DOM and triggering + // Player hotkeys. + if (keycode.isEventKey(event, 'Space') || keycode.isEventKey(event, 'Enter')) { + event.preventDefault(); + event.stopPropagation(); + this.trigger('click'); + } else { + // Pass keypress handling up for unsupported keys + super.handleKeyDown(event); + } + } +} +Component$1.registerComponent('ClickableComponent', ClickableComponent); + +/** + * @file poster-image.js + */ + +/** + * A `ClickableComponent` that handles showing the poster image for the player. + * + * @extends ClickableComponent + */ +class PosterImage extends ClickableComponent { + /** + * Create an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should attach to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.update(); + this.update_ = e => this.update(e); + player.on('posterchange', this.update_); + } + + /** + * Clean up and dispose of the `PosterImage`. + */ + dispose() { + this.player().off('posterchange', this.update_); + super.dispose(); + } + + /** + * Create the `PosterImage`s DOM element. + * + * @return {Element} + * The element that gets created. + */ + createEl() { + // The el is an empty div to keep position in the DOM + // A picture and img el will be inserted when a source is set + return createEl('div', { + className: 'vjs-poster' + }); + } + + /** + * Get or set the `PosterImage`'s crossOrigin option. + * + * @param {string|null} [value] + * The value to set the crossOrigin to. If an argument is + * given, must be one of `'anonymous'` or `'use-credentials'`, or 'null'. + * + * @return {string|null} + * - The current crossOrigin value of the `Player` when getting. + * - undefined when setting + */ + crossOrigin(value) { + // `null` can be set to unset a value + if (typeof value === 'undefined') { + if (this.$('img')) { + // If the poster's element exists, give its value + return this.$('img').crossOrigin; + } else if (this.player_.tech_ && this.player_.tech_.isReady_) { + // If not but the tech is ready, query the tech + return this.player_.crossOrigin(); + } + // Otherwise check options as the poster is usually set before the state of crossorigin + // can be retrieved by the getter + return this.player_.options_.crossOrigin || this.player_.options_.crossorigin || null; + } + if (value !== null && value !== 'anonymous' && value !== 'use-credentials') { + this.player_.log.warn(`crossOrigin must be null, "anonymous" or "use-credentials", given "${value}"`); + return; + } + if (this.$('img')) { + this.$('img').crossOrigin = value; + } + return; + } + + /** + * An {@link EventTarget~EventListener} for {@link Player#posterchange} events. + * + * @listens Player#posterchange + * + * @param {Event} [event] + * The `Player#posterchange` event that triggered this function. + */ + update(event) { + const url = this.player().poster(); + this.setSrc(url); + + // If there's no poster source we should display:none on this component + // so it's not still clickable or right-clickable + if (url) { + this.show(); + } else { + this.hide(); + } + } + + /** + * Set the source of the `PosterImage` depending on the display method. (Re)creates + * the inner picture and img elementss when needed. + * + * @param {string} [url] + * The URL to the source for the `PosterImage`. If not specified or falsy, + * any source and ant inner picture/img are removed. + */ + setSrc(url) { + if (!url) { + this.el_.textContent = ''; + return; + } + if (!this.$('img')) { + this.el_.appendChild(createEl('picture', { + className: 'vjs-poster', + // Don't want poster to be tabbable. + tabIndex: -1 + }, {}, createEl('img', { + loading: 'lazy', + crossOrigin: this.crossOrigin() + }, { + alt: '' + }))); + } + this.$('img').src = url; + } + + /** + * An {@link EventTarget~EventListener} for clicks on the `PosterImage`. See + * {@link ClickableComponent#handleClick} for instances where this will be triggered. + * + * @listens tap + * @listens click + * @listens keydown + * + * @param {Event} event + + The `click`, `tap` or `keydown` event that caused this function to be called. + */ + handleClick(event) { + // We don't want a click to trigger playback when controls are disabled + if (!this.player_.controls()) { + return; + } + if (this.player_.tech(true)) { + this.player_.tech(true).focus(); + } + if (this.player_.paused()) { + silencePromise(this.player_.play()); + } else { + this.player_.pause(); + } + } +} + +/** + * Get or set the `PosterImage`'s crossorigin option. For the HTML5 player, this + * sets the `crossOrigin` property on the `<img>` tag to control the CORS + * behavior. + * + * @param {string|null} [value] + * The value to set the `PosterImages`'s crossorigin to. If an argument is + * given, must be one of `anonymous` or `use-credentials`. + * + * @return {string|null|undefined} + * - The current crossorigin value of the `Player` when getting. + * - undefined when setting + */ +PosterImage.prototype.crossorigin = PosterImage.prototype.crossOrigin; +Component$1.registerComponent('PosterImage', PosterImage); + +/** + * @file text-track-display.js + */ +const darkGray = '#222'; +const lightGray = '#ccc'; +const fontMap = { + monospace: 'monospace', + sansSerif: 'sans-serif', + serif: 'serif', + monospaceSansSerif: '"Andale Mono", "Lucida Console", monospace', + monospaceSerif: '"Courier New", monospace', + proportionalSansSerif: 'sans-serif', + proportionalSerif: 'serif', + casual: '"Comic Sans MS", Impact, fantasy', + script: '"Monotype Corsiva", cursive', + smallcaps: '"Andale Mono", "Lucida Console", monospace, sans-serif' +}; + +/** + * Construct an rgba color from a given hex color code. + * + * @param {number} color + * Hex number for color, like #f0e or #f604e2. + * + * @param {number} opacity + * Value for opacity, 0.0 - 1.0. + * + * @return {string} + * The rgba color that was created, like 'rgba(255, 0, 0, 0.3)'. + */ +function constructColor(color, opacity) { + let hex; + if (color.length === 4) { + // color looks like "#f0e" + hex = color[1] + color[1] + color[2] + color[2] + color[3] + color[3]; + } else if (color.length === 7) { + // color looks like "#f604e2" + hex = color.slice(1); + } else { + throw new Error('Invalid color code provided, ' + color + '; must be formatted as e.g. #f0e or #f604e2.'); + } + return 'rgba(' + parseInt(hex.slice(0, 2), 16) + ',' + parseInt(hex.slice(2, 4), 16) + ',' + parseInt(hex.slice(4, 6), 16) + ',' + opacity + ')'; +} + +/** + * Try to update the style of a DOM element. Some style changes will throw an error, + * particularly in IE8. Those should be noops. + * + * @param {Element} el + * The DOM element to be styled. + * + * @param {string} style + * The CSS property on the element that should be styled. + * + * @param {string} rule + * The style rule that should be applied to the property. + * + * @private + */ +function tryUpdateStyle(el, style, rule) { + try { + el.style[style] = rule; + } catch (e) { + // Satisfies linter. + return; + } +} + +/** + * Converts the CSS top/right/bottom/left property numeric value to string in pixels. + * + * @param {number} position + * The CSS top/right/bottom/left property value. + * + * @return {string} + * The CSS property value that was created, like '10px'. + * + * @private + */ +function getCSSPositionValue(position) { + return position ? `${position}px` : ''; +} + +/** + * The component for displaying text track cues. + * + * @extends Component + */ +class TextTrackDisplay extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + * + * @param {Function} [ready] + * The function to call when `TextTrackDisplay` is ready. + */ + constructor(player, options, ready) { + super(player, options, ready); + const updateDisplayTextHandler = e => this.updateDisplay(e); + const updateDisplayHandler = e => { + this.updateDisplayOverlay(); + this.updateDisplay(e); + }; + player.on('loadstart', e => this.toggleDisplay(e)); + player.on('texttrackchange', updateDisplayTextHandler); + player.on('loadedmetadata', e => { + this.updateDisplayOverlay(); + this.preselectTrack(e); + }); + + // This used to be called during player init, but was causing an error + // if a track should show by default and the display hadn't loaded yet. + // Should probably be moved to an external track loader when we support + // tracks that don't need a display. + player.ready(bind_(this, function () { + if (player.tech_ && player.tech_.featuresNativeTextTracks) { + this.hide(); + return; + } + player.on('fullscreenchange', updateDisplayHandler); + player.on('playerresize', updateDisplayHandler); + const screenOrientation = window$1.screen.orientation || window$1; + const changeOrientationEvent = window$1.screen.orientation ? 'change' : 'orientationchange'; + screenOrientation.addEventListener(changeOrientationEvent, updateDisplayHandler); + player.on('dispose', () => screenOrientation.removeEventListener(changeOrientationEvent, updateDisplayHandler)); + const tracks = this.options_.playerOptions.tracks || []; + for (let i = 0; i < tracks.length; i++) { + this.player_.addRemoteTextTrack(tracks[i], true); + } + this.preselectTrack(); + })); + } + + /** + * Preselect a track following this precedence: + * - matches the previously selected {@link TextTrack}'s language and kind + * - matches the previously selected {@link TextTrack}'s language only + * - is the first default captions track + * - is the first default descriptions track + * + * @listens Player#loadstart + */ + preselectTrack() { + const modes = { + captions: 1, + subtitles: 1 + }; + const trackList = this.player_.textTracks(); + const userPref = this.player_.cache_.selectedLanguage; + let firstDesc; + let firstCaptions; + let preferredTrack; + for (let i = 0; i < trackList.length; i++) { + const track = trackList[i]; + if (userPref && userPref.enabled && userPref.language && userPref.language === track.language && track.kind in modes) { + // Always choose the track that matches both language and kind + if (track.kind === userPref.kind) { + preferredTrack = track; + // or choose the first track that matches language + } else if (!preferredTrack) { + preferredTrack = track; + } + + // clear everything if offTextTrackMenuItem was clicked + } else if (userPref && !userPref.enabled) { + preferredTrack = null; + firstDesc = null; + firstCaptions = null; + } else if (track.default) { + if (track.kind === 'descriptions' && !firstDesc) { + firstDesc = track; + } else if (track.kind in modes && !firstCaptions) { + firstCaptions = track; + } + } + } + + // The preferredTrack matches the user preference and takes + // precedence over all the other tracks. + // So, display the preferredTrack before the first default track + // and the subtitles/captions track before the descriptions track + if (preferredTrack) { + preferredTrack.mode = 'showing'; + } else if (firstCaptions) { + firstCaptions.mode = 'showing'; + } else if (firstDesc) { + firstDesc.mode = 'showing'; + } + } + + /** + * Turn display of {@link TextTrack}'s from the current state into the other state. + * There are only two states: + * - 'shown' + * - 'hidden' + * + * @listens Player#loadstart + */ + toggleDisplay() { + if (this.player_.tech_ && this.player_.tech_.featuresNativeTextTracks) { + this.hide(); + } else { + this.show(); + } + } + + /** + * Create the {@link Component}'s DOM element. + * + * @return {Element} + * The element that was created. + */ + createEl() { + return super.createEl('div', { + className: 'vjs-text-track-display' + }, { + 'translate': 'yes', + 'aria-live': 'off', + 'aria-atomic': 'true' + }); + } + + /** + * Clear all displayed {@link TextTrack}s. + */ + clearDisplay() { + if (typeof window$1.WebVTT === 'function') { + window$1.WebVTT.processCues(window$1, [], this.el_); + } + } + + /** + * Update the displayed TextTrack when a either a {@link Player#texttrackchange} or + * a {@link Player#fullscreenchange} is fired. + * + * @listens Player#texttrackchange + * @listens Player#fullscreenchange + */ + updateDisplay() { + const tracks = this.player_.textTracks(); + const allowMultipleShowingTracks = this.options_.allowMultipleShowingTracks; + this.clearDisplay(); + if (allowMultipleShowingTracks) { + const showingTracks = []; + for (let i = 0; i < tracks.length; ++i) { + const track = tracks[i]; + if (track.mode !== 'showing') { + continue; + } + showingTracks.push(track); + } + this.updateForTrack(showingTracks); + return; + } + + // Track display prioritization model: if multiple tracks are 'showing', + // display the first 'subtitles' or 'captions' track which is 'showing', + // otherwise display the first 'descriptions' track which is 'showing' + + let descriptionsTrack = null; + let captionsSubtitlesTrack = null; + let i = tracks.length; + while (i--) { + const track = tracks[i]; + if (track.mode === 'showing') { + if (track.kind === 'descriptions') { + descriptionsTrack = track; + } else { + captionsSubtitlesTrack = track; + } + } + } + if (captionsSubtitlesTrack) { + if (this.getAttribute('aria-live') !== 'off') { + this.setAttribute('aria-live', 'off'); + } + this.updateForTrack(captionsSubtitlesTrack); + } else if (descriptionsTrack) { + if (this.getAttribute('aria-live') !== 'assertive') { + this.setAttribute('aria-live', 'assertive'); + } + this.updateForTrack(descriptionsTrack); + } + } + + /** + * Updates the displayed TextTrack to be sure it overlays the video when a either + * a {@link Player#texttrackchange} or a {@link Player#fullscreenchange} is fired. + */ + updateDisplayOverlay() { + // inset-inline and inset-block are not supprted on old chrome, but these are + // only likely to be used on TV devices + if (!this.player_.videoHeight() || !window$1.CSS.supports('inset-inline: 10px')) { + return; + } + const playerWidth = this.player_.currentWidth(); + const playerHeight = this.player_.currentHeight(); + const playerAspectRatio = playerWidth / playerHeight; + const videoAspectRatio = this.player_.videoWidth() / this.player_.videoHeight(); + let insetInlineMatch = 0; + let insetBlockMatch = 0; + if (Math.abs(playerAspectRatio - videoAspectRatio) > 0.1) { + if (playerAspectRatio > videoAspectRatio) { + insetInlineMatch = Math.round((playerWidth - playerHeight * videoAspectRatio) / 2); + } else { + insetBlockMatch = Math.round((playerHeight - playerWidth / videoAspectRatio) / 2); + } + } + tryUpdateStyle(this.el_, 'insetInline', getCSSPositionValue(insetInlineMatch)); + tryUpdateStyle(this.el_, 'insetBlock', getCSSPositionValue(insetBlockMatch)); + } + + /** + * Style {@Link TextTrack} activeCues according to {@Link TextTrackSettings}. + * + * @param {TextTrack} track + * Text track object containing active cues to style. + */ + updateDisplayState(track) { + const overrides = this.player_.textTrackSettings.getValues(); + const cues = track.activeCues; + let i = cues.length; + while (i--) { + const cue = cues[i]; + if (!cue) { + continue; + } + const cueDiv = cue.displayState; + if (overrides.color) { + cueDiv.firstChild.style.color = overrides.color; + } + if (overrides.textOpacity) { + tryUpdateStyle(cueDiv.firstChild, 'color', constructColor(overrides.color || '#fff', overrides.textOpacity)); + } + if (overrides.backgroundColor) { + cueDiv.firstChild.style.backgroundColor = overrides.backgroundColor; + } + if (overrides.backgroundOpacity) { + tryUpdateStyle(cueDiv.firstChild, 'backgroundColor', constructColor(overrides.backgroundColor || '#000', overrides.backgroundOpacity)); + } + if (overrides.windowColor) { + if (overrides.windowOpacity) { + tryUpdateStyle(cueDiv, 'backgroundColor', constructColor(overrides.windowColor, overrides.windowOpacity)); + } else { + cueDiv.style.backgroundColor = overrides.windowColor; + } + } + if (overrides.edgeStyle) { + if (overrides.edgeStyle === 'dropshadow') { + cueDiv.firstChild.style.textShadow = `2px 2px 3px ${darkGray}, 2px 2px 4px ${darkGray}, 2px 2px 5px ${darkGray}`; + } else if (overrides.edgeStyle === 'raised') { + cueDiv.firstChild.style.textShadow = `1px 1px ${darkGray}, 2px 2px ${darkGray}, 3px 3px ${darkGray}`; + } else if (overrides.edgeStyle === 'depressed') { + cueDiv.firstChild.style.textShadow = `1px 1px ${lightGray}, 0 1px ${lightGray}, -1px -1px ${darkGray}, 0 -1px ${darkGray}`; + } else if (overrides.edgeStyle === 'uniform') { + cueDiv.firstChild.style.textShadow = `0 0 4px ${darkGray}, 0 0 4px ${darkGray}, 0 0 4px ${darkGray}, 0 0 4px ${darkGray}`; + } + } + if (overrides.fontPercent && overrides.fontPercent !== 1) { + const fontSize = window$1.parseFloat(cueDiv.style.fontSize); + cueDiv.style.fontSize = fontSize * overrides.fontPercent + 'px'; + cueDiv.style.height = 'auto'; + cueDiv.style.top = 'auto'; + } + if (overrides.fontFamily && overrides.fontFamily !== 'default') { + if (overrides.fontFamily === 'small-caps') { + cueDiv.firstChild.style.fontVariant = 'small-caps'; + } else { + cueDiv.firstChild.style.fontFamily = fontMap[overrides.fontFamily]; + } + } + } + } + + /** + * Add an {@link TextTrack} to to the {@link Tech}s {@link TextTrackList}. + * + * @param {TextTrack|TextTrack[]} tracks + * Text track object or text track array to be added to the list. + */ + updateForTrack(tracks) { + if (!Array.isArray(tracks)) { + tracks = [tracks]; + } + if (typeof window$1.WebVTT !== 'function' || tracks.every(track => { + return !track.activeCues; + })) { + return; + } + const cues = []; + + // push all active track cues + for (let i = 0; i < tracks.length; ++i) { + const track = tracks[i]; + for (let j = 0; j < track.activeCues.length; ++j) { + cues.push(track.activeCues[j]); + } + } + + // removes all cues before it processes new ones + window$1.WebVTT.processCues(window$1, cues, this.el_); + + // add unique class to each language text track & add settings styling if necessary + for (let i = 0; i < tracks.length; ++i) { + const track = tracks[i]; + for (let j = 0; j < track.activeCues.length; ++j) { + const cueEl = track.activeCues[j].displayState; + addClass(cueEl, 'vjs-text-track-cue', 'vjs-text-track-cue-' + (track.language ? track.language : i)); + if (track.language) { + setAttribute(cueEl, 'lang', track.language); + } + } + if (this.player_.textTrackSettings) { + this.updateDisplayState(track); + } + } + } +} +Component$1.registerComponent('TextTrackDisplay', TextTrackDisplay); + +/** + * @file loading-spinner.js + */ + +/** + * A loading spinner for use during waiting/loading events. + * + * @extends Component + */ +class LoadingSpinner extends Component$1 { + /** + * Create the `LoadingSpinner`s DOM element. + * + * @return {Element} + * The dom element that gets created. + */ + createEl() { + const isAudio = this.player_.isAudio(); + const playerType = this.localize(isAudio ? 'Audio Player' : 'Video Player'); + const controlText = createEl('span', { + className: 'vjs-control-text', + textContent: this.localize('{1} is loading.', [playerType]) + }); + const el = super.createEl('div', { + className: 'vjs-loading-spinner', + dir: 'ltr' + }); + el.appendChild(controlText); + return el; + } + + /** + * Update control text on languagechange + */ + handleLanguagechange() { + this.$('.vjs-control-text').textContent = this.localize('{1} is loading.', [this.player_.isAudio() ? 'Audio Player' : 'Video Player']); + } +} +Component$1.registerComponent('LoadingSpinner', LoadingSpinner); + +/** + * @file button.js + */ + +/** + * Base class for all buttons. + * + * @extends ClickableComponent + */ +class Button extends ClickableComponent { + /** + * Create the `Button`s DOM element. + * + * @param {string} [tag="button"] + * The element's node type. This argument is IGNORED: no matter what + * is passed, it will always create a `button` element. + * + * @param {Object} [props={}] + * An object of properties that should be set on the element. + * + * @param {Object} [attributes={}] + * An object of attributes that should be set on the element. + * + * @return {Element} + * The element that gets created. + */ + createEl(tag, props = {}, attributes = {}) { + tag = 'button'; + props = Object.assign({ + className: this.buildCSSClass() + }, props); + + // Add attributes for button element + attributes = Object.assign({ + // Necessary since the default button type is "submit" + type: 'button' + }, attributes); + const el = createEl(tag, props, attributes); + if (!this.player_.options_.experimentalSvgIcons) { + el.appendChild(createEl('span', { + className: 'vjs-icon-placeholder' + }, { + 'aria-hidden': true + })); + } + this.createControlTextEl(el); + return el; + } + + /** + * Add a child `Component` inside of this `Button`. + * + * @param {string|Component} child + * The name or instance of a child to add. + * + * @param {Object} [options={}] + * The key/value store of options that will get passed to children of + * the child. + * + * @return {Component} + * The `Component` that gets added as a child. When using a string the + * `Component` will get created by this process. + * + * @deprecated since version 5 + */ + addChild(child, options = {}) { + const className = this.constructor.name; + log$1.warn(`Adding an actionable (user controllable) child to a Button (${className}) is not supported; use a ClickableComponent instead.`); + + // Avoid the error message generated by ClickableComponent's addChild method + return Component$1.prototype.addChild.call(this, child, options); + } + + /** + * Enable the `Button` element so that it can be activated or clicked. Use this with + * {@link Button#disable}. + */ + enable() { + super.enable(); + this.el_.removeAttribute('disabled'); + } + + /** + * Disable the `Button` element so that it cannot be activated or clicked. Use this with + * {@link Button#enable}. + */ + disable() { + super.disable(); + this.el_.setAttribute('disabled', 'disabled'); + } + + /** + * This gets called when a `Button` has focus and `keydown` is triggered via a key + * press. + * + * @param {KeyboardEvent} event + * The event that caused this function to get called. + * + * @listens keydown + */ + handleKeyDown(event) { + // Ignore Space or Enter key operation, which is handled by the browser for + // a button - though not for its super class, ClickableComponent. Also, + // prevent the event from propagating through the DOM and triggering Player + // hotkeys. We do not preventDefault here because we _want_ the browser to + // handle it. + if (keycode.isEventKey(event, 'Space') || keycode.isEventKey(event, 'Enter')) { + event.stopPropagation(); + return; + } + + // Pass keypress handling up for unsupported keys + super.handleKeyDown(event); + } +} +Component$1.registerComponent('Button', Button); + +/** + * @file big-play-button.js + */ + +/** + * The initial play button that shows before the video has played. The hiding of the + * `BigPlayButton` get done via CSS and `Player` states. + * + * @extends Button + */ +class BigPlayButton extends Button { + constructor(player, options) { + super(player, options); + this.mouseused_ = false; + this.setIcon('play'); + this.on('mousedown', e => this.handleMouseDown(e)); + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. Always returns 'vjs-big-play-button'. + */ + buildCSSClass() { + return 'vjs-big-play-button'; + } + + /** + * This gets called when a `BigPlayButton` "clicked". See {@link ClickableComponent} + * for more detailed information on what a click can be. + * + * @param {KeyboardEvent|MouseEvent|TouchEvent} event + * The `keydown`, `tap`, or `click` event that caused this function to be + * called. + * + * @listens tap + * @listens click + */ + handleClick(event) { + const playPromise = this.player_.play(); + + // exit early if clicked via the mouse + if (this.mouseused_ && 'clientX' in event && 'clientY' in event) { + silencePromise(playPromise); + if (this.player_.tech(true)) { + this.player_.tech(true).focus(); + } + return; + } + const cb = this.player_.getChild('controlBar'); + const playToggle = cb && cb.getChild('playToggle'); + if (!playToggle) { + this.player_.tech(true).focus(); + return; + } + const playFocus = () => playToggle.focus(); + if (isPromise(playPromise)) { + playPromise.then(playFocus, () => {}); + } else { + this.setTimeout(playFocus, 1); + } + } + + /** + * Event handler that is called when a `BigPlayButton` receives a + * `keydown` event. + * + * @param {KeyboardEvent} event + * The `keydown` event that caused this function to be called. + * + * @listens keydown + */ + handleKeyDown(event) { + this.mouseused_ = false; + super.handleKeyDown(event); + } + + /** + * Handle `mousedown` events on the `BigPlayButton`. + * + * @param {MouseEvent} event + * `mousedown` or `touchstart` event that triggered this function + * + * @listens mousedown + */ + handleMouseDown(event) { + this.mouseused_ = true; + } +} + +/** + * The text that should display over the `BigPlayButton`s controls. Added to for localization. + * + * @type {string} + * @protected + */ +BigPlayButton.prototype.controlText_ = 'Play Video'; +Component$1.registerComponent('BigPlayButton', BigPlayButton); + +/** + * @file close-button.js + */ + +/** + * The `CloseButton` is a `{@link Button}` that fires a `close` event when + * it gets clicked. + * + * @extends Button + */ +class CloseButton extends Button { + /** + * Creates an instance of the this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.setIcon('cancel'); + this.controlText(options && options.controlText || this.localize('Close')); + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-close-button ${super.buildCSSClass()}`; + } + + /** + * This gets called when a `CloseButton` gets clicked. See + * {@link ClickableComponent#handleClick} for more information on when + * this will be triggered + * + * @param {Event} event + * The `keydown`, `tap`, or `click` event that caused this function to be + * called. + * + * @listens tap + * @listens click + * @fires CloseButton#close + */ + handleClick(event) { + /** + * Triggered when the a `CloseButton` is clicked. + * + * @event CloseButton#close + * @type {Event} + * + * @property {boolean} [bubbles=false] + * set to false so that the close event does not + * bubble up to parents if there is no listener + */ + this.trigger({ + type: 'close', + bubbles: false + }); + } + /** + * Event handler that is called when a `CloseButton` receives a + * `keydown` event. + * + * By default, if the key is Esc, it will trigger a `click` event. + * + * @param {KeyboardEvent} event + * The `keydown` event that caused this function to be called. + * + * @listens keydown + */ + handleKeyDown(event) { + // Esc button will trigger `click` event + if (keycode.isEventKey(event, 'Esc')) { + event.preventDefault(); + event.stopPropagation(); + this.trigger('click'); + } else { + // Pass keypress handling up for unsupported keys + super.handleKeyDown(event); + } + } +} +Component$1.registerComponent('CloseButton', CloseButton); + +/** + * @file play-toggle.js + */ + +/** + * Button to toggle between play and pause. + * + * @extends Button + */ +class PlayToggle extends Button { + /** + * Creates an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options={}] + * The key/value store of player options. + */ + constructor(player, options = {}) { + super(player, options); + + // show or hide replay icon + options.replay = options.replay === undefined || options.replay; + this.setIcon('play'); + this.on(player, 'play', e => this.handlePlay(e)); + this.on(player, 'pause', e => this.handlePause(e)); + if (options.replay) { + this.on(player, 'ended', e => this.handleEnded(e)); + } + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-play-control ${super.buildCSSClass()}`; + } + + /** + * This gets called when an `PlayToggle` is "clicked". See + * {@link ClickableComponent} for more detailed information on what a click can be. + * + * @param {Event} [event] + * The `keydown`, `tap`, or `click` event that caused this function to be + * called. + * + * @listens tap + * @listens click + */ + handleClick(event) { + if (this.player_.paused()) { + silencePromise(this.player_.play()); + } else { + this.player_.pause(); + } + } + + /** + * This gets called once after the video has ended and the user seeks so that + * we can change the replay button back to a play button. + * + * @param {Event} [event] + * The event that caused this function to run. + * + * @listens Player#seeked + */ + handleSeeked(event) { + this.removeClass('vjs-ended'); + if (this.player_.paused()) { + this.handlePause(event); + } else { + this.handlePlay(event); + } + } + + /** + * Add the vjs-playing class to the element so it can change appearance. + * + * @param {Event} [event] + * The event that caused this function to run. + * + * @listens Player#play + */ + handlePlay(event) { + this.removeClass('vjs-ended', 'vjs-paused'); + this.addClass('vjs-playing'); + // change the button text to "Pause" + this.setIcon('pause'); + this.controlText('Pause'); + } + + /** + * Add the vjs-paused class to the element so it can change appearance. + * + * @param {Event} [event] + * The event that caused this function to run. + * + * @listens Player#pause + */ + handlePause(event) { + this.removeClass('vjs-playing'); + this.addClass('vjs-paused'); + // change the button text to "Play" + this.setIcon('play'); + this.controlText('Play'); + } + + /** + * Add the vjs-ended class to the element so it can change appearance + * + * @param {Event} [event] + * The event that caused this function to run. + * + * @listens Player#ended + */ + handleEnded(event) { + this.removeClass('vjs-playing'); + this.addClass('vjs-ended'); + // change the button text to "Replay" + this.setIcon('replay'); + this.controlText('Replay'); + + // on the next seek remove the replay button + this.one(this.player_, 'seeked', e => this.handleSeeked(e)); + } +} + +/** + * The text that should display over the `PlayToggle`s controls. Added for localization. + * + * @type {string} + * @protected + */ +PlayToggle.prototype.controlText_ = 'Play'; +Component$1.registerComponent('PlayToggle', PlayToggle); + +/** + * @file time-display.js + */ + +/** + * Displays time information about the video + * + * @extends Component + */ +class TimeDisplay extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.on(player, ['timeupdate', 'ended', 'seeking'], e => this.update(e)); + this.updateTextNode_(); + } + + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + const className = this.buildCSSClass(); + const el = super.createEl('div', { + className: `${className} vjs-time-control vjs-control` + }); + const span = createEl('span', { + className: 'vjs-control-text', + textContent: `${this.localize(this.labelText_)}\u00a0` + }, { + role: 'presentation' + }); + el.appendChild(span); + this.contentEl_ = createEl('span', { + className: `${className}-display` + }, { + // span elements have no implicit role, but some screen readers (notably VoiceOver) + // treat them as a break between items in the DOM when using arrow keys + // (or left-to-right swipes on iOS) to read contents of a page. Using + // role='presentation' causes VoiceOver to NOT treat this span as a break. + role: 'presentation' + }); + el.appendChild(this.contentEl_); + return el; + } + dispose() { + this.contentEl_ = null; + this.textNode_ = null; + super.dispose(); + } + + /** + * Updates the displayed time according to the `updateContent` function which is defined in the child class. + * + * @param {Event} [event] + * The `timeupdate`, `ended` or `seeking` (if enableSmoothSeeking is true) event that caused this function to be called. + */ + update(event) { + if (!this.player_.options_.enableSmoothSeeking && event.type === 'seeking') { + return; + } + this.updateContent(event); + } + + /** + * Updates the time display text node with a new time + * + * @param {number} [time=0] the time to update to + * + * @private + */ + updateTextNode_(time = 0) { + time = formatTime(time); + if (this.formattedTime_ === time) { + return; + } + this.formattedTime_ = time; + this.requestNamedAnimationFrame('TimeDisplay#updateTextNode_', () => { + if (!this.contentEl_) { + return; + } + let oldNode = this.textNode_; + if (oldNode && this.contentEl_.firstChild !== oldNode) { + oldNode = null; + log$1.warn('TimeDisplay#updateTextnode_: Prevented replacement of text node element since it was no longer a child of this node. Appending a new node instead.'); + } + this.textNode_ = document.createTextNode(this.formattedTime_); + if (!this.textNode_) { + return; + } + if (oldNode) { + this.contentEl_.replaceChild(this.textNode_, oldNode); + } else { + this.contentEl_.appendChild(this.textNode_); + } + }); + } + + /** + * To be filled out in the child class, should update the displayed time + * in accordance with the fact that the current time has changed. + * + * @param {Event} [event] + * The `timeupdate` event that caused this to run. + * + * @listens Player#timeupdate + */ + updateContent(event) {} +} + +/** + * The text that is added to the `TimeDisplay` for screen reader users. + * + * @type {string} + * @private + */ +TimeDisplay.prototype.labelText_ = 'Time'; + +/** + * The text that should display over the `TimeDisplay`s controls. Added to for localization. + * + * @type {string} + * @protected + * + * @deprecated in v7; controlText_ is not used in non-active display Components + */ +TimeDisplay.prototype.controlText_ = 'Time'; +Component$1.registerComponent('TimeDisplay', TimeDisplay); + +/** + * @file current-time-display.js + */ + +/** + * Displays the current time + * + * @extends Component + */ +class CurrentTimeDisplay extends TimeDisplay { + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return 'vjs-current-time'; + } + + /** + * Update current time display + * + * @param {Event} [event] + * The `timeupdate` event that caused this function to run. + * + * @listens Player#timeupdate + */ + updateContent(event) { + // Allows for smooth scrubbing, when player can't keep up. + let time; + if (this.player_.ended()) { + time = this.player_.duration(); + } else { + time = this.player_.scrubbing() ? this.player_.getCache().currentTime : this.player_.currentTime(); + } + this.updateTextNode_(time); + } +} + +/** + * The text that is added to the `CurrentTimeDisplay` for screen reader users. + * + * @type {string} + * @private + */ +CurrentTimeDisplay.prototype.labelText_ = 'Current Time'; + +/** + * The text that should display over the `CurrentTimeDisplay`s controls. Added to for localization. + * + * @type {string} + * @protected + * + * @deprecated in v7; controlText_ is not used in non-active display Components + */ +CurrentTimeDisplay.prototype.controlText_ = 'Current Time'; +Component$1.registerComponent('CurrentTimeDisplay', CurrentTimeDisplay); + +/** + * @file duration-display.js + */ + +/** + * Displays the duration + * + * @extends Component + */ +class DurationDisplay extends TimeDisplay { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + const updateContent = e => this.updateContent(e); + + // we do not want to/need to throttle duration changes, + // as they should always display the changed duration as + // it has changed + this.on(player, 'durationchange', updateContent); + + // Listen to loadstart because the player duration is reset when a new media element is loaded, + // but the durationchange on the user agent will not fire. + // @see [Spec]{@link https://www.w3.org/TR/2011/WD-html5-20110113/video.html#media-element-load-algorithm} + this.on(player, 'loadstart', updateContent); + + // Also listen for timeupdate (in the parent) and loadedmetadata because removing those + // listeners could have broken dependent applications/libraries. These + // can likely be removed for 7.0. + this.on(player, 'loadedmetadata', updateContent); + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return 'vjs-duration'; + } + + /** + * Update duration time display. + * + * @param {Event} [event] + * The `durationchange`, `timeupdate`, or `loadedmetadata` event that caused + * this function to be called. + * + * @listens Player#durationchange + * @listens Player#timeupdate + * @listens Player#loadedmetadata + */ + updateContent(event) { + const duration = this.player_.duration(); + this.updateTextNode_(duration); + } +} + +/** + * The text that is added to the `DurationDisplay` for screen reader users. + * + * @type {string} + * @private + */ +DurationDisplay.prototype.labelText_ = 'Duration'; + +/** + * The text that should display over the `DurationDisplay`s controls. Added to for localization. + * + * @type {string} + * @protected + * + * @deprecated in v7; controlText_ is not used in non-active display Components + */ +DurationDisplay.prototype.controlText_ = 'Duration'; +Component$1.registerComponent('DurationDisplay', DurationDisplay); + +/** + * @file time-divider.js + */ + +/** + * The separator between the current time and duration. + * Can be hidden if it's not needed in the design. + * + * @extends Component + */ +class TimeDivider extends Component$1 { + /** + * Create the component's DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + const el = super.createEl('div', { + className: 'vjs-time-control vjs-time-divider' + }, { + // this element and its contents can be hidden from assistive techs since + // it is made extraneous by the announcement of the control text + // for the current time and duration displays + 'aria-hidden': true + }); + const div = super.createEl('div'); + const span = super.createEl('span', { + textContent: '/' + }); + div.appendChild(span); + el.appendChild(div); + return el; + } +} +Component$1.registerComponent('TimeDivider', TimeDivider); + +/** + * @file remaining-time-display.js + */ + +/** + * Displays the time left in the video + * + * @extends Component + */ +class RemainingTimeDisplay extends TimeDisplay { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.on(player, 'durationchange', e => this.updateContent(e)); + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return 'vjs-remaining-time'; + } + + /** + * Create the `Component`'s DOM element with the "minus" character prepend to the time + * + * @return {Element} + * The element that was created. + */ + createEl() { + const el = super.createEl(); + if (this.options_.displayNegative !== false) { + el.insertBefore(createEl('span', {}, { + 'aria-hidden': true + }, '-'), this.contentEl_); + } + return el; + } + + /** + * Update remaining time display. + * + * @param {Event} [event] + * The `timeupdate` or `durationchange` event that caused this to run. + * + * @listens Player#timeupdate + * @listens Player#durationchange + */ + updateContent(event) { + if (typeof this.player_.duration() !== 'number') { + return; + } + let time; + + // @deprecated We should only use remainingTimeDisplay + // as of video.js 7 + if (this.player_.ended()) { + time = 0; + } else if (this.player_.remainingTimeDisplay) { + time = this.player_.remainingTimeDisplay(); + } else { + time = this.player_.remainingTime(); + } + this.updateTextNode_(time); + } +} + +/** + * The text that is added to the `RemainingTimeDisplay` for screen reader users. + * + * @type {string} + * @private + */ +RemainingTimeDisplay.prototype.labelText_ = 'Remaining Time'; + +/** + * The text that should display over the `RemainingTimeDisplay`s controls. Added to for localization. + * + * @type {string} + * @protected + * + * @deprecated in v7; controlText_ is not used in non-active display Components + */ +RemainingTimeDisplay.prototype.controlText_ = 'Remaining Time'; +Component$1.registerComponent('RemainingTimeDisplay', RemainingTimeDisplay); + +/** + * @file live-display.js + */ + +// TODO - Future make it click to snap to live + +/** + * Displays the live indicator when duration is Infinity. + * + * @extends Component + */ +class LiveDisplay extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.updateShowing(); + this.on(this.player(), 'durationchange', e => this.updateShowing(e)); + } + + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + const el = super.createEl('div', { + className: 'vjs-live-control vjs-control' + }); + this.contentEl_ = createEl('div', { + className: 'vjs-live-display' + }, { + 'aria-live': 'off' + }); + this.contentEl_.appendChild(createEl('span', { + className: 'vjs-control-text', + textContent: `${this.localize('Stream Type')}\u00a0` + })); + this.contentEl_.appendChild(document.createTextNode(this.localize('LIVE'))); + el.appendChild(this.contentEl_); + return el; + } + dispose() { + this.contentEl_ = null; + super.dispose(); + } + + /** + * Check the duration to see if the LiveDisplay should be showing or not. Then show/hide + * it accordingly + * + * @param {Event} [event] + * The {@link Player#durationchange} event that caused this function to run. + * + * @listens Player#durationchange + */ + updateShowing(event) { + if (this.player().duration() === Infinity) { + this.show(); + } else { + this.hide(); + } + } +} +Component$1.registerComponent('LiveDisplay', LiveDisplay); + +/** + * @file seek-to-live.js + */ + +/** + * Displays the live indicator when duration is Infinity. + * + * @extends Component + */ +class SeekToLive extends Button { + /** + * Creates an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.updateLiveEdgeStatus(); + if (this.player_.liveTracker) { + this.updateLiveEdgeStatusHandler_ = e => this.updateLiveEdgeStatus(e); + this.on(this.player_.liveTracker, 'liveedgechange', this.updateLiveEdgeStatusHandler_); + } + } + + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + const el = super.createEl('button', { + className: 'vjs-seek-to-live-control vjs-control' + }); + this.setIcon('circle', el); + this.textEl_ = createEl('span', { + className: 'vjs-seek-to-live-text', + textContent: this.localize('LIVE') + }, { + 'aria-hidden': 'true' + }); + el.appendChild(this.textEl_); + return el; + } + + /** + * Update the state of this button if we are at the live edge + * or not + */ + updateLiveEdgeStatus() { + // default to live edge + if (!this.player_.liveTracker || this.player_.liveTracker.atLiveEdge()) { + this.setAttribute('aria-disabled', true); + this.addClass('vjs-at-live-edge'); + this.controlText('Seek to live, currently playing live'); + } else { + this.setAttribute('aria-disabled', false); + this.removeClass('vjs-at-live-edge'); + this.controlText('Seek to live, currently behind live'); + } + } + + /** + * On click bring us as near to the live point as possible. + * This requires that we wait for the next `live-seekable-change` + * event which will happen every segment length seconds. + */ + handleClick() { + this.player_.liveTracker.seekToLiveEdge(); + } + + /** + * Dispose of the element and stop tracking + */ + dispose() { + if (this.player_.liveTracker) { + this.off(this.player_.liveTracker, 'liveedgechange', this.updateLiveEdgeStatusHandler_); + } + this.textEl_ = null; + super.dispose(); + } +} +/** + * The text that should display over the `SeekToLive`s control. Added for localization. + * + * @type {string} + * @protected + */ +SeekToLive.prototype.controlText_ = 'Seek to live, currently playing live'; +Component$1.registerComponent('SeekToLive', SeekToLive); + +/** + * @file num.js + * @module num + */ + +/** + * Keep a number between a min and a max value + * + * @param {number} number + * The number to clamp + * + * @param {number} min + * The minimum value + * @param {number} max + * The maximum value + * + * @return {number} + * the clamped number + */ +function clamp(number, min, max) { + number = Number(number); + return Math.min(max, Math.max(min, isNaN(number) ? min : number)); +} + +var Num = /*#__PURE__*/Object.freeze({ + __proto__: null, + clamp: clamp +}); + +/** + * @file slider.js + */ + +/** + * The base functionality for a slider. Can be vertical or horizontal. + * For instance the volume bar or the seek bar on a video is a slider. + * + * @extends Component + */ +class Slider extends Component$1 { + /** + * Create an instance of this class + * + * @param { import('../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.handleMouseDown_ = e => this.handleMouseDown(e); + this.handleMouseUp_ = e => this.handleMouseUp(e); + this.handleKeyDown_ = e => this.handleKeyDown(e); + this.handleClick_ = e => this.handleClick(e); + this.handleMouseMove_ = e => this.handleMouseMove(e); + this.update_ = e => this.update(e); + + // Set property names to bar to match with the child Slider class is looking for + this.bar = this.getChild(this.options_.barName); + + // Set a horizontal or vertical class on the slider depending on the slider type + this.vertical(!!this.options_.vertical); + this.enable(); + } + + /** + * Are controls are currently enabled for this slider or not. + * + * @return {boolean} + * true if controls are enabled, false otherwise + */ + enabled() { + return this.enabled_; + } + + /** + * Enable controls for this slider if they are disabled + */ + enable() { + if (this.enabled()) { + return; + } + this.on('mousedown', this.handleMouseDown_); + this.on('touchstart', this.handleMouseDown_); + this.on('keydown', this.handleKeyDown_); + this.on('click', this.handleClick_); + + // TODO: deprecated, controlsvisible does not seem to be fired + this.on(this.player_, 'controlsvisible', this.update); + if (this.playerEvent) { + this.on(this.player_, this.playerEvent, this.update); + } + this.removeClass('disabled'); + this.setAttribute('tabindex', 0); + this.enabled_ = true; + } + + /** + * Disable controls for this slider if they are enabled + */ + disable() { + if (!this.enabled()) { + return; + } + const doc = this.bar.el_.ownerDocument; + this.off('mousedown', this.handleMouseDown_); + this.off('touchstart', this.handleMouseDown_); + this.off('keydown', this.handleKeyDown_); + this.off('click', this.handleClick_); + this.off(this.player_, 'controlsvisible', this.update_); + this.off(doc, 'mousemove', this.handleMouseMove_); + this.off(doc, 'mouseup', this.handleMouseUp_); + this.off(doc, 'touchmove', this.handleMouseMove_); + this.off(doc, 'touchend', this.handleMouseUp_); + this.removeAttribute('tabindex'); + this.addClass('disabled'); + if (this.playerEvent) { + this.off(this.player_, this.playerEvent, this.update); + } + this.enabled_ = false; + } + + /** + * Create the `Slider`s DOM element. + * + * @param {string} type + * Type of element to create. + * + * @param {Object} [props={}] + * List of properties in Object form. + * + * @param {Object} [attributes={}] + * list of attributes in Object form. + * + * @return {Element} + * The element that gets created. + */ + createEl(type, props = {}, attributes = {}) { + // Add the slider element class to all sub classes + props.className = props.className + ' vjs-slider'; + props = Object.assign({ + tabIndex: 0 + }, props); + attributes = Object.assign({ + 'role': 'slider', + 'aria-valuenow': 0, + 'aria-valuemin': 0, + 'aria-valuemax': 100 + }, attributes); + return super.createEl(type, props, attributes); + } + + /** + * Handle `mousedown` or `touchstart` events on the `Slider`. + * + * @param {MouseEvent} event + * `mousedown` or `touchstart` event that triggered this function + * + * @listens mousedown + * @listens touchstart + * @fires Slider#slideractive + */ + handleMouseDown(event) { + const doc = this.bar.el_.ownerDocument; + if (event.type === 'mousedown') { + event.preventDefault(); + } + // Do not call preventDefault() on touchstart in Chrome + // to avoid console warnings. Use a 'touch-action: none' style + // instead to prevent unintended scrolling. + // https://developers.google.com/web/updates/2017/01/scrolling-intervention + if (event.type === 'touchstart' && !IS_CHROME) { + event.preventDefault(); + } + blockTextSelection(); + this.addClass('vjs-sliding'); + /** + * Triggered when the slider is in an active state + * + * @event Slider#slideractive + * @type {MouseEvent} + */ + this.trigger('slideractive'); + this.on(doc, 'mousemove', this.handleMouseMove_); + this.on(doc, 'mouseup', this.handleMouseUp_); + this.on(doc, 'touchmove', this.handleMouseMove_); + this.on(doc, 'touchend', this.handleMouseUp_); + this.handleMouseMove(event, true); + } + + /** + * Handle the `mousemove`, `touchmove`, and `mousedown` events on this `Slider`. + * The `mousemove` and `touchmove` events will only only trigger this function during + * `mousedown` and `touchstart`. This is due to {@link Slider#handleMouseDown} and + * {@link Slider#handleMouseUp}. + * + * @param {MouseEvent} event + * `mousedown`, `mousemove`, `touchstart`, or `touchmove` event that triggered + * this function + * @param {boolean} mouseDown this is a flag that should be set to true if `handleMouseMove` is called directly. It allows us to skip things that should not happen if coming from mouse down but should happen on regular mouse move handler. Defaults to false. + * + * @listens mousemove + * @listens touchmove + */ + handleMouseMove(event) {} + + /** + * Handle `mouseup` or `touchend` events on the `Slider`. + * + * @param {MouseEvent} event + * `mouseup` or `touchend` event that triggered this function. + * + * @listens touchend + * @listens mouseup + * @fires Slider#sliderinactive + */ + handleMouseUp(event) { + const doc = this.bar.el_.ownerDocument; + unblockTextSelection(); + this.removeClass('vjs-sliding'); + /** + * Triggered when the slider is no longer in an active state. + * + * @event Slider#sliderinactive + * @type {Event} + */ + this.trigger('sliderinactive'); + this.off(doc, 'mousemove', this.handleMouseMove_); + this.off(doc, 'mouseup', this.handleMouseUp_); + this.off(doc, 'touchmove', this.handleMouseMove_); + this.off(doc, 'touchend', this.handleMouseUp_); + this.update(); + } + + /** + * Update the progress bar of the `Slider`. + * + * @return {number} + * The percentage of progress the progress bar represents as a + * number from 0 to 1. + */ + update() { + // In VolumeBar init we have a setTimeout for update that pops and update + // to the end of the execution stack. The player is destroyed before then + // update will cause an error + // If there's no bar... + if (!this.el_ || !this.bar) { + return; + } + + // clamp progress between 0 and 1 + // and only round to four decimal places, as we round to two below + const progress = this.getProgress(); + if (progress === this.progress_) { + return progress; + } + this.progress_ = progress; + this.requestNamedAnimationFrame('Slider#update', () => { + // Set the new bar width or height + const sizeKey = this.vertical() ? 'height' : 'width'; + + // Convert to a percentage for css value + this.bar.el().style[sizeKey] = (progress * 100).toFixed(2) + '%'; + }); + return progress; + } + + /** + * Get the percentage of the bar that should be filled + * but clamped and rounded. + * + * @return {number} + * percentage filled that the slider is + */ + getProgress() { + return Number(clamp(this.getPercent(), 0, 1).toFixed(4)); + } + + /** + * Calculate distance for slider + * + * @param {Event} event + * The event that caused this function to run. + * + * @return {number} + * The current position of the Slider. + * - position.x for vertical `Slider`s + * - position.y for horizontal `Slider`s + */ + calculateDistance(event) { + const position = getPointerPosition(this.el_, event); + if (this.vertical()) { + return position.y; + } + return position.x; + } + + /** + * Handle a `keydown` event on the `Slider`. Watches for left, right, up, and down + * arrow keys. This function will only be called when the slider has focus. See + * {@link Slider#handleFocus} and {@link Slider#handleBlur}. + * + * @param {KeyboardEvent} event + * the `keydown` event that caused this function to run. + * + * @listens keydown + */ + handleKeyDown(event) { + // Left and Down Arrows + if (keycode.isEventKey(event, 'Left') || keycode.isEventKey(event, 'Down')) { + event.preventDefault(); + event.stopPropagation(); + this.stepBack(); + + // Up and Right Arrows + } else if (keycode.isEventKey(event, 'Right') || keycode.isEventKey(event, 'Up')) { + event.preventDefault(); + event.stopPropagation(); + this.stepForward(); + } else { + // Pass keydown handling up for unsupported keys + super.handleKeyDown(event); + } + } + + /** + * Listener for click events on slider, used to prevent clicks + * from bubbling up to parent elements like button menus. + * + * @param {Object} event + * Event that caused this object to run + */ + handleClick(event) { + event.stopPropagation(); + event.preventDefault(); + } + + /** + * Get/set if slider is horizontal for vertical + * + * @param {boolean} [bool] + * - true if slider is vertical, + * - false is horizontal + * + * @return {boolean} + * - true if slider is vertical, and getting + * - false if the slider is horizontal, and getting + */ + vertical(bool) { + if (bool === undefined) { + return this.vertical_ || false; + } + this.vertical_ = !!bool; + if (this.vertical_) { + this.addClass('vjs-slider-vertical'); + } else { + this.addClass('vjs-slider-horizontal'); + } + } +} +Component$1.registerComponent('Slider', Slider); + +/** + * @file load-progress-bar.js + */ + +// get the percent width of a time compared to the total end +const percentify = (time, end) => clamp(time / end * 100, 0, 100).toFixed(2) + '%'; + +/** + * Shows loading progress + * + * @extends Component + */ +class LoadProgressBar extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.partEls_ = []; + this.on(player, 'progress', e => this.update(e)); + } + + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + const el = super.createEl('div', { + className: 'vjs-load-progress' + }); + const wrapper = createEl('span', { + className: 'vjs-control-text' + }); + const loadedText = createEl('span', { + textContent: this.localize('Loaded') + }); + const separator = document.createTextNode(': '); + this.percentageEl_ = createEl('span', { + className: 'vjs-control-text-loaded-percentage', + textContent: '0%' + }); + el.appendChild(wrapper); + wrapper.appendChild(loadedText); + wrapper.appendChild(separator); + wrapper.appendChild(this.percentageEl_); + return el; + } + dispose() { + this.partEls_ = null; + this.percentageEl_ = null; + super.dispose(); + } + + /** + * Update progress bar + * + * @param {Event} [event] + * The `progress` event that caused this function to run. + * + * @listens Player#progress + */ + update(event) { + this.requestNamedAnimationFrame('LoadProgressBar#update', () => { + const liveTracker = this.player_.liveTracker; + const buffered = this.player_.buffered(); + const duration = liveTracker && liveTracker.isLive() ? liveTracker.seekableEnd() : this.player_.duration(); + const bufferedEnd = this.player_.bufferedEnd(); + const children = this.partEls_; + const percent = percentify(bufferedEnd, duration); + if (this.percent_ !== percent) { + // update the width of the progress bar + this.el_.style.width = percent; + // update the control-text + textContent(this.percentageEl_, percent); + this.percent_ = percent; + } + + // add child elements to represent the individual buffered time ranges + for (let i = 0; i < buffered.length; i++) { + const start = buffered.start(i); + const end = buffered.end(i); + let part = children[i]; + if (!part) { + part = this.el_.appendChild(createEl()); + children[i] = part; + } + + // only update if changed + if (part.dataset.start === start && part.dataset.end === end) { + continue; + } + part.dataset.start = start; + part.dataset.end = end; + + // set the percent based on the width of the progress bar (bufferedEnd) + part.style.left = percentify(start, bufferedEnd); + part.style.width = percentify(end - start, bufferedEnd); + } + + // remove unused buffered range elements + for (let i = children.length; i > buffered.length; i--) { + this.el_.removeChild(children[i - 1]); + } + children.length = buffered.length; + }); + } +} +Component$1.registerComponent('LoadProgressBar', LoadProgressBar); + +/** + * @file time-tooltip.js + */ + +/** + * Time tooltips display a time above the progress bar. + * + * @extends Component + */ +class TimeTooltip extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The {@link Player} that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.update = throttle(bind_(this, this.update), UPDATE_REFRESH_INTERVAL); + } + + /** + * Create the time tooltip DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + return super.createEl('div', { + className: 'vjs-time-tooltip' + }, { + 'aria-hidden': 'true' + }); + } + + /** + * Updates the position of the time tooltip relative to the `SeekBar`. + * + * @param {Object} seekBarRect + * The `ClientRect` for the {@link SeekBar} element. + * + * @param {number} seekBarPoint + * A number from 0 to 1, representing a horizontal reference point + * from the left edge of the {@link SeekBar} + */ + update(seekBarRect, seekBarPoint, content) { + const tooltipRect = findPosition(this.el_); + const playerRect = getBoundingClientRect(this.player_.el()); + const seekBarPointPx = seekBarRect.width * seekBarPoint; + + // do nothing if either rect isn't available + // for example, if the player isn't in the DOM for testing + if (!playerRect || !tooltipRect) { + return; + } + + // This is the space left of the `seekBarPoint` available within the bounds + // of the player. We calculate any gap between the left edge of the player + // and the left edge of the `SeekBar` and add the number of pixels in the + // `SeekBar` before hitting the `seekBarPoint` + let spaceLeftOfPoint = seekBarRect.left - playerRect.left + seekBarPointPx; + + // This is the space right of the `seekBarPoint` available within the bounds + // of the player. We calculate the number of pixels from the `seekBarPoint` + // to the right edge of the `SeekBar` and add to that any gap between the + // right edge of the `SeekBar` and the player. + let spaceRightOfPoint = seekBarRect.width - seekBarPointPx + (playerRect.right - seekBarRect.right); + + // spaceRightOfPoint is always NaN for mouse time display + // because the seekbarRect does not have a right property. This causes + // the mouse tool tip to be truncated when it's close to the right edge of the player. + // In such cases, we ignore the `playerRect.right - seekBarRect.right` value when calculating. + // For the sake of consistency, we ignore seekBarRect.left - playerRect.left for the left edge. + if (!spaceRightOfPoint) { + spaceRightOfPoint = seekBarRect.width - seekBarPointPx; + spaceLeftOfPoint = seekBarPointPx; + } + // This is the number of pixels by which the tooltip will need to be pulled + // further to the right to center it over the `seekBarPoint`. + let pullTooltipBy = tooltipRect.width / 2; + + // Adjust the `pullTooltipBy` distance to the left or right depending on + // the results of the space calculations above. + if (spaceLeftOfPoint < pullTooltipBy) { + pullTooltipBy += pullTooltipBy - spaceLeftOfPoint; + } else if (spaceRightOfPoint < pullTooltipBy) { + pullTooltipBy = spaceRightOfPoint; + } + + // Due to the imprecision of decimal/ratio based calculations and varying + // rounding behaviors, there are cases where the spacing adjustment is off + // by a pixel or two. This adds insurance to these calculations. + if (pullTooltipBy < 0) { + pullTooltipBy = 0; + } else if (pullTooltipBy > tooltipRect.width) { + pullTooltipBy = tooltipRect.width; + } + + // prevent small width fluctuations within 0.4px from + // changing the value below. + // This really helps for live to prevent the play + // progress time tooltip from jittering + pullTooltipBy = Math.round(pullTooltipBy); + this.el_.style.right = `-${pullTooltipBy}px`; + this.write(content); + } + + /** + * Write the time to the tooltip DOM element. + * + * @param {string} content + * The formatted time for the tooltip. + */ + write(content) { + textContent(this.el_, content); + } + + /** + * Updates the position of the time tooltip relative to the `SeekBar`. + * + * @param {Object} seekBarRect + * The `ClientRect` for the {@link SeekBar} element. + * + * @param {number} seekBarPoint + * A number from 0 to 1, representing a horizontal reference point + * from the left edge of the {@link SeekBar} + * + * @param {number} time + * The time to update the tooltip to, not used during live playback + * + * @param {Function} cb + * A function that will be called during the request animation frame + * for tooltips that need to do additional animations from the default + */ + updateTime(seekBarRect, seekBarPoint, time, cb) { + this.requestNamedAnimationFrame('TimeTooltip#updateTime', () => { + let content; + const duration = this.player_.duration(); + if (this.player_.liveTracker && this.player_.liveTracker.isLive()) { + const liveWindow = this.player_.liveTracker.liveWindow(); + const secondsBehind = liveWindow - seekBarPoint * liveWindow; + content = (secondsBehind < 1 ? '' : '-') + formatTime(secondsBehind, liveWindow); + } else { + content = formatTime(time, duration); + } + this.update(seekBarRect, seekBarPoint, content); + if (cb) { + cb(); + } + }); + } +} +Component$1.registerComponent('TimeTooltip', TimeTooltip); + +/** + * @file play-progress-bar.js + */ + +/** + * Used by {@link SeekBar} to display media playback progress as part of the + * {@link ProgressControl}. + * + * @extends Component + */ +class PlayProgressBar extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The {@link Player} that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.setIcon('circle'); + this.update = throttle(bind_(this, this.update), UPDATE_REFRESH_INTERVAL); + } + + /** + * Create the the DOM element for this class. + * + * @return {Element} + * The element that was created. + */ + createEl() { + return super.createEl('div', { + className: 'vjs-play-progress vjs-slider-bar' + }, { + 'aria-hidden': 'true' + }); + } + + /** + * Enqueues updates to its own DOM as well as the DOM of its + * {@link TimeTooltip} child. + * + * @param {Object} seekBarRect + * The `ClientRect` for the {@link SeekBar} element. + * + * @param {number} seekBarPoint + * A number from 0 to 1, representing a horizontal reference point + * from the left edge of the {@link SeekBar} + */ + update(seekBarRect, seekBarPoint) { + const timeTooltip = this.getChild('timeTooltip'); + if (!timeTooltip) { + return; + } + const time = this.player_.scrubbing() ? this.player_.getCache().currentTime : this.player_.currentTime(); + timeTooltip.updateTime(seekBarRect, seekBarPoint, time); + } +} + +/** + * Default options for {@link PlayProgressBar}. + * + * @type {Object} + * @private + */ +PlayProgressBar.prototype.options_ = { + children: [] +}; + +// Time tooltips should not be added to a player on mobile devices +if (!IS_IOS && !IS_ANDROID) { + PlayProgressBar.prototype.options_.children.push('timeTooltip'); +} +Component$1.registerComponent('PlayProgressBar', PlayProgressBar); + +/** + * @file mouse-time-display.js + */ + +/** + * The {@link MouseTimeDisplay} component tracks mouse movement over the + * {@link ProgressControl}. It displays an indicator and a {@link TimeTooltip} + * indicating the time which is represented by a given point in the + * {@link ProgressControl}. + * + * @extends Component + */ +class MouseTimeDisplay extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The {@link Player} that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.update = throttle(bind_(this, this.update), UPDATE_REFRESH_INTERVAL); + } + + /** + * Create the DOM element for this class. + * + * @return {Element} + * The element that was created. + */ + createEl() { + return super.createEl('div', { + className: 'vjs-mouse-display' + }); + } + + /** + * Enqueues updates to its own DOM as well as the DOM of its + * {@link TimeTooltip} child. + * + * @param {Object} seekBarRect + * The `ClientRect` for the {@link SeekBar} element. + * + * @param {number} seekBarPoint + * A number from 0 to 1, representing a horizontal reference point + * from the left edge of the {@link SeekBar} + */ + update(seekBarRect, seekBarPoint) { + const time = seekBarPoint * this.player_.duration(); + this.getChild('timeTooltip').updateTime(seekBarRect, seekBarPoint, time, () => { + this.el_.style.left = `${seekBarRect.width * seekBarPoint}px`; + }); + } +} + +/** + * Default options for `MouseTimeDisplay` + * + * @type {Object} + * @private + */ +MouseTimeDisplay.prototype.options_ = { + children: ['timeTooltip'] +}; +Component$1.registerComponent('MouseTimeDisplay', MouseTimeDisplay); + +/** + * @file seek-bar.js + */ + +// The number of seconds the `step*` functions move the timeline. +const STEP_SECONDS = 5; + +// The multiplier of STEP_SECONDS that PgUp/PgDown move the timeline. +const PAGE_KEY_MULTIPLIER = 12; + +/** + * Seek bar and container for the progress bars. Uses {@link PlayProgressBar} + * as its `bar`. + * + * @extends Slider + */ +class SeekBar extends Slider { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.setEventHandlers_(); + } + + /** + * Sets the event handlers + * + * @private + */ + setEventHandlers_() { + this.update_ = bind_(this, this.update); + this.update = throttle(this.update_, UPDATE_REFRESH_INTERVAL); + this.on(this.player_, ['ended', 'durationchange', 'timeupdate'], this.update); + if (this.player_.liveTracker) { + this.on(this.player_.liveTracker, 'liveedgechange', this.update); + } + + // when playing, let's ensure we smoothly update the play progress bar + // via an interval + this.updateInterval = null; + this.enableIntervalHandler_ = e => this.enableInterval_(e); + this.disableIntervalHandler_ = e => this.disableInterval_(e); + this.on(this.player_, ['playing'], this.enableIntervalHandler_); + this.on(this.player_, ['ended', 'pause', 'waiting'], this.disableIntervalHandler_); + + // we don't need to update the play progress if the document is hidden, + // also, this causes the CPU to spike and eventually crash the page on IE11. + if ('hidden' in document && 'visibilityState' in document) { + this.on(document, 'visibilitychange', this.toggleVisibility_); + } + } + toggleVisibility_(e) { + if (document.visibilityState === 'hidden') { + this.cancelNamedAnimationFrame('SeekBar#update'); + this.cancelNamedAnimationFrame('Slider#update'); + this.disableInterval_(e); + } else { + if (!this.player_.ended() && !this.player_.paused()) { + this.enableInterval_(); + } + + // we just switched back to the page and someone may be looking, so, update ASAP + this.update(); + } + } + enableInterval_() { + if (this.updateInterval) { + return; + } + this.updateInterval = this.setInterval(this.update, UPDATE_REFRESH_INTERVAL); + } + disableInterval_(e) { + if (this.player_.liveTracker && this.player_.liveTracker.isLive() && e && e.type !== 'ended') { + return; + } + if (!this.updateInterval) { + return; + } + this.clearInterval(this.updateInterval); + this.updateInterval = null; + } + + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + return super.createEl('div', { + className: 'vjs-progress-holder' + }, { + 'aria-label': this.localize('Progress Bar') + }); + } + + /** + * This function updates the play progress bar and accessibility + * attributes to whatever is passed in. + * + * @param {Event} [event] + * The `timeupdate` or `ended` event that caused this to run. + * + * @listens Player#timeupdate + * + * @return {number} + * The current percent at a number from 0-1 + */ + update(event) { + // ignore updates while the tab is hidden + if (document.visibilityState === 'hidden') { + return; + } + const percent = super.update(); + this.requestNamedAnimationFrame('SeekBar#update', () => { + const currentTime = this.player_.ended() ? this.player_.duration() : this.getCurrentTime_(); + const liveTracker = this.player_.liveTracker; + let duration = this.player_.duration(); + if (liveTracker && liveTracker.isLive()) { + duration = this.player_.liveTracker.liveCurrentTime(); + } + if (this.percent_ !== percent) { + // machine readable value of progress bar (percentage complete) + this.el_.setAttribute('aria-valuenow', (percent * 100).toFixed(2)); + this.percent_ = percent; + } + if (this.currentTime_ !== currentTime || this.duration_ !== duration) { + // human readable value of progress bar (time complete) + this.el_.setAttribute('aria-valuetext', this.localize('progress bar timing: currentTime={1} duration={2}', [formatTime(currentTime, duration), formatTime(duration, duration)], '{1} of {2}')); + this.currentTime_ = currentTime; + this.duration_ = duration; + } + + // update the progress bar time tooltip with the current time + if (this.bar) { + this.bar.update(getBoundingClientRect(this.el()), this.getProgress()); + } + }); + return percent; + } + + /** + * Prevent liveThreshold from causing seeks to seem like they + * are not happening from a user perspective. + * + * @param {number} ct + * current time to seek to + */ + userSeek_(ct) { + if (this.player_.liveTracker && this.player_.liveTracker.isLive()) { + this.player_.liveTracker.nextSeekedFromUser(); + } + this.player_.currentTime(ct); + } + + /** + * Get the value of current time but allows for smooth scrubbing, + * when player can't keep up. + * + * @return {number} + * The current time value to display + * + * @private + */ + getCurrentTime_() { + return this.player_.scrubbing() ? this.player_.getCache().currentTime : this.player_.currentTime(); + } + + /** + * Get the percentage of media played so far. + * + * @return {number} + * The percentage of media played so far (0 to 1). + */ + getPercent() { + const currentTime = this.getCurrentTime_(); + let percent; + const liveTracker = this.player_.liveTracker; + if (liveTracker && liveTracker.isLive()) { + percent = (currentTime - liveTracker.seekableStart()) / liveTracker.liveWindow(); + + // prevent the percent from changing at the live edge + if (liveTracker.atLiveEdge()) { + percent = 1; + } + } else { + percent = currentTime / this.player_.duration(); + } + return percent; + } + + /** + * Handle mouse down on seek bar + * + * @param {MouseEvent} event + * The `mousedown` event that caused this to run. + * + * @listens mousedown + */ + handleMouseDown(event) { + if (!isSingleLeftClick(event)) { + return; + } + + // Stop event propagation to prevent double fire in progress-control.js + event.stopPropagation(); + this.videoWasPlaying = !this.player_.paused(); + this.player_.pause(); + super.handleMouseDown(event); + } + + /** + * Handle mouse move on seek bar + * + * @param {MouseEvent} event + * The `mousemove` event that caused this to run. + * @param {boolean} mouseDown this is a flag that should be set to true if `handleMouseMove` is called directly. It allows us to skip things that should not happen if coming from mouse down but should happen on regular mouse move handler. Defaults to false + * + * @listens mousemove + */ + handleMouseMove(event, mouseDown = false) { + if (!isSingleLeftClick(event) || isNaN(this.player_.duration())) { + return; + } + if (!mouseDown && !this.player_.scrubbing()) { + this.player_.scrubbing(true); + } + let newTime; + const distance = this.calculateDistance(event); + const liveTracker = this.player_.liveTracker; + if (!liveTracker || !liveTracker.isLive()) { + newTime = distance * this.player_.duration(); + + // Don't let video end while scrubbing. + if (newTime === this.player_.duration()) { + newTime = newTime - 0.1; + } + } else { + if (distance >= 0.99) { + liveTracker.seekToLiveEdge(); + return; + } + const seekableStart = liveTracker.seekableStart(); + const seekableEnd = liveTracker.liveCurrentTime(); + newTime = seekableStart + distance * liveTracker.liveWindow(); + + // Don't let video end while scrubbing. + if (newTime >= seekableEnd) { + newTime = seekableEnd; + } + + // Compensate for precision differences so that currentTime is not less + // than seekable start + if (newTime <= seekableStart) { + newTime = seekableStart + 0.1; + } + + // On android seekableEnd can be Infinity sometimes, + // this will cause newTime to be Infinity, which is + // not a valid currentTime. + if (newTime === Infinity) { + return; + } + } + + // Set new time (tell player to seek to new time) + this.userSeek_(newTime); + if (this.player_.options_.enableSmoothSeeking) { + this.update(); + } + } + enable() { + super.enable(); + const mouseTimeDisplay = this.getChild('mouseTimeDisplay'); + if (!mouseTimeDisplay) { + return; + } + mouseTimeDisplay.show(); + } + disable() { + super.disable(); + const mouseTimeDisplay = this.getChild('mouseTimeDisplay'); + if (!mouseTimeDisplay) { + return; + } + mouseTimeDisplay.hide(); + } + + /** + * Handle mouse up on seek bar + * + * @param {MouseEvent} event + * The `mouseup` event that caused this to run. + * + * @listens mouseup + */ + handleMouseUp(event) { + super.handleMouseUp(event); + + // Stop event propagation to prevent double fire in progress-control.js + if (event) { + event.stopPropagation(); + } + this.player_.scrubbing(false); + + /** + * Trigger timeupdate because we're done seeking and the time has changed. + * This is particularly useful for if the player is paused to time the time displays. + * + * @event Tech#timeupdate + * @type {Event} + */ + this.player_.trigger({ + type: 'timeupdate', + target: this, + manuallyTriggered: true + }); + if (this.videoWasPlaying) { + silencePromise(this.player_.play()); + } else { + // We're done seeking and the time has changed. + // If the player is paused, make sure we display the correct time on the seek bar. + this.update_(); + } + } + + /** + * Move more quickly fast forward for keyboard-only users + */ + stepForward() { + this.userSeek_(this.player_.currentTime() + STEP_SECONDS); + } + + /** + * Move more quickly rewind for keyboard-only users + */ + stepBack() { + this.userSeek_(this.player_.currentTime() - STEP_SECONDS); + } + + /** + * Toggles the playback state of the player + * This gets called when enter or space is used on the seekbar + * + * @param {KeyboardEvent} event + * The `keydown` event that caused this function to be called + * + */ + handleAction(event) { + if (this.player_.paused()) { + this.player_.play(); + } else { + this.player_.pause(); + } + } + + /** + * Called when this SeekBar has focus and a key gets pressed down. + * Supports the following keys: + * + * Space or Enter key fire a click event + * Home key moves to start of the timeline + * End key moves to end of the timeline + * Digit "0" through "9" keys move to 0%, 10% ... 80%, 90% of the timeline + * PageDown key moves back a larger step than ArrowDown + * PageUp key moves forward a large step + * + * @param {KeyboardEvent} event + * The `keydown` event that caused this function to be called. + * + * @listens keydown + */ + handleKeyDown(event) { + const liveTracker = this.player_.liveTracker; + if (keycode.isEventKey(event, 'Space') || keycode.isEventKey(event, 'Enter')) { + event.preventDefault(); + event.stopPropagation(); + this.handleAction(event); + } else if (keycode.isEventKey(event, 'Home')) { + event.preventDefault(); + event.stopPropagation(); + this.userSeek_(0); + } else if (keycode.isEventKey(event, 'End')) { + event.preventDefault(); + event.stopPropagation(); + if (liveTracker && liveTracker.isLive()) { + this.userSeek_(liveTracker.liveCurrentTime()); + } else { + this.userSeek_(this.player_.duration()); + } + } else if (/^[0-9]$/.test(keycode(event))) { + event.preventDefault(); + event.stopPropagation(); + const gotoFraction = (keycode.codes[keycode(event)] - keycode.codes['0']) * 10.0 / 100.0; + if (liveTracker && liveTracker.isLive()) { + this.userSeek_(liveTracker.seekableStart() + liveTracker.liveWindow() * gotoFraction); + } else { + this.userSeek_(this.player_.duration() * gotoFraction); + } + } else if (keycode.isEventKey(event, 'PgDn')) { + event.preventDefault(); + event.stopPropagation(); + this.userSeek_(this.player_.currentTime() - STEP_SECONDS * PAGE_KEY_MULTIPLIER); + } else if (keycode.isEventKey(event, 'PgUp')) { + event.preventDefault(); + event.stopPropagation(); + this.userSeek_(this.player_.currentTime() + STEP_SECONDS * PAGE_KEY_MULTIPLIER); + } else { + // Pass keydown handling up for unsupported keys + super.handleKeyDown(event); + } + } + dispose() { + this.disableInterval_(); + this.off(this.player_, ['ended', 'durationchange', 'timeupdate'], this.update); + if (this.player_.liveTracker) { + this.off(this.player_.liveTracker, 'liveedgechange', this.update); + } + this.off(this.player_, ['playing'], this.enableIntervalHandler_); + this.off(this.player_, ['ended', 'pause', 'waiting'], this.disableIntervalHandler_); + + // we don't need to update the play progress if the document is hidden, + // also, this causes the CPU to spike and eventually crash the page on IE11. + if ('hidden' in document && 'visibilityState' in document) { + this.off(document, 'visibilitychange', this.toggleVisibility_); + } + super.dispose(); + } +} + +/** + * Default options for the `SeekBar` + * + * @type {Object} + * @private + */ +SeekBar.prototype.options_ = { + children: ['loadProgressBar', 'playProgressBar'], + barName: 'playProgressBar' +}; + +// MouseTimeDisplay tooltips should not be added to a player on mobile devices +if (!IS_IOS && !IS_ANDROID) { + SeekBar.prototype.options_.children.splice(1, 0, 'mouseTimeDisplay'); +} +Component$1.registerComponent('SeekBar', SeekBar); + +/** + * @file progress-control.js + */ + +/** + * The Progress Control component contains the seek bar, load progress, + * and play progress. + * + * @extends Component + */ +class ProgressControl extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.handleMouseMove = throttle(bind_(this, this.handleMouseMove), UPDATE_REFRESH_INTERVAL); + this.throttledHandleMouseSeek = throttle(bind_(this, this.handleMouseSeek), UPDATE_REFRESH_INTERVAL); + this.handleMouseUpHandler_ = e => this.handleMouseUp(e); + this.handleMouseDownHandler_ = e => this.handleMouseDown(e); + this.enable(); + } + + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + return super.createEl('div', { + className: 'vjs-progress-control vjs-control' + }); + } + + /** + * When the mouse moves over the `ProgressControl`, the pointer position + * gets passed down to the `MouseTimeDisplay` component. + * + * @param {Event} event + * The `mousemove` event that caused this function to run. + * + * @listen mousemove + */ + handleMouseMove(event) { + const seekBar = this.getChild('seekBar'); + if (!seekBar) { + return; + } + const playProgressBar = seekBar.getChild('playProgressBar'); + const mouseTimeDisplay = seekBar.getChild('mouseTimeDisplay'); + if (!playProgressBar && !mouseTimeDisplay) { + return; + } + const seekBarEl = seekBar.el(); + const seekBarRect = findPosition(seekBarEl); + let seekBarPoint = getPointerPosition(seekBarEl, event).x; + + // The default skin has a gap on either side of the `SeekBar`. This means + // that it's possible to trigger this behavior outside the boundaries of + // the `SeekBar`. This ensures we stay within it at all times. + seekBarPoint = clamp(seekBarPoint, 0, 1); + if (mouseTimeDisplay) { + mouseTimeDisplay.update(seekBarRect, seekBarPoint); + } + if (playProgressBar) { + playProgressBar.update(seekBarRect, seekBar.getProgress()); + } + } + + /** + * A throttled version of the {@link ProgressControl#handleMouseSeek} listener. + * + * @method ProgressControl#throttledHandleMouseSeek + * @param {Event} event + * The `mousemove` event that caused this function to run. + * + * @listen mousemove + * @listen touchmove + */ + + /** + * Handle `mousemove` or `touchmove` events on the `ProgressControl`. + * + * @param {Event} event + * `mousedown` or `touchstart` event that triggered this function + * + * @listens mousemove + * @listens touchmove + */ + handleMouseSeek(event) { + const seekBar = this.getChild('seekBar'); + if (seekBar) { + seekBar.handleMouseMove(event); + } + } + + /** + * Are controls are currently enabled for this progress control. + * + * @return {boolean} + * true if controls are enabled, false otherwise + */ + enabled() { + return this.enabled_; + } + + /** + * Disable all controls on the progress control and its children + */ + disable() { + this.children().forEach(child => child.disable && child.disable()); + if (!this.enabled()) { + return; + } + this.off(['mousedown', 'touchstart'], this.handleMouseDownHandler_); + this.off(this.el_, 'mousemove', this.handleMouseMove); + this.removeListenersAddedOnMousedownAndTouchstart(); + this.addClass('disabled'); + this.enabled_ = false; + + // Restore normal playback state if controls are disabled while scrubbing + if (this.player_.scrubbing()) { + const seekBar = this.getChild('seekBar'); + this.player_.scrubbing(false); + if (seekBar.videoWasPlaying) { + silencePromise(this.player_.play()); + } + } + } + + /** + * Enable all controls on the progress control and its children + */ + enable() { + this.children().forEach(child => child.enable && child.enable()); + if (this.enabled()) { + return; + } + this.on(['mousedown', 'touchstart'], this.handleMouseDownHandler_); + this.on(this.el_, 'mousemove', this.handleMouseMove); + this.removeClass('disabled'); + this.enabled_ = true; + } + + /** + * Cleanup listeners after the user finishes interacting with the progress controls + */ + removeListenersAddedOnMousedownAndTouchstart() { + const doc = this.el_.ownerDocument; + this.off(doc, 'mousemove', this.throttledHandleMouseSeek); + this.off(doc, 'touchmove', this.throttledHandleMouseSeek); + this.off(doc, 'mouseup', this.handleMouseUpHandler_); + this.off(doc, 'touchend', this.handleMouseUpHandler_); + } + + /** + * Handle `mousedown` or `touchstart` events on the `ProgressControl`. + * + * @param {Event} event + * `mousedown` or `touchstart` event that triggered this function + * + * @listens mousedown + * @listens touchstart + */ + handleMouseDown(event) { + const doc = this.el_.ownerDocument; + const seekBar = this.getChild('seekBar'); + if (seekBar) { + seekBar.handleMouseDown(event); + } + this.on(doc, 'mousemove', this.throttledHandleMouseSeek); + this.on(doc, 'touchmove', this.throttledHandleMouseSeek); + this.on(doc, 'mouseup', this.handleMouseUpHandler_); + this.on(doc, 'touchend', this.handleMouseUpHandler_); + } + + /** + * Handle `mouseup` or `touchend` events on the `ProgressControl`. + * + * @param {Event} event + * `mouseup` or `touchend` event that triggered this function. + * + * @listens touchend + * @listens mouseup + */ + handleMouseUp(event) { + const seekBar = this.getChild('seekBar'); + if (seekBar) { + seekBar.handleMouseUp(event); + } + this.removeListenersAddedOnMousedownAndTouchstart(); + } +} + +/** + * Default options for `ProgressControl` + * + * @type {Object} + * @private + */ +ProgressControl.prototype.options_ = { + children: ['seekBar'] +}; +Component$1.registerComponent('ProgressControl', ProgressControl); + +/** + * @file picture-in-picture-toggle.js + */ + +/** + * Toggle Picture-in-Picture mode + * + * @extends Button + */ +class PictureInPictureToggle extends Button { + /** + * Creates an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + * + * @listens Player#enterpictureinpicture + * @listens Player#leavepictureinpicture + */ + constructor(player, options) { + super(player, options); + this.setIcon('picture-in-picture-enter'); + this.on(player, ['enterpictureinpicture', 'leavepictureinpicture'], e => this.handlePictureInPictureChange(e)); + this.on(player, ['disablepictureinpicturechanged', 'loadedmetadata'], e => this.handlePictureInPictureEnabledChange(e)); + this.on(player, ['loadedmetadata', 'audioonlymodechange', 'audiopostermodechange'], () => this.handlePictureInPictureAudioModeChange()); + + // TODO: Deactivate button on player emptied event. + this.disable(); + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-picture-in-picture-control vjs-hidden ${super.buildCSSClass()}`; + } + + /** + * Displays or hides the button depending on the audio mode detection. + * Exits picture-in-picture if it is enabled when switching to audio mode. + */ + handlePictureInPictureAudioModeChange() { + // This audio detection will not detect HLS or DASH audio-only streams because there was no reliable way to detect them at the time + const isSourceAudio = this.player_.currentType().substring(0, 5) === 'audio'; + const isAudioMode = isSourceAudio || this.player_.audioPosterMode() || this.player_.audioOnlyMode(); + if (!isAudioMode) { + this.show(); + return; + } + if (this.player_.isInPictureInPicture()) { + this.player_.exitPictureInPicture(); + } + this.hide(); + } + + /** + * Enables or disables button based on availability of a Picture-In-Picture mode. + * + * Enabled if + * - `player.options().enableDocumentPictureInPicture` is true and + * window.documentPictureInPicture is available; or + * - `player.disablePictureInPicture()` is false and + * element.requestPictureInPicture is available + */ + handlePictureInPictureEnabledChange() { + if (document.pictureInPictureEnabled && this.player_.disablePictureInPicture() === false || this.player_.options_.enableDocumentPictureInPicture && 'documentPictureInPicture' in window$1) { + this.enable(); + } else { + this.disable(); + } + } + + /** + * Handles enterpictureinpicture and leavepictureinpicture on the player and change control text accordingly. + * + * @param {Event} [event] + * The {@link Player#enterpictureinpicture} or {@link Player#leavepictureinpicture} event that caused this function to be + * called. + * + * @listens Player#enterpictureinpicture + * @listens Player#leavepictureinpicture + */ + handlePictureInPictureChange(event) { + if (this.player_.isInPictureInPicture()) { + this.setIcon('picture-in-picture-exit'); + this.controlText('Exit Picture-in-Picture'); + } else { + this.setIcon('picture-in-picture-enter'); + this.controlText('Picture-in-Picture'); + } + this.handlePictureInPictureEnabledChange(); + } + + /** + * This gets called when an `PictureInPictureToggle` is "clicked". See + * {@link ClickableComponent} for more detailed information on what a click can be. + * + * @param {Event} [event] + * The `keydown`, `tap`, or `click` event that caused this function to be + * called. + * + * @listens tap + * @listens click + */ + handleClick(event) { + if (!this.player_.isInPictureInPicture()) { + this.player_.requestPictureInPicture(); + } else { + this.player_.exitPictureInPicture(); + } + } + + /** + * Show the `Component`s element if it is hidden by removing the + * 'vjs-hidden' class name from it only in browsers that support the Picture-in-Picture API. + */ + show() { + // Does not allow to display the pictureInPictureToggle in browsers that do not support the Picture-in-Picture API, e.g. Firefox. + if (typeof document.exitPictureInPicture !== 'function') { + return; + } + super.show(); + } +} + +/** + * The text that should display over the `PictureInPictureToggle`s controls. Added for localization. + * + * @type {string} + * @protected + */ +PictureInPictureToggle.prototype.controlText_ = 'Picture-in-Picture'; +Component$1.registerComponent('PictureInPictureToggle', PictureInPictureToggle); + +/** + * @file fullscreen-toggle.js + */ + +/** + * Toggle fullscreen video + * + * @extends Button + */ +class FullscreenToggle extends Button { + /** + * Creates an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.setIcon('fullscreen-enter'); + this.on(player, 'fullscreenchange', e => this.handleFullscreenChange(e)); + if (document[player.fsApi_.fullscreenEnabled] === false) { + this.disable(); + } + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-fullscreen-control ${super.buildCSSClass()}`; + } + + /** + * Handles fullscreenchange on the player and change control text accordingly. + * + * @param {Event} [event] + * The {@link Player#fullscreenchange} event that caused this function to be + * called. + * + * @listens Player#fullscreenchange + */ + handleFullscreenChange(event) { + if (this.player_.isFullscreen()) { + this.controlText('Exit Fullscreen'); + this.setIcon('fullscreen-exit'); + } else { + this.controlText('Fullscreen'); + this.setIcon('fullscreen-enter'); + } + } + + /** + * This gets called when an `FullscreenToggle` is "clicked". See + * {@link ClickableComponent} for more detailed information on what a click can be. + * + * @param {Event} [event] + * The `keydown`, `tap`, or `click` event that caused this function to be + * called. + * + * @listens tap + * @listens click + */ + handleClick(event) { + if (!this.player_.isFullscreen()) { + this.player_.requestFullscreen(); + } else { + this.player_.exitFullscreen(); + } + } +} + +/** + * The text that should display over the `FullscreenToggle`s controls. Added for localization. + * + * @type {string} + * @protected + */ +FullscreenToggle.prototype.controlText_ = 'Fullscreen'; +Component$1.registerComponent('FullscreenToggle', FullscreenToggle); + +/** + * Check if volume control is supported and if it isn't hide the + * `Component` that was passed using the `vjs-hidden` class. + * + * @param { import('../../component').default } self + * The component that should be hidden if volume is unsupported + * + * @param { import('../../player').default } player + * A reference to the player + * + * @private + */ +const checkVolumeSupport = function (self, player) { + // hide volume controls when they're not supported by the current tech + if (player.tech_ && !player.tech_.featuresVolumeControl) { + self.addClass('vjs-hidden'); + } + self.on(player, 'loadstart', function () { + if (!player.tech_.featuresVolumeControl) { + self.addClass('vjs-hidden'); + } else { + self.removeClass('vjs-hidden'); + } + }); +}; + +/** + * @file volume-level.js + */ + +/** + * Shows volume level + * + * @extends Component + */ +class VolumeLevel extends Component$1 { + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + const el = super.createEl('div', { + className: 'vjs-volume-level' + }); + this.setIcon('circle', el); + el.appendChild(super.createEl('span', { + className: 'vjs-control-text' + })); + return el; + } +} +Component$1.registerComponent('VolumeLevel', VolumeLevel); + +/** + * @file volume-level-tooltip.js + */ + +/** + * Volume level tooltips display a volume above or side by side the volume bar. + * + * @extends Component + */ +class VolumeLevelTooltip extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The {@link Player} that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.update = throttle(bind_(this, this.update), UPDATE_REFRESH_INTERVAL); + } + + /** + * Create the volume tooltip DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + return super.createEl('div', { + className: 'vjs-volume-tooltip' + }, { + 'aria-hidden': 'true' + }); + } + + /** + * Updates the position of the tooltip relative to the `VolumeBar` and + * its content text. + * + * @param {Object} rangeBarRect + * The `ClientRect` for the {@link VolumeBar} element. + * + * @param {number} rangeBarPoint + * A number from 0 to 1, representing a horizontal/vertical reference point + * from the left edge of the {@link VolumeBar} + * + * @param {boolean} vertical + * Referees to the Volume control position + * in the control bar{@link VolumeControl} + * + */ + update(rangeBarRect, rangeBarPoint, vertical, content) { + if (!vertical) { + const tooltipRect = getBoundingClientRect(this.el_); + const playerRect = getBoundingClientRect(this.player_.el()); + const volumeBarPointPx = rangeBarRect.width * rangeBarPoint; + if (!playerRect || !tooltipRect) { + return; + } + const spaceLeftOfPoint = rangeBarRect.left - playerRect.left + volumeBarPointPx; + const spaceRightOfPoint = rangeBarRect.width - volumeBarPointPx + (playerRect.right - rangeBarRect.right); + let pullTooltipBy = tooltipRect.width / 2; + if (spaceLeftOfPoint < pullTooltipBy) { + pullTooltipBy += pullTooltipBy - spaceLeftOfPoint; + } else if (spaceRightOfPoint < pullTooltipBy) { + pullTooltipBy = spaceRightOfPoint; + } + if (pullTooltipBy < 0) { + pullTooltipBy = 0; + } else if (pullTooltipBy > tooltipRect.width) { + pullTooltipBy = tooltipRect.width; + } + this.el_.style.right = `-${pullTooltipBy}px`; + } + this.write(`${content}%`); + } + + /** + * Write the volume to the tooltip DOM element. + * + * @param {string} content + * The formatted volume for the tooltip. + */ + write(content) { + textContent(this.el_, content); + } + + /** + * Updates the position of the volume tooltip relative to the `VolumeBar`. + * + * @param {Object} rangeBarRect + * The `ClientRect` for the {@link VolumeBar} element. + * + * @param {number} rangeBarPoint + * A number from 0 to 1, representing a horizontal/vertical reference point + * from the left edge of the {@link VolumeBar} + * + * @param {boolean} vertical + * Referees to the Volume control position + * in the control bar{@link VolumeControl} + * + * @param {number} volume + * The volume level to update the tooltip to + * + * @param {Function} cb + * A function that will be called during the request animation frame + * for tooltips that need to do additional animations from the default + */ + updateVolume(rangeBarRect, rangeBarPoint, vertical, volume, cb) { + this.requestNamedAnimationFrame('VolumeLevelTooltip#updateVolume', () => { + this.update(rangeBarRect, rangeBarPoint, vertical, volume.toFixed(0)); + if (cb) { + cb(); + } + }); + } +} +Component$1.registerComponent('VolumeLevelTooltip', VolumeLevelTooltip); + +/** + * @file mouse-volume-level-display.js + */ + +/** + * The {@link MouseVolumeLevelDisplay} component tracks mouse movement over the + * {@link VolumeControl}. It displays an indicator and a {@link VolumeLevelTooltip} + * indicating the volume level which is represented by a given point in the + * {@link VolumeBar}. + * + * @extends Component + */ +class MouseVolumeLevelDisplay extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The {@link Player} that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.update = throttle(bind_(this, this.update), UPDATE_REFRESH_INTERVAL); + } + + /** + * Create the DOM element for this class. + * + * @return {Element} + * The element that was created. + */ + createEl() { + return super.createEl('div', { + className: 'vjs-mouse-display' + }); + } + + /** + * Enquires updates to its own DOM as well as the DOM of its + * {@link VolumeLevelTooltip} child. + * + * @param {Object} rangeBarRect + * The `ClientRect` for the {@link VolumeBar} element. + * + * @param {number} rangeBarPoint + * A number from 0 to 1, representing a horizontal/vertical reference point + * from the left edge of the {@link VolumeBar} + * + * @param {boolean} vertical + * Referees to the Volume control position + * in the control bar{@link VolumeControl} + * + */ + update(rangeBarRect, rangeBarPoint, vertical) { + const volume = 100 * rangeBarPoint; + this.getChild('volumeLevelTooltip').updateVolume(rangeBarRect, rangeBarPoint, vertical, volume, () => { + if (vertical) { + this.el_.style.bottom = `${rangeBarRect.height * rangeBarPoint}px`; + } else { + this.el_.style.left = `${rangeBarRect.width * rangeBarPoint}px`; + } + }); + } +} + +/** + * Default options for `MouseVolumeLevelDisplay` + * + * @type {Object} + * @private + */ +MouseVolumeLevelDisplay.prototype.options_ = { + children: ['volumeLevelTooltip'] +}; +Component$1.registerComponent('MouseVolumeLevelDisplay', MouseVolumeLevelDisplay); + +/** + * @file volume-bar.js + */ + +/** + * The bar that contains the volume level and can be clicked on to adjust the level + * + * @extends Slider + */ +class VolumeBar extends Slider { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.on('slideractive', e => this.updateLastVolume_(e)); + this.on(player, 'volumechange', e => this.updateARIAAttributes(e)); + player.ready(() => this.updateARIAAttributes()); + } + + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + return super.createEl('div', { + className: 'vjs-volume-bar vjs-slider-bar' + }, { + 'aria-label': this.localize('Volume Level'), + 'aria-live': 'polite' + }); + } + + /** + * Handle mouse down on volume bar + * + * @param {Event} event + * The `mousedown` event that caused this to run. + * + * @listens mousedown + */ + handleMouseDown(event) { + if (!isSingleLeftClick(event)) { + return; + } + super.handleMouseDown(event); + } + + /** + * Handle movement events on the {@link VolumeMenuButton}. + * + * @param {Event} event + * The event that caused this function to run. + * + * @listens mousemove + */ + handleMouseMove(event) { + const mouseVolumeLevelDisplay = this.getChild('mouseVolumeLevelDisplay'); + if (mouseVolumeLevelDisplay) { + const volumeBarEl = this.el(); + const volumeBarRect = getBoundingClientRect(volumeBarEl); + const vertical = this.vertical(); + let volumeBarPoint = getPointerPosition(volumeBarEl, event); + volumeBarPoint = vertical ? volumeBarPoint.y : volumeBarPoint.x; + // The default skin has a gap on either side of the `VolumeBar`. This means + // that it's possible to trigger this behavior outside the boundaries of + // the `VolumeBar`. This ensures we stay within it at all times. + volumeBarPoint = clamp(volumeBarPoint, 0, 1); + mouseVolumeLevelDisplay.update(volumeBarRect, volumeBarPoint, vertical); + } + if (!isSingleLeftClick(event)) { + return; + } + this.checkMuted(); + this.player_.volume(this.calculateDistance(event)); + } + + /** + * If the player is muted unmute it. + */ + checkMuted() { + if (this.player_.muted()) { + this.player_.muted(false); + } + } + + /** + * Get percent of volume level + * + * @return {number} + * Volume level percent as a decimal number. + */ + getPercent() { + if (this.player_.muted()) { + return 0; + } + return this.player_.volume(); + } + + /** + * Increase volume level for keyboard users + */ + stepForward() { + this.checkMuted(); + this.player_.volume(this.player_.volume() + 0.1); + } + + /** + * Decrease volume level for keyboard users + */ + stepBack() { + this.checkMuted(); + this.player_.volume(this.player_.volume() - 0.1); + } + + /** + * Update ARIA accessibility attributes + * + * @param {Event} [event] + * The `volumechange` event that caused this function to run. + * + * @listens Player#volumechange + */ + updateARIAAttributes(event) { + const ariaValue = this.player_.muted() ? 0 : this.volumeAsPercentage_(); + this.el_.setAttribute('aria-valuenow', ariaValue); + this.el_.setAttribute('aria-valuetext', ariaValue + '%'); + } + + /** + * Returns the current value of the player volume as a percentage + * + * @private + */ + volumeAsPercentage_() { + return Math.round(this.player_.volume() * 100); + } + + /** + * When user starts dragging the VolumeBar, store the volume and listen for + * the end of the drag. When the drag ends, if the volume was set to zero, + * set lastVolume to the stored volume. + * + * @listens slideractive + * @private + */ + updateLastVolume_() { + const volumeBeforeDrag = this.player_.volume(); + this.one('sliderinactive', () => { + if (this.player_.volume() === 0) { + this.player_.lastVolume_(volumeBeforeDrag); + } + }); + } +} + +/** + * Default options for the `VolumeBar` + * + * @type {Object} + * @private + */ +VolumeBar.prototype.options_ = { + children: ['volumeLevel'], + barName: 'volumeLevel' +}; + +// MouseVolumeLevelDisplay tooltip should not be added to a player on mobile devices +if (!IS_IOS && !IS_ANDROID) { + VolumeBar.prototype.options_.children.splice(0, 0, 'mouseVolumeLevelDisplay'); +} + +/** + * Call the update event for this Slider when this event happens on the player. + * + * @type {string} + */ +VolumeBar.prototype.playerEvent = 'volumechange'; +Component$1.registerComponent('VolumeBar', VolumeBar); + +/** + * @file volume-control.js + */ + +/** + * The component for controlling the volume level + * + * @extends Component + */ +class VolumeControl extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options={}] + * The key/value store of player options. + */ + constructor(player, options = {}) { + options.vertical = options.vertical || false; + + // Pass the vertical option down to the VolumeBar if + // the VolumeBar is turned on. + if (typeof options.volumeBar === 'undefined' || isPlain(options.volumeBar)) { + options.volumeBar = options.volumeBar || {}; + options.volumeBar.vertical = options.vertical; + } + super(player, options); + + // hide this control if volume support is missing + checkVolumeSupport(this, player); + this.throttledHandleMouseMove = throttle(bind_(this, this.handleMouseMove), UPDATE_REFRESH_INTERVAL); + this.handleMouseUpHandler_ = e => this.handleMouseUp(e); + this.on('mousedown', e => this.handleMouseDown(e)); + this.on('touchstart', e => this.handleMouseDown(e)); + this.on('mousemove', e => this.handleMouseMove(e)); + + // while the slider is active (the mouse has been pressed down and + // is dragging) or in focus we do not want to hide the VolumeBar + this.on(this.volumeBar, ['focus', 'slideractive'], () => { + this.volumeBar.addClass('vjs-slider-active'); + this.addClass('vjs-slider-active'); + this.trigger('slideractive'); + }); + this.on(this.volumeBar, ['blur', 'sliderinactive'], () => { + this.volumeBar.removeClass('vjs-slider-active'); + this.removeClass('vjs-slider-active'); + this.trigger('sliderinactive'); + }); + } + + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + let orientationClass = 'vjs-volume-horizontal'; + if (this.options_.vertical) { + orientationClass = 'vjs-volume-vertical'; + } + return super.createEl('div', { + className: `vjs-volume-control vjs-control ${orientationClass}` + }); + } + + /** + * Handle `mousedown` or `touchstart` events on the `VolumeControl`. + * + * @param {Event} event + * `mousedown` or `touchstart` event that triggered this function + * + * @listens mousedown + * @listens touchstart + */ + handleMouseDown(event) { + const doc = this.el_.ownerDocument; + this.on(doc, 'mousemove', this.throttledHandleMouseMove); + this.on(doc, 'touchmove', this.throttledHandleMouseMove); + this.on(doc, 'mouseup', this.handleMouseUpHandler_); + this.on(doc, 'touchend', this.handleMouseUpHandler_); + } + + /** + * Handle `mouseup` or `touchend` events on the `VolumeControl`. + * + * @param {Event} event + * `mouseup` or `touchend` event that triggered this function. + * + * @listens touchend + * @listens mouseup + */ + handleMouseUp(event) { + const doc = this.el_.ownerDocument; + this.off(doc, 'mousemove', this.throttledHandleMouseMove); + this.off(doc, 'touchmove', this.throttledHandleMouseMove); + this.off(doc, 'mouseup', this.handleMouseUpHandler_); + this.off(doc, 'touchend', this.handleMouseUpHandler_); + } + + /** + * Handle `mousedown` or `touchstart` events on the `VolumeControl`. + * + * @param {Event} event + * `mousedown` or `touchstart` event that triggered this function + * + * @listens mousedown + * @listens touchstart + */ + handleMouseMove(event) { + this.volumeBar.handleMouseMove(event); + } +} + +/** + * Default options for the `VolumeControl` + * + * @type {Object} + * @private + */ +VolumeControl.prototype.options_ = { + children: ['volumeBar'] +}; +Component$1.registerComponent('VolumeControl', VolumeControl); + +/** + * Check if muting volume is supported and if it isn't hide the mute toggle + * button. + * + * @param { import('../../component').default } self + * A reference to the mute toggle button + * + * @param { import('../../player').default } player + * A reference to the player + * + * @private + */ +const checkMuteSupport = function (self, player) { + // hide mute toggle button if it's not supported by the current tech + if (player.tech_ && !player.tech_.featuresMuteControl) { + self.addClass('vjs-hidden'); + } + self.on(player, 'loadstart', function () { + if (!player.tech_.featuresMuteControl) { + self.addClass('vjs-hidden'); + } else { + self.removeClass('vjs-hidden'); + } + }); +}; + +/** + * @file mute-toggle.js + */ + +/** + * A button component for muting the audio. + * + * @extends Button + */ +class MuteToggle extends Button { + /** + * Creates an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + + // hide this control if volume support is missing + checkMuteSupport(this, player); + this.on(player, ['loadstart', 'volumechange'], e => this.update(e)); + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-mute-control ${super.buildCSSClass()}`; + } + + /** + * This gets called when an `MuteToggle` is "clicked". See + * {@link ClickableComponent} for more detailed information on what a click can be. + * + * @param {Event} [event] + * The `keydown`, `tap`, or `click` event that caused this function to be + * called. + * + * @listens tap + * @listens click + */ + handleClick(event) { + const vol = this.player_.volume(); + const lastVolume = this.player_.lastVolume_(); + if (vol === 0) { + const volumeToSet = lastVolume < 0.1 ? 0.1 : lastVolume; + this.player_.volume(volumeToSet); + this.player_.muted(false); + } else { + this.player_.muted(this.player_.muted() ? false : true); + } + } + + /** + * Update the `MuteToggle` button based on the state of `volume` and `muted` + * on the player. + * + * @param {Event} [event] + * The {@link Player#loadstart} event if this function was called + * through an event. + * + * @listens Player#loadstart + * @listens Player#volumechange + */ + update(event) { + this.updateIcon_(); + this.updateControlText_(); + } + + /** + * Update the appearance of the `MuteToggle` icon. + * + * Possible states (given `level` variable below): + * - 0: crossed out + * - 1: zero bars of volume + * - 2: one bar of volume + * - 3: two bars of volume + * + * @private + */ + updateIcon_() { + const vol = this.player_.volume(); + let level = 3; + this.setIcon('volume-high'); + + // in iOS when a player is loaded with muted attribute + // and volume is changed with a native mute button + // we want to make sure muted state is updated + if (IS_IOS && this.player_.tech_ && this.player_.tech_.el_) { + this.player_.muted(this.player_.tech_.el_.muted); + } + if (vol === 0 || this.player_.muted()) { + this.setIcon('volume-mute'); + level = 0; + } else if (vol < 0.33) { + this.setIcon('volume-low'); + level = 1; + } else if (vol < 0.67) { + this.setIcon('volume-medium'); + level = 2; + } + removeClass(this.el_, [0, 1, 2, 3].reduce((str, i) => str + `${i ? ' ' : ''}vjs-vol-${i}`, '')); + addClass(this.el_, `vjs-vol-${level}`); + } + + /** + * If `muted` has changed on the player, update the control text + * (`title` attribute on `vjs-mute-control` element and content of + * `vjs-control-text` element). + * + * @private + */ + updateControlText_() { + const soundOff = this.player_.muted() || this.player_.volume() === 0; + const text = soundOff ? 'Unmute' : 'Mute'; + if (this.controlText() !== text) { + this.controlText(text); + } + } +} + +/** + * The text that should display over the `MuteToggle`s controls. Added for localization. + * + * @type {string} + * @protected + */ +MuteToggle.prototype.controlText_ = 'Mute'; +Component$1.registerComponent('MuteToggle', MuteToggle); + +/** + * @file volume-control.js + */ + +/** + * A Component to contain the MuteToggle and VolumeControl so that + * they can work together. + * + * @extends Component + */ +class VolumePanel extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options={}] + * The key/value store of player options. + */ + constructor(player, options = {}) { + if (typeof options.inline !== 'undefined') { + options.inline = options.inline; + } else { + options.inline = true; + } + + // pass the inline option down to the VolumeControl as vertical if + // the VolumeControl is on. + if (typeof options.volumeControl === 'undefined' || isPlain(options.volumeControl)) { + options.volumeControl = options.volumeControl || {}; + options.volumeControl.vertical = !options.inline; + } + super(player, options); + + // this handler is used by mouse handler methods below + this.handleKeyPressHandler_ = e => this.handleKeyPress(e); + this.on(player, ['loadstart'], e => this.volumePanelState_(e)); + this.on(this.muteToggle, 'keyup', e => this.handleKeyPress(e)); + this.on(this.volumeControl, 'keyup', e => this.handleVolumeControlKeyUp(e)); + this.on('keydown', e => this.handleKeyPress(e)); + this.on('mouseover', e => this.handleMouseOver(e)); + this.on('mouseout', e => this.handleMouseOut(e)); + + // while the slider is active (the mouse has been pressed down and + // is dragging) we do not want to hide the VolumeBar + this.on(this.volumeControl, ['slideractive'], this.sliderActive_); + this.on(this.volumeControl, ['sliderinactive'], this.sliderInactive_); + } + + /** + * Add vjs-slider-active class to the VolumePanel + * + * @listens VolumeControl#slideractive + * @private + */ + sliderActive_() { + this.addClass('vjs-slider-active'); + } + + /** + * Removes vjs-slider-active class to the VolumePanel + * + * @listens VolumeControl#sliderinactive + * @private + */ + sliderInactive_() { + this.removeClass('vjs-slider-active'); + } + + /** + * Adds vjs-hidden or vjs-mute-toggle-only to the VolumePanel + * depending on MuteToggle and VolumeControl state + * + * @listens Player#loadstart + * @private + */ + volumePanelState_() { + // hide volume panel if neither volume control or mute toggle + // are displayed + if (this.volumeControl.hasClass('vjs-hidden') && this.muteToggle.hasClass('vjs-hidden')) { + this.addClass('vjs-hidden'); + } + + // if only mute toggle is visible we don't want + // volume panel expanding when hovered or active + if (this.volumeControl.hasClass('vjs-hidden') && !this.muteToggle.hasClass('vjs-hidden')) { + this.addClass('vjs-mute-toggle-only'); + } + } + + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + let orientationClass = 'vjs-volume-panel-horizontal'; + if (!this.options_.inline) { + orientationClass = 'vjs-volume-panel-vertical'; + } + return super.createEl('div', { + className: `vjs-volume-panel vjs-control ${orientationClass}` + }); + } + + /** + * Dispose of the `volume-panel` and all child components. + */ + dispose() { + this.handleMouseOut(); + super.dispose(); + } + + /** + * Handles `keyup` events on the `VolumeControl`, looking for ESC, which closes + * the volume panel and sets focus on `MuteToggle`. + * + * @param {Event} event + * The `keyup` event that caused this function to be called. + * + * @listens keyup + */ + handleVolumeControlKeyUp(event) { + if (keycode.isEventKey(event, 'Esc')) { + this.muteToggle.focus(); + } + } + + /** + * This gets called when a `VolumePanel` gains hover via a `mouseover` event. + * Turns on listening for `mouseover` event. When they happen it + * calls `this.handleMouseOver`. + * + * @param {Event} event + * The `mouseover` event that caused this function to be called. + * + * @listens mouseover + */ + handleMouseOver(event) { + this.addClass('vjs-hover'); + on(document, 'keyup', this.handleKeyPressHandler_); + } + + /** + * This gets called when a `VolumePanel` gains hover via a `mouseout` event. + * Turns on listening for `mouseout` event. When they happen it + * calls `this.handleMouseOut`. + * + * @param {Event} event + * The `mouseout` event that caused this function to be called. + * + * @listens mouseout + */ + handleMouseOut(event) { + this.removeClass('vjs-hover'); + off(document, 'keyup', this.handleKeyPressHandler_); + } + + /** + * Handles `keyup` event on the document or `keydown` event on the `VolumePanel`, + * looking for ESC, which hides the `VolumeControl`. + * + * @param {Event} event + * The keypress that triggered this event. + * + * @listens keydown | keyup + */ + handleKeyPress(event) { + if (keycode.isEventKey(event, 'Esc')) { + this.handleMouseOut(); + } + } +} + +/** + * Default options for the `VolumeControl` + * + * @type {Object} + * @private + */ +VolumePanel.prototype.options_ = { + children: ['muteToggle', 'volumeControl'] +}; +Component$1.registerComponent('VolumePanel', VolumePanel); + +/** + * Button to skip forward a configurable amount of time + * through a video. Renders in the control bar. + * + * e.g. options: {controlBar: {skipButtons: forward: 5}} + * + * @extends Button + */ +class SkipForward extends Button { + constructor(player, options) { + super(player, options); + this.validOptions = [5, 10, 30]; + this.skipTime = this.getSkipForwardTime(); + if (this.skipTime && this.validOptions.includes(this.skipTime)) { + this.setIcon(`forward-${this.skipTime}`); + this.controlText(this.localize('Skip forward {1} seconds', [this.skipTime.toLocaleString(player.language())])); + this.show(); + } else { + this.hide(); + } + } + getSkipForwardTime() { + const playerOptions = this.options_.playerOptions; + return playerOptions.controlBar && playerOptions.controlBar.skipButtons && playerOptions.controlBar.skipButtons.forward; + } + buildCSSClass() { + return `vjs-skip-forward-${this.getSkipForwardTime()} ${super.buildCSSClass()}`; + } + + /** + * On click, skips forward in the duration/seekable range by a configurable amount of seconds. + * If the time left in the duration/seekable range is less than the configured 'skip forward' time, + * skips to end of duration/seekable range. + * + * Handle a click on a `SkipForward` button + * + * @param {EventTarget~Event} event + * The `click` event that caused this function + * to be called + */ + handleClick(event) { + if (isNaN(this.player_.duration())) { + return; + } + const currentVideoTime = this.player_.currentTime(); + const liveTracker = this.player_.liveTracker; + const duration = liveTracker && liveTracker.isLive() ? liveTracker.seekableEnd() : this.player_.duration(); + let newTime; + if (currentVideoTime + this.skipTime <= duration) { + newTime = currentVideoTime + this.skipTime; + } else { + newTime = duration; + } + this.player_.currentTime(newTime); + } + + /** + * Update control text on languagechange + */ + handleLanguagechange() { + this.controlText(this.localize('Skip forward {1} seconds', [this.skipTime])); + } +} +SkipForward.prototype.controlText_ = 'Skip Forward'; +Component$1.registerComponent('SkipForward', SkipForward); + +/** + * Button to skip backward a configurable amount of time + * through a video. Renders in the control bar. + * + * * e.g. options: {controlBar: {skipButtons: backward: 5}} + * + * @extends Button + */ +class SkipBackward extends Button { + constructor(player, options) { + super(player, options); + this.validOptions = [5, 10, 30]; + this.skipTime = this.getSkipBackwardTime(); + if (this.skipTime && this.validOptions.includes(this.skipTime)) { + this.setIcon(`replay-${this.skipTime}`); + this.controlText(this.localize('Skip backward {1} seconds', [this.skipTime.toLocaleString(player.language())])); + this.show(); + } else { + this.hide(); + } + } + getSkipBackwardTime() { + const playerOptions = this.options_.playerOptions; + return playerOptions.controlBar && playerOptions.controlBar.skipButtons && playerOptions.controlBar.skipButtons.backward; + } + buildCSSClass() { + return `vjs-skip-backward-${this.getSkipBackwardTime()} ${super.buildCSSClass()}`; + } + + /** + * On click, skips backward in the video by a configurable amount of seconds. + * If the current time in the video is less than the configured 'skip backward' time, + * skips to beginning of video or seekable range. + * + * Handle a click on a `SkipBackward` button + * + * @param {EventTarget~Event} event + * The `click` event that caused this function + * to be called + */ + handleClick(event) { + const currentVideoTime = this.player_.currentTime(); + const liveTracker = this.player_.liveTracker; + const seekableStart = liveTracker && liveTracker.isLive() && liveTracker.seekableStart(); + let newTime; + if (seekableStart && currentVideoTime - this.skipTime <= seekableStart) { + newTime = seekableStart; + } else if (currentVideoTime >= this.skipTime) { + newTime = currentVideoTime - this.skipTime; + } else { + newTime = 0; + } + this.player_.currentTime(newTime); + } + + /** + * Update control text on languagechange + */ + handleLanguagechange() { + this.controlText(this.localize('Skip backward {1} seconds', [this.skipTime])); + } +} +SkipBackward.prototype.controlText_ = 'Skip Backward'; +Component$1.registerComponent('SkipBackward', SkipBackward); + +/** + * @file menu.js + */ + +/** + * The Menu component is used to build popup menus, including subtitle and + * captions selection menus. + * + * @extends Component + */ +class Menu extends Component$1 { + /** + * Create an instance of this class. + * + * @param { import('../player').default } player + * the player that this component should attach to + * + * @param {Object} [options] + * Object of option names and values + * + */ + constructor(player, options) { + super(player, options); + if (options) { + this.menuButton_ = options.menuButton; + } + this.focusedChild_ = -1; + this.on('keydown', e => this.handleKeyDown(e)); + + // All the menu item instances share the same blur handler provided by the menu container. + this.boundHandleBlur_ = e => this.handleBlur(e); + this.boundHandleTapClick_ = e => this.handleTapClick(e); + } + + /** + * Add event listeners to the {@link MenuItem}. + * + * @param {Object} component + * The instance of the `MenuItem` to add listeners to. + * + */ + addEventListenerForItem(component) { + if (!(component instanceof Component$1)) { + return; + } + this.on(component, 'blur', this.boundHandleBlur_); + this.on(component, ['tap', 'click'], this.boundHandleTapClick_); + } + + /** + * Remove event listeners from the {@link MenuItem}. + * + * @param {Object} component + * The instance of the `MenuItem` to remove listeners. + * + */ + removeEventListenerForItem(component) { + if (!(component instanceof Component$1)) { + return; + } + this.off(component, 'blur', this.boundHandleBlur_); + this.off(component, ['tap', 'click'], this.boundHandleTapClick_); + } + + /** + * This method will be called indirectly when the component has been added + * before the component adds to the new menu instance by `addItem`. + * In this case, the original menu instance will remove the component + * by calling `removeChild`. + * + * @param {Object} component + * The instance of the `MenuItem` + */ + removeChild(component) { + if (typeof component === 'string') { + component = this.getChild(component); + } + this.removeEventListenerForItem(component); + super.removeChild(component); + } + + /** + * Add a {@link MenuItem} to the menu. + * + * @param {Object|string} component + * The name or instance of the `MenuItem` to add. + * + */ + addItem(component) { + const childComponent = this.addChild(component); + if (childComponent) { + this.addEventListenerForItem(childComponent); + } + } + + /** + * Create the `Menu`s DOM element. + * + * @return {Element} + * the element that was created + */ + createEl() { + const contentElType = this.options_.contentElType || 'ul'; + this.contentEl_ = createEl(contentElType, { + className: 'vjs-menu-content' + }); + this.contentEl_.setAttribute('role', 'menu'); + const el = super.createEl('div', { + append: this.contentEl_, + className: 'vjs-menu' + }); + el.appendChild(this.contentEl_); + + // Prevent clicks from bubbling up. Needed for Menu Buttons, + // where a click on the parent is significant + on(el, 'click', function (event) { + event.preventDefault(); + event.stopImmediatePropagation(); + }); + return el; + } + dispose() { + this.contentEl_ = null; + this.boundHandleBlur_ = null; + this.boundHandleTapClick_ = null; + super.dispose(); + } + + /** + * Called when a `MenuItem` loses focus. + * + * @param {Event} event + * The `blur` event that caused this function to be called. + * + * @listens blur + */ + handleBlur(event) { + const relatedTarget = event.relatedTarget || document.activeElement; + + // Close menu popup when a user clicks outside the menu + if (!this.children().some(element => { + return element.el() === relatedTarget; + })) { + const btn = this.menuButton_; + if (btn && btn.buttonPressed_ && relatedTarget !== btn.el().firstChild) { + btn.unpressButton(); + } + } + } + + /** + * Called when a `MenuItem` gets clicked or tapped. + * + * @param {Event} event + * The `click` or `tap` event that caused this function to be called. + * + * @listens click,tap + */ + handleTapClick(event) { + // Unpress the associated MenuButton, and move focus back to it + if (this.menuButton_) { + this.menuButton_.unpressButton(); + const childComponents = this.children(); + if (!Array.isArray(childComponents)) { + return; + } + const foundComponent = childComponents.filter(component => component.el() === event.target)[0]; + if (!foundComponent) { + return; + } + + // don't focus menu button if item is a caption settings item + // because focus will move elsewhere + if (foundComponent.name() !== 'CaptionSettingsMenuItem') { + this.menuButton_.focus(); + } + } + } + + /** + * Handle a `keydown` event on this menu. This listener is added in the constructor. + * + * @param {KeyboardEvent} event + * A `keydown` event that happened on the menu. + * + * @listens keydown + */ + handleKeyDown(event) { + // Left and Down Arrows + if (keycode.isEventKey(event, 'Left') || keycode.isEventKey(event, 'Down')) { + event.preventDefault(); + event.stopPropagation(); + this.stepForward(); + + // Up and Right Arrows + } else if (keycode.isEventKey(event, 'Right') || keycode.isEventKey(event, 'Up')) { + event.preventDefault(); + event.stopPropagation(); + this.stepBack(); + } + } + + /** + * Move to next (lower) menu item for keyboard users. + */ + stepForward() { + let stepChild = 0; + if (this.focusedChild_ !== undefined) { + stepChild = this.focusedChild_ + 1; + } + this.focus(stepChild); + } + + /** + * Move to previous (higher) menu item for keyboard users. + */ + stepBack() { + let stepChild = 0; + if (this.focusedChild_ !== undefined) { + stepChild = this.focusedChild_ - 1; + } + this.focus(stepChild); + } + + /** + * Set focus on a {@link MenuItem} in the `Menu`. + * + * @param {Object|string} [item=0] + * Index of child item set focus on. + */ + focus(item = 0) { + const children = this.children().slice(); + const haveTitle = children.length && children[0].hasClass('vjs-menu-title'); + if (haveTitle) { + children.shift(); + } + if (children.length > 0) { + if (item < 0) { + item = 0; + } else if (item >= children.length) { + item = children.length - 1; + } + this.focusedChild_ = item; + children[item].el_.focus(); + } + } +} +Component$1.registerComponent('Menu', Menu); + +/** + * @file menu-button.js + */ + +/** + * A `MenuButton` class for any popup {@link Menu}. + * + * @extends Component + */ +class MenuButton extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options={}] + * The key/value store of player options. + */ + constructor(player, options = {}) { + super(player, options); + this.menuButton_ = new Button(player, options); + this.menuButton_.controlText(this.controlText_); + this.menuButton_.el_.setAttribute('aria-haspopup', 'true'); + + // Add buildCSSClass values to the button, not the wrapper + const buttonClass = Button.prototype.buildCSSClass(); + this.menuButton_.el_.className = this.buildCSSClass() + ' ' + buttonClass; + this.menuButton_.removeClass('vjs-control'); + this.addChild(this.menuButton_); + this.update(); + this.enabled_ = true; + const handleClick = e => this.handleClick(e); + this.handleMenuKeyUp_ = e => this.handleMenuKeyUp(e); + this.on(this.menuButton_, 'tap', handleClick); + this.on(this.menuButton_, 'click', handleClick); + this.on(this.menuButton_, 'keydown', e => this.handleKeyDown(e)); + this.on(this.menuButton_, 'mouseenter', () => { + this.addClass('vjs-hover'); + this.menu.show(); + on(document, 'keyup', this.handleMenuKeyUp_); + }); + this.on('mouseleave', e => this.handleMouseLeave(e)); + this.on('keydown', e => this.handleSubmenuKeyDown(e)); + } + + /** + * Update the menu based on the current state of its items. + */ + update() { + const menu = this.createMenu(); + if (this.menu) { + this.menu.dispose(); + this.removeChild(this.menu); + } + this.menu = menu; + this.addChild(menu); + + /** + * Track the state of the menu button + * + * @type {Boolean} + * @private + */ + this.buttonPressed_ = false; + this.menuButton_.el_.setAttribute('aria-expanded', 'false'); + if (this.items && this.items.length <= this.hideThreshold_) { + this.hide(); + this.menu.contentEl_.removeAttribute('role'); + } else { + this.show(); + this.menu.contentEl_.setAttribute('role', 'menu'); + } + } + + /** + * Create the menu and add all items to it. + * + * @return {Menu} + * The constructed menu + */ + createMenu() { + const menu = new Menu(this.player_, { + menuButton: this + }); + + /** + * Hide the menu if the number of items is less than or equal to this threshold. This defaults + * to 0 and whenever we add items which can be hidden to the menu we'll increment it. We list + * it here because every time we run `createMenu` we need to reset the value. + * + * @protected + * @type {Number} + */ + this.hideThreshold_ = 0; + + // Add a title list item to the top + if (this.options_.title) { + const titleEl = createEl('li', { + className: 'vjs-menu-title', + textContent: toTitleCase$1(this.options_.title), + tabIndex: -1 + }); + const titleComponent = new Component$1(this.player_, { + el: titleEl + }); + menu.addItem(titleComponent); + } + this.items = this.createItems(); + if (this.items) { + // Add menu items to the menu + for (let i = 0; i < this.items.length; i++) { + menu.addItem(this.items[i]); + } + } + return menu; + } + + /** + * Create the list of menu items. Specific to each subclass. + * + * @abstract + */ + createItems() {} + + /** + * Create the `MenuButtons`s DOM element. + * + * @return {Element} + * The element that gets created. + */ + createEl() { + return super.createEl('div', { + className: this.buildWrapperCSSClass() + }, {}); + } + + /** + * Overwrites the `setIcon` method from `Component`. + * In this case, we want the icon to be appended to the menuButton. + * + * @param {string} name + * The icon name to be added. + */ + setIcon(name) { + super.setIcon(name, this.menuButton_.el_); + } + + /** + * Allow sub components to stack CSS class names for the wrapper element + * + * @return {string} + * The constructed wrapper DOM `className` + */ + buildWrapperCSSClass() { + let menuButtonClass = 'vjs-menu-button'; + + // If the inline option is passed, we want to use different styles altogether. + if (this.options_.inline === true) { + menuButtonClass += '-inline'; + } else { + menuButtonClass += '-popup'; + } + + // TODO: Fix the CSS so that this isn't necessary + const buttonClass = Button.prototype.buildCSSClass(); + return `vjs-menu-button ${menuButtonClass} ${buttonClass} ${super.buildCSSClass()}`; + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + let menuButtonClass = 'vjs-menu-button'; + + // If the inline option is passed, we want to use different styles altogether. + if (this.options_.inline === true) { + menuButtonClass += '-inline'; + } else { + menuButtonClass += '-popup'; + } + return `vjs-menu-button ${menuButtonClass} ${super.buildCSSClass()}`; + } + + /** + * Get or set the localized control text that will be used for accessibility. + * + * > NOTE: This will come from the internal `menuButton_` element. + * + * @param {string} [text] + * Control text for element. + * + * @param {Element} [el=this.menuButton_.el()] + * Element to set the title on. + * + * @return {string} + * - The control text when getting + */ + controlText(text, el = this.menuButton_.el()) { + return this.menuButton_.controlText(text, el); + } + + /** + * Dispose of the `menu-button` and all child components. + */ + dispose() { + this.handleMouseLeave(); + super.dispose(); + } + + /** + * Handle a click on a `MenuButton`. + * See {@link ClickableComponent#handleClick} for instances where this is called. + * + * @param {Event} event + * The `keydown`, `tap`, or `click` event that caused this function to be + * called. + * + * @listens tap + * @listens click + */ + handleClick(event) { + if (this.buttonPressed_) { + this.unpressButton(); + } else { + this.pressButton(); + } + } + + /** + * Handle `mouseleave` for `MenuButton`. + * + * @param {Event} event + * The `mouseleave` event that caused this function to be called. + * + * @listens mouseleave + */ + handleMouseLeave(event) { + this.removeClass('vjs-hover'); + off(document, 'keyup', this.handleMenuKeyUp_); + } + + /** + * Set the focus to the actual button, not to this element + */ + focus() { + this.menuButton_.focus(); + } + + /** + * Remove the focus from the actual button, not this element + */ + blur() { + this.menuButton_.blur(); + } + + /** + * Handle tab, escape, down arrow, and up arrow keys for `MenuButton`. See + * {@link ClickableComponent#handleKeyDown} for instances where this is called. + * + * @param {Event} event + * The `keydown` event that caused this function to be called. + * + * @listens keydown + */ + handleKeyDown(event) { + // Escape or Tab unpress the 'button' + if (keycode.isEventKey(event, 'Esc') || keycode.isEventKey(event, 'Tab')) { + if (this.buttonPressed_) { + this.unpressButton(); + } + + // Don't preventDefault for Tab key - we still want to lose focus + if (!keycode.isEventKey(event, 'Tab')) { + event.preventDefault(); + // Set focus back to the menu button's button + this.menuButton_.focus(); + } + // Up Arrow or Down Arrow also 'press' the button to open the menu + } else if (keycode.isEventKey(event, 'Up') || keycode.isEventKey(event, 'Down')) { + if (!this.buttonPressed_) { + event.preventDefault(); + this.pressButton(); + } + } + } + + /** + * Handle a `keyup` event on a `MenuButton`. The listener for this is added in + * the constructor. + * + * @param {Event} event + * Key press event + * + * @listens keyup + */ + handleMenuKeyUp(event) { + // Escape hides popup menu + if (keycode.isEventKey(event, 'Esc') || keycode.isEventKey(event, 'Tab')) { + this.removeClass('vjs-hover'); + } + } + + /** + * This method name now delegates to `handleSubmenuKeyDown`. This means + * anyone calling `handleSubmenuKeyPress` will not see their method calls + * stop working. + * + * @param {Event} event + * The event that caused this function to be called. + */ + handleSubmenuKeyPress(event) { + this.handleSubmenuKeyDown(event); + } + + /** + * Handle a `keydown` event on a sub-menu. The listener for this is added in + * the constructor. + * + * @param {Event} event + * Key press event + * + * @listens keydown + */ + handleSubmenuKeyDown(event) { + // Escape or Tab unpress the 'button' + if (keycode.isEventKey(event, 'Esc') || keycode.isEventKey(event, 'Tab')) { + if (this.buttonPressed_) { + this.unpressButton(); + } + // Don't preventDefault for Tab key - we still want to lose focus + if (!keycode.isEventKey(event, 'Tab')) { + event.preventDefault(); + // Set focus back to the menu button's button + this.menuButton_.focus(); + } + } + } + + /** + * Put the current `MenuButton` into a pressed state. + */ + pressButton() { + if (this.enabled_) { + this.buttonPressed_ = true; + this.menu.show(); + this.menu.lockShowing(); + this.menuButton_.el_.setAttribute('aria-expanded', 'true'); + + // set the focus into the submenu, except on iOS where it is resulting in + // undesired scrolling behavior when the player is in an iframe + if (IS_IOS && isInFrame()) { + // Return early so that the menu isn't focused + return; + } + this.menu.focus(); + } + } + + /** + * Take the current `MenuButton` out of a pressed state. + */ + unpressButton() { + if (this.enabled_) { + this.buttonPressed_ = false; + this.menu.unlockShowing(); + this.menu.hide(); + this.menuButton_.el_.setAttribute('aria-expanded', 'false'); + } + } + + /** + * Disable the `MenuButton`. Don't allow it to be clicked. + */ + disable() { + this.unpressButton(); + this.enabled_ = false; + this.addClass('vjs-disabled'); + this.menuButton_.disable(); + } + + /** + * Enable the `MenuButton`. Allow it to be clicked. + */ + enable() { + this.enabled_ = true; + this.removeClass('vjs-disabled'); + this.menuButton_.enable(); + } +} +Component$1.registerComponent('MenuButton', MenuButton); + +/** + * @file track-button.js + */ + +/** + * The base class for buttons that toggle specific track types (e.g. subtitles). + * + * @extends MenuButton + */ +class TrackButton extends MenuButton { + /** + * Creates an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + const tracks = options.tracks; + super(player, options); + if (this.items.length <= 1) { + this.hide(); + } + if (!tracks) { + return; + } + const updateHandler = bind_(this, this.update); + tracks.addEventListener('removetrack', updateHandler); + tracks.addEventListener('addtrack', updateHandler); + tracks.addEventListener('labelchange', updateHandler); + this.player_.on('ready', updateHandler); + this.player_.on('dispose', function () { + tracks.removeEventListener('removetrack', updateHandler); + tracks.removeEventListener('addtrack', updateHandler); + tracks.removeEventListener('labelchange', updateHandler); + }); + } +} +Component$1.registerComponent('TrackButton', TrackButton); + +/** + * @file menu-keys.js + */ + +/** + * All keys used for operation of a menu (`MenuButton`, `Menu`, and `MenuItem`) + * Note that 'Enter' and 'Space' are not included here (otherwise they would + * prevent the `MenuButton` and `MenuItem` from being keyboard-clickable) + * + * @typedef MenuKeys + * @array + */ +const MenuKeys = ['Tab', 'Esc', 'Up', 'Down', 'Right', 'Left']; + +/** + * @file menu-item.js + */ + +/** + * The component for a menu item. `<li>` + * + * @extends ClickableComponent + */ +class MenuItem extends ClickableComponent { + /** + * Creates an instance of the this class. + * + * @param { import('../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options={}] + * The key/value store of player options. + * + */ + constructor(player, options) { + super(player, options); + this.selectable = options.selectable; + this.isSelected_ = options.selected || false; + this.multiSelectable = options.multiSelectable; + this.selected(this.isSelected_); + if (this.selectable) { + if (this.multiSelectable) { + this.el_.setAttribute('role', 'menuitemcheckbox'); + } else { + this.el_.setAttribute('role', 'menuitemradio'); + } + } else { + this.el_.setAttribute('role', 'menuitem'); + } + } + + /** + * Create the `MenuItem's DOM element + * + * @param {string} [type=li] + * Element's node type, not actually used, always set to `li`. + * + * @param {Object} [props={}] + * An object of properties that should be set on the element + * + * @param {Object} [attrs={}] + * An object of attributes that should be set on the element + * + * @return {Element} + * The element that gets created. + */ + createEl(type, props, attrs) { + // The control is textual, not just an icon + this.nonIconControl = true; + const el = super.createEl('li', Object.assign({ + className: 'vjs-menu-item', + tabIndex: -1 + }, props), attrs); + + // swap icon with menu item text. + const menuItemEl = createEl('span', { + className: 'vjs-menu-item-text', + textContent: this.localize(this.options_.label) + }); + + // If using SVG icons, the element with vjs-icon-placeholder will be added separately. + if (this.player_.options_.experimentalSvgIcons) { + el.appendChild(menuItemEl); + } else { + el.replaceChild(menuItemEl, el.querySelector('.vjs-icon-placeholder')); + } + return el; + } + + /** + * Ignore keys which are used by the menu, but pass any other ones up. See + * {@link ClickableComponent#handleKeyDown} for instances where this is called. + * + * @param {KeyboardEvent} event + * The `keydown` event that caused this function to be called. + * + * @listens keydown + */ + handleKeyDown(event) { + if (!MenuKeys.some(key => keycode.isEventKey(event, key))) { + // Pass keydown handling up for unused keys + super.handleKeyDown(event); + } + } + + /** + * Any click on a `MenuItem` puts it into the selected state. + * See {@link ClickableComponent#handleClick} for instances where this is called. + * + * @param {Event} event + * The `keydown`, `tap`, or `click` event that caused this function to be + * called. + * + * @listens tap + * @listens click + */ + handleClick(event) { + this.selected(true); + } + + /** + * Set the state for this menu item as selected or not. + * + * @param {boolean} selected + * if the menu item is selected or not + */ + selected(selected) { + if (this.selectable) { + if (selected) { + this.addClass('vjs-selected'); + this.el_.setAttribute('aria-checked', 'true'); + // aria-checked isn't fully supported by browsers/screen readers, + // so indicate selected state to screen reader in the control text. + this.controlText(', selected'); + this.isSelected_ = true; + } else { + this.removeClass('vjs-selected'); + this.el_.setAttribute('aria-checked', 'false'); + // Indicate un-selected state to screen reader + this.controlText(''); + this.isSelected_ = false; + } + } + } +} +Component$1.registerComponent('MenuItem', MenuItem); + +/** + * @file text-track-menu-item.js + */ + +/** + * The specific menu item type for selecting a language within a text track kind + * + * @extends MenuItem + */ +class TextTrackMenuItem extends MenuItem { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + const track = options.track; + const tracks = player.textTracks(); + + // Modify options for parent MenuItem class's init. + options.label = track.label || track.language || 'Unknown'; + options.selected = track.mode === 'showing'; + super(player, options); + this.track = track; + // Determine the relevant kind(s) of tracks for this component and filter + // out empty kinds. + this.kinds = (options.kinds || [options.kind || this.track.kind]).filter(Boolean); + const changeHandler = (...args) => { + this.handleTracksChange.apply(this, args); + }; + const selectedLanguageChangeHandler = (...args) => { + this.handleSelectedLanguageChange.apply(this, args); + }; + player.on(['loadstart', 'texttrackchange'], changeHandler); + tracks.addEventListener('change', changeHandler); + tracks.addEventListener('selectedlanguagechange', selectedLanguageChangeHandler); + this.on('dispose', function () { + player.off(['loadstart', 'texttrackchange'], changeHandler); + tracks.removeEventListener('change', changeHandler); + tracks.removeEventListener('selectedlanguagechange', selectedLanguageChangeHandler); + }); + + // iOS7 doesn't dispatch change events to TextTrackLists when an + // associated track's mode changes. Without something like + // Object.observe() (also not present on iOS7), it's not + // possible to detect changes to the mode attribute and polyfill + // the change event. As a poor substitute, we manually dispatch + // change events whenever the controls modify the mode. + if (tracks.onchange === undefined) { + let event; + this.on(['tap', 'click'], function () { + if (typeof window$1.Event !== 'object') { + // Android 2.3 throws an Illegal Constructor error for window.Event + try { + event = new window$1.Event('change'); + } catch (err) { + // continue regardless of error + } + } + if (!event) { + event = document.createEvent('Event'); + event.initEvent('change', true, true); + } + tracks.dispatchEvent(event); + }); + } + + // set the default state based on current tracks + this.handleTracksChange(); + } + + /** + * This gets called when an `TextTrackMenuItem` is "clicked". See + * {@link ClickableComponent} for more detailed information on what a click can be. + * + * @param {Event} event + * The `keydown`, `tap`, or `click` event that caused this function to be + * called. + * + * @listens tap + * @listens click + */ + handleClick(event) { + const referenceTrack = this.track; + const tracks = this.player_.textTracks(); + super.handleClick(event); + if (!tracks) { + return; + } + for (let i = 0; i < tracks.length; i++) { + const track = tracks[i]; + + // If the track from the text tracks list is not of the right kind, + // skip it. We do not want to affect tracks of incompatible kind(s). + if (this.kinds.indexOf(track.kind) === -1) { + continue; + } + + // If this text track is the component's track and it is not showing, + // set it to showing. + if (track === referenceTrack) { + if (track.mode !== 'showing') { + track.mode = 'showing'; + } + + // If this text track is not the component's track and it is not + // disabled, set it to disabled. + } else if (track.mode !== 'disabled') { + track.mode = 'disabled'; + } + } + } + + /** + * Handle text track list change + * + * @param {Event} event + * The `change` event that caused this function to be called. + * + * @listens TextTrackList#change + */ + handleTracksChange(event) { + const shouldBeSelected = this.track.mode === 'showing'; + + // Prevent redundant selected() calls because they may cause + // screen readers to read the appended control text unnecessarily + if (shouldBeSelected !== this.isSelected_) { + this.selected(shouldBeSelected); + } + } + handleSelectedLanguageChange(event) { + if (this.track.mode === 'showing') { + const selectedLanguage = this.player_.cache_.selectedLanguage; + + // Don't replace the kind of track across the same language + if (selectedLanguage && selectedLanguage.enabled && selectedLanguage.language === this.track.language && selectedLanguage.kind !== this.track.kind) { + return; + } + this.player_.cache_.selectedLanguage = { + enabled: true, + language: this.track.language, + kind: this.track.kind + }; + } + } + dispose() { + // remove reference to track object on dispose + this.track = null; + super.dispose(); + } +} +Component$1.registerComponent('TextTrackMenuItem', TextTrackMenuItem); + +/** + * @file off-text-track-menu-item.js + */ + +/** + * A special menu item for turning off a specific type of text track + * + * @extends TextTrackMenuItem + */ +class OffTextTrackMenuItem extends TextTrackMenuItem { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + // Create pseudo track info + // Requires options['kind'] + options.track = { + player, + // it is no longer necessary to store `kind` or `kinds` on the track itself + // since they are now stored in the `kinds` property of all instances of + // TextTrackMenuItem, but this will remain for backwards compatibility + kind: options.kind, + kinds: options.kinds, + default: false, + mode: 'disabled' + }; + if (!options.kinds) { + options.kinds = [options.kind]; + } + if (options.label) { + options.track.label = options.label; + } else { + options.track.label = options.kinds.join(' and ') + ' off'; + } + + // MenuItem is selectable + options.selectable = true; + // MenuItem is NOT multiSelectable (i.e. only one can be marked "selected" at a time) + options.multiSelectable = false; + super(player, options); + } + + /** + * Handle text track change + * + * @param {Event} event + * The event that caused this function to run + */ + handleTracksChange(event) { + const tracks = this.player().textTracks(); + let shouldBeSelected = true; + for (let i = 0, l = tracks.length; i < l; i++) { + const track = tracks[i]; + if (this.options_.kinds.indexOf(track.kind) > -1 && track.mode === 'showing') { + shouldBeSelected = false; + break; + } + } + + // Prevent redundant selected() calls because they may cause + // screen readers to read the appended control text unnecessarily + if (shouldBeSelected !== this.isSelected_) { + this.selected(shouldBeSelected); + } + } + handleSelectedLanguageChange(event) { + const tracks = this.player().textTracks(); + let allHidden = true; + for (let i = 0, l = tracks.length; i < l; i++) { + const track = tracks[i]; + if (['captions', 'descriptions', 'subtitles'].indexOf(track.kind) > -1 && track.mode === 'showing') { + allHidden = false; + break; + } + } + if (allHidden) { + this.player_.cache_.selectedLanguage = { + enabled: false + }; + } + } + + /** + * Update control text and label on languagechange + */ + handleLanguagechange() { + this.$('.vjs-menu-item-text').textContent = this.player_.localize(this.options_.label); + super.handleLanguagechange(); + } +} +Component$1.registerComponent('OffTextTrackMenuItem', OffTextTrackMenuItem); + +/** + * @file text-track-button.js + */ + +/** + * The base class for buttons that toggle specific text track types (e.g. subtitles) + * + * @extends MenuButton + */ +class TextTrackButton extends TrackButton { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options={}] + * The key/value store of player options. + */ + constructor(player, options = {}) { + options.tracks = player.textTracks(); + super(player, options); + } + + /** + * Create a menu item for each text track + * + * @param {TextTrackMenuItem[]} [items=[]] + * Existing array of items to use during creation + * + * @return {TextTrackMenuItem[]} + * Array of menu items that were created + */ + createItems(items = [], TrackMenuItem = TextTrackMenuItem) { + // Label is an override for the [track] off label + // USed to localise captions/subtitles + let label; + if (this.label_) { + label = `${this.label_} off`; + } + // Add an OFF menu item to turn all tracks off + items.push(new OffTextTrackMenuItem(this.player_, { + kinds: this.kinds_, + kind: this.kind_, + label + })); + this.hideThreshold_ += 1; + const tracks = this.player_.textTracks(); + if (!Array.isArray(this.kinds_)) { + this.kinds_ = [this.kind_]; + } + for (let i = 0; i < tracks.length; i++) { + const track = tracks[i]; + + // only add tracks that are of an appropriate kind and have a label + if (this.kinds_.indexOf(track.kind) > -1) { + const item = new TrackMenuItem(this.player_, { + track, + kinds: this.kinds_, + kind: this.kind_, + // MenuItem is selectable + selectable: true, + // MenuItem is NOT multiSelectable (i.e. only one can be marked "selected" at a time) + multiSelectable: false + }); + item.addClass(`vjs-${track.kind}-menu-item`); + items.push(item); + } + } + return items; + } +} +Component$1.registerComponent('TextTrackButton', TextTrackButton); + +/** + * @file chapters-track-menu-item.js + */ + +/** + * The chapter track menu item + * + * @extends MenuItem + */ +class ChaptersTrackMenuItem extends MenuItem { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + const track = options.track; + const cue = options.cue; + const currentTime = player.currentTime(); + + // Modify options for parent MenuItem class's init. + options.selectable = true; + options.multiSelectable = false; + options.label = cue.text; + options.selected = cue.startTime <= currentTime && currentTime < cue.endTime; + super(player, options); + this.track = track; + this.cue = cue; + } + + /** + * This gets called when an `ChaptersTrackMenuItem` is "clicked". See + * {@link ClickableComponent} for more detailed information on what a click can be. + * + * @param {Event} [event] + * The `keydown`, `tap`, or `click` event that caused this function to be + * called. + * + * @listens tap + * @listens click + */ + handleClick(event) { + super.handleClick(); + this.player_.currentTime(this.cue.startTime); + } +} +Component$1.registerComponent('ChaptersTrackMenuItem', ChaptersTrackMenuItem); + +/** + * @file chapters-button.js + */ + +/** + * The button component for toggling and selecting chapters + * Chapters act much differently than other text tracks + * Cues are navigation vs. other tracks of alternative languages + * + * @extends TextTrackButton + */ +class ChaptersButton extends TextTrackButton { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + * + * @param {Function} [ready] + * The function to call when this function is ready. + */ + constructor(player, options, ready) { + super(player, options, ready); + this.setIcon('chapters'); + this.selectCurrentItem_ = () => { + this.items.forEach(item => { + item.selected(this.track_.activeCues[0] === item.cue); + }); + }; + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-chapters-button ${super.buildCSSClass()}`; + } + buildWrapperCSSClass() { + return `vjs-chapters-button ${super.buildWrapperCSSClass()}`; + } + + /** + * Update the menu based on the current state of its items. + * + * @param {Event} [event] + * An event that triggered this function to run. + * + * @listens TextTrackList#addtrack + * @listens TextTrackList#removetrack + * @listens TextTrackList#change + */ + update(event) { + if (event && event.track && event.track.kind !== 'chapters') { + return; + } + const track = this.findChaptersTrack(); + if (track !== this.track_) { + this.setTrack(track); + super.update(); + } else if (!this.items || track && track.cues && track.cues.length !== this.items.length) { + // Update the menu initially or if the number of cues has changed since set + super.update(); + } + } + + /** + * Set the currently selected track for the chapters button. + * + * @param {TextTrack} track + * The new track to select. Nothing will change if this is the currently selected + * track. + */ + setTrack(track) { + if (this.track_ === track) { + return; + } + if (!this.updateHandler_) { + this.updateHandler_ = this.update.bind(this); + } + + // here this.track_ refers to the old track instance + if (this.track_) { + const remoteTextTrackEl = this.player_.remoteTextTrackEls().getTrackElementByTrack_(this.track_); + if (remoteTextTrackEl) { + remoteTextTrackEl.removeEventListener('load', this.updateHandler_); + } + this.track_.removeEventListener('cuechange', this.selectCurrentItem_); + this.track_ = null; + } + this.track_ = track; + + // here this.track_ refers to the new track instance + if (this.track_) { + this.track_.mode = 'hidden'; + const remoteTextTrackEl = this.player_.remoteTextTrackEls().getTrackElementByTrack_(this.track_); + if (remoteTextTrackEl) { + remoteTextTrackEl.addEventListener('load', this.updateHandler_); + } + this.track_.addEventListener('cuechange', this.selectCurrentItem_); + } + } + + /** + * Find the track object that is currently in use by this ChaptersButton + * + * @return {TextTrack|undefined} + * The current track or undefined if none was found. + */ + findChaptersTrack() { + const tracks = this.player_.textTracks() || []; + for (let i = tracks.length - 1; i >= 0; i--) { + // We will always choose the last track as our chaptersTrack + const track = tracks[i]; + if (track.kind === this.kind_) { + return track; + } + } + } + + /** + * Get the caption for the ChaptersButton based on the track label. This will also + * use the current tracks localized kind as a fallback if a label does not exist. + * + * @return {string} + * The tracks current label or the localized track kind. + */ + getMenuCaption() { + if (this.track_ && this.track_.label) { + return this.track_.label; + } + return this.localize(toTitleCase$1(this.kind_)); + } + + /** + * Create menu from chapter track + * + * @return { import('../../menu/menu').default } + * New menu for the chapter buttons + */ + createMenu() { + this.options_.title = this.getMenuCaption(); + return super.createMenu(); + } + + /** + * Create a menu item for each text track + * + * @return { import('./text-track-menu-item').default[] } + * Array of menu items + */ + createItems() { + const items = []; + if (!this.track_) { + return items; + } + const cues = this.track_.cues; + if (!cues) { + return items; + } + for (let i = 0, l = cues.length; i < l; i++) { + const cue = cues[i]; + const mi = new ChaptersTrackMenuItem(this.player_, { + track: this.track_, + cue + }); + items.push(mi); + } + return items; + } +} + +/** + * `kind` of TextTrack to look for to associate it with this menu. + * + * @type {string} + * @private + */ +ChaptersButton.prototype.kind_ = 'chapters'; + +/** + * The text that should display over the `ChaptersButton`s controls. Added for localization. + * + * @type {string} + * @protected + */ +ChaptersButton.prototype.controlText_ = 'Chapters'; +Component$1.registerComponent('ChaptersButton', ChaptersButton); + +/** + * @file descriptions-button.js + */ + +/** + * The button component for toggling and selecting descriptions + * + * @extends TextTrackButton + */ +class DescriptionsButton extends TextTrackButton { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + * + * @param {Function} [ready] + * The function to call when this component is ready. + */ + constructor(player, options, ready) { + super(player, options, ready); + this.setIcon('audio-description'); + const tracks = player.textTracks(); + const changeHandler = bind_(this, this.handleTracksChange); + tracks.addEventListener('change', changeHandler); + this.on('dispose', function () { + tracks.removeEventListener('change', changeHandler); + }); + } + + /** + * Handle text track change + * + * @param {Event} event + * The event that caused this function to run + * + * @listens TextTrackList#change + */ + handleTracksChange(event) { + const tracks = this.player().textTracks(); + let disabled = false; + + // Check whether a track of a different kind is showing + for (let i = 0, l = tracks.length; i < l; i++) { + const track = tracks[i]; + if (track.kind !== this.kind_ && track.mode === 'showing') { + disabled = true; + break; + } + } + + // If another track is showing, disable this menu button + if (disabled) { + this.disable(); + } else { + this.enable(); + } + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-descriptions-button ${super.buildCSSClass()}`; + } + buildWrapperCSSClass() { + return `vjs-descriptions-button ${super.buildWrapperCSSClass()}`; + } +} + +/** + * `kind` of TextTrack to look for to associate it with this menu. + * + * @type {string} + * @private + */ +DescriptionsButton.prototype.kind_ = 'descriptions'; + +/** + * The text that should display over the `DescriptionsButton`s controls. Added for localization. + * + * @type {string} + * @protected + */ +DescriptionsButton.prototype.controlText_ = 'Descriptions'; +Component$1.registerComponent('DescriptionsButton', DescriptionsButton); + +/** + * @file subtitles-button.js + */ + +/** + * The button component for toggling and selecting subtitles + * + * @extends TextTrackButton + */ +class SubtitlesButton extends TextTrackButton { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + * + * @param {Function} [ready] + * The function to call when this component is ready. + */ + constructor(player, options, ready) { + super(player, options, ready); + this.setIcon('subtitles'); + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-subtitles-button ${super.buildCSSClass()}`; + } + buildWrapperCSSClass() { + return `vjs-subtitles-button ${super.buildWrapperCSSClass()}`; + } +} + +/** + * `kind` of TextTrack to look for to associate it with this menu. + * + * @type {string} + * @private + */ +SubtitlesButton.prototype.kind_ = 'subtitles'; + +/** + * The text that should display over the `SubtitlesButton`s controls. Added for localization. + * + * @type {string} + * @protected + */ +SubtitlesButton.prototype.controlText_ = 'Subtitles'; +Component$1.registerComponent('SubtitlesButton', SubtitlesButton); + +/** + * @file caption-settings-menu-item.js + */ + +/** + * The menu item for caption track settings menu + * + * @extends TextTrackMenuItem + */ +class CaptionSettingsMenuItem extends TextTrackMenuItem { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + options.track = { + player, + kind: options.kind, + label: options.kind + ' settings', + selectable: false, + default: false, + mode: 'disabled' + }; + + // CaptionSettingsMenuItem has no concept of 'selected' + options.selectable = false; + options.name = 'CaptionSettingsMenuItem'; + super(player, options); + this.addClass('vjs-texttrack-settings'); + this.controlText(', opens ' + options.kind + ' settings dialog'); + } + + /** + * This gets called when an `CaptionSettingsMenuItem` is "clicked". See + * {@link ClickableComponent} for more detailed information on what a click can be. + * + * @param {Event} [event] + * The `keydown`, `tap`, or `click` event that caused this function to be + * called. + * + * @listens tap + * @listens click + */ + handleClick(event) { + this.player().getChild('textTrackSettings').open(); + } + + /** + * Update control text and label on languagechange + */ + handleLanguagechange() { + this.$('.vjs-menu-item-text').textContent = this.player_.localize(this.options_.kind + ' settings'); + super.handleLanguagechange(); + } +} +Component$1.registerComponent('CaptionSettingsMenuItem', CaptionSettingsMenuItem); + +/** + * @file captions-button.js + */ + +/** + * The button component for toggling and selecting captions + * + * @extends TextTrackButton + */ +class CaptionsButton extends TextTrackButton { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + * + * @param {Function} [ready] + * The function to call when this component is ready. + */ + constructor(player, options, ready) { + super(player, options, ready); + this.setIcon('captions'); + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-captions-button ${super.buildCSSClass()}`; + } + buildWrapperCSSClass() { + return `vjs-captions-button ${super.buildWrapperCSSClass()}`; + } + + /** + * Create caption menu items + * + * @return {CaptionSettingsMenuItem[]} + * The array of current menu items. + */ + createItems() { + const items = []; + if (!(this.player().tech_ && this.player().tech_.featuresNativeTextTracks) && this.player().getChild('textTrackSettings')) { + items.push(new CaptionSettingsMenuItem(this.player_, { + kind: this.kind_ + })); + this.hideThreshold_ += 1; + } + return super.createItems(items); + } +} + +/** + * `kind` of TextTrack to look for to associate it with this menu. + * + * @type {string} + * @private + */ +CaptionsButton.prototype.kind_ = 'captions'; + +/** + * The text that should display over the `CaptionsButton`s controls. Added for localization. + * + * @type {string} + * @protected + */ +CaptionsButton.prototype.controlText_ = 'Captions'; +Component$1.registerComponent('CaptionsButton', CaptionsButton); + +/** + * @file subs-caps-menu-item.js + */ + +/** + * SubsCapsMenuItem has an [cc] icon to distinguish captions from subtitles + * in the SubsCapsMenu. + * + * @extends TextTrackMenuItem + */ +class SubsCapsMenuItem extends TextTrackMenuItem { + createEl(type, props, attrs) { + const el = super.createEl(type, props, attrs); + const parentSpan = el.querySelector('.vjs-menu-item-text'); + if (this.options_.track.kind === 'captions') { + if (this.player_.options_.experimentalSvgIcons) { + this.setIcon('captions', el); + } else { + parentSpan.appendChild(createEl('span', { + className: 'vjs-icon-placeholder' + }, { + 'aria-hidden': true + })); + } + parentSpan.appendChild(createEl('span', { + className: 'vjs-control-text', + // space added as the text will visually flow with the + // label + textContent: ` ${this.localize('Captions')}` + })); + } + return el; + } +} +Component$1.registerComponent('SubsCapsMenuItem', SubsCapsMenuItem); + +/** + * @file sub-caps-button.js + */ + +/** + * The button component for toggling and selecting captions and/or subtitles + * + * @extends TextTrackButton + */ +class SubsCapsButton extends TextTrackButton { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + * + * @param {Function} [ready] + * The function to call when this component is ready. + */ + constructor(player, options = {}) { + super(player, options); + + // Although North America uses "captions" in most cases for + // "captions and subtitles" other locales use "subtitles" + this.label_ = 'subtitles'; + this.setIcon('subtitles'); + if (['en', 'en-us', 'en-ca', 'fr-ca'].indexOf(this.player_.language_) > -1) { + this.label_ = 'captions'; + this.setIcon('captions'); + } + this.menuButton_.controlText(toTitleCase$1(this.label_)); + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-subs-caps-button ${super.buildCSSClass()}`; + } + buildWrapperCSSClass() { + return `vjs-subs-caps-button ${super.buildWrapperCSSClass()}`; + } + + /** + * Create caption/subtitles menu items + * + * @return {CaptionSettingsMenuItem[]} + * The array of current menu items. + */ + createItems() { + let items = []; + if (!(this.player().tech_ && this.player().tech_.featuresNativeTextTracks) && this.player().getChild('textTrackSettings')) { + items.push(new CaptionSettingsMenuItem(this.player_, { + kind: this.label_ + })); + this.hideThreshold_ += 1; + } + items = super.createItems(items, SubsCapsMenuItem); + return items; + } +} + +/** + * `kind`s of TextTrack to look for to associate it with this menu. + * + * @type {array} + * @private + */ +SubsCapsButton.prototype.kinds_ = ['captions', 'subtitles']; + +/** + * The text that should display over the `SubsCapsButton`s controls. + * + * + * @type {string} + * @protected + */ +SubsCapsButton.prototype.controlText_ = 'Subtitles'; +Component$1.registerComponent('SubsCapsButton', SubsCapsButton); + +/** + * @file audio-track-menu-item.js + */ + +/** + * An {@link AudioTrack} {@link MenuItem} + * + * @extends MenuItem + */ +class AudioTrackMenuItem extends MenuItem { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + const track = options.track; + const tracks = player.audioTracks(); + + // Modify options for parent MenuItem class's init. + options.label = track.label || track.language || 'Unknown'; + options.selected = track.enabled; + super(player, options); + this.track = track; + this.addClass(`vjs-${track.kind}-menu-item`); + const changeHandler = (...args) => { + this.handleTracksChange.apply(this, args); + }; + tracks.addEventListener('change', changeHandler); + this.on('dispose', () => { + tracks.removeEventListener('change', changeHandler); + }); + } + createEl(type, props, attrs) { + const el = super.createEl(type, props, attrs); + const parentSpan = el.querySelector('.vjs-menu-item-text'); + if (['main-desc', 'description'].indexOf(this.options_.track.kind) >= 0) { + parentSpan.appendChild(createEl('span', { + className: 'vjs-icon-placeholder' + }, { + 'aria-hidden': true + })); + parentSpan.appendChild(createEl('span', { + className: 'vjs-control-text', + textContent: ' ' + this.localize('Descriptions') + })); + } + return el; + } + + /** + * This gets called when an `AudioTrackMenuItem is "clicked". See {@link ClickableComponent} + * for more detailed information on what a click can be. + * + * @param {Event} [event] + * The `keydown`, `tap`, or `click` event that caused this function to be + * called. + * + * @listens tap + * @listens click + */ + handleClick(event) { + super.handleClick(event); + + // the audio track list will automatically toggle other tracks + // off for us. + this.track.enabled = true; + + // when native audio tracks are used, we want to make sure that other tracks are turned off + if (this.player_.tech_.featuresNativeAudioTracks) { + const tracks = this.player_.audioTracks(); + for (let i = 0; i < tracks.length; i++) { + const track = tracks[i]; + + // skip the current track since we enabled it above + if (track === this.track) { + continue; + } + track.enabled = track === this.track; + } + } + } + + /** + * Handle any {@link AudioTrack} change. + * + * @param {Event} [event] + * The {@link AudioTrackList#change} event that caused this to run. + * + * @listens AudioTrackList#change + */ + handleTracksChange(event) { + this.selected(this.track.enabled); + } +} +Component$1.registerComponent('AudioTrackMenuItem', AudioTrackMenuItem); + +/** + * @file audio-track-button.js + */ + +/** + * The base class for buttons that toggle specific {@link AudioTrack} types. + * + * @extends TrackButton + */ +class AudioTrackButton extends TrackButton { + /** + * Creates an instance of this class. + * + * @param {Player} player + * The `Player` that this class should be attached to. + * + * @param {Object} [options={}] + * The key/value store of player options. + */ + constructor(player, options = {}) { + options.tracks = player.audioTracks(); + super(player, options); + this.setIcon('audio'); + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-audio-button ${super.buildCSSClass()}`; + } + buildWrapperCSSClass() { + return `vjs-audio-button ${super.buildWrapperCSSClass()}`; + } + + /** + * Create a menu item for each audio track + * + * @param {AudioTrackMenuItem[]} [items=[]] + * An array of existing menu items to use. + * + * @return {AudioTrackMenuItem[]} + * An array of menu items + */ + createItems(items = []) { + // if there's only one audio track, there no point in showing it + this.hideThreshold_ = 1; + const tracks = this.player_.audioTracks(); + for (let i = 0; i < tracks.length; i++) { + const track = tracks[i]; + items.push(new AudioTrackMenuItem(this.player_, { + track, + // MenuItem is selectable + selectable: true, + // MenuItem is NOT multiSelectable (i.e. only one can be marked "selected" at a time) + multiSelectable: false + })); + } + return items; + } +} + +/** + * The text that should display over the `AudioTrackButton`s controls. Added for localization. + * + * @type {string} + * @protected + */ +AudioTrackButton.prototype.controlText_ = 'Audio Track'; +Component$1.registerComponent('AudioTrackButton', AudioTrackButton); + +/** + * @file playback-rate-menu-item.js + */ + +/** + * The specific menu item type for selecting a playback rate. + * + * @extends MenuItem + */ +class PlaybackRateMenuItem extends MenuItem { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + const label = options.rate; + const rate = parseFloat(label, 10); + + // Modify options for parent MenuItem class's init. + options.label = label; + options.selected = rate === player.playbackRate(); + options.selectable = true; + options.multiSelectable = false; + super(player, options); + this.label = label; + this.rate = rate; + this.on(player, 'ratechange', e => this.update(e)); + } + + /** + * This gets called when an `PlaybackRateMenuItem` is "clicked". See + * {@link ClickableComponent} for more detailed information on what a click can be. + * + * @param {Event} [event] + * The `keydown`, `tap`, or `click` event that caused this function to be + * called. + * + * @listens tap + * @listens click + */ + handleClick(event) { + super.handleClick(); + this.player().playbackRate(this.rate); + } + + /** + * Update the PlaybackRateMenuItem when the playbackrate changes. + * + * @param {Event} [event] + * The `ratechange` event that caused this function to run. + * + * @listens Player#ratechange + */ + update(event) { + this.selected(this.player().playbackRate() === this.rate); + } +} + +/** + * The text that should display over the `PlaybackRateMenuItem`s controls. Added for localization. + * + * @type {string} + * @private + */ +PlaybackRateMenuItem.prototype.contentElType = 'button'; +Component$1.registerComponent('PlaybackRateMenuItem', PlaybackRateMenuItem); + +/** + * @file playback-rate-menu-button.js + */ + +/** + * The component for controlling the playback rate. + * + * @extends MenuButton + */ +class PlaybackRateMenuButton extends MenuButton { + /** + * Creates an instance of this class. + * + * @param { import('../../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.menuButton_.el_.setAttribute('aria-describedby', this.labelElId_); + this.updateVisibility(); + this.updateLabel(); + this.on(player, 'loadstart', e => this.updateVisibility(e)); + this.on(player, 'ratechange', e => this.updateLabel(e)); + this.on(player, 'playbackrateschange', e => this.handlePlaybackRateschange(e)); + } + + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + const el = super.createEl(); + this.labelElId_ = 'vjs-playback-rate-value-label-' + this.id_; + this.labelEl_ = createEl('div', { + className: 'vjs-playback-rate-value', + id: this.labelElId_, + textContent: '1x' + }); + el.appendChild(this.labelEl_); + return el; + } + dispose() { + this.labelEl_ = null; + super.dispose(); + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-playback-rate ${super.buildCSSClass()}`; + } + buildWrapperCSSClass() { + return `vjs-playback-rate ${super.buildWrapperCSSClass()}`; + } + + /** + * Create the list of menu items. Specific to each subclass. + * + */ + createItems() { + const rates = this.playbackRates(); + const items = []; + for (let i = rates.length - 1; i >= 0; i--) { + items.push(new PlaybackRateMenuItem(this.player(), { + rate: rates[i] + 'x' + })); + } + return items; + } + + /** + * On playbackrateschange, update the menu to account for the new items. + * + * @listens Player#playbackrateschange + */ + handlePlaybackRateschange(event) { + this.update(); + } + + /** + * Get possible playback rates + * + * @return {Array} + * All possible playback rates + */ + playbackRates() { + const player = this.player(); + return player.playbackRates && player.playbackRates() || []; + } + + /** + * Get whether playback rates is supported by the tech + * and an array of playback rates exists + * + * @return {boolean} + * Whether changing playback rate is supported + */ + playbackRateSupported() { + return this.player().tech_ && this.player().tech_.featuresPlaybackRate && this.playbackRates() && this.playbackRates().length > 0; + } + + /** + * Hide playback rate controls when they're no playback rate options to select + * + * @param {Event} [event] + * The event that caused this function to run. + * + * @listens Player#loadstart + */ + updateVisibility(event) { + if (this.playbackRateSupported()) { + this.removeClass('vjs-hidden'); + } else { + this.addClass('vjs-hidden'); + } + } + + /** + * Update button label when rate changed + * + * @param {Event} [event] + * The event that caused this function to run. + * + * @listens Player#ratechange + */ + updateLabel(event) { + if (this.playbackRateSupported()) { + this.labelEl_.textContent = this.player().playbackRate() + 'x'; + } + } +} + +/** + * The text that should display over the `PlaybackRateMenuButton`s controls. + * + * Added for localization. + * + * @type {string} + * @protected + */ +PlaybackRateMenuButton.prototype.controlText_ = 'Playback Rate'; +Component$1.registerComponent('PlaybackRateMenuButton', PlaybackRateMenuButton); + +/** + * @file spacer.js + */ + +/** + * Just an empty spacer element that can be used as an append point for plugins, etc. + * Also can be used to create space between elements when necessary. + * + * @extends Component + */ +class Spacer extends Component$1 { + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-spacer ${super.buildCSSClass()}`; + } + + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl(tag = 'div', props = {}, attributes = {}) { + if (!props.className) { + props.className = this.buildCSSClass(); + } + return super.createEl(tag, props, attributes); + } +} +Component$1.registerComponent('Spacer', Spacer); + +/** + * @file custom-control-spacer.js + */ + +/** + * Spacer specifically meant to be used as an insertion point for new plugins, etc. + * + * @extends Spacer + */ +class CustomControlSpacer extends Spacer { + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + */ + buildCSSClass() { + return `vjs-custom-control-spacer ${super.buildCSSClass()}`; + } + + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + return super.createEl('div', { + className: this.buildCSSClass(), + // No-flex/table-cell mode requires there be some content + // in the cell to fill the remaining space of the table. + textContent: '\u00a0' + }); + } +} +Component$1.registerComponent('CustomControlSpacer', CustomControlSpacer); + +/** + * @file control-bar.js + */ + +/** + * Container of main controls. + * + * @extends Component + */ +class ControlBar extends Component$1 { + /** + * Create the `Component`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + return super.createEl('div', { + className: 'vjs-control-bar', + dir: 'ltr' + }); + } +} + +/** + * Default options for `ControlBar` + * + * @type {Object} + * @private + */ +ControlBar.prototype.options_ = { + children: ['playToggle', 'skipBackward', 'skipForward', 'volumePanel', 'currentTimeDisplay', 'timeDivider', 'durationDisplay', 'progressControl', 'liveDisplay', 'seekToLive', 'remainingTimeDisplay', 'customControlSpacer', 'playbackRateMenuButton', 'chaptersButton', 'descriptionsButton', 'subsCapsButton', 'audioTrackButton', 'pictureInPictureToggle', 'fullscreenToggle'] +}; +Component$1.registerComponent('ControlBar', ControlBar); + +/** + * @file error-display.js + */ + +/** + * A display that indicates an error has occurred. This means that the video + * is unplayable. + * + * @extends ModalDialog + */ +class ErrorDisplay extends ModalDialog { + /** + * Creates an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + super(player, options); + this.on(player, 'error', e => { + this.open(e); + }); + } + + /** + * Builds the default DOM `className`. + * + * @return {string} + * The DOM `className` for this object. + * + * @deprecated Since version 5. + */ + buildCSSClass() { + return `vjs-error-display ${super.buildCSSClass()}`; + } + + /** + * Gets the localized error message based on the `Player`s error. + * + * @return {string} + * The `Player`s error message localized or an empty string. + */ + content() { + const error = this.player().error(); + return error ? this.localize(error.message) : ''; + } +} + +/** + * The default options for an `ErrorDisplay`. + * + * @private + */ +ErrorDisplay.prototype.options_ = Object.assign({}, ModalDialog.prototype.options_, { + pauseOnOpen: false, + fillAlways: true, + temporary: false, + uncloseable: true +}); +Component$1.registerComponent('ErrorDisplay', ErrorDisplay); + +/** + * @file text-track-settings.js + */ +const LOCAL_STORAGE_KEY$1 = 'vjs-text-track-settings'; +const COLOR_BLACK = ['#000', 'Black']; +const COLOR_BLUE = ['#00F', 'Blue']; +const COLOR_CYAN = ['#0FF', 'Cyan']; +const COLOR_GREEN = ['#0F0', 'Green']; +const COLOR_MAGENTA = ['#F0F', 'Magenta']; +const COLOR_RED = ['#F00', 'Red']; +const COLOR_WHITE = ['#FFF', 'White']; +const COLOR_YELLOW = ['#FF0', 'Yellow']; +const OPACITY_OPAQUE = ['1', 'Opaque']; +const OPACITY_SEMI = ['0.5', 'Semi-Transparent']; +const OPACITY_TRANS = ['0', 'Transparent']; + +// Configuration for the various <select> elements in the DOM of this component. +// +// Possible keys include: +// +// `default`: +// The default option index. Only needs to be provided if not zero. +// `parser`: +// A function which is used to parse the value from the selected option in +// a customized way. +// `selector`: +// The selector used to find the associated <select> element. +const selectConfigs = { + backgroundColor: { + selector: '.vjs-bg-color > select', + id: 'captions-background-color-%s', + label: 'Color', + options: [COLOR_BLACK, COLOR_WHITE, COLOR_RED, COLOR_GREEN, COLOR_BLUE, COLOR_YELLOW, COLOR_MAGENTA, COLOR_CYAN] + }, + backgroundOpacity: { + selector: '.vjs-bg-opacity > select', + id: 'captions-background-opacity-%s', + label: 'Opacity', + options: [OPACITY_OPAQUE, OPACITY_SEMI, OPACITY_TRANS] + }, + color: { + selector: '.vjs-text-color > select', + id: 'captions-foreground-color-%s', + label: 'Color', + options: [COLOR_WHITE, COLOR_BLACK, COLOR_RED, COLOR_GREEN, COLOR_BLUE, COLOR_YELLOW, COLOR_MAGENTA, COLOR_CYAN] + }, + edgeStyle: { + selector: '.vjs-edge-style > select', + id: '%s', + label: 'Text Edge Style', + options: [['none', 'None'], ['raised', 'Raised'], ['depressed', 'Depressed'], ['uniform', 'Uniform'], ['dropshadow', 'Drop shadow']] + }, + fontFamily: { + selector: '.vjs-font-family > select', + id: 'captions-font-family-%s', + label: 'Font Family', + options: [['proportionalSansSerif', 'Proportional Sans-Serif'], ['monospaceSansSerif', 'Monospace Sans-Serif'], ['proportionalSerif', 'Proportional Serif'], ['monospaceSerif', 'Monospace Serif'], ['casual', 'Casual'], ['script', 'Script'], ['small-caps', 'Small Caps']] + }, + fontPercent: { + selector: '.vjs-font-percent > select', + id: 'captions-font-size-%s', + label: 'Font Size', + options: [['0.50', '50%'], ['0.75', '75%'], ['1.00', '100%'], ['1.25', '125%'], ['1.50', '150%'], ['1.75', '175%'], ['2.00', '200%'], ['3.00', '300%'], ['4.00', '400%']], + default: 2, + parser: v => v === '1.00' ? null : Number(v) + }, + textOpacity: { + selector: '.vjs-text-opacity > select', + id: 'captions-foreground-opacity-%s', + label: 'Opacity', + options: [OPACITY_OPAQUE, OPACITY_SEMI] + }, + // Options for this object are defined below. + windowColor: { + selector: '.vjs-window-color > select', + id: 'captions-window-color-%s', + label: 'Color' + }, + // Options for this object are defined below. + windowOpacity: { + selector: '.vjs-window-opacity > select', + id: 'captions-window-opacity-%s', + label: 'Opacity', + options: [OPACITY_TRANS, OPACITY_SEMI, OPACITY_OPAQUE] + } +}; +selectConfigs.windowColor.options = selectConfigs.backgroundColor.options; + +/** + * Get the actual value of an option. + * + * @param {string} value + * The value to get + * + * @param {Function} [parser] + * Optional function to adjust the value. + * + * @return {*} + * - Will be `undefined` if no value exists + * - Will be `undefined` if the given value is "none". + * - Will be the actual value otherwise. + * + * @private + */ +function parseOptionValue(value, parser) { + if (parser) { + value = parser(value); + } + if (value && value !== 'none') { + return value; + } +} + +/** + * Gets the value of the selected <option> element within a <select> element. + * + * @param {Element} el + * the element to look in + * + * @param {Function} [parser] + * Optional function to adjust the value. + * + * @return {*} + * - Will be `undefined` if no value exists + * - Will be `undefined` if the given value is "none". + * - Will be the actual value otherwise. + * + * @private + */ +function getSelectedOptionValue(el, parser) { + const value = el.options[el.options.selectedIndex].value; + return parseOptionValue(value, parser); +} + +/** + * Sets the selected <option> element within a <select> element based on a + * given value. + * + * @param {Element} el + * The element to look in. + * + * @param {string} value + * the property to look on. + * + * @param {Function} [parser] + * Optional function to adjust the value before comparing. + * + * @private + */ +function setSelectedOption(el, value, parser) { + if (!value) { + return; + } + for (let i = 0; i < el.options.length; i++) { + if (parseOptionValue(el.options[i].value, parser) === value) { + el.selectedIndex = i; + break; + } + } +} + +/** + * Manipulate Text Tracks settings. + * + * @extends ModalDialog + */ +class TextTrackSettings extends ModalDialog { + /** + * Creates an instance of this class. + * + * @param { import('../player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + */ + constructor(player, options) { + options.temporary = false; + super(player, options); + this.updateDisplay = this.updateDisplay.bind(this); + + // fill the modal and pretend we have opened it + this.fill(); + this.hasBeenOpened_ = this.hasBeenFilled_ = true; + this.endDialog = createEl('p', { + className: 'vjs-control-text', + textContent: this.localize('End of dialog window.') + }); + this.el().appendChild(this.endDialog); + this.setDefaults(); + + // Grab `persistTextTrackSettings` from the player options if not passed in child options + if (options.persistTextTrackSettings === undefined) { + this.options_.persistTextTrackSettings = this.options_.playerOptions.persistTextTrackSettings; + } + this.on(this.$('.vjs-done-button'), 'click', () => { + this.saveSettings(); + this.close(); + }); + this.on(this.$('.vjs-default-button'), 'click', () => { + this.setDefaults(); + this.updateDisplay(); + }); + each(selectConfigs, config => { + this.on(this.$(config.selector), 'change', this.updateDisplay); + }); + if (this.options_.persistTextTrackSettings) { + this.restoreSettings(); + } + } + dispose() { + this.endDialog = null; + super.dispose(); + } + + /** + * Create a <select> element with configured options. + * + * @param {string} key + * Configuration key to use during creation. + * + * @param {string} [legendId] + * Id of associated <legend>. + * + * @param {string} [type=label] + * Type of labelling element, `label` or `legend` + * + * @return {string} + * An HTML string. + * + * @private + */ + createElSelect_(key, legendId = '', type = 'label') { + const config = selectConfigs[key]; + const id = config.id.replace('%s', this.id_); + const selectLabelledbyIds = [legendId, id].join(' ').trim(); + const guid = `vjs_select_${newGUID()}`; + return [`<${type} id="${id}"${type === 'label' ? ` for="${guid}" class="vjs-label"` : ''}>`, this.localize(config.label), `</${type}>`, `<select aria-labelledby="${selectLabelledbyIds}" id="${guid}">`].concat(config.options.map(o => { + const optionId = id + '-' + o[1].replace(/\W+/g, ''); + return [`<option id="${optionId}" value="${o[0]}" `, `aria-labelledby="${selectLabelledbyIds} ${optionId}">`, this.localize(o[1]), '</option>'].join(''); + })).concat('</select>').join(''); + } + + /** + * Create foreground color element for the component + * + * @return {string} + * An HTML string. + * + * @private + */ + createElFgColor_() { + const legendId = `captions-text-legend-${this.id_}`; + return ['<fieldset class="vjs-fg vjs-track-setting">', `<legend id="${legendId}">`, this.localize('Text'), '</legend>', '<span class="vjs-text-color">', this.createElSelect_('color', legendId), '</span>', '<span class="vjs-text-opacity vjs-opacity">', this.createElSelect_('textOpacity', legendId), '</span>', '</fieldset>'].join(''); + } + + /** + * Create background color element for the component + * + * @return {string} + * An HTML string. + * + * @private + */ + createElBgColor_() { + const legendId = `captions-background-${this.id_}`; + return ['<fieldset class="vjs-bg vjs-track-setting">', `<legend id="${legendId}">`, this.localize('Text Background'), '</legend>', '<span class="vjs-bg-color">', this.createElSelect_('backgroundColor', legendId), '</span>', '<span class="vjs-bg-opacity vjs-opacity">', this.createElSelect_('backgroundOpacity', legendId), '</span>', '</fieldset>'].join(''); + } + + /** + * Create window color element for the component + * + * @return {string} + * An HTML string. + * + * @private + */ + createElWinColor_() { + const legendId = `captions-window-${this.id_}`; + return ['<fieldset class="vjs-window vjs-track-setting">', `<legend id="${legendId}">`, this.localize('Caption Area Background'), '</legend>', '<span class="vjs-window-color">', this.createElSelect_('windowColor', legendId), '</span>', '<span class="vjs-window-opacity vjs-opacity">', this.createElSelect_('windowOpacity', legendId), '</span>', '</fieldset>'].join(''); + } + + /** + * Create color elements for the component + * + * @return {Element} + * The element that was created + * + * @private + */ + createElColors_() { + return createEl('div', { + className: 'vjs-track-settings-colors', + innerHTML: [this.createElFgColor_(), this.createElBgColor_(), this.createElWinColor_()].join('') + }); + } + + /** + * Create font elements for the component + * + * @return {Element} + * The element that was created. + * + * @private + */ + createElFont_() { + return createEl('div', { + className: 'vjs-track-settings-font', + innerHTML: ['<fieldset class="vjs-font-percent vjs-track-setting">', this.createElSelect_('fontPercent', '', 'legend'), '</fieldset>', '<fieldset class="vjs-edge-style vjs-track-setting">', this.createElSelect_('edgeStyle', '', 'legend'), '</fieldset>', '<fieldset class="vjs-font-family vjs-track-setting">', this.createElSelect_('fontFamily', '', 'legend'), '</fieldset>'].join('') + }); + } + + /** + * Create controls for the component + * + * @return {Element} + * The element that was created. + * + * @private + */ + createElControls_() { + const defaultsDescription = this.localize('restore all settings to the default values'); + return createEl('div', { + className: 'vjs-track-settings-controls', + innerHTML: [`<button type="button" class="vjs-default-button" title="${defaultsDescription}">`, this.localize('Reset'), `<span class="vjs-control-text"> ${defaultsDescription}</span>`, '</button>', `<button type="button" class="vjs-done-button">${this.localize('Done')}</button>`].join('') + }); + } + content() { + return [this.createElColors_(), this.createElFont_(), this.createElControls_()]; + } + label() { + return this.localize('Caption Settings Dialog'); + } + description() { + return this.localize('Beginning of dialog window. Escape will cancel and close the window.'); + } + buildCSSClass() { + return super.buildCSSClass() + ' vjs-text-track-settings'; + } + + /** + * Gets an object of text track settings (or null). + * + * @return {Object} + * An object with config values parsed from the DOM or localStorage. + */ + getValues() { + return reduce(selectConfigs, (accum, config, key) => { + const value = getSelectedOptionValue(this.$(config.selector), config.parser); + if (value !== undefined) { + accum[key] = value; + } + return accum; + }, {}); + } + + /** + * Sets text track settings from an object of values. + * + * @param {Object} values + * An object with config values parsed from the DOM or localStorage. + */ + setValues(values) { + each(selectConfigs, (config, key) => { + setSelectedOption(this.$(config.selector), values[key], config.parser); + }); + } + + /** + * Sets all `<select>` elements to their default values. + */ + setDefaults() { + each(selectConfigs, config => { + const index = config.hasOwnProperty('default') ? config.default : 0; + this.$(config.selector).selectedIndex = index; + }); + } + + /** + * Restore texttrack settings from localStorage + */ + restoreSettings() { + let values; + try { + values = JSON.parse(window$1.localStorage.getItem(LOCAL_STORAGE_KEY$1)); + } catch (err) { + log$1.warn(err); + } + if (values) { + this.setValues(values); + } + } + + /** + * Save text track settings to localStorage + */ + saveSettings() { + if (!this.options_.persistTextTrackSettings) { + return; + } + const values = this.getValues(); + try { + if (Object.keys(values).length) { + window$1.localStorage.setItem(LOCAL_STORAGE_KEY$1, JSON.stringify(values)); + } else { + window$1.localStorage.removeItem(LOCAL_STORAGE_KEY$1); + } + } catch (err) { + log$1.warn(err); + } + } + + /** + * Update display of text track settings + */ + updateDisplay() { + const ttDisplay = this.player_.getChild('textTrackDisplay'); + if (ttDisplay) { + ttDisplay.updateDisplay(); + } + } + + /** + * conditionally blur the element and refocus the captions button + * + * @private + */ + conditionalBlur_() { + this.previouslyActiveEl_ = null; + const cb = this.player_.controlBar; + const subsCapsBtn = cb && cb.subsCapsButton; + const ccBtn = cb && cb.captionsButton; + if (subsCapsBtn) { + subsCapsBtn.focus(); + } else if (ccBtn) { + ccBtn.focus(); + } + } + + /** + * Repopulate dialog with new localizations on languagechange + */ + handleLanguagechange() { + this.fill(); + } +} +Component$1.registerComponent('TextTrackSettings', TextTrackSettings); + +/** + * @file resize-manager.js + */ + +/** + * A Resize Manager. It is in charge of triggering `playerresize` on the player in the right conditions. + * + * It'll either create an iframe and use a debounced resize handler on it or use the new {@link https://wicg.github.io/ResizeObserver/|ResizeObserver}. + * + * If the ResizeObserver is available natively, it will be used. A polyfill can be passed in as an option. + * If a `playerresize` event is not needed, the ResizeManager component can be removed from the player, see the example below. + * + * @example <caption>How to disable the resize manager</caption> + * const player = videojs('#vid', { + * resizeManager: false + * }); + * + * @see {@link https://wicg.github.io/ResizeObserver/|ResizeObserver specification} + * + * @extends Component + */ +class ResizeManager extends Component$1 { + /** + * Create the ResizeManager. + * + * @param {Object} player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of ResizeManager options. + * + * @param {Object} [options.ResizeObserver] + * A polyfill for ResizeObserver can be passed in here. + * If this is set to null it will ignore the native ResizeObserver and fall back to the iframe fallback. + */ + constructor(player, options) { + let RESIZE_OBSERVER_AVAILABLE = options.ResizeObserver || window$1.ResizeObserver; + + // if `null` was passed, we want to disable the ResizeObserver + if (options.ResizeObserver === null) { + RESIZE_OBSERVER_AVAILABLE = false; + } + + // Only create an element when ResizeObserver isn't available + const options_ = merge$1({ + createEl: !RESIZE_OBSERVER_AVAILABLE, + reportTouchActivity: false + }, options); + super(player, options_); + this.ResizeObserver = options.ResizeObserver || window$1.ResizeObserver; + this.loadListener_ = null; + this.resizeObserver_ = null; + this.debouncedHandler_ = debounce(() => { + this.resizeHandler(); + }, 100, false, this); + if (RESIZE_OBSERVER_AVAILABLE) { + this.resizeObserver_ = new this.ResizeObserver(this.debouncedHandler_); + this.resizeObserver_.observe(player.el()); + } else { + this.loadListener_ = () => { + if (!this.el_ || !this.el_.contentWindow) { + return; + } + const debouncedHandler_ = this.debouncedHandler_; + let unloadListener_ = this.unloadListener_ = function () { + off(this, 'resize', debouncedHandler_); + off(this, 'unload', unloadListener_); + unloadListener_ = null; + }; + + // safari and edge can unload the iframe before resizemanager dispose + // we have to dispose of event handlers correctly before that happens + on(this.el_.contentWindow, 'unload', unloadListener_); + on(this.el_.contentWindow, 'resize', debouncedHandler_); + }; + this.one('load', this.loadListener_); + } + } + createEl() { + return super.createEl('iframe', { + className: 'vjs-resize-manager', + tabIndex: -1, + title: this.localize('No content') + }, { + 'aria-hidden': 'true' + }); + } + + /** + * Called when a resize is triggered on the iframe or a resize is observed via the ResizeObserver + * + * @fires Player#playerresize + */ + resizeHandler() { + /** + * Called when the player size has changed + * + * @event Player#playerresize + * @type {Event} + */ + // make sure player is still around to trigger + // prevents this from causing an error after dispose + if (!this.player_ || !this.player_.trigger) { + return; + } + this.player_.trigger('playerresize'); + } + dispose() { + if (this.debouncedHandler_) { + this.debouncedHandler_.cancel(); + } + if (this.resizeObserver_) { + if (this.player_.el()) { + this.resizeObserver_.unobserve(this.player_.el()); + } + this.resizeObserver_.disconnect(); + } + if (this.loadListener_) { + this.off('load', this.loadListener_); + } + if (this.el_ && this.el_.contentWindow && this.unloadListener_) { + this.unloadListener_.call(this.el_.contentWindow); + } + this.ResizeObserver = null; + this.resizeObserver = null; + this.debouncedHandler_ = null; + this.loadListener_ = null; + super.dispose(); + } +} +Component$1.registerComponent('ResizeManager', ResizeManager); + +const defaults = { + trackingThreshold: 20, + liveTolerance: 15 +}; + +/* + track when we are at the live edge, and other helpers for live playback */ + +/** + * A class for checking live current time and determining when the player + * is at or behind the live edge. + */ +class LiveTracker extends Component$1 { + /** + * Creates an instance of this class. + * + * @param { import('./player').default } player + * The `Player` that this class should be attached to. + * + * @param {Object} [options] + * The key/value store of player options. + * + * @param {number} [options.trackingThreshold=20] + * Number of seconds of live window (seekableEnd - seekableStart) that + * media needs to have before the liveui will be shown. + * + * @param {number} [options.liveTolerance=15] + * Number of seconds behind live that we have to be + * before we will be considered non-live. Note that this will only + * be used when playing at the live edge. This allows large seekable end + * changes to not effect whether we are live or not. + */ + constructor(player, options) { + // LiveTracker does not need an element + const options_ = merge$1(defaults, options, { + createEl: false + }); + super(player, options_); + this.trackLiveHandler_ = () => this.trackLive_(); + this.handlePlay_ = e => this.handlePlay(e); + this.handleFirstTimeupdate_ = e => this.handleFirstTimeupdate(e); + this.handleSeeked_ = e => this.handleSeeked(e); + this.seekToLiveEdge_ = e => this.seekToLiveEdge(e); + this.reset_(); + this.on(this.player_, 'durationchange', e => this.handleDurationchange(e)); + // we should try to toggle tracking on canplay as native playback engines, like Safari + // may not have the proper values for things like seekableEnd until then + this.on(this.player_, 'canplay', () => this.toggleTracking()); + } + + /** + * all the functionality for tracking when seek end changes + * and for tracking how far past seek end we should be + */ + trackLive_() { + const seekable = this.player_.seekable(); + + // skip undefined seekable + if (!seekable || !seekable.length) { + return; + } + const newTime = Number(window$1.performance.now().toFixed(4)); + const deltaTime = this.lastTime_ === -1 ? 0 : (newTime - this.lastTime_) / 1000; + this.lastTime_ = newTime; + this.pastSeekEnd_ = this.pastSeekEnd() + deltaTime; + const liveCurrentTime = this.liveCurrentTime(); + const currentTime = this.player_.currentTime(); + + // we are behind live if any are true + // 1. the player is paused + // 2. the user seeked to a location 2 seconds away from live + // 3. the difference between live and current time is greater + // liveTolerance which defaults to 15s + let isBehind = this.player_.paused() || this.seekedBehindLive_ || Math.abs(liveCurrentTime - currentTime) > this.options_.liveTolerance; + + // we cannot be behind if + // 1. until we have not seen a timeupdate yet + // 2. liveCurrentTime is Infinity, which happens on Android and Native Safari + if (!this.timeupdateSeen_ || liveCurrentTime === Infinity) { + isBehind = false; + } + if (isBehind !== this.behindLiveEdge_) { + this.behindLiveEdge_ = isBehind; + this.trigger('liveedgechange'); + } + } + + /** + * handle a durationchange event on the player + * and start/stop tracking accordingly. + */ + handleDurationchange() { + this.toggleTracking(); + } + + /** + * start/stop tracking + */ + toggleTracking() { + if (this.player_.duration() === Infinity && this.liveWindow() >= this.options_.trackingThreshold) { + if (this.player_.options_.liveui) { + this.player_.addClass('vjs-liveui'); + } + this.startTracking(); + } else { + this.player_.removeClass('vjs-liveui'); + this.stopTracking(); + } + } + + /** + * start tracking live playback + */ + startTracking() { + if (this.isTracking()) { + return; + } + + // If we haven't seen a timeupdate, we need to check whether playback + // began before this component started tracking. This can happen commonly + // when using autoplay. + if (!this.timeupdateSeen_) { + this.timeupdateSeen_ = this.player_.hasStarted(); + } + this.trackingInterval_ = this.setInterval(this.trackLiveHandler_, UPDATE_REFRESH_INTERVAL); + this.trackLive_(); + this.on(this.player_, ['play', 'pause'], this.trackLiveHandler_); + if (!this.timeupdateSeen_) { + this.one(this.player_, 'play', this.handlePlay_); + this.one(this.player_, 'timeupdate', this.handleFirstTimeupdate_); + } else { + this.on(this.player_, 'seeked', this.handleSeeked_); + } + } + + /** + * handle the first timeupdate on the player if it wasn't already playing + * when live tracker started tracking. + */ + handleFirstTimeupdate() { + this.timeupdateSeen_ = true; + this.on(this.player_, 'seeked', this.handleSeeked_); + } + + /** + * Keep track of what time a seek starts, and listen for seeked + * to find where a seek ends. + */ + handleSeeked() { + const timeDiff = Math.abs(this.liveCurrentTime() - this.player_.currentTime()); + this.seekedBehindLive_ = this.nextSeekedFromUser_ && timeDiff > 2; + this.nextSeekedFromUser_ = false; + this.trackLive_(); + } + + /** + * handle the first play on the player, and make sure that we seek + * right to the live edge. + */ + handlePlay() { + this.one(this.player_, 'timeupdate', this.seekToLiveEdge_); + } + + /** + * Stop tracking, and set all internal variables to + * their initial value. + */ + reset_() { + this.lastTime_ = -1; + this.pastSeekEnd_ = 0; + this.lastSeekEnd_ = -1; + this.behindLiveEdge_ = true; + this.timeupdateSeen_ = false; + this.seekedBehindLive_ = false; + this.nextSeekedFromUser_ = false; + this.clearInterval(this.trackingInterval_); + this.trackingInterval_ = null; + this.off(this.player_, ['play', 'pause'], this.trackLiveHandler_); + this.off(this.player_, 'seeked', this.handleSeeked_); + this.off(this.player_, 'play', this.handlePlay_); + this.off(this.player_, 'timeupdate', this.handleFirstTimeupdate_); + this.off(this.player_, 'timeupdate', this.seekToLiveEdge_); + } + + /** + * The next seeked event is from the user. Meaning that any seek + * > 2s behind live will be considered behind live for real and + * liveTolerance will be ignored. + */ + nextSeekedFromUser() { + this.nextSeekedFromUser_ = true; + } + + /** + * stop tracking live playback + */ + stopTracking() { + if (!this.isTracking()) { + return; + } + this.reset_(); + this.trigger('liveedgechange'); + } + + /** + * A helper to get the player seekable end + * so that we don't have to null check everywhere + * + * @return {number} + * The furthest seekable end or Infinity. + */ + seekableEnd() { + const seekable = this.player_.seekable(); + const seekableEnds = []; + let i = seekable ? seekable.length : 0; + while (i--) { + seekableEnds.push(seekable.end(i)); + } + + // grab the furthest seekable end after sorting, or if there are none + // default to Infinity + return seekableEnds.length ? seekableEnds.sort()[seekableEnds.length - 1] : Infinity; + } + + /** + * A helper to get the player seekable start + * so that we don't have to null check everywhere + * + * @return {number} + * The earliest seekable start or 0. + */ + seekableStart() { + const seekable = this.player_.seekable(); + const seekableStarts = []; + let i = seekable ? seekable.length : 0; + while (i--) { + seekableStarts.push(seekable.start(i)); + } + + // grab the first seekable start after sorting, or if there are none + // default to 0 + return seekableStarts.length ? seekableStarts.sort()[0] : 0; + } + + /** + * Get the live time window aka + * the amount of time between seekable start and + * live current time. + * + * @return {number} + * The amount of seconds that are seekable in + * the live video. + */ + liveWindow() { + const liveCurrentTime = this.liveCurrentTime(); + + // if liveCurrenTime is Infinity then we don't have a liveWindow at all + if (liveCurrentTime === Infinity) { + return 0; + } + return liveCurrentTime - this.seekableStart(); + } + + /** + * Determines if the player is live, only checks if this component + * is tracking live playback or not + * + * @return {boolean} + * Whether liveTracker is tracking + */ + isLive() { + return this.isTracking(); + } + + /** + * Determines if currentTime is at the live edge and won't fall behind + * on each seekableendchange + * + * @return {boolean} + * Whether playback is at the live edge + */ + atLiveEdge() { + return !this.behindLiveEdge(); + } + + /** + * get what we expect the live current time to be + * + * @return {number} + * The expected live current time + */ + liveCurrentTime() { + return this.pastSeekEnd() + this.seekableEnd(); + } + + /** + * The number of seconds that have occurred after seekable end + * changed. This will be reset to 0 once seekable end changes. + * + * @return {number} + * Seconds past the current seekable end + */ + pastSeekEnd() { + const seekableEnd = this.seekableEnd(); + if (this.lastSeekEnd_ !== -1 && seekableEnd !== this.lastSeekEnd_) { + this.pastSeekEnd_ = 0; + } + this.lastSeekEnd_ = seekableEnd; + return this.pastSeekEnd_; + } + + /** + * If we are currently behind the live edge, aka currentTime will be + * behind on a seekableendchange + * + * @return {boolean} + * If we are behind the live edge + */ + behindLiveEdge() { + return this.behindLiveEdge_; + } + + /** + * Whether live tracker is currently tracking or not. + */ + isTracking() { + return typeof this.trackingInterval_ === 'number'; + } + + /** + * Seek to the live edge if we are behind the live edge + */ + seekToLiveEdge() { + this.seekedBehindLive_ = false; + if (this.atLiveEdge()) { + return; + } + this.nextSeekedFromUser_ = false; + this.player_.currentTime(this.liveCurrentTime()); + } + + /** + * Dispose of liveTracker + */ + dispose() { + this.stopTracking(); + super.dispose(); + } +} +Component$1.registerComponent('LiveTracker', LiveTracker); + +/** + * Displays an element over the player which contains an optional title and + * description for the current content. + * + * Much of the code for this component originated in the now obsolete + * videojs-dock plugin: https://github.com/brightcove/videojs-dock/ + * + * @extends Component + */ +class TitleBar extends Component$1 { + constructor(player, options) { + super(player, options); + this.on('statechanged', e => this.updateDom_()); + this.updateDom_(); + } + + /** + * Create the `TitleBar`'s DOM element + * + * @return {Element} + * The element that was created. + */ + createEl() { + this.els = { + title: createEl('div', { + className: 'vjs-title-bar-title', + id: `vjs-title-bar-title-${newGUID()}` + }), + description: createEl('div', { + className: 'vjs-title-bar-description', + id: `vjs-title-bar-description-${newGUID()}` + }) + }; + return createEl('div', { + className: 'vjs-title-bar' + }, {}, values(this.els)); + } + + /** + * Updates the DOM based on the component's state object. + */ + updateDom_() { + const tech = this.player_.tech_; + const techEl = tech && tech.el_; + const techAriaAttrs = { + title: 'aria-labelledby', + description: 'aria-describedby' + }; + ['title', 'description'].forEach(k => { + const value = this.state[k]; + const el = this.els[k]; + const techAriaAttr = techAriaAttrs[k]; + emptyEl(el); + if (value) { + textContent(el, value); + } + + // If there is a tech element available, update its ARIA attributes + // according to whether a title and/or description have been provided. + if (techEl) { + techEl.removeAttribute(techAriaAttr); + if (value) { + techEl.setAttribute(techAriaAttr, el.id); + } + } + }); + if (this.state.title || this.state.description) { + this.show(); + } else { + this.hide(); + } + } + + /** + * Update the contents of the title bar component with new title and + * description text. + * + * If both title and description are missing, the title bar will be hidden. + * + * If either title or description are present, the title bar will be visible. + * + * NOTE: Any previously set value will be preserved. To unset a previously + * set value, you must pass an empty string or null. + * + * For example: + * + * ``` + * update({title: 'foo', description: 'bar'}) // title: 'foo', description: 'bar' + * update({description: 'bar2'}) // title: 'foo', description: 'bar2' + * update({title: ''}) // title: '', description: 'bar2' + * update({title: 'foo', description: null}) // title: 'foo', description: null + * ``` + * + * @param {Object} [options={}] + * An options object. When empty, the title bar will be hidden. + * + * @param {string} [options.title] + * A title to display in the title bar. + * + * @param {string} [options.description] + * A description to display in the title bar. + */ + update(options) { + this.setState(options); + } + + /** + * Dispose the component. + */ + dispose() { + const tech = this.player_.tech_; + const techEl = tech && tech.el_; + if (techEl) { + techEl.removeAttribute('aria-labelledby'); + techEl.removeAttribute('aria-describedby'); + } + super.dispose(); + this.els = null; + } +} +Component$1.registerComponent('TitleBar', TitleBar); + +/** + * This function is used to fire a sourceset when there is something + * similar to `mediaEl.load()` being called. It will try to find the source via + * the `src` attribute and then the `<source>` elements. It will then fire `sourceset` + * with the source that was found or empty string if we cannot know. If it cannot + * find a source then `sourceset` will not be fired. + * + * @param { import('./html5').default } tech + * The tech object that sourceset was setup on + * + * @return {boolean} + * returns false if the sourceset was not fired and true otherwise. + */ +const sourcesetLoad = tech => { + const el = tech.el(); + + // if `el.src` is set, that source will be loaded. + if (el.hasAttribute('src')) { + tech.triggerSourceset(el.src); + return true; + } + + /** + * Since there isn't a src property on the media element, source elements will be used for + * implementing the source selection algorithm. This happens asynchronously and + * for most cases were there is more than one source we cannot tell what source will + * be loaded, without re-implementing the source selection algorithm. At this time we are not + * going to do that. There are three special cases that we do handle here though: + * + * 1. If there are no sources, do not fire `sourceset`. + * 2. If there is only one `<source>` with a `src` property/attribute that is our `src` + * 3. If there is more than one `<source>` but all of them have the same `src` url. + * That will be our src. + */ + const sources = tech.$$('source'); + const srcUrls = []; + let src = ''; + + // if there are no sources, do not fire sourceset + if (!sources.length) { + return false; + } + + // only count valid/non-duplicate source elements + for (let i = 0; i < sources.length; i++) { + const url = sources[i].src; + if (url && srcUrls.indexOf(url) === -1) { + srcUrls.push(url); + } + } + + // there were no valid sources + if (!srcUrls.length) { + return false; + } + + // there is only one valid source element url + // use that + if (srcUrls.length === 1) { + src = srcUrls[0]; + } + tech.triggerSourceset(src); + return true; +}; + +/** + * our implementation of an `innerHTML` descriptor for browsers + * that do not have one. + */ +const innerHTMLDescriptorPolyfill = Object.defineProperty({}, 'innerHTML', { + get() { + return this.cloneNode(true).innerHTML; + }, + set(v) { + // make a dummy node to use innerHTML on + const dummy = document.createElement(this.nodeName.toLowerCase()); + + // set innerHTML to the value provided + dummy.innerHTML = v; + + // make a document fragment to hold the nodes from dummy + const docFrag = document.createDocumentFragment(); + + // copy all of the nodes created by the innerHTML on dummy + // to the document fragment + while (dummy.childNodes.length) { + docFrag.appendChild(dummy.childNodes[0]); + } + + // remove content + this.innerText = ''; + + // now we add all of that html in one by appending the + // document fragment. This is how innerHTML does it. + window$1.Element.prototype.appendChild.call(this, docFrag); + + // then return the result that innerHTML's setter would + return this.innerHTML; + } +}); + +/** + * Get a property descriptor given a list of priorities and the + * property to get. + */ +const getDescriptor = (priority, prop) => { + let descriptor = {}; + for (let i = 0; i < priority.length; i++) { + descriptor = Object.getOwnPropertyDescriptor(priority[i], prop); + if (descriptor && descriptor.set && descriptor.get) { + break; + } + } + descriptor.enumerable = true; + descriptor.configurable = true; + return descriptor; +}; +const getInnerHTMLDescriptor = tech => getDescriptor([tech.el(), window$1.HTMLMediaElement.prototype, window$1.Element.prototype, innerHTMLDescriptorPolyfill], 'innerHTML'); + +/** + * Patches browser internal functions so that we can tell synchronously + * if a `<source>` was appended to the media element. For some reason this + * causes a `sourceset` if the the media element is ready and has no source. + * This happens when: + * - The page has just loaded and the media element does not have a source. + * - The media element was emptied of all sources, then `load()` was called. + * + * It does this by patching the following functions/properties when they are supported: + * + * - `append()` - can be used to add a `<source>` element to the media element + * - `appendChild()` - can be used to add a `<source>` element to the media element + * - `insertAdjacentHTML()` - can be used to add a `<source>` element to the media element + * - `innerHTML` - can be used to add a `<source>` element to the media element + * + * @param {Html5} tech + * The tech object that sourceset is being setup on. + */ +const firstSourceWatch = function (tech) { + const el = tech.el(); + + // make sure firstSourceWatch isn't setup twice. + if (el.resetSourceWatch_) { + return; + } + const old = {}; + const innerDescriptor = getInnerHTMLDescriptor(tech); + const appendWrapper = appendFn => (...args) => { + const retval = appendFn.apply(el, args); + sourcesetLoad(tech); + return retval; + }; + ['append', 'appendChild', 'insertAdjacentHTML'].forEach(k => { + if (!el[k]) { + return; + } + + // store the old function + old[k] = el[k]; + + // call the old function with a sourceset if a source + // was loaded + el[k] = appendWrapper(old[k]); + }); + Object.defineProperty(el, 'innerHTML', merge$1(innerDescriptor, { + set: appendWrapper(innerDescriptor.set) + })); + el.resetSourceWatch_ = () => { + el.resetSourceWatch_ = null; + Object.keys(old).forEach(k => { + el[k] = old[k]; + }); + Object.defineProperty(el, 'innerHTML', innerDescriptor); + }; + + // on the first sourceset, we need to revert our changes + tech.one('sourceset', el.resetSourceWatch_); +}; + +/** + * our implementation of a `src` descriptor for browsers + * that do not have one + */ +const srcDescriptorPolyfill = Object.defineProperty({}, 'src', { + get() { + if (this.hasAttribute('src')) { + return getAbsoluteURL(window$1.Element.prototype.getAttribute.call(this, 'src')); + } + return ''; + }, + set(v) { + window$1.Element.prototype.setAttribute.call(this, 'src', v); + return v; + } +}); +const getSrcDescriptor = tech => getDescriptor([tech.el(), window$1.HTMLMediaElement.prototype, srcDescriptorPolyfill], 'src'); + +/** + * setup `sourceset` handling on the `Html5` tech. This function + * patches the following element properties/functions: + * + * - `src` - to determine when `src` is set + * - `setAttribute()` - to determine when `src` is set + * - `load()` - this re-triggers the source selection algorithm, and can + * cause a sourceset. + * + * If there is no source when we are adding `sourceset` support or during a `load()` + * we also patch the functions listed in `firstSourceWatch`. + * + * @param {Html5} tech + * The tech to patch + */ +const setupSourceset = function (tech) { + if (!tech.featuresSourceset) { + return; + } + const el = tech.el(); + + // make sure sourceset isn't setup twice. + if (el.resetSourceset_) { + return; + } + const srcDescriptor = getSrcDescriptor(tech); + const oldSetAttribute = el.setAttribute; + const oldLoad = el.load; + Object.defineProperty(el, 'src', merge$1(srcDescriptor, { + set: v => { + const retval = srcDescriptor.set.call(el, v); + + // we use the getter here to get the actual value set on src + tech.triggerSourceset(el.src); + return retval; + } + })); + el.setAttribute = (n, v) => { + const retval = oldSetAttribute.call(el, n, v); + if (/src/i.test(n)) { + tech.triggerSourceset(el.src); + } + return retval; + }; + el.load = () => { + const retval = oldLoad.call(el); + + // if load was called, but there was no source to fire + // sourceset on. We have to watch for a source append + // as that can trigger a `sourceset` when the media element + // has no source + if (!sourcesetLoad(tech)) { + tech.triggerSourceset(''); + firstSourceWatch(tech); + } + return retval; + }; + if (el.currentSrc) { + tech.triggerSourceset(el.currentSrc); + } else if (!sourcesetLoad(tech)) { + firstSourceWatch(tech); + } + el.resetSourceset_ = () => { + el.resetSourceset_ = null; + el.load = oldLoad; + el.setAttribute = oldSetAttribute; + Object.defineProperty(el, 'src', srcDescriptor); + if (el.resetSourceWatch_) { + el.resetSourceWatch_(); + } + }; +}; + +/** + * @file html5.js + */ + +/** + * HTML5 Media Controller - Wrapper for HTML5 Media API + * + * @mixes Tech~SourceHandlerAdditions + * @extends Tech + */ +class Html5 extends Tech { + /** + * Create an instance of this Tech. + * + * @param {Object} [options] + * The key/value store of player options. + * + * @param {Function} [ready] + * Callback function to call when the `HTML5` Tech is ready. + */ + constructor(options, ready) { + super(options, ready); + const source = options.source; + let crossoriginTracks = false; + this.featuresVideoFrameCallback = this.featuresVideoFrameCallback && this.el_.tagName === 'VIDEO'; + + // Set the source if one is provided + // 1) Check if the source is new (if not, we want to keep the original so playback isn't interrupted) + // 2) Check to see if the network state of the tag was failed at init, and if so, reset the source + // anyway so the error gets fired. + if (source && (this.el_.currentSrc !== source.src || options.tag && options.tag.initNetworkState_ === 3)) { + this.setSource(source); + } else { + this.handleLateInit_(this.el_); + } + + // setup sourceset after late sourceset/init + if (options.enableSourceset) { + this.setupSourcesetHandling_(); + } + this.isScrubbing_ = false; + if (this.el_.hasChildNodes()) { + const nodes = this.el_.childNodes; + let nodesLength = nodes.length; + const removeNodes = []; + while (nodesLength--) { + const node = nodes[nodesLength]; + const nodeName = node.nodeName.toLowerCase(); + if (nodeName === 'track') { + if (!this.featuresNativeTextTracks) { + // Empty video tag tracks so the built-in player doesn't use them also. + // This may not be fast enough to stop HTML5 browsers from reading the tags + // so we'll need to turn off any default tracks if we're manually doing + // captions and subtitles. videoElement.textTracks + removeNodes.push(node); + } else { + // store HTMLTrackElement and TextTrack to remote list + this.remoteTextTrackEls().addTrackElement_(node); + this.remoteTextTracks().addTrack(node.track); + this.textTracks().addTrack(node.track); + if (!crossoriginTracks && !this.el_.hasAttribute('crossorigin') && isCrossOrigin(node.src)) { + crossoriginTracks = true; + } + } + } + } + for (let i = 0; i < removeNodes.length; i++) { + this.el_.removeChild(removeNodes[i]); + } + } + this.proxyNativeTracks_(); + if (this.featuresNativeTextTracks && crossoriginTracks) { + log$1.warn('Text Tracks are being loaded from another origin but the crossorigin attribute isn\'t used.\n' + 'This may prevent text tracks from loading.'); + } + + // prevent iOS Safari from disabling metadata text tracks during native playback + this.restoreMetadataTracksInIOSNativePlayer_(); + + // Determine if native controls should be used + // Our goal should be to get the custom controls on mobile solid everywhere + // so we can remove this all together. Right now this will block custom + // controls on touch enabled laptops like the Chrome Pixel + if ((TOUCH_ENABLED || IS_IPHONE) && options.nativeControlsForTouch === true) { + this.setControls(true); + } + + // on iOS, we want to proxy `webkitbeginfullscreen` and `webkitendfullscreen` + // into a `fullscreenchange` event + this.proxyWebkitFullscreen_(); + this.triggerReady(); + } + + /** + * Dispose of `HTML5` media element and remove all tracks. + */ + dispose() { + if (this.el_ && this.el_.resetSourceset_) { + this.el_.resetSourceset_(); + } + Html5.disposeMediaElement(this.el_); + this.options_ = null; + + // tech will handle clearing of the emulated track list + super.dispose(); + } + + /** + * Modify the media element so that we can detect when + * the source is changed. Fires `sourceset` just after the source has changed + */ + setupSourcesetHandling_() { + setupSourceset(this); + } + + /** + * When a captions track is enabled in the iOS Safari native player, all other + * tracks are disabled (including metadata tracks), which nulls all of their + * associated cue points. This will restore metadata tracks to their pre-fullscreen + * state in those cases so that cue points are not needlessly lost. + * + * @private + */ + restoreMetadataTracksInIOSNativePlayer_() { + const textTracks = this.textTracks(); + let metadataTracksPreFullscreenState; + + // captures a snapshot of every metadata track's current state + const takeMetadataTrackSnapshot = () => { + metadataTracksPreFullscreenState = []; + for (let i = 0; i < textTracks.length; i++) { + const track = textTracks[i]; + if (track.kind === 'metadata') { + metadataTracksPreFullscreenState.push({ + track, + storedMode: track.mode + }); + } + } + }; + + // snapshot each metadata track's initial state, and update the snapshot + // each time there is a track 'change' event + takeMetadataTrackSnapshot(); + textTracks.addEventListener('change', takeMetadataTrackSnapshot); + this.on('dispose', () => textTracks.removeEventListener('change', takeMetadataTrackSnapshot)); + const restoreTrackMode = () => { + for (let i = 0; i < metadataTracksPreFullscreenState.length; i++) { + const storedTrack = metadataTracksPreFullscreenState[i]; + if (storedTrack.track.mode === 'disabled' && storedTrack.track.mode !== storedTrack.storedMode) { + storedTrack.track.mode = storedTrack.storedMode; + } + } + // we only want this handler to be executed on the first 'change' event + textTracks.removeEventListener('change', restoreTrackMode); + }; + + // when we enter fullscreen playback, stop updating the snapshot and + // restore all track modes to their pre-fullscreen state + this.on('webkitbeginfullscreen', () => { + textTracks.removeEventListener('change', takeMetadataTrackSnapshot); + + // remove the listener before adding it just in case it wasn't previously removed + textTracks.removeEventListener('change', restoreTrackMode); + textTracks.addEventListener('change', restoreTrackMode); + }); + + // start updating the snapshot again after leaving fullscreen + this.on('webkitendfullscreen', () => { + // remove the listener before adding it just in case it wasn't previously removed + textTracks.removeEventListener('change', takeMetadataTrackSnapshot); + textTracks.addEventListener('change', takeMetadataTrackSnapshot); + + // remove the restoreTrackMode handler in case it wasn't triggered during fullscreen playback + textTracks.removeEventListener('change', restoreTrackMode); + }); + } + + /** + * Attempt to force override of tracks for the given type + * + * @param {string} type - Track type to override, possible values include 'Audio', + * 'Video', and 'Text'. + * @param {boolean} override - If set to true native audio/video will be overridden, + * otherwise native audio/video will potentially be used. + * @private + */ + overrideNative_(type, override) { + // If there is no behavioral change don't add/remove listeners + if (override !== this[`featuresNative${type}Tracks`]) { + return; + } + const lowerCaseType = type.toLowerCase(); + if (this[`${lowerCaseType}TracksListeners_`]) { + Object.keys(this[`${lowerCaseType}TracksListeners_`]).forEach(eventName => { + const elTracks = this.el()[`${lowerCaseType}Tracks`]; + elTracks.removeEventListener(eventName, this[`${lowerCaseType}TracksListeners_`][eventName]); + }); + } + this[`featuresNative${type}Tracks`] = !override; + this[`${lowerCaseType}TracksListeners_`] = null; + this.proxyNativeTracksForType_(lowerCaseType); + } + + /** + * Attempt to force override of native audio tracks. + * + * @param {boolean} override - If set to true native audio will be overridden, + * otherwise native audio will potentially be used. + */ + overrideNativeAudioTracks(override) { + this.overrideNative_('Audio', override); + } + + /** + * Attempt to force override of native video tracks. + * + * @param {boolean} override - If set to true native video will be overridden, + * otherwise native video will potentially be used. + */ + overrideNativeVideoTracks(override) { + this.overrideNative_('Video', override); + } + + /** + * Proxy native track list events for the given type to our track + * lists if the browser we are playing in supports that type of track list. + * + * @param {string} name - Track type; values include 'audio', 'video', and 'text' + * @private + */ + proxyNativeTracksForType_(name) { + const props = NORMAL[name]; + const elTracks = this.el()[props.getterName]; + const techTracks = this[props.getterName](); + if (!this[`featuresNative${props.capitalName}Tracks`] || !elTracks || !elTracks.addEventListener) { + return; + } + const listeners = { + change: e => { + const event = { + type: 'change', + target: techTracks, + currentTarget: techTracks, + srcElement: techTracks + }; + techTracks.trigger(event); + + // if we are a text track change event, we should also notify the + // remote text track list. This can potentially cause a false positive + // if we were to get a change event on a non-remote track and + // we triggered the event on the remote text track list which doesn't + // contain that track. However, best practices mean looping through the + // list of tracks and searching for the appropriate mode value, so, + // this shouldn't pose an issue + if (name === 'text') { + this[REMOTE.remoteText.getterName]().trigger(event); + } + }, + addtrack(e) { + techTracks.addTrack(e.track); + }, + removetrack(e) { + techTracks.removeTrack(e.track); + } + }; + const removeOldTracks = function () { + const removeTracks = []; + for (let i = 0; i < techTracks.length; i++) { + let found = false; + for (let j = 0; j < elTracks.length; j++) { + if (elTracks[j] === techTracks[i]) { + found = true; + break; + } + } + if (!found) { + removeTracks.push(techTracks[i]); + } + } + while (removeTracks.length) { + techTracks.removeTrack(removeTracks.shift()); + } + }; + this[props.getterName + 'Listeners_'] = listeners; + Object.keys(listeners).forEach(eventName => { + const listener = listeners[eventName]; + elTracks.addEventListener(eventName, listener); + this.on('dispose', e => elTracks.removeEventListener(eventName, listener)); + }); + + // Remove (native) tracks that are not used anymore + this.on('loadstart', removeOldTracks); + this.on('dispose', e => this.off('loadstart', removeOldTracks)); + } + + /** + * Proxy all native track list events to our track lists if the browser we are playing + * in supports that type of track list. + * + * @private + */ + proxyNativeTracks_() { + NORMAL.names.forEach(name => { + this.proxyNativeTracksForType_(name); + }); + } + + /** + * Create the `Html5` Tech's DOM element. + * + * @return {Element} + * The element that gets created. + */ + createEl() { + let el = this.options_.tag; + + // Check if this browser supports moving the element into the box. + // On the iPhone video will break if you move the element, + // So we have to create a brand new element. + // If we ingested the player div, we do not need to move the media element. + if (!el || !(this.options_.playerElIngest || this.movingMediaElementInDOM)) { + // If the original tag is still there, clone and remove it. + if (el) { + const clone = el.cloneNode(true); + if (el.parentNode) { + el.parentNode.insertBefore(clone, el); + } + Html5.disposeMediaElement(el); + el = clone; + } else { + el = document.createElement('video'); + + // determine if native controls should be used + const tagAttributes = this.options_.tag && getAttributes(this.options_.tag); + const attributes = merge$1({}, tagAttributes); + if (!TOUCH_ENABLED || this.options_.nativeControlsForTouch !== true) { + delete attributes.controls; + } + setAttributes(el, Object.assign(attributes, { + id: this.options_.techId, + class: 'vjs-tech' + })); + } + el.playerId = this.options_.playerId; + } + if (typeof this.options_.preload !== 'undefined') { + setAttribute(el, 'preload', this.options_.preload); + } + if (this.options_.disablePictureInPicture !== undefined) { + el.disablePictureInPicture = this.options_.disablePictureInPicture; + } + + // Update specific tag settings, in case they were overridden + // `autoplay` has to be *last* so that `muted` and `playsinline` are present + // when iOS/Safari or other browsers attempt to autoplay. + const settingsAttrs = ['loop', 'muted', 'playsinline', 'autoplay']; + for (let i = 0; i < settingsAttrs.length; i++) { + const attr = settingsAttrs[i]; + const value = this.options_[attr]; + if (typeof value !== 'undefined') { + if (value) { + setAttribute(el, attr, attr); + } else { + removeAttribute(el, attr); + } + el[attr] = value; + } + } + return el; + } + + /** + * This will be triggered if the loadstart event has already fired, before videojs was + * ready. Two known examples of when this can happen are: + * 1. If we're loading the playback object after it has started loading + * 2. The media is already playing the (often with autoplay on) then + * + * This function will fire another loadstart so that videojs can catchup. + * + * @fires Tech#loadstart + * + * @return {undefined} + * returns nothing. + */ + handleLateInit_(el) { + if (el.networkState === 0 || el.networkState === 3) { + // The video element hasn't started loading the source yet + // or didn't find a source + return; + } + if (el.readyState === 0) { + // NetworkState is set synchronously BUT loadstart is fired at the + // end of the current stack, usually before setInterval(fn, 0). + // So at this point we know loadstart may have already fired or is + // about to fire, and either way the player hasn't seen it yet. + // We don't want to fire loadstart prematurely here and cause a + // double loadstart so we'll wait and see if it happens between now + // and the next loop, and fire it if not. + // HOWEVER, we also want to make sure it fires before loadedmetadata + // which could also happen between now and the next loop, so we'll + // watch for that also. + let loadstartFired = false; + const setLoadstartFired = function () { + loadstartFired = true; + }; + this.on('loadstart', setLoadstartFired); + const triggerLoadstart = function () { + // We did miss the original loadstart. Make sure the player + // sees loadstart before loadedmetadata + if (!loadstartFired) { + this.trigger('loadstart'); + } + }; + this.on('loadedmetadata', triggerLoadstart); + this.ready(function () { + this.off('loadstart', setLoadstartFired); + this.off('loadedmetadata', triggerLoadstart); + if (!loadstartFired) { + // We did miss the original native loadstart. Fire it now. + this.trigger('loadstart'); + } + }); + return; + } + + // From here on we know that loadstart already fired and we missed it. + // The other readyState events aren't as much of a problem if we double + // them, so not going to go to as much trouble as loadstart to prevent + // that unless we find reason to. + const eventsToTrigger = ['loadstart']; + + // loadedmetadata: newly equal to HAVE_METADATA (1) or greater + eventsToTrigger.push('loadedmetadata'); + + // loadeddata: newly increased to HAVE_CURRENT_DATA (2) or greater + if (el.readyState >= 2) { + eventsToTrigger.push('loadeddata'); + } + + // canplay: newly increased to HAVE_FUTURE_DATA (3) or greater + if (el.readyState >= 3) { + eventsToTrigger.push('canplay'); + } + + // canplaythrough: newly equal to HAVE_ENOUGH_DATA (4) + if (el.readyState >= 4) { + eventsToTrigger.push('canplaythrough'); + } + + // We still need to give the player time to add event listeners + this.ready(function () { + eventsToTrigger.forEach(function (type) { + this.trigger(type); + }, this); + }); + } + + /** + * Set whether we are scrubbing or not. + * This is used to decide whether we should use `fastSeek` or not. + * `fastSeek` is used to provide trick play on Safari browsers. + * + * @param {boolean} isScrubbing + * - true for we are currently scrubbing + * - false for we are no longer scrubbing + */ + setScrubbing(isScrubbing) { + this.isScrubbing_ = isScrubbing; + } + + /** + * Get whether we are scrubbing or not. + * + * @return {boolean} isScrubbing + * - true for we are currently scrubbing + * - false for we are no longer scrubbing + */ + scrubbing() { + return this.isScrubbing_; + } + + /** + * Set current time for the `HTML5` tech. + * + * @param {number} seconds + * Set the current time of the media to this. + */ + setCurrentTime(seconds) { + try { + if (this.isScrubbing_ && this.el_.fastSeek && IS_ANY_SAFARI) { + this.el_.fastSeek(seconds); + } else { + this.el_.currentTime = seconds; + } + } catch (e) { + log$1(e, 'Video is not ready. (Video.js)'); + // this.warning(VideoJS.warnings.videoNotReady); + } + } + + /** + * Get the current duration of the HTML5 media element. + * + * @return {number} + * The duration of the media or 0 if there is no duration. + */ + duration() { + // Android Chrome will report duration as Infinity for VOD HLS until after + // playback has started, which triggers the live display erroneously. + // Return NaN if playback has not started and trigger a durationupdate once + // the duration can be reliably known. + if (this.el_.duration === Infinity && IS_ANDROID && IS_CHROME && this.el_.currentTime === 0) { + // Wait for the first `timeupdate` with currentTime > 0 - there may be + // several with 0 + const checkProgress = () => { + if (this.el_.currentTime > 0) { + // Trigger durationchange for genuinely live video + if (this.el_.duration === Infinity) { + this.trigger('durationchange'); + } + this.off('timeupdate', checkProgress); + } + }; + this.on('timeupdate', checkProgress); + return NaN; + } + return this.el_.duration || NaN; + } + + /** + * Get the current width of the HTML5 media element. + * + * @return {number} + * The width of the HTML5 media element. + */ + width() { + return this.el_.offsetWidth; + } + + /** + * Get the current height of the HTML5 media element. + * + * @return {number} + * The height of the HTML5 media element. + */ + height() { + return this.el_.offsetHeight; + } + + /** + * Proxy iOS `webkitbeginfullscreen` and `webkitendfullscreen` into + * `fullscreenchange` event. + * + * @private + * @fires fullscreenchange + * @listens webkitendfullscreen + * @listens webkitbeginfullscreen + * @listens webkitbeginfullscreen + */ + proxyWebkitFullscreen_() { + if (!('webkitDisplayingFullscreen' in this.el_)) { + return; + } + const endFn = function () { + this.trigger('fullscreenchange', { + isFullscreen: false + }); + // Safari will sometimes set controls on the videoelement when existing fullscreen. + if (this.el_.controls && !this.options_.nativeControlsForTouch && this.controls()) { + this.el_.controls = false; + } + }; + const beginFn = function () { + if ('webkitPresentationMode' in this.el_ && this.el_.webkitPresentationMode !== 'picture-in-picture') { + this.one('webkitendfullscreen', endFn); + this.trigger('fullscreenchange', { + isFullscreen: true, + // set a flag in case another tech triggers fullscreenchange + nativeIOSFullscreen: true + }); + } + }; + this.on('webkitbeginfullscreen', beginFn); + this.on('dispose', () => { + this.off('webkitbeginfullscreen', beginFn); + this.off('webkitendfullscreen', endFn); + }); + } + + /** + * Check if fullscreen is supported on the video el. + * + * @return {boolean} + * - True if fullscreen is supported. + * - False if fullscreen is not supported. + */ + supportsFullScreen() { + return typeof this.el_.webkitEnterFullScreen === 'function'; + } + + /** + * Request that the `HTML5` Tech enter fullscreen. + */ + enterFullScreen() { + const video = this.el_; + if (video.paused && video.networkState <= video.HAVE_METADATA) { + // attempt to prime the video element for programmatic access + // this isn't necessary on the desktop but shouldn't hurt + silencePromise(this.el_.play()); + + // playing and pausing synchronously during the transition to fullscreen + // can get iOS ~6.1 devices into a play/pause loop + this.setTimeout(function () { + video.pause(); + try { + video.webkitEnterFullScreen(); + } catch (e) { + this.trigger('fullscreenerror', e); + } + }, 0); + } else { + try { + video.webkitEnterFullScreen(); + } catch (e) { + this.trigger('fullscreenerror', e); + } + } + } + + /** + * Request that the `HTML5` Tech exit fullscreen. + */ + exitFullScreen() { + if (!this.el_.webkitDisplayingFullscreen) { + this.trigger('fullscreenerror', new Error('The video is not fullscreen')); + return; + } + this.el_.webkitExitFullScreen(); + } + + /** + * Create a floating video window always on top of other windows so that users may + * continue consuming media while they interact with other content sites, or + * applications on their device. + * + * @see [Spec]{@link https://wicg.github.io/picture-in-picture} + * + * @return {Promise} + * A promise with a Picture-in-Picture window. + */ + requestPictureInPicture() { + return this.el_.requestPictureInPicture(); + } + + /** + * Native requestVideoFrameCallback if supported by browser/tech, or fallback + * Don't use rVCF on Safari when DRM is playing, as it doesn't fire + * Needs to be checked later than the constructor + * This will be a false positive for clear sources loaded after a Fairplay source + * + * @param {function} cb function to call + * @return {number} id of request + */ + requestVideoFrameCallback(cb) { + if (this.featuresVideoFrameCallback && !this.el_.webkitKeys) { + return this.el_.requestVideoFrameCallback(cb); + } + return super.requestVideoFrameCallback(cb); + } + + /** + * Native or fallback requestVideoFrameCallback + * + * @param {number} id request id to cancel + */ + cancelVideoFrameCallback(id) { + if (this.featuresVideoFrameCallback && !this.el_.webkitKeys) { + this.el_.cancelVideoFrameCallback(id); + } else { + super.cancelVideoFrameCallback(id); + } + } + + /** + * A getter/setter for the `Html5` Tech's source object. + * > Note: Please use {@link Html5#setSource} + * + * @param {Tech~SourceObject} [src] + * The source object you want to set on the `HTML5` techs element. + * + * @return {Tech~SourceObject|undefined} + * - The current source object when a source is not passed in. + * - undefined when setting + * + * @deprecated Since version 5. + */ + src(src) { + if (src === undefined) { + return this.el_.src; + } + + // Setting src through `src` instead of `setSrc` will be deprecated + this.setSrc(src); + } + + /** + * Reset the tech by removing all sources and then calling + * {@link Html5.resetMediaElement}. + */ + reset() { + Html5.resetMediaElement(this.el_); + } + + /** + * Get the current source on the `HTML5` Tech. Falls back to returning the source from + * the HTML5 media element. + * + * @return {Tech~SourceObject} + * The current source object from the HTML5 tech. With a fallback to the + * elements source. + */ + currentSrc() { + if (this.currentSource_) { + return this.currentSource_.src; + } + return this.el_.currentSrc; + } + + /** + * Set controls attribute for the HTML5 media Element. + * + * @param {string} val + * Value to set the controls attribute to + */ + setControls(val) { + this.el_.controls = !!val; + } + + /** + * Create and returns a remote {@link TextTrack} object. + * + * @param {string} kind + * `TextTrack` kind (subtitles, captions, descriptions, chapters, or metadata) + * + * @param {string} [label] + * Label to identify the text track + * + * @param {string} [language] + * Two letter language abbreviation + * + * @return {TextTrack} + * The TextTrack that gets created. + */ + addTextTrack(kind, label, language) { + if (!this.featuresNativeTextTracks) { + return super.addTextTrack(kind, label, language); + } + return this.el_.addTextTrack(kind, label, language); + } + + /** + * Creates either native TextTrack or an emulated TextTrack depending + * on the value of `featuresNativeTextTracks` + * + * @param {Object} options + * The object should contain the options to initialize the TextTrack with. + * + * @param {string} [options.kind] + * `TextTrack` kind (subtitles, captions, descriptions, chapters, or metadata). + * + * @param {string} [options.label] + * Label to identify the text track + * + * @param {string} [options.language] + * Two letter language abbreviation. + * + * @param {boolean} [options.default] + * Default this track to on. + * + * @param {string} [options.id] + * The internal id to assign this track. + * + * @param {string} [options.src] + * A source url for the track. + * + * @return {HTMLTrackElement} + * The track element that gets created. + */ + createRemoteTextTrack(options) { + if (!this.featuresNativeTextTracks) { + return super.createRemoteTextTrack(options); + } + const htmlTrackElement = document.createElement('track'); + if (options.kind) { + htmlTrackElement.kind = options.kind; + } + if (options.label) { + htmlTrackElement.label = options.label; + } + if (options.language || options.srclang) { + htmlTrackElement.srclang = options.language || options.srclang; + } + if (options.default) { + htmlTrackElement.default = options.default; + } + if (options.id) { + htmlTrackElement.id = options.id; + } + if (options.src) { + htmlTrackElement.src = options.src; + } + return htmlTrackElement; + } + + /** + * Creates a remote text track object and returns an html track element. + * + * @param {Object} options The object should contain values for + * kind, language, label, and src (location of the WebVTT file) + * @param {boolean} [manualCleanup=false] if set to true, the TextTrack + * will not be removed from the TextTrackList and HtmlTrackElementList + * after a source change + * @return {HTMLTrackElement} An Html Track Element. + * This can be an emulated {@link HTMLTrackElement} or a native one. + * + */ + addRemoteTextTrack(options, manualCleanup) { + const htmlTrackElement = super.addRemoteTextTrack(options, manualCleanup); + if (this.featuresNativeTextTracks) { + this.el().appendChild(htmlTrackElement); + } + return htmlTrackElement; + } + + /** + * Remove remote `TextTrack` from `TextTrackList` object + * + * @param {TextTrack} track + * `TextTrack` object to remove + */ + removeRemoteTextTrack(track) { + super.removeRemoteTextTrack(track); + if (this.featuresNativeTextTracks) { + const tracks = this.$$('track'); + let i = tracks.length; + while (i--) { + if (track === tracks[i] || track === tracks[i].track) { + this.el().removeChild(tracks[i]); + } + } + } + } + + /** + * Gets available media playback quality metrics as specified by the W3C's Media + * Playback Quality API. + * + * @see [Spec]{@link https://wicg.github.io/media-playback-quality} + * + * @return {Object} + * An object with supported media playback quality metrics + */ + getVideoPlaybackQuality() { + if (typeof this.el().getVideoPlaybackQuality === 'function') { + return this.el().getVideoPlaybackQuality(); + } + const videoPlaybackQuality = {}; + if (typeof this.el().webkitDroppedFrameCount !== 'undefined' && typeof this.el().webkitDecodedFrameCount !== 'undefined') { + videoPlaybackQuality.droppedVideoFrames = this.el().webkitDroppedFrameCount; + videoPlaybackQuality.totalVideoFrames = this.el().webkitDecodedFrameCount; + } + if (window$1.performance) { + videoPlaybackQuality.creationTime = window$1.performance.now(); + } + return videoPlaybackQuality; + } +} + +/* HTML5 Support Testing ---------------------------------------------------- */ + +/** + * Element for testing browser HTML5 media capabilities + * + * @type {Element} + * @constant + * @private + */ +defineLazyProperty(Html5, 'TEST_VID', function () { + if (!isReal()) { + return; + } + const video = document.createElement('video'); + const track = document.createElement('track'); + track.kind = 'captions'; + track.srclang = 'en'; + track.label = 'English'; + video.appendChild(track); + return video; +}); + +/** + * Check if HTML5 media is supported by this browser/device. + * + * @return {boolean} + * - True if HTML5 media is supported. + * - False if HTML5 media is not supported. + */ +Html5.isSupported = function () { + // IE with no Media Player is a LIAR! (#984) + try { + Html5.TEST_VID.volume = 0.5; + } catch (e) { + return false; + } + return !!(Html5.TEST_VID && Html5.TEST_VID.canPlayType); +}; + +/** + * Check if the tech can support the given type + * + * @param {string} type + * The mimetype to check + * @return {string} 'probably', 'maybe', or '' (empty string) + */ +Html5.canPlayType = function (type) { + return Html5.TEST_VID.canPlayType(type); +}; + +/** + * Check if the tech can support the given source + * + * @param {Object} srcObj + * The source object + * @param {Object} options + * The options passed to the tech + * @return {string} 'probably', 'maybe', or '' (empty string) + */ +Html5.canPlaySource = function (srcObj, options) { + return Html5.canPlayType(srcObj.type); +}; + +/** + * Check if the volume can be changed in this browser/device. + * Volume cannot be changed in a lot of mobile devices. + * Specifically, it can't be changed from 1 on iOS. + * + * @return {boolean} + * - True if volume can be controlled + * - False otherwise + */ +Html5.canControlVolume = function () { + // IE will error if Windows Media Player not installed #3315 + try { + const volume = Html5.TEST_VID.volume; + Html5.TEST_VID.volume = volume / 2 + 0.1; + const canControl = volume !== Html5.TEST_VID.volume; + + // With the introduction of iOS 15, there are cases where the volume is read as + // changed but reverts back to its original state at the start of the next tick. + // To determine whether volume can be controlled on iOS, + // a timeout is set and the volume is checked asynchronously. + // Since `features` doesn't currently work asynchronously, the value is manually set. + if (canControl && IS_IOS) { + window$1.setTimeout(() => { + if (Html5 && Html5.prototype) { + Html5.prototype.featuresVolumeControl = volume !== Html5.TEST_VID.volume; + } + }); + + // default iOS to false, which will be updated in the timeout above. + return false; + } + return canControl; + } catch (e) { + return false; + } +}; + +/** + * Check if the volume can be muted in this browser/device. + * Some devices, e.g. iOS, don't allow changing volume + * but permits muting/unmuting. + * + * @return {boolean} + * - True if volume can be muted + * - False otherwise + */ +Html5.canMuteVolume = function () { + try { + const muted = Html5.TEST_VID.muted; + + // in some versions of iOS muted property doesn't always + // work, so we want to set both property and attribute + Html5.TEST_VID.muted = !muted; + if (Html5.TEST_VID.muted) { + setAttribute(Html5.TEST_VID, 'muted', 'muted'); + } else { + removeAttribute(Html5.TEST_VID, 'muted', 'muted'); + } + return muted !== Html5.TEST_VID.muted; + } catch (e) { + return false; + } +}; + +/** + * Check if the playback rate can be changed in this browser/device. + * + * @return {boolean} + * - True if playback rate can be controlled + * - False otherwise + */ +Html5.canControlPlaybackRate = function () { + // Playback rate API is implemented in Android Chrome, but doesn't do anything + // https://github.com/videojs/video.js/issues/3180 + if (IS_ANDROID && IS_CHROME && CHROME_VERSION < 58) { + return false; + } + // IE will error if Windows Media Player not installed #3315 + try { + const playbackRate = Html5.TEST_VID.playbackRate; + Html5.TEST_VID.playbackRate = playbackRate / 2 + 0.1; + return playbackRate !== Html5.TEST_VID.playbackRate; + } catch (e) { + return false; + } +}; + +/** + * Check if we can override a video/audio elements attributes, with + * Object.defineProperty. + * + * @return {boolean} + * - True if builtin attributes can be overridden + * - False otherwise + */ +Html5.canOverrideAttributes = function () { + // if we cannot overwrite the src/innerHTML property, there is no support + // iOS 7 safari for instance cannot do this. + try { + const noop = () => {}; + Object.defineProperty(document.createElement('video'), 'src', { + get: noop, + set: noop + }); + Object.defineProperty(document.createElement('audio'), 'src', { + get: noop, + set: noop + }); + Object.defineProperty(document.createElement('video'), 'innerHTML', { + get: noop, + set: noop + }); + Object.defineProperty(document.createElement('audio'), 'innerHTML', { + get: noop, + set: noop + }); + } catch (e) { + return false; + } + return true; +}; + +/** + * Check to see if native `TextTrack`s are supported by this browser/device. + * + * @return {boolean} + * - True if native `TextTrack`s are supported. + * - False otherwise + */ +Html5.supportsNativeTextTracks = function () { + return IS_ANY_SAFARI || IS_IOS && IS_CHROME; +}; + +/** + * Check to see if native `VideoTrack`s are supported by this browser/device + * + * @return {boolean} + * - True if native `VideoTrack`s are supported. + * - False otherwise + */ +Html5.supportsNativeVideoTracks = function () { + return !!(Html5.TEST_VID && Html5.TEST_VID.videoTracks); +}; + +/** + * Check to see if native `AudioTrack`s are supported by this browser/device + * + * @return {boolean} + * - True if native `AudioTrack`s are supported. + * - False otherwise + */ +Html5.supportsNativeAudioTracks = function () { + return !!(Html5.TEST_VID && Html5.TEST_VID.audioTracks); +}; + +/** + * An array of events available on the Html5 tech. + * + * @private + * @type {Array} + */ +Html5.Events = ['loadstart', 'suspend', 'abort', 'error', 'emptied', 'stalled', 'loadedmetadata', 'loadeddata', 'canplay', 'canplaythrough', 'playing', 'waiting', 'seeking', 'seeked', 'ended', 'durationchange', 'timeupdate', 'progress', 'play', 'pause', 'ratechange', 'resize', 'volumechange']; + +/** + * Boolean indicating whether the `Tech` supports volume control. + * + * @type {boolean} + * @default {@link Html5.canControlVolume} + */ +/** + * Boolean indicating whether the `Tech` supports muting volume. + * + * @type {boolean} + * @default {@link Html5.canMuteVolume} + */ + +/** + * Boolean indicating whether the `Tech` supports changing the speed at which the media + * plays. Examples: + * - Set player to play 2x (twice) as fast + * - Set player to play 0.5x (half) as fast + * + * @type {boolean} + * @default {@link Html5.canControlPlaybackRate} + */ + +/** + * Boolean indicating whether the `Tech` supports the `sourceset` event. + * + * @type {boolean} + * @default + */ +/** + * Boolean indicating whether the `HTML5` tech currently supports native `TextTrack`s. + * + * @type {boolean} + * @default {@link Html5.supportsNativeTextTracks} + */ +/** + * Boolean indicating whether the `HTML5` tech currently supports native `VideoTrack`s. + * + * @type {boolean} + * @default {@link Html5.supportsNativeVideoTracks} + */ +/** + * Boolean indicating whether the `HTML5` tech currently supports native `AudioTrack`s. + * + * @type {boolean} + * @default {@link Html5.supportsNativeAudioTracks} + */ +[['featuresMuteControl', 'canMuteVolume'], ['featuresPlaybackRate', 'canControlPlaybackRate'], ['featuresSourceset', 'canOverrideAttributes'], ['featuresNativeTextTracks', 'supportsNativeTextTracks'], ['featuresNativeVideoTracks', 'supportsNativeVideoTracks'], ['featuresNativeAudioTracks', 'supportsNativeAudioTracks']].forEach(function ([key, fn]) { + defineLazyProperty(Html5.prototype, key, () => Html5[fn](), true); +}); +Html5.prototype.featuresVolumeControl = Html5.canControlVolume(); + +/** + * Boolean indicating whether the `HTML5` tech currently supports the media element + * moving in the DOM. iOS breaks if you move the media element, so this is set this to + * false there. Everywhere else this should be true. + * + * @type {boolean} + * @default + */ +Html5.prototype.movingMediaElementInDOM = !IS_IOS; + +// TODO: Previous comment: No longer appears to be used. Can probably be removed. +// Is this true? +/** + * Boolean indicating whether the `HTML5` tech currently supports automatic media resize + * when going into fullscreen. + * + * @type {boolean} + * @default + */ +Html5.prototype.featuresFullscreenResize = true; + +/** + * Boolean indicating whether the `HTML5` tech currently supports the progress event. + * If this is false, manual `progress` events will be triggered instead. + * + * @type {boolean} + * @default + */ +Html5.prototype.featuresProgressEvents = true; + +/** + * Boolean indicating whether the `HTML5` tech currently supports the timeupdate event. + * If this is false, manual `timeupdate` events will be triggered instead. + * + * @default + */ +Html5.prototype.featuresTimeupdateEvents = true; + +/** + * Whether the HTML5 el supports `requestVideoFrameCallback` + * + * @type {boolean} + */ +Html5.prototype.featuresVideoFrameCallback = !!(Html5.TEST_VID && Html5.TEST_VID.requestVideoFrameCallback); +Html5.disposeMediaElement = function (el) { + if (!el) { + return; + } + if (el.parentNode) { + el.parentNode.removeChild(el); + } + + // remove any child track or source nodes to prevent their loading + while (el.hasChildNodes()) { + el.removeChild(el.firstChild); + } + + // remove any src reference. not setting `src=''` because that causes a warning + // in firefox + el.removeAttribute('src'); + + // force the media element to update its loading state by calling load() + // however IE on Windows 7N has a bug that throws an error so need a try/catch (#793) + if (typeof el.load === 'function') { + // wrapping in an iife so it's not deoptimized (#1060#discussion_r10324473) + (function () { + try { + el.load(); + } catch (e) { + // not supported + } + })(); + } +}; +Html5.resetMediaElement = function (el) { + if (!el) { + return; + } + const sources = el.querySelectorAll('source'); + let i = sources.length; + while (i--) { + el.removeChild(sources[i]); + } + + // remove any src reference. + // not setting `src=''` because that throws an error + el.removeAttribute('src'); + if (typeof el.load === 'function') { + // wrapping in an iife so it's not deoptimized (#1060#discussion_r10324473) + (function () { + try { + el.load(); + } catch (e) { + // satisfy linter + } + })(); + } +}; + +/* Native HTML5 element property wrapping ----------------------------------- */ +// Wrap native boolean attributes with getters that check both property and attribute +// The list is as followed: +// muted, defaultMuted, autoplay, controls, loop, playsinline +[ +/** + * Get the value of `muted` from the media element. `muted` indicates + * that the volume for the media should be set to silent. This does not actually change + * the `volume` attribute. + * + * @method Html5#muted + * @return {boolean} + * - True if the value of `volume` should be ignored and the audio set to silent. + * - False if the value of `volume` should be used. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-muted} + */ +'muted', +/** + * Get the value of `defaultMuted` from the media element. `defaultMuted` indicates + * whether the media should start muted or not. Only changes the default state of the + * media. `muted` and `defaultMuted` can have different values. {@link Html5#muted} indicates the + * current state. + * + * @method Html5#defaultMuted + * @return {boolean} + * - The value of `defaultMuted` from the media element. + * - True indicates that the media should start muted. + * - False indicates that the media should not start muted + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-defaultmuted} + */ +'defaultMuted', +/** + * Get the value of `autoplay` from the media element. `autoplay` indicates + * that the media should start to play as soon as the page is ready. + * + * @method Html5#autoplay + * @return {boolean} + * - The value of `autoplay` from the media element. + * - True indicates that the media should start as soon as the page loads. + * - False indicates that the media should not start as soon as the page loads. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#attr-media-autoplay} + */ +'autoplay', +/** + * Get the value of `controls` from the media element. `controls` indicates + * whether the native media controls should be shown or hidden. + * + * @method Html5#controls + * @return {boolean} + * - The value of `controls` from the media element. + * - True indicates that native controls should be showing. + * - False indicates that native controls should be hidden. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#attr-media-controls} + */ +'controls', +/** + * Get the value of `loop` from the media element. `loop` indicates + * that the media should return to the start of the media and continue playing once + * it reaches the end. + * + * @method Html5#loop + * @return {boolean} + * - The value of `loop` from the media element. + * - True indicates that playback should seek back to start once + * the end of a media is reached. + * - False indicates that playback should not loop back to the start when the + * end of the media is reached. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#attr-media-loop} + */ +'loop', +/** + * Get the value of `playsinline` from the media element. `playsinline` indicates + * to the browser that non-fullscreen playback is preferred when fullscreen + * playback is the native default, such as in iOS Safari. + * + * @method Html5#playsinline + * @return {boolean} + * - The value of `playsinline` from the media element. + * - True indicates that the media should play inline. + * - False indicates that the media should not play inline. + * + * @see [Spec]{@link https://html.spec.whatwg.org/#attr-video-playsinline} + */ +'playsinline'].forEach(function (prop) { + Html5.prototype[prop] = function () { + return this.el_[prop] || this.el_.hasAttribute(prop); + }; +}); + +// Wrap native boolean attributes with setters that set both property and attribute +// The list is as followed: +// setMuted, setDefaultMuted, setAutoplay, setLoop, setPlaysinline +// setControls is special-cased above +[ +/** + * Set the value of `muted` on the media element. `muted` indicates that the current + * audio level should be silent. + * + * @method Html5#setMuted + * @param {boolean} muted + * - True if the audio should be set to silent + * - False otherwise + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-muted} + */ +'muted', +/** + * Set the value of `defaultMuted` on the media element. `defaultMuted` indicates that the current + * audio level should be silent, but will only effect the muted level on initial playback.. + * + * @method Html5.prototype.setDefaultMuted + * @param {boolean} defaultMuted + * - True if the audio should be set to silent + * - False otherwise + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-defaultmuted} + */ +'defaultMuted', +/** + * Set the value of `autoplay` on the media element. `autoplay` indicates + * that the media should start to play as soon as the page is ready. + * + * @method Html5#setAutoplay + * @param {boolean} autoplay + * - True indicates that the media should start as soon as the page loads. + * - False indicates that the media should not start as soon as the page loads. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#attr-media-autoplay} + */ +'autoplay', +/** + * Set the value of `loop` on the media element. `loop` indicates + * that the media should return to the start of the media and continue playing once + * it reaches the end. + * + * @method Html5#setLoop + * @param {boolean} loop + * - True indicates that playback should seek back to start once + * the end of a media is reached. + * - False indicates that playback should not loop back to the start when the + * end of the media is reached. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#attr-media-loop} + */ +'loop', +/** + * Set the value of `playsinline` from the media element. `playsinline` indicates + * to the browser that non-fullscreen playback is preferred when fullscreen + * playback is the native default, such as in iOS Safari. + * + * @method Html5#setPlaysinline + * @param {boolean} playsinline + * - True indicates that the media should play inline. + * - False indicates that the media should not play inline. + * + * @see [Spec]{@link https://html.spec.whatwg.org/#attr-video-playsinline} + */ +'playsinline'].forEach(function (prop) { + Html5.prototype['set' + toTitleCase$1(prop)] = function (v) { + this.el_[prop] = v; + if (v) { + this.el_.setAttribute(prop, prop); + } else { + this.el_.removeAttribute(prop); + } + }; +}); + +// Wrap native properties with a getter +// The list is as followed +// paused, currentTime, buffered, volume, poster, preload, error, seeking +// seekable, ended, playbackRate, defaultPlaybackRate, disablePictureInPicture +// played, networkState, readyState, videoWidth, videoHeight, crossOrigin +[ +/** + * Get the value of `paused` from the media element. `paused` indicates whether the media element + * is currently paused or not. + * + * @method Html5#paused + * @return {boolean} + * The value of `paused` from the media element. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-paused} + */ +'paused', +/** + * Get the value of `currentTime` from the media element. `currentTime` indicates + * the current second that the media is at in playback. + * + * @method Html5#currentTime + * @return {number} + * The value of `currentTime` from the media element. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-currenttime} + */ +'currentTime', +/** + * Get the value of `buffered` from the media element. `buffered` is a `TimeRange` + * object that represents the parts of the media that are already downloaded and + * available for playback. + * + * @method Html5#buffered + * @return {TimeRange} + * The value of `buffered` from the media element. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-buffered} + */ +'buffered', +/** + * Get the value of `volume` from the media element. `volume` indicates + * the current playback volume of audio for a media. `volume` will be a value from 0 + * (silent) to 1 (loudest and default). + * + * @method Html5#volume + * @return {number} + * The value of `volume` from the media element. Value will be between 0-1. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-a-volume} + */ +'volume', +/** + * Get the value of `poster` from the media element. `poster` indicates + * that the url of an image file that can/will be shown when no media data is available. + * + * @method Html5#poster + * @return {string} + * The value of `poster` from the media element. Value will be a url to an + * image. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#attr-video-poster} + */ +'poster', +/** + * Get the value of `preload` from the media element. `preload` indicates + * what should download before the media is interacted with. It can have the following + * values: + * - none: nothing should be downloaded + * - metadata: poster and the first few frames of the media may be downloaded to get + * media dimensions and other metadata + * - auto: allow the media and metadata for the media to be downloaded before + * interaction + * + * @method Html5#preload + * @return {string} + * The value of `preload` from the media element. Will be 'none', 'metadata', + * or 'auto'. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#attr-media-preload} + */ +'preload', +/** + * Get the value of the `error` from the media element. `error` indicates any + * MediaError that may have occurred during playback. If error returns null there is no + * current error. + * + * @method Html5#error + * @return {MediaError|null} + * The value of `error` from the media element. Will be `MediaError` if there + * is a current error and null otherwise. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-error} + */ +'error', +/** + * Get the value of `seeking` from the media element. `seeking` indicates whether the + * media is currently seeking to a new position or not. + * + * @method Html5#seeking + * @return {boolean} + * - The value of `seeking` from the media element. + * - True indicates that the media is currently seeking to a new position. + * - False indicates that the media is not seeking to a new position at this time. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-seeking} + */ +'seeking', +/** + * Get the value of `seekable` from the media element. `seekable` returns a + * `TimeRange` object indicating ranges of time that can currently be `seeked` to. + * + * @method Html5#seekable + * @return {TimeRange} + * The value of `seekable` from the media element. A `TimeRange` object + * indicating the current ranges of time that can be seeked to. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-seekable} + */ +'seekable', +/** + * Get the value of `ended` from the media element. `ended` indicates whether + * the media has reached the end or not. + * + * @method Html5#ended + * @return {boolean} + * - The value of `ended` from the media element. + * - True indicates that the media has ended. + * - False indicates that the media has not ended. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-ended} + */ +'ended', +/** + * Get the value of `playbackRate` from the media element. `playbackRate` indicates + * the rate at which the media is currently playing back. Examples: + * - if playbackRate is set to 2, media will play twice as fast. + * - if playbackRate is set to 0.5, media will play half as fast. + * + * @method Html5#playbackRate + * @return {number} + * The value of `playbackRate` from the media element. A number indicating + * the current playback speed of the media, where 1 is normal speed. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-playbackrate} + */ +'playbackRate', +/** + * Get the value of `defaultPlaybackRate` from the media element. `defaultPlaybackRate` indicates + * the rate at which the media is currently playing back. This value will not indicate the current + * `playbackRate` after playback has started, use {@link Html5#playbackRate} for that. + * + * Examples: + * - if defaultPlaybackRate is set to 2, media will play twice as fast. + * - if defaultPlaybackRate is set to 0.5, media will play half as fast. + * + * @method Html5.prototype.defaultPlaybackRate + * @return {number} + * The value of `defaultPlaybackRate` from the media element. A number indicating + * the current playback speed of the media, where 1 is normal speed. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-playbackrate} + */ +'defaultPlaybackRate', +/** + * Get the value of 'disablePictureInPicture' from the video element. + * + * @method Html5#disablePictureInPicture + * @return {boolean} value + * - The value of `disablePictureInPicture` from the video element. + * - True indicates that the video can't be played in Picture-In-Picture mode + * - False indicates that the video can be played in Picture-In-Picture mode + * + * @see [Spec]{@link https://w3c.github.io/picture-in-picture/#disable-pip} + */ +'disablePictureInPicture', +/** + * Get the value of `played` from the media element. `played` returns a `TimeRange` + * object representing points in the media timeline that have been played. + * + * @method Html5#played + * @return {TimeRange} + * The value of `played` from the media element. A `TimeRange` object indicating + * the ranges of time that have been played. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-played} + */ +'played', +/** + * Get the value of `networkState` from the media element. `networkState` indicates + * the current network state. It returns an enumeration from the following list: + * - 0: NETWORK_EMPTY + * - 1: NETWORK_IDLE + * - 2: NETWORK_LOADING + * - 3: NETWORK_NO_SOURCE + * + * @method Html5#networkState + * @return {number} + * The value of `networkState` from the media element. This will be a number + * from the list in the description. + * + * @see [Spec] {@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-networkstate} + */ +'networkState', +/** + * Get the value of `readyState` from the media element. `readyState` indicates + * the current state of the media element. It returns an enumeration from the + * following list: + * - 0: HAVE_NOTHING + * - 1: HAVE_METADATA + * - 2: HAVE_CURRENT_DATA + * - 3: HAVE_FUTURE_DATA + * - 4: HAVE_ENOUGH_DATA + * + * @method Html5#readyState + * @return {number} + * The value of `readyState` from the media element. This will be a number + * from the list in the description. + * + * @see [Spec] {@link https://www.w3.org/TR/html5/embedded-content-0.html#ready-states} + */ +'readyState', +/** + * Get the value of `videoWidth` from the video element. `videoWidth` indicates + * the current width of the video in css pixels. + * + * @method Html5#videoWidth + * @return {number} + * The value of `videoWidth` from the video element. This will be a number + * in css pixels. + * + * @see [Spec] {@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-video-videowidth} + */ +'videoWidth', +/** + * Get the value of `videoHeight` from the video element. `videoHeight` indicates + * the current height of the video in css pixels. + * + * @method Html5#videoHeight + * @return {number} + * The value of `videoHeight` from the video element. This will be a number + * in css pixels. + * + * @see [Spec] {@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-video-videowidth} + */ +'videoHeight', +/** + * Get the value of `crossOrigin` from the media element. `crossOrigin` indicates + * to the browser that should sent the cookies along with the requests for the + * different assets/playlists + * + * @method Html5#crossOrigin + * @return {string} + * - anonymous indicates that the media should not sent cookies. + * - use-credentials indicates that the media should sent cookies along the requests. + * + * @see [Spec]{@link https://html.spec.whatwg.org/#attr-media-crossorigin} + */ +'crossOrigin'].forEach(function (prop) { + Html5.prototype[prop] = function () { + return this.el_[prop]; + }; +}); + +// Wrap native properties with a setter in this format: +// set + toTitleCase(name) +// The list is as follows: +// setVolume, setSrc, setPoster, setPreload, setPlaybackRate, setDefaultPlaybackRate, +// setDisablePictureInPicture, setCrossOrigin +[ +/** + * Set the value of `volume` on the media element. `volume` indicates the current + * audio level as a percentage in decimal form. This means that 1 is 100%, 0.5 is 50%, and + * so on. + * + * @method Html5#setVolume + * @param {number} percentAsDecimal + * The volume percent as a decimal. Valid range is from 0-1. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-a-volume} + */ +'volume', +/** + * Set the value of `src` on the media element. `src` indicates the current + * {@link Tech~SourceObject} for the media. + * + * @method Html5#setSrc + * @param {Tech~SourceObject} src + * The source object to set as the current source. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-src} + */ +'src', +/** + * Set the value of `poster` on the media element. `poster` is the url to + * an image file that can/will be shown when no media data is available. + * + * @method Html5#setPoster + * @param {string} poster + * The url to an image that should be used as the `poster` for the media + * element. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#attr-media-poster} + */ +'poster', +/** + * Set the value of `preload` on the media element. `preload` indicates + * what should download before the media is interacted with. It can have the following + * values: + * - none: nothing should be downloaded + * - metadata: poster and the first few frames of the media may be downloaded to get + * media dimensions and other metadata + * - auto: allow the media and metadata for the media to be downloaded before + * interaction + * + * @method Html5#setPreload + * @param {string} preload + * The value of `preload` to set on the media element. Must be 'none', 'metadata', + * or 'auto'. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#attr-media-preload} + */ +'preload', +/** + * Set the value of `playbackRate` on the media element. `playbackRate` indicates + * the rate at which the media should play back. Examples: + * - if playbackRate is set to 2, media will play twice as fast. + * - if playbackRate is set to 0.5, media will play half as fast. + * + * @method Html5#setPlaybackRate + * @return {number} + * The value of `playbackRate` from the media element. A number indicating + * the current playback speed of the media, where 1 is normal speed. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-playbackrate} + */ +'playbackRate', +/** + * Set the value of `defaultPlaybackRate` on the media element. `defaultPlaybackRate` indicates + * the rate at which the media should play back upon initial startup. Changing this value + * after a video has started will do nothing. Instead you should used {@link Html5#setPlaybackRate}. + * + * Example Values: + * - if playbackRate is set to 2, media will play twice as fast. + * - if playbackRate is set to 0.5, media will play half as fast. + * + * @method Html5.prototype.setDefaultPlaybackRate + * @return {number} + * The value of `defaultPlaybackRate` from the media element. A number indicating + * the current playback speed of the media, where 1 is normal speed. + * + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-defaultplaybackrate} + */ +'defaultPlaybackRate', +/** + * Prevents the browser from suggesting a Picture-in-Picture context menu + * or to request Picture-in-Picture automatically in some cases. + * + * @method Html5#setDisablePictureInPicture + * @param {boolean} value + * The true value will disable Picture-in-Picture mode. + * + * @see [Spec]{@link https://w3c.github.io/picture-in-picture/#disable-pip} + */ +'disablePictureInPicture', +/** + * Set the value of `crossOrigin` from the media element. `crossOrigin` indicates + * to the browser that should sent the cookies along with the requests for the + * different assets/playlists + * + * @method Html5#setCrossOrigin + * @param {string} crossOrigin + * - anonymous indicates that the media should not sent cookies. + * - use-credentials indicates that the media should sent cookies along the requests. + * + * @see [Spec]{@link https://html.spec.whatwg.org/#attr-media-crossorigin} + */ +'crossOrigin'].forEach(function (prop) { + Html5.prototype['set' + toTitleCase$1(prop)] = function (v) { + this.el_[prop] = v; + }; +}); + +// wrap native functions with a function +// The list is as follows: +// pause, load, play +[ +/** + * A wrapper around the media elements `pause` function. This will call the `HTML5` + * media elements `pause` function. + * + * @method Html5#pause + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-pause} + */ +'pause', +/** + * A wrapper around the media elements `load` function. This will call the `HTML5`s + * media element `load` function. + * + * @method Html5#load + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-load} + */ +'load', +/** + * A wrapper around the media elements `play` function. This will call the `HTML5`s + * media element `play` function. + * + * @method Html5#play + * @see [Spec]{@link https://www.w3.org/TR/html5/embedded-content-0.html#dom-media-play} + */ +'play'].forEach(function (prop) { + Html5.prototype[prop] = function () { + return this.el_[prop](); + }; +}); +Tech.withSourceHandlers(Html5); + +/** + * Native source handler for Html5, simply passes the source to the media element. + * + * @property {Tech~SourceObject} source + * The source object + * + * @property {Html5} tech + * The instance of the HTML5 tech. + */ +Html5.nativeSourceHandler = {}; + +/** + * Check if the media element can play the given mime type. + * + * @param {string} type + * The mimetype to check + * + * @return {string} + * 'probably', 'maybe', or '' (empty string) + */ +Html5.nativeSourceHandler.canPlayType = function (type) { + // IE without MediaPlayer throws an error (#519) + try { + return Html5.TEST_VID.canPlayType(type); + } catch (e) { + return ''; + } +}; + +/** + * Check if the media element can handle a source natively. + * + * @param {Tech~SourceObject} source + * The source object + * + * @param {Object} [options] + * Options to be passed to the tech. + * + * @return {string} + * 'probably', 'maybe', or '' (empty string). + */ +Html5.nativeSourceHandler.canHandleSource = function (source, options) { + // If a type was provided we should rely on that + if (source.type) { + return Html5.nativeSourceHandler.canPlayType(source.type); + + // If no type, fall back to checking 'video/[EXTENSION]' + } else if (source.src) { + const ext = getFileExtension(source.src); + return Html5.nativeSourceHandler.canPlayType(`video/${ext}`); + } + return ''; +}; + +/** + * Pass the source to the native media element. + * + * @param {Tech~SourceObject} source + * The source object + * + * @param {Html5} tech + * The instance of the Html5 tech + * + * @param {Object} [options] + * The options to pass to the source + */ +Html5.nativeSourceHandler.handleSource = function (source, tech, options) { + tech.setSrc(source.src); +}; + +/** + * A noop for the native dispose function, as cleanup is not needed. + */ +Html5.nativeSourceHandler.dispose = function () {}; + +// Register the native source handler +Html5.registerSourceHandler(Html5.nativeSourceHandler); +Tech.registerTech('Html5', Html5); + +/** + * @file player.js + */ + +// The following tech events are simply re-triggered +// on the player when they happen +const TECH_EVENTS_RETRIGGER = [ +/** + * Fired while the user agent is downloading media data. + * + * @event Player#progress + * @type {Event} + */ +/** + * Retrigger the `progress` event that was triggered by the {@link Tech}. + * + * @private + * @method Player#handleTechProgress_ + * @fires Player#progress + * @listens Tech#progress + */ +'progress', +/** + * Fires when the loading of an audio/video is aborted. + * + * @event Player#abort + * @type {Event} + */ +/** + * Retrigger the `abort` event that was triggered by the {@link Tech}. + * + * @private + * @method Player#handleTechAbort_ + * @fires Player#abort + * @listens Tech#abort + */ +'abort', +/** + * Fires when the browser is intentionally not getting media data. + * + * @event Player#suspend + * @type {Event} + */ +/** + * Retrigger the `suspend` event that was triggered by the {@link Tech}. + * + * @private + * @method Player#handleTechSuspend_ + * @fires Player#suspend + * @listens Tech#suspend + */ +'suspend', +/** + * Fires when the current playlist is empty. + * + * @event Player#emptied + * @type {Event} + */ +/** + * Retrigger the `emptied` event that was triggered by the {@link Tech}. + * + * @private + * @method Player#handleTechEmptied_ + * @fires Player#emptied + * @listens Tech#emptied + */ +'emptied', +/** + * Fires when the browser is trying to get media data, but data is not available. + * + * @event Player#stalled + * @type {Event} + */ +/** + * Retrigger the `stalled` event that was triggered by the {@link Tech}. + * + * @private + * @method Player#handleTechStalled_ + * @fires Player#stalled + * @listens Tech#stalled + */ +'stalled', +/** + * Fires when the browser has loaded meta data for the audio/video. + * + * @event Player#loadedmetadata + * @type {Event} + */ +/** + * Retrigger the `loadedmetadata` event that was triggered by the {@link Tech}. + * + * @private + * @method Player#handleTechLoadedmetadata_ + * @fires Player#loadedmetadata + * @listens Tech#loadedmetadata + */ +'loadedmetadata', +/** + * Fires when the browser has loaded the current frame of the audio/video. + * + * @event Player#loadeddata + * @type {event} + */ +/** + * Retrigger the `loadeddata` event that was triggered by the {@link Tech}. + * + * @private + * @method Player#handleTechLoaddeddata_ + * @fires Player#loadeddata + * @listens Tech#loadeddata + */ +'loadeddata', +/** + * Fires when the current playback position has changed. + * + * @event Player#timeupdate + * @type {event} + */ +/** + * Retrigger the `timeupdate` event that was triggered by the {@link Tech}. + * + * @private + * @method Player#handleTechTimeUpdate_ + * @fires Player#timeupdate + * @listens Tech#timeupdate + */ +'timeupdate', +/** + * Fires when the video's intrinsic dimensions change + * + * @event Player#resize + * @type {event} + */ +/** + * Retrigger the `resize` event that was triggered by the {@link Tech}. + * + * @private + * @method Player#handleTechResize_ + * @fires Player#resize + * @listens Tech#resize + */ +'resize', +/** + * Fires when the volume has been changed + * + * @event Player#volumechange + * @type {event} + */ +/** + * Retrigger the `volumechange` event that was triggered by the {@link Tech}. + * + * @private + * @method Player#handleTechVolumechange_ + * @fires Player#volumechange + * @listens Tech#volumechange + */ +'volumechange', +/** + * Fires when the text track has been changed + * + * @event Player#texttrackchange + * @type {event} + */ +/** + * Retrigger the `texttrackchange` event that was triggered by the {@link Tech}. + * + * @private + * @method Player#handleTechTexttrackchange_ + * @fires Player#texttrackchange + * @listens Tech#texttrackchange + */ +'texttrackchange']; + +// events to queue when playback rate is zero +// this is a hash for the sole purpose of mapping non-camel-cased event names +// to camel-cased function names +const TECH_EVENTS_QUEUE = { + canplay: 'CanPlay', + canplaythrough: 'CanPlayThrough', + playing: 'Playing', + seeked: 'Seeked' +}; +const BREAKPOINT_ORDER = ['tiny', 'xsmall', 'small', 'medium', 'large', 'xlarge', 'huge']; +const BREAKPOINT_CLASSES = {}; + +// grep: vjs-layout-tiny +// grep: vjs-layout-x-small +// grep: vjs-layout-small +// grep: vjs-layout-medium +// grep: vjs-layout-large +// grep: vjs-layout-x-large +// grep: vjs-layout-huge +BREAKPOINT_ORDER.forEach(k => { + const v = k.charAt(0) === 'x' ? `x-${k.substring(1)}` : k; + BREAKPOINT_CLASSES[k] = `vjs-layout-${v}`; +}); +const DEFAULT_BREAKPOINTS = { + tiny: 210, + xsmall: 320, + small: 425, + medium: 768, + large: 1440, + xlarge: 2560, + huge: Infinity +}; + +/** + * An instance of the `Player` class is created when any of the Video.js setup methods + * are used to initialize a video. + * + * After an instance has been created it can be accessed globally in three ways: + * 1. By calling `videojs.getPlayer('example_video_1');` + * 2. By calling `videojs('example_video_1');` (not recommended) + * 2. By using it directly via `videojs.players.example_video_1;` + * + * @extends Component + * @global + */ +class Player extends Component$1 { + /** + * Create an instance of this class. + * + * @param {Element} tag + * The original video DOM element used for configuring options. + * + * @param {Object} [options] + * Object of option names and values. + * + * @param {Function} [ready] + * Ready callback function. + */ + constructor(tag, options, ready) { + // Make sure tag ID exists + // also here.. probably better + tag.id = tag.id || options.id || `vjs_video_${newGUID()}`; + + // Set Options + // The options argument overrides options set in the video tag + // which overrides globally set options. + // This latter part coincides with the load order + // (tag must exist before Player) + options = Object.assign(Player.getTagSettings(tag), options); + + // Delay the initialization of children because we need to set up + // player properties first, and can't use `this` before `super()` + options.initChildren = false; + + // Same with creating the element + options.createEl = false; + + // don't auto mixin the evented mixin + options.evented = false; + + // we don't want the player to report touch activity on itself + // see enableTouchActivity in Component + options.reportTouchActivity = false; + + // If language is not set, get the closest lang attribute + if (!options.language) { + const closest = tag.closest('[lang]'); + if (closest) { + options.language = closest.getAttribute('lang'); + } + } + + // Run base component initializing with new options + super(null, options, ready); + + // Create bound methods for document listeners. + this.boundDocumentFullscreenChange_ = e => this.documentFullscreenChange_(e); + this.boundFullWindowOnEscKey_ = e => this.fullWindowOnEscKey(e); + this.boundUpdateStyleEl_ = e => this.updateStyleEl_(e); + this.boundApplyInitTime_ = e => this.applyInitTime_(e); + this.boundUpdateCurrentBreakpoint_ = e => this.updateCurrentBreakpoint_(e); + this.boundHandleTechClick_ = e => this.handleTechClick_(e); + this.boundHandleTechDoubleClick_ = e => this.handleTechDoubleClick_(e); + this.boundHandleTechTouchStart_ = e => this.handleTechTouchStart_(e); + this.boundHandleTechTouchMove_ = e => this.handleTechTouchMove_(e); + this.boundHandleTechTouchEnd_ = e => this.handleTechTouchEnd_(e); + this.boundHandleTechTap_ = e => this.handleTechTap_(e); + + // default isFullscreen_ to false + this.isFullscreen_ = false; + + // create logger + this.log = createLogger(this.id_); + + // Hold our own reference to fullscreen api so it can be mocked in tests + this.fsApi_ = FullscreenApi; + + // Tracks when a tech changes the poster + this.isPosterFromTech_ = false; + + // Holds callback info that gets queued when playback rate is zero + // and a seek is happening + this.queuedCallbacks_ = []; + + // Turn off API access because we're loading a new tech that might load asynchronously + this.isReady_ = false; + + // Init state hasStarted_ + this.hasStarted_ = false; + + // Init state userActive_ + this.userActive_ = false; + + // Init debugEnabled_ + this.debugEnabled_ = false; + + // Init state audioOnlyMode_ + this.audioOnlyMode_ = false; + + // Init state audioPosterMode_ + this.audioPosterMode_ = false; + + // Init state audioOnlyCache_ + this.audioOnlyCache_ = { + playerHeight: null, + hiddenChildren: [] + }; + + // if the global option object was accidentally blown away by + // someone, bail early with an informative error + if (!this.options_ || !this.options_.techOrder || !this.options_.techOrder.length) { + throw new Error('No techOrder specified. Did you overwrite ' + 'videojs.options instead of just changing the ' + 'properties you want to override?'); + } + + // Store the original tag used to set options + this.tag = tag; + + // Store the tag attributes used to restore html5 element + this.tagAttributes = tag && getAttributes(tag); + + // Update current language + this.language(this.options_.language); + + // Update Supported Languages + if (options.languages) { + // Normalise player option languages to lowercase + const languagesToLower = {}; + Object.getOwnPropertyNames(options.languages).forEach(function (name) { + languagesToLower[name.toLowerCase()] = options.languages[name]; + }); + this.languages_ = languagesToLower; + } else { + this.languages_ = Player.prototype.options_.languages; + } + this.resetCache_(); + + // Set poster + /** @type string */ + this.poster_ = options.poster || ''; + + // Set controls + /** @type {boolean} */ + this.controls_ = !!options.controls; + + // Original tag settings stored in options + // now remove immediately so native controls don't flash. + // May be turned back on by HTML5 tech if nativeControlsForTouch is true + tag.controls = false; + tag.removeAttribute('controls'); + this.changingSrc_ = false; + this.playCallbacks_ = []; + this.playTerminatedQueue_ = []; + + // the attribute overrides the option + if (tag.hasAttribute('autoplay')) { + this.autoplay(true); + } else { + // otherwise use the setter to validate and + // set the correct value. + this.autoplay(this.options_.autoplay); + } + + // check plugins + if (options.plugins) { + Object.keys(options.plugins).forEach(name => { + if (typeof this[name] !== 'function') { + throw new Error(`plugin "${name}" does not exist`); + } + }); + } + + /* + * Store the internal state of scrubbing + * + * @private + * @return {Boolean} True if the user is scrubbing + */ + this.scrubbing_ = false; + this.el_ = this.createEl(); + + // Make this an evented object and use `el_` as its event bus. + evented(this, { + eventBusKey: 'el_' + }); + + // listen to document and player fullscreenchange handlers so we receive those events + // before a user can receive them so we can update isFullscreen appropriately. + // make sure that we listen to fullscreenchange events before everything else to make sure that + // our isFullscreen method is updated properly for internal components as well as external. + if (this.fsApi_.requestFullscreen) { + on(document, this.fsApi_.fullscreenchange, this.boundDocumentFullscreenChange_); + this.on(this.fsApi_.fullscreenchange, this.boundDocumentFullscreenChange_); + } + if (this.fluid_) { + this.on(['playerreset', 'resize'], this.boundUpdateStyleEl_); + } + // We also want to pass the original player options to each component and plugin + // as well so they don't need to reach back into the player for options later. + // We also need to do another copy of this.options_ so we don't end up with + // an infinite loop. + const playerOptionsCopy = merge$1(this.options_); + + // Load plugins + if (options.plugins) { + Object.keys(options.plugins).forEach(name => { + this[name](options.plugins[name]); + }); + } + + // Enable debug mode to fire debugon event for all plugins. + if (options.debug) { + this.debug(true); + } + this.options_.playerOptions = playerOptionsCopy; + this.middleware_ = []; + this.playbackRates(options.playbackRates); + if (options.experimentalSvgIcons) { + // Add SVG Sprite to the DOM + const parser = new window$1.DOMParser(); + const parsedSVG = parser.parseFromString(icons, 'image/svg+xml'); + const errorNode = parsedSVG.querySelector('parsererror'); + if (errorNode) { + log$1.warn('Failed to load SVG Icons. Falling back to Font Icons.'); + this.options_.experimentalSvgIcons = null; + } else { + const sprite = parsedSVG.documentElement; + sprite.style.display = 'none'; + this.el_.appendChild(sprite); + this.addClass('vjs-svg-icons-enabled'); + } + } + this.initChildren(); + + // Set isAudio based on whether or not an audio tag was used + this.isAudio(tag.nodeName.toLowerCase() === 'audio'); + + // Update controls className. Can't do this when the controls are initially + // set because the element doesn't exist yet. + if (this.controls()) { + this.addClass('vjs-controls-enabled'); + } else { + this.addClass('vjs-controls-disabled'); + } + + // Set ARIA label and region role depending on player type + this.el_.setAttribute('role', 'region'); + if (this.isAudio()) { + this.el_.setAttribute('aria-label', this.localize('Audio Player')); + } else { + this.el_.setAttribute('aria-label', this.localize('Video Player')); + } + if (this.isAudio()) { + this.addClass('vjs-audio'); + } + + // TODO: Make this smarter. Toggle user state between touching/mousing + // using events, since devices can have both touch and mouse events. + // TODO: Make this check be performed again when the window switches between monitors + // (See https://github.com/videojs/video.js/issues/5683) + if (TOUCH_ENABLED) { + this.addClass('vjs-touch-enabled'); + } + + // iOS Safari has broken hover handling + if (!IS_IOS) { + this.addClass('vjs-workinghover'); + } + + // Make player easily findable by ID + Player.players[this.id_] = this; + + // Add a major version class to aid css in plugins + const majorVersion = version$6.split('.')[0]; + this.addClass(`vjs-v${majorVersion}`); + + // When the player is first initialized, trigger activity so components + // like the control bar show themselves if needed + this.userActive(true); + this.reportUserActivity(); + this.one('play', e => this.listenForUserActivity_(e)); + this.on('keydown', e => this.handleKeyDown(e)); + this.on('languagechange', e => this.handleLanguagechange(e)); + this.breakpoints(this.options_.breakpoints); + this.responsive(this.options_.responsive); + + // Calling both the audio mode methods after the player is fully + // setup to be able to listen to the events triggered by them + this.on('ready', () => { + // Calling the audioPosterMode method first so that + // the audioOnlyMode can take precedence when both options are set to true + this.audioPosterMode(this.options_.audioPosterMode); + this.audioOnlyMode(this.options_.audioOnlyMode); + }); + } + + /** + * Destroys the video player and does any necessary cleanup. + * + * This is especially helpful if you are dynamically adding and removing videos + * to/from the DOM. + * + * @fires Player#dispose + */ + dispose() { + /** + * Called when the player is being disposed of. + * + * @event Player#dispose + * @type {Event} + */ + this.trigger('dispose'); + // prevent dispose from being called twice + this.off('dispose'); + + // Make sure all player-specific document listeners are unbound. This is + off(document, this.fsApi_.fullscreenchange, this.boundDocumentFullscreenChange_); + off(document, 'keydown', this.boundFullWindowOnEscKey_); + if (this.styleEl_ && this.styleEl_.parentNode) { + this.styleEl_.parentNode.removeChild(this.styleEl_); + this.styleEl_ = null; + } + + // Kill reference to this player + Player.players[this.id_] = null; + if (this.tag && this.tag.player) { + this.tag.player = null; + } + if (this.el_ && this.el_.player) { + this.el_.player = null; + } + if (this.tech_) { + this.tech_.dispose(); + this.isPosterFromTech_ = false; + this.poster_ = ''; + } + if (this.playerElIngest_) { + this.playerElIngest_ = null; + } + if (this.tag) { + this.tag = null; + } + clearCacheForPlayer(this); + + // remove all event handlers for track lists + // all tracks and track listeners are removed on + // tech dispose + ALL.names.forEach(name => { + const props = ALL[name]; + const list = this[props.getterName](); + + // if it is not a native list + // we have to manually remove event listeners + if (list && list.off) { + list.off(); + } + }); + + // the actual .el_ is removed here, or replaced if + super.dispose({ + restoreEl: this.options_.restoreEl + }); + } + + /** + * Create the `Player`'s DOM element. + * + * @return {Element} + * The DOM element that gets created. + */ + createEl() { + let tag = this.tag; + let el; + let playerElIngest = this.playerElIngest_ = tag.parentNode && tag.parentNode.hasAttribute && tag.parentNode.hasAttribute('data-vjs-player'); + const divEmbed = this.tag.tagName.toLowerCase() === 'video-js'; + if (playerElIngest) { + el = this.el_ = tag.parentNode; + } else if (!divEmbed) { + el = this.el_ = super.createEl('div'); + } + + // Copy over all the attributes from the tag, including ID and class + // ID will now reference player box, not the video tag + const attrs = getAttributes(tag); + if (divEmbed) { + el = this.el_ = tag; + tag = this.tag = document.createElement('video'); + while (el.children.length) { + tag.appendChild(el.firstChild); + } + if (!hasClass(el, 'video-js')) { + addClass(el, 'video-js'); + } + el.appendChild(tag); + playerElIngest = this.playerElIngest_ = el; + // move properties over from our custom `video-js` element + // to our new `video` element. This will move things like + // `src` or `controls` that were set via js before the player + // was initialized. + Object.keys(el).forEach(k => { + try { + tag[k] = el[k]; + } catch (e) { + // we got a a property like outerHTML which we can't actually copy, ignore it + } + }); + } + + // set tabindex to -1 to remove the video element from the focus order + tag.setAttribute('tabindex', '-1'); + attrs.tabindex = '-1'; + + // Workaround for #4583 on Chrome (on Windows) with JAWS. + // See https://github.com/FreedomScientific/VFO-standards-support/issues/78 + // Note that we can't detect if JAWS is being used, but this ARIA attribute + // doesn't change behavior of Chrome if JAWS is not being used + if (IS_CHROME && IS_WINDOWS) { + tag.setAttribute('role', 'application'); + attrs.role = 'application'; + } + + // Remove width/height attrs from tag so CSS can make it 100% width/height + tag.removeAttribute('width'); + tag.removeAttribute('height'); + if ('width' in attrs) { + delete attrs.width; + } + if ('height' in attrs) { + delete attrs.height; + } + Object.getOwnPropertyNames(attrs).forEach(function (attr) { + // don't copy over the class attribute to the player element when we're in a div embed + // the class is already set up properly in the divEmbed case + // and we want to make sure that the `video-js` class doesn't get lost + if (!(divEmbed && attr === 'class')) { + el.setAttribute(attr, attrs[attr]); + } + if (divEmbed) { + tag.setAttribute(attr, attrs[attr]); + } + }); + + // Update tag id/class for use as HTML5 playback tech + // Might think we should do this after embedding in container so .vjs-tech class + // doesn't flash 100% width/height, but class only applies with .video-js parent + tag.playerId = tag.id; + tag.id += '_html5_api'; + tag.className = 'vjs-tech'; + + // Make player findable on elements + tag.player = el.player = this; + // Default state of video is paused + this.addClass('vjs-paused'); + const deviceClassNames = ['IS_SMART_TV', 'IS_TIZEN', 'IS_WEBOS', 'IS_ANDROID', 'IS_IPAD', 'IS_IPHONE'].filter(key => browser[key]).map(key => { + return 'vjs-device-' + key.substring(3).toLowerCase().replace(/\_/g, '-'); + }); + this.addClass(...deviceClassNames); + + // Add a style element in the player that we'll use to set the width/height + // of the player in a way that's still overridable by CSS, just like the + // video element + if (window$1.VIDEOJS_NO_DYNAMIC_STYLE !== true) { + this.styleEl_ = createStyleElement('vjs-styles-dimensions'); + const defaultsStyleEl = $('.vjs-styles-defaults'); + const head = $('head'); + head.insertBefore(this.styleEl_, defaultsStyleEl ? defaultsStyleEl.nextSibling : head.firstChild); + } + this.fill_ = false; + this.fluid_ = false; + + // Pass in the width/height/aspectRatio options which will update the style el + this.width(this.options_.width); + this.height(this.options_.height); + this.fill(this.options_.fill); + this.fluid(this.options_.fluid); + this.aspectRatio(this.options_.aspectRatio); + // support both crossOrigin and crossorigin to reduce confusion and issues around the name + this.crossOrigin(this.options_.crossOrigin || this.options_.crossorigin); + + // Hide any links within the video/audio tag, + // because IE doesn't hide them completely from screen readers. + const links = tag.getElementsByTagName('a'); + for (let i = 0; i < links.length; i++) { + const linkEl = links.item(i); + addClass(linkEl, 'vjs-hidden'); + linkEl.setAttribute('hidden', 'hidden'); + } + + // insertElFirst seems to cause the networkState to flicker from 3 to 2, so + // keep track of the original for later so we can know if the source originally failed + tag.initNetworkState_ = tag.networkState; + + // Wrap video tag in div (el/box) container + if (tag.parentNode && !playerElIngest) { + tag.parentNode.insertBefore(el, tag); + } + + // insert the tag as the first child of the player element + // then manually add it to the children array so that this.addChild + // will work properly for other components + // + // Breaks iPhone, fixed in HTML5 setup. + prependTo(tag, el); + this.children_.unshift(tag); + + // Set lang attr on player to ensure CSS :lang() in consistent with player + // if it's been set to something different to the doc + this.el_.setAttribute('lang', this.language_); + this.el_.setAttribute('translate', 'no'); + this.el_ = el; + return el; + } + + /** + * Get or set the `Player`'s crossOrigin option. For the HTML5 player, this + * sets the `crossOrigin` property on the `<video>` tag to control the CORS + * behavior. + * + * @see [Video Element Attributes]{@link https://developer.mozilla.org/en-US/docs/Web/HTML/Element/video#attr-crossorigin} + * + * @param {string|null} [value] + * The value to set the `Player`'s crossOrigin to. If an argument is + * given, must be one of `'anonymous'` or `'use-credentials'`, or 'null'. + * + * @return {string|null|undefined} + * - The current crossOrigin value of the `Player` when getting. + * - undefined when setting + */ + crossOrigin(value) { + // `null` can be set to unset a value + if (typeof value === 'undefined') { + return this.techGet_('crossOrigin'); + } + if (value !== null && value !== 'anonymous' && value !== 'use-credentials') { + log$1.warn(`crossOrigin must be null, "anonymous" or "use-credentials", given "${value}"`); + return; + } + this.techCall_('setCrossOrigin', value); + if (this.posterImage) { + this.posterImage.crossOrigin(value); + } + return; + } + + /** + * A getter/setter for the `Player`'s width. Returns the player's configured value. + * To get the current width use `currentWidth()`. + * + * @param {number|string} [value] + * CSS value to set the `Player`'s width to. + * + * @return {number|undefined} + * - The current width of the `Player` when getting. + * - Nothing when setting + */ + width(value) { + return this.dimension('width', value); + } + + /** + * A getter/setter for the `Player`'s height. Returns the player's configured value. + * To get the current height use `currentheight()`. + * + * @param {number|string} [value] + * CSS value to set the `Player`'s height to. + * + * @return {number|undefined} + * - The current height of the `Player` when getting. + * - Nothing when setting + */ + height(value) { + return this.dimension('height', value); + } + + /** + * A getter/setter for the `Player`'s width & height. + * + * @param {string} dimension + * This string can be: + * - 'width' + * - 'height' + * + * @param {number|string} [value] + * Value for dimension specified in the first argument. + * + * @return {number} + * The dimension arguments value when getting (width/height). + */ + dimension(dimension, value) { + const privDimension = dimension + '_'; + if (value === undefined) { + return this[privDimension] || 0; + } + if (value === '' || value === 'auto') { + // If an empty string is given, reset the dimension to be automatic + this[privDimension] = undefined; + this.updateStyleEl_(); + return; + } + const parsedVal = parseFloat(value); + if (isNaN(parsedVal)) { + log$1.error(`Improper value "${value}" supplied for for ${dimension}`); + return; + } + this[privDimension] = parsedVal; + this.updateStyleEl_(); + } + + /** + * A getter/setter/toggler for the vjs-fluid `className` on the `Player`. + * + * Turning this on will turn off fill mode. + * + * @param {boolean} [bool] + * - A value of true adds the class. + * - A value of false removes the class. + * - No value will be a getter. + * + * @return {boolean|undefined} + * - The value of fluid when getting. + * - `undefined` when setting. + */ + fluid(bool) { + if (bool === undefined) { + return !!this.fluid_; + } + this.fluid_ = !!bool; + if (isEvented(this)) { + this.off(['playerreset', 'resize'], this.boundUpdateStyleEl_); + } + if (bool) { + this.addClass('vjs-fluid'); + this.fill(false); + addEventedCallback(this, () => { + this.on(['playerreset', 'resize'], this.boundUpdateStyleEl_); + }); + } else { + this.removeClass('vjs-fluid'); + } + this.updateStyleEl_(); + } + + /** + * A getter/setter/toggler for the vjs-fill `className` on the `Player`. + * + * Turning this on will turn off fluid mode. + * + * @param {boolean} [bool] + * - A value of true adds the class. + * - A value of false removes the class. + * - No value will be a getter. + * + * @return {boolean|undefined} + * - The value of fluid when getting. + * - `undefined` when setting. + */ + fill(bool) { + if (bool === undefined) { + return !!this.fill_; + } + this.fill_ = !!bool; + if (bool) { + this.addClass('vjs-fill'); + this.fluid(false); + } else { + this.removeClass('vjs-fill'); + } + } + + /** + * Get/Set the aspect ratio + * + * @param {string} [ratio] + * Aspect ratio for player + * + * @return {string|undefined} + * returns the current aspect ratio when getting + */ + + /** + * A getter/setter for the `Player`'s aspect ratio. + * + * @param {string} [ratio] + * The value to set the `Player`'s aspect ratio to. + * + * @return {string|undefined} + * - The current aspect ratio of the `Player` when getting. + * - undefined when setting + */ + aspectRatio(ratio) { + if (ratio === undefined) { + return this.aspectRatio_; + } + + // Check for width:height format + if (!/^\d+\:\d+$/.test(ratio)) { + throw new Error('Improper value supplied for aspect ratio. The format should be width:height, for example 16:9.'); + } + this.aspectRatio_ = ratio; + + // We're assuming if you set an aspect ratio you want fluid mode, + // because in fixed mode you could calculate width and height yourself. + this.fluid(true); + this.updateStyleEl_(); + } + + /** + * Update styles of the `Player` element (height, width and aspect ratio). + * + * @private + * @listens Tech#loadedmetadata + */ + updateStyleEl_() { + if (window$1.VIDEOJS_NO_DYNAMIC_STYLE === true) { + const width = typeof this.width_ === 'number' ? this.width_ : this.options_.width; + const height = typeof this.height_ === 'number' ? this.height_ : this.options_.height; + const techEl = this.tech_ && this.tech_.el(); + if (techEl) { + if (width >= 0) { + techEl.width = width; + } + if (height >= 0) { + techEl.height = height; + } + } + return; + } + let width; + let height; + let aspectRatio; + let idClass; + + // The aspect ratio is either used directly or to calculate width and height. + if (this.aspectRatio_ !== undefined && this.aspectRatio_ !== 'auto') { + // Use any aspectRatio that's been specifically set + aspectRatio = this.aspectRatio_; + } else if (this.videoWidth() > 0) { + // Otherwise try to get the aspect ratio from the video metadata + aspectRatio = this.videoWidth() + ':' + this.videoHeight(); + } else { + // Or use a default. The video element's is 2:1, but 16:9 is more common. + aspectRatio = '16:9'; + } + + // Get the ratio as a decimal we can use to calculate dimensions + const ratioParts = aspectRatio.split(':'); + const ratioMultiplier = ratioParts[1] / ratioParts[0]; + if (this.width_ !== undefined) { + // Use any width that's been specifically set + width = this.width_; + } else if (this.height_ !== undefined) { + // Or calculate the width from the aspect ratio if a height has been set + width = this.height_ / ratioMultiplier; + } else { + // Or use the video's metadata, or use the video el's default of 300 + width = this.videoWidth() || 300; + } + if (this.height_ !== undefined) { + // Use any height that's been specifically set + height = this.height_; + } else { + // Otherwise calculate the height from the ratio and the width + height = width * ratioMultiplier; + } + + // Ensure the CSS class is valid by starting with an alpha character + if (/^[^a-zA-Z]/.test(this.id())) { + idClass = 'dimensions-' + this.id(); + } else { + idClass = this.id() + '-dimensions'; + } + + // Ensure the right class is still on the player for the style element + this.addClass(idClass); + setTextContent(this.styleEl_, ` + .${idClass} { + width: ${width}px; + height: ${height}px; + } + + .${idClass}.vjs-fluid:not(.vjs-audio-only-mode) { + padding-top: ${ratioMultiplier * 100}%; + } + `); + } + + /** + * Load/Create an instance of playback {@link Tech} including element + * and API methods. Then append the `Tech` element in `Player` as a child. + * + * @param {string} techName + * name of the playback technology + * + * @param {string} source + * video source + * + * @private + */ + loadTech_(techName, source) { + // Pause and remove current playback technology + if (this.tech_) { + this.unloadTech_(); + } + const titleTechName = toTitleCase$1(techName); + const camelTechName = techName.charAt(0).toLowerCase() + techName.slice(1); + + // get rid of the HTML5 video tag as soon as we are using another tech + if (titleTechName !== 'Html5' && this.tag) { + Tech.getTech('Html5').disposeMediaElement(this.tag); + this.tag.player = null; + this.tag = null; + } + this.techName_ = titleTechName; + + // Turn off API access because we're loading a new tech that might load asynchronously + this.isReady_ = false; + let autoplay = this.autoplay(); + + // if autoplay is a string (or `true` with normalizeAutoplay: true) we pass false to the tech + // because the player is going to handle autoplay on `loadstart` + if (typeof this.autoplay() === 'string' || this.autoplay() === true && this.options_.normalizeAutoplay) { + autoplay = false; + } + + // Grab tech-specific options from player options and add source and parent element to use. + const techOptions = { + source, + autoplay, + 'nativeControlsForTouch': this.options_.nativeControlsForTouch, + 'playerId': this.id(), + 'techId': `${this.id()}_${camelTechName}_api`, + 'playsinline': this.options_.playsinline, + 'preload': this.options_.preload, + 'loop': this.options_.loop, + 'disablePictureInPicture': this.options_.disablePictureInPicture, + 'muted': this.options_.muted, + 'poster': this.poster(), + 'language': this.language(), + 'playerElIngest': this.playerElIngest_ || false, + 'vtt.js': this.options_['vtt.js'], + 'canOverridePoster': !!this.options_.techCanOverridePoster, + 'enableSourceset': this.options_.enableSourceset + }; + ALL.names.forEach(name => { + const props = ALL[name]; + techOptions[props.getterName] = this[props.privateName]; + }); + Object.assign(techOptions, this.options_[titleTechName]); + Object.assign(techOptions, this.options_[camelTechName]); + Object.assign(techOptions, this.options_[techName.toLowerCase()]); + if (this.tag) { + techOptions.tag = this.tag; + } + if (source && source.src === this.cache_.src && this.cache_.currentTime > 0) { + techOptions.startTime = this.cache_.currentTime; + } + + // Initialize tech instance + const TechClass = Tech.getTech(techName); + if (!TechClass) { + throw new Error(`No Tech named '${titleTechName}' exists! '${titleTechName}' should be registered using videojs.registerTech()'`); + } + this.tech_ = new TechClass(techOptions); + + // player.triggerReady is always async, so don't need this to be async + this.tech_.ready(bind_(this, this.handleTechReady_), true); + textTrackConverter.jsonToTextTracks(this.textTracksJson_ || [], this.tech_); + + // Listen to all HTML5-defined events and trigger them on the player + TECH_EVENTS_RETRIGGER.forEach(event => { + this.on(this.tech_, event, e => this[`handleTech${toTitleCase$1(event)}_`](e)); + }); + Object.keys(TECH_EVENTS_QUEUE).forEach(event => { + this.on(this.tech_, event, eventObj => { + if (this.tech_.playbackRate() === 0 && this.tech_.seeking()) { + this.queuedCallbacks_.push({ + callback: this[`handleTech${TECH_EVENTS_QUEUE[event]}_`].bind(this), + event: eventObj + }); + return; + } + this[`handleTech${TECH_EVENTS_QUEUE[event]}_`](eventObj); + }); + }); + this.on(this.tech_, 'loadstart', e => this.handleTechLoadStart_(e)); + this.on(this.tech_, 'sourceset', e => this.handleTechSourceset_(e)); + this.on(this.tech_, 'waiting', e => this.handleTechWaiting_(e)); + this.on(this.tech_, 'ended', e => this.handleTechEnded_(e)); + this.on(this.tech_, 'seeking', e => this.handleTechSeeking_(e)); + this.on(this.tech_, 'play', e => this.handleTechPlay_(e)); + this.on(this.tech_, 'pause', e => this.handleTechPause_(e)); + this.on(this.tech_, 'durationchange', e => this.handleTechDurationChange_(e)); + this.on(this.tech_, 'fullscreenchange', (e, data) => this.handleTechFullscreenChange_(e, data)); + this.on(this.tech_, 'fullscreenerror', (e, err) => this.handleTechFullscreenError_(e, err)); + this.on(this.tech_, 'enterpictureinpicture', e => this.handleTechEnterPictureInPicture_(e)); + this.on(this.tech_, 'leavepictureinpicture', e => this.handleTechLeavePictureInPicture_(e)); + this.on(this.tech_, 'error', e => this.handleTechError_(e)); + this.on(this.tech_, 'posterchange', e => this.handleTechPosterChange_(e)); + this.on(this.tech_, 'textdata', e => this.handleTechTextData_(e)); + this.on(this.tech_, 'ratechange', e => this.handleTechRateChange_(e)); + this.on(this.tech_, 'loadedmetadata', this.boundUpdateStyleEl_); + this.usingNativeControls(this.techGet_('controls')); + if (this.controls() && !this.usingNativeControls()) { + this.addTechControlsListeners_(); + } + + // Add the tech element in the DOM if it was not already there + // Make sure to not insert the original video element if using Html5 + if (this.tech_.el().parentNode !== this.el() && (titleTechName !== 'Html5' || !this.tag)) { + prependTo(this.tech_.el(), this.el()); + } + + // Get rid of the original video tag reference after the first tech is loaded + if (this.tag) { + this.tag.player = null; + this.tag = null; + } + } + + /** + * Unload and dispose of the current playback {@link Tech}. + * + * @private + */ + unloadTech_() { + // Save the current text tracks so that we can reuse the same text tracks with the next tech + ALL.names.forEach(name => { + const props = ALL[name]; + this[props.privateName] = this[props.getterName](); + }); + this.textTracksJson_ = textTrackConverter.textTracksToJson(this.tech_); + this.isReady_ = false; + this.tech_.dispose(); + this.tech_ = false; + if (this.isPosterFromTech_) { + this.poster_ = ''; + this.trigger('posterchange'); + } + this.isPosterFromTech_ = false; + } + + /** + * Return a reference to the current {@link Tech}. + * It will print a warning by default about the danger of using the tech directly + * but any argument that is passed in will silence the warning. + * + * @param {*} [safety] + * Anything passed in to silence the warning + * + * @return {Tech} + * The Tech + */ + tech(safety) { + if (safety === undefined) { + log$1.warn('Using the tech directly can be dangerous. I hope you know what you\'re doing.\n' + 'See https://github.com/videojs/video.js/issues/2617 for more info.\n'); + } + return this.tech_; + } + + /** + * An object that contains Video.js version. + * + * @typedef {Object} PlayerVersion + * + * @property {string} 'video.js' - Video.js version + */ + + /** + * Returns an object with Video.js version. + * + * @return {PlayerVersion} + * An object with Video.js version. + */ + version() { + return { + 'video.js': version$6 + }; + } + + /** + * Set up click and touch listeners for the playback element + * + * - On desktops: a click on the video itself will toggle playback + * - On mobile devices: a click on the video toggles controls + * which is done by toggling the user state between active and + * inactive + * - A tap can signal that a user has become active or has become inactive + * e.g. a quick tap on an iPhone movie should reveal the controls. Another + * quick tap should hide them again (signaling the user is in an inactive + * viewing state) + * - In addition to this, we still want the user to be considered inactive after + * a few seconds of inactivity. + * + * > Note: the only part of iOS interaction we can't mimic with this setup + * is a touch and hold on the video element counting as activity in order to + * keep the controls showing, but that shouldn't be an issue. A touch and hold + * on any controls will still keep the user active + * + * @private + */ + addTechControlsListeners_() { + // Make sure to remove all the previous listeners in case we are called multiple times. + this.removeTechControlsListeners_(); + this.on(this.tech_, 'click', this.boundHandleTechClick_); + this.on(this.tech_, 'dblclick', this.boundHandleTechDoubleClick_); + + // If the controls were hidden we don't want that to change without a tap event + // so we'll check if the controls were already showing before reporting user + // activity + this.on(this.tech_, 'touchstart', this.boundHandleTechTouchStart_); + this.on(this.tech_, 'touchmove', this.boundHandleTechTouchMove_); + this.on(this.tech_, 'touchend', this.boundHandleTechTouchEnd_); + + // The tap listener needs to come after the touchend listener because the tap + // listener cancels out any reportedUserActivity when setting userActive(false) + this.on(this.tech_, 'tap', this.boundHandleTechTap_); + } + + /** + * Remove the listeners used for click and tap controls. This is needed for + * toggling to controls disabled, where a tap/touch should do nothing. + * + * @private + */ + removeTechControlsListeners_() { + // We don't want to just use `this.off()` because there might be other needed + // listeners added by techs that extend this. + this.off(this.tech_, 'tap', this.boundHandleTechTap_); + this.off(this.tech_, 'touchstart', this.boundHandleTechTouchStart_); + this.off(this.tech_, 'touchmove', this.boundHandleTechTouchMove_); + this.off(this.tech_, 'touchend', this.boundHandleTechTouchEnd_); + this.off(this.tech_, 'click', this.boundHandleTechClick_); + this.off(this.tech_, 'dblclick', this.boundHandleTechDoubleClick_); + } + + /** + * Player waits for the tech to be ready + * + * @private + */ + handleTechReady_() { + this.triggerReady(); + + // Keep the same volume as before + if (this.cache_.volume) { + this.techCall_('setVolume', this.cache_.volume); + } + + // Look if the tech found a higher resolution poster while loading + this.handleTechPosterChange_(); + + // Update the duration if available + this.handleTechDurationChange_(); + } + + /** + * Retrigger the `loadstart` event that was triggered by the {@link Tech}. + * + * @fires Player#loadstart + * @listens Tech#loadstart + * @private + */ + handleTechLoadStart_() { + // TODO: Update to use `emptied` event instead. See #1277. + + this.removeClass('vjs-ended', 'vjs-seeking'); + + // reset the error state + this.error(null); + + // Update the duration + this.handleTechDurationChange_(); + if (!this.paused()) { + /** + * Fired when the user agent begins looking for media data + * + * @event Player#loadstart + * @type {Event} + */ + this.trigger('loadstart'); + } else { + // reset the hasStarted state + this.hasStarted(false); + this.trigger('loadstart'); + } + + // autoplay happens after loadstart for the browser, + // so we mimic that behavior + this.manualAutoplay_(this.autoplay() === true && this.options_.normalizeAutoplay ? 'play' : this.autoplay()); + } + + /** + * Handle autoplay string values, rather than the typical boolean + * values that should be handled by the tech. Note that this is not + * part of any specification. Valid values and what they do can be + * found on the autoplay getter at Player#autoplay() + */ + manualAutoplay_(type) { + if (!this.tech_ || typeof type !== 'string') { + return; + } + + // Save original muted() value, set muted to true, and attempt to play(). + // On promise rejection, restore muted from saved value + const resolveMuted = () => { + const previouslyMuted = this.muted(); + this.muted(true); + const restoreMuted = () => { + this.muted(previouslyMuted); + }; + + // restore muted on play terminatation + this.playTerminatedQueue_.push(restoreMuted); + const mutedPromise = this.play(); + if (!isPromise(mutedPromise)) { + return; + } + return mutedPromise.catch(err => { + restoreMuted(); + throw new Error(`Rejection at manualAutoplay. Restoring muted value. ${err ? err : ''}`); + }); + }; + let promise; + + // if muted defaults to true + // the only thing we can do is call play + if (type === 'any' && !this.muted()) { + promise = this.play(); + if (isPromise(promise)) { + promise = promise.catch(resolveMuted); + } + } else if (type === 'muted' && !this.muted()) { + promise = resolveMuted(); + } else { + promise = this.play(); + } + if (!isPromise(promise)) { + return; + } + return promise.then(() => { + this.trigger({ + type: 'autoplay-success', + autoplay: type + }); + }).catch(() => { + this.trigger({ + type: 'autoplay-failure', + autoplay: type + }); + }); + } + + /** + * Update the internal source caches so that we return the correct source from + * `src()`, `currentSource()`, and `currentSources()`. + * + * > Note: `currentSources` will not be updated if the source that is passed in exists + * in the current `currentSources` cache. + * + * + * @param {Tech~SourceObject} srcObj + * A string or object source to update our caches to. + */ + updateSourceCaches_(srcObj = '') { + let src = srcObj; + let type = ''; + if (typeof src !== 'string') { + src = srcObj.src; + type = srcObj.type; + } + + // make sure all the caches are set to default values + // to prevent null checking + this.cache_.source = this.cache_.source || {}; + this.cache_.sources = this.cache_.sources || []; + + // try to get the type of the src that was passed in + if (src && !type) { + type = findMimetype(this, src); + } + + // update `currentSource` cache always + this.cache_.source = merge$1({}, srcObj, { + src, + type + }); + const matchingSources = this.cache_.sources.filter(s => s.src && s.src === src); + const sourceElSources = []; + const sourceEls = this.$$('source'); + const matchingSourceEls = []; + for (let i = 0; i < sourceEls.length; i++) { + const sourceObj = getAttributes(sourceEls[i]); + sourceElSources.push(sourceObj); + if (sourceObj.src && sourceObj.src === src) { + matchingSourceEls.push(sourceObj.src); + } + } + + // if we have matching source els but not matching sources + // the current source cache is not up to date + if (matchingSourceEls.length && !matchingSources.length) { + this.cache_.sources = sourceElSources; + // if we don't have matching source or source els set the + // sources cache to the `currentSource` cache + } else if (!matchingSources.length) { + this.cache_.sources = [this.cache_.source]; + } + + // update the tech `src` cache + this.cache_.src = src; + } + + /** + * *EXPERIMENTAL* Fired when the source is set or changed on the {@link Tech} + * causing the media element to reload. + * + * It will fire for the initial source and each subsequent source. + * This event is a custom event from Video.js and is triggered by the {@link Tech}. + * + * The event object for this event contains a `src` property that will contain the source + * that was available when the event was triggered. This is generally only necessary if Video.js + * is switching techs while the source was being changed. + * + * It is also fired when `load` is called on the player (or media element) + * because the {@link https://html.spec.whatwg.org/multipage/media.html#dom-media-load|specification for `load`} + * says that the resource selection algorithm needs to be aborted and restarted. + * In this case, it is very likely that the `src` property will be set to the + * empty string `""` to indicate we do not know what the source will be but + * that it is changing. + * + * *This event is currently still experimental and may change in minor releases.* + * __To use this, pass `enableSourceset` option to the player.__ + * + * @event Player#sourceset + * @type {Event} + * @prop {string} src + * The source url available when the `sourceset` was triggered. + * It will be an empty string if we cannot know what the source is + * but know that the source will change. + */ + /** + * Retrigger the `sourceset` event that was triggered by the {@link Tech}. + * + * @fires Player#sourceset + * @listens Tech#sourceset + * @private + */ + handleTechSourceset_(event) { + // only update the source cache when the source + // was not updated using the player api + if (!this.changingSrc_) { + let updateSourceCaches = src => this.updateSourceCaches_(src); + const playerSrc = this.currentSource().src; + const eventSrc = event.src; + + // if we have a playerSrc that is not a blob, and a tech src that is a blob + if (playerSrc && !/^blob:/.test(playerSrc) && /^blob:/.test(eventSrc)) { + // if both the tech source and the player source were updated we assume + // something like @videojs/http-streaming did the sourceset and skip updating the source cache. + if (!this.lastSource_ || this.lastSource_.tech !== eventSrc && this.lastSource_.player !== playerSrc) { + updateSourceCaches = () => {}; + } + } + + // update the source to the initial source right away + // in some cases this will be empty string + updateSourceCaches(eventSrc); + + // if the `sourceset` `src` was an empty string + // wait for a `loadstart` to update the cache to `currentSrc`. + // If a sourceset happens before a `loadstart`, we reset the state + if (!event.src) { + this.tech_.any(['sourceset', 'loadstart'], e => { + // if a sourceset happens before a `loadstart` there + // is nothing to do as this `handleTechSourceset_` + // will be called again and this will be handled there. + if (e.type === 'sourceset') { + return; + } + const techSrc = this.techGet_('currentSrc'); + this.lastSource_.tech = techSrc; + this.updateSourceCaches_(techSrc); + }); + } + } + this.lastSource_ = { + player: this.currentSource().src, + tech: event.src + }; + this.trigger({ + src: event.src, + type: 'sourceset' + }); + } + + /** + * Add/remove the vjs-has-started class + * + * + * @param {boolean} request + * - true: adds the class + * - false: remove the class + * + * @return {boolean} + * the boolean value of hasStarted_ + */ + hasStarted(request) { + if (request === undefined) { + // act as getter, if we have no request to change + return this.hasStarted_; + } + if (request === this.hasStarted_) { + return; + } + this.hasStarted_ = request; + if (this.hasStarted_) { + this.addClass('vjs-has-started'); + } else { + this.removeClass('vjs-has-started'); + } + } + + /** + * Fired whenever the media begins or resumes playback + * + * @see [Spec]{@link https://html.spec.whatwg.org/multipage/embedded-content.html#dom-media-play} + * @fires Player#play + * @listens Tech#play + * @private + */ + handleTechPlay_() { + this.removeClass('vjs-ended', 'vjs-paused'); + this.addClass('vjs-playing'); + + // hide the poster when the user hits play + this.hasStarted(true); + /** + * Triggered whenever an {@link Tech#play} event happens. Indicates that + * playback has started or resumed. + * + * @event Player#play + * @type {Event} + */ + this.trigger('play'); + } + + /** + * Retrigger the `ratechange` event that was triggered by the {@link Tech}. + * + * If there were any events queued while the playback rate was zero, fire + * those events now. + * + * @private + * @method Player#handleTechRateChange_ + * @fires Player#ratechange + * @listens Tech#ratechange + */ + handleTechRateChange_() { + if (this.tech_.playbackRate() > 0 && this.cache_.lastPlaybackRate === 0) { + this.queuedCallbacks_.forEach(queued => queued.callback(queued.event)); + this.queuedCallbacks_ = []; + } + this.cache_.lastPlaybackRate = this.tech_.playbackRate(); + /** + * Fires when the playing speed of the audio/video is changed + * + * @event Player#ratechange + * @type {event} + */ + this.trigger('ratechange'); + } + + /** + * Retrigger the `waiting` event that was triggered by the {@link Tech}. + * + * @fires Player#waiting + * @listens Tech#waiting + * @private + */ + handleTechWaiting_() { + this.addClass('vjs-waiting'); + /** + * A readyState change on the DOM element has caused playback to stop. + * + * @event Player#waiting + * @type {Event} + */ + this.trigger('waiting'); + + // Browsers may emit a timeupdate event after a waiting event. In order to prevent + // premature removal of the waiting class, wait for the time to change. + const timeWhenWaiting = this.currentTime(); + const timeUpdateListener = () => { + if (timeWhenWaiting !== this.currentTime()) { + this.removeClass('vjs-waiting'); + this.off('timeupdate', timeUpdateListener); + } + }; + this.on('timeupdate', timeUpdateListener); + } + + /** + * Retrigger the `canplay` event that was triggered by the {@link Tech}. + * > Note: This is not consistent between browsers. See #1351 + * + * @fires Player#canplay + * @listens Tech#canplay + * @private + */ + handleTechCanPlay_() { + this.removeClass('vjs-waiting'); + /** + * The media has a readyState of HAVE_FUTURE_DATA or greater. + * + * @event Player#canplay + * @type {Event} + */ + this.trigger('canplay'); + } + + /** + * Retrigger the `canplaythrough` event that was triggered by the {@link Tech}. + * + * @fires Player#canplaythrough + * @listens Tech#canplaythrough + * @private + */ + handleTechCanPlayThrough_() { + this.removeClass('vjs-waiting'); + /** + * The media has a readyState of HAVE_ENOUGH_DATA or greater. This means that the + * entire media file can be played without buffering. + * + * @event Player#canplaythrough + * @type {Event} + */ + this.trigger('canplaythrough'); + } + + /** + * Retrigger the `playing` event that was triggered by the {@link Tech}. + * + * @fires Player#playing + * @listens Tech#playing + * @private + */ + handleTechPlaying_() { + this.removeClass('vjs-waiting'); + /** + * The media is no longer blocked from playback, and has started playing. + * + * @event Player#playing + * @type {Event} + */ + this.trigger('playing'); + } + + /** + * Retrigger the `seeking` event that was triggered by the {@link Tech}. + * + * @fires Player#seeking + * @listens Tech#seeking + * @private + */ + handleTechSeeking_() { + this.addClass('vjs-seeking'); + /** + * Fired whenever the player is jumping to a new time + * + * @event Player#seeking + * @type {Event} + */ + this.trigger('seeking'); + } + + /** + * Retrigger the `seeked` event that was triggered by the {@link Tech}. + * + * @fires Player#seeked + * @listens Tech#seeked + * @private + */ + handleTechSeeked_() { + this.removeClass('vjs-seeking', 'vjs-ended'); + /** + * Fired when the player has finished jumping to a new time + * + * @event Player#seeked + * @type {Event} + */ + this.trigger('seeked'); + } + + /** + * Retrigger the `pause` event that was triggered by the {@link Tech}. + * + * @fires Player#pause + * @listens Tech#pause + * @private + */ + handleTechPause_() { + this.removeClass('vjs-playing'); + this.addClass('vjs-paused'); + /** + * Fired whenever the media has been paused + * + * @event Player#pause + * @type {Event} + */ + this.trigger('pause'); + } + + /** + * Retrigger the `ended` event that was triggered by the {@link Tech}. + * + * @fires Player#ended + * @listens Tech#ended + * @private + */ + handleTechEnded_() { + this.addClass('vjs-ended'); + this.removeClass('vjs-waiting'); + if (this.options_.loop) { + this.currentTime(0); + this.play(); + } else if (!this.paused()) { + this.pause(); + } + + /** + * Fired when the end of the media resource is reached (currentTime == duration) + * + * @event Player#ended + * @type {Event} + */ + this.trigger('ended'); + } + + /** + * Fired when the duration of the media resource is first known or changed + * + * @listens Tech#durationchange + * @private + */ + handleTechDurationChange_() { + this.duration(this.techGet_('duration')); + } + + /** + * Handle a click on the media element to play/pause + * + * @param {Event} event + * the event that caused this function to trigger + * + * @listens Tech#click + * @private + */ + handleTechClick_(event) { + // When controls are disabled a click should not toggle playback because + // the click is considered a control + if (!this.controls_) { + return; + } + if (this.options_ === undefined || this.options_.userActions === undefined || this.options_.userActions.click === undefined || this.options_.userActions.click !== false) { + if (this.options_ !== undefined && this.options_.userActions !== undefined && typeof this.options_.userActions.click === 'function') { + this.options_.userActions.click.call(this, event); + } else if (this.paused()) { + silencePromise(this.play()); + } else { + this.pause(); + } + } + } + + /** + * Handle a double-click on the media element to enter/exit fullscreen + * + * @param {Event} event + * the event that caused this function to trigger + * + * @listens Tech#dblclick + * @private + */ + handleTechDoubleClick_(event) { + if (!this.controls_) { + return; + } + + // we do not want to toggle fullscreen state + // when double-clicking inside a control bar or a modal + const inAllowedEls = Array.prototype.some.call(this.$$('.vjs-control-bar, .vjs-modal-dialog'), el => el.contains(event.target)); + if (!inAllowedEls) { + /* + * options.userActions.doubleClick + * + * If `undefined` or `true`, double-click toggles fullscreen if controls are present + * Set to `false` to disable double-click handling + * Set to a function to substitute an external double-click handler + */ + if (this.options_ === undefined || this.options_.userActions === undefined || this.options_.userActions.doubleClick === undefined || this.options_.userActions.doubleClick !== false) { + if (this.options_ !== undefined && this.options_.userActions !== undefined && typeof this.options_.userActions.doubleClick === 'function') { + this.options_.userActions.doubleClick.call(this, event); + } else if (this.isFullscreen()) { + this.exitFullscreen(); + } else { + this.requestFullscreen(); + } + } + } + } + + /** + * Handle a tap on the media element. It will toggle the user + * activity state, which hides and shows the controls. + * + * @listens Tech#tap + * @private + */ + handleTechTap_() { + this.userActive(!this.userActive()); + } + + /** + * Handle touch to start + * + * @listens Tech#touchstart + * @private + */ + handleTechTouchStart_() { + this.userWasActive = this.userActive(); + } + + /** + * Handle touch to move + * + * @listens Tech#touchmove + * @private + */ + handleTechTouchMove_() { + if (this.userWasActive) { + this.reportUserActivity(); + } + } + + /** + * Handle touch to end + * + * @param {Event} event + * the touchend event that triggered + * this function + * + * @listens Tech#touchend + * @private + */ + handleTechTouchEnd_(event) { + // Stop the mouse events from also happening + if (event.cancelable) { + event.preventDefault(); + } + } + + /** + * @private + */ + toggleFullscreenClass_() { + if (this.isFullscreen()) { + this.addClass('vjs-fullscreen'); + } else { + this.removeClass('vjs-fullscreen'); + } + } + + /** + * when the document fschange event triggers it calls this + */ + documentFullscreenChange_(e) { + const targetPlayer = e.target.player; + + // if another player was fullscreen + // do a null check for targetPlayer because older firefox's would put document as e.target + if (targetPlayer && targetPlayer !== this) { + return; + } + const el = this.el(); + let isFs = document[this.fsApi_.fullscreenElement] === el; + if (!isFs && el.matches) { + isFs = el.matches(':' + this.fsApi_.fullscreen); + } + this.isFullscreen(isFs); + } + + /** + * Handle Tech Fullscreen Change + * + * @param {Event} event + * the fullscreenchange event that triggered this function + * + * @param {Object} data + * the data that was sent with the event + * + * @private + * @listens Tech#fullscreenchange + * @fires Player#fullscreenchange + */ + handleTechFullscreenChange_(event, data) { + if (data) { + if (data.nativeIOSFullscreen) { + this.addClass('vjs-ios-native-fs'); + this.tech_.one('webkitendfullscreen', () => { + this.removeClass('vjs-ios-native-fs'); + }); + } + this.isFullscreen(data.isFullscreen); + } + } + handleTechFullscreenError_(event, err) { + this.trigger('fullscreenerror', err); + } + + /** + * @private + */ + togglePictureInPictureClass_() { + if (this.isInPictureInPicture()) { + this.addClass('vjs-picture-in-picture'); + } else { + this.removeClass('vjs-picture-in-picture'); + } + } + + /** + * Handle Tech Enter Picture-in-Picture. + * + * @param {Event} event + * the enterpictureinpicture event that triggered this function + * + * @private + * @listens Tech#enterpictureinpicture + */ + handleTechEnterPictureInPicture_(event) { + this.isInPictureInPicture(true); + } + + /** + * Handle Tech Leave Picture-in-Picture. + * + * @param {Event} event + * the leavepictureinpicture event that triggered this function + * + * @private + * @listens Tech#leavepictureinpicture + */ + handleTechLeavePictureInPicture_(event) { + this.isInPictureInPicture(false); + } + + /** + * Fires when an error occurred during the loading of an audio/video. + * + * @private + * @listens Tech#error + */ + handleTechError_() { + const error = this.tech_.error(); + if (error) { + this.error(error); + } + } + + /** + * Retrigger the `textdata` event that was triggered by the {@link Tech}. + * + * @fires Player#textdata + * @listens Tech#textdata + * @private + */ + handleTechTextData_() { + let data = null; + if (arguments.length > 1) { + data = arguments[1]; + } + + /** + * Fires when we get a textdata event from tech + * + * @event Player#textdata + * @type {Event} + */ + this.trigger('textdata', data); + } + + /** + * Get object for cached values. + * + * @return {Object} + * get the current object cache + */ + getCache() { + return this.cache_; + } + + /** + * Resets the internal cache object. + * + * Using this function outside the player constructor or reset method may + * have unintended side-effects. + * + * @private + */ + resetCache_() { + this.cache_ = { + // Right now, the currentTime is not _really_ cached because it is always + // retrieved from the tech (see: currentTime). However, for completeness, + // we set it to zero here to ensure that if we do start actually caching + // it, we reset it along with everything else. + currentTime: 0, + initTime: 0, + inactivityTimeout: this.options_.inactivityTimeout, + duration: NaN, + lastVolume: 1, + lastPlaybackRate: this.defaultPlaybackRate(), + media: null, + src: '', + source: {}, + sources: [], + playbackRates: [], + volume: 1 + }; + } + + /** + * Pass values to the playback tech + * + * @param {string} [method] + * the method to call + * + * @param {Object} [arg] + * the argument to pass + * + * @private + */ + techCall_(method, arg) { + // If it's not ready yet, call method when it is + + this.ready(function () { + if (method in allowedSetters) { + return set(this.middleware_, this.tech_, method, arg); + } else if (method in allowedMediators) { + return mediate(this.middleware_, this.tech_, method, arg); + } + try { + if (this.tech_) { + this.tech_[method](arg); + } + } catch (e) { + log$1(e); + throw e; + } + }, true); + } + + /** + * Mediate attempt to call playback tech method + * and return the value of the method called. + * + * @param {string} method + * Tech method + * + * @return {*} + * Value returned by the tech method called, undefined if tech + * is not ready or tech method is not present + * + * @private + */ + techGet_(method) { + if (!this.tech_ || !this.tech_.isReady_) { + return; + } + if (method in allowedGetters) { + return get(this.middleware_, this.tech_, method); + } else if (method in allowedMediators) { + return mediate(this.middleware_, this.tech_, method); + } + + // Log error when playback tech object is present but method + // is undefined or unavailable + try { + return this.tech_[method](); + } catch (e) { + // When building additional tech libs, an expected method may not be defined yet + if (this.tech_[method] === undefined) { + log$1(`Video.js: ${method} method not defined for ${this.techName_} playback technology.`, e); + throw e; + } + + // When a method isn't available on the object it throws a TypeError + if (e.name === 'TypeError') { + log$1(`Video.js: ${method} unavailable on ${this.techName_} playback technology element.`, e); + this.tech_.isReady_ = false; + throw e; + } + + // If error unknown, just log and throw + log$1(e); + throw e; + } + } + + /** + * Attempt to begin playback at the first opportunity. + * + * @return {Promise|undefined} + * Returns a promise if the browser supports Promises (or one + * was passed in as an option). This promise will be resolved on + * the return value of play. If this is undefined it will fulfill the + * promise chain otherwise the promise chain will be fulfilled when + * the promise from play is fulfilled. + */ + play() { + return new Promise(resolve => { + this.play_(resolve); + }); + } + + /** + * The actual logic for play, takes a callback that will be resolved on the + * return value of play. This allows us to resolve to the play promise if there + * is one on modern browsers. + * + * @private + * @param {Function} [callback] + * The callback that should be called when the techs play is actually called + */ + play_(callback = silencePromise) { + this.playCallbacks_.push(callback); + const isSrcReady = Boolean(!this.changingSrc_ && (this.src() || this.currentSrc())); + const isSafariOrIOS = Boolean(IS_ANY_SAFARI || IS_IOS); + + // treat calls to play_ somewhat like the `one` event function + if (this.waitToPlay_) { + this.off(['ready', 'loadstart'], this.waitToPlay_); + this.waitToPlay_ = null; + } + + // if the player/tech is not ready or the src itself is not ready + // queue up a call to play on `ready` or `loadstart` + if (!this.isReady_ || !isSrcReady) { + this.waitToPlay_ = e => { + this.play_(); + }; + this.one(['ready', 'loadstart'], this.waitToPlay_); + + // if we are in Safari, there is a high chance that loadstart will trigger after the gesture timeperiod + // in that case, we need to prime the video element by calling load so it'll be ready in time + if (!isSrcReady && isSafariOrIOS) { + this.load(); + } + return; + } + + // If the player/tech is ready and we have a source, we can attempt playback. + const val = this.techGet_('play'); + + // For native playback, reset the progress bar if we get a play call from a replay. + const isNativeReplay = isSafariOrIOS && this.hasClass('vjs-ended'); + if (isNativeReplay) { + this.resetProgressBar_(); + } + // play was terminated if the returned value is null + if (val === null) { + this.runPlayTerminatedQueue_(); + } else { + this.runPlayCallbacks_(val); + } + } + + /** + * These functions will be run when if play is terminated. If play + * runPlayCallbacks_ is run these function will not be run. This allows us + * to differentiate between a terminated play and an actual call to play. + */ + runPlayTerminatedQueue_() { + const queue = this.playTerminatedQueue_.slice(0); + this.playTerminatedQueue_ = []; + queue.forEach(function (q) { + q(); + }); + } + + /** + * When a callback to play is delayed we have to run these + * callbacks when play is actually called on the tech. This function + * runs the callbacks that were delayed and accepts the return value + * from the tech. + * + * @param {undefined|Promise} val + * The return value from the tech. + */ + runPlayCallbacks_(val) { + const callbacks = this.playCallbacks_.slice(0); + this.playCallbacks_ = []; + // clear play terminatedQueue since we finished a real play + this.playTerminatedQueue_ = []; + callbacks.forEach(function (cb) { + cb(val); + }); + } + + /** + * Pause the video playback + */ + pause() { + this.techCall_('pause'); + } + + /** + * Check if the player is paused or has yet to play + * + * @return {boolean} + * - false: if the media is currently playing + * - true: if media is not currently playing + */ + paused() { + // The initial state of paused should be true (in Safari it's actually false) + return this.techGet_('paused') === false ? false : true; + } + + /** + * Get a TimeRange object representing the current ranges of time that the user + * has played. + * + * @return { import('./utils/time').TimeRange } + * A time range object that represents all the increments of time that have + * been played. + */ + played() { + return this.techGet_('played') || createTimeRanges$1(0, 0); + } + + /** + * Sets or returns whether or not the user is "scrubbing". Scrubbing is + * when the user has clicked the progress bar handle and is + * dragging it along the progress bar. + * + * @param {boolean} [isScrubbing] + * whether the user is or is not scrubbing + * + * @return {boolean|undefined} + * - The value of scrubbing when getting + * - Nothing when setting + */ + scrubbing(isScrubbing) { + if (typeof isScrubbing === 'undefined') { + return this.scrubbing_; + } + this.scrubbing_ = !!isScrubbing; + this.techCall_('setScrubbing', this.scrubbing_); + if (isScrubbing) { + this.addClass('vjs-scrubbing'); + } else { + this.removeClass('vjs-scrubbing'); + } + } + + /** + * Get or set the current time (in seconds) + * + * @param {number|string} [seconds] + * The time to seek to in seconds + * + * @return {number|undefined} + * - the current time in seconds when getting + * - Nothing when setting + */ + currentTime(seconds) { + if (seconds === undefined) { + // cache last currentTime and return. default to 0 seconds + // + // Caching the currentTime is meant to prevent a massive amount of reads on the tech's + // currentTime when scrubbing, but may not provide much performance benefit after all. + // Should be tested. Also something has to read the actual current time or the cache will + // never get updated. + this.cache_.currentTime = this.techGet_('currentTime') || 0; + return this.cache_.currentTime; + } + if (seconds < 0) { + seconds = 0; + } + if (!this.isReady_ || this.changingSrc_ || !this.tech_ || !this.tech_.isReady_) { + this.cache_.initTime = seconds; + this.off('canplay', this.boundApplyInitTime_); + this.one('canplay', this.boundApplyInitTime_); + return; + } + this.techCall_('setCurrentTime', seconds); + this.cache_.initTime = 0; + if (isFinite(seconds)) { + this.cache_.currentTime = Number(seconds); + } + } + + /** + * Apply the value of initTime stored in cache as currentTime. + * + * @private + */ + applyInitTime_() { + this.currentTime(this.cache_.initTime); + } + + /** + * Normally gets the length in time of the video in seconds; + * in all but the rarest use cases an argument will NOT be passed to the method + * + * > **NOTE**: The video must have started loading before the duration can be + * known, and depending on preload behaviour may not be known until the video starts + * playing. + * + * @fires Player#durationchange + * + * @param {number} [seconds] + * The duration of the video to set in seconds + * + * @return {number|undefined} + * - The duration of the video in seconds when getting + * - Nothing when setting + */ + duration(seconds) { + if (seconds === undefined) { + // return NaN if the duration is not known + return this.cache_.duration !== undefined ? this.cache_.duration : NaN; + } + seconds = parseFloat(seconds); + + // Standardize on Infinity for signaling video is live + if (seconds < 0) { + seconds = Infinity; + } + if (seconds !== this.cache_.duration) { + // Cache the last set value for optimized scrubbing + this.cache_.duration = seconds; + if (seconds === Infinity) { + this.addClass('vjs-live'); + } else { + this.removeClass('vjs-live'); + } + if (!isNaN(seconds)) { + // Do not fire durationchange unless the duration value is known. + // @see [Spec]{@link https://www.w3.org/TR/2011/WD-html5-20110113/video.html#media-element-load-algorithm} + + /** + * @event Player#durationchange + * @type {Event} + */ + this.trigger('durationchange'); + } + } + } + + /** + * Calculates how much time is left in the video. Not part + * of the native video API. + * + * @return {number} + * The time remaining in seconds + */ + remainingTime() { + return this.duration() - this.currentTime(); + } + + /** + * A remaining time function that is intended to be used when + * the time is to be displayed directly to the user. + * + * @return {number} + * The rounded time remaining in seconds + */ + remainingTimeDisplay() { + return Math.floor(this.duration()) - Math.floor(this.currentTime()); + } + + // + // Kind of like an array of portions of the video that have been downloaded. + + /** + * Get a TimeRange object with an array of the times of the video + * that have been downloaded. If you just want the percent of the + * video that's been downloaded, use bufferedPercent. + * + * @see [Buffered Spec]{@link http://dev.w3.org/html5/spec/video.html#dom-media-buffered} + * + * @return { import('./utils/time').TimeRange } + * A mock {@link TimeRanges} object (following HTML spec) + */ + buffered() { + let buffered = this.techGet_('buffered'); + if (!buffered || !buffered.length) { + buffered = createTimeRanges$1(0, 0); + } + return buffered; + } + + /** + * Get the TimeRanges of the media that are currently available + * for seeking to. + * + * @see [Seekable Spec]{@link https://html.spec.whatwg.org/multipage/media.html#dom-media-seekable} + * + * @return { import('./utils/time').TimeRange } + * A mock {@link TimeRanges} object (following HTML spec) + */ + seekable() { + let seekable = this.techGet_('seekable'); + if (!seekable || !seekable.length) { + seekable = createTimeRanges$1(0, 0); + } + return seekable; + } + + /** + * Returns whether the player is in the "seeking" state. + * + * @return {boolean} True if the player is in the seeking state, false if not. + */ + seeking() { + return this.techGet_('seeking'); + } + + /** + * Returns whether the player is in the "ended" state. + * + * @return {boolean} True if the player is in the ended state, false if not. + */ + ended() { + return this.techGet_('ended'); + } + + /** + * Returns the current state of network activity for the element, from + * the codes in the list below. + * - NETWORK_EMPTY (numeric value 0) + * The element has not yet been initialised. All attributes are in + * their initial states. + * - NETWORK_IDLE (numeric value 1) + * The element's resource selection algorithm is active and has + * selected a resource, but it is not actually using the network at + * this time. + * - NETWORK_LOADING (numeric value 2) + * The user agent is actively trying to download data. + * - NETWORK_NO_SOURCE (numeric value 3) + * The element's resource selection algorithm is active, but it has + * not yet found a resource to use. + * + * @see https://html.spec.whatwg.org/multipage/embedded-content.html#network-states + * @return {number} the current network activity state + */ + networkState() { + return this.techGet_('networkState'); + } + + /** + * Returns a value that expresses the current state of the element + * with respect to rendering the current playback position, from the + * codes in the list below. + * - HAVE_NOTHING (numeric value 0) + * No information regarding the media resource is available. + * - HAVE_METADATA (numeric value 1) + * Enough of the resource has been obtained that the duration of the + * resource is available. + * - HAVE_CURRENT_DATA (numeric value 2) + * Data for the immediate current playback position is available. + * - HAVE_FUTURE_DATA (numeric value 3) + * Data for the immediate current playback position is available, as + * well as enough data for the user agent to advance the current + * playback position in the direction of playback. + * - HAVE_ENOUGH_DATA (numeric value 4) + * The user agent estimates that enough data is available for + * playback to proceed uninterrupted. + * + * @see https://html.spec.whatwg.org/multipage/embedded-content.html#dom-media-readystate + * @return {number} the current playback rendering state + */ + readyState() { + return this.techGet_('readyState'); + } + + /** + * Get the percent (as a decimal) of the video that's been downloaded. + * This method is not a part of the native HTML video API. + * + * @return {number} + * A decimal between 0 and 1 representing the percent + * that is buffered 0 being 0% and 1 being 100% + */ + bufferedPercent() { + return bufferedPercent(this.buffered(), this.duration()); + } + + /** + * Get the ending time of the last buffered time range + * This is used in the progress bar to encapsulate all time ranges. + * + * @return {number} + * The end of the last buffered time range + */ + bufferedEnd() { + const buffered = this.buffered(); + const duration = this.duration(); + let end = buffered.end(buffered.length - 1); + if (end > duration) { + end = duration; + } + return end; + } + + /** + * Get or set the current volume of the media + * + * @param {number} [percentAsDecimal] + * The new volume as a decimal percent: + * - 0 is muted/0%/off + * - 1.0 is 100%/full + * - 0.5 is half volume or 50% + * + * @return {number|undefined} + * The current volume as a percent when getting + */ + volume(percentAsDecimal) { + let vol; + if (percentAsDecimal !== undefined) { + // Force value to between 0 and 1 + vol = Math.max(0, Math.min(1, percentAsDecimal)); + this.cache_.volume = vol; + this.techCall_('setVolume', vol); + if (vol > 0) { + this.lastVolume_(vol); + } + return; + } + + // Default to 1 when returning current volume. + vol = parseFloat(this.techGet_('volume')); + return isNaN(vol) ? 1 : vol; + } + + /** + * Get the current muted state, or turn mute on or off + * + * @param {boolean} [muted] + * - true to mute + * - false to unmute + * + * @return {boolean|undefined} + * - true if mute is on and getting + * - false if mute is off and getting + * - nothing if setting + */ + muted(muted) { + if (muted !== undefined) { + this.techCall_('setMuted', muted); + return; + } + return this.techGet_('muted') || false; + } + + /** + * Get the current defaultMuted state, or turn defaultMuted on or off. defaultMuted + * indicates the state of muted on initial playback. + * + * ```js + * var myPlayer = videojs('some-player-id'); + * + * myPlayer.src("http://www.example.com/path/to/video.mp4"); + * + * // get, should be false + * console.log(myPlayer.defaultMuted()); + * // set to true + * myPlayer.defaultMuted(true); + * // get should be true + * console.log(myPlayer.defaultMuted()); + * ``` + * + * @param {boolean} [defaultMuted] + * - true to mute + * - false to unmute + * + * @return {boolean|undefined} + * - true if defaultMuted is on and getting + * - false if defaultMuted is off and getting + * - Nothing when setting + */ + defaultMuted(defaultMuted) { + if (defaultMuted !== undefined) { + this.techCall_('setDefaultMuted', defaultMuted); + } + return this.techGet_('defaultMuted') || false; + } + + /** + * Get the last volume, or set it + * + * @param {number} [percentAsDecimal] + * The new last volume as a decimal percent: + * - 0 is muted/0%/off + * - 1.0 is 100%/full + * - 0.5 is half volume or 50% + * + * @return {number|undefined} + * - The current value of lastVolume as a percent when getting + * - Nothing when setting + * + * @private + */ + lastVolume_(percentAsDecimal) { + if (percentAsDecimal !== undefined && percentAsDecimal !== 0) { + this.cache_.lastVolume = percentAsDecimal; + return; + } + return this.cache_.lastVolume; + } + + /** + * Check if current tech can support native fullscreen + * (e.g. with built in controls like iOS) + * + * @return {boolean} + * if native fullscreen is supported + */ + supportsFullScreen() { + return this.techGet_('supportsFullScreen') || false; + } + + /** + * Check if the player is in fullscreen mode or tell the player that it + * is or is not in fullscreen mode. + * + * > NOTE: As of the latest HTML5 spec, isFullscreen is no longer an official + * property and instead document.fullscreenElement is used. But isFullscreen is + * still a valuable property for internal player workings. + * + * @param {boolean} [isFS] + * Set the players current fullscreen state + * + * @return {boolean|undefined} + * - true if fullscreen is on and getting + * - false if fullscreen is off and getting + * - Nothing when setting + */ + isFullscreen(isFS) { + if (isFS !== undefined) { + const oldValue = this.isFullscreen_; + this.isFullscreen_ = Boolean(isFS); + + // if we changed fullscreen state and we're in prefixed mode, trigger fullscreenchange + // this is the only place where we trigger fullscreenchange events for older browsers + // fullWindow mode is treated as a prefixed event and will get a fullscreenchange event as well + if (this.isFullscreen_ !== oldValue && this.fsApi_.prefixed) { + /** + * @event Player#fullscreenchange + * @type {Event} + */ + this.trigger('fullscreenchange'); + } + this.toggleFullscreenClass_(); + return; + } + return this.isFullscreen_; + } + + /** + * Increase the size of the video to full screen + * In some browsers, full screen is not supported natively, so it enters + * "full window mode", where the video fills the browser window. + * In browsers and devices that support native full screen, sometimes the + * browser's default controls will be shown, and not the Video.js custom skin. + * This includes most mobile devices (iOS, Android) and older versions of + * Safari. + * + * @param {Object} [fullscreenOptions] + * Override the player fullscreen options + * + * @fires Player#fullscreenchange + */ + requestFullscreen(fullscreenOptions) { + if (this.isInPictureInPicture()) { + this.exitPictureInPicture(); + } + const self = this; + return new Promise((resolve, reject) => { + function offHandler() { + self.off('fullscreenerror', errorHandler); + self.off('fullscreenchange', changeHandler); + } + function changeHandler() { + offHandler(); + resolve(); + } + function errorHandler(e, err) { + offHandler(); + reject(err); + } + self.one('fullscreenchange', changeHandler); + self.one('fullscreenerror', errorHandler); + const promise = self.requestFullscreenHelper_(fullscreenOptions); + if (promise) { + promise.then(offHandler, offHandler); + promise.then(resolve, reject); + } + }); + } + requestFullscreenHelper_(fullscreenOptions) { + let fsOptions; + + // Only pass fullscreen options to requestFullscreen in spec-compliant browsers. + // Use defaults or player configured option unless passed directly to this method. + if (!this.fsApi_.prefixed) { + fsOptions = this.options_.fullscreen && this.options_.fullscreen.options || {}; + if (fullscreenOptions !== undefined) { + fsOptions = fullscreenOptions; + } + } + + // This method works as follows: + // 1. if a fullscreen api is available, use it + // 1. call requestFullscreen with potential options + // 2. if we got a promise from above, use it to update isFullscreen() + // 2. otherwise, if the tech supports fullscreen, call `enterFullScreen` on it. + // This is particularly used for iPhone, older iPads, and non-safari browser on iOS. + // 3. otherwise, use "fullWindow" mode + if (this.fsApi_.requestFullscreen) { + const promise = this.el_[this.fsApi_.requestFullscreen](fsOptions); + + // Even on browsers with promise support this may not return a promise + if (promise) { + promise.then(() => this.isFullscreen(true), () => this.isFullscreen(false)); + } + return promise; + } else if (this.tech_.supportsFullScreen() && !this.options_.preferFullWindow === true) { + // we can't take the video.js controls fullscreen but we can go fullscreen + // with native controls + this.techCall_('enterFullScreen'); + } else { + // fullscreen isn't supported so we'll just stretch the video element to + // fill the viewport + this.enterFullWindow(); + } + } + + /** + * Return the video to its normal size after having been in full screen mode + * + * @fires Player#fullscreenchange + */ + exitFullscreen() { + const self = this; + return new Promise((resolve, reject) => { + function offHandler() { + self.off('fullscreenerror', errorHandler); + self.off('fullscreenchange', changeHandler); + } + function changeHandler() { + offHandler(); + resolve(); + } + function errorHandler(e, err) { + offHandler(); + reject(err); + } + self.one('fullscreenchange', changeHandler); + self.one('fullscreenerror', errorHandler); + const promise = self.exitFullscreenHelper_(); + if (promise) { + promise.then(offHandler, offHandler); + // map the promise to our resolve/reject methods + promise.then(resolve, reject); + } + }); + } + exitFullscreenHelper_() { + if (this.fsApi_.requestFullscreen) { + const promise = document[this.fsApi_.exitFullscreen](); + + // Even on browsers with promise support this may not return a promise + if (promise) { + // we're splitting the promise here, so, we want to catch the + // potential error so that this chain doesn't have unhandled errors + silencePromise(promise.then(() => this.isFullscreen(false))); + } + return promise; + } else if (this.tech_.supportsFullScreen() && !this.options_.preferFullWindow === true) { + this.techCall_('exitFullScreen'); + } else { + this.exitFullWindow(); + } + } + + /** + * When fullscreen isn't supported we can stretch the + * video container to as wide as the browser will let us. + * + * @fires Player#enterFullWindow + */ + enterFullWindow() { + this.isFullscreen(true); + this.isFullWindow = true; + + // Storing original doc overflow value to return to when fullscreen is off + this.docOrigOverflow = document.documentElement.style.overflow; + + // Add listener for esc key to exit fullscreen + on(document, 'keydown', this.boundFullWindowOnEscKey_); + + // Hide any scroll bars + document.documentElement.style.overflow = 'hidden'; + + // Apply fullscreen styles + addClass(document.body, 'vjs-full-window'); + + /** + * @event Player#enterFullWindow + * @type {Event} + */ + this.trigger('enterFullWindow'); + } + + /** + * Check for call to either exit full window or + * full screen on ESC key + * + * @param {string} event + * Event to check for key press + */ + fullWindowOnEscKey(event) { + if (keycode.isEventKey(event, 'Esc')) { + if (this.isFullscreen() === true) { + if (!this.isFullWindow) { + this.exitFullscreen(); + } else { + this.exitFullWindow(); + } + } + } + } + + /** + * Exit full window + * + * @fires Player#exitFullWindow + */ + exitFullWindow() { + this.isFullscreen(false); + this.isFullWindow = false; + off(document, 'keydown', this.boundFullWindowOnEscKey_); + + // Unhide scroll bars. + document.documentElement.style.overflow = this.docOrigOverflow; + + // Remove fullscreen styles + removeClass(document.body, 'vjs-full-window'); + + // Resize the box, controller, and poster to original sizes + // this.positionAll(); + /** + * @event Player#exitFullWindow + * @type {Event} + */ + this.trigger('exitFullWindow'); + } + + /** + * Get or set disable Picture-in-Picture mode. + * + * @param {boolean} [value] + * - true will disable Picture-in-Picture mode + * - false will enable Picture-in-Picture mode + */ + disablePictureInPicture(value) { + if (value === undefined) { + return this.techGet_('disablePictureInPicture'); + } + this.techCall_('setDisablePictureInPicture', value); + this.options_.disablePictureInPicture = value; + this.trigger('disablepictureinpicturechanged'); + } + + /** + * Check if the player is in Picture-in-Picture mode or tell the player that it + * is or is not in Picture-in-Picture mode. + * + * @param {boolean} [isPiP] + * Set the players current Picture-in-Picture state + * + * @return {boolean|undefined} + * - true if Picture-in-Picture is on and getting + * - false if Picture-in-Picture is off and getting + * - nothing if setting + */ + isInPictureInPicture(isPiP) { + if (isPiP !== undefined) { + this.isInPictureInPicture_ = !!isPiP; + this.togglePictureInPictureClass_(); + return; + } + return !!this.isInPictureInPicture_; + } + + /** + * Create a floating video window always on top of other windows so that users may + * continue consuming media while they interact with other content sites, or + * applications on their device. + * + * This can use document picture-in-picture or element picture in picture + * + * Set `enableDocumentPictureInPicture` to `true` to use docPiP on a supported browser + * Else set `disablePictureInPicture` to `false` to disable elPiP on a supported browser + * + * + * @see [Spec]{@link https://w3c.github.io/picture-in-picture/} + * @see [Spec]{@link https://wicg.github.io/document-picture-in-picture/} + * + * @fires Player#enterpictureinpicture + * + * @return {Promise} + * A promise with a Picture-in-Picture window. + */ + requestPictureInPicture() { + if (this.options_.enableDocumentPictureInPicture && window$1.documentPictureInPicture) { + const pipContainer = document.createElement(this.el().tagName); + pipContainer.classList = this.el().classList; + pipContainer.classList.add('vjs-pip-container'); + if (this.posterImage) { + pipContainer.appendChild(this.posterImage.el().cloneNode(true)); + } + if (this.titleBar) { + pipContainer.appendChild(this.titleBar.el().cloneNode(true)); + } + pipContainer.appendChild(createEl('p', { + className: 'vjs-pip-text' + }, {}, this.localize('Playing in picture-in-picture'))); + return window$1.documentPictureInPicture.requestWindow({ + // The aspect ratio won't be correct, Chrome bug https://crbug.com/1407629 + width: this.videoWidth(), + height: this.videoHeight() + }).then(pipWindow => { + copyStyleSheetsToWindow(pipWindow); + this.el_.parentNode.insertBefore(pipContainer, this.el_); + pipWindow.document.body.appendChild(this.el_); + pipWindow.document.body.classList.add('vjs-pip-window'); + this.player_.isInPictureInPicture(true); + this.player_.trigger({ + type: 'enterpictureinpicture', + pipWindow + }); + + // Listen for the PiP closing event to move the video back. + pipWindow.addEventListener('pagehide', event => { + const pipVideo = event.target.querySelector('.video-js'); + pipContainer.parentNode.replaceChild(pipVideo, pipContainer); + this.player_.isInPictureInPicture(false); + this.player_.trigger('leavepictureinpicture'); + }); + return pipWindow; + }); + } + if ('pictureInPictureEnabled' in document && this.disablePictureInPicture() === false) { + /** + * This event fires when the player enters picture in picture mode + * + * @event Player#enterpictureinpicture + * @type {Event} + */ + return this.techGet_('requestPictureInPicture'); + } + return Promise.reject('No PiP mode is available'); + } + + /** + * Exit Picture-in-Picture mode. + * + * @see [Spec]{@link https://wicg.github.io/picture-in-picture} + * + * @fires Player#leavepictureinpicture + * + * @return {Promise} + * A promise. + */ + exitPictureInPicture() { + if (window$1.documentPictureInPicture && window$1.documentPictureInPicture.window) { + // With documentPictureInPicture, Player#leavepictureinpicture is fired in the pagehide handler + window$1.documentPictureInPicture.window.close(); + return Promise.resolve(); + } + if ('pictureInPictureEnabled' in document) { + /** + * This event fires when the player leaves picture in picture mode + * + * @event Player#leavepictureinpicture + * @type {Event} + */ + return document.exitPictureInPicture(); + } + } + + /** + * Called when this Player has focus and a key gets pressed down, or when + * any Component of this player receives a key press that it doesn't handle. + * This allows player-wide hotkeys (either as defined below, or optionally + * by an external function). + * + * @param {KeyboardEvent} event + * The `keydown` event that caused this function to be called. + * + * @listens keydown + */ + handleKeyDown(event) { + const { + userActions + } = this.options_; + + // Bail out if hotkeys are not configured. + if (!userActions || !userActions.hotkeys) { + return; + } + + // Function that determines whether or not to exclude an element from + // hotkeys handling. + const excludeElement = el => { + const tagName = el.tagName.toLowerCase(); + + // The first and easiest test is for `contenteditable` elements. + if (el.isContentEditable) { + return true; + } + + // Inputs matching these types will still trigger hotkey handling as + // they are not text inputs. + const allowedInputTypes = ['button', 'checkbox', 'hidden', 'radio', 'reset', 'submit']; + if (tagName === 'input') { + return allowedInputTypes.indexOf(el.type) === -1; + } + + // The final test is by tag name. These tags will be excluded entirely. + const excludedTags = ['textarea']; + return excludedTags.indexOf(tagName) !== -1; + }; + + // Bail out if the user is focused on an interactive form element. + if (excludeElement(this.el_.ownerDocument.activeElement)) { + return; + } + if (typeof userActions.hotkeys === 'function') { + userActions.hotkeys.call(this, event); + } else { + this.handleHotkeys(event); + } + } + + /** + * Called when this Player receives a hotkey keydown event. + * Supported player-wide hotkeys are: + * + * f - toggle fullscreen + * m - toggle mute + * k or Space - toggle play/pause + * + * @param {Event} event + * The `keydown` event that caused this function to be called. + */ + handleHotkeys(event) { + const hotkeys = this.options_.userActions ? this.options_.userActions.hotkeys : {}; + + // set fullscreenKey, muteKey, playPauseKey from `hotkeys`, use defaults if not set + const { + fullscreenKey = keydownEvent => keycode.isEventKey(keydownEvent, 'f'), + muteKey = keydownEvent => keycode.isEventKey(keydownEvent, 'm'), + playPauseKey = keydownEvent => keycode.isEventKey(keydownEvent, 'k') || keycode.isEventKey(keydownEvent, 'Space') + } = hotkeys; + if (fullscreenKey.call(this, event)) { + event.preventDefault(); + event.stopPropagation(); + const FSToggle = Component$1.getComponent('FullscreenToggle'); + if (document[this.fsApi_.fullscreenEnabled] !== false) { + FSToggle.prototype.handleClick.call(this, event); + } + } else if (muteKey.call(this, event)) { + event.preventDefault(); + event.stopPropagation(); + const MuteToggle = Component$1.getComponent('MuteToggle'); + MuteToggle.prototype.handleClick.call(this, event); + } else if (playPauseKey.call(this, event)) { + event.preventDefault(); + event.stopPropagation(); + const PlayToggle = Component$1.getComponent('PlayToggle'); + PlayToggle.prototype.handleClick.call(this, event); + } + } + + /** + * Check whether the player can play a given mimetype + * + * @see https://www.w3.org/TR/2011/WD-html5-20110113/video.html#dom-navigator-canplaytype + * + * @param {string} type + * The mimetype to check + * + * @return {string} + * 'probably', 'maybe', or '' (empty string) + */ + canPlayType(type) { + let can; + + // Loop through each playback technology in the options order + for (let i = 0, j = this.options_.techOrder; i < j.length; i++) { + const techName = j[i]; + let tech = Tech.getTech(techName); + + // Support old behavior of techs being registered as components. + // Remove once that deprecated behavior is removed. + if (!tech) { + tech = Component$1.getComponent(techName); + } + + // Check if the current tech is defined before continuing + if (!tech) { + log$1.error(`The "${techName}" tech is undefined. Skipped browser support check for that tech.`); + continue; + } + + // Check if the browser supports this technology + if (tech.isSupported()) { + can = tech.canPlayType(type); + if (can) { + return can; + } + } + } + return ''; + } + + /** + * Select source based on tech-order or source-order + * Uses source-order selection if `options.sourceOrder` is truthy. Otherwise, + * defaults to tech-order selection + * + * @param {Array} sources + * The sources for a media asset + * + * @return {Object|boolean} + * Object of source and tech order or false + */ + selectSource(sources) { + // Get only the techs specified in `techOrder` that exist and are supported by the + // current platform + const techs = this.options_.techOrder.map(techName => { + return [techName, Tech.getTech(techName)]; + }).filter(([techName, tech]) => { + // Check if the current tech is defined before continuing + if (tech) { + // Check if the browser supports this technology + return tech.isSupported(); + } + log$1.error(`The "${techName}" tech is undefined. Skipped browser support check for that tech.`); + return false; + }); + + // Iterate over each `innerArray` element once per `outerArray` element and execute + // `tester` with both. If `tester` returns a non-falsy value, exit early and return + // that value. + const findFirstPassingTechSourcePair = function (outerArray, innerArray, tester) { + let found; + outerArray.some(outerChoice => { + return innerArray.some(innerChoice => { + found = tester(outerChoice, innerChoice); + if (found) { + return true; + } + }); + }); + return found; + }; + let foundSourceAndTech; + const flip = fn => (a, b) => fn(b, a); + const finder = ([techName, tech], source) => { + if (tech.canPlaySource(source, this.options_[techName.toLowerCase()])) { + return { + source, + tech: techName + }; + } + }; + + // Depending on the truthiness of `options.sourceOrder`, we swap the order of techs and sources + // to select from them based on their priority. + if (this.options_.sourceOrder) { + // Source-first ordering + foundSourceAndTech = findFirstPassingTechSourcePair(sources, techs, flip(finder)); + } else { + // Tech-first ordering + foundSourceAndTech = findFirstPassingTechSourcePair(techs, sources, finder); + } + return foundSourceAndTech || false; + } + + /** + * Executes source setting and getting logic + * + * @param {Tech~SourceObject|Tech~SourceObject[]|string} [source] + * A SourceObject, an array of SourceObjects, or a string referencing + * a URL to a media source. It is _highly recommended_ that an object + * or array of objects is used here, so that source selection + * algorithms can take the `type` into account. + * + * If not provided, this method acts as a getter. + * @param {boolean} [isRetry] + * Indicates whether this is being called internally as a result of a retry + * + * @return {string|undefined} + * If the `source` argument is missing, returns the current source + * URL. Otherwise, returns nothing/undefined. + */ + handleSrc_(source, isRetry) { + // getter usage + if (typeof source === 'undefined') { + return this.cache_.src || ''; + } + + // Reset retry behavior for new source + if (this.resetRetryOnError_) { + this.resetRetryOnError_(); + } + + // filter out invalid sources and turn our source into + // an array of source objects + const sources = filterSource(source); + + // if a source was passed in then it is invalid because + // it was filtered to a zero length Array. So we have to + // show an error + if (!sources.length) { + this.setTimeout(function () { + this.error({ + code: 4, + message: this.options_.notSupportedMessage + }); + }, 0); + return; + } + + // initial sources + this.changingSrc_ = true; + + // Only update the cached source list if we are not retrying a new source after error, + // since in that case we want to include the failed source(s) in the cache + if (!isRetry) { + this.cache_.sources = sources; + } + this.updateSourceCaches_(sources[0]); + + // middlewareSource is the source after it has been changed by middleware + setSource(this, sources[0], (middlewareSource, mws) => { + this.middleware_ = mws; + + // since sourceSet is async we have to update the cache again after we select a source since + // the source that is selected could be out of order from the cache update above this callback. + if (!isRetry) { + this.cache_.sources = sources; + } + this.updateSourceCaches_(middlewareSource); + const err = this.src_(middlewareSource); + if (err) { + if (sources.length > 1) { + return this.handleSrc_(sources.slice(1)); + } + this.changingSrc_ = false; + + // We need to wrap this in a timeout to give folks a chance to add error event handlers + this.setTimeout(function () { + this.error({ + code: 4, + message: this.options_.notSupportedMessage + }); + }, 0); + + // we could not find an appropriate tech, but let's still notify the delegate that this is it + // this needs a better comment about why this is needed + this.triggerReady(); + return; + } + setTech(mws, this.tech_); + }); + + // Try another available source if this one fails before playback. + if (sources.length > 1) { + const retry = () => { + // Remove the error modal + this.error(null); + this.handleSrc_(sources.slice(1), true); + }; + const stopListeningForErrors = () => { + this.off('error', retry); + }; + this.one('error', retry); + this.one('playing', stopListeningForErrors); + this.resetRetryOnError_ = () => { + this.off('error', retry); + this.off('playing', stopListeningForErrors); + }; + } + } + + /** + * Get or set the video source. + * + * @param {Tech~SourceObject|Tech~SourceObject[]|string} [source] + * A SourceObject, an array of SourceObjects, or a string referencing + * a URL to a media source. It is _highly recommended_ that an object + * or array of objects is used here, so that source selection + * algorithms can take the `type` into account. + * + * If not provided, this method acts as a getter. + * + * @return {string|undefined} + * If the `source` argument is missing, returns the current source + * URL. Otherwise, returns nothing/undefined. + */ + src(source) { + return this.handleSrc_(source, false); + } + + /** + * Set the source object on the tech, returns a boolean that indicates whether + * there is a tech that can play the source or not + * + * @param {Tech~SourceObject} source + * The source object to set on the Tech + * + * @return {boolean} + * - True if there is no Tech to playback this source + * - False otherwise + * + * @private + */ + src_(source) { + const sourceTech = this.selectSource([source]); + if (!sourceTech) { + return true; + } + if (!titleCaseEquals(sourceTech.tech, this.techName_)) { + this.changingSrc_ = true; + // load this technology with the chosen source + this.loadTech_(sourceTech.tech, sourceTech.source); + this.tech_.ready(() => { + this.changingSrc_ = false; + }); + return false; + } + + // wait until the tech is ready to set the source + // and set it synchronously if possible (#2326) + this.ready(function () { + // The setSource tech method was added with source handlers + // so older techs won't support it + // We need to check the direct prototype for the case where subclasses + // of the tech do not support source handlers + if (this.tech_.constructor.prototype.hasOwnProperty('setSource')) { + this.techCall_('setSource', source); + } else { + this.techCall_('src', source.src); + } + this.changingSrc_ = false; + }, true); + return false; + } + + /** + * Begin loading the src data. + */ + load() { + // Workaround to use the load method with the VHS. + // Does not cover the case when the load method is called directly from the mediaElement. + if (this.tech_ && this.tech_.vhs) { + this.src(this.currentSource()); + return; + } + this.techCall_('load'); + } + + /** + * Reset the player. Loads the first tech in the techOrder, + * removes all the text tracks in the existing `tech`, + * and calls `reset` on the `tech`. + */ + reset() { + if (this.paused()) { + this.doReset_(); + } else { + const playPromise = this.play(); + silencePromise(playPromise.then(() => this.doReset_())); + } + } + doReset_() { + if (this.tech_) { + this.tech_.clearTracks('text'); + } + this.removeClass('vjs-playing'); + this.addClass('vjs-paused'); + this.resetCache_(); + this.poster(''); + this.loadTech_(this.options_.techOrder[0], null); + this.techCall_('reset'); + this.resetControlBarUI_(); + this.error(null); + if (this.titleBar) { + this.titleBar.update({ + title: undefined, + description: undefined + }); + } + if (isEvented(this)) { + this.trigger('playerreset'); + } + } + + /** + * Reset Control Bar's UI by calling sub-methods that reset + * all of Control Bar's components + */ + resetControlBarUI_() { + this.resetProgressBar_(); + this.resetPlaybackRate_(); + this.resetVolumeBar_(); + } + + /** + * Reset tech's progress so progress bar is reset in the UI + */ + resetProgressBar_() { + this.currentTime(0); + const { + currentTimeDisplay, + durationDisplay, + progressControl, + remainingTimeDisplay + } = this.controlBar || {}; + const { + seekBar + } = progressControl || {}; + if (currentTimeDisplay) { + currentTimeDisplay.updateContent(); + } + if (durationDisplay) { + durationDisplay.updateContent(); + } + if (remainingTimeDisplay) { + remainingTimeDisplay.updateContent(); + } + if (seekBar) { + seekBar.update(); + if (seekBar.loadProgressBar) { + seekBar.loadProgressBar.update(); + } + } + } + + /** + * Reset Playback ratio + */ + resetPlaybackRate_() { + this.playbackRate(this.defaultPlaybackRate()); + this.handleTechRateChange_(); + } + + /** + * Reset Volume bar + */ + resetVolumeBar_() { + this.volume(1.0); + this.trigger('volumechange'); + } + + /** + * Returns all of the current source objects. + * + * @return {Tech~SourceObject[]} + * The current source objects + */ + currentSources() { + const source = this.currentSource(); + const sources = []; + + // assume `{}` or `{ src }` + if (Object.keys(source).length !== 0) { + sources.push(source); + } + return this.cache_.sources || sources; + } + + /** + * Returns the current source object. + * + * @return {Tech~SourceObject} + * The current source object + */ + currentSource() { + return this.cache_.source || {}; + } + + /** + * Returns the fully qualified URL of the current source value e.g. http://mysite.com/video.mp4 + * Can be used in conjunction with `currentType` to assist in rebuilding the current source object. + * + * @return {string} + * The current source + */ + currentSrc() { + return this.currentSource() && this.currentSource().src || ''; + } + + /** + * Get the current source type e.g. video/mp4 + * This can allow you rebuild the current source object so that you could load the same + * source and tech later + * + * @return {string} + * The source MIME type + */ + currentType() { + return this.currentSource() && this.currentSource().type || ''; + } + + /** + * Get or set the preload attribute + * + * @param {'none'|'auto'|'metadata'} [value] + * Preload mode to pass to tech + * + * @return {string|undefined} + * - The preload attribute value when getting + * - Nothing when setting + */ + preload(value) { + if (value !== undefined) { + this.techCall_('setPreload', value); + this.options_.preload = value; + return; + } + return this.techGet_('preload'); + } + + /** + * Get or set the autoplay option. When this is a boolean it will + * modify the attribute on the tech. When this is a string the attribute on + * the tech will be removed and `Player` will handle autoplay on loadstarts. + * + * @param {boolean|'play'|'muted'|'any'} [value] + * - true: autoplay using the browser behavior + * - false: do not autoplay + * - 'play': call play() on every loadstart + * - 'muted': call muted() then play() on every loadstart + * - 'any': call play() on every loadstart. if that fails call muted() then play(). + * - *: values other than those listed here will be set `autoplay` to true + * + * @return {boolean|string|undefined} + * - The current value of autoplay when getting + * - Nothing when setting + */ + autoplay(value) { + // getter usage + if (value === undefined) { + return this.options_.autoplay || false; + } + let techAutoplay; + + // if the value is a valid string set it to that, or normalize `true` to 'play', if need be + if (typeof value === 'string' && /(any|play|muted)/.test(value) || value === true && this.options_.normalizeAutoplay) { + this.options_.autoplay = value; + this.manualAutoplay_(typeof value === 'string' ? value : 'play'); + techAutoplay = false; + + // any falsy value sets autoplay to false in the browser, + // lets do the same + } else if (!value) { + this.options_.autoplay = false; + + // any other value (ie truthy) sets autoplay to true + } else { + this.options_.autoplay = true; + } + techAutoplay = typeof techAutoplay === 'undefined' ? this.options_.autoplay : techAutoplay; + + // if we don't have a tech then we do not queue up + // a setAutoplay call on tech ready. We do this because the + // autoplay option will be passed in the constructor and we + // do not need to set it twice + if (this.tech_) { + this.techCall_('setAutoplay', techAutoplay); + } + } + + /** + * Set or unset the playsinline attribute. + * Playsinline tells the browser that non-fullscreen playback is preferred. + * + * @param {boolean} [value] + * - true means that we should try to play inline by default + * - false means that we should use the browser's default playback mode, + * which in most cases is inline. iOS Safari is a notable exception + * and plays fullscreen by default. + * + * @return {string|undefined} + * - the current value of playsinline + * - Nothing when setting + * + * @see [Spec]{@link https://html.spec.whatwg.org/#attr-video-playsinline} + */ + playsinline(value) { + if (value !== undefined) { + this.techCall_('setPlaysinline', value); + this.options_.playsinline = value; + } + return this.techGet_('playsinline'); + } + + /** + * Get or set the loop attribute on the video element. + * + * @param {boolean} [value] + * - true means that we should loop the video + * - false means that we should not loop the video + * + * @return {boolean|undefined} + * - The current value of loop when getting + * - Nothing when setting + */ + loop(value) { + if (value !== undefined) { + this.techCall_('setLoop', value); + this.options_.loop = value; + return; + } + return this.techGet_('loop'); + } + + /** + * Get or set the poster image source url + * + * @fires Player#posterchange + * + * @param {string} [src] + * Poster image source URL + * + * @return {string|undefined} + * - The current value of poster when getting + * - Nothing when setting + */ + poster(src) { + if (src === undefined) { + return this.poster_; + } + + // The correct way to remove a poster is to set as an empty string + // other falsey values will throw errors + if (!src) { + src = ''; + } + if (src === this.poster_) { + return; + } + + // update the internal poster variable + this.poster_ = src; + + // update the tech's poster + this.techCall_('setPoster', src); + this.isPosterFromTech_ = false; + + // alert components that the poster has been set + /** + * This event fires when the poster image is changed on the player. + * + * @event Player#posterchange + * @type {Event} + */ + this.trigger('posterchange'); + } + + /** + * Some techs (e.g. YouTube) can provide a poster source in an + * asynchronous way. We want the poster component to use this + * poster source so that it covers up the tech's controls. + * (YouTube's play button). However we only want to use this + * source if the player user hasn't set a poster through + * the normal APIs. + * + * @fires Player#posterchange + * @listens Tech#posterchange + * @private + */ + handleTechPosterChange_() { + if ((!this.poster_ || this.options_.techCanOverridePoster) && this.tech_ && this.tech_.poster) { + const newPoster = this.tech_.poster() || ''; + if (newPoster !== this.poster_) { + this.poster_ = newPoster; + this.isPosterFromTech_ = true; + + // Let components know the poster has changed + this.trigger('posterchange'); + } + } + } + + /** + * Get or set whether or not the controls are showing. + * + * @fires Player#controlsenabled + * + * @param {boolean} [bool] + * - true to turn controls on + * - false to turn controls off + * + * @return {boolean|undefined} + * - The current value of controls when getting + * - Nothing when setting + */ + controls(bool) { + if (bool === undefined) { + return !!this.controls_; + } + bool = !!bool; + + // Don't trigger a change event unless it actually changed + if (this.controls_ === bool) { + return; + } + this.controls_ = bool; + if (this.usingNativeControls()) { + this.techCall_('setControls', bool); + } + if (this.controls_) { + this.removeClass('vjs-controls-disabled'); + this.addClass('vjs-controls-enabled'); + /** + * @event Player#controlsenabled + * @type {Event} + */ + this.trigger('controlsenabled'); + if (!this.usingNativeControls()) { + this.addTechControlsListeners_(); + } + } else { + this.removeClass('vjs-controls-enabled'); + this.addClass('vjs-controls-disabled'); + /** + * @event Player#controlsdisabled + * @type {Event} + */ + this.trigger('controlsdisabled'); + if (!this.usingNativeControls()) { + this.removeTechControlsListeners_(); + } + } + } + + /** + * Toggle native controls on/off. Native controls are the controls built into + * devices (e.g. default iPhone controls) or other techs + * (e.g. Vimeo Controls) + * **This should only be set by the current tech, because only the tech knows + * if it can support native controls** + * + * @fires Player#usingnativecontrols + * @fires Player#usingcustomcontrols + * + * @param {boolean} [bool] + * - true to turn native controls on + * - false to turn native controls off + * + * @return {boolean|undefined} + * - The current value of native controls when getting + * - Nothing when setting + */ + usingNativeControls(bool) { + if (bool === undefined) { + return !!this.usingNativeControls_; + } + bool = !!bool; + + // Don't trigger a change event unless it actually changed + if (this.usingNativeControls_ === bool) { + return; + } + this.usingNativeControls_ = bool; + if (this.usingNativeControls_) { + this.addClass('vjs-using-native-controls'); + + /** + * player is using the native device controls + * + * @event Player#usingnativecontrols + * @type {Event} + */ + this.trigger('usingnativecontrols'); + } else { + this.removeClass('vjs-using-native-controls'); + + /** + * player is using the custom HTML controls + * + * @event Player#usingcustomcontrols + * @type {Event} + */ + this.trigger('usingcustomcontrols'); + } + } + + /** + * Set or get the current MediaError + * + * @fires Player#error + * + * @param {MediaError|string|number} [err] + * A MediaError or a string/number to be turned + * into a MediaError + * + * @return {MediaError|null|undefined} + * - The current MediaError when getting (or null) + * - Nothing when setting + */ + error(err) { + if (err === undefined) { + return this.error_ || null; + } + + // allow hooks to modify error object + hooks('beforeerror').forEach(hookFunction => { + const newErr = hookFunction(this, err); + if (!(isObject(newErr) && !Array.isArray(newErr) || typeof newErr === 'string' || typeof newErr === 'number' || newErr === null)) { + this.log.error('please return a value that MediaError expects in beforeerror hooks'); + return; + } + err = newErr; + }); + + // Suppress the first error message for no compatible source until + // user interaction + if (this.options_.suppressNotSupportedError && err && err.code === 4) { + const triggerSuppressedError = function () { + this.error(err); + }; + this.options_.suppressNotSupportedError = false; + this.any(['click', 'touchstart'], triggerSuppressedError); + this.one('loadstart', function () { + this.off(['click', 'touchstart'], triggerSuppressedError); + }); + return; + } + + // restoring to default + if (err === null) { + this.error_ = null; + this.removeClass('vjs-error'); + if (this.errorDisplay) { + this.errorDisplay.close(); + } + return; + } + this.error_ = new MediaError(err); + + // add the vjs-error classname to the player + this.addClass('vjs-error'); + + // log the name of the error type and any message + // IE11 logs "[object object]" and required you to expand message to see error object + log$1.error(`(CODE:${this.error_.code} ${MediaError.errorTypes[this.error_.code]})`, this.error_.message, this.error_); + + /** + * @event Player#error + * @type {Event} + */ + this.trigger('error'); + + // notify hooks of the per player error + hooks('error').forEach(hookFunction => hookFunction(this, this.error_)); + return; + } + + /** + * Report user activity + * + * @param {Object} event + * Event object + */ + reportUserActivity(event) { + this.userActivity_ = true; + } + + /** + * Get/set if user is active + * + * @fires Player#useractive + * @fires Player#userinactive + * + * @param {boolean} [bool] + * - true if the user is active + * - false if the user is inactive + * + * @return {boolean|undefined} + * - The current value of userActive when getting + * - Nothing when setting + */ + userActive(bool) { + if (bool === undefined) { + return this.userActive_; + } + bool = !!bool; + if (bool === this.userActive_) { + return; + } + this.userActive_ = bool; + if (this.userActive_) { + this.userActivity_ = true; + this.removeClass('vjs-user-inactive'); + this.addClass('vjs-user-active'); + /** + * @event Player#useractive + * @type {Event} + */ + this.trigger('useractive'); + return; + } + + // Chrome/Safari/IE have bugs where when you change the cursor it can + // trigger a mousemove event. This causes an issue when you're hiding + // the cursor when the user is inactive, and a mousemove signals user + // activity. Making it impossible to go into inactive mode. Specifically + // this happens in fullscreen when we really need to hide the cursor. + // + // When this gets resolved in ALL browsers it can be removed + // https://code.google.com/p/chromium/issues/detail?id=103041 + if (this.tech_) { + this.tech_.one('mousemove', function (e) { + e.stopPropagation(); + e.preventDefault(); + }); + } + this.userActivity_ = false; + this.removeClass('vjs-user-active'); + this.addClass('vjs-user-inactive'); + /** + * @event Player#userinactive + * @type {Event} + */ + this.trigger('userinactive'); + } + + /** + * Listen for user activity based on timeout value + * + * @private + */ + listenForUserActivity_() { + let mouseInProgress; + let lastMoveX; + let lastMoveY; + const handleActivity = bind_(this, this.reportUserActivity); + const handleMouseMove = function (e) { + // #1068 - Prevent mousemove spamming + // Chrome Bug: https://code.google.com/p/chromium/issues/detail?id=366970 + if (e.screenX !== lastMoveX || e.screenY !== lastMoveY) { + lastMoveX = e.screenX; + lastMoveY = e.screenY; + handleActivity(); + } + }; + const handleMouseDown = function () { + handleActivity(); + // For as long as the they are touching the device or have their mouse down, + // we consider them active even if they're not moving their finger or mouse. + // So we want to continue to update that they are active + this.clearInterval(mouseInProgress); + // Setting userActivity=true now and setting the interval to the same time + // as the activityCheck interval (250) should ensure we never miss the + // next activityCheck + mouseInProgress = this.setInterval(handleActivity, 250); + }; + const handleMouseUpAndMouseLeave = function (event) { + handleActivity(); + // Stop the interval that maintains activity if the mouse/touch is down + this.clearInterval(mouseInProgress); + }; + + // Any mouse movement will be considered user activity + this.on('mousedown', handleMouseDown); + this.on('mousemove', handleMouseMove); + this.on('mouseup', handleMouseUpAndMouseLeave); + this.on('mouseleave', handleMouseUpAndMouseLeave); + const controlBar = this.getChild('controlBar'); + + // Fixes bug on Android & iOS where when tapping progressBar (when control bar is displayed) + // controlBar would no longer be hidden by default timeout. + if (controlBar && !IS_IOS && !IS_ANDROID) { + controlBar.on('mouseenter', function (event) { + if (this.player().options_.inactivityTimeout !== 0) { + this.player().cache_.inactivityTimeout = this.player().options_.inactivityTimeout; + } + this.player().options_.inactivityTimeout = 0; + }); + controlBar.on('mouseleave', function (event) { + this.player().options_.inactivityTimeout = this.player().cache_.inactivityTimeout; + }); + } + + // Listen for keyboard navigation + // Shouldn't need to use inProgress interval because of key repeat + this.on('keydown', handleActivity); + this.on('keyup', handleActivity); + + // Run an interval every 250 milliseconds instead of stuffing everything into + // the mousemove/touchmove function itself, to prevent performance degradation. + // `this.reportUserActivity` simply sets this.userActivity_ to true, which + // then gets picked up by this loop + // http://ejohn.org/blog/learning-from-twitter/ + let inactivityTimeout; + + /** @this Player */ + const activityCheck = function () { + // Check to see if mouse/touch activity has happened + if (!this.userActivity_) { + return; + } + + // Reset the activity tracker + this.userActivity_ = false; + + // If the user state was inactive, set the state to active + this.userActive(true); + + // Clear any existing inactivity timeout to start the timer over + this.clearTimeout(inactivityTimeout); + const timeout = this.options_.inactivityTimeout; + if (timeout <= 0) { + return; + } + + // In <timeout> milliseconds, if no more activity has occurred the + // user will be considered inactive + inactivityTimeout = this.setTimeout(function () { + // Protect against the case where the inactivityTimeout can trigger just + // before the next user activity is picked up by the activity check loop + // causing a flicker + if (!this.userActivity_) { + this.userActive(false); + } + }, timeout); + }; + this.setInterval(activityCheck, 250); + } + + /** + * Gets or sets the current playback rate. A playback rate of + * 1.0 represents normal speed and 0.5 would indicate half-speed + * playback, for instance. + * + * @see https://html.spec.whatwg.org/multipage/embedded-content.html#dom-media-playbackrate + * + * @param {number} [rate] + * New playback rate to set. + * + * @return {number|undefined} + * - The current playback rate when getting or 1.0 + * - Nothing when setting + */ + playbackRate(rate) { + if (rate !== undefined) { + // NOTE: this.cache_.lastPlaybackRate is set from the tech handler + // that is registered above + this.techCall_('setPlaybackRate', rate); + return; + } + if (this.tech_ && this.tech_.featuresPlaybackRate) { + return this.cache_.lastPlaybackRate || this.techGet_('playbackRate'); + } + return 1.0; + } + + /** + * Gets or sets the current default playback rate. A default playback rate of + * 1.0 represents normal speed and 0.5 would indicate half-speed playback, for instance. + * defaultPlaybackRate will only represent what the initial playbackRate of a video was, not + * not the current playbackRate. + * + * @see https://html.spec.whatwg.org/multipage/embedded-content.html#dom-media-defaultplaybackrate + * + * @param {number} [rate] + * New default playback rate to set. + * + * @return {number|undefined} + * - The default playback rate when getting or 1.0 + * - Nothing when setting + */ + defaultPlaybackRate(rate) { + if (rate !== undefined) { + return this.techCall_('setDefaultPlaybackRate', rate); + } + if (this.tech_ && this.tech_.featuresPlaybackRate) { + return this.techGet_('defaultPlaybackRate'); + } + return 1.0; + } + + /** + * Gets or sets the audio flag + * + * @param {boolean} [bool] + * - true signals that this is an audio player + * - false signals that this is not an audio player + * + * @return {boolean|undefined} + * - The current value of isAudio when getting + * - Nothing when setting + */ + isAudio(bool) { + if (bool !== undefined) { + this.isAudio_ = !!bool; + return; + } + return !!this.isAudio_; + } + enableAudioOnlyUI_() { + // Update styling immediately to show the control bar so we can get its height + this.addClass('vjs-audio-only-mode'); + const playerChildren = this.children(); + const controlBar = this.getChild('ControlBar'); + const controlBarHeight = controlBar && controlBar.currentHeight(); + + // Hide all player components except the control bar. Control bar components + // needed only for video are hidden with CSS + playerChildren.forEach(child => { + if (child === controlBar) { + return; + } + if (child.el_ && !child.hasClass('vjs-hidden')) { + child.hide(); + this.audioOnlyCache_.hiddenChildren.push(child); + } + }); + this.audioOnlyCache_.playerHeight = this.currentHeight(); + + // Set the player height the same as the control bar + this.height(controlBarHeight); + this.trigger('audioonlymodechange'); + } + disableAudioOnlyUI_() { + this.removeClass('vjs-audio-only-mode'); + + // Show player components that were previously hidden + this.audioOnlyCache_.hiddenChildren.forEach(child => child.show()); + + // Reset player height + this.height(this.audioOnlyCache_.playerHeight); + this.trigger('audioonlymodechange'); + } + + /** + * Get the current audioOnlyMode state or set audioOnlyMode to true or false. + * + * Setting this to `true` will hide all player components except the control bar, + * as well as control bar components needed only for video. + * + * @param {boolean} [value] + * The value to set audioOnlyMode to. + * + * @return {Promise|boolean} + * A Promise is returned when setting the state, and a boolean when getting + * the present state + */ + audioOnlyMode(value) { + if (typeof value !== 'boolean' || value === this.audioOnlyMode_) { + return this.audioOnlyMode_; + } + this.audioOnlyMode_ = value; + + // Enable Audio Only Mode + if (value) { + const exitPromises = []; + + // Fullscreen and PiP are not supported in audioOnlyMode, so exit if we need to. + if (this.isInPictureInPicture()) { + exitPromises.push(this.exitPictureInPicture()); + } + if (this.isFullscreen()) { + exitPromises.push(this.exitFullscreen()); + } + if (this.audioPosterMode()) { + exitPromises.push(this.audioPosterMode(false)); + } + return Promise.all(exitPromises).then(() => this.enableAudioOnlyUI_()); + } + + // Disable Audio Only Mode + return Promise.resolve().then(() => this.disableAudioOnlyUI_()); + } + enablePosterModeUI_() { + // Hide the video element and show the poster image to enable posterModeUI + const tech = this.tech_ && this.tech_; + tech.hide(); + this.addClass('vjs-audio-poster-mode'); + this.trigger('audiopostermodechange'); + } + disablePosterModeUI_() { + // Show the video element and hide the poster image to disable posterModeUI + const tech = this.tech_ && this.tech_; + tech.show(); + this.removeClass('vjs-audio-poster-mode'); + this.trigger('audiopostermodechange'); + } + + /** + * Get the current audioPosterMode state or set audioPosterMode to true or false + * + * @param {boolean} [value] + * The value to set audioPosterMode to. + * + * @return {Promise|boolean} + * A Promise is returned when setting the state, and a boolean when getting + * the present state + */ + audioPosterMode(value) { + if (typeof value !== 'boolean' || value === this.audioPosterMode_) { + return this.audioPosterMode_; + } + this.audioPosterMode_ = value; + if (value) { + if (this.audioOnlyMode()) { + const audioOnlyModePromise = this.audioOnlyMode(false); + return audioOnlyModePromise.then(() => { + // enable audio poster mode after audio only mode is disabled + this.enablePosterModeUI_(); + }); + } + return Promise.resolve().then(() => { + // enable audio poster mode + this.enablePosterModeUI_(); + }); + } + return Promise.resolve().then(() => { + // disable audio poster mode + this.disablePosterModeUI_(); + }); + } + + /** + * A helper method for adding a {@link TextTrack} to our + * {@link TextTrackList}. + * + * In addition to the W3C settings we allow adding additional info through options. + * + * @see http://www.w3.org/html/wg/drafts/html/master/embedded-content-0.html#dom-media-addtexttrack + * + * @param {string} [kind] + * the kind of TextTrack you are adding + * + * @param {string} [label] + * the label to give the TextTrack label + * + * @param {string} [language] + * the language to set on the TextTrack + * + * @return {TextTrack|undefined} + * the TextTrack that was added or undefined + * if there is no tech + */ + addTextTrack(kind, label, language) { + if (this.tech_) { + return this.tech_.addTextTrack(kind, label, language); + } + } + + /** + * Create a remote {@link TextTrack} and an {@link HTMLTrackElement}. + * + * @param {Object} options + * Options to pass to {@link HTMLTrackElement} during creation. See + * {@link HTMLTrackElement} for object properties that you should use. + * + * @param {boolean} [manualCleanup=false] if set to true, the TextTrack will not be removed + * from the TextTrackList and HtmlTrackElementList + * after a source change + * + * @return { import('./tracks/html-track-element').default } + * the HTMLTrackElement that was created and added + * to the HtmlTrackElementList and the remote + * TextTrackList + * + */ + addRemoteTextTrack(options, manualCleanup) { + if (this.tech_) { + return this.tech_.addRemoteTextTrack(options, manualCleanup); + } + } + + /** + * Remove a remote {@link TextTrack} from the respective + * {@link TextTrackList} and {@link HtmlTrackElementList}. + * + * @param {Object} track + * Remote {@link TextTrack} to remove + * + * @return {undefined} + * does not return anything + */ + removeRemoteTextTrack(obj = {}) { + let { + track + } = obj; + if (!track) { + track = obj; + } + + // destructure the input into an object with a track argument, defaulting to arguments[0] + // default the whole argument to an empty object if nothing was passed in + + if (this.tech_) { + return this.tech_.removeRemoteTextTrack(track); + } + } + + /** + * Gets available media playback quality metrics as specified by the W3C's Media + * Playback Quality API. + * + * @see [Spec]{@link https://wicg.github.io/media-playback-quality} + * + * @return {Object|undefined} + * An object with supported media playback quality metrics or undefined if there + * is no tech or the tech does not support it. + */ + getVideoPlaybackQuality() { + return this.techGet_('getVideoPlaybackQuality'); + } + + /** + * Get video width + * + * @return {number} + * current video width + */ + videoWidth() { + return this.tech_ && this.tech_.videoWidth && this.tech_.videoWidth() || 0; + } + + /** + * Get video height + * + * @return {number} + * current video height + */ + videoHeight() { + return this.tech_ && this.tech_.videoHeight && this.tech_.videoHeight() || 0; + } + + /** + * Set or get the player's language code. + * + * Changing the language will trigger + * [languagechange]{@link Player#event:languagechange} + * which Components can use to update control text. + * ClickableComponent will update its control text by default on + * [languagechange]{@link Player#event:languagechange}. + * + * @fires Player#languagechange + * + * @param {string} [code] + * the language code to set the player to + * + * @return {string|undefined} + * - The current language code when getting + * - Nothing when setting + */ + language(code) { + if (code === undefined) { + return this.language_; + } + if (this.language_ !== String(code).toLowerCase()) { + this.language_ = String(code).toLowerCase(); + + // during first init, it's possible some things won't be evented + if (isEvented(this)) { + /** + * fires when the player language change + * + * @event Player#languagechange + * @type {Event} + */ + this.trigger('languagechange'); + } + } + } + + /** + * Get the player's language dictionary + * Merge every time, because a newly added plugin might call videojs.addLanguage() at any time + * Languages specified directly in the player options have precedence + * + * @return {Array} + * An array of of supported languages + */ + languages() { + return merge$1(Player.prototype.options_.languages, this.languages_); + } + + /** + * returns a JavaScript object representing the current track + * information. **DOES not return it as JSON** + * + * @return {Object} + * Object representing the current of track info + */ + toJSON() { + const options = merge$1(this.options_); + const tracks = options.tracks; + options.tracks = []; + for (let i = 0; i < tracks.length; i++) { + let track = tracks[i]; + + // deep merge tracks and null out player so no circular references + track = merge$1(track); + track.player = undefined; + options.tracks[i] = track; + } + return options; + } + + /** + * Creates a simple modal dialog (an instance of the {@link ModalDialog} + * component) that immediately overlays the player with arbitrary + * content and removes itself when closed. + * + * @param {string|Function|Element|Array|null} content + * Same as {@link ModalDialog#content}'s param of the same name. + * The most straight-forward usage is to provide a string or DOM + * element. + * + * @param {Object} [options] + * Extra options which will be passed on to the {@link ModalDialog}. + * + * @return {ModalDialog} + * the {@link ModalDialog} that was created + */ + createModal(content, options) { + options = options || {}; + options.content = content || ''; + const modal = new ModalDialog(this, options); + this.addChild(modal); + modal.on('dispose', () => { + this.removeChild(modal); + }); + modal.open(); + return modal; + } + + /** + * Change breakpoint classes when the player resizes. + * + * @private + */ + updateCurrentBreakpoint_() { + if (!this.responsive()) { + return; + } + const currentBreakpoint = this.currentBreakpoint(); + const currentWidth = this.currentWidth(); + for (let i = 0; i < BREAKPOINT_ORDER.length; i++) { + const candidateBreakpoint = BREAKPOINT_ORDER[i]; + const maxWidth = this.breakpoints_[candidateBreakpoint]; + if (currentWidth <= maxWidth) { + // The current breakpoint did not change, nothing to do. + if (currentBreakpoint === candidateBreakpoint) { + return; + } + + // Only remove a class if there is a current breakpoint. + if (currentBreakpoint) { + this.removeClass(BREAKPOINT_CLASSES[currentBreakpoint]); + } + this.addClass(BREAKPOINT_CLASSES[candidateBreakpoint]); + this.breakpoint_ = candidateBreakpoint; + break; + } + } + } + + /** + * Removes the current breakpoint. + * + * @private + */ + removeCurrentBreakpoint_() { + const className = this.currentBreakpointClass(); + this.breakpoint_ = ''; + if (className) { + this.removeClass(className); + } + } + + /** + * Get or set breakpoints on the player. + * + * Calling this method with an object or `true` will remove any previous + * custom breakpoints and start from the defaults again. + * + * @param {Object|boolean} [breakpoints] + * If an object is given, it can be used to provide custom + * breakpoints. If `true` is given, will set default breakpoints. + * If this argument is not given, will simply return the current + * breakpoints. + * + * @param {number} [breakpoints.tiny] + * The maximum width for the "vjs-layout-tiny" class. + * + * @param {number} [breakpoints.xsmall] + * The maximum width for the "vjs-layout-x-small" class. + * + * @param {number} [breakpoints.small] + * The maximum width for the "vjs-layout-small" class. + * + * @param {number} [breakpoints.medium] + * The maximum width for the "vjs-layout-medium" class. + * + * @param {number} [breakpoints.large] + * The maximum width for the "vjs-layout-large" class. + * + * @param {number} [breakpoints.xlarge] + * The maximum width for the "vjs-layout-x-large" class. + * + * @param {number} [breakpoints.huge] + * The maximum width for the "vjs-layout-huge" class. + * + * @return {Object} + * An object mapping breakpoint names to maximum width values. + */ + breakpoints(breakpoints) { + // Used as a getter. + if (breakpoints === undefined) { + return Object.assign(this.breakpoints_); + } + this.breakpoint_ = ''; + this.breakpoints_ = Object.assign({}, DEFAULT_BREAKPOINTS, breakpoints); + + // When breakpoint definitions change, we need to update the currently + // selected breakpoint. + this.updateCurrentBreakpoint_(); + + // Clone the breakpoints before returning. + return Object.assign(this.breakpoints_); + } + + /** + * Get or set a flag indicating whether or not this player should adjust + * its UI based on its dimensions. + * + * @param {boolean} [value] + * Should be `true` if the player should adjust its UI based on its + * dimensions; otherwise, should be `false`. + * + * @return {boolean|undefined} + * Will be `true` if this player should adjust its UI based on its + * dimensions; otherwise, will be `false`. + * Nothing if setting + */ + responsive(value) { + // Used as a getter. + if (value === undefined) { + return this.responsive_; + } + value = Boolean(value); + const current = this.responsive_; + + // Nothing changed. + if (value === current) { + return; + } + + // The value actually changed, set it. + this.responsive_ = value; + + // Start listening for breakpoints and set the initial breakpoint if the + // player is now responsive. + if (value) { + this.on('playerresize', this.boundUpdateCurrentBreakpoint_); + this.updateCurrentBreakpoint_(); + + // Stop listening for breakpoints if the player is no longer responsive. + } else { + this.off('playerresize', this.boundUpdateCurrentBreakpoint_); + this.removeCurrentBreakpoint_(); + } + return value; + } + + /** + * Get current breakpoint name, if any. + * + * @return {string} + * If there is currently a breakpoint set, returns a the key from the + * breakpoints object matching it. Otherwise, returns an empty string. + */ + currentBreakpoint() { + return this.breakpoint_; + } + + /** + * Get the current breakpoint class name. + * + * @return {string} + * The matching class name (e.g. `"vjs-layout-tiny"` or + * `"vjs-layout-large"`) for the current breakpoint. Empty string if + * there is no current breakpoint. + */ + currentBreakpointClass() { + return BREAKPOINT_CLASSES[this.breakpoint_] || ''; + } + + /** + * An object that describes a single piece of media. + * + * Properties that are not part of this type description will be retained; so, + * this can be viewed as a generic metadata storage mechanism as well. + * + * @see {@link https://wicg.github.io/mediasession/#the-mediametadata-interface} + * @typedef {Object} Player~MediaObject + * + * @property {string} [album] + * Unused, except if this object is passed to the `MediaSession` + * API. + * + * @property {string} [artist] + * Unused, except if this object is passed to the `MediaSession` + * API. + * + * @property {Object[]} [artwork] + * Unused, except if this object is passed to the `MediaSession` + * API. If not specified, will be populated via the `poster`, if + * available. + * + * @property {string} [poster] + * URL to an image that will display before playback. + * + * @property {Tech~SourceObject|Tech~SourceObject[]|string} [src] + * A single source object, an array of source objects, or a string + * referencing a URL to a media source. It is _highly recommended_ + * that an object or array of objects is used here, so that source + * selection algorithms can take the `type` into account. + * + * @property {string} [title] + * Unused, except if this object is passed to the `MediaSession` + * API. + * + * @property {Object[]} [textTracks] + * An array of objects to be used to create text tracks, following + * the {@link https://www.w3.org/TR/html50/embedded-content-0.html#the-track-element|native track element format}. + * For ease of removal, these will be created as "remote" text + * tracks and set to automatically clean up on source changes. + * + * These objects may have properties like `src`, `kind`, `label`, + * and `language`, see {@link Tech#createRemoteTextTrack}. + */ + + /** + * Populate the player using a {@link Player~MediaObject|MediaObject}. + * + * @param {Player~MediaObject} media + * A media object. + * + * @param {Function} ready + * A callback to be called when the player is ready. + */ + loadMedia(media, ready) { + if (!media || typeof media !== 'object') { + return; + } + const crossOrigin = this.crossOrigin(); + this.reset(); + + // Clone the media object so it cannot be mutated from outside. + this.cache_.media = merge$1(media); + const { + artist, + artwork, + description, + poster, + src, + textTracks, + title + } = this.cache_.media; + + // If `artwork` is not given, create it using `poster`. + if (!artwork && poster) { + this.cache_.media.artwork = [{ + src: poster, + type: getMimetype(poster) + }]; + } + if (crossOrigin) { + this.crossOrigin(crossOrigin); + } + if (src) { + this.src(src); + } + if (poster) { + this.poster(poster); + } + if (Array.isArray(textTracks)) { + textTracks.forEach(tt => this.addRemoteTextTrack(tt, false)); + } + if (this.titleBar) { + this.titleBar.update({ + title, + description: description || artist || '' + }); + } + this.ready(ready); + } + + /** + * Get a clone of the current {@link Player~MediaObject} for this player. + * + * If the `loadMedia` method has not been used, will attempt to return a + * {@link Player~MediaObject} based on the current state of the player. + * + * @return {Player~MediaObject} + */ + getMedia() { + if (!this.cache_.media) { + const poster = this.poster(); + const src = this.currentSources(); + const textTracks = Array.prototype.map.call(this.remoteTextTracks(), tt => ({ + kind: tt.kind, + label: tt.label, + language: tt.language, + src: tt.src + })); + const media = { + src, + textTracks + }; + if (poster) { + media.poster = poster; + media.artwork = [{ + src: media.poster, + type: getMimetype(media.poster) + }]; + } + return media; + } + return merge$1(this.cache_.media); + } + + /** + * Gets tag settings + * + * @param {Element} tag + * The player tag + * + * @return {Object} + * An object containing all of the settings + * for a player tag + */ + static getTagSettings(tag) { + const baseOptions = { + sources: [], + tracks: [] + }; + const tagOptions = getAttributes(tag); + const dataSetup = tagOptions['data-setup']; + if (hasClass(tag, 'vjs-fill')) { + tagOptions.fill = true; + } + if (hasClass(tag, 'vjs-fluid')) { + tagOptions.fluid = true; + } + + // Check if data-setup attr exists. + if (dataSetup !== null) { + // Parse options JSON + // If empty string, make it a parsable json object. + const [err, data] = safeParseTuple(dataSetup || '{}'); + if (err) { + log$1.error(err); + } + Object.assign(tagOptions, data); + } + Object.assign(baseOptions, tagOptions); + + // Get tag children settings + if (tag.hasChildNodes()) { + const children = tag.childNodes; + for (let i = 0, j = children.length; i < j; i++) { + const child = children[i]; + // Change case needed: http://ejohn.org/blog/nodename-case-sensitivity/ + const childName = child.nodeName.toLowerCase(); + if (childName === 'source') { + baseOptions.sources.push(getAttributes(child)); + } else if (childName === 'track') { + baseOptions.tracks.push(getAttributes(child)); + } + } + } + return baseOptions; + } + + /** + * Set debug mode to enable/disable logs at info level. + * + * @param {boolean} enabled + * @fires Player#debugon + * @fires Player#debugoff + * @return {boolean|undefined} + */ + debug(enabled) { + if (enabled === undefined) { + return this.debugEnabled_; + } + if (enabled) { + this.trigger('debugon'); + this.previousLogLevel_ = this.log.level; + this.log.level('debug'); + this.debugEnabled_ = true; + } else { + this.trigger('debugoff'); + this.log.level(this.previousLogLevel_); + this.previousLogLevel_ = undefined; + this.debugEnabled_ = false; + } + } + + /** + * Set or get current playback rates. + * Takes an array and updates the playback rates menu with the new items. + * Pass in an empty array to hide the menu. + * Values other than arrays are ignored. + * + * @fires Player#playbackrateschange + * @param {number[]} [newRates] + * The new rates that the playback rates menu should update to. + * An empty array will hide the menu + * @return {number[]} When used as a getter will return the current playback rates + */ + playbackRates(newRates) { + if (newRates === undefined) { + return this.cache_.playbackRates; + } + + // ignore any value that isn't an array + if (!Array.isArray(newRates)) { + return; + } + + // ignore any arrays that don't only contain numbers + if (!newRates.every(rate => typeof rate === 'number')) { + return; + } + this.cache_.playbackRates = newRates; + + /** + * fires when the playback rates in a player are changed + * + * @event Player#playbackrateschange + * @type {Event} + */ + this.trigger('playbackrateschange'); + } +} + +/** + * Get the {@link VideoTrackList} + * + * @link https://html.spec.whatwg.org/multipage/embedded-content.html#videotracklist + * + * @return {VideoTrackList} + * the current video track list + * + * @method Player.prototype.videoTracks + */ + +/** + * Get the {@link AudioTrackList} + * + * @link https://html.spec.whatwg.org/multipage/embedded-content.html#audiotracklist + * + * @return {AudioTrackList} + * the current audio track list + * + * @method Player.prototype.audioTracks + */ + +/** + * Get the {@link TextTrackList} + * + * @link http://www.w3.org/html/wg/drafts/html/master/embedded-content-0.html#dom-media-texttracks + * + * @return {TextTrackList} + * the current text track list + * + * @method Player.prototype.textTracks + */ + +/** + * Get the remote {@link TextTrackList} + * + * @return {TextTrackList} + * The current remote text track list + * + * @method Player.prototype.remoteTextTracks + */ + +/** + * Get the remote {@link HtmlTrackElementList} tracks. + * + * @return {HtmlTrackElementList} + * The current remote text track element list + * + * @method Player.prototype.remoteTextTrackEls + */ + +ALL.names.forEach(function (name) { + const props = ALL[name]; + Player.prototype[props.getterName] = function () { + if (this.tech_) { + return this.tech_[props.getterName](); + } + + // if we have not yet loadTech_, we create {video,audio,text}Tracks_ + // these will be passed to the tech during loading + this[props.privateName] = this[props.privateName] || new props.ListClass(); + return this[props.privateName]; + }; +}); + +/** + * Get or set the `Player`'s crossorigin option. For the HTML5 player, this + * sets the `crossOrigin` property on the `<video>` tag to control the CORS + * behavior. + * + * @see [Video Element Attributes]{@link https://developer.mozilla.org/en-US/docs/Web/HTML/Element/video#attr-crossorigin} + * + * @param {string} [value] + * The value to set the `Player`'s crossorigin to. If an argument is + * given, must be one of `anonymous` or `use-credentials`. + * + * @return {string|undefined} + * - The current crossorigin value of the `Player` when getting. + * - undefined when setting + */ +Player.prototype.crossorigin = Player.prototype.crossOrigin; + +/** + * Global enumeration of players. + * + * The keys are the player IDs and the values are either the {@link Player} + * instance or `null` for disposed players. + * + * @type {Object} + */ +Player.players = {}; +const navigator = window$1.navigator; + +/* + * Player instance options, surfaced using options + * options = Player.prototype.options_ + * Make changes in options, not here. + * + * @type {Object} + * @private + */ +Player.prototype.options_ = { + // Default order of fallback technology + techOrder: Tech.defaultTechOrder_, + html5: {}, + // enable sourceset by default + enableSourceset: true, + // default inactivity timeout + inactivityTimeout: 2000, + // default playback rates + playbackRates: [], + // Add playback rate selection by adding rates + // 'playbackRates': [0.5, 1, 1.5, 2], + liveui: false, + // Included control sets + children: ['mediaLoader', 'posterImage', 'titleBar', 'textTrackDisplay', 'loadingSpinner', 'bigPlayButton', 'liveTracker', 'controlBar', 'errorDisplay', 'textTrackSettings', 'resizeManager'], + language: navigator && (navigator.languages && navigator.languages[0] || navigator.userLanguage || navigator.language) || 'en', + // locales and their language translations + languages: {}, + // Default message to show when a video cannot be played. + notSupportedMessage: 'No compatible source was found for this media.', + normalizeAutoplay: false, + fullscreen: { + options: { + navigationUI: 'hide' + } + }, + breakpoints: {}, + responsive: false, + audioOnlyMode: false, + audioPosterMode: false, + // Default smooth seeking to false + enableSmoothSeeking: false +}; +TECH_EVENTS_RETRIGGER.forEach(function (event) { + Player.prototype[`handleTech${toTitleCase$1(event)}_`] = function () { + return this.trigger(event); + }; +}); + +/** + * Fired when the player has initial duration and dimension information + * + * @event Player#loadedmetadata + * @type {Event} + */ + +/** + * Fired when the player has downloaded data at the current playback position + * + * @event Player#loadeddata + * @type {Event} + */ + +/** + * Fired when the current playback position has changed * + * During playback this is fired every 15-250 milliseconds, depending on the + * playback technology in use. + * + * @event Player#timeupdate + * @type {Event} + */ + +/** + * Fired when the volume changes + * + * @event Player#volumechange + * @type {Event} + */ + +/** + * Reports whether or not a player has a plugin available. + * + * This does not report whether or not the plugin has ever been initialized + * on this player. For that, [usingPlugin]{@link Player#usingPlugin}. + * + * @method Player#hasPlugin + * @param {string} name + * The name of a plugin. + * + * @return {boolean} + * Whether or not this player has the requested plugin available. + */ + +/** + * Reports whether or not a player is using a plugin by name. + * + * For basic plugins, this only reports whether the plugin has _ever_ been + * initialized on this player. + * + * @method Player#usingPlugin + * @param {string} name + * The name of a plugin. + * + * @return {boolean} + * Whether or not this player is using the requested plugin. + */ + +Component$1.registerComponent('Player', Player); + +/** + * @file plugin.js + */ + +/** + * The base plugin name. + * + * @private + * @constant + * @type {string} + */ +const BASE_PLUGIN_NAME = 'plugin'; + +/** + * The key on which a player's active plugins cache is stored. + * + * @private + * @constant + * @type {string} + */ +const PLUGIN_CACHE_KEY = 'activePlugins_'; + +/** + * Stores registered plugins in a private space. + * + * @private + * @type {Object} + */ +const pluginStorage = {}; + +/** + * Reports whether or not a plugin has been registered. + * + * @private + * @param {string} name + * The name of a plugin. + * + * @return {boolean} + * Whether or not the plugin has been registered. + */ +const pluginExists = name => pluginStorage.hasOwnProperty(name); + +/** + * Get a single registered plugin by name. + * + * @private + * @param {string} name + * The name of a plugin. + * + * @return {typeof Plugin|Function|undefined} + * The plugin (or undefined). + */ +const getPlugin = name => pluginExists(name) ? pluginStorage[name] : undefined; + +/** + * Marks a plugin as "active" on a player. + * + * Also, ensures that the player has an object for tracking active plugins. + * + * @private + * @param {Player} player + * A Video.js player instance. + * + * @param {string} name + * The name of a plugin. + */ +const markPluginAsActive = (player, name) => { + player[PLUGIN_CACHE_KEY] = player[PLUGIN_CACHE_KEY] || {}; + player[PLUGIN_CACHE_KEY][name] = true; +}; + +/** + * Triggers a pair of plugin setup events. + * + * @private + * @param {Player} player + * A Video.js player instance. + * + * @param {PluginEventHash} hash + * A plugin event hash. + * + * @param {boolean} [before] + * If true, prefixes the event name with "before". In other words, + * use this to trigger "beforepluginsetup" instead of "pluginsetup". + */ +const triggerSetupEvent = (player, hash, before) => { + const eventName = (before ? 'before' : '') + 'pluginsetup'; + player.trigger(eventName, hash); + player.trigger(eventName + ':' + hash.name, hash); +}; + +/** + * Takes a basic plugin function and returns a wrapper function which marks + * on the player that the plugin has been activated. + * + * @private + * @param {string} name + * The name of the plugin. + * + * @param {Function} plugin + * The basic plugin. + * + * @return {Function} + * A wrapper function for the given plugin. + */ +const createBasicPlugin = function (name, plugin) { + const basicPluginWrapper = function () { + // We trigger the "beforepluginsetup" and "pluginsetup" events on the player + // regardless, but we want the hash to be consistent with the hash provided + // for advanced plugins. + // + // The only potentially counter-intuitive thing here is the `instance` in + // the "pluginsetup" event is the value returned by the `plugin` function. + triggerSetupEvent(this, { + name, + plugin, + instance: null + }, true); + const instance = plugin.apply(this, arguments); + markPluginAsActive(this, name); + triggerSetupEvent(this, { + name, + plugin, + instance + }); + return instance; + }; + Object.keys(plugin).forEach(function (prop) { + basicPluginWrapper[prop] = plugin[prop]; + }); + return basicPluginWrapper; +}; + +/** + * Takes a plugin sub-class and returns a factory function for generating + * instances of it. + * + * This factory function will replace itself with an instance of the requested + * sub-class of Plugin. + * + * @private + * @param {string} name + * The name of the plugin. + * + * @param {Plugin} PluginSubClass + * The advanced plugin. + * + * @return {Function} + */ +const createPluginFactory = (name, PluginSubClass) => { + // Add a `name` property to the plugin prototype so that each plugin can + // refer to itself by name. + PluginSubClass.prototype.name = name; + return function (...args) { + triggerSetupEvent(this, { + name, + plugin: PluginSubClass, + instance: null + }, true); + const instance = new PluginSubClass(...[this, ...args]); + + // The plugin is replaced by a function that returns the current instance. + this[name] = () => instance; + triggerSetupEvent(this, instance.getEventHash()); + return instance; + }; +}; + +/** + * Parent class for all advanced plugins. + * + * @mixes module:evented~EventedMixin + * @mixes module:stateful~StatefulMixin + * @fires Player#beforepluginsetup + * @fires Player#beforepluginsetup:$name + * @fires Player#pluginsetup + * @fires Player#pluginsetup:$name + * @listens Player#dispose + * @throws {Error} + * If attempting to instantiate the base {@link Plugin} class + * directly instead of via a sub-class. + */ +class Plugin { + /** + * Creates an instance of this class. + * + * Sub-classes should call `super` to ensure plugins are properly initialized. + * + * @param {Player} player + * A Video.js player instance. + */ + constructor(player) { + if (this.constructor === Plugin) { + throw new Error('Plugin must be sub-classed; not directly instantiated.'); + } + this.player = player; + if (!this.log) { + this.log = this.player.log.createLogger(this.name); + } + + // Make this object evented, but remove the added `trigger` method so we + // use the prototype version instead. + evented(this); + delete this.trigger; + stateful(this, this.constructor.defaultState); + markPluginAsActive(player, this.name); + + // Auto-bind the dispose method so we can use it as a listener and unbind + // it later easily. + this.dispose = this.dispose.bind(this); + + // If the player is disposed, dispose the plugin. + player.on('dispose', this.dispose); + } + + /** + * Get the version of the plugin that was set on <pluginName>.VERSION + */ + version() { + return this.constructor.VERSION; + } + + /** + * Each event triggered by plugins includes a hash of additional data with + * conventional properties. + * + * This returns that object or mutates an existing hash. + * + * @param {Object} [hash={}] + * An object to be used as event an event hash. + * + * @return {PluginEventHash} + * An event hash object with provided properties mixed-in. + */ + getEventHash(hash = {}) { + hash.name = this.name; + hash.plugin = this.constructor; + hash.instance = this; + return hash; + } + + /** + * Triggers an event on the plugin object and overrides + * {@link module:evented~EventedMixin.trigger|EventedMixin.trigger}. + * + * @param {string|Object} event + * An event type or an object with a type property. + * + * @param {Object} [hash={}] + * Additional data hash to merge with a + * {@link PluginEventHash|PluginEventHash}. + * + * @return {boolean} + * Whether or not default was prevented. + */ + trigger(event, hash = {}) { + return trigger(this.eventBusEl_, event, this.getEventHash(hash)); + } + + /** + * Handles "statechanged" events on the plugin. No-op by default, override by + * subclassing. + * + * @abstract + * @param {Event} e + * An event object provided by a "statechanged" event. + * + * @param {Object} e.changes + * An object describing changes that occurred with the "statechanged" + * event. + */ + handleStateChanged(e) {} + + /** + * Disposes a plugin. + * + * Subclasses can override this if they want, but for the sake of safety, + * it's probably best to subscribe the "dispose" event. + * + * @fires Plugin#dispose + */ + dispose() { + const { + name, + player + } = this; + + /** + * Signals that a advanced plugin is about to be disposed. + * + * @event Plugin#dispose + * @type {Event} + */ + this.trigger('dispose'); + this.off(); + player.off('dispose', this.dispose); + + // Eliminate any possible sources of leaking memory by clearing up + // references between the player and the plugin instance and nulling out + // the plugin's state and replacing methods with a function that throws. + player[PLUGIN_CACHE_KEY][name] = false; + this.player = this.state = null; + + // Finally, replace the plugin name on the player with a new factory + // function, so that the plugin is ready to be set up again. + player[name] = createPluginFactory(name, pluginStorage[name]); + } + + /** + * Determines if a plugin is a basic plugin (i.e. not a sub-class of `Plugin`). + * + * @param {string|Function} plugin + * If a string, matches the name of a plugin. If a function, will be + * tested directly. + * + * @return {boolean} + * Whether or not a plugin is a basic plugin. + */ + static isBasic(plugin) { + const p = typeof plugin === 'string' ? getPlugin(plugin) : plugin; + return typeof p === 'function' && !Plugin.prototype.isPrototypeOf(p.prototype); + } + + /** + * Register a Video.js plugin. + * + * @param {string} name + * The name of the plugin to be registered. Must be a string and + * must not match an existing plugin or a method on the `Player` + * prototype. + * + * @param {typeof Plugin|Function} plugin + * A sub-class of `Plugin` or a function for basic plugins. + * + * @return {typeof Plugin|Function} + * For advanced plugins, a factory function for that plugin. For + * basic plugins, a wrapper function that initializes the plugin. + */ + static registerPlugin(name, plugin) { + if (typeof name !== 'string') { + throw new Error(`Illegal plugin name, "${name}", must be a string, was ${typeof name}.`); + } + if (pluginExists(name)) { + log$1.warn(`A plugin named "${name}" already exists. You may want to avoid re-registering plugins!`); + } else if (Player.prototype.hasOwnProperty(name)) { + throw new Error(`Illegal plugin name, "${name}", cannot share a name with an existing player method!`); + } + if (typeof plugin !== 'function') { + throw new Error(`Illegal plugin for "${name}", must be a function, was ${typeof plugin}.`); + } + pluginStorage[name] = plugin; + + // Add a player prototype method for all sub-classed plugins (but not for + // the base Plugin class). + if (name !== BASE_PLUGIN_NAME) { + if (Plugin.isBasic(plugin)) { + Player.prototype[name] = createBasicPlugin(name, plugin); + } else { + Player.prototype[name] = createPluginFactory(name, plugin); + } + } + return plugin; + } + + /** + * De-register a Video.js plugin. + * + * @param {string} name + * The name of the plugin to be de-registered. Must be a string that + * matches an existing plugin. + * + * @throws {Error} + * If an attempt is made to de-register the base plugin. + */ + static deregisterPlugin(name) { + if (name === BASE_PLUGIN_NAME) { + throw new Error('Cannot de-register base plugin.'); + } + if (pluginExists(name)) { + delete pluginStorage[name]; + delete Player.prototype[name]; + } + } + + /** + * Gets an object containing multiple Video.js plugins. + * + * @param {Array} [names] + * If provided, should be an array of plugin names. Defaults to _all_ + * plugin names. + * + * @return {Object|undefined} + * An object containing plugin(s) associated with their name(s) or + * `undefined` if no matching plugins exist). + */ + static getPlugins(names = Object.keys(pluginStorage)) { + let result; + names.forEach(name => { + const plugin = getPlugin(name); + if (plugin) { + result = result || {}; + result[name] = plugin; + } + }); + return result; + } + + /** + * Gets a plugin's version, if available + * + * @param {string} name + * The name of a plugin. + * + * @return {string} + * The plugin's version or an empty string. + */ + static getPluginVersion(name) { + const plugin = getPlugin(name); + return plugin && plugin.VERSION || ''; + } +} + +/** + * Gets a plugin by name if it exists. + * + * @static + * @method getPlugin + * @memberOf Plugin + * @param {string} name + * The name of a plugin. + * + * @returns {typeof Plugin|Function|undefined} + * The plugin (or `undefined`). + */ +Plugin.getPlugin = getPlugin; + +/** + * The name of the base plugin class as it is registered. + * + * @type {string} + */ +Plugin.BASE_PLUGIN_NAME = BASE_PLUGIN_NAME; +Plugin.registerPlugin(BASE_PLUGIN_NAME, Plugin); + +/** + * Documented in player.js + * + * @ignore + */ +Player.prototype.usingPlugin = function (name) { + return !!this[PLUGIN_CACHE_KEY] && this[PLUGIN_CACHE_KEY][name] === true; +}; + +/** + * Documented in player.js + * + * @ignore + */ +Player.prototype.hasPlugin = function (name) { + return !!pluginExists(name); +}; + +/** + * Signals that a plugin is about to be set up on a player. + * + * @event Player#beforepluginsetup + * @type {PluginEventHash} + */ + +/** + * Signals that a plugin is about to be set up on a player - by name. The name + * is the name of the plugin. + * + * @event Player#beforepluginsetup:$name + * @type {PluginEventHash} + */ + +/** + * Signals that a plugin has just been set up on a player. + * + * @event Player#pluginsetup + * @type {PluginEventHash} + */ + +/** + * Signals that a plugin has just been set up on a player - by name. The name + * is the name of the plugin. + * + * @event Player#pluginsetup:$name + * @type {PluginEventHash} + */ + +/** + * @typedef {Object} PluginEventHash + * + * @property {string} instance + * For basic plugins, the return value of the plugin function. For + * advanced plugins, the plugin instance on which the event is fired. + * + * @property {string} name + * The name of the plugin. + * + * @property {string} plugin + * For basic plugins, the plugin function. For advanced plugins, the + * plugin class/constructor. + */ + +/** + * @file deprecate.js + * @module deprecate + */ + +/** + * Decorate a function with a deprecation message the first time it is called. + * + * @param {string} message + * A deprecation message to log the first time the returned function + * is called. + * + * @param {Function} fn + * The function to be deprecated. + * + * @return {Function} + * A wrapper function that will log a deprecation warning the first + * time it is called. The return value will be the return value of + * the wrapped function. + */ +function deprecate(message, fn) { + let warned = false; + return function (...args) { + if (!warned) { + log$1.warn(message); + } + warned = true; + return fn.apply(this, args); + }; +} + +/** + * Internal function used to mark a function as deprecated in the next major + * version with consistent messaging. + * + * @param {number} major The major version where it will be removed + * @param {string} oldName The old function name + * @param {string} newName The new function name + * @param {Function} fn The function to deprecate + * @return {Function} The decorated function + */ +function deprecateForMajor(major, oldName, newName, fn) { + return deprecate(`${oldName} is deprecated and will be removed in ${major}.0; please use ${newName} instead.`, fn); +} + +var VjsErrors = { + UnsupportedSidxContainer: 'unsupported-sidx-container-error', + DashManifestSidxParsingError: 'dash-manifest-sidx-parsing-error', + HlsPlaylistRequestError: 'hls-playlist-request-error', + SegmentUnsupportedMediaFormat: 'segment-unsupported-media-format-error', + UnsupportedMediaInitialization: 'unsupported-media-initialization-error', + SegmentSwitchError: 'segment-switch-error', + SegmentExceedsSourceBufferQuota: 'segment-exceeds-source-buffer-quota-error', + SegmentAppendError: 'segment-append-error', + VttLoadError: 'vtt-load-error', + VttCueParsingError: 'vtt-cue-parsing-error', + // Errors used in contrib-ads: + AdsBeforePrerollError: 'ads-before-preroll-error', + AdsPrerollError: 'ads-preroll-error', + AdsMidrollError: 'ads-midroll-error', + AdsPostrollError: 'ads-postroll-error', + AdsMacroReplacementFailed: 'ads-macro-replacement-failed', + AdsResumeContentFailed: 'ads-resume-content-failed', + // Errors used in contrib-eme: + EMEFailedToRequestMediaKeySystemAccess: 'eme-failed-request-media-key-system-access', + EMEFailedToCreateMediaKeys: 'eme-failed-create-media-keys', + EMEFailedToAttachMediaKeysToVideoElement: 'eme-failed-attach-media-keys-to-video', + EMEFailedToCreateMediaKeySession: 'eme-failed-create-media-key-session', + EMEFailedToSetServerCertificate: 'eme-failed-set-server-certificate', + EMEFailedToGenerateLicenseRequest: 'eme-failed-generate-license-request', + EMEFailedToUpdateSessionWithReceivedLicenseKeys: 'eme-failed-update-session', + EMEFailedToCloseSession: 'eme-failed-close-session', + EMEFailedToRemoveKeysFromSession: 'eme-failed-remove-keys', + EMEFailedToLoadSessionBySessionId: 'eme-failed-load-session' +}; + +/** + * @file video.js + * @module videojs + */ + +/** + * Normalize an `id` value by trimming off a leading `#` + * + * @private + * @param {string} id + * A string, maybe with a leading `#`. + * + * @return {string} + * The string, without any leading `#`. + */ +const normalizeId = id => id.indexOf('#') === 0 ? id.slice(1) : id; + +/** + * A callback that is called when a component is ready. Does not have any + * parameters and any callback value will be ignored. See: {@link Component~ReadyCallback} + * + * @callback ReadyCallback + */ + +/** + * The `videojs()` function doubles as the main function for users to create a + * {@link Player} instance as well as the main library namespace. + * + * It can also be used as a getter for a pre-existing {@link Player} instance. + * However, we _strongly_ recommend using `videojs.getPlayer()` for this + * purpose because it avoids any potential for unintended initialization. + * + * Due to [limitations](https://github.com/jsdoc3/jsdoc/issues/955#issuecomment-313829149) + * of our JSDoc template, we cannot properly document this as both a function + * and a namespace, so its function signature is documented here. + * + * #### Arguments + * ##### id + * string|Element, **required** + * + * Video element or video element ID. + * + * ##### options + * Object, optional + * + * Options object for providing settings. + * See: [Options Guide](https://docs.videojs.com/tutorial-options.html). + * + * ##### ready + * {@link Component~ReadyCallback}, optional + * + * A function to be called when the {@link Player} and {@link Tech} are ready. + * + * #### Return Value + * + * The `videojs()` function returns a {@link Player} instance. + * + * @namespace + * + * @borrows AudioTrack as AudioTrack + * @borrows Component.getComponent as getComponent + * @borrows module:events.on as on + * @borrows module:events.one as one + * @borrows module:events.off as off + * @borrows module:events.trigger as trigger + * @borrows EventTarget as EventTarget + * @borrows module:middleware.use as use + * @borrows Player.players as players + * @borrows Plugin.registerPlugin as registerPlugin + * @borrows Plugin.deregisterPlugin as deregisterPlugin + * @borrows Plugin.getPlugins as getPlugins + * @borrows Plugin.getPlugin as getPlugin + * @borrows Plugin.getPluginVersion as getPluginVersion + * @borrows Tech.getTech as getTech + * @borrows Tech.registerTech as registerTech + * @borrows TextTrack as TextTrack + * @borrows VideoTrack as VideoTrack + * + * @param {string|Element} id + * Video element or video element ID. + * + * @param {Object} [options] + * Options object for providing settings. + * See: [Options Guide](https://docs.videojs.com/tutorial-options.html). + * + * @param {ReadyCallback} [ready] + * A function to be called when the {@link Player} and {@link Tech} are + * ready. + * + * @return {Player} + * The `videojs()` function returns a {@link Player|Player} instance. + */ +function videojs(id, options, ready) { + let player = videojs.getPlayer(id); + if (player) { + if (options) { + log$1.warn(`Player "${id}" is already initialised. Options will not be applied.`); + } + if (ready) { + player.ready(ready); + } + return player; + } + const el = typeof id === 'string' ? $('#' + normalizeId(id)) : id; + if (!isEl(el)) { + throw new TypeError('The element or ID supplied is not valid. (videojs)'); + } + + // document.body.contains(el) will only check if el is contained within that one document. + // This causes problems for elements in iframes. + // Instead, use the element's ownerDocument instead of the global document. + // This will make sure that the element is indeed in the dom of that document. + // Additionally, check that the document in question has a default view. + // If the document is no longer attached to the dom, the defaultView of the document will be null. + // If element is inside Shadow DOM (e.g. is part of a Custom element), ownerDocument.body + // always returns false. Instead, use the Shadow DOM root. + const inShadowDom = 'getRootNode' in el ? el.getRootNode() instanceof window$1.ShadowRoot : false; + const rootNode = inShadowDom ? el.getRootNode() : el.ownerDocument.body; + if (!el.ownerDocument.defaultView || !rootNode.contains(el)) { + log$1.warn('The element supplied is not included in the DOM'); + } + options = options || {}; + + // Store a copy of the el before modification, if it is to be restored in destroy() + // If div ingest, store the parent div + if (options.restoreEl === true) { + options.restoreEl = (el.parentNode && el.parentNode.hasAttribute('data-vjs-player') ? el.parentNode : el).cloneNode(true); + } + hooks('beforesetup').forEach(hookFunction => { + const opts = hookFunction(el, merge$1(options)); + if (!isObject(opts) || Array.isArray(opts)) { + log$1.error('please return an object in beforesetup hooks'); + return; + } + options = merge$1(options, opts); + }); + + // We get the current "Player" component here in case an integration has + // replaced it with a custom player. + const PlayerComponent = Component$1.getComponent('Player'); + player = new PlayerComponent(el, options, ready); + hooks('setup').forEach(hookFunction => hookFunction(player)); + return player; +} +videojs.hooks_ = hooks_; +videojs.hooks = hooks; +videojs.hook = hook; +videojs.hookOnce = hookOnce; +videojs.removeHook = removeHook; + +// Add default styles +if (window$1.VIDEOJS_NO_DYNAMIC_STYLE !== true && isReal()) { + let style = $('.vjs-styles-defaults'); + if (!style) { + style = createStyleElement('vjs-styles-defaults'); + const head = $('head'); + if (head) { + head.insertBefore(style, head.firstChild); + } + setTextContent(style, ` + .video-js { + width: 300px; + height: 150px; + } + + .vjs-fluid:not(.vjs-audio-only-mode) { + padding-top: 56.25% + } + `); + } +} + +// Run Auto-load players +// You have to wait at least once in case this script is loaded after your +// video in the DOM (weird behavior only with minified version) +autoSetupTimeout(1, videojs); + +/** + * Current Video.js version. Follows [semantic versioning](https://semver.org/). + * + * @type {string} + */ +videojs.VERSION = version$6; + +/** + * The global options object. These are the settings that take effect + * if no overrides are specified when the player is created. + * + * @type {Object} + */ +videojs.options = Player.prototype.options_; + +/** + * Get an object with the currently created players, keyed by player ID + * + * @return {Object} + * The created players + */ +videojs.getPlayers = () => Player.players; + +/** + * Get a single player based on an ID or DOM element. + * + * This is useful if you want to check if an element or ID has an associated + * Video.js player, but not create one if it doesn't. + * + * @param {string|Element} id + * An HTML element - `<video>`, `<audio>`, or `<video-js>` - + * or a string matching the `id` of such an element. + * + * @return {Player|undefined} + * A player instance or `undefined` if there is no player instance + * matching the argument. + */ +videojs.getPlayer = id => { + const players = Player.players; + let tag; + if (typeof id === 'string') { + const nId = normalizeId(id); + const player = players[nId]; + if (player) { + return player; + } + tag = $('#' + nId); + } else { + tag = id; + } + if (isEl(tag)) { + const { + player, + playerId + } = tag; + + // Element may have a `player` property referring to an already created + // player instance. If so, return that. + if (player || players[playerId]) { + return player || players[playerId]; + } + } +}; + +/** + * Returns an array of all current players. + * + * @return {Array} + * An array of all players. The array will be in the order that + * `Object.keys` provides, which could potentially vary between + * JavaScript engines. + * + */ +videojs.getAllPlayers = () => +// Disposed players leave a key with a `null` value, so we need to make sure +// we filter those out. +Object.keys(Player.players).map(k => Player.players[k]).filter(Boolean); +videojs.players = Player.players; +videojs.getComponent = Component$1.getComponent; + +/** + * Register a component so it can referred to by name. Used when adding to other + * components, either through addChild `component.addChild('myComponent')` or through + * default children options `{ children: ['myComponent'] }`. + * + * > NOTE: You could also just initialize the component before adding. + * `component.addChild(new MyComponent());` + * + * @param {string} name + * The class name of the component + * + * @param {typeof Component} comp + * The component class + * + * @return {typeof Component} + * The newly registered component + */ +videojs.registerComponent = (name, comp) => { + if (Tech.isTech(comp)) { + log$1.warn(`The ${name} tech was registered as a component. It should instead be registered using videojs.registerTech(name, tech)`); + } + return Component$1.registerComponent.call(Component$1, name, comp); +}; +videojs.getTech = Tech.getTech; +videojs.registerTech = Tech.registerTech; +videojs.use = use; + +/** + * An object that can be returned by a middleware to signify + * that the middleware is being terminated. + * + * @type {object} + * @property {object} middleware.TERMINATOR + */ +Object.defineProperty(videojs, 'middleware', { + value: {}, + writeable: false, + enumerable: true +}); +Object.defineProperty(videojs.middleware, 'TERMINATOR', { + value: TERMINATOR, + writeable: false, + enumerable: true +}); + +/** + * A reference to the {@link module:browser|browser utility module} as an object. + * + * @type {Object} + * @see {@link module:browser|browser} + */ +videojs.browser = browser; + +/** + * A reference to the {@link module:obj|obj utility module} as an object. + * + * @type {Object} + * @see {@link module:obj|obj} + */ +videojs.obj = Obj; + +/** + * Deprecated reference to the {@link module:obj.merge|merge function} + * + * @type {Function} + * @see {@link module:obj.merge|merge} + * @deprecated Deprecated and will be removed in 9.0. Please use videojs.obj.merge instead. + */ +videojs.mergeOptions = deprecateForMajor(9, 'videojs.mergeOptions', 'videojs.obj.merge', merge$1); + +/** + * Deprecated reference to the {@link module:obj.defineLazyProperty|defineLazyProperty function} + * + * @type {Function} + * @see {@link module:obj.defineLazyProperty|defineLazyProperty} + * @deprecated Deprecated and will be removed in 9.0. Please use videojs.obj.defineLazyProperty instead. + */ +videojs.defineLazyProperty = deprecateForMajor(9, 'videojs.defineLazyProperty', 'videojs.obj.defineLazyProperty', defineLazyProperty); + +/** + * Deprecated reference to the {@link module:fn.bind_|fn.bind_ function} + * + * @type {Function} + * @see {@link module:fn.bind_|fn.bind_} + * @deprecated Deprecated and will be removed in 9.0. Please use native Function.prototype.bind instead. + */ +videojs.bind = deprecateForMajor(9, 'videojs.bind', 'native Function.prototype.bind', bind_); +videojs.registerPlugin = Plugin.registerPlugin; +videojs.deregisterPlugin = Plugin.deregisterPlugin; + +/** + * Deprecated method to register a plugin with Video.js + * + * @deprecated Deprecated and will be removed in 9.0. Use videojs.registerPlugin() instead. + * + * @param {string} name + * The plugin name +* + * @param {typeof Plugin|Function} plugin + * The plugin sub-class or function + * + * @return {typeof Plugin|Function} + */ +videojs.plugin = (name, plugin) => { + log$1.warn('videojs.plugin() is deprecated; use videojs.registerPlugin() instead'); + return Plugin.registerPlugin(name, plugin); +}; +videojs.getPlugins = Plugin.getPlugins; +videojs.getPlugin = Plugin.getPlugin; +videojs.getPluginVersion = Plugin.getPluginVersion; + +/** + * Adding languages so that they're available to all players. + * Example: `videojs.addLanguage('es', { 'Hello': 'Hola' });` + * + * @param {string} code + * The language code or dictionary property + * + * @param {Object} data + * The data values to be translated + * + * @return {Object} + * The resulting language dictionary object + */ +videojs.addLanguage = function (code, data) { + code = ('' + code).toLowerCase(); + videojs.options.languages = merge$1(videojs.options.languages, { + [code]: data + }); + return videojs.options.languages[code]; +}; + +/** + * A reference to the {@link module:log|log utility module} as an object. + * + * @type {Function} + * @see {@link module:log|log} + */ +videojs.log = log$1; +videojs.createLogger = createLogger; + +/** + * A reference to the {@link module:time|time utility module} as an object. + * + * @type {Object} + * @see {@link module:time|time} + */ +videojs.time = Time; + +/** + * Deprecated reference to the {@link module:time.createTimeRanges|createTimeRanges function} + * + * @type {Function} + * @see {@link module:time.createTimeRanges|createTimeRanges} + * @deprecated Deprecated and will be removed in 9.0. Please use videojs.time.createTimeRanges instead. + */ +videojs.createTimeRange = deprecateForMajor(9, 'videojs.createTimeRange', 'videojs.time.createTimeRanges', createTimeRanges$1); + +/** + * Deprecated reference to the {@link module:time.createTimeRanges|createTimeRanges function} + * + * @type {Function} + * @see {@link module:time.createTimeRanges|createTimeRanges} + * @deprecated Deprecated and will be removed in 9.0. Please use videojs.time.createTimeRanges instead. + */ +videojs.createTimeRanges = deprecateForMajor(9, 'videojs.createTimeRanges', 'videojs.time.createTimeRanges', createTimeRanges$1); + +/** + * Deprecated reference to the {@link module:time.formatTime|formatTime function} + * + * @type {Function} + * @see {@link module:time.formatTime|formatTime} + * @deprecated Deprecated and will be removed in 9.0. Please use videojs.time.format instead. + */ +videojs.formatTime = deprecateForMajor(9, 'videojs.formatTime', 'videojs.time.formatTime', formatTime); + +/** + * Deprecated reference to the {@link module:time.setFormatTime|setFormatTime function} + * + * @type {Function} + * @see {@link module:time.setFormatTime|setFormatTime} + * @deprecated Deprecated and will be removed in 9.0. Please use videojs.time.setFormat instead. + */ +videojs.setFormatTime = deprecateForMajor(9, 'videojs.setFormatTime', 'videojs.time.setFormatTime', setFormatTime); + +/** + * Deprecated reference to the {@link module:time.resetFormatTime|resetFormatTime function} + * + * @type {Function} + * @see {@link module:time.resetFormatTime|resetFormatTime} + * @deprecated Deprecated and will be removed in 9.0. Please use videojs.time.resetFormat instead. + */ +videojs.resetFormatTime = deprecateForMajor(9, 'videojs.resetFormatTime', 'videojs.time.resetFormatTime', resetFormatTime); + +/** + * Deprecated reference to the {@link module:url.parseUrl|Url.parseUrl function} + * + * @type {Function} + * @see {@link module:url.parseUrl|parseUrl} + * @deprecated Deprecated and will be removed in 9.0. Please use videojs.url.parseUrl instead. + */ +videojs.parseUrl = deprecateForMajor(9, 'videojs.parseUrl', 'videojs.url.parseUrl', parseUrl); + +/** + * Deprecated reference to the {@link module:url.isCrossOrigin|Url.isCrossOrigin function} + * + * @type {Function} + * @see {@link module:url.isCrossOrigin|isCrossOrigin} + * @deprecated Deprecated and will be removed in 9.0. Please use videojs.url.isCrossOrigin instead. + */ +videojs.isCrossOrigin = deprecateForMajor(9, 'videojs.isCrossOrigin', 'videojs.url.isCrossOrigin', isCrossOrigin); +videojs.EventTarget = EventTarget$2; +videojs.any = any; +videojs.on = on; +videojs.one = one; +videojs.off = off; +videojs.trigger = trigger; + +/** + * A cross-browser XMLHttpRequest wrapper. + * + * @function + * @param {Object} options + * Settings for the request. + * + * @return {XMLHttpRequest|XDomainRequest} + * The request object. + * + * @see https://github.com/Raynos/xhr + */ +videojs.xhr = XHR; +videojs.TextTrack = TextTrack; +videojs.AudioTrack = AudioTrack; +videojs.VideoTrack = VideoTrack; +['isEl', 'isTextNode', 'createEl', 'hasClass', 'addClass', 'removeClass', 'toggleClass', 'setAttributes', 'getAttributes', 'emptyEl', 'appendContent', 'insertContent'].forEach(k => { + videojs[k] = function () { + log$1.warn(`videojs.${k}() is deprecated; use videojs.dom.${k}() instead`); + return Dom[k].apply(null, arguments); + }; +}); +videojs.computedStyle = deprecateForMajor(9, 'videojs.computedStyle', 'videojs.dom.computedStyle', computedStyle); + +/** + * A reference to the {@link module:dom|DOM utility module} as an object. + * + * @type {Object} + * @see {@link module:dom|dom} + */ +videojs.dom = Dom; + +/** + * A reference to the {@link module:fn|fn utility module} as an object. + * + * @type {Object} + * @see {@link module:fn|fn} + */ +videojs.fn = Fn; + +/** + * A reference to the {@link module:num|num utility module} as an object. + * + * @type {Object} + * @see {@link module:num|num} + */ +videojs.num = Num; + +/** + * A reference to the {@link module:str|str utility module} as an object. + * + * @type {Object} + * @see {@link module:str|str} + */ +videojs.str = Str; + +/** + * A reference to the {@link module:url|URL utility module} as an object. + * + * @type {Object} + * @see {@link module:url|url} + */ +videojs.url = Url; + +// The list of possible error types to occur in video.js +videojs.Error = VjsErrors; + +/*! @name videojs-contrib-quality-levels @version 4.1.0 @license Apache-2.0 */ + +/** + * A single QualityLevel. + * + * interface QualityLevel { + * readonly attribute DOMString id; + * attribute DOMString label; + * readonly attribute long width; + * readonly attribute long height; + * readonly attribute long bitrate; + * attribute boolean enabled; + * }; + * + * @class QualityLevel + */ +class QualityLevel { + /** + * Creates a QualityLevel + * + * @param {Representation|Object} representation The representation of the quality level + * @param {string} representation.id Unique id of the QualityLevel + * @param {number=} representation.width Resolution width of the QualityLevel + * @param {number=} representation.height Resolution height of the QualityLevel + * @param {number} representation.bandwidth Bitrate of the QualityLevel + * @param {number=} representation.frameRate Frame-rate of the QualityLevel + * @param {Function} representation.enabled Callback to enable/disable QualityLevel + */ + constructor(representation) { + let level = this; // eslint-disable-line + + level.id = representation.id; + level.label = level.id; + level.width = representation.width; + level.height = representation.height; + level.bitrate = representation.bandwidth; + level.frameRate = representation.frameRate; + level.enabled_ = representation.enabled; + Object.defineProperty(level, 'enabled', { + /** + * Get whether the QualityLevel is enabled. + * + * @return {boolean} True if the QualityLevel is enabled. + */ + get() { + return level.enabled_(); + }, + /** + * Enable or disable the QualityLevel. + * + * @param {boolean} enable true to enable QualityLevel, false to disable. + */ + set(enable) { + level.enabled_(enable); + } + }); + return level; + } +} + +/** + * A list of QualityLevels. + * + * interface QualityLevelList : EventTarget { + * getter QualityLevel (unsigned long index); + * readonly attribute unsigned long length; + * readonly attribute long selectedIndex; + * + * void addQualityLevel(QualityLevel qualityLevel) + * void removeQualityLevel(QualityLevel remove) + * QualityLevel? getQualityLevelById(DOMString id); + * + * attribute EventHandler onchange; + * attribute EventHandler onaddqualitylevel; + * attribute EventHandler onremovequalitylevel; + * }; + * + * @extends videojs.EventTarget + * @class QualityLevelList + */ +class QualityLevelList extends videojs.EventTarget { + /** + * Creates a QualityLevelList. + */ + constructor() { + super(); + let list = this; // eslint-disable-line + + list.levels_ = []; + list.selectedIndex_ = -1; + + /** + * Get the index of the currently selected QualityLevel. + * + * @returns {number} The index of the selected QualityLevel. -1 if none selected. + * @readonly + */ + Object.defineProperty(list, 'selectedIndex', { + get() { + return list.selectedIndex_; + } + }); + + /** + * Get the length of the list of QualityLevels. + * + * @returns {number} The length of the list. + * @readonly + */ + Object.defineProperty(list, 'length', { + get() { + return list.levels_.length; + } + }); + list[Symbol.iterator] = () => list.levels_.values(); + return list; + } + + /** + * Adds a quality level to the list. + * + * @param {Representation|Object} representation The representation of the quality level + * @param {string} representation.id Unique id of the QualityLevel + * @param {number=} representation.width Resolution width of the QualityLevel + * @param {number=} representation.height Resolution height of the QualityLevel + * @param {number} representation.bandwidth Bitrate of the QualityLevel + * @param {number=} representation.frameRate Frame-rate of the QualityLevel + * @param {Function} representation.enabled Callback to enable/disable QualityLevel + * @return {QualityLevel} the QualityLevel added to the list + * @method addQualityLevel + */ + addQualityLevel(representation) { + let qualityLevel = this.getQualityLevelById(representation.id); + + // Do not add duplicate quality levels + if (qualityLevel) { + return qualityLevel; + } + const index = this.levels_.length; + qualityLevel = new QualityLevel(representation); + if (!('' + index in this)) { + Object.defineProperty(this, index, { + get() { + return this.levels_[index]; + } + }); + } + this.levels_.push(qualityLevel); + this.trigger({ + qualityLevel, + type: 'addqualitylevel' + }); + return qualityLevel; + } + + /** + * Removes a quality level from the list. + * + * @param {QualityLevel} qualityLevel The QualityLevel to remove from the list. + * @return {QualityLevel|null} the QualityLevel removed or null if nothing removed + * @method removeQualityLevel + */ + removeQualityLevel(qualityLevel) { + let removed = null; + for (let i = 0, l = this.length; i < l; i++) { + if (this[i] === qualityLevel) { + removed = this.levels_.splice(i, 1)[0]; + if (this.selectedIndex_ === i) { + this.selectedIndex_ = -1; + } else if (this.selectedIndex_ > i) { + this.selectedIndex_--; + } + break; + } + } + if (removed) { + this.trigger({ + qualityLevel, + type: 'removequalitylevel' + }); + } + return removed; + } + + /** + * Searches for a QualityLevel with the given id. + * + * @param {string} id The id of the QualityLevel to find. + * @return {QualityLevel|null} The QualityLevel with id, or null if not found. + * @method getQualityLevelById + */ + getQualityLevelById(id) { + for (let i = 0, l = this.length; i < l; i++) { + const level = this[i]; + if (level.id === id) { + return level; + } + } + return null; + } + + /** + * Resets the list of QualityLevels to empty + * + * @method dispose + */ + dispose() { + this.selectedIndex_ = -1; + this.levels_.length = 0; + } +} + +/** + * change - The selected QualityLevel has changed. + * addqualitylevel - A QualityLevel has been added to the QualityLevelList. + * removequalitylevel - A QualityLevel has been removed from the QualityLevelList. + */ +QualityLevelList.prototype.allowedEvents_ = { + change: 'change', + addqualitylevel: 'addqualitylevel', + removequalitylevel: 'removequalitylevel' +}; + +// emulate attribute EventHandler support to allow for feature detection +for (const event in QualityLevelList.prototype.allowedEvents_) { + QualityLevelList.prototype['on' + event] = null; +} +var version$5 = "4.1.0"; + +/** + * Initialization function for the qualityLevels plugin. Sets up the QualityLevelList and + * event handlers. + * + * @param {Player} player Player object. + * @param {Object} options Plugin options object. + * @return {QualityLevelList} a list of QualityLevels + */ +const initPlugin$1 = function (player, options) { + const originalPluginFn = player.qualityLevels; + const qualityLevelList = new QualityLevelList(); + const disposeHandler = function () { + qualityLevelList.dispose(); + player.qualityLevels = originalPluginFn; + player.off('dispose', disposeHandler); + }; + player.on('dispose', disposeHandler); + player.qualityLevels = () => qualityLevelList; + player.qualityLevels.VERSION = version$5; + return qualityLevelList; +}; + +/** + * A video.js plugin. + * + * In the plugin function, the value of `this` is a video.js `Player` + * instance. You cannot rely on the player being in a "ready" state here, + * depending on how the plugin is invoked. This may or may not be important + * to you; if not, remove the wait for "ready"! + * + * @param {Object} options Plugin options object + * @return {QualityLevelList} a list of QualityLevels + */ +const qualityLevels = function (options) { + return initPlugin$1(this, videojs.obj.merge({}, options)); +}; + +// Register the plugin with video.js. +videojs.registerPlugin('qualityLevels', qualityLevels); + +// Include the version number. +qualityLevels.VERSION = version$5; + +/*! @name @videojs/http-streaming @version 3.12.1 @license Apache-2.0 */ + +/** + * @file resolve-url.js - Handling how URLs are resolved and manipulated + */ +const resolveUrl = _resolveUrl; +/** + * If the xhr request was redirected, return the responseURL, otherwise, + * return the original url. + * + * @api private + * + * @param {string} url - an url being requested + * @param {XMLHttpRequest} req - xhr request result + * + * @return {string} + */ + +const resolveManifestRedirect = (url, req) => { + // To understand how the responseURL below is set and generated: + // - https://fetch.spec.whatwg.org/#concept-response-url + // - https://fetch.spec.whatwg.org/#atomic-http-redirect-handling + if (req && req.responseURL && url !== req.responseURL) { + return req.responseURL; + } + return url; +}; +const logger = source => { + if (videojs.log.debug) { + return videojs.log.debug.bind(videojs, 'VHS:', `${source} >`); + } + return function () {}; +}; + +/** + * Provides a compatibility layer between Video.js 7 and 8 API changes for VHS. + */ +/** + * Delegates to videojs.obj.merge (Video.js 8) or + * videojs.mergeOptions (Video.js 7). + */ + +function merge(...args) { + const context = videojs.obj || videojs; + const fn = context.merge || context.mergeOptions; + return fn.apply(context, args); +} +/** + * Delegates to videojs.time.createTimeRanges (Video.js 8) or + * videojs.createTimeRanges (Video.js 7). + */ + +function createTimeRanges(...args) { + const context = videojs.time || videojs; + const fn = context.createTimeRanges || context.createTimeRanges; + return fn.apply(context, args); +} +/** + * Converts provided buffered ranges to a descriptive string + * + * @param {TimeRanges} buffered - received buffered time ranges + * + * @return {string} - descriptive string + */ + +function bufferedRangesToString(buffered) { + if (buffered.length === 0) { + return 'Buffered Ranges are empty'; + } + let bufferedRangesStr = 'Buffered Ranges: \n'; + for (let i = 0; i < buffered.length; i++) { + const start = buffered.start(i); + const end = buffered.end(i); + bufferedRangesStr += `${start} --> ${end}. Duration (${end - start})\n`; + } + return bufferedRangesStr; +} + +/** + * ranges + * + * Utilities for working with TimeRanges. + * + */ + +const TIME_FUDGE_FACTOR = 1 / 30; // Comparisons between time values such as current time and the end of the buffered range +// can be misleading because of precision differences or when the current media has poorly +// aligned audio and video, which can cause values to be slightly off from what you would +// expect. This value is what we consider to be safe to use in such comparisons to account +// for these scenarios. + +const SAFE_TIME_DELTA = TIME_FUDGE_FACTOR * 3; +const filterRanges = function (timeRanges, predicate) { + const results = []; + let i; + if (timeRanges && timeRanges.length) { + // Search for ranges that match the predicate + for (i = 0; i < timeRanges.length; i++) { + if (predicate(timeRanges.start(i), timeRanges.end(i))) { + results.push([timeRanges.start(i), timeRanges.end(i)]); + } + } + } + return createTimeRanges(results); +}; +/** + * Attempts to find the buffered TimeRange that contains the specified + * time. + * + * @param {TimeRanges} buffered - the TimeRanges object to query + * @param {number} time - the time to filter on. + * @return {TimeRanges} a new TimeRanges object + */ + +const findRange = function (buffered, time) { + return filterRanges(buffered, function (start, end) { + return start - SAFE_TIME_DELTA <= time && end + SAFE_TIME_DELTA >= time; + }); +}; +/** + * Returns the TimeRanges that begin later than the specified time. + * + * @param {TimeRanges} timeRanges - the TimeRanges object to query + * @param {number} time - the time to filter on. + * @return {TimeRanges} a new TimeRanges object. + */ + +const findNextRange = function (timeRanges, time) { + return filterRanges(timeRanges, function (start) { + return start - TIME_FUDGE_FACTOR >= time; + }); +}; +/** + * Returns gaps within a list of TimeRanges + * + * @param {TimeRanges} buffered - the TimeRanges object + * @return {TimeRanges} a TimeRanges object of gaps + */ + +const findGaps = function (buffered) { + if (buffered.length < 2) { + return createTimeRanges(); + } + const ranges = []; + for (let i = 1; i < buffered.length; i++) { + const start = buffered.end(i - 1); + const end = buffered.start(i); + ranges.push([start, end]); + } + return createTimeRanges(ranges); +}; +/** + * Calculate the intersection of two TimeRanges + * + * @param {TimeRanges} bufferA + * @param {TimeRanges} bufferB + * @return {TimeRanges} The interesection of `bufferA` with `bufferB` + */ + +const bufferIntersection = function (bufferA, bufferB) { + let start = null; + let end = null; + let arity = 0; + const extents = []; + const ranges = []; + if (!bufferA || !bufferA.length || !bufferB || !bufferB.length) { + return createTimeRanges(); + } // Handle the case where we have both buffers and create an + // intersection of the two + + let count = bufferA.length; // A) Gather up all start and end times + + while (count--) { + extents.push({ + time: bufferA.start(count), + type: 'start' + }); + extents.push({ + time: bufferA.end(count), + type: 'end' + }); + } + count = bufferB.length; + while (count--) { + extents.push({ + time: bufferB.start(count), + type: 'start' + }); + extents.push({ + time: bufferB.end(count), + type: 'end' + }); + } // B) Sort them by time + + extents.sort(function (a, b) { + return a.time - b.time; + }); // C) Go along one by one incrementing arity for start and decrementing + // arity for ends + + for (count = 0; count < extents.length; count++) { + if (extents[count].type === 'start') { + arity++; // D) If arity is ever incremented to 2 we are entering an + // overlapping range + + if (arity === 2) { + start = extents[count].time; + } + } else if (extents[count].type === 'end') { + arity--; // E) If arity is ever decremented to 1 we leaving an + // overlapping range + + if (arity === 1) { + end = extents[count].time; + } + } // F) Record overlapping ranges + + if (start !== null && end !== null) { + ranges.push([start, end]); + start = null; + end = null; + } + } + return createTimeRanges(ranges); +}; +/** + * Gets a human readable string for a TimeRange + * + * @param {TimeRange} range + * @return {string} a human readable string + */ + +const printableRange = range => { + const strArr = []; + if (!range || !range.length) { + return ''; + } + for (let i = 0; i < range.length; i++) { + strArr.push(range.start(i) + ' => ' + range.end(i)); + } + return strArr.join(', '); +}; +/** + * Calculates the amount of time left in seconds until the player hits the end of the + * buffer and causes a rebuffer + * + * @param {TimeRange} buffered + * The state of the buffer + * @param {Numnber} currentTime + * The current time of the player + * @param {number} playbackRate + * The current playback rate of the player. Defaults to 1. + * @return {number} + * Time until the player has to start rebuffering in seconds. + * @function timeUntilRebuffer + */ + +const timeUntilRebuffer = function (buffered, currentTime, playbackRate = 1) { + const bufferedEnd = buffered.length ? buffered.end(buffered.length - 1) : 0; + return (bufferedEnd - currentTime) / playbackRate; +}; +/** + * Converts a TimeRanges object into an array representation + * + * @param {TimeRanges} timeRanges + * @return {Array} + */ + +const timeRangesToArray = timeRanges => { + const timeRangesList = []; + for (let i = 0; i < timeRanges.length; i++) { + timeRangesList.push({ + start: timeRanges.start(i), + end: timeRanges.end(i) + }); + } + return timeRangesList; +}; +/** + * Determines if two time range objects are different. + * + * @param {TimeRange} a + * the first time range object to check + * + * @param {TimeRange} b + * the second time range object to check + * + * @return {Boolean} + * Whether the time range objects differ + */ + +const isRangeDifferent = function (a, b) { + // same object + if (a === b) { + return false; + } // one or the other is undefined + + if (!a && b || !b && a) { + return true; + } // length is different + + if (a.length !== b.length) { + return true; + } // see if any start/end pair is different + + for (let i = 0; i < a.length; i++) { + if (a.start(i) !== b.start(i) || a.end(i) !== b.end(i)) { + return true; + } + } // if the length and every pair is the same + // this is the same time range + + return false; +}; +const lastBufferedEnd = function (a) { + if (!a || !a.length || !a.end) { + return; + } + return a.end(a.length - 1); +}; +/** + * A utility function to add up the amount of time in a timeRange + * after a specified startTime. + * ie:[[0, 10], [20, 40], [50, 60]] with a startTime 0 + * would return 40 as there are 40s seconds after 0 in the timeRange + * + * @param {TimeRange} range + * The range to check against + * @param {number} startTime + * The time in the time range that you should start counting from + * + * @return {number} + * The number of seconds in the buffer passed the specified time. + */ + +const timeAheadOf = function (range, startTime) { + let time = 0; + if (!range || !range.length) { + return time; + } + for (let i = 0; i < range.length; i++) { + const start = range.start(i); + const end = range.end(i); // startTime is after this range entirely + + if (startTime > end) { + continue; + } // startTime is within this range + + if (startTime > start && startTime <= end) { + time += end - startTime; + continue; + } // startTime is before this range. + + time += end - start; + } + return time; +}; + +/** + * @file playlist.js + * + * Playlist related utilities. + */ +/** + * Get the duration of a segment, with special cases for + * llhls segments that do not have a duration yet. + * + * @param {Object} playlist + * the playlist that the segment belongs to. + * @param {Object} segment + * the segment to get a duration for. + * + * @return {number} + * the segment duration + */ + +const segmentDurationWithParts = (playlist, segment) => { + // if this isn't a preload segment + // then we will have a segment duration that is accurate. + if (!segment.preload) { + return segment.duration; + } // otherwise we have to add up parts and preload hints + // to get an up to date duration. + + let result = 0; + (segment.parts || []).forEach(function (p) { + result += p.duration; + }); // for preload hints we have to use partTargetDuration + // as they won't even have a duration yet. + + (segment.preloadHints || []).forEach(function (p) { + if (p.type === 'PART') { + result += playlist.partTargetDuration; + } + }); + return result; +}; +/** + * A function to get a combined list of parts and segments with durations + * and indexes. + * + * @param {Playlist} playlist the playlist to get the list for. + * + * @return {Array} The part/segment list. + */ + +const getPartsAndSegments = playlist => (playlist.segments || []).reduce((acc, segment, si) => { + if (segment.parts) { + segment.parts.forEach(function (part, pi) { + acc.push({ + duration: part.duration, + segmentIndex: si, + partIndex: pi, + part, + segment + }); + }); + } else { + acc.push({ + duration: segment.duration, + segmentIndex: si, + partIndex: null, + segment, + part: null + }); + } + return acc; +}, []); +const getLastParts = media => { + const lastSegment = media.segments && media.segments.length && media.segments[media.segments.length - 1]; + return lastSegment && lastSegment.parts || []; +}; +const getKnownPartCount = ({ + preloadSegment +}) => { + if (!preloadSegment) { + return; + } + const { + parts, + preloadHints + } = preloadSegment; + let partCount = (preloadHints || []).reduce((count, hint) => count + (hint.type === 'PART' ? 1 : 0), 0); + partCount += parts && parts.length ? parts.length : 0; + return partCount; +}; +/** + * Get the number of seconds to delay from the end of a + * live playlist. + * + * @param {Playlist} main the main playlist + * @param {Playlist} media the media playlist + * @return {number} the hold back in seconds. + */ + +const liveEdgeDelay = (main, media) => { + if (media.endList) { + return 0; + } // dash suggestedPresentationDelay trumps everything + + if (main && main.suggestedPresentationDelay) { + return main.suggestedPresentationDelay; + } + const hasParts = getLastParts(media).length > 0; // look for "part" delays from ll-hls first + + if (hasParts && media.serverControl && media.serverControl.partHoldBack) { + return media.serverControl.partHoldBack; + } else if (hasParts && media.partTargetDuration) { + return media.partTargetDuration * 3; // finally look for full segment delays + } else if (media.serverControl && media.serverControl.holdBack) { + return media.serverControl.holdBack; + } else if (media.targetDuration) { + return media.targetDuration * 3; + } + return 0; +}; +/** + * walk backward until we find a duration we can use + * or return a failure + * + * @param {Playlist} playlist the playlist to walk through + * @param {Number} endSequence the mediaSequence to stop walking on + */ + +const backwardDuration = function (playlist, endSequence) { + let result = 0; + let i = endSequence - playlist.mediaSequence; // if a start time is available for segment immediately following + // the interval, use it + + let segment = playlist.segments[i]; // Walk backward until we find the latest segment with timeline + // information that is earlier than endSequence + + if (segment) { + if (typeof segment.start !== 'undefined') { + return { + result: segment.start, + precise: true + }; + } + if (typeof segment.end !== 'undefined') { + return { + result: segment.end - segment.duration, + precise: true + }; + } + } + while (i--) { + segment = playlist.segments[i]; + if (typeof segment.end !== 'undefined') { + return { + result: result + segment.end, + precise: true + }; + } + result += segmentDurationWithParts(playlist, segment); + if (typeof segment.start !== 'undefined') { + return { + result: result + segment.start, + precise: true + }; + } + } + return { + result, + precise: false + }; +}; +/** + * walk forward until we find a duration we can use + * or return a failure + * + * @param {Playlist} playlist the playlist to walk through + * @param {number} endSequence the mediaSequence to stop walking on + */ + +const forwardDuration = function (playlist, endSequence) { + let result = 0; + let segment; + let i = endSequence - playlist.mediaSequence; // Walk forward until we find the earliest segment with timeline + // information + + for (; i < playlist.segments.length; i++) { + segment = playlist.segments[i]; + if (typeof segment.start !== 'undefined') { + return { + result: segment.start - result, + precise: true + }; + } + result += segmentDurationWithParts(playlist, segment); + if (typeof segment.end !== 'undefined') { + return { + result: segment.end - result, + precise: true + }; + } + } // indicate we didn't find a useful duration estimate + + return { + result: -1, + precise: false + }; +}; +/** + * Calculate the media duration from the segments associated with a + * playlist. The duration of a subinterval of the available segments + * may be calculated by specifying an end index. + * + * @param {Object} playlist a media playlist object + * @param {number=} endSequence an exclusive upper boundary + * for the playlist. Defaults to playlist length. + * @param {number} expired the amount of time that has dropped + * off the front of the playlist in a live scenario + * @return {number} the duration between the first available segment + * and end index. + */ + +const intervalDuration = function (playlist, endSequence, expired) { + if (typeof endSequence === 'undefined') { + endSequence = playlist.mediaSequence + playlist.segments.length; + } + if (endSequence < playlist.mediaSequence) { + return 0; + } // do a backward walk to estimate the duration + + const backward = backwardDuration(playlist, endSequence); + if (backward.precise) { + // if we were able to base our duration estimate on timing + // information provided directly from the Media Source, return + // it + return backward.result; + } // walk forward to see if a precise duration estimate can be made + // that way + + const forward = forwardDuration(playlist, endSequence); + if (forward.precise) { + // we found a segment that has been buffered and so it's + // position is known precisely + return forward.result; + } // return the less-precise, playlist-based duration estimate + + return backward.result + expired; +}; +/** + * Calculates the duration of a playlist. If a start and end index + * are specified, the duration will be for the subset of the media + * timeline between those two indices. The total duration for live + * playlists is always Infinity. + * + * @param {Object} playlist a media playlist object + * @param {number=} endSequence an exclusive upper + * boundary for the playlist. Defaults to the playlist media + * sequence number plus its length. + * @param {number=} expired the amount of time that has + * dropped off the front of the playlist in a live scenario + * @return {number} the duration between the start index and end + * index. + */ + +const duration = function (playlist, endSequence, expired) { + if (!playlist) { + return 0; + } + if (typeof expired !== 'number') { + expired = 0; + } // if a slice of the total duration is not requested, use + // playlist-level duration indicators when they're present + + if (typeof endSequence === 'undefined') { + // if present, use the duration specified in the playlist + if (playlist.totalDuration) { + return playlist.totalDuration; + } // duration should be Infinity for live playlists + + if (!playlist.endList) { + return window$1.Infinity; + } + } // calculate the total duration based on the segment durations + + return intervalDuration(playlist, endSequence, expired); +}; +/** + * Calculate the time between two indexes in the current playlist + * neight the start- nor the end-index need to be within the current + * playlist in which case, the targetDuration of the playlist is used + * to approximate the durations of the segments + * + * @param {Array} options.durationList list to iterate over for durations. + * @param {number} options.defaultDuration duration to use for elements before or after the durationList + * @param {number} options.startIndex partsAndSegments index to start + * @param {number} options.endIndex partsAndSegments index to end. + * @return {number} the number of seconds between startIndex and endIndex + */ + +const sumDurations = function ({ + defaultDuration, + durationList, + startIndex, + endIndex +}) { + let durations = 0; + if (startIndex > endIndex) { + [startIndex, endIndex] = [endIndex, startIndex]; + } + if (startIndex < 0) { + for (let i = startIndex; i < Math.min(0, endIndex); i++) { + durations += defaultDuration; + } + startIndex = 0; + } + for (let i = startIndex; i < endIndex; i++) { + durations += durationList[i].duration; + } + return durations; +}; +/** + * Calculates the playlist end time + * + * @param {Object} playlist a media playlist object + * @param {number=} expired the amount of time that has + * dropped off the front of the playlist in a live scenario + * @param {boolean|false} useSafeLiveEnd a boolean value indicating whether or not the + * playlist end calculation should consider the safe live end + * (truncate the playlist end by three segments). This is normally + * used for calculating the end of the playlist's seekable range. + * This takes into account the value of liveEdgePadding. + * Setting liveEdgePadding to 0 is equivalent to setting this to false. + * @param {number} liveEdgePadding a number indicating how far from the end of the playlist we should be in seconds. + * If this is provided, it is used in the safe live end calculation. + * Setting useSafeLiveEnd=false or liveEdgePadding=0 are equivalent. + * Corresponds to suggestedPresentationDelay in DASH manifests. + * @return {number} the end time of playlist + * @function playlistEnd + */ + +const playlistEnd = function (playlist, expired, useSafeLiveEnd, liveEdgePadding) { + if (!playlist || !playlist.segments) { + return null; + } + if (playlist.endList) { + return duration(playlist); + } + if (expired === null) { + return null; + } + expired = expired || 0; + let lastSegmentEndTime = intervalDuration(playlist, playlist.mediaSequence + playlist.segments.length, expired); + if (useSafeLiveEnd) { + liveEdgePadding = typeof liveEdgePadding === 'number' ? liveEdgePadding : liveEdgeDelay(null, playlist); + lastSegmentEndTime -= liveEdgePadding; + } // don't return a time less than zero + + return Math.max(0, lastSegmentEndTime); +}; +/** + * Calculates the interval of time that is currently seekable in a + * playlist. The returned time ranges are relative to the earliest + * moment in the specified playlist that is still available. A full + * seekable implementation for live streams would need to offset + * these values by the duration of content that has expired from the + * stream. + * + * @param {Object} playlist a media playlist object + * dropped off the front of the playlist in a live scenario + * @param {number=} expired the amount of time that has + * dropped off the front of the playlist in a live scenario + * @param {number} liveEdgePadding how far from the end of the playlist we should be in seconds. + * Corresponds to suggestedPresentationDelay in DASH manifests. + * @return {TimeRanges} the periods of time that are valid targets + * for seeking + */ + +const seekable = function (playlist, expired, liveEdgePadding) { + const useSafeLiveEnd = true; + const seekableStart = expired || 0; + let seekableEnd = playlistEnd(playlist, expired, useSafeLiveEnd, liveEdgePadding); + if (seekableEnd === null) { + return createTimeRanges(); + } // Clamp seekable end since it can not be less than the seekable start + + if (seekableEnd < seekableStart) { + seekableEnd = seekableStart; + } + return createTimeRanges(seekableStart, seekableEnd); +}; +/** + * Determine the index and estimated starting time of the segment that + * contains a specified playback position in a media playlist. + * + * @param {Object} options.playlist the media playlist to query + * @param {number} options.currentTime The number of seconds since the earliest + * possible position to determine the containing segment for + * @param {number} options.startTime the time when the segment/part starts + * @param {number} options.startingSegmentIndex the segment index to start looking at. + * @param {number?} [options.startingPartIndex] the part index to look at within the segment. + * + * @return {Object} an object with partIndex, segmentIndex, and startTime. + */ + +const getMediaInfoForTime = function ({ + playlist, + currentTime, + startingSegmentIndex, + startingPartIndex, + startTime, + exactManifestTimings +}) { + let time = currentTime - startTime; + const partsAndSegments = getPartsAndSegments(playlist); + let startIndex = 0; + for (let i = 0; i < partsAndSegments.length; i++) { + const partAndSegment = partsAndSegments[i]; + if (startingSegmentIndex !== partAndSegment.segmentIndex) { + continue; + } // skip this if part index does not match. + + if (typeof startingPartIndex === 'number' && typeof partAndSegment.partIndex === 'number' && startingPartIndex !== partAndSegment.partIndex) { + continue; + } + startIndex = i; + break; + } + if (time < 0) { + // Walk backward from startIndex in the playlist, adding durations + // until we find a segment that contains `time` and return it + if (startIndex > 0) { + for (let i = startIndex - 1; i >= 0; i--) { + const partAndSegment = partsAndSegments[i]; + time += partAndSegment.duration; + if (exactManifestTimings) { + if (time < 0) { + continue; + } + } else if (time + TIME_FUDGE_FACTOR <= 0) { + continue; + } + return { + partIndex: partAndSegment.partIndex, + segmentIndex: partAndSegment.segmentIndex, + startTime: startTime - sumDurations({ + defaultDuration: playlist.targetDuration, + durationList: partsAndSegments, + startIndex, + endIndex: i + }) + }; + } + } // We were unable to find a good segment within the playlist + // so select the first segment + + return { + partIndex: partsAndSegments[0] && partsAndSegments[0].partIndex || null, + segmentIndex: partsAndSegments[0] && partsAndSegments[0].segmentIndex || 0, + startTime: currentTime + }; + } // When startIndex is negative, we first walk forward to first segment + // adding target durations. If we "run out of time" before getting to + // the first segment, return the first segment + + if (startIndex < 0) { + for (let i = startIndex; i < 0; i++) { + time -= playlist.targetDuration; + if (time < 0) { + return { + partIndex: partsAndSegments[0] && partsAndSegments[0].partIndex || null, + segmentIndex: partsAndSegments[0] && partsAndSegments[0].segmentIndex || 0, + startTime: currentTime + }; + } + } + startIndex = 0; + } // Walk forward from startIndex in the playlist, subtracting durations + // until we find a segment that contains `time` and return it + + for (let i = startIndex; i < partsAndSegments.length; i++) { + const partAndSegment = partsAndSegments[i]; + time -= partAndSegment.duration; + const canUseFudgeFactor = partAndSegment.duration > TIME_FUDGE_FACTOR; + const isExactlyAtTheEnd = time === 0; + const isExtremelyCloseToTheEnd = canUseFudgeFactor && time + TIME_FUDGE_FACTOR >= 0; + if (isExactlyAtTheEnd || isExtremelyCloseToTheEnd) { + // 1) We are exactly at the end of the current segment. + // 2) We are extremely close to the end of the current segment (The difference is less than 1 / 30). + // We may encounter this situation when + // we don't have exact match between segment duration info in the manifest and the actual duration of the segment + // For example: + // We appended 3 segments 10 seconds each, meaning we should have 30 sec buffered, + // but we the actual buffered is 29.99999 + // + // In both cases: + // if we passed current time -> it means that we already played current segment + // if we passed buffered.end -> it means that this segment is already loaded and buffered + // we should select the next segment if we have one: + if (i !== partsAndSegments.length - 1) { + continue; + } + } + if (exactManifestTimings) { + if (time > 0) { + continue; + } + } else if (time - TIME_FUDGE_FACTOR >= 0) { + continue; + } + return { + partIndex: partAndSegment.partIndex, + segmentIndex: partAndSegment.segmentIndex, + startTime: startTime + sumDurations({ + defaultDuration: playlist.targetDuration, + durationList: partsAndSegments, + startIndex, + endIndex: i + }) + }; + } // We are out of possible candidates so load the last one... + + return { + segmentIndex: partsAndSegments[partsAndSegments.length - 1].segmentIndex, + partIndex: partsAndSegments[partsAndSegments.length - 1].partIndex, + startTime: currentTime + }; +}; +/** + * Check whether the playlist is excluded or not. + * + * @param {Object} playlist the media playlist object + * @return {boolean} whether the playlist is excluded or not + * @function isExcluded + */ + +const isExcluded = function (playlist) { + return playlist.excludeUntil && playlist.excludeUntil > Date.now(); +}; +/** + * Check whether the playlist is compatible with current playback configuration or has + * been excluded permanently for being incompatible. + * + * @param {Object} playlist the media playlist object + * @return {boolean} whether the playlist is incompatible or not + * @function isIncompatible + */ + +const isIncompatible = function (playlist) { + return playlist.excludeUntil && playlist.excludeUntil === Infinity; +}; +/** + * Check whether the playlist is enabled or not. + * + * @param {Object} playlist the media playlist object + * @return {boolean} whether the playlist is enabled or not + * @function isEnabled + */ + +const isEnabled = function (playlist) { + const excluded = isExcluded(playlist); + return !playlist.disabled && !excluded; +}; +/** + * Check whether the playlist has been manually disabled through the representations api. + * + * @param {Object} playlist the media playlist object + * @return {boolean} whether the playlist is disabled manually or not + * @function isDisabled + */ + +const isDisabled = function (playlist) { + return playlist.disabled; +}; +/** + * Returns whether the current playlist is an AES encrypted HLS stream + * + * @return {boolean} true if it's an AES encrypted HLS stream + */ + +const isAes = function (media) { + for (let i = 0; i < media.segments.length; i++) { + if (media.segments[i].key) { + return true; + } + } + return false; +}; +/** + * Checks if the playlist has a value for the specified attribute + * + * @param {string} attr + * Attribute to check for + * @param {Object} playlist + * The media playlist object + * @return {boolean} + * Whether the playlist contains a value for the attribute or not + * @function hasAttribute + */ + +const hasAttribute = function (attr, playlist) { + return playlist.attributes && playlist.attributes[attr]; +}; +/** + * Estimates the time required to complete a segment download from the specified playlist + * + * @param {number} segmentDuration + * Duration of requested segment + * @param {number} bandwidth + * Current measured bandwidth of the player + * @param {Object} playlist + * The media playlist object + * @param {number=} bytesReceived + * Number of bytes already received for the request. Defaults to 0 + * @return {number|NaN} + * The estimated time to request the segment. NaN if bandwidth information for + * the given playlist is unavailable + * @function estimateSegmentRequestTime + */ + +const estimateSegmentRequestTime = function (segmentDuration, bandwidth, playlist, bytesReceived = 0) { + if (!hasAttribute('BANDWIDTH', playlist)) { + return NaN; + } + const size = segmentDuration * playlist.attributes.BANDWIDTH; + return (size - bytesReceived * 8) / bandwidth; +}; +/* + * Returns whether the current playlist is the lowest rendition + * + * @return {Boolean} true if on lowest rendition + */ + +const isLowestEnabledRendition = (main, media) => { + if (main.playlists.length === 1) { + return true; + } + const currentBandwidth = media.attributes.BANDWIDTH || Number.MAX_VALUE; + return main.playlists.filter(playlist => { + if (!isEnabled(playlist)) { + return false; + } + return (playlist.attributes.BANDWIDTH || 0) < currentBandwidth; + }).length === 0; +}; +const playlistMatch = (a, b) => { + // both playlits are null + // or only one playlist is non-null + // no match + if (!a && !b || !a && b || a && !b) { + return false; + } // playlist objects are the same, match + + if (a === b) { + return true; + } // first try to use id as it should be the most + // accurate + + if (a.id && b.id && a.id === b.id) { + return true; + } // next try to use reslovedUri as it should be the + // second most accurate. + + if (a.resolvedUri && b.resolvedUri && a.resolvedUri === b.resolvedUri) { + return true; + } // finally try to use uri as it should be accurate + // but might miss a few cases for relative uris + + if (a.uri && b.uri && a.uri === b.uri) { + return true; + } + return false; +}; +const someAudioVariant = function (main, callback) { + const AUDIO = main && main.mediaGroups && main.mediaGroups.AUDIO || {}; + let found = false; + for (const groupName in AUDIO) { + for (const label in AUDIO[groupName]) { + found = callback(AUDIO[groupName][label]); + if (found) { + break; + } + } + if (found) { + break; + } + } + return !!found; +}; +const isAudioOnly = main => { + // we are audio only if we have no main playlists but do + // have media group playlists. + if (!main || !main.playlists || !main.playlists.length) { + // without audio variants or playlists this + // is not an audio only main. + const found = someAudioVariant(main, variant => variant.playlists && variant.playlists.length || variant.uri); + return found; + } // if every playlist has only an audio codec it is audio only + + for (let i = 0; i < main.playlists.length; i++) { + const playlist = main.playlists[i]; + const CODECS = playlist.attributes && playlist.attributes.CODECS; // all codecs are audio, this is an audio playlist. + + if (CODECS && CODECS.split(',').every(c => isAudioCodec(c))) { + continue; + } // playlist is in an audio group it is audio only + + const found = someAudioVariant(main, variant => playlistMatch(playlist, variant)); + if (found) { + continue; + } // if we make it here this playlist isn't audio and we + // are not audio only + + return false; + } // if we make it past every playlist without returning, then + // this is an audio only playlist. + + return true; +}; // exports + +var Playlist = { + liveEdgeDelay, + duration, + seekable, + getMediaInfoForTime, + isEnabled, + isDisabled, + isExcluded, + isIncompatible, + playlistEnd, + isAes, + hasAttribute, + estimateSegmentRequestTime, + isLowestEnabledRendition, + isAudioOnly, + playlistMatch, + segmentDurationWithParts +}; +const { + log +} = videojs; +const createPlaylistID = (index, uri) => { + return `${index}-${uri}`; +}; // default function for creating a group id + +const groupID = (type, group, label) => { + return `placeholder-uri-${type}-${group}-${label}`; +}; +/** + * Parses a given m3u8 playlist + * + * @param {Function} [onwarn] + * a function to call when the parser triggers a warning event. + * @param {Function} [oninfo] + * a function to call when the parser triggers an info event. + * @param {string} manifestString + * The downloaded manifest string + * @param {Object[]} [customTagParsers] + * An array of custom tag parsers for the m3u8-parser instance + * @param {Object[]} [customTagMappers] + * An array of custom tag mappers for the m3u8-parser instance + * @param {boolean} [llhls] + * Whether to keep ll-hls features in the manifest after parsing. + * @return {Object} + * The manifest object + */ + +const parseManifest = ({ + onwarn, + oninfo, + manifestString, + customTagParsers = [], + customTagMappers = [], + llhls +}) => { + const parser = new Parser(); + if (onwarn) { + parser.on('warn', onwarn); + } + if (oninfo) { + parser.on('info', oninfo); + } + customTagParsers.forEach(customParser => parser.addParser(customParser)); + customTagMappers.forEach(mapper => parser.addTagMapper(mapper)); + parser.push(manifestString); + parser.end(); + const manifest = parser.manifest; // remove llhls features from the parsed manifest + // if we don't want llhls support. + + if (!llhls) { + ['preloadSegment', 'skip', 'serverControl', 'renditionReports', 'partInf', 'partTargetDuration'].forEach(function (k) { + if (manifest.hasOwnProperty(k)) { + delete manifest[k]; + } + }); + if (manifest.segments) { + manifest.segments.forEach(function (segment) { + ['parts', 'preloadHints'].forEach(function (k) { + if (segment.hasOwnProperty(k)) { + delete segment[k]; + } + }); + }); + } + } + if (!manifest.targetDuration) { + let targetDuration = 10; + if (manifest.segments && manifest.segments.length) { + targetDuration = manifest.segments.reduce((acc, s) => Math.max(acc, s.duration), 0); + } + if (onwarn) { + onwarn({ + message: `manifest has no targetDuration defaulting to ${targetDuration}` + }); + } + manifest.targetDuration = targetDuration; + } + const parts = getLastParts(manifest); + if (parts.length && !manifest.partTargetDuration) { + const partTargetDuration = parts.reduce((acc, p) => Math.max(acc, p.duration), 0); + if (onwarn) { + onwarn({ + message: `manifest has no partTargetDuration defaulting to ${partTargetDuration}` + }); + log.error('LL-HLS manifest has parts but lacks required #EXT-X-PART-INF:PART-TARGET value. See https://datatracker.ietf.org/doc/html/draft-pantos-hls-rfc8216bis-09#section-4.4.3.7. Playback is not guaranteed.'); + } + manifest.partTargetDuration = partTargetDuration; + } + return manifest; +}; +/** + * Loops through all supported media groups in main and calls the provided + * callback for each group + * + * @param {Object} main + * The parsed main manifest object + * @param {Function} callback + * Callback to call for each media group + */ + +const forEachMediaGroup = (main, callback) => { + if (!main.mediaGroups) { + return; + } + ['AUDIO', 'SUBTITLES'].forEach(mediaType => { + if (!main.mediaGroups[mediaType]) { + return; + } + for (const groupKey in main.mediaGroups[mediaType]) { + for (const labelKey in main.mediaGroups[mediaType][groupKey]) { + const mediaProperties = main.mediaGroups[mediaType][groupKey][labelKey]; + callback(mediaProperties, mediaType, groupKey, labelKey); + } + } + }); +}; +/** + * Adds properties and attributes to the playlist to keep consistent functionality for + * playlists throughout VHS. + * + * @param {Object} config + * Arguments object + * @param {Object} config.playlist + * The media playlist + * @param {string} [config.uri] + * The uri to the media playlist (if media playlist is not from within a main + * playlist) + * @param {string} id + * ID to use for the playlist + */ + +const setupMediaPlaylist = ({ + playlist, + uri, + id +}) => { + playlist.id = id; + playlist.playlistErrors_ = 0; + if (uri) { + // For media playlists, m3u8-parser does not have access to a URI, as HLS media + // playlists do not contain their own source URI, but one is needed for consistency in + // VHS. + playlist.uri = uri; + } // For HLS main playlists, even though certain attributes MUST be defined, the + // stream may still be played without them. + // For HLS media playlists, m3u8-parser does not attach an attributes object to the + // manifest. + // + // To avoid undefined reference errors through the project, and make the code easier + // to write/read, add an empty attributes object for these cases. + + playlist.attributes = playlist.attributes || {}; +}; +/** + * Adds ID, resolvedUri, and attributes properties to each playlist of the main, where + * necessary. In addition, creates playlist IDs for each playlist and adds playlist ID to + * playlist references to the playlists array. + * + * @param {Object} main + * The main playlist + */ + +const setupMediaPlaylists = main => { + let i = main.playlists.length; + while (i--) { + const playlist = main.playlists[i]; + setupMediaPlaylist({ + playlist, + id: createPlaylistID(i, playlist.uri) + }); + playlist.resolvedUri = resolveUrl(main.uri, playlist.uri); + main.playlists[playlist.id] = playlist; // URI reference added for backwards compatibility + + main.playlists[playlist.uri] = playlist; // Although the spec states an #EXT-X-STREAM-INF tag MUST have a BANDWIDTH attribute, + // the stream can be played without it. Although an attributes property may have been + // added to the playlist to prevent undefined references, issue a warning to fix the + // manifest. + + if (!playlist.attributes.BANDWIDTH) { + log.warn('Invalid playlist STREAM-INF detected. Missing BANDWIDTH attribute.'); + } + } +}; +/** + * Adds resolvedUri properties to each media group. + * + * @param {Object} main + * The main playlist + */ + +const resolveMediaGroupUris = main => { + forEachMediaGroup(main, properties => { + if (properties.uri) { + properties.resolvedUri = resolveUrl(main.uri, properties.uri); + } + }); +}; +/** + * Creates a main playlist wrapper to insert a sole media playlist into. + * + * @param {Object} media + * Media playlist + * @param {string} uri + * The media URI + * + * @return {Object} + * main playlist + */ + +const mainForMedia = (media, uri) => { + const id = createPlaylistID(0, uri); + const main = { + mediaGroups: { + 'AUDIO': {}, + 'VIDEO': {}, + 'CLOSED-CAPTIONS': {}, + 'SUBTITLES': {} + }, + uri: window$1.location.href, + resolvedUri: window$1.location.href, + playlists: [{ + uri, + id, + resolvedUri: uri, + // m3u8-parser does not attach an attributes property to media playlists so make + // sure that the property is attached to avoid undefined reference errors + attributes: {} + }] + }; // set up ID reference + + main.playlists[id] = main.playlists[0]; // URI reference added for backwards compatibility + + main.playlists[uri] = main.playlists[0]; + return main; +}; +/** + * Does an in-place update of the main manifest to add updated playlist URI references + * as well as other properties needed by VHS that aren't included by the parser. + * + * @param {Object} main + * main manifest object + * @param {string} uri + * The source URI + * @param {function} createGroupID + * A function to determine how to create the groupID for mediaGroups + */ + +const addPropertiesToMain = (main, uri, createGroupID = groupID) => { + main.uri = uri; + for (let i = 0; i < main.playlists.length; i++) { + if (!main.playlists[i].uri) { + // Set up phony URIs for the playlists since playlists are referenced by their URIs + // throughout VHS, but some formats (e.g., DASH) don't have external URIs + // TODO: consider adding dummy URIs in mpd-parser + const phonyUri = `placeholder-uri-${i}`; + main.playlists[i].uri = phonyUri; + } + } + const audioOnlyMain = isAudioOnly(main); + forEachMediaGroup(main, (properties, mediaType, groupKey, labelKey) => { + // add a playlist array under properties + if (!properties.playlists || !properties.playlists.length) { + // If the manifest is audio only and this media group does not have a uri, check + // if the media group is located in the main list of playlists. If it is, don't add + // placeholder properties as it shouldn't be considered an alternate audio track. + if (audioOnlyMain && mediaType === 'AUDIO' && !properties.uri) { + for (let i = 0; i < main.playlists.length; i++) { + const p = main.playlists[i]; + if (p.attributes && p.attributes.AUDIO && p.attributes.AUDIO === groupKey) { + return; + } + } + } + properties.playlists = [_extends({}, properties)]; + } + properties.playlists.forEach(function (p, i) { + const groupId = createGroupID(mediaType, groupKey, labelKey, p); + const id = createPlaylistID(i, groupId); + if (p.uri) { + p.resolvedUri = p.resolvedUri || resolveUrl(main.uri, p.uri); + } else { + // DEPRECATED, this has been added to prevent a breaking change. + // previously we only ever had a single media group playlist, so + // we mark the first playlist uri without prepending the index as we used to + // ideally we would do all of the playlists the same way. + p.uri = i === 0 ? groupId : id; // don't resolve a placeholder uri to an absolute url, just use + // the placeholder again + + p.resolvedUri = p.uri; + } + p.id = p.id || id; // add an empty attributes object, all playlists are + // expected to have this. + + p.attributes = p.attributes || {}; // setup ID and URI references (URI for backwards compatibility) + + main.playlists[p.id] = p; + main.playlists[p.uri] = p; + }); + }); + setupMediaPlaylists(main); + resolveMediaGroupUris(main); +}; +class DateRangesStorage { + constructor() { + this.offset_ = null; + this.pendingDateRanges_ = new Map(); + this.processedDateRanges_ = new Map(); + } + setOffset(segments = []) { + // already set + if (this.offset_ !== null) { + return; + } // no segment to process + + if (!segments.length) { + return; + } + const [firstSegment] = segments; // no program date time + + if (firstSegment.programDateTime === undefined) { + return; + } // Set offset as ProgramDateTime for the very first segment of the very first playlist load: + + this.offset_ = firstSegment.programDateTime / 1000; + } + setPendingDateRanges(dateRanges = []) { + if (!dateRanges.length) { + return; + } + const [dateRange] = dateRanges; + const startTime = dateRange.startDate.getTime(); + this.trimProcessedDateRanges_(startTime); + this.pendingDateRanges_ = dateRanges.reduce((map, pendingDateRange) => { + map.set(pendingDateRange.id, pendingDateRange); + return map; + }, new Map()); + } + processDateRange(dateRange) { + this.pendingDateRanges_.delete(dateRange.id); + this.processedDateRanges_.set(dateRange.id, dateRange); + } + getDateRangesToProcess() { + if (this.offset_ === null) { + return []; + } + const dateRangeClasses = {}; + const dateRangesToProcess = []; + this.pendingDateRanges_.forEach((dateRange, id) => { + if (this.processedDateRanges_.has(id)) { + return; + } + dateRange.startTime = dateRange.startDate.getTime() / 1000 - this.offset_; + dateRange.processDateRange = () => this.processDateRange(dateRange); + dateRangesToProcess.push(dateRange); + if (!dateRange.class) { + return; + } + if (dateRangeClasses[dateRange.class]) { + const length = dateRangeClasses[dateRange.class].push(dateRange); + dateRange.classListIndex = length - 1; + } else { + dateRangeClasses[dateRange.class] = [dateRange]; + dateRange.classListIndex = 0; + } + }); + for (const dateRange of dateRangesToProcess) { + const classList = dateRangeClasses[dateRange.class] || []; + if (dateRange.endDate) { + dateRange.endTime = dateRange.endDate.getTime() / 1000 - this.offset_; + } else if (dateRange.endOnNext && classList[dateRange.classListIndex + 1]) { + dateRange.endTime = classList[dateRange.classListIndex + 1].startTime; + } else if (dateRange.duration) { + dateRange.endTime = dateRange.startTime + dateRange.duration; + } else if (dateRange.plannedDuration) { + dateRange.endTime = dateRange.startTime + dateRange.plannedDuration; + } else { + dateRange.endTime = dateRange.startTime; + } + } + return dateRangesToProcess; + } + trimProcessedDateRanges_(startTime) { + const copy = new Map(this.processedDateRanges_); + copy.forEach((dateRange, id) => { + if (dateRange.startDate.getTime() < startTime) { + this.processedDateRanges_.delete(id); + } + }); + } +} +const { + EventTarget: EventTarget$1 +} = videojs; +const addLLHLSQueryDirectives = (uri, media) => { + if (media.endList || !media.serverControl) { + return uri; + } + const parameters = {}; + if (media.serverControl.canBlockReload) { + const { + preloadSegment + } = media; // next msn is a zero based value, length is not. + + let nextMSN = media.mediaSequence + media.segments.length; // If preload segment has parts then it is likely + // that we are going to request a part of that preload segment. + // the logic below is used to determine that. + + if (preloadSegment) { + const parts = preloadSegment.parts || []; // _HLS_part is a zero based index + + const nextPart = getKnownPartCount(media) - 1; // if nextPart is > -1 and not equal to just the + // length of parts, then we know we had part preload hints + // and we need to add the _HLS_part= query + + if (nextPart > -1 && nextPart !== parts.length - 1) { + // add existing parts to our preload hints + // eslint-disable-next-line + parameters._HLS_part = nextPart; + } // this if statement makes sure that we request the msn + // of the preload segment if: + // 1. the preload segment had parts (and was not yet a full segment) + // but was added to our segments array + // 2. the preload segment had preload hints for parts that are not in + // the manifest yet. + // in all other cases we want the segment after the preload segment + // which will be given by using media.segments.length because it is 1 based + // rather than 0 based. + + if (nextPart > -1 || parts.length) { + nextMSN--; + } + } // add _HLS_msn= in front of any _HLS_part query + // eslint-disable-next-line + + parameters._HLS_msn = nextMSN; + } + if (media.serverControl && media.serverControl.canSkipUntil) { + // add _HLS_skip= infront of all other queries. + // eslint-disable-next-line + parameters._HLS_skip = media.serverControl.canSkipDateranges ? 'v2' : 'YES'; + } + if (Object.keys(parameters).length) { + const parsedUri = new window$1.URL(uri); + ['_HLS_skip', '_HLS_msn', '_HLS_part'].forEach(function (name) { + if (!parameters.hasOwnProperty(name)) { + return; + } + parsedUri.searchParams.set(name, parameters[name]); + }); + uri = parsedUri.toString(); + } + return uri; +}; +/** + * Returns a new segment object with properties and + * the parts array merged. + * + * @param {Object} a the old segment + * @param {Object} b the new segment + * + * @return {Object} the merged segment + */ + +const updateSegment = (a, b) => { + if (!a) { + return b; + } + const result = merge(a, b); // if only the old segment has preload hints + // and the new one does not, remove preload hints. + + if (a.preloadHints && !b.preloadHints) { + delete result.preloadHints; + } // if only the old segment has parts + // then the parts are no longer valid + + if (a.parts && !b.parts) { + delete result.parts; // if both segments have parts + // copy part propeties from the old segment + // to the new one. + } else if (a.parts && b.parts) { + for (let i = 0; i < b.parts.length; i++) { + if (a.parts && a.parts[i]) { + result.parts[i] = merge(a.parts[i], b.parts[i]); + } + } + } // set skipped to false for segments that have + // have had information merged from the old segment. + + if (!a.skipped && b.skipped) { + result.skipped = false; + } // set preload to false for segments that have + // had information added in the new segment. + + if (a.preload && !b.preload) { + result.preload = false; + } + return result; +}; +/** + * Returns a new array of segments that is the result of merging + * properties from an older list of segments onto an updated + * list. No properties on the updated playlist will be ovewritten. + * + * @param {Array} original the outdated list of segments + * @param {Array} update the updated list of segments + * @param {number=} offset the index of the first update + * segment in the original segment list. For non-live playlists, + * this should always be zero and does not need to be + * specified. For live playlists, it should be the difference + * between the media sequence numbers in the original and updated + * playlists. + * @return {Array} a list of merged segment objects + */ + +const updateSegments = (original, update, offset) => { + const oldSegments = original.slice(); + const newSegments = update.slice(); + offset = offset || 0; + const result = []; + let currentMap; + for (let newIndex = 0; newIndex < newSegments.length; newIndex++) { + const oldSegment = oldSegments[newIndex + offset]; + const newSegment = newSegments[newIndex]; + if (oldSegment) { + currentMap = oldSegment.map || currentMap; + result.push(updateSegment(oldSegment, newSegment)); + } else { + // carry over map to new segment if it is missing + if (currentMap && !newSegment.map) { + newSegment.map = currentMap; + } + result.push(newSegment); + } + } + return result; +}; +const resolveSegmentUris = (segment, baseUri) => { + // preloadSegment will not have a uri at all + // as the segment isn't actually in the manifest yet, only parts + if (!segment.resolvedUri && segment.uri) { + segment.resolvedUri = resolveUrl(baseUri, segment.uri); + } + if (segment.key && !segment.key.resolvedUri) { + segment.key.resolvedUri = resolveUrl(baseUri, segment.key.uri); + } + if (segment.map && !segment.map.resolvedUri) { + segment.map.resolvedUri = resolveUrl(baseUri, segment.map.uri); + } + if (segment.map && segment.map.key && !segment.map.key.resolvedUri) { + segment.map.key.resolvedUri = resolveUrl(baseUri, segment.map.key.uri); + } + if (segment.parts && segment.parts.length) { + segment.parts.forEach(p => { + if (p.resolvedUri) { + return; + } + p.resolvedUri = resolveUrl(baseUri, p.uri); + }); + } + if (segment.preloadHints && segment.preloadHints.length) { + segment.preloadHints.forEach(p => { + if (p.resolvedUri) { + return; + } + p.resolvedUri = resolveUrl(baseUri, p.uri); + }); + } +}; +const getAllSegments = function (media) { + const segments = media.segments || []; + const preloadSegment = media.preloadSegment; // a preloadSegment with only preloadHints is not currently + // a usable segment, only include a preloadSegment that has + // parts. + + if (preloadSegment && preloadSegment.parts && preloadSegment.parts.length) { + // if preloadHints has a MAP that means that the + // init segment is going to change. We cannot use any of the parts + // from this preload segment. + if (preloadSegment.preloadHints) { + for (let i = 0; i < preloadSegment.preloadHints.length; i++) { + if (preloadSegment.preloadHints[i].type === 'MAP') { + return segments; + } + } + } // set the duration for our preload segment to target duration. + + preloadSegment.duration = media.targetDuration; + preloadSegment.preload = true; + segments.push(preloadSegment); + } + return segments; +}; // consider the playlist unchanged if the playlist object is the same or +// the number of segments is equal, the media sequence number is unchanged, +// and this playlist hasn't become the end of the playlist + +const isPlaylistUnchanged = (a, b) => a === b || a.segments && b.segments && a.segments.length === b.segments.length && a.endList === b.endList && a.mediaSequence === b.mediaSequence && a.preloadSegment === b.preloadSegment; +/** + * Returns a new main playlist that is the result of merging an + * updated media playlist into the original version. If the + * updated media playlist does not match any of the playlist + * entries in the original main playlist, null is returned. + * + * @param {Object} main a parsed main M3U8 object + * @param {Object} media a parsed media M3U8 object + * @return {Object} a new object that represents the original + * main playlist with the updated media playlist merged in, or + * null if the merge produced no change. + */ + +const updateMain$1 = (main, newMedia, unchangedCheck = isPlaylistUnchanged) => { + const result = merge(main, {}); + const oldMedia = result.playlists[newMedia.id]; + if (!oldMedia) { + return null; + } + if (unchangedCheck(oldMedia, newMedia)) { + return null; + } + newMedia.segments = getAllSegments(newMedia); + const mergedPlaylist = merge(oldMedia, newMedia); // always use the new media's preload segment + + if (mergedPlaylist.preloadSegment && !newMedia.preloadSegment) { + delete mergedPlaylist.preloadSegment; + } // if the update could overlap existing segment information, merge the two segment lists + + if (oldMedia.segments) { + if (newMedia.skip) { + newMedia.segments = newMedia.segments || []; // add back in objects for skipped segments, so that we merge + // old properties into the new segments + + for (let i = 0; i < newMedia.skip.skippedSegments; i++) { + newMedia.segments.unshift({ + skipped: true + }); + } + } + mergedPlaylist.segments = updateSegments(oldMedia.segments, newMedia.segments, newMedia.mediaSequence - oldMedia.mediaSequence); + } // resolve any segment URIs to prevent us from having to do it later + + mergedPlaylist.segments.forEach(segment => { + resolveSegmentUris(segment, mergedPlaylist.resolvedUri); + }); // TODO Right now in the playlists array there are two references to each playlist, one + // that is referenced by index, and one by URI. The index reference may no longer be + // necessary. + + for (let i = 0; i < result.playlists.length; i++) { + if (result.playlists[i].id === newMedia.id) { + result.playlists[i] = mergedPlaylist; + } + } + result.playlists[newMedia.id] = mergedPlaylist; // URI reference added for backwards compatibility + + result.playlists[newMedia.uri] = mergedPlaylist; // update media group playlist references. + + forEachMediaGroup(main, (properties, mediaType, groupKey, labelKey) => { + if (!properties.playlists) { + return; + } + for (let i = 0; i < properties.playlists.length; i++) { + if (newMedia.id === properties.playlists[i].id) { + properties.playlists[i] = mergedPlaylist; + } + } + }); + return result; +}; +/** + * Calculates the time to wait before refreshing a live playlist + * + * @param {Object} media + * The current media + * @param {boolean} update + * True if there were any updates from the last refresh, false otherwise + * @return {number} + * The time in ms to wait before refreshing the live playlist + */ + +const refreshDelay = (media, update) => { + const segments = media.segments || []; + const lastSegment = segments[segments.length - 1]; + const lastPart = lastSegment && lastSegment.parts && lastSegment.parts[lastSegment.parts.length - 1]; + const lastDuration = lastPart && lastPart.duration || lastSegment && lastSegment.duration; + if (update && lastDuration) { + return lastDuration * 1000; + } // if the playlist is unchanged since the last reload or last segment duration + // cannot be determined, try again after half the target duration + + return (media.partTargetDuration || media.targetDuration || 10) * 500; +}; +/** + * Load a playlist from a remote location + * + * @class PlaylistLoader + * @extends Stream + * @param {string|Object} src url or object of manifest + * @param {boolean} withCredentials the withCredentials xhr option + * @class + */ + +class PlaylistLoader extends EventTarget$1 { + constructor(src, vhs, options = {}) { + super(); + if (!src) { + throw new Error('A non-empty playlist URL or object is required'); + } + this.logger_ = logger('PlaylistLoader'); + const { + withCredentials = false + } = options; + this.src = src; + this.vhs_ = vhs; + this.withCredentials = withCredentials; + this.addDateRangesToTextTrack_ = options.addDateRangesToTextTrack; + const vhsOptions = vhs.options_; + this.customTagParsers = vhsOptions && vhsOptions.customTagParsers || []; + this.customTagMappers = vhsOptions && vhsOptions.customTagMappers || []; + this.llhls = vhsOptions && vhsOptions.llhls; + this.dateRangesStorage_ = new DateRangesStorage(); // initialize the loader state + + this.state = 'HAVE_NOTHING'; // live playlist staleness timeout + + this.handleMediaupdatetimeout_ = this.handleMediaupdatetimeout_.bind(this); + this.on('mediaupdatetimeout', this.handleMediaupdatetimeout_); + this.on('loadedplaylist', this.handleLoadedPlaylist_.bind(this)); + } + handleLoadedPlaylist_() { + const mediaPlaylist = this.media(); + if (!mediaPlaylist) { + return; + } + this.dateRangesStorage_.setOffset(mediaPlaylist.segments); + this.dateRangesStorage_.setPendingDateRanges(mediaPlaylist.dateRanges); + const availableDateRanges = this.dateRangesStorage_.getDateRangesToProcess(); + if (!availableDateRanges.length || !this.addDateRangesToTextTrack_) { + return; + } + this.addDateRangesToTextTrack_(availableDateRanges); + } + handleMediaupdatetimeout_() { + if (this.state !== 'HAVE_METADATA') { + // only refresh the media playlist if no other activity is going on + return; + } + const media = this.media(); + let uri = resolveUrl(this.main.uri, media.uri); + if (this.llhls) { + uri = addLLHLSQueryDirectives(uri, media); + } + this.state = 'HAVE_CURRENT_METADATA'; + this.request = this.vhs_.xhr({ + uri, + withCredentials: this.withCredentials, + requestType: 'hls-playlist' + }, (error, req) => { + // disposed + if (!this.request) { + return; + } + if (error) { + return this.playlistRequestError(this.request, this.media(), 'HAVE_METADATA'); + } + this.haveMetadata({ + playlistString: this.request.responseText, + url: this.media().uri, + id: this.media().id + }); + }); + } + playlistRequestError(xhr, playlist, startingState) { + const { + uri, + id + } = playlist; // any in-flight request is now finished + + this.request = null; + if (startingState) { + this.state = startingState; + } + this.error = { + playlist: this.main.playlists[id], + status: xhr.status, + message: `HLS playlist request error at URL: ${uri}.`, + responseText: xhr.responseText, + code: xhr.status >= 500 ? 4 : 2, + metadata: { + errorType: videojs.Error.HlsPlaylistRequestError + } + }; + this.trigger('error'); + } + parseManifest_({ + url, + manifestString + }) { + return parseManifest({ + onwarn: ({ + message + }) => this.logger_(`m3u8-parser warn for ${url}: ${message}`), + oninfo: ({ + message + }) => this.logger_(`m3u8-parser info for ${url}: ${message}`), + manifestString, + customTagParsers: this.customTagParsers, + customTagMappers: this.customTagMappers, + llhls: this.llhls + }); + } + /** + * Update the playlist loader's state in response to a new or updated playlist. + * + * @param {string} [playlistString] + * Playlist string (if playlistObject is not provided) + * @param {Object} [playlistObject] + * Playlist object (if playlistString is not provided) + * @param {string} url + * URL of playlist + * @param {string} id + * ID to use for playlist + */ + + haveMetadata({ + playlistString, + playlistObject, + url, + id + }) { + // any in-flight request is now finished + this.request = null; + this.state = 'HAVE_METADATA'; + const playlist = playlistObject || this.parseManifest_({ + url, + manifestString: playlistString + }); + playlist.lastRequest = Date.now(); + setupMediaPlaylist({ + playlist, + uri: url, + id + }); // merge this playlist into the main manifest + + const update = updateMain$1(this.main, playlist); + this.targetDuration = playlist.partTargetDuration || playlist.targetDuration; + this.pendingMedia_ = null; + if (update) { + this.main = update; + this.media_ = this.main.playlists[id]; + } else { + this.trigger('playlistunchanged'); + } + this.updateMediaUpdateTimeout_(refreshDelay(this.media(), !!update)); + this.trigger('loadedplaylist'); + } + /** + * Abort any outstanding work and clean up. + */ + + dispose() { + this.trigger('dispose'); + this.stopRequest(); + window$1.clearTimeout(this.mediaUpdateTimeout); + window$1.clearTimeout(this.finalRenditionTimeout); + this.dateRangesStorage_ = new DateRangesStorage(); + this.off(); + } + stopRequest() { + if (this.request) { + const oldRequest = this.request; + this.request = null; + oldRequest.onreadystatechange = null; + oldRequest.abort(); + } + } + /** + * When called without any arguments, returns the currently + * active media playlist. When called with a single argument, + * triggers the playlist loader to asynchronously switch to the + * specified media playlist. Calling this method while the + * loader is in the HAVE_NOTHING causes an error to be emitted + * but otherwise has no effect. + * + * @param {Object=} playlist the parsed media playlist + * object to switch to + * @param {boolean=} shouldDelay whether we should delay the request by half target duration + * + * @return {Playlist} the current loaded media + */ + + media(playlist, shouldDelay) { + // getter + if (!playlist) { + return this.media_; + } // setter + + if (this.state === 'HAVE_NOTHING') { + throw new Error('Cannot switch media playlist from ' + this.state); + } // find the playlist object if the target playlist has been + // specified by URI + + if (typeof playlist === 'string') { + if (!this.main.playlists[playlist]) { + throw new Error('Unknown playlist URI: ' + playlist); + } + playlist = this.main.playlists[playlist]; + } + window$1.clearTimeout(this.finalRenditionTimeout); + if (shouldDelay) { + const delay = (playlist.partTargetDuration || playlist.targetDuration) / 2 * 1000 || 5 * 1000; + this.finalRenditionTimeout = window$1.setTimeout(this.media.bind(this, playlist, false), delay); + return; + } + const startingState = this.state; + const mediaChange = !this.media_ || playlist.id !== this.media_.id; + const mainPlaylistRef = this.main.playlists[playlist.id]; // switch to fully loaded playlists immediately + + if (mainPlaylistRef && mainPlaylistRef.endList || + // handle the case of a playlist object (e.g., if using vhs-json with a resolved + // media playlist or, for the case of demuxed audio, a resolved audio media group) + playlist.endList && playlist.segments.length) { + // abort outstanding playlist requests + if (this.request) { + this.request.onreadystatechange = null; + this.request.abort(); + this.request = null; + } + this.state = 'HAVE_METADATA'; + this.media_ = playlist; // trigger media change if the active media has been updated + + if (mediaChange) { + this.trigger('mediachanging'); + if (startingState === 'HAVE_MAIN_MANIFEST') { + // The initial playlist was a main manifest, and the first media selected was + // also provided (in the form of a resolved playlist object) as part of the + // source object (rather than just a URL). Therefore, since the media playlist + // doesn't need to be requested, loadedmetadata won't trigger as part of the + // normal flow, and needs an explicit trigger here. + this.trigger('loadedmetadata'); + } else { + this.trigger('mediachange'); + } + } + return; + } // We update/set the timeout here so that live playlists + // that are not a media change will "start" the loader as expected. + // We expect that this function will start the media update timeout + // cycle again. This also prevents a playlist switch failure from + // causing us to stall during live. + + this.updateMediaUpdateTimeout_(refreshDelay(playlist, true)); // switching to the active playlist is a no-op + + if (!mediaChange) { + return; + } + this.state = 'SWITCHING_MEDIA'; // there is already an outstanding playlist request + + if (this.request) { + if (playlist.resolvedUri === this.request.url) { + // requesting to switch to the same playlist multiple times + // has no effect after the first + return; + } + this.request.onreadystatechange = null; + this.request.abort(); + this.request = null; + } // request the new playlist + + if (this.media_) { + this.trigger('mediachanging'); + } + this.pendingMedia_ = playlist; + this.request = this.vhs_.xhr({ + uri: playlist.resolvedUri, + withCredentials: this.withCredentials, + requestType: 'hls-playlist' + }, (error, req) => { + // disposed + if (!this.request) { + return; + } + playlist.lastRequest = Date.now(); + playlist.resolvedUri = resolveManifestRedirect(playlist.resolvedUri, req); + if (error) { + return this.playlistRequestError(this.request, playlist, startingState); + } + this.haveMetadata({ + playlistString: req.responseText, + url: playlist.uri, + id: playlist.id + }); // fire loadedmetadata the first time a media playlist is loaded + + if (startingState === 'HAVE_MAIN_MANIFEST') { + this.trigger('loadedmetadata'); + } else { + this.trigger('mediachange'); + } + }); + } + /** + * pause loading of the playlist + */ + + pause() { + if (this.mediaUpdateTimeout) { + window$1.clearTimeout(this.mediaUpdateTimeout); + this.mediaUpdateTimeout = null; + } + this.stopRequest(); + if (this.state === 'HAVE_NOTHING') { + // If we pause the loader before any data has been retrieved, its as if we never + // started, so reset to an unstarted state. + this.started = false; + } // Need to restore state now that no activity is happening + + if (this.state === 'SWITCHING_MEDIA') { + // if the loader was in the process of switching media, it should either return to + // HAVE_MAIN_MANIFEST or HAVE_METADATA depending on if the loader has loaded a media + // playlist yet. This is determined by the existence of loader.media_ + if (this.media_) { + this.state = 'HAVE_METADATA'; + } else { + this.state = 'HAVE_MAIN_MANIFEST'; + } + } else if (this.state === 'HAVE_CURRENT_METADATA') { + this.state = 'HAVE_METADATA'; + } + } + /** + * start loading of the playlist + */ + + load(shouldDelay) { + if (this.mediaUpdateTimeout) { + window$1.clearTimeout(this.mediaUpdateTimeout); + this.mediaUpdateTimeout = null; + } + const media = this.media(); + if (shouldDelay) { + const delay = media ? (media.partTargetDuration || media.targetDuration) / 2 * 1000 : 5 * 1000; + this.mediaUpdateTimeout = window$1.setTimeout(() => { + this.mediaUpdateTimeout = null; + this.load(); + }, delay); + return; + } + if (!this.started) { + this.start(); + return; + } + if (media && !media.endList) { + this.trigger('mediaupdatetimeout'); + } else { + this.trigger('loadedplaylist'); + } + } + updateMediaUpdateTimeout_(delay) { + if (this.mediaUpdateTimeout) { + window$1.clearTimeout(this.mediaUpdateTimeout); + this.mediaUpdateTimeout = null; + } // we only have use mediaupdatetimeout for live playlists. + + if (!this.media() || this.media().endList) { + return; + } + this.mediaUpdateTimeout = window$1.setTimeout(() => { + this.mediaUpdateTimeout = null; + this.trigger('mediaupdatetimeout'); + this.updateMediaUpdateTimeout_(delay); + }, delay); + } + /** + * start loading of the playlist + */ + + start() { + this.started = true; + if (typeof this.src === 'object') { + // in the case of an entirely constructed manifest object (meaning there's no actual + // manifest on a server), default the uri to the page's href + if (!this.src.uri) { + this.src.uri = window$1.location.href; + } // resolvedUri is added on internally after the initial request. Since there's no + // request for pre-resolved manifests, add on resolvedUri here. + + this.src.resolvedUri = this.src.uri; // Since a manifest object was passed in as the source (instead of a URL), the first + // request can be skipped (since the top level of the manifest, at a minimum, is + // already available as a parsed manifest object). However, if the manifest object + // represents a main playlist, some media playlists may need to be resolved before + // the starting segment list is available. Therefore, go directly to setup of the + // initial playlist, and let the normal flow continue from there. + // + // Note that the call to setup is asynchronous, as other sections of VHS may assume + // that the first request is asynchronous. + + setTimeout(() => { + this.setupInitialPlaylist(this.src); + }, 0); + return; + } // request the specified URL + + this.request = this.vhs_.xhr({ + uri: this.src, + withCredentials: this.withCredentials, + requestType: 'hls-playlist' + }, (error, req) => { + // disposed + if (!this.request) { + return; + } // clear the loader's request reference + + this.request = null; + if (error) { + this.error = { + status: req.status, + message: `HLS playlist request error at URL: ${this.src}.`, + responseText: req.responseText, + // MEDIA_ERR_NETWORK + code: 2, + metadata: { + errorType: videojs.Error.HlsPlaylistRequestError + } + }; + if (this.state === 'HAVE_NOTHING') { + this.started = false; + } + return this.trigger('error'); + } + this.src = resolveManifestRedirect(this.src, req); + const manifest = this.parseManifest_({ + manifestString: req.responseText, + url: this.src + }); + this.setupInitialPlaylist(manifest); + }); + } + srcUri() { + return typeof this.src === 'string' ? this.src : this.src.uri; + } + /** + * Given a manifest object that's either a main or media playlist, trigger the proper + * events and set the state of the playlist loader. + * + * If the manifest object represents a main playlist, `loadedplaylist` will be + * triggered to allow listeners to select a playlist. If none is selected, the loader + * will default to the first one in the playlists array. + * + * If the manifest object represents a media playlist, `loadedplaylist` will be + * triggered followed by `loadedmetadata`, as the only available playlist is loaded. + * + * In the case of a media playlist, a main playlist object wrapper with one playlist + * will be created so that all logic can handle playlists in the same fashion (as an + * assumed manifest object schema). + * + * @param {Object} manifest + * The parsed manifest object + */ + + setupInitialPlaylist(manifest) { + this.state = 'HAVE_MAIN_MANIFEST'; + if (manifest.playlists) { + this.main = manifest; + addPropertiesToMain(this.main, this.srcUri()); // If the initial main playlist has playlists wtih segments already resolved, + // then resolve URIs in advance, as they are usually done after a playlist request, + // which may not happen if the playlist is resolved. + + manifest.playlists.forEach(playlist => { + playlist.segments = getAllSegments(playlist); + playlist.segments.forEach(segment => { + resolveSegmentUris(segment, playlist.resolvedUri); + }); + }); + this.trigger('loadedplaylist'); + if (!this.request) { + // no media playlist was specifically selected so start + // from the first listed one + this.media(this.main.playlists[0]); + } + return; + } // In order to support media playlists passed in as vhs-json, the case where the uri + // is not provided as part of the manifest should be considered, and an appropriate + // default used. + + const uri = this.srcUri() || window$1.location.href; + this.main = mainForMedia(manifest, uri); + this.haveMetadata({ + playlistObject: manifest, + url: uri, + id: this.main.playlists[0].id + }); + this.trigger('loadedmetadata'); + } + /** + * Updates or deletes a preexisting pathway clone. + * Ensures that all playlists related to the old pathway clone are + * either updated or deleted. + * + * @param {Object} clone On update, the pathway clone object for the newly updated pathway clone. + * On delete, the old pathway clone object to be deleted. + * @param {boolean} isUpdate True if the pathway is to be updated, + * false if it is meant to be deleted. + */ + + updateOrDeleteClone(clone, isUpdate) { + const main = this.main; + const pathway = clone.ID; + let i = main.playlists.length; // Iterate backwards through the playlist so we can remove playlists if necessary. + + while (i--) { + const p = main.playlists[i]; + if (p.attributes['PATHWAY-ID'] === pathway) { + const oldPlaylistUri = p.resolvedUri; + const oldPlaylistId = p.id; // update the indexed playlist and add new playlists by ID and URI + + if (isUpdate) { + const newPlaylistUri = this.createCloneURI_(p.resolvedUri, clone); + const newPlaylistId = createPlaylistID(pathway, newPlaylistUri); + const attributes = this.createCloneAttributes_(pathway, p.attributes); + const updatedPlaylist = this.createClonePlaylist_(p, newPlaylistId, clone, attributes); + main.playlists[i] = updatedPlaylist; + main.playlists[newPlaylistId] = updatedPlaylist; + main.playlists[newPlaylistUri] = updatedPlaylist; + } else { + // Remove the indexed playlist. + main.playlists.splice(i, 1); + } // Remove playlists by the old ID and URI. + + delete main.playlists[oldPlaylistId]; + delete main.playlists[oldPlaylistUri]; + } + } + this.updateOrDeleteCloneMedia(clone, isUpdate); + } + /** + * Updates or deletes media data based on the pathway clone object. + * Due to the complexity of the media groups and playlists, in all cases + * we remove all of the old media groups and playlists. + * On updates, we then create new media groups and playlists based on the + * new pathway clone object. + * + * @param {Object} clone The pathway clone object for the newly updated pathway clone. + * @param {boolean} isUpdate True if the pathway is to be updated, + * false if it is meant to be deleted. + */ + + updateOrDeleteCloneMedia(clone, isUpdate) { + const main = this.main; + const id = clone.ID; + ['AUDIO', 'SUBTITLES', 'CLOSED-CAPTIONS'].forEach(mediaType => { + if (!main.mediaGroups[mediaType] || !main.mediaGroups[mediaType][id]) { + return; + } + for (const groupKey in main.mediaGroups[mediaType]) { + // Remove all media playlists for the media group for this pathway clone. + if (groupKey === id) { + for (const labelKey in main.mediaGroups[mediaType][groupKey]) { + const oldMedia = main.mediaGroups[mediaType][groupKey][labelKey]; + oldMedia.playlists.forEach((p, i) => { + const oldMediaPlaylist = main.playlists[p.id]; + const oldPlaylistId = oldMediaPlaylist.id; + const oldPlaylistUri = oldMediaPlaylist.resolvedUri; + delete main.playlists[oldPlaylistId]; + delete main.playlists[oldPlaylistUri]; + }); + } // Delete the old media group. + + delete main.mediaGroups[mediaType][groupKey]; + } + } + }); // Create the new media groups and playlists if there is an update. + + if (isUpdate) { + this.createClonedMediaGroups_(clone); + } + } + /** + * Given a pathway clone object, clones all necessary playlists. + * + * @param {Object} clone The pathway clone object. + * @param {Object} basePlaylist The original playlist to clone from. + */ + + addClonePathway(clone, basePlaylist = {}) { + const main = this.main; + const index = main.playlists.length; + const uri = this.createCloneURI_(basePlaylist.resolvedUri, clone); + const playlistId = createPlaylistID(clone.ID, uri); + const attributes = this.createCloneAttributes_(clone.ID, basePlaylist.attributes); + const playlist = this.createClonePlaylist_(basePlaylist, playlistId, clone, attributes); + main.playlists[index] = playlist; // add playlist by ID and URI + + main.playlists[playlistId] = playlist; + main.playlists[uri] = playlist; + this.createClonedMediaGroups_(clone); + } + /** + * Given a pathway clone object we create clones of all media. + * In this function, all necessary information and updated playlists + * are added to the `mediaGroup` object. + * Playlists are also added to the `playlists` array so the media groups + * will be properly linked. + * + * @param {Object} clone The pathway clone object. + */ + + createClonedMediaGroups_(clone) { + const id = clone.ID; + const baseID = clone['BASE-ID']; + const main = this.main; + ['AUDIO', 'SUBTITLES', 'CLOSED-CAPTIONS'].forEach(mediaType => { + // If the media type doesn't exist, or there is already a clone, skip + // to the next media type. + if (!main.mediaGroups[mediaType] || main.mediaGroups[mediaType][id]) { + return; + } + for (const groupKey in main.mediaGroups[mediaType]) { + if (groupKey === baseID) { + // Create the group. + main.mediaGroups[mediaType][id] = {}; + } else { + // There is no need to iterate over label keys in this case. + continue; + } + for (const labelKey in main.mediaGroups[mediaType][groupKey]) { + const oldMedia = main.mediaGroups[mediaType][groupKey][labelKey]; + main.mediaGroups[mediaType][id][labelKey] = _extends({}, oldMedia); + const newMedia = main.mediaGroups[mediaType][id][labelKey]; // update URIs on the media + + const newUri = this.createCloneURI_(oldMedia.resolvedUri, clone); + newMedia.resolvedUri = newUri; + newMedia.uri = newUri; // Reset playlists in the new media group. + + newMedia.playlists = []; // Create new playlists in the newly cloned media group. + + oldMedia.playlists.forEach((p, i) => { + const oldMediaPlaylist = main.playlists[p.id]; + const group = groupID(mediaType, id, labelKey); + const newPlaylistID = createPlaylistID(id, group); // Check to see if it already exists + + if (oldMediaPlaylist && !main.playlists[newPlaylistID]) { + const newMediaPlaylist = this.createClonePlaylist_(oldMediaPlaylist, newPlaylistID, clone); + const newPlaylistUri = newMediaPlaylist.resolvedUri; + main.playlists[newPlaylistID] = newMediaPlaylist; + main.playlists[newPlaylistUri] = newMediaPlaylist; + } + newMedia.playlists[i] = this.createClonePlaylist_(p, newPlaylistID, clone); + }); + } + } + }); + } + /** + * Using the original playlist to be cloned, and the pathway clone object + * information, we create a new playlist. + * + * @param {Object} basePlaylist The original playlist to be cloned from. + * @param {string} id The desired id of the newly cloned playlist. + * @param {Object} clone The pathway clone object. + * @param {Object} attributes An optional object to populate the `attributes` property in the playlist. + * + * @return {Object} The combined cloned playlist. + */ + + createClonePlaylist_(basePlaylist, id, clone, attributes) { + const uri = this.createCloneURI_(basePlaylist.resolvedUri, clone); + const newProps = { + resolvedUri: uri, + uri, + id + }; // Remove all segments from previous playlist in the clone. + + if (basePlaylist.segments) { + newProps.segments = []; + } + if (attributes) { + newProps.attributes = attributes; + } + return merge(basePlaylist, newProps); + } + /** + * Generates an updated URI for a cloned pathway based on the original + * pathway's URI and the paramaters from the pathway clone object in the + * content steering server response. + * + * @param {string} baseUri URI to be updated in the cloned pathway. + * @param {Object} clone The pathway clone object. + * + * @return {string} The updated URI for the cloned pathway. + */ + + createCloneURI_(baseURI, clone) { + const uri = new URL(baseURI); + uri.hostname = clone['URI-REPLACEMENT'].HOST; + const params = clone['URI-REPLACEMENT'].PARAMS; // Add params to the cloned URL. + + for (const key of Object.keys(params)) { + uri.searchParams.set(key, params[key]); + } + return uri.href; + } + /** + * Helper function to create the attributes needed for the new clone. + * This mainly adds the necessary media attributes. + * + * @param {string} id The pathway clone object ID. + * @param {Object} oldAttributes The old attributes to compare to. + * @return {Object} The new attributes to add to the playlist. + */ + + createCloneAttributes_(id, oldAttributes) { + const attributes = { + ['PATHWAY-ID']: id + }; + ['AUDIO', 'SUBTITLES', 'CLOSED-CAPTIONS'].forEach(mediaType => { + if (oldAttributes[mediaType]) { + attributes[mediaType] = id; + } + }); + return attributes; + } + /** + * Returns the key ID set from a playlist + * + * @param {playlist} playlist to fetch the key ID set from. + * @return a Set of 32 digit hex strings that represent the unique keyIds for that playlist. + */ + + getKeyIdSet(playlist) { + if (playlist.contentProtection) { + const keyIds = new Set(); + for (const keysystem in playlist.contentProtection) { + const keyId = playlist.contentProtection[keysystem].attributes.keyId; + if (keyId) { + keyIds.add(keyId.toLowerCase()); + } + } + return keyIds; + } + } +} + +/** + * @file xhr.js + */ + +const callbackWrapper = function (request, error, response, callback) { + const reqResponse = request.responseType === 'arraybuffer' ? request.response : request.responseText; + if (!error && reqResponse) { + request.responseTime = Date.now(); + request.roundTripTime = request.responseTime - request.requestTime; + request.bytesReceived = reqResponse.byteLength || reqResponse.length; + if (!request.bandwidth) { + request.bandwidth = Math.floor(request.bytesReceived / request.roundTripTime * 8 * 1000); + } + } + if (response.headers) { + request.responseHeaders = response.headers; + } // videojs.xhr now uses a specific code on the error + // object to signal that a request has timed out instead + // of setting a boolean on the request object + + if (error && error.code === 'ETIMEDOUT') { + request.timedout = true; + } // videojs.xhr no longer considers status codes outside of 200 and 0 + // (for file uris) to be errors, but the old XHR did, so emulate that + // behavior. Status 206 may be used in response to byterange requests. + + if (!error && !request.aborted && response.statusCode !== 200 && response.statusCode !== 206 && response.statusCode !== 0) { + error = new Error('XHR Failed with a response of: ' + (request && (reqResponse || request.responseText))); + } + callback(error, request); +}; +/** + * Iterates over the request hooks Set and calls them in order + * + * @param {Set} hooks the hook Set to iterate over + * @param {Object} options the request options to pass to the xhr wrapper + * @return the callback hook function return value, the modified or new options Object. + */ + +const callAllRequestHooks = (requestSet, options) => { + if (!requestSet || !requestSet.size) { + return; + } + let newOptions = options; + requestSet.forEach(requestCallback => { + newOptions = requestCallback(newOptions); + }); + return newOptions; +}; +/** + * Iterates over the response hooks Set and calls them in order. + * + * @param {Set} hooks the hook Set to iterate over + * @param {Object} request the xhr request object + * @param {Object} error the xhr error object + * @param {Object} response the xhr response object + */ + +const callAllResponseHooks = (responseSet, request, error, response) => { + if (!responseSet || !responseSet.size) { + return; + } + responseSet.forEach(responseCallback => { + responseCallback(request, error, response); + }); +}; +const xhrFactory = function () { + const xhr = function XhrFunction(options, callback) { + // Add a default timeout + options = merge({ + timeout: 45e3 + }, options); // Allow an optional user-specified function to modify the option + // object before we construct the xhr request + // TODO: Remove beforeRequest in the next major release. + + const beforeRequest = XhrFunction.beforeRequest || videojs.Vhs.xhr.beforeRequest; // onRequest and onResponse hooks as a Set, at either the player or global level. + // TODO: new Set added here for beforeRequest alias. Remove this when beforeRequest is removed. + + const _requestCallbackSet = XhrFunction._requestCallbackSet || videojs.Vhs.xhr._requestCallbackSet || new Set(); + const _responseCallbackSet = XhrFunction._responseCallbackSet || videojs.Vhs.xhr._responseCallbackSet; + if (beforeRequest && typeof beforeRequest === 'function') { + videojs.log.warn('beforeRequest is deprecated, use onRequest instead.'); + _requestCallbackSet.add(beforeRequest); + } // Use the standard videojs.xhr() method unless `videojs.Vhs.xhr` has been overriden + // TODO: switch back to videojs.Vhs.xhr.name === 'XhrFunction' when we drop IE11 + + const xhrMethod = videojs.Vhs.xhr.original === true ? videojs.xhr : videojs.Vhs.xhr; // call all registered onRequest hooks, assign new options. + + const beforeRequestOptions = callAllRequestHooks(_requestCallbackSet, options); // Remove the beforeRequest function from the hooks set so stale beforeRequest functions are not called. + + _requestCallbackSet.delete(beforeRequest); // xhrMethod will call XMLHttpRequest.open and XMLHttpRequest.send + + const request = xhrMethod(beforeRequestOptions || options, function (error, response) { + // call all registered onResponse hooks + callAllResponseHooks(_responseCallbackSet, request, error, response); + return callbackWrapper(request, error, response, callback); + }); + const originalAbort = request.abort; + request.abort = function () { + request.aborted = true; + return originalAbort.apply(request, arguments); + }; + request.uri = options.uri; + request.requestTime = Date.now(); + return request; + }; + xhr.original = true; + return xhr; +}; +/** + * Turns segment byterange into a string suitable for use in + * HTTP Range requests + * + * @param {Object} byterange - an object with two values defining the start and end + * of a byte-range + */ + +const byterangeStr = function (byterange) { + // `byterangeEnd` is one less than `offset + length` because the HTTP range + // header uses inclusive ranges + let byterangeEnd; + const byterangeStart = byterange.offset; + if (typeof byterange.offset === 'bigint' || typeof byterange.length === 'bigint') { + byterangeEnd = window$1.BigInt(byterange.offset) + window$1.BigInt(byterange.length) - window$1.BigInt(1); + } else { + byterangeEnd = byterange.offset + byterange.length - 1; + } + return 'bytes=' + byterangeStart + '-' + byterangeEnd; +}; +/** + * Defines headers for use in the xhr request for a particular segment. + * + * @param {Object} segment - a simplified copy of the segmentInfo object + * from SegmentLoader + */ + +const segmentXhrHeaders = function (segment) { + const headers = {}; + if (segment.byterange) { + headers.Range = byterangeStr(segment.byterange); + } + return headers; +}; + +/** + * @file bin-utils.js + */ + +/** + * convert a TimeRange to text + * + * @param {TimeRange} range the timerange to use for conversion + * @param {number} i the iterator on the range to convert + * @return {string} the range in string format + */ + +const textRange = function (range, i) { + return range.start(i) + '-' + range.end(i); +}; +/** + * format a number as hex string + * + * @param {number} e The number + * @param {number} i the iterator + * @return {string} the hex formatted number as a string + */ + +const formatHexString = function (e, i) { + const value = e.toString(16); + return '00'.substring(0, 2 - value.length) + value + (i % 2 ? ' ' : ''); +}; +const formatAsciiString = function (e) { + if (e >= 0x20 && e < 0x7e) { + return String.fromCharCode(e); + } + return '.'; +}; +/** + * Creates an object for sending to a web worker modifying properties that are TypedArrays + * into a new object with seperated properties for the buffer, byteOffset, and byteLength. + * + * @param {Object} message + * Object of properties and values to send to the web worker + * @return {Object} + * Modified message with TypedArray values expanded + * @function createTransferableMessage + */ + +const createTransferableMessage = function (message) { + const transferable = {}; + Object.keys(message).forEach(key => { + const value = message[key]; + if (isArrayBufferView(value)) { + transferable[key] = { + bytes: value.buffer, + byteOffset: value.byteOffset, + byteLength: value.byteLength + }; + } else { + transferable[key] = value; + } + }); + return transferable; +}; +/** + * Returns a unique string identifier for a media initialization + * segment. + * + * @param {Object} initSegment + * the init segment object. + * + * @return {string} the generated init segment id + */ + +const initSegmentId = function (initSegment) { + const byterange = initSegment.byterange || { + length: Infinity, + offset: 0 + }; + return [byterange.length, byterange.offset, initSegment.resolvedUri].join(','); +}; +/** + * Returns a unique string identifier for a media segment key. + * + * @param {Object} key the encryption key + * @return {string} the unique id for the media segment key. + */ + +const segmentKeyId = function (key) { + return key.resolvedUri; +}; +/** + * utils to help dump binary data to the console + * + * @param {Array|TypedArray} data + * data to dump to a string + * + * @return {string} the data as a hex string. + */ + +const hexDump = data => { + const bytes = Array.prototype.slice.call(data); + const step = 16; + let result = ''; + let hex; + let ascii; + for (let j = 0; j < bytes.length / step; j++) { + hex = bytes.slice(j * step, j * step + step).map(formatHexString).join(''); + ascii = bytes.slice(j * step, j * step + step).map(formatAsciiString).join(''); + result += hex + ' ' + ascii + '\n'; + } + return result; +}; +const tagDump = ({ + bytes +}) => hexDump(bytes); +const textRanges = ranges => { + let result = ''; + let i; + for (i = 0; i < ranges.length; i++) { + result += textRange(ranges, i) + ' '; + } + return result; +}; +var utils = /*#__PURE__*/Object.freeze({ + __proto__: null, + createTransferableMessage: createTransferableMessage, + initSegmentId: initSegmentId, + segmentKeyId: segmentKeyId, + hexDump: hexDump, + tagDump: tagDump, + textRanges: textRanges +}); + +// TODO handle fmp4 case where the timing info is accurate and doesn't involve transmux +// 25% was arbitrarily chosen, and may need to be refined over time. + +const SEGMENT_END_FUDGE_PERCENT = 0.25; +/** + * Converts a player time (any time that can be gotten/set from player.currentTime(), + * e.g., any time within player.seekable().start(0) to player.seekable().end(0)) to a + * program time (any time referencing the real world (e.g., EXT-X-PROGRAM-DATE-TIME)). + * + * The containing segment is required as the EXT-X-PROGRAM-DATE-TIME serves as an "anchor + * point" (a point where we have a mapping from program time to player time, with player + * time being the post transmux start of the segment). + * + * For more details, see [this doc](../../docs/program-time-from-player-time.md). + * + * @param {number} playerTime the player time + * @param {Object} segment the segment which contains the player time + * @return {Date} program time + */ + +const playerTimeToProgramTime = (playerTime, segment) => { + if (!segment.dateTimeObject) { + // Can't convert without an "anchor point" for the program time (i.e., a time that can + // be used to map the start of a segment with a real world time). + return null; + } + const transmuxerPrependedSeconds = segment.videoTimingInfo.transmuxerPrependedSeconds; + const transmuxedStart = segment.videoTimingInfo.transmuxedPresentationStart; // get the start of the content from before old content is prepended + + const startOfSegment = transmuxedStart + transmuxerPrependedSeconds; + const offsetFromSegmentStart = playerTime - startOfSegment; + return new Date(segment.dateTimeObject.getTime() + offsetFromSegmentStart * 1000); +}; +const originalSegmentVideoDuration = videoTimingInfo => { + return videoTimingInfo.transmuxedPresentationEnd - videoTimingInfo.transmuxedPresentationStart - videoTimingInfo.transmuxerPrependedSeconds; +}; +/** + * Finds a segment that contains the time requested given as an ISO-8601 string. The + * returned segment might be an estimate or an accurate match. + * + * @param {string} programTime The ISO-8601 programTime to find a match for + * @param {Object} playlist A playlist object to search within + */ + +const findSegmentForProgramTime = (programTime, playlist) => { + // Assumptions: + // - verifyProgramDateTimeTags has already been run + // - live streams have been started + let dateTimeObject; + try { + dateTimeObject = new Date(programTime); + } catch (e) { + return null; + } + if (!playlist || !playlist.segments || playlist.segments.length === 0) { + return null; + } + let segment = playlist.segments[0]; + if (dateTimeObject < new Date(segment.dateTimeObject)) { + // Requested time is before stream start. + return null; + } + for (let i = 0; i < playlist.segments.length - 1; i++) { + segment = playlist.segments[i]; + const nextSegmentStart = new Date(playlist.segments[i + 1].dateTimeObject); + if (dateTimeObject < nextSegmentStart) { + break; + } + } + const lastSegment = playlist.segments[playlist.segments.length - 1]; + const lastSegmentStart = lastSegment.dateTimeObject; + const lastSegmentDuration = lastSegment.videoTimingInfo ? originalSegmentVideoDuration(lastSegment.videoTimingInfo) : lastSegment.duration + lastSegment.duration * SEGMENT_END_FUDGE_PERCENT; + const lastSegmentEnd = new Date(lastSegmentStart.getTime() + lastSegmentDuration * 1000); + if (dateTimeObject > lastSegmentEnd) { + // Beyond the end of the stream, or our best guess of the end of the stream. + return null; + } + if (dateTimeObject > new Date(lastSegmentStart)) { + segment = lastSegment; + } + return { + segment, + estimatedStart: segment.videoTimingInfo ? segment.videoTimingInfo.transmuxedPresentationStart : Playlist.duration(playlist, playlist.mediaSequence + playlist.segments.indexOf(segment)), + // Although, given that all segments have accurate date time objects, the segment + // selected should be accurate, unless the video has been transmuxed at some point + // (determined by the presence of the videoTimingInfo object), the segment's "player + // time" (the start time in the player) can't be considered accurate. + type: segment.videoTimingInfo ? 'accurate' : 'estimate' + }; +}; +/** + * Finds a segment that contains the given player time(in seconds). + * + * @param {number} time The player time to find a match for + * @param {Object} playlist A playlist object to search within + */ + +const findSegmentForPlayerTime = (time, playlist) => { + // Assumptions: + // - there will always be a segment.duration + // - we can start from zero + // - segments are in time order + if (!playlist || !playlist.segments || playlist.segments.length === 0) { + return null; + } + let segmentEnd = 0; + let segment; + for (let i = 0; i < playlist.segments.length; i++) { + segment = playlist.segments[i]; // videoTimingInfo is set after the segment is downloaded and transmuxed, and + // should contain the most accurate values we have for the segment's player times. + // + // Use the accurate transmuxedPresentationEnd value if it is available, otherwise fall + // back to an estimate based on the manifest derived (inaccurate) segment.duration, to + // calculate an end value. + + segmentEnd = segment.videoTimingInfo ? segment.videoTimingInfo.transmuxedPresentationEnd : segmentEnd + segment.duration; + if (time <= segmentEnd) { + break; + } + } + const lastSegment = playlist.segments[playlist.segments.length - 1]; + if (lastSegment.videoTimingInfo && lastSegment.videoTimingInfo.transmuxedPresentationEnd < time) { + // The time requested is beyond the stream end. + return null; + } + if (time > segmentEnd) { + // The time is within or beyond the last segment. + // + // Check to see if the time is beyond a reasonable guess of the end of the stream. + if (time > segmentEnd + lastSegment.duration * SEGMENT_END_FUDGE_PERCENT) { + // Technically, because the duration value is only an estimate, the time may still + // exist in the last segment, however, there isn't enough information to make even + // a reasonable estimate. + return null; + } + segment = lastSegment; + } + return { + segment, + estimatedStart: segment.videoTimingInfo ? segment.videoTimingInfo.transmuxedPresentationStart : segmentEnd - segment.duration, + // Because videoTimingInfo is only set after transmux, it is the only way to get + // accurate timing values. + type: segment.videoTimingInfo ? 'accurate' : 'estimate' + }; +}; +/** + * Gives the offset of the comparisonTimestamp from the programTime timestamp in seconds. + * If the offset returned is positive, the programTime occurs after the + * comparisonTimestamp. + * If the offset is negative, the programTime occurs before the comparisonTimestamp. + * + * @param {string} comparisonTimeStamp An ISO-8601 timestamp to compare against + * @param {string} programTime The programTime as an ISO-8601 string + * @return {number} offset + */ + +const getOffsetFromTimestamp = (comparisonTimeStamp, programTime) => { + let segmentDateTime; + let programDateTime; + try { + segmentDateTime = new Date(comparisonTimeStamp); + programDateTime = new Date(programTime); + } catch (e) {// TODO handle error + } + const segmentTimeEpoch = segmentDateTime.getTime(); + const programTimeEpoch = programDateTime.getTime(); + return (programTimeEpoch - segmentTimeEpoch) / 1000; +}; +/** + * Checks that all segments in this playlist have programDateTime tags. + * + * @param {Object} playlist A playlist object + */ + +const verifyProgramDateTimeTags = playlist => { + if (!playlist.segments || playlist.segments.length === 0) { + return false; + } + for (let i = 0; i < playlist.segments.length; i++) { + const segment = playlist.segments[i]; + if (!segment.dateTimeObject) { + return false; + } + } + return true; +}; +/** + * Returns the programTime of the media given a playlist and a playerTime. + * The playlist must have programDateTime tags for a programDateTime tag to be returned. + * If the segments containing the time requested have not been buffered yet, an estimate + * may be returned to the callback. + * + * @param {Object} args + * @param {Object} args.playlist A playlist object to search within + * @param {number} time A playerTime in seconds + * @param {Function} callback(err, programTime) + * @return {string} err.message A detailed error message + * @return {Object} programTime + * @return {number} programTime.mediaSeconds The streamTime in seconds + * @return {string} programTime.programDateTime The programTime as an ISO-8601 String + */ + +const getProgramTime = ({ + playlist, + time = undefined, + callback +}) => { + if (!callback) { + throw new Error('getProgramTime: callback must be provided'); + } + if (!playlist || time === undefined) { + return callback({ + message: 'getProgramTime: playlist and time must be provided' + }); + } + const matchedSegment = findSegmentForPlayerTime(time, playlist); + if (!matchedSegment) { + return callback({ + message: 'valid programTime was not found' + }); + } + if (matchedSegment.type === 'estimate') { + return callback({ + message: 'Accurate programTime could not be determined.' + ' Please seek to e.seekTime and try again', + seekTime: matchedSegment.estimatedStart + }); + } + const programTimeObject = { + mediaSeconds: time + }; + const programTime = playerTimeToProgramTime(time, matchedSegment.segment); + if (programTime) { + programTimeObject.programDateTime = programTime.toISOString(); + } + return callback(null, programTimeObject); +}; +/** + * Seeks in the player to a time that matches the given programTime ISO-8601 string. + * + * @param {Object} args + * @param {string} args.programTime A programTime to seek to as an ISO-8601 String + * @param {Object} args.playlist A playlist to look within + * @param {number} args.retryCount The number of times to try for an accurate seek. Default is 2. + * @param {Function} args.seekTo A method to perform a seek + * @param {boolean} args.pauseAfterSeek Whether to end in a paused state after seeking. Default is true. + * @param {Object} args.tech The tech to seek on + * @param {Function} args.callback(err, newTime) A callback to return the new time to + * @return {string} err.message A detailed error message + * @return {number} newTime The exact time that was seeked to in seconds + */ + +const seekToProgramTime = ({ + programTime, + playlist, + retryCount = 2, + seekTo, + pauseAfterSeek = true, + tech, + callback +}) => { + if (!callback) { + throw new Error('seekToProgramTime: callback must be provided'); + } + if (typeof programTime === 'undefined' || !playlist || !seekTo) { + return callback({ + message: 'seekToProgramTime: programTime, seekTo and playlist must be provided' + }); + } + if (!playlist.endList && !tech.hasStarted_) { + return callback({ + message: 'player must be playing a live stream to start buffering' + }); + } + if (!verifyProgramDateTimeTags(playlist)) { + return callback({ + message: 'programDateTime tags must be provided in the manifest ' + playlist.resolvedUri + }); + } + const matchedSegment = findSegmentForProgramTime(programTime, playlist); // no match + + if (!matchedSegment) { + return callback({ + message: `${programTime} was not found in the stream` + }); + } + const segment = matchedSegment.segment; + const mediaOffset = getOffsetFromTimestamp(segment.dateTimeObject, programTime); + if (matchedSegment.type === 'estimate') { + // we've run out of retries + if (retryCount === 0) { + return callback({ + message: `${programTime} is not buffered yet. Try again` + }); + } + seekTo(matchedSegment.estimatedStart + mediaOffset); + tech.one('seeked', () => { + seekToProgramTime({ + programTime, + playlist, + retryCount: retryCount - 1, + seekTo, + pauseAfterSeek, + tech, + callback + }); + }); + return; + } // Since the segment.start value is determined from the buffered end or ending time + // of the prior segment, the seekToTime doesn't need to account for any transmuxer + // modifications. + + const seekToTime = segment.start + mediaOffset; + const seekedCallback = () => { + return callback(null, tech.currentTime()); + }; // listen for seeked event + + tech.one('seeked', seekedCallback); // pause before seeking as video.js will restore this state + + if (pauseAfterSeek) { + tech.pause(); + } + seekTo(seekToTime); +}; + +// which will only happen if the request is complete. + +const callbackOnCompleted = (request, cb) => { + if (request.readyState === 4) { + return cb(); + } + return; +}; +const containerRequest = (uri, xhr, cb) => { + let bytes = []; + let id3Offset; + let finished = false; + const endRequestAndCallback = function (err, req, type, _bytes) { + req.abort(); + finished = true; + return cb(err, req, type, _bytes); + }; + const progressListener = function (error, request) { + if (finished) { + return; + } + if (error) { + return endRequestAndCallback(error, request, '', bytes); + } // grap the new part of content that was just downloaded + + const newPart = request.responseText.substring(bytes && bytes.byteLength || 0, request.responseText.length); // add that onto bytes + + bytes = concatTypedArrays(bytes, stringToBytes(newPart, true)); + id3Offset = id3Offset || getId3Offset(bytes); // we need at least 10 bytes to determine a type + // or we need at least two bytes after an id3Offset + + if (bytes.length < 10 || id3Offset && bytes.length < id3Offset + 2) { + return callbackOnCompleted(request, () => endRequestAndCallback(error, request, '', bytes)); + } + const type = detectContainerForBytes(bytes); // if this looks like a ts segment but we don't have enough data + // to see the second sync byte, wait until we have enough data + // before declaring it ts + + if (type === 'ts' && bytes.length < 188) { + return callbackOnCompleted(request, () => endRequestAndCallback(error, request, '', bytes)); + } // this may be an unsynced ts segment + // wait for 376 bytes before detecting no container + + if (!type && bytes.length < 376) { + return callbackOnCompleted(request, () => endRequestAndCallback(error, request, '', bytes)); + } + return endRequestAndCallback(null, request, type, bytes); + }; + const options = { + uri, + beforeSend(request) { + // this forces the browser to pass the bytes to us unprocessed + request.overrideMimeType('text/plain; charset=x-user-defined'); + request.addEventListener('progress', function ({ + total, + loaded + }) { + return callbackWrapper(request, null, { + statusCode: request.status + }, progressListener); + }); + } + }; + const request = xhr(options, function (error, response) { + return callbackWrapper(request, error, response, progressListener); + }); + return request; +}; +const { + EventTarget +} = videojs; +const dashPlaylistUnchanged = function (a, b) { + if (!isPlaylistUnchanged(a, b)) { + return false; + } // for dash the above check will often return true in scenarios where + // the playlist actually has changed because mediaSequence isn't a + // dash thing, and we often set it to 1. So if the playlists have the same amount + // of segments we return true. + // So for dash we need to make sure that the underlying segments are different. + // if sidx changed then the playlists are different. + + if (a.sidx && b.sidx && (a.sidx.offset !== b.sidx.offset || a.sidx.length !== b.sidx.length)) { + return false; + } else if (!a.sidx && b.sidx || a.sidx && !b.sidx) { + return false; + } // one or the other does not have segments + // there was a change. + + if (a.segments && !b.segments || !a.segments && b.segments) { + return false; + } // neither has segments nothing changed + + if (!a.segments && !b.segments) { + return true; + } // check segments themselves + + for (let i = 0; i < a.segments.length; i++) { + const aSegment = a.segments[i]; + const bSegment = b.segments[i]; // if uris are different between segments there was a change + + if (aSegment.uri !== bSegment.uri) { + return false; + } // neither segment has a byterange, there will be no byterange change. + + if (!aSegment.byterange && !bSegment.byterange) { + continue; + } + const aByterange = aSegment.byterange; + const bByterange = bSegment.byterange; // if byterange only exists on one of the segments, there was a change. + + if (aByterange && !bByterange || !aByterange && bByterange) { + return false; + } // if both segments have byterange with different offsets, there was a change. + + if (aByterange.offset !== bByterange.offset || aByterange.length !== bByterange.length) { + return false; + } + } // if everything was the same with segments, this is the same playlist. + + return true; +}; +/** + * Use the representation IDs from the mpd object to create groupIDs, the NAME is set to mandatory representation + * ID in the parser. This allows for continuous playout across periods with the same representation IDs + * (continuous periods as defined in DASH-IF 3.2.12). This is assumed in the mpd-parser as well. If we want to support + * periods without continuous playback this function may need modification as well as the parser. + */ + +const dashGroupId = (type, group, label, playlist) => { + // If the manifest somehow does not have an ID (non-dash compliant), use the label. + const playlistId = playlist.attributes.NAME || label; + return `placeholder-uri-${type}-${group}-${playlistId}`; +}; +/** + * Parses the main XML string and updates playlist URI references. + * + * @param {Object} config + * Object of arguments + * @param {string} config.mainXml + * The mpd XML + * @param {string} config.srcUrl + * The mpd URL + * @param {Date} config.clientOffset + * A time difference between server and client + * @param {Object} config.sidxMapping + * SIDX mappings for moof/mdat URIs and byte ranges + * @return {Object} + * The parsed mpd manifest object + */ + +const parseMainXml = ({ + mainXml, + srcUrl, + clientOffset, + sidxMapping, + previousManifest +}) => { + const manifest = parse(mainXml, { + manifestUri: srcUrl, + clientOffset, + sidxMapping, + previousManifest + }); + addPropertiesToMain(manifest, srcUrl, dashGroupId); + return manifest; +}; +/** + * Removes any mediaGroup labels that no longer exist in the newMain + * + * @param {Object} update + * The previous mpd object being updated + * @param {Object} newMain + * The new mpd object + */ + +const removeOldMediaGroupLabels = (update, newMain) => { + forEachMediaGroup(update, (properties, type, group, label) => { + if (!(label in newMain.mediaGroups[type][group])) { + delete update.mediaGroups[type][group][label]; + } + }); +}; +/** + * Returns a new main manifest that is the result of merging an updated main manifest + * into the original version. + * + * @param {Object} oldMain + * The old parsed mpd object + * @param {Object} newMain + * The updated parsed mpd object + * @return {Object} + * A new object representing the original main manifest with the updated media + * playlists merged in + */ + +const updateMain = (oldMain, newMain, sidxMapping) => { + let noChanges = true; + let update = merge(oldMain, { + // These are top level properties that can be updated + duration: newMain.duration, + minimumUpdatePeriod: newMain.minimumUpdatePeriod, + timelineStarts: newMain.timelineStarts + }); // First update the playlists in playlist list + + for (let i = 0; i < newMain.playlists.length; i++) { + const playlist = newMain.playlists[i]; + if (playlist.sidx) { + const sidxKey = generateSidxKey(playlist.sidx); // add sidx segments to the playlist if we have all the sidx info already + + if (sidxMapping && sidxMapping[sidxKey] && sidxMapping[sidxKey].sidx) { + addSidxSegmentsToPlaylist(playlist, sidxMapping[sidxKey].sidx, playlist.sidx.resolvedUri); + } + } + const playlistUpdate = updateMain$1(update, playlist, dashPlaylistUnchanged); + if (playlistUpdate) { + update = playlistUpdate; + noChanges = false; + } + } // Then update media group playlists + + forEachMediaGroup(newMain, (properties, type, group, label) => { + if (properties.playlists && properties.playlists.length) { + const id = properties.playlists[0].id; + const playlistUpdate = updateMain$1(update, properties.playlists[0], dashPlaylistUnchanged); + if (playlistUpdate) { + update = playlistUpdate; // add new mediaGroup label if it doesn't exist and assign the new mediaGroup. + + if (!(label in update.mediaGroups[type][group])) { + update.mediaGroups[type][group][label] = properties; + } // update the playlist reference within media groups + + update.mediaGroups[type][group][label].playlists[0] = update.playlists[id]; + noChanges = false; + } + } + }); // remove mediaGroup labels and references that no longer exist in the newMain + + removeOldMediaGroupLabels(update, newMain); + if (newMain.minimumUpdatePeriod !== oldMain.minimumUpdatePeriod) { + noChanges = false; + } + if (noChanges) { + return null; + } + return update; +}; // SIDX should be equivalent if the URI and byteranges of the SIDX match. +// If the SIDXs have maps, the two maps should match, +// both `a` and `b` missing SIDXs is considered matching. +// If `a` or `b` but not both have a map, they aren't matching. + +const equivalentSidx = (a, b) => { + const neitherMap = Boolean(!a.map && !b.map); + const equivalentMap = neitherMap || Boolean(a.map && b.map && a.map.byterange.offset === b.map.byterange.offset && a.map.byterange.length === b.map.byterange.length); + return equivalentMap && a.uri === b.uri && a.byterange.offset === b.byterange.offset && a.byterange.length === b.byterange.length; +}; // exported for testing + +const compareSidxEntry = (playlists, oldSidxMapping) => { + const newSidxMapping = {}; + for (const id in playlists) { + const playlist = playlists[id]; + const currentSidxInfo = playlist.sidx; + if (currentSidxInfo) { + const key = generateSidxKey(currentSidxInfo); + if (!oldSidxMapping[key]) { + break; + } + const savedSidxInfo = oldSidxMapping[key].sidxInfo; + if (equivalentSidx(savedSidxInfo, currentSidxInfo)) { + newSidxMapping[key] = oldSidxMapping[key]; + } + } + } + return newSidxMapping; +}; +/** + * A function that filters out changed items as they need to be requested separately. + * + * The method is exported for testing + * + * @param {Object} main the parsed mpd XML returned via mpd-parser + * @param {Object} oldSidxMapping the SIDX to compare against + */ + +const filterChangedSidxMappings = (main, oldSidxMapping) => { + const videoSidx = compareSidxEntry(main.playlists, oldSidxMapping); + let mediaGroupSidx = videoSidx; + forEachMediaGroup(main, (properties, mediaType, groupKey, labelKey) => { + if (properties.playlists && properties.playlists.length) { + const playlists = properties.playlists; + mediaGroupSidx = merge(mediaGroupSidx, compareSidxEntry(playlists, oldSidxMapping)); + } + }); + return mediaGroupSidx; +}; +class DashPlaylistLoader extends EventTarget { + // DashPlaylistLoader must accept either a src url or a playlist because subsequent + // playlist loader setups from media groups will expect to be able to pass a playlist + // (since there aren't external URLs to media playlists with DASH) + constructor(srcUrlOrPlaylist, vhs, options = {}, mainPlaylistLoader) { + super(); + this.mainPlaylistLoader_ = mainPlaylistLoader || this; + if (!mainPlaylistLoader) { + this.isMain_ = true; + } + const { + withCredentials = false + } = options; + this.vhs_ = vhs; + this.withCredentials = withCredentials; + this.addMetadataToTextTrack = options.addMetadataToTextTrack; + if (!srcUrlOrPlaylist) { + throw new Error('A non-empty playlist URL or object is required'); + } // event naming? + + this.on('minimumUpdatePeriod', () => { + this.refreshXml_(); + }); // live playlist staleness timeout + + this.on('mediaupdatetimeout', () => { + this.refreshMedia_(this.media().id); + }); + this.state = 'HAVE_NOTHING'; + this.loadedPlaylists_ = {}; + this.logger_ = logger('DashPlaylistLoader'); // initialize the loader state + // The mainPlaylistLoader will be created with a string + + if (this.isMain_) { + this.mainPlaylistLoader_.srcUrl = srcUrlOrPlaylist; // TODO: reset sidxMapping between period changes + // once multi-period is refactored + + this.mainPlaylistLoader_.sidxMapping_ = {}; + } else { + this.childPlaylist_ = srcUrlOrPlaylist; + } + } + requestErrored_(err, request, startingState) { + // disposed + if (!this.request) { + return true; + } // pending request is cleared + + this.request = null; + if (err) { + // use the provided error object or create one + // based on the request/response + this.error = typeof err === 'object' && !(err instanceof Error) ? err : { + status: request.status, + message: 'DASH request error at URL: ' + request.uri, + response: request.response, + // MEDIA_ERR_NETWORK + code: 2, + metadata: err.metadata + }; + if (startingState) { + this.state = startingState; + } + this.trigger('error'); + return true; + } + } + /** + * Verify that the container of the sidx segment can be parsed + * and if it can, get and parse that segment. + */ + + addSidxSegments_(playlist, startingState, cb) { + const sidxKey = playlist.sidx && generateSidxKey(playlist.sidx); // playlist lacks sidx or sidx segments were added to this playlist already. + + if (!playlist.sidx || !sidxKey || this.mainPlaylistLoader_.sidxMapping_[sidxKey]) { + // keep this function async + this.mediaRequest_ = window$1.setTimeout(() => cb(false), 0); + return; + } // resolve the segment URL relative to the playlist + + const uri = resolveManifestRedirect(playlist.sidx.resolvedUri); + const fin = (err, request) => { + // TODO: add error metdata here once we create an error type in video.js + if (this.requestErrored_(err, request, startingState)) { + return; + } + const sidxMapping = this.mainPlaylistLoader_.sidxMapping_; + let sidx; + try { + sidx = parseSidx(toUint8(request.response).subarray(8)); + } catch (e) { + e.metadata = { + errorType: videojs.Error.DashManifestSidxParsingError + }; // sidx parsing failed. + + this.requestErrored_(e, request, startingState); + return; + } + sidxMapping[sidxKey] = { + sidxInfo: playlist.sidx, + sidx + }; + addSidxSegmentsToPlaylist(playlist, sidx, playlist.sidx.resolvedUri); + return cb(true); + }; + this.request = containerRequest(uri, this.vhs_.xhr, (err, request, container, bytes) => { + if (err) { + return fin(err, request); + } + if (!container || container !== 'mp4') { + const sidxContainer = container || 'unknown'; + return fin({ + status: request.status, + message: `Unsupported ${sidxContainer} container type for sidx segment at URL: ${uri}`, + // response is just bytes in this case + // but we really don't want to return that. + response: '', + playlist, + internal: true, + playlistExclusionDuration: Infinity, + // MEDIA_ERR_NETWORK + code: 2, + metadata: { + errorType: videojs.Error.UnsupportedSidxContainer, + sidxContainer + } + }, request); + } // if we already downloaded the sidx bytes in the container request, use them + + const { + offset, + length + } = playlist.sidx.byterange; + if (bytes.length >= length + offset) { + return fin(err, { + response: bytes.subarray(offset, offset + length), + status: request.status, + uri: request.uri + }); + } // otherwise request sidx bytes + + this.request = this.vhs_.xhr({ + uri, + responseType: 'arraybuffer', + headers: segmentXhrHeaders({ + byterange: playlist.sidx.byterange + }) + }, fin); + }); + } + dispose() { + this.trigger('dispose'); + this.stopRequest(); + this.loadedPlaylists_ = {}; + window$1.clearTimeout(this.minimumUpdatePeriodTimeout_); + window$1.clearTimeout(this.mediaRequest_); + window$1.clearTimeout(this.mediaUpdateTimeout); + this.mediaUpdateTimeout = null; + this.mediaRequest_ = null; + this.minimumUpdatePeriodTimeout_ = null; + if (this.mainPlaylistLoader_.createMupOnMedia_) { + this.off('loadedmetadata', this.mainPlaylistLoader_.createMupOnMedia_); + this.mainPlaylistLoader_.createMupOnMedia_ = null; + } + this.off(); + } + hasPendingRequest() { + return this.request || this.mediaRequest_; + } + stopRequest() { + if (this.request) { + const oldRequest = this.request; + this.request = null; + oldRequest.onreadystatechange = null; + oldRequest.abort(); + } + } + media(playlist) { + // getter + if (!playlist) { + return this.media_; + } // setter + + if (this.state === 'HAVE_NOTHING') { + throw new Error('Cannot switch media playlist from ' + this.state); + } + const startingState = this.state; // find the playlist object if the target playlist has been specified by URI + + if (typeof playlist === 'string') { + if (!this.mainPlaylistLoader_.main.playlists[playlist]) { + throw new Error('Unknown playlist URI: ' + playlist); + } + playlist = this.mainPlaylistLoader_.main.playlists[playlist]; + } + const mediaChange = !this.media_ || playlist.id !== this.media_.id; // switch to previously loaded playlists immediately + + if (mediaChange && this.loadedPlaylists_[playlist.id] && this.loadedPlaylists_[playlist.id].endList) { + this.state = 'HAVE_METADATA'; + this.media_ = playlist; // trigger media change if the active media has been updated + + if (mediaChange) { + this.trigger('mediachanging'); + this.trigger('mediachange'); + } + return; + } // switching to the active playlist is a no-op + + if (!mediaChange) { + return; + } // switching from an already loaded playlist + + if (this.media_) { + this.trigger('mediachanging'); + } + this.addSidxSegments_(playlist, startingState, sidxChanged => { + // everything is ready just continue to haveMetadata + this.haveMetadata({ + startingState, + playlist + }); + }); + } + haveMetadata({ + startingState, + playlist + }) { + this.state = 'HAVE_METADATA'; + this.loadedPlaylists_[playlist.id] = playlist; + this.mediaRequest_ = null; // This will trigger loadedplaylist + + this.refreshMedia_(playlist.id); // fire loadedmetadata the first time a media playlist is loaded + // to resolve setup of media groups + + if (startingState === 'HAVE_MAIN_MANIFEST') { + this.trigger('loadedmetadata'); + } else { + // trigger media change if the active media has been updated + this.trigger('mediachange'); + } + } + pause() { + if (this.mainPlaylistLoader_.createMupOnMedia_) { + this.off('loadedmetadata', this.mainPlaylistLoader_.createMupOnMedia_); + this.mainPlaylistLoader_.createMupOnMedia_ = null; + } + this.stopRequest(); + window$1.clearTimeout(this.mediaUpdateTimeout); + this.mediaUpdateTimeout = null; + if (this.isMain_) { + window$1.clearTimeout(this.mainPlaylistLoader_.minimumUpdatePeriodTimeout_); + this.mainPlaylistLoader_.minimumUpdatePeriodTimeout_ = null; + } + if (this.state === 'HAVE_NOTHING') { + // If we pause the loader before any data has been retrieved, its as if we never + // started, so reset to an unstarted state. + this.started = false; + } + } + load(isFinalRendition) { + window$1.clearTimeout(this.mediaUpdateTimeout); + this.mediaUpdateTimeout = null; + const media = this.media(); + if (isFinalRendition) { + const delay = media ? media.targetDuration / 2 * 1000 : 5 * 1000; + this.mediaUpdateTimeout = window$1.setTimeout(() => this.load(), delay); + return; + } // because the playlists are internal to the manifest, load should either load the + // main manifest, or do nothing but trigger an event + + if (!this.started) { + this.start(); + return; + } + if (media && !media.endList) { + // Check to see if this is the main loader and the MUP was cleared (this happens + // when the loader was paused). `media` should be set at this point since one is always + // set during `start()`. + if (this.isMain_ && !this.minimumUpdatePeriodTimeout_) { + // Trigger minimumUpdatePeriod to refresh the main manifest + this.trigger('minimumUpdatePeriod'); // Since there was no prior minimumUpdatePeriodTimeout it should be recreated + + this.updateMinimumUpdatePeriodTimeout_(); + } + this.trigger('mediaupdatetimeout'); + } else { + this.trigger('loadedplaylist'); + } + } + start() { + this.started = true; // We don't need to request the main manifest again + // Call this asynchronously to match the xhr request behavior below + + if (!this.isMain_) { + this.mediaRequest_ = window$1.setTimeout(() => this.haveMain_(), 0); + return; + } + this.requestMain_((req, mainChanged) => { + this.haveMain_(); + if (!this.hasPendingRequest() && !this.media_) { + this.media(this.mainPlaylistLoader_.main.playlists[0]); + } + }); + } + requestMain_(cb) { + this.request = this.vhs_.xhr({ + uri: this.mainPlaylistLoader_.srcUrl, + withCredentials: this.withCredentials, + requestType: 'dash-manifest' + }, (error, req) => { + if (this.requestErrored_(error, req)) { + if (this.state === 'HAVE_NOTHING') { + this.started = false; + } + return; + } + const mainChanged = req.responseText !== this.mainPlaylistLoader_.mainXml_; + this.mainPlaylistLoader_.mainXml_ = req.responseText; + if (req.responseHeaders && req.responseHeaders.date) { + this.mainLoaded_ = Date.parse(req.responseHeaders.date); + } else { + this.mainLoaded_ = Date.now(); + } + this.mainPlaylistLoader_.srcUrl = resolveManifestRedirect(this.mainPlaylistLoader_.srcUrl, req); + if (mainChanged) { + this.handleMain_(); + this.syncClientServerClock_(() => { + return cb(req, mainChanged); + }); + return; + } + return cb(req, mainChanged); + }); + } + /** + * Parses the main xml for UTCTiming node to sync the client clock to the server + * clock. If the UTCTiming node requires a HEAD or GET request, that request is made. + * + * @param {Function} done + * Function to call when clock sync has completed + */ + + syncClientServerClock_(done) { + const utcTiming = parseUTCTiming(this.mainPlaylistLoader_.mainXml_); // No UTCTiming element found in the mpd. Use Date header from mpd request as the + // server clock + + if (utcTiming === null) { + this.mainPlaylistLoader_.clientOffset_ = this.mainLoaded_ - Date.now(); + return done(); + } + if (utcTiming.method === 'DIRECT') { + this.mainPlaylistLoader_.clientOffset_ = utcTiming.value - Date.now(); + return done(); + } + this.request = this.vhs_.xhr({ + uri: resolveUrl(this.mainPlaylistLoader_.srcUrl, utcTiming.value), + method: utcTiming.method, + withCredentials: this.withCredentials, + requestType: 'dash-clock-sync' + }, (error, req) => { + // disposed + if (!this.request) { + return; + } + if (error) { + // sync request failed, fall back to using date header from mpd + // TODO: log warning + this.mainPlaylistLoader_.clientOffset_ = this.mainLoaded_ - Date.now(); + return done(); + } + let serverTime; + if (utcTiming.method === 'HEAD') { + if (!req.responseHeaders || !req.responseHeaders.date) { + // expected date header not preset, fall back to using date header from mpd + // TODO: log warning + serverTime = this.mainLoaded_; + } else { + serverTime = Date.parse(req.responseHeaders.date); + } + } else { + serverTime = Date.parse(req.responseText); + } + this.mainPlaylistLoader_.clientOffset_ = serverTime - Date.now(); + done(); + }); + } + haveMain_() { + this.state = 'HAVE_MAIN_MANIFEST'; + if (this.isMain_) { + // We have the main playlist at this point, so + // trigger this to allow PlaylistController + // to make an initial playlist selection + this.trigger('loadedplaylist'); + } else if (!this.media_) { + // no media playlist was specifically selected so select + // the one the child playlist loader was created with + this.media(this.childPlaylist_); + } + } + handleMain_() { + // clear media request + this.mediaRequest_ = null; + const oldMain = this.mainPlaylistLoader_.main; + let newMain = parseMainXml({ + mainXml: this.mainPlaylistLoader_.mainXml_, + srcUrl: this.mainPlaylistLoader_.srcUrl, + clientOffset: this.mainPlaylistLoader_.clientOffset_, + sidxMapping: this.mainPlaylistLoader_.sidxMapping_, + previousManifest: oldMain + }); // if we have an old main to compare the new main against + + if (oldMain) { + newMain = updateMain(oldMain, newMain, this.mainPlaylistLoader_.sidxMapping_); + } // only update main if we have a new main + + this.mainPlaylistLoader_.main = newMain ? newMain : oldMain; + const location = this.mainPlaylistLoader_.main.locations && this.mainPlaylistLoader_.main.locations[0]; + if (location && location !== this.mainPlaylistLoader_.srcUrl) { + this.mainPlaylistLoader_.srcUrl = location; + } + if (!oldMain || newMain && newMain.minimumUpdatePeriod !== oldMain.minimumUpdatePeriod) { + this.updateMinimumUpdatePeriodTimeout_(); + } + this.addEventStreamToMetadataTrack_(newMain); + return Boolean(newMain); + } + updateMinimumUpdatePeriodTimeout_() { + const mpl = this.mainPlaylistLoader_; // cancel any pending creation of mup on media + // a new one will be added if needed. + + if (mpl.createMupOnMedia_) { + mpl.off('loadedmetadata', mpl.createMupOnMedia_); + mpl.createMupOnMedia_ = null; + } // clear any pending timeouts + + if (mpl.minimumUpdatePeriodTimeout_) { + window$1.clearTimeout(mpl.minimumUpdatePeriodTimeout_); + mpl.minimumUpdatePeriodTimeout_ = null; + } + let mup = mpl.main && mpl.main.minimumUpdatePeriod; // If the minimumUpdatePeriod has a value of 0, that indicates that the current + // MPD has no future validity, so a new one will need to be acquired when new + // media segments are to be made available. Thus, we use the target duration + // in this case + + if (mup === 0) { + if (mpl.media()) { + mup = mpl.media().targetDuration * 1000; + } else { + mpl.createMupOnMedia_ = mpl.updateMinimumUpdatePeriodTimeout_; + mpl.one('loadedmetadata', mpl.createMupOnMedia_); + } + } // if minimumUpdatePeriod is invalid or <= zero, which + // can happen when a live video becomes VOD. skip timeout + // creation. + + if (typeof mup !== 'number' || mup <= 0) { + if (mup < 0) { + this.logger_(`found invalid minimumUpdatePeriod of ${mup}, not setting a timeout`); + } + return; + } + this.createMUPTimeout_(mup); + } + createMUPTimeout_(mup) { + const mpl = this.mainPlaylistLoader_; + mpl.minimumUpdatePeriodTimeout_ = window$1.setTimeout(() => { + mpl.minimumUpdatePeriodTimeout_ = null; + mpl.trigger('minimumUpdatePeriod'); + mpl.createMUPTimeout_(mup); + }, mup); + } + /** + * Sends request to refresh the main xml and updates the parsed main manifest + */ + + refreshXml_() { + this.requestMain_((req, mainChanged) => { + if (!mainChanged) { + return; + } + if (this.media_) { + this.media_ = this.mainPlaylistLoader_.main.playlists[this.media_.id]; + } // This will filter out updated sidx info from the mapping + + this.mainPlaylistLoader_.sidxMapping_ = filterChangedSidxMappings(this.mainPlaylistLoader_.main, this.mainPlaylistLoader_.sidxMapping_); + this.addSidxSegments_(this.media(), this.state, sidxChanged => { + // TODO: do we need to reload the current playlist? + this.refreshMedia_(this.media().id); + }); + }); + } + /** + * Refreshes the media playlist by re-parsing the main xml and updating playlist + * references. If this is an alternate loader, the updated parsed manifest is retrieved + * from the main loader. + */ + + refreshMedia_(mediaID) { + if (!mediaID) { + throw new Error('refreshMedia_ must take a media id'); + } // for main we have to reparse the main xml + // to re-create segments based on current timing values + // which may change media. We only skip updating the main manifest + // if this is the first time this.media_ is being set. + // as main was just parsed in that case. + + if (this.media_ && this.isMain_) { + this.handleMain_(); + } + const playlists = this.mainPlaylistLoader_.main.playlists; + const mediaChanged = !this.media_ || this.media_ !== playlists[mediaID]; + if (mediaChanged) { + this.media_ = playlists[mediaID]; + } else { + this.trigger('playlistunchanged'); + } + if (!this.mediaUpdateTimeout) { + const createMediaUpdateTimeout = () => { + if (this.media().endList) { + return; + } + this.mediaUpdateTimeout = window$1.setTimeout(() => { + this.trigger('mediaupdatetimeout'); + createMediaUpdateTimeout(); + }, refreshDelay(this.media(), Boolean(mediaChanged))); + }; + createMediaUpdateTimeout(); + } + this.trigger('loadedplaylist'); + } + /** + * Takes eventstream data from a parsed DASH manifest and adds it to the metadata text track. + * + * @param {manifest} newMain the newly parsed manifest + */ + + addEventStreamToMetadataTrack_(newMain) { + // Only add new event stream metadata if we have a new manifest. + if (newMain && this.mainPlaylistLoader_.main.eventStream) { + // convert EventStream to ID3-like data. + const metadataArray = this.mainPlaylistLoader_.main.eventStream.map(eventStreamNode => { + return { + cueTime: eventStreamNode.start, + frames: [{ + data: eventStreamNode.messageData + }] + }; + }); + this.addMetadataToTextTrack('EventStream', metadataArray, this.mainPlaylistLoader_.main.duration); + } + } + /** + * Returns the key ID set from a playlist + * + * @param {playlist} playlist to fetch the key ID set from. + * @return a Set of 32 digit hex strings that represent the unique keyIds for that playlist. + */ + + getKeyIdSet(playlist) { + if (playlist.contentProtection) { + const keyIds = new Set(); + for (const keysystem in playlist.contentProtection) { + const defaultKID = playlist.contentProtection[keysystem].attributes['cenc:default_KID']; + if (defaultKID) { + // DASH keyIds are separated by dashes. + keyIds.add(defaultKID.replace(/-/g, '').toLowerCase()); + } + } + return keyIds; + } + } +} +var Config = { + GOAL_BUFFER_LENGTH: 30, + MAX_GOAL_BUFFER_LENGTH: 60, + BACK_BUFFER_LENGTH: 30, + GOAL_BUFFER_LENGTH_RATE: 1, + // 0.5 MB/s + INITIAL_BANDWIDTH: 4194304, + // A fudge factor to apply to advertised playlist bitrates to account for + // temporary flucations in client bandwidth + BANDWIDTH_VARIANCE: 1.2, + // How much of the buffer must be filled before we consider upswitching + BUFFER_LOW_WATER_LINE: 0, + MAX_BUFFER_LOW_WATER_LINE: 30, + // TODO: Remove this when experimentalBufferBasedABR is removed + EXPERIMENTAL_MAX_BUFFER_LOW_WATER_LINE: 16, + BUFFER_LOW_WATER_LINE_RATE: 1, + // If the buffer is greater than the high water line, we won't switch down + BUFFER_HIGH_WATER_LINE: 30 +}; +const stringToArrayBuffer = string => { + const view = new Uint8Array(new ArrayBuffer(string.length)); + for (let i = 0; i < string.length; i++) { + view[i] = string.charCodeAt(i); + } + return view.buffer; +}; + +/* global Blob, BlobBuilder, Worker */ +// unify worker interface +const browserWorkerPolyFill = function (workerObj) { + // node only supports on/off + workerObj.on = workerObj.addEventListener; + workerObj.off = workerObj.removeEventListener; + return workerObj; +}; +const createObjectURL = function (str) { + try { + return URL.createObjectURL(new Blob([str], { + type: 'application/javascript' + })); + } catch (e) { + const blob = new BlobBuilder(); + blob.append(str); + return URL.createObjectURL(blob.getBlob()); + } +}; +const factory = function (code) { + return function () { + const objectUrl = createObjectURL(code); + const worker = browserWorkerPolyFill(new Worker(objectUrl)); + worker.objURL = objectUrl; + const terminate = worker.terminate; + worker.on = worker.addEventListener; + worker.off = worker.removeEventListener; + worker.terminate = function () { + URL.revokeObjectURL(objectUrl); + return terminate.call(this); + }; + return worker; + }; +}; +const transform = function (code) { + return `var browserWorkerPolyFill = ${browserWorkerPolyFill.toString()};\n` + 'browserWorkerPolyFill(self);\n' + code; +}; +const getWorkerString = function (fn) { + return fn.toString().replace(/^function.+?{/, '').slice(0, -1); +}; + +/* rollup-plugin-worker-factory start for worker!/home/runner/work/http-streaming/http-streaming/src/transmuxer-worker.js */ +const workerCode$1 = transform(getWorkerString(function () { + var commonjsGlobal = typeof globalThis !== 'undefined' ? globalThis : typeof window !== 'undefined' ? window : typeof global !== 'undefined' ? global : typeof self !== 'undefined' ? self : {}; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * A lightweight readable stream implemention that handles event dispatching. + * Objects that inherit from streams should call init in their constructors. + */ + + var Stream$8 = function () { + this.init = function () { + var listeners = {}; + /** + * Add a listener for a specified event type. + * @param type {string} the event name + * @param listener {function} the callback to be invoked when an event of + * the specified type occurs + */ + + this.on = function (type, listener) { + if (!listeners[type]) { + listeners[type] = []; + } + listeners[type] = listeners[type].concat(listener); + }; + /** + * Remove a listener for a specified event type. + * @param type {string} the event name + * @param listener {function} a function previously registered for this + * type of event through `on` + */ + + this.off = function (type, listener) { + var index; + if (!listeners[type]) { + return false; + } + index = listeners[type].indexOf(listener); + listeners[type] = listeners[type].slice(); + listeners[type].splice(index, 1); + return index > -1; + }; + /** + * Trigger an event of the specified type on this stream. Any additional + * arguments to this function are passed as parameters to event listeners. + * @param type {string} the event name + */ + + this.trigger = function (type) { + var callbacks, i, length, args; + callbacks = listeners[type]; + if (!callbacks) { + return; + } // Slicing the arguments on every invocation of this method + // can add a significant amount of overhead. Avoid the + // intermediate object creation for the common case of a + // single callback argument + + if (arguments.length === 2) { + length = callbacks.length; + for (i = 0; i < length; ++i) { + callbacks[i].call(this, arguments[1]); + } + } else { + args = []; + i = arguments.length; + for (i = 1; i < arguments.length; ++i) { + args.push(arguments[i]); + } + length = callbacks.length; + for (i = 0; i < length; ++i) { + callbacks[i].apply(this, args); + } + } + }; + /** + * Destroys the stream and cleans up. + */ + + this.dispose = function () { + listeners = {}; + }; + }; + }; + /** + * Forwards all `data` events on this stream to the destination stream. The + * destination stream should provide a method `push` to receive the data + * events as they arrive. + * @param destination {stream} the stream that will receive all `data` events + * @param autoFlush {boolean} if false, we will not call `flush` on the destination + * when the current stream emits a 'done' event + * @see http://nodejs.org/api/stream.html#stream_readable_pipe_destination_options + */ + + Stream$8.prototype.pipe = function (destination) { + this.on('data', function (data) { + destination.push(data); + }); + this.on('done', function (flushSource) { + destination.flush(flushSource); + }); + this.on('partialdone', function (flushSource) { + destination.partialFlush(flushSource); + }); + this.on('endedtimeline', function (flushSource) { + destination.endTimeline(flushSource); + }); + this.on('reset', function (flushSource) { + destination.reset(flushSource); + }); + return destination; + }; // Default stream functions that are expected to be overridden to perform + // actual work. These are provided by the prototype as a sort of no-op + // implementation so that we don't have to check for their existence in the + // `pipe` function above. + + Stream$8.prototype.push = function (data) { + this.trigger('data', data); + }; + Stream$8.prototype.flush = function (flushSource) { + this.trigger('done', flushSource); + }; + Stream$8.prototype.partialFlush = function (flushSource) { + this.trigger('partialdone', flushSource); + }; + Stream$8.prototype.endTimeline = function (flushSource) { + this.trigger('endedtimeline', flushSource); + }; + Stream$8.prototype.reset = function (flushSource) { + this.trigger('reset', flushSource); + }; + var stream = Stream$8; + var MAX_UINT32$1 = Math.pow(2, 32); + var getUint64$3 = function (uint8) { + var dv = new DataView(uint8.buffer, uint8.byteOffset, uint8.byteLength); + var value; + if (dv.getBigUint64) { + value = dv.getBigUint64(0); + if (value < Number.MAX_SAFE_INTEGER) { + return Number(value); + } + return value; + } + return dv.getUint32(0) * MAX_UINT32$1 + dv.getUint32(4); + }; + var numbers = { + getUint64: getUint64$3, + MAX_UINT32: MAX_UINT32$1 + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * Functions that generate fragmented MP4s suitable for use with Media + * Source Extensions. + */ + + var MAX_UINT32 = numbers.MAX_UINT32; + var box, dinf, esds, ftyp, mdat, mfhd, minf, moof, moov, mvex, mvhd, trak, tkhd, mdia, mdhd, hdlr, sdtp, stbl, stsd, traf, trex, trun$1, types, MAJOR_BRAND, MINOR_VERSION, AVC1_BRAND, VIDEO_HDLR, AUDIO_HDLR, HDLR_TYPES, VMHD, SMHD, DREF, STCO, STSC, STSZ, STTS; // pre-calculate constants + + (function () { + var i; + types = { + avc1: [], + // codingname + avcC: [], + btrt: [], + dinf: [], + dref: [], + esds: [], + ftyp: [], + hdlr: [], + mdat: [], + mdhd: [], + mdia: [], + mfhd: [], + minf: [], + moof: [], + moov: [], + mp4a: [], + // codingname + mvex: [], + mvhd: [], + pasp: [], + sdtp: [], + smhd: [], + stbl: [], + stco: [], + stsc: [], + stsd: [], + stsz: [], + stts: [], + styp: [], + tfdt: [], + tfhd: [], + traf: [], + trak: [], + trun: [], + trex: [], + tkhd: [], + vmhd: [] + }; // In environments where Uint8Array is undefined (e.g., IE8), skip set up so that we + // don't throw an error + + if (typeof Uint8Array === 'undefined') { + return; + } + for (i in types) { + if (types.hasOwnProperty(i)) { + types[i] = [i.charCodeAt(0), i.charCodeAt(1), i.charCodeAt(2), i.charCodeAt(3)]; + } + } + MAJOR_BRAND = new Uint8Array(['i'.charCodeAt(0), 's'.charCodeAt(0), 'o'.charCodeAt(0), 'm'.charCodeAt(0)]); + AVC1_BRAND = new Uint8Array(['a'.charCodeAt(0), 'v'.charCodeAt(0), 'c'.charCodeAt(0), '1'.charCodeAt(0)]); + MINOR_VERSION = new Uint8Array([0, 0, 0, 1]); + VIDEO_HDLR = new Uint8Array([0x00, + // version 0 + 0x00, 0x00, 0x00, + // flags + 0x00, 0x00, 0x00, 0x00, + // pre_defined + 0x76, 0x69, 0x64, 0x65, + // handler_type: 'vide' + 0x00, 0x00, 0x00, 0x00, + // reserved + 0x00, 0x00, 0x00, 0x00, + // reserved + 0x00, 0x00, 0x00, 0x00, + // reserved + 0x56, 0x69, 0x64, 0x65, 0x6f, 0x48, 0x61, 0x6e, 0x64, 0x6c, 0x65, 0x72, 0x00 // name: 'VideoHandler' + ]); + + AUDIO_HDLR = new Uint8Array([0x00, + // version 0 + 0x00, 0x00, 0x00, + // flags + 0x00, 0x00, 0x00, 0x00, + // pre_defined + 0x73, 0x6f, 0x75, 0x6e, + // handler_type: 'soun' + 0x00, 0x00, 0x00, 0x00, + // reserved + 0x00, 0x00, 0x00, 0x00, + // reserved + 0x00, 0x00, 0x00, 0x00, + // reserved + 0x53, 0x6f, 0x75, 0x6e, 0x64, 0x48, 0x61, 0x6e, 0x64, 0x6c, 0x65, 0x72, 0x00 // name: 'SoundHandler' + ]); + + HDLR_TYPES = { + video: VIDEO_HDLR, + audio: AUDIO_HDLR + }; + DREF = new Uint8Array([0x00, + // version 0 + 0x00, 0x00, 0x00, + // flags + 0x00, 0x00, 0x00, 0x01, + // entry_count + 0x00, 0x00, 0x00, 0x0c, + // entry_size + 0x75, 0x72, 0x6c, 0x20, + // 'url' type + 0x00, + // version 0 + 0x00, 0x00, 0x01 // entry_flags + ]); + + SMHD = new Uint8Array([0x00, + // version + 0x00, 0x00, 0x00, + // flags + 0x00, 0x00, + // balance, 0 means centered + 0x00, 0x00 // reserved + ]); + + STCO = new Uint8Array([0x00, + // version + 0x00, 0x00, 0x00, + // flags + 0x00, 0x00, 0x00, 0x00 // entry_count + ]); + + STSC = STCO; + STSZ = new Uint8Array([0x00, + // version + 0x00, 0x00, 0x00, + // flags + 0x00, 0x00, 0x00, 0x00, + // sample_size + 0x00, 0x00, 0x00, 0x00 // sample_count + ]); + + STTS = STCO; + VMHD = new Uint8Array([0x00, + // version + 0x00, 0x00, 0x01, + // flags + 0x00, 0x00, + // graphicsmode + 0x00, 0x00, 0x00, 0x00, 0x00, 0x00 // opcolor + ]); + })(); + + box = function (type) { + var payload = [], + size = 0, + i, + result, + view; + for (i = 1; i < arguments.length; i++) { + payload.push(arguments[i]); + } + i = payload.length; // calculate the total size we need to allocate + + while (i--) { + size += payload[i].byteLength; + } + result = new Uint8Array(size + 8); + view = new DataView(result.buffer, result.byteOffset, result.byteLength); + view.setUint32(0, result.byteLength); + result.set(type, 4); // copy the payload into the result + + for (i = 0, size = 8; i < payload.length; i++) { + result.set(payload[i], size); + size += payload[i].byteLength; + } + return result; + }; + dinf = function () { + return box(types.dinf, box(types.dref, DREF)); + }; + esds = function (track) { + return box(types.esds, new Uint8Array([0x00, + // version + 0x00, 0x00, 0x00, + // flags + // ES_Descriptor + 0x03, + // tag, ES_DescrTag + 0x19, + // length + 0x00, 0x00, + // ES_ID + 0x00, + // streamDependenceFlag, URL_flag, reserved, streamPriority + // DecoderConfigDescriptor + 0x04, + // tag, DecoderConfigDescrTag + 0x11, + // length + 0x40, + // object type + 0x15, + // streamType + 0x00, 0x06, 0x00, + // bufferSizeDB + 0x00, 0x00, 0xda, 0xc0, + // maxBitrate + 0x00, 0x00, 0xda, 0xc0, + // avgBitrate + // DecoderSpecificInfo + 0x05, + // tag, DecoderSpecificInfoTag + 0x02, + // length + // ISO/IEC 14496-3, AudioSpecificConfig + // for samplingFrequencyIndex see ISO/IEC 13818-7:2006, 8.1.3.2.2, Table 35 + track.audioobjecttype << 3 | track.samplingfrequencyindex >>> 1, track.samplingfrequencyindex << 7 | track.channelcount << 3, 0x06, 0x01, 0x02 // GASpecificConfig + ])); + }; + + ftyp = function () { + return box(types.ftyp, MAJOR_BRAND, MINOR_VERSION, MAJOR_BRAND, AVC1_BRAND); + }; + hdlr = function (type) { + return box(types.hdlr, HDLR_TYPES[type]); + }; + mdat = function (data) { + return box(types.mdat, data); + }; + mdhd = function (track) { + var result = new Uint8Array([0x00, + // version 0 + 0x00, 0x00, 0x00, + // flags + 0x00, 0x00, 0x00, 0x02, + // creation_time + 0x00, 0x00, 0x00, 0x03, + // modification_time + 0x00, 0x01, 0x5f, 0x90, + // timescale, 90,000 "ticks" per second + track.duration >>> 24 & 0xFF, track.duration >>> 16 & 0xFF, track.duration >>> 8 & 0xFF, track.duration & 0xFF, + // duration + 0x55, 0xc4, + // 'und' language (undetermined) + 0x00, 0x00]); // Use the sample rate from the track metadata, when it is + // defined. The sample rate can be parsed out of an ADTS header, for + // instance. + + if (track.samplerate) { + result[12] = track.samplerate >>> 24 & 0xFF; + result[13] = track.samplerate >>> 16 & 0xFF; + result[14] = track.samplerate >>> 8 & 0xFF; + result[15] = track.samplerate & 0xFF; + } + return box(types.mdhd, result); + }; + mdia = function (track) { + return box(types.mdia, mdhd(track), hdlr(track.type), minf(track)); + }; + mfhd = function (sequenceNumber) { + return box(types.mfhd, new Uint8Array([0x00, 0x00, 0x00, 0x00, + // flags + (sequenceNumber & 0xFF000000) >> 24, (sequenceNumber & 0xFF0000) >> 16, (sequenceNumber & 0xFF00) >> 8, sequenceNumber & 0xFF // sequence_number + ])); + }; + + minf = function (track) { + return box(types.minf, track.type === 'video' ? box(types.vmhd, VMHD) : box(types.smhd, SMHD), dinf(), stbl(track)); + }; + moof = function (sequenceNumber, tracks) { + var trackFragments = [], + i = tracks.length; // build traf boxes for each track fragment + + while (i--) { + trackFragments[i] = traf(tracks[i]); + } + return box.apply(null, [types.moof, mfhd(sequenceNumber)].concat(trackFragments)); + }; + /** + * Returns a movie box. + * @param tracks {array} the tracks associated with this movie + * @see ISO/IEC 14496-12:2012(E), section 8.2.1 + */ + + moov = function (tracks) { + var i = tracks.length, + boxes = []; + while (i--) { + boxes[i] = trak(tracks[i]); + } + return box.apply(null, [types.moov, mvhd(0xffffffff)].concat(boxes).concat(mvex(tracks))); + }; + mvex = function (tracks) { + var i = tracks.length, + boxes = []; + while (i--) { + boxes[i] = trex(tracks[i]); + } + return box.apply(null, [types.mvex].concat(boxes)); + }; + mvhd = function (duration) { + var bytes = new Uint8Array([0x00, + // version 0 + 0x00, 0x00, 0x00, + // flags + 0x00, 0x00, 0x00, 0x01, + // creation_time + 0x00, 0x00, 0x00, 0x02, + // modification_time + 0x00, 0x01, 0x5f, 0x90, + // timescale, 90,000 "ticks" per second + (duration & 0xFF000000) >> 24, (duration & 0xFF0000) >> 16, (duration & 0xFF00) >> 8, duration & 0xFF, + // duration + 0x00, 0x01, 0x00, 0x00, + // 1.0 rate + 0x01, 0x00, + // 1.0 volume + 0x00, 0x00, + // reserved + 0x00, 0x00, 0x00, 0x00, + // reserved + 0x00, 0x00, 0x00, 0x00, + // reserved + 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x40, 0x00, 0x00, 0x00, + // transformation: unity matrix + 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, + // pre_defined + 0xff, 0xff, 0xff, 0xff // next_track_ID + ]); + + return box(types.mvhd, bytes); + }; + sdtp = function (track) { + var samples = track.samples || [], + bytes = new Uint8Array(4 + samples.length), + flags, + i; // leave the full box header (4 bytes) all zero + // write the sample table + + for (i = 0; i < samples.length; i++) { + flags = samples[i].flags; + bytes[i + 4] = flags.dependsOn << 4 | flags.isDependedOn << 2 | flags.hasRedundancy; + } + return box(types.sdtp, bytes); + }; + stbl = function (track) { + return box(types.stbl, stsd(track), box(types.stts, STTS), box(types.stsc, STSC), box(types.stsz, STSZ), box(types.stco, STCO)); + }; + (function () { + var videoSample, audioSample; + stsd = function (track) { + return box(types.stsd, new Uint8Array([0x00, + // version 0 + 0x00, 0x00, 0x00, + // flags + 0x00, 0x00, 0x00, 0x01]), track.type === 'video' ? videoSample(track) : audioSample(track)); + }; + videoSample = function (track) { + var sps = track.sps || [], + pps = track.pps || [], + sequenceParameterSets = [], + pictureParameterSets = [], + i, + avc1Box; // assemble the SPSs + + for (i = 0; i < sps.length; i++) { + sequenceParameterSets.push((sps[i].byteLength & 0xFF00) >>> 8); + sequenceParameterSets.push(sps[i].byteLength & 0xFF); // sequenceParameterSetLength + + sequenceParameterSets = sequenceParameterSets.concat(Array.prototype.slice.call(sps[i])); // SPS + } // assemble the PPSs + + for (i = 0; i < pps.length; i++) { + pictureParameterSets.push((pps[i].byteLength & 0xFF00) >>> 8); + pictureParameterSets.push(pps[i].byteLength & 0xFF); + pictureParameterSets = pictureParameterSets.concat(Array.prototype.slice.call(pps[i])); + } + avc1Box = [types.avc1, new Uint8Array([0x00, 0x00, 0x00, 0x00, 0x00, 0x00, + // reserved + 0x00, 0x01, + // data_reference_index + 0x00, 0x00, + // pre_defined + 0x00, 0x00, + // reserved + 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, + // pre_defined + (track.width & 0xff00) >> 8, track.width & 0xff, + // width + (track.height & 0xff00) >> 8, track.height & 0xff, + // height + 0x00, 0x48, 0x00, 0x00, + // horizresolution + 0x00, 0x48, 0x00, 0x00, + // vertresolution + 0x00, 0x00, 0x00, 0x00, + // reserved + 0x00, 0x01, + // frame_count + 0x13, 0x76, 0x69, 0x64, 0x65, 0x6f, 0x6a, 0x73, 0x2d, 0x63, 0x6f, 0x6e, 0x74, 0x72, 0x69, 0x62, 0x2d, 0x68, 0x6c, 0x73, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, + // compressorname + 0x00, 0x18, + // depth = 24 + 0x11, 0x11 // pre_defined = -1 + ]), box(types.avcC, new Uint8Array([0x01, + // configurationVersion + track.profileIdc, + // AVCProfileIndication + track.profileCompatibility, + // profile_compatibility + track.levelIdc, + // AVCLevelIndication + 0xff // lengthSizeMinusOne, hard-coded to 4 bytes + ].concat([sps.length], + // numOfSequenceParameterSets + sequenceParameterSets, + // "SPS" + [pps.length], + // numOfPictureParameterSets + pictureParameterSets // "PPS" + ))), box(types.btrt, new Uint8Array([0x00, 0x1c, 0x9c, 0x80, + // bufferSizeDB + 0x00, 0x2d, 0xc6, 0xc0, + // maxBitrate + 0x00, 0x2d, 0xc6, 0xc0 // avgBitrate + ]))]; + + if (track.sarRatio) { + var hSpacing = track.sarRatio[0], + vSpacing = track.sarRatio[1]; + avc1Box.push(box(types.pasp, new Uint8Array([(hSpacing & 0xFF000000) >> 24, (hSpacing & 0xFF0000) >> 16, (hSpacing & 0xFF00) >> 8, hSpacing & 0xFF, (vSpacing & 0xFF000000) >> 24, (vSpacing & 0xFF0000) >> 16, (vSpacing & 0xFF00) >> 8, vSpacing & 0xFF]))); + } + return box.apply(null, avc1Box); + }; + audioSample = function (track) { + return box(types.mp4a, new Uint8Array([ + // SampleEntry, ISO/IEC 14496-12 + 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, + // reserved + 0x00, 0x01, + // data_reference_index + // AudioSampleEntry, ISO/IEC 14496-12 + 0x00, 0x00, 0x00, 0x00, + // reserved + 0x00, 0x00, 0x00, 0x00, + // reserved + (track.channelcount & 0xff00) >> 8, track.channelcount & 0xff, + // channelcount + (track.samplesize & 0xff00) >> 8, track.samplesize & 0xff, + // samplesize + 0x00, 0x00, + // pre_defined + 0x00, 0x00, + // reserved + (track.samplerate & 0xff00) >> 8, track.samplerate & 0xff, 0x00, 0x00 // samplerate, 16.16 + // MP4AudioSampleEntry, ISO/IEC 14496-14 + ]), esds(track)); + }; + })(); + tkhd = function (track) { + var result = new Uint8Array([0x00, + // version 0 + 0x00, 0x00, 0x07, + // flags + 0x00, 0x00, 0x00, 0x00, + // creation_time + 0x00, 0x00, 0x00, 0x00, + // modification_time + (track.id & 0xFF000000) >> 24, (track.id & 0xFF0000) >> 16, (track.id & 0xFF00) >> 8, track.id & 0xFF, + // track_ID + 0x00, 0x00, 0x00, 0x00, + // reserved + (track.duration & 0xFF000000) >> 24, (track.duration & 0xFF0000) >> 16, (track.duration & 0xFF00) >> 8, track.duration & 0xFF, + // duration + 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, + // reserved + 0x00, 0x00, + // layer + 0x00, 0x00, + // alternate_group + 0x01, 0x00, + // non-audio track volume + 0x00, 0x00, + // reserved + 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x40, 0x00, 0x00, 0x00, + // transformation: unity matrix + (track.width & 0xFF00) >> 8, track.width & 0xFF, 0x00, 0x00, + // width + (track.height & 0xFF00) >> 8, track.height & 0xFF, 0x00, 0x00 // height + ]); + + return box(types.tkhd, result); + }; + /** + * Generate a track fragment (traf) box. A traf box collects metadata + * about tracks in a movie fragment (moof) box. + */ + + traf = function (track) { + var trackFragmentHeader, trackFragmentDecodeTime, trackFragmentRun, sampleDependencyTable, dataOffset, upperWordBaseMediaDecodeTime, lowerWordBaseMediaDecodeTime; + trackFragmentHeader = box(types.tfhd, new Uint8Array([0x00, + // version 0 + 0x00, 0x00, 0x3a, + // flags + (track.id & 0xFF000000) >> 24, (track.id & 0xFF0000) >> 16, (track.id & 0xFF00) >> 8, track.id & 0xFF, + // track_ID + 0x00, 0x00, 0x00, 0x01, + // sample_description_index + 0x00, 0x00, 0x00, 0x00, + // default_sample_duration + 0x00, 0x00, 0x00, 0x00, + // default_sample_size + 0x00, 0x00, 0x00, 0x00 // default_sample_flags + ])); + + upperWordBaseMediaDecodeTime = Math.floor(track.baseMediaDecodeTime / MAX_UINT32); + lowerWordBaseMediaDecodeTime = Math.floor(track.baseMediaDecodeTime % MAX_UINT32); + trackFragmentDecodeTime = box(types.tfdt, new Uint8Array([0x01, + // version 1 + 0x00, 0x00, 0x00, + // flags + // baseMediaDecodeTime + upperWordBaseMediaDecodeTime >>> 24 & 0xFF, upperWordBaseMediaDecodeTime >>> 16 & 0xFF, upperWordBaseMediaDecodeTime >>> 8 & 0xFF, upperWordBaseMediaDecodeTime & 0xFF, lowerWordBaseMediaDecodeTime >>> 24 & 0xFF, lowerWordBaseMediaDecodeTime >>> 16 & 0xFF, lowerWordBaseMediaDecodeTime >>> 8 & 0xFF, lowerWordBaseMediaDecodeTime & 0xFF])); // the data offset specifies the number of bytes from the start of + // the containing moof to the first payload byte of the associated + // mdat + + dataOffset = 32 + + // tfhd + 20 + + // tfdt + 8 + + // traf header + 16 + + // mfhd + 8 + + // moof header + 8; // mdat header + // audio tracks require less metadata + + if (track.type === 'audio') { + trackFragmentRun = trun$1(track, dataOffset); + return box(types.traf, trackFragmentHeader, trackFragmentDecodeTime, trackFragmentRun); + } // video tracks should contain an independent and disposable samples + // box (sdtp) + // generate one and adjust offsets to match + + sampleDependencyTable = sdtp(track); + trackFragmentRun = trun$1(track, sampleDependencyTable.length + dataOffset); + return box(types.traf, trackFragmentHeader, trackFragmentDecodeTime, trackFragmentRun, sampleDependencyTable); + }; + /** + * Generate a track box. + * @param track {object} a track definition + * @return {Uint8Array} the track box + */ + + trak = function (track) { + track.duration = track.duration || 0xffffffff; + return box(types.trak, tkhd(track), mdia(track)); + }; + trex = function (track) { + var result = new Uint8Array([0x00, + // version 0 + 0x00, 0x00, 0x00, + // flags + (track.id & 0xFF000000) >> 24, (track.id & 0xFF0000) >> 16, (track.id & 0xFF00) >> 8, track.id & 0xFF, + // track_ID + 0x00, 0x00, 0x00, 0x01, + // default_sample_description_index + 0x00, 0x00, 0x00, 0x00, + // default_sample_duration + 0x00, 0x00, 0x00, 0x00, + // default_sample_size + 0x00, 0x01, 0x00, 0x01 // default_sample_flags + ]); // the last two bytes of default_sample_flags is the sample + // degradation priority, a hint about the importance of this sample + // relative to others. Lower the degradation priority for all sample + // types other than video. + + if (track.type !== 'video') { + result[result.length - 1] = 0x00; + } + return box(types.trex, result); + }; + (function () { + var audioTrun, videoTrun, trunHeader; // This method assumes all samples are uniform. That is, if a + // duration is present for the first sample, it will be present for + // all subsequent samples. + // see ISO/IEC 14496-12:2012, Section 8.8.8.1 + + trunHeader = function (samples, offset) { + var durationPresent = 0, + sizePresent = 0, + flagsPresent = 0, + compositionTimeOffset = 0; // trun flag constants + + if (samples.length) { + if (samples[0].duration !== undefined) { + durationPresent = 0x1; + } + if (samples[0].size !== undefined) { + sizePresent = 0x2; + } + if (samples[0].flags !== undefined) { + flagsPresent = 0x4; + } + if (samples[0].compositionTimeOffset !== undefined) { + compositionTimeOffset = 0x8; + } + } + return [0x00, + // version 0 + 0x00, durationPresent | sizePresent | flagsPresent | compositionTimeOffset, 0x01, + // flags + (samples.length & 0xFF000000) >>> 24, (samples.length & 0xFF0000) >>> 16, (samples.length & 0xFF00) >>> 8, samples.length & 0xFF, + // sample_count + (offset & 0xFF000000) >>> 24, (offset & 0xFF0000) >>> 16, (offset & 0xFF00) >>> 8, offset & 0xFF // data_offset + ]; + }; + + videoTrun = function (track, offset) { + var bytesOffest, bytes, header, samples, sample, i; + samples = track.samples || []; + offset += 8 + 12 + 16 * samples.length; + header = trunHeader(samples, offset); + bytes = new Uint8Array(header.length + samples.length * 16); + bytes.set(header); + bytesOffest = header.length; + for (i = 0; i < samples.length; i++) { + sample = samples[i]; + bytes[bytesOffest++] = (sample.duration & 0xFF000000) >>> 24; + bytes[bytesOffest++] = (sample.duration & 0xFF0000) >>> 16; + bytes[bytesOffest++] = (sample.duration & 0xFF00) >>> 8; + bytes[bytesOffest++] = sample.duration & 0xFF; // sample_duration + + bytes[bytesOffest++] = (sample.size & 0xFF000000) >>> 24; + bytes[bytesOffest++] = (sample.size & 0xFF0000) >>> 16; + bytes[bytesOffest++] = (sample.size & 0xFF00) >>> 8; + bytes[bytesOffest++] = sample.size & 0xFF; // sample_size + + bytes[bytesOffest++] = sample.flags.isLeading << 2 | sample.flags.dependsOn; + bytes[bytesOffest++] = sample.flags.isDependedOn << 6 | sample.flags.hasRedundancy << 4 | sample.flags.paddingValue << 1 | sample.flags.isNonSyncSample; + bytes[bytesOffest++] = sample.flags.degradationPriority & 0xF0 << 8; + bytes[bytesOffest++] = sample.flags.degradationPriority & 0x0F; // sample_flags + + bytes[bytesOffest++] = (sample.compositionTimeOffset & 0xFF000000) >>> 24; + bytes[bytesOffest++] = (sample.compositionTimeOffset & 0xFF0000) >>> 16; + bytes[bytesOffest++] = (sample.compositionTimeOffset & 0xFF00) >>> 8; + bytes[bytesOffest++] = sample.compositionTimeOffset & 0xFF; // sample_composition_time_offset + } + + return box(types.trun, bytes); + }; + audioTrun = function (track, offset) { + var bytes, bytesOffest, header, samples, sample, i; + samples = track.samples || []; + offset += 8 + 12 + 8 * samples.length; + header = trunHeader(samples, offset); + bytes = new Uint8Array(header.length + samples.length * 8); + bytes.set(header); + bytesOffest = header.length; + for (i = 0; i < samples.length; i++) { + sample = samples[i]; + bytes[bytesOffest++] = (sample.duration & 0xFF000000) >>> 24; + bytes[bytesOffest++] = (sample.duration & 0xFF0000) >>> 16; + bytes[bytesOffest++] = (sample.duration & 0xFF00) >>> 8; + bytes[bytesOffest++] = sample.duration & 0xFF; // sample_duration + + bytes[bytesOffest++] = (sample.size & 0xFF000000) >>> 24; + bytes[bytesOffest++] = (sample.size & 0xFF0000) >>> 16; + bytes[bytesOffest++] = (sample.size & 0xFF00) >>> 8; + bytes[bytesOffest++] = sample.size & 0xFF; // sample_size + } + + return box(types.trun, bytes); + }; + trun$1 = function (track, offset) { + if (track.type === 'audio') { + return audioTrun(track, offset); + } + return videoTrun(track, offset); + }; + })(); + var mp4Generator = { + ftyp: ftyp, + mdat: mdat, + moof: moof, + moov: moov, + initSegment: function (tracks) { + var fileType = ftyp(), + movie = moov(tracks), + result; + result = new Uint8Array(fileType.byteLength + movie.byteLength); + result.set(fileType); + result.set(movie, fileType.byteLength); + return result; + } + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + */ + // composed of the nal units that make up that frame + // Also keep track of cummulative data about the frame from the nal units such + // as the frame duration, starting pts, etc. + + var groupNalsIntoFrames = function (nalUnits) { + var i, + currentNal, + currentFrame = [], + frames = []; // TODO added for LHLS, make sure this is OK + + frames.byteLength = 0; + frames.nalCount = 0; + frames.duration = 0; + currentFrame.byteLength = 0; + for (i = 0; i < nalUnits.length; i++) { + currentNal = nalUnits[i]; // Split on 'aud'-type nal units + + if (currentNal.nalUnitType === 'access_unit_delimiter_rbsp') { + // Since the very first nal unit is expected to be an AUD + // only push to the frames array when currentFrame is not empty + if (currentFrame.length) { + currentFrame.duration = currentNal.dts - currentFrame.dts; // TODO added for LHLS, make sure this is OK + + frames.byteLength += currentFrame.byteLength; + frames.nalCount += currentFrame.length; + frames.duration += currentFrame.duration; + frames.push(currentFrame); + } + currentFrame = [currentNal]; + currentFrame.byteLength = currentNal.data.byteLength; + currentFrame.pts = currentNal.pts; + currentFrame.dts = currentNal.dts; + } else { + // Specifically flag key frames for ease of use later + if (currentNal.nalUnitType === 'slice_layer_without_partitioning_rbsp_idr') { + currentFrame.keyFrame = true; + } + currentFrame.duration = currentNal.dts - currentFrame.dts; + currentFrame.byteLength += currentNal.data.byteLength; + currentFrame.push(currentNal); + } + } // For the last frame, use the duration of the previous frame if we + // have nothing better to go on + + if (frames.length && (!currentFrame.duration || currentFrame.duration <= 0)) { + currentFrame.duration = frames[frames.length - 1].duration; + } // Push the final frame + // TODO added for LHLS, make sure this is OK + + frames.byteLength += currentFrame.byteLength; + frames.nalCount += currentFrame.length; + frames.duration += currentFrame.duration; + frames.push(currentFrame); + return frames; + }; // Convert an array of frames into an array of Gop with each Gop being composed + // of the frames that make up that Gop + // Also keep track of cummulative data about the Gop from the frames such as the + // Gop duration, starting pts, etc. + + var groupFramesIntoGops = function (frames) { + var i, + currentFrame, + currentGop = [], + gops = []; // We must pre-set some of the values on the Gop since we + // keep running totals of these values + + currentGop.byteLength = 0; + currentGop.nalCount = 0; + currentGop.duration = 0; + currentGop.pts = frames[0].pts; + currentGop.dts = frames[0].dts; // store some metadata about all the Gops + + gops.byteLength = 0; + gops.nalCount = 0; + gops.duration = 0; + gops.pts = frames[0].pts; + gops.dts = frames[0].dts; + for (i = 0; i < frames.length; i++) { + currentFrame = frames[i]; + if (currentFrame.keyFrame) { + // Since the very first frame is expected to be an keyframe + // only push to the gops array when currentGop is not empty + if (currentGop.length) { + gops.push(currentGop); + gops.byteLength += currentGop.byteLength; + gops.nalCount += currentGop.nalCount; + gops.duration += currentGop.duration; + } + currentGop = [currentFrame]; + currentGop.nalCount = currentFrame.length; + currentGop.byteLength = currentFrame.byteLength; + currentGop.pts = currentFrame.pts; + currentGop.dts = currentFrame.dts; + currentGop.duration = currentFrame.duration; + } else { + currentGop.duration += currentFrame.duration; + currentGop.nalCount += currentFrame.length; + currentGop.byteLength += currentFrame.byteLength; + currentGop.push(currentFrame); + } + } + if (gops.length && currentGop.duration <= 0) { + currentGop.duration = gops[gops.length - 1].duration; + } + gops.byteLength += currentGop.byteLength; + gops.nalCount += currentGop.nalCount; + gops.duration += currentGop.duration; // push the final Gop + + gops.push(currentGop); + return gops; + }; + /* + * Search for the first keyframe in the GOPs and throw away all frames + * until that keyframe. Then extend the duration of the pulled keyframe + * and pull the PTS and DTS of the keyframe so that it covers the time + * range of the frames that were disposed. + * + * @param {Array} gops video GOPs + * @returns {Array} modified video GOPs + */ + + var extendFirstKeyFrame = function (gops) { + var currentGop; + if (!gops[0][0].keyFrame && gops.length > 1) { + // Remove the first GOP + currentGop = gops.shift(); + gops.byteLength -= currentGop.byteLength; + gops.nalCount -= currentGop.nalCount; // Extend the first frame of what is now the + // first gop to cover the time period of the + // frames we just removed + + gops[0][0].dts = currentGop.dts; + gops[0][0].pts = currentGop.pts; + gops[0][0].duration += currentGop.duration; + } + return gops; + }; + /** + * Default sample object + * see ISO/IEC 14496-12:2012, section 8.6.4.3 + */ + + var createDefaultSample = function () { + return { + size: 0, + flags: { + isLeading: 0, + dependsOn: 1, + isDependedOn: 0, + hasRedundancy: 0, + degradationPriority: 0, + isNonSyncSample: 1 + } + }; + }; + /* + * Collates information from a video frame into an object for eventual + * entry into an MP4 sample table. + * + * @param {Object} frame the video frame + * @param {Number} dataOffset the byte offset to position the sample + * @return {Object} object containing sample table info for a frame + */ + + var sampleForFrame = function (frame, dataOffset) { + var sample = createDefaultSample(); + sample.dataOffset = dataOffset; + sample.compositionTimeOffset = frame.pts - frame.dts; + sample.duration = frame.duration; + sample.size = 4 * frame.length; // Space for nal unit size + + sample.size += frame.byteLength; + if (frame.keyFrame) { + sample.flags.dependsOn = 2; + sample.flags.isNonSyncSample = 0; + } + return sample; + }; // generate the track's sample table from an array of gops + + var generateSampleTable$1 = function (gops, baseDataOffset) { + var h, + i, + sample, + currentGop, + currentFrame, + dataOffset = baseDataOffset || 0, + samples = []; + for (h = 0; h < gops.length; h++) { + currentGop = gops[h]; + for (i = 0; i < currentGop.length; i++) { + currentFrame = currentGop[i]; + sample = sampleForFrame(currentFrame, dataOffset); + dataOffset += sample.size; + samples.push(sample); + } + } + return samples; + }; // generate the track's raw mdat data from an array of gops + + var concatenateNalData = function (gops) { + var h, + i, + j, + currentGop, + currentFrame, + currentNal, + dataOffset = 0, + nalsByteLength = gops.byteLength, + numberOfNals = gops.nalCount, + totalByteLength = nalsByteLength + 4 * numberOfNals, + data = new Uint8Array(totalByteLength), + view = new DataView(data.buffer); // For each Gop.. + + for (h = 0; h < gops.length; h++) { + currentGop = gops[h]; // For each Frame.. + + for (i = 0; i < currentGop.length; i++) { + currentFrame = currentGop[i]; // For each NAL.. + + for (j = 0; j < currentFrame.length; j++) { + currentNal = currentFrame[j]; + view.setUint32(dataOffset, currentNal.data.byteLength); + dataOffset += 4; + data.set(currentNal.data, dataOffset); + dataOffset += currentNal.data.byteLength; + } + } + } + return data; + }; // generate the track's sample table from a frame + + var generateSampleTableForFrame = function (frame, baseDataOffset) { + var sample, + dataOffset = baseDataOffset || 0, + samples = []; + sample = sampleForFrame(frame, dataOffset); + samples.push(sample); + return samples; + }; // generate the track's raw mdat data from a frame + + var concatenateNalDataForFrame = function (frame) { + var i, + currentNal, + dataOffset = 0, + nalsByteLength = frame.byteLength, + numberOfNals = frame.length, + totalByteLength = nalsByteLength + 4 * numberOfNals, + data = new Uint8Array(totalByteLength), + view = new DataView(data.buffer); // For each NAL.. + + for (i = 0; i < frame.length; i++) { + currentNal = frame[i]; + view.setUint32(dataOffset, currentNal.data.byteLength); + dataOffset += 4; + data.set(currentNal.data, dataOffset); + dataOffset += currentNal.data.byteLength; + } + return data; + }; + var frameUtils$1 = { + groupNalsIntoFrames: groupNalsIntoFrames, + groupFramesIntoGops: groupFramesIntoGops, + extendFirstKeyFrame: extendFirstKeyFrame, + generateSampleTable: generateSampleTable$1, + concatenateNalData: concatenateNalData, + generateSampleTableForFrame: generateSampleTableForFrame, + concatenateNalDataForFrame: concatenateNalDataForFrame + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + */ + + var highPrefix = [33, 16, 5, 32, 164, 27]; + var lowPrefix = [33, 65, 108, 84, 1, 2, 4, 8, 168, 2, 4, 8, 17, 191, 252]; + var zeroFill = function (count) { + var a = []; + while (count--) { + a.push(0); + } + return a; + }; + var makeTable = function (metaTable) { + return Object.keys(metaTable).reduce(function (obj, key) { + obj[key] = new Uint8Array(metaTable[key].reduce(function (arr, part) { + return arr.concat(part); + }, [])); + return obj; + }, {}); + }; + var silence; + var silence_1 = function () { + if (!silence) { + // Frames-of-silence to use for filling in missing AAC frames + var coneOfSilence = { + 96000: [highPrefix, [227, 64], zeroFill(154), [56]], + 88200: [highPrefix, [231], zeroFill(170), [56]], + 64000: [highPrefix, [248, 192], zeroFill(240), [56]], + 48000: [highPrefix, [255, 192], zeroFill(268), [55, 148, 128], zeroFill(54), [112]], + 44100: [highPrefix, [255, 192], zeroFill(268), [55, 163, 128], zeroFill(84), [112]], + 32000: [highPrefix, [255, 192], zeroFill(268), [55, 234], zeroFill(226), [112]], + 24000: [highPrefix, [255, 192], zeroFill(268), [55, 255, 128], zeroFill(268), [111, 112], zeroFill(126), [224]], + 16000: [highPrefix, [255, 192], zeroFill(268), [55, 255, 128], zeroFill(268), [111, 255], zeroFill(269), [223, 108], zeroFill(195), [1, 192]], + 12000: [lowPrefix, zeroFill(268), [3, 127, 248], zeroFill(268), [6, 255, 240], zeroFill(268), [13, 255, 224], zeroFill(268), [27, 253, 128], zeroFill(259), [56]], + 11025: [lowPrefix, zeroFill(268), [3, 127, 248], zeroFill(268), [6, 255, 240], zeroFill(268), [13, 255, 224], zeroFill(268), [27, 255, 192], zeroFill(268), [55, 175, 128], zeroFill(108), [112]], + 8000: [lowPrefix, zeroFill(268), [3, 121, 16], zeroFill(47), [7]] + }; + silence = makeTable(coneOfSilence); + } + return silence; + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + */ + + var ONE_SECOND_IN_TS$4 = 90000, + // 90kHz clock + secondsToVideoTs, + secondsToAudioTs, + videoTsToSeconds, + audioTsToSeconds, + audioTsToVideoTs, + videoTsToAudioTs, + metadataTsToSeconds; + secondsToVideoTs = function (seconds) { + return seconds * ONE_SECOND_IN_TS$4; + }; + secondsToAudioTs = function (seconds, sampleRate) { + return seconds * sampleRate; + }; + videoTsToSeconds = function (timestamp) { + return timestamp / ONE_SECOND_IN_TS$4; + }; + audioTsToSeconds = function (timestamp, sampleRate) { + return timestamp / sampleRate; + }; + audioTsToVideoTs = function (timestamp, sampleRate) { + return secondsToVideoTs(audioTsToSeconds(timestamp, sampleRate)); + }; + videoTsToAudioTs = function (timestamp, sampleRate) { + return secondsToAudioTs(videoTsToSeconds(timestamp), sampleRate); + }; + /** + * Adjust ID3 tag or caption timing information by the timeline pts values + * (if keepOriginalTimestamps is false) and convert to seconds + */ + + metadataTsToSeconds = function (timestamp, timelineStartPts, keepOriginalTimestamps) { + return videoTsToSeconds(keepOriginalTimestamps ? timestamp : timestamp - timelineStartPts); + }; + var clock$2 = { + ONE_SECOND_IN_TS: ONE_SECOND_IN_TS$4, + secondsToVideoTs: secondsToVideoTs, + secondsToAudioTs: secondsToAudioTs, + videoTsToSeconds: videoTsToSeconds, + audioTsToSeconds: audioTsToSeconds, + audioTsToVideoTs: audioTsToVideoTs, + videoTsToAudioTs: videoTsToAudioTs, + metadataTsToSeconds: metadataTsToSeconds + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + */ + + var coneOfSilence = silence_1; + var clock$1 = clock$2; + /** + * Sum the `byteLength` properties of the data in each AAC frame + */ + + var sumFrameByteLengths = function (array) { + var i, + currentObj, + sum = 0; // sum the byteLength's all each nal unit in the frame + + for (i = 0; i < array.length; i++) { + currentObj = array[i]; + sum += currentObj.data.byteLength; + } + return sum; + }; // Possibly pad (prefix) the audio track with silence if appending this track + // would lead to the introduction of a gap in the audio buffer + + var prefixWithSilence = function (track, frames, audioAppendStartTs, videoBaseMediaDecodeTime) { + var baseMediaDecodeTimeTs, + frameDuration = 0, + audioGapDuration = 0, + audioFillFrameCount = 0, + audioFillDuration = 0, + silentFrame, + i, + firstFrame; + if (!frames.length) { + return; + } + baseMediaDecodeTimeTs = clock$1.audioTsToVideoTs(track.baseMediaDecodeTime, track.samplerate); // determine frame clock duration based on sample rate, round up to avoid overfills + + frameDuration = Math.ceil(clock$1.ONE_SECOND_IN_TS / (track.samplerate / 1024)); + if (audioAppendStartTs && videoBaseMediaDecodeTime) { + // insert the shortest possible amount (audio gap or audio to video gap) + audioGapDuration = baseMediaDecodeTimeTs - Math.max(audioAppendStartTs, videoBaseMediaDecodeTime); // number of full frames in the audio gap + + audioFillFrameCount = Math.floor(audioGapDuration / frameDuration); + audioFillDuration = audioFillFrameCount * frameDuration; + } // don't attempt to fill gaps smaller than a single frame or larger + // than a half second + + if (audioFillFrameCount < 1 || audioFillDuration > clock$1.ONE_SECOND_IN_TS / 2) { + return; + } + silentFrame = coneOfSilence()[track.samplerate]; + if (!silentFrame) { + // we don't have a silent frame pregenerated for the sample rate, so use a frame + // from the content instead + silentFrame = frames[0].data; + } + for (i = 0; i < audioFillFrameCount; i++) { + firstFrame = frames[0]; + frames.splice(0, 0, { + data: silentFrame, + dts: firstFrame.dts - frameDuration, + pts: firstFrame.pts - frameDuration + }); + } + track.baseMediaDecodeTime -= Math.floor(clock$1.videoTsToAudioTs(audioFillDuration, track.samplerate)); + return audioFillDuration; + }; // If the audio segment extends before the earliest allowed dts + // value, remove AAC frames until starts at or after the earliest + // allowed DTS so that we don't end up with a negative baseMedia- + // DecodeTime for the audio track + + var trimAdtsFramesByEarliestDts = function (adtsFrames, track, earliestAllowedDts) { + if (track.minSegmentDts >= earliestAllowedDts) { + return adtsFrames; + } // We will need to recalculate the earliest segment Dts + + track.minSegmentDts = Infinity; + return adtsFrames.filter(function (currentFrame) { + // If this is an allowed frame, keep it and record it's Dts + if (currentFrame.dts >= earliestAllowedDts) { + track.minSegmentDts = Math.min(track.minSegmentDts, currentFrame.dts); + track.minSegmentPts = track.minSegmentDts; + return true; + } // Otherwise, discard it + + return false; + }); + }; // generate the track's raw mdat data from an array of frames + + var generateSampleTable = function (frames) { + var i, + currentFrame, + samples = []; + for (i = 0; i < frames.length; i++) { + currentFrame = frames[i]; + samples.push({ + size: currentFrame.data.byteLength, + duration: 1024 // For AAC audio, all samples contain 1024 samples + }); + } + + return samples; + }; // generate the track's sample table from an array of frames + + var concatenateFrameData = function (frames) { + var i, + currentFrame, + dataOffset = 0, + data = new Uint8Array(sumFrameByteLengths(frames)); + for (i = 0; i < frames.length; i++) { + currentFrame = frames[i]; + data.set(currentFrame.data, dataOffset); + dataOffset += currentFrame.data.byteLength; + } + return data; + }; + var audioFrameUtils$1 = { + prefixWithSilence: prefixWithSilence, + trimAdtsFramesByEarliestDts: trimAdtsFramesByEarliestDts, + generateSampleTable: generateSampleTable, + concatenateFrameData: concatenateFrameData + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + */ + + var ONE_SECOND_IN_TS$3 = clock$2.ONE_SECOND_IN_TS; + /** + * Store information about the start and end of the track and the + * duration for each frame/sample we process in order to calculate + * the baseMediaDecodeTime + */ + + var collectDtsInfo = function (track, data) { + if (typeof data.pts === 'number') { + if (track.timelineStartInfo.pts === undefined) { + track.timelineStartInfo.pts = data.pts; + } + if (track.minSegmentPts === undefined) { + track.minSegmentPts = data.pts; + } else { + track.minSegmentPts = Math.min(track.minSegmentPts, data.pts); + } + if (track.maxSegmentPts === undefined) { + track.maxSegmentPts = data.pts; + } else { + track.maxSegmentPts = Math.max(track.maxSegmentPts, data.pts); + } + } + if (typeof data.dts === 'number') { + if (track.timelineStartInfo.dts === undefined) { + track.timelineStartInfo.dts = data.dts; + } + if (track.minSegmentDts === undefined) { + track.minSegmentDts = data.dts; + } else { + track.minSegmentDts = Math.min(track.minSegmentDts, data.dts); + } + if (track.maxSegmentDts === undefined) { + track.maxSegmentDts = data.dts; + } else { + track.maxSegmentDts = Math.max(track.maxSegmentDts, data.dts); + } + } + }; + /** + * Clear values used to calculate the baseMediaDecodeTime between + * tracks + */ + + var clearDtsInfo = function (track) { + delete track.minSegmentDts; + delete track.maxSegmentDts; + delete track.minSegmentPts; + delete track.maxSegmentPts; + }; + /** + * Calculate the track's baseMediaDecodeTime based on the earliest + * DTS the transmuxer has ever seen and the minimum DTS for the + * current track + * @param track {object} track metadata configuration + * @param keepOriginalTimestamps {boolean} If true, keep the timestamps + * in the source; false to adjust the first segment to start at 0. + */ + + var calculateTrackBaseMediaDecodeTime = function (track, keepOriginalTimestamps) { + var baseMediaDecodeTime, + scale, + minSegmentDts = track.minSegmentDts; // Optionally adjust the time so the first segment starts at zero. + + if (!keepOriginalTimestamps) { + minSegmentDts -= track.timelineStartInfo.dts; + } // track.timelineStartInfo.baseMediaDecodeTime is the location, in time, where + // we want the start of the first segment to be placed + + baseMediaDecodeTime = track.timelineStartInfo.baseMediaDecodeTime; // Add to that the distance this segment is from the very first + + baseMediaDecodeTime += minSegmentDts; // baseMediaDecodeTime must not become negative + + baseMediaDecodeTime = Math.max(0, baseMediaDecodeTime); + if (track.type === 'audio') { + // Audio has a different clock equal to the sampling_rate so we need to + // scale the PTS values into the clock rate of the track + scale = track.samplerate / ONE_SECOND_IN_TS$3; + baseMediaDecodeTime *= scale; + baseMediaDecodeTime = Math.floor(baseMediaDecodeTime); + } + return baseMediaDecodeTime; + }; + var trackDecodeInfo$1 = { + clearDtsInfo: clearDtsInfo, + calculateTrackBaseMediaDecodeTime: calculateTrackBaseMediaDecodeTime, + collectDtsInfo: collectDtsInfo + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * Reads in-band caption information from a video elementary + * stream. Captions must follow the CEA-708 standard for injection + * into an MPEG-2 transport streams. + * @see https://en.wikipedia.org/wiki/CEA-708 + * @see https://www.gpo.gov/fdsys/pkg/CFR-2007-title47-vol1/pdf/CFR-2007-title47-vol1-sec15-119.pdf + */ + // payload type field to indicate how they are to be + // interpreted. CEAS-708 caption content is always transmitted with + // payload type 0x04. + + var USER_DATA_REGISTERED_ITU_T_T35 = 4, + RBSP_TRAILING_BITS = 128; + /** + * Parse a supplemental enhancement information (SEI) NAL unit. + * Stops parsing once a message of type ITU T T35 has been found. + * + * @param bytes {Uint8Array} the bytes of a SEI NAL unit + * @return {object} the parsed SEI payload + * @see Rec. ITU-T H.264, 7.3.2.3.1 + */ + + var parseSei = function (bytes) { + var i = 0, + result = { + payloadType: -1, + payloadSize: 0 + }, + payloadType = 0, + payloadSize = 0; // go through the sei_rbsp parsing each each individual sei_message + + while (i < bytes.byteLength) { + // stop once we have hit the end of the sei_rbsp + if (bytes[i] === RBSP_TRAILING_BITS) { + break; + } // Parse payload type + + while (bytes[i] === 0xFF) { + payloadType += 255; + i++; + } + payloadType += bytes[i++]; // Parse payload size + + while (bytes[i] === 0xFF) { + payloadSize += 255; + i++; + } + payloadSize += bytes[i++]; // this sei_message is a 608/708 caption so save it and break + // there can only ever be one caption message in a frame's sei + + if (!result.payload && payloadType === USER_DATA_REGISTERED_ITU_T_T35) { + var userIdentifier = String.fromCharCode(bytes[i + 3], bytes[i + 4], bytes[i + 5], bytes[i + 6]); + if (userIdentifier === 'GA94') { + result.payloadType = payloadType; + result.payloadSize = payloadSize; + result.payload = bytes.subarray(i, i + payloadSize); + break; + } else { + result.payload = void 0; + } + } // skip the payload and parse the next message + + i += payloadSize; + payloadType = 0; + payloadSize = 0; + } + return result; + }; // see ANSI/SCTE 128-1 (2013), section 8.1 + + var parseUserData = function (sei) { + // itu_t_t35_contry_code must be 181 (United States) for + // captions + if (sei.payload[0] !== 181) { + return null; + } // itu_t_t35_provider_code should be 49 (ATSC) for captions + + if ((sei.payload[1] << 8 | sei.payload[2]) !== 49) { + return null; + } // the user_identifier should be "GA94" to indicate ATSC1 data + + if (String.fromCharCode(sei.payload[3], sei.payload[4], sei.payload[5], sei.payload[6]) !== 'GA94') { + return null; + } // finally, user_data_type_code should be 0x03 for caption data + + if (sei.payload[7] !== 0x03) { + return null; + } // return the user_data_type_structure and strip the trailing + // marker bits + + return sei.payload.subarray(8, sei.payload.length - 1); + }; // see CEA-708-D, section 4.4 + + var parseCaptionPackets = function (pts, userData) { + var results = [], + i, + count, + offset, + data; // if this is just filler, return immediately + + if (!(userData[0] & 0x40)) { + return results; + } // parse out the cc_data_1 and cc_data_2 fields + + count = userData[0] & 0x1f; + for (i = 0; i < count; i++) { + offset = i * 3; + data = { + type: userData[offset + 2] & 0x03, + pts: pts + }; // capture cc data when cc_valid is 1 + + if (userData[offset + 2] & 0x04) { + data.ccData = userData[offset + 3] << 8 | userData[offset + 4]; + results.push(data); + } + } + return results; + }; + var discardEmulationPreventionBytes$1 = function (data) { + var length = data.byteLength, + emulationPreventionBytesPositions = [], + i = 1, + newLength, + newData; // Find all `Emulation Prevention Bytes` + + while (i < length - 2) { + if (data[i] === 0 && data[i + 1] === 0 && data[i + 2] === 0x03) { + emulationPreventionBytesPositions.push(i + 2); + i += 2; + } else { + i++; + } + } // If no Emulation Prevention Bytes were found just return the original + // array + + if (emulationPreventionBytesPositions.length === 0) { + return data; + } // Create a new array to hold the NAL unit data + + newLength = length - emulationPreventionBytesPositions.length; + newData = new Uint8Array(newLength); + var sourceIndex = 0; + for (i = 0; i < newLength; sourceIndex++, i++) { + if (sourceIndex === emulationPreventionBytesPositions[0]) { + // Skip this byte + sourceIndex++; // Remove this position index + + emulationPreventionBytesPositions.shift(); + } + newData[i] = data[sourceIndex]; + } + return newData; + }; // exports + + var captionPacketParser = { + parseSei: parseSei, + parseUserData: parseUserData, + parseCaptionPackets: parseCaptionPackets, + discardEmulationPreventionBytes: discardEmulationPreventionBytes$1, + USER_DATA_REGISTERED_ITU_T_T35: USER_DATA_REGISTERED_ITU_T_T35 + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * Reads in-band caption information from a video elementary + * stream. Captions must follow the CEA-708 standard for injection + * into an MPEG-2 transport streams. + * @see https://en.wikipedia.org/wiki/CEA-708 + * @see https://www.gpo.gov/fdsys/pkg/CFR-2007-title47-vol1/pdf/CFR-2007-title47-vol1-sec15-119.pdf + */ + // Link To Transport + // ----------------- + + var Stream$7 = stream; + var cea708Parser = captionPacketParser; + var CaptionStream$2 = function (options) { + options = options || {}; + CaptionStream$2.prototype.init.call(this); // parse708captions flag, default to true + + this.parse708captions_ = typeof options.parse708captions === 'boolean' ? options.parse708captions : true; + this.captionPackets_ = []; + this.ccStreams_ = [new Cea608Stream(0, 0), + // eslint-disable-line no-use-before-define + new Cea608Stream(0, 1), + // eslint-disable-line no-use-before-define + new Cea608Stream(1, 0), + // eslint-disable-line no-use-before-define + new Cea608Stream(1, 1) // eslint-disable-line no-use-before-define + ]; + + if (this.parse708captions_) { + this.cc708Stream_ = new Cea708Stream({ + captionServices: options.captionServices + }); // eslint-disable-line no-use-before-define + } + + this.reset(); // forward data and done events from CCs to this CaptionStream + + this.ccStreams_.forEach(function (cc) { + cc.on('data', this.trigger.bind(this, 'data')); + cc.on('partialdone', this.trigger.bind(this, 'partialdone')); + cc.on('done', this.trigger.bind(this, 'done')); + }, this); + if (this.parse708captions_) { + this.cc708Stream_.on('data', this.trigger.bind(this, 'data')); + this.cc708Stream_.on('partialdone', this.trigger.bind(this, 'partialdone')); + this.cc708Stream_.on('done', this.trigger.bind(this, 'done')); + } + }; + CaptionStream$2.prototype = new Stream$7(); + CaptionStream$2.prototype.push = function (event) { + var sei, userData, newCaptionPackets; // only examine SEI NALs + + if (event.nalUnitType !== 'sei_rbsp') { + return; + } // parse the sei + + sei = cea708Parser.parseSei(event.escapedRBSP); // no payload data, skip + + if (!sei.payload) { + return; + } // ignore everything but user_data_registered_itu_t_t35 + + if (sei.payloadType !== cea708Parser.USER_DATA_REGISTERED_ITU_T_T35) { + return; + } // parse out the user data payload + + userData = cea708Parser.parseUserData(sei); // ignore unrecognized userData + + if (!userData) { + return; + } // Sometimes, the same segment # will be downloaded twice. To stop the + // caption data from being processed twice, we track the latest dts we've + // received and ignore everything with a dts before that. However, since + // data for a specific dts can be split across packets on either side of + // a segment boundary, we need to make sure we *don't* ignore the packets + // from the *next* segment that have dts === this.latestDts_. By constantly + // tracking the number of packets received with dts === this.latestDts_, we + // know how many should be ignored once we start receiving duplicates. + + if (event.dts < this.latestDts_) { + // We've started getting older data, so set the flag. + this.ignoreNextEqualDts_ = true; + return; + } else if (event.dts === this.latestDts_ && this.ignoreNextEqualDts_) { + this.numSameDts_--; + if (!this.numSameDts_) { + // We've received the last duplicate packet, time to start processing again + this.ignoreNextEqualDts_ = false; + } + return; + } // parse out CC data packets and save them for later + + newCaptionPackets = cea708Parser.parseCaptionPackets(event.pts, userData); + this.captionPackets_ = this.captionPackets_.concat(newCaptionPackets); + if (this.latestDts_ !== event.dts) { + this.numSameDts_ = 0; + } + this.numSameDts_++; + this.latestDts_ = event.dts; + }; + CaptionStream$2.prototype.flushCCStreams = function (flushType) { + this.ccStreams_.forEach(function (cc) { + return flushType === 'flush' ? cc.flush() : cc.partialFlush(); + }, this); + }; + CaptionStream$2.prototype.flushStream = function (flushType) { + // make sure we actually parsed captions before proceeding + if (!this.captionPackets_.length) { + this.flushCCStreams(flushType); + return; + } // In Chrome, the Array#sort function is not stable so add a + // presortIndex that we can use to ensure we get a stable-sort + + this.captionPackets_.forEach(function (elem, idx) { + elem.presortIndex = idx; + }); // sort caption byte-pairs based on their PTS values + + this.captionPackets_.sort(function (a, b) { + if (a.pts === b.pts) { + return a.presortIndex - b.presortIndex; + } + return a.pts - b.pts; + }); + this.captionPackets_.forEach(function (packet) { + if (packet.type < 2) { + // Dispatch packet to the right Cea608Stream + this.dispatchCea608Packet(packet); + } else { + // Dispatch packet to the Cea708Stream + this.dispatchCea708Packet(packet); + } + }, this); + this.captionPackets_.length = 0; + this.flushCCStreams(flushType); + }; + CaptionStream$2.prototype.flush = function () { + return this.flushStream('flush'); + }; // Only called if handling partial data + + CaptionStream$2.prototype.partialFlush = function () { + return this.flushStream('partialFlush'); + }; + CaptionStream$2.prototype.reset = function () { + this.latestDts_ = null; + this.ignoreNextEqualDts_ = false; + this.numSameDts_ = 0; + this.activeCea608Channel_ = [null, null]; + this.ccStreams_.forEach(function (ccStream) { + ccStream.reset(); + }); + }; // From the CEA-608 spec: + + /* + * When XDS sub-packets are interleaved with other services, the end of each sub-packet shall be followed + * by a control pair to change to a different service. When any of the control codes from 0x10 to 0x1F is + * used to begin a control code pair, it indicates the return to captioning or Text data. The control code pair + * and subsequent data should then be processed according to the FCC rules. It may be necessary for the + * line 21 data encoder to automatically insert a control code pair (i.e. RCL, RU2, RU3, RU4, RDC, or RTD) + * to switch to captioning or Text. + */ + // With that in mind, we ignore any data between an XDS control code and a + // subsequent closed-captioning control code. + + CaptionStream$2.prototype.dispatchCea608Packet = function (packet) { + // NOTE: packet.type is the CEA608 field + if (this.setsTextOrXDSActive(packet)) { + this.activeCea608Channel_[packet.type] = null; + } else if (this.setsChannel1Active(packet)) { + this.activeCea608Channel_[packet.type] = 0; + } else if (this.setsChannel2Active(packet)) { + this.activeCea608Channel_[packet.type] = 1; + } + if (this.activeCea608Channel_[packet.type] === null) { + // If we haven't received anything to set the active channel, or the + // packets are Text/XDS data, discard the data; we don't want jumbled + // captions + return; + } + this.ccStreams_[(packet.type << 1) + this.activeCea608Channel_[packet.type]].push(packet); + }; + CaptionStream$2.prototype.setsChannel1Active = function (packet) { + return (packet.ccData & 0x7800) === 0x1000; + }; + CaptionStream$2.prototype.setsChannel2Active = function (packet) { + return (packet.ccData & 0x7800) === 0x1800; + }; + CaptionStream$2.prototype.setsTextOrXDSActive = function (packet) { + return (packet.ccData & 0x7100) === 0x0100 || (packet.ccData & 0x78fe) === 0x102a || (packet.ccData & 0x78fe) === 0x182a; + }; + CaptionStream$2.prototype.dispatchCea708Packet = function (packet) { + if (this.parse708captions_) { + this.cc708Stream_.push(packet); + } + }; // ---------------------- + // Session to Application + // ---------------------- + // This hash maps special and extended character codes to their + // proper Unicode equivalent. The first one-byte key is just a + // non-standard character code. The two-byte keys that follow are + // the extended CEA708 character codes, along with the preceding + // 0x10 extended character byte to distinguish these codes from + // non-extended character codes. Every CEA708 character code that + // is not in this object maps directly to a standard unicode + // character code. + // The transparent space and non-breaking transparent space are + // technically not fully supported since there is no code to + // make them transparent, so they have normal non-transparent + // stand-ins. + // The special closed caption (CC) character isn't a standard + // unicode character, so a fairly similar unicode character was + // chosen in it's place. + + var CHARACTER_TRANSLATION_708 = { + 0x7f: 0x266a, + // ♪ + 0x1020: 0x20, + // Transparent Space + 0x1021: 0xa0, + // Nob-breaking Transparent Space + 0x1025: 0x2026, + // … + 0x102a: 0x0160, + // Š + 0x102c: 0x0152, + // Œ + 0x1030: 0x2588, + // █ + 0x1031: 0x2018, + // ‘ + 0x1032: 0x2019, + // ’ + 0x1033: 0x201c, + // “ + 0x1034: 0x201d, + // ” + 0x1035: 0x2022, + // • + 0x1039: 0x2122, + // ™ + 0x103a: 0x0161, + // š + 0x103c: 0x0153, + // œ + 0x103d: 0x2120, + // ℠ + 0x103f: 0x0178, + // Ÿ + 0x1076: 0x215b, + // ⅛ + 0x1077: 0x215c, + // ⅜ + 0x1078: 0x215d, + // ⅝ + 0x1079: 0x215e, + // ⅞ + 0x107a: 0x23d0, + // ⏐ + 0x107b: 0x23a4, + // ⎤ + 0x107c: 0x23a3, + // ⎣ + 0x107d: 0x23af, + // ⎯ + 0x107e: 0x23a6, + // ⎦ + 0x107f: 0x23a1, + // ⎡ + 0x10a0: 0x3138 // ㄸ (CC char) + }; + + var get708CharFromCode = function (code) { + var newCode = CHARACTER_TRANSLATION_708[code] || code; + if (code & 0x1000 && code === newCode) { + // Invalid extended code + return ''; + } + return String.fromCharCode(newCode); + }; + var within708TextBlock = function (b) { + return 0x20 <= b && b <= 0x7f || 0xa0 <= b && b <= 0xff; + }; + var Cea708Window = function (windowNum) { + this.windowNum = windowNum; + this.reset(); + }; + Cea708Window.prototype.reset = function () { + this.clearText(); + this.pendingNewLine = false; + this.winAttr = {}; + this.penAttr = {}; + this.penLoc = {}; + this.penColor = {}; // These default values are arbitrary, + // defineWindow will usually override them + + this.visible = 0; + this.rowLock = 0; + this.columnLock = 0; + this.priority = 0; + this.relativePositioning = 0; + this.anchorVertical = 0; + this.anchorHorizontal = 0; + this.anchorPoint = 0; + this.rowCount = 1; + this.virtualRowCount = this.rowCount + 1; + this.columnCount = 41; + this.windowStyle = 0; + this.penStyle = 0; + }; + Cea708Window.prototype.getText = function () { + return this.rows.join('\n'); + }; + Cea708Window.prototype.clearText = function () { + this.rows = ['']; + this.rowIdx = 0; + }; + Cea708Window.prototype.newLine = function (pts) { + if (this.rows.length >= this.virtualRowCount && typeof this.beforeRowOverflow === 'function') { + this.beforeRowOverflow(pts); + } + if (this.rows.length > 0) { + this.rows.push(''); + this.rowIdx++; + } // Show all virtual rows since there's no visible scrolling + + while (this.rows.length > this.virtualRowCount) { + this.rows.shift(); + this.rowIdx--; + } + }; + Cea708Window.prototype.isEmpty = function () { + if (this.rows.length === 0) { + return true; + } else if (this.rows.length === 1) { + return this.rows[0] === ''; + } + return false; + }; + Cea708Window.prototype.addText = function (text) { + this.rows[this.rowIdx] += text; + }; + Cea708Window.prototype.backspace = function () { + if (!this.isEmpty()) { + var row = this.rows[this.rowIdx]; + this.rows[this.rowIdx] = row.substr(0, row.length - 1); + } + }; + var Cea708Service = function (serviceNum, encoding, stream) { + this.serviceNum = serviceNum; + this.text = ''; + this.currentWindow = new Cea708Window(-1); + this.windows = []; + this.stream = stream; // Try to setup a TextDecoder if an `encoding` value was provided + + if (typeof encoding === 'string') { + this.createTextDecoder(encoding); + } + }; + /** + * Initialize service windows + * Must be run before service use + * + * @param {Integer} pts PTS value + * @param {Function} beforeRowOverflow Function to execute before row overflow of a window + */ + + Cea708Service.prototype.init = function (pts, beforeRowOverflow) { + this.startPts = pts; + for (var win = 0; win < 8; win++) { + this.windows[win] = new Cea708Window(win); + if (typeof beforeRowOverflow === 'function') { + this.windows[win].beforeRowOverflow = beforeRowOverflow; + } + } + }; + /** + * Set current window of service to be affected by commands + * + * @param {Integer} windowNum Window number + */ + + Cea708Service.prototype.setCurrentWindow = function (windowNum) { + this.currentWindow = this.windows[windowNum]; + }; + /** + * Try to create a TextDecoder if it is natively supported + */ + + Cea708Service.prototype.createTextDecoder = function (encoding) { + if (typeof TextDecoder === 'undefined') { + this.stream.trigger('log', { + level: 'warn', + message: 'The `encoding` option is unsupported without TextDecoder support' + }); + } else { + try { + this.textDecoder_ = new TextDecoder(encoding); + } catch (error) { + this.stream.trigger('log', { + level: 'warn', + message: 'TextDecoder could not be created with ' + encoding + ' encoding. ' + error + }); + } + } + }; + var Cea708Stream = function (options) { + options = options || {}; + Cea708Stream.prototype.init.call(this); + var self = this; + var captionServices = options.captionServices || {}; + var captionServiceEncodings = {}; + var serviceProps; // Get service encodings from captionServices option block + + Object.keys(captionServices).forEach(serviceName => { + serviceProps = captionServices[serviceName]; + if (/^SERVICE/.test(serviceName)) { + captionServiceEncodings[serviceName] = serviceProps.encoding; + } + }); + this.serviceEncodings = captionServiceEncodings; + this.current708Packet = null; + this.services = {}; + this.push = function (packet) { + if (packet.type === 3) { + // 708 packet start + self.new708Packet(); + self.add708Bytes(packet); + } else { + if (self.current708Packet === null) { + // This should only happen at the start of a file if there's no packet start. + self.new708Packet(); + } + self.add708Bytes(packet); + } + }; + }; + Cea708Stream.prototype = new Stream$7(); + /** + * Push current 708 packet, create new 708 packet. + */ + + Cea708Stream.prototype.new708Packet = function () { + if (this.current708Packet !== null) { + this.push708Packet(); + } + this.current708Packet = { + data: [], + ptsVals: [] + }; + }; + /** + * Add pts and both bytes from packet into current 708 packet. + */ + + Cea708Stream.prototype.add708Bytes = function (packet) { + var data = packet.ccData; + var byte0 = data >>> 8; + var byte1 = data & 0xff; // I would just keep a list of packets instead of bytes, but it isn't clear in the spec + // that service blocks will always line up with byte pairs. + + this.current708Packet.ptsVals.push(packet.pts); + this.current708Packet.data.push(byte0); + this.current708Packet.data.push(byte1); + }; + /** + * Parse completed 708 packet into service blocks and push each service block. + */ + + Cea708Stream.prototype.push708Packet = function () { + var packet708 = this.current708Packet; + var packetData = packet708.data; + var serviceNum = null; + var blockSize = null; + var i = 0; + var b = packetData[i++]; + packet708.seq = b >> 6; + packet708.sizeCode = b & 0x3f; // 0b00111111; + + for (; i < packetData.length; i++) { + b = packetData[i++]; + serviceNum = b >> 5; + blockSize = b & 0x1f; // 0b00011111 + + if (serviceNum === 7 && blockSize > 0) { + // Extended service num + b = packetData[i++]; + serviceNum = b; + } + this.pushServiceBlock(serviceNum, i, blockSize); + if (blockSize > 0) { + i += blockSize - 1; + } + } + }; + /** + * Parse service block, execute commands, read text. + * + * Note: While many of these commands serve important purposes, + * many others just parse out the parameters or attributes, but + * nothing is done with them because this is not a full and complete + * implementation of the entire 708 spec. + * + * @param {Integer} serviceNum Service number + * @param {Integer} start Start index of the 708 packet data + * @param {Integer} size Block size + */ + + Cea708Stream.prototype.pushServiceBlock = function (serviceNum, start, size) { + var b; + var i = start; + var packetData = this.current708Packet.data; + var service = this.services[serviceNum]; + if (!service) { + service = this.initService(serviceNum, i); + } + for (; i < start + size && i < packetData.length; i++) { + b = packetData[i]; + if (within708TextBlock(b)) { + i = this.handleText(i, service); + } else if (b === 0x18) { + i = this.multiByteCharacter(i, service); + } else if (b === 0x10) { + i = this.extendedCommands(i, service); + } else if (0x80 <= b && b <= 0x87) { + i = this.setCurrentWindow(i, service); + } else if (0x98 <= b && b <= 0x9f) { + i = this.defineWindow(i, service); + } else if (b === 0x88) { + i = this.clearWindows(i, service); + } else if (b === 0x8c) { + i = this.deleteWindows(i, service); + } else if (b === 0x89) { + i = this.displayWindows(i, service); + } else if (b === 0x8a) { + i = this.hideWindows(i, service); + } else if (b === 0x8b) { + i = this.toggleWindows(i, service); + } else if (b === 0x97) { + i = this.setWindowAttributes(i, service); + } else if (b === 0x90) { + i = this.setPenAttributes(i, service); + } else if (b === 0x91) { + i = this.setPenColor(i, service); + } else if (b === 0x92) { + i = this.setPenLocation(i, service); + } else if (b === 0x8f) { + service = this.reset(i, service); + } else if (b === 0x08) { + // BS: Backspace + service.currentWindow.backspace(); + } else if (b === 0x0c) { + // FF: Form feed + service.currentWindow.clearText(); + } else if (b === 0x0d) { + // CR: Carriage return + service.currentWindow.pendingNewLine = true; + } else if (b === 0x0e) { + // HCR: Horizontal carriage return + service.currentWindow.clearText(); + } else if (b === 0x8d) { + // DLY: Delay, nothing to do + i++; + } else ; + } + }; + /** + * Execute an extended command + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.extendedCommands = function (i, service) { + var packetData = this.current708Packet.data; + var b = packetData[++i]; + if (within708TextBlock(b)) { + i = this.handleText(i, service, { + isExtended: true + }); + } + return i; + }; + /** + * Get PTS value of a given byte index + * + * @param {Integer} byteIndex Index of the byte + * @return {Integer} PTS + */ + + Cea708Stream.prototype.getPts = function (byteIndex) { + // There's 1 pts value per 2 bytes + return this.current708Packet.ptsVals[Math.floor(byteIndex / 2)]; + }; + /** + * Initializes a service + * + * @param {Integer} serviceNum Service number + * @return {Service} Initialized service object + */ + + Cea708Stream.prototype.initService = function (serviceNum, i) { + var serviceName = 'SERVICE' + serviceNum; + var self = this; + var serviceName; + var encoding; + if (serviceName in this.serviceEncodings) { + encoding = this.serviceEncodings[serviceName]; + } + this.services[serviceNum] = new Cea708Service(serviceNum, encoding, self); + this.services[serviceNum].init(this.getPts(i), function (pts) { + self.flushDisplayed(pts, self.services[serviceNum]); + }); + return this.services[serviceNum]; + }; + /** + * Execute text writing to current window + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.handleText = function (i, service, options) { + var isExtended = options && options.isExtended; + var isMultiByte = options && options.isMultiByte; + var packetData = this.current708Packet.data; + var extended = isExtended ? 0x1000 : 0x0000; + var currentByte = packetData[i]; + var nextByte = packetData[i + 1]; + var win = service.currentWindow; + var char; + var charCodeArray; // Converts an array of bytes to a unicode hex string. + + function toHexString(byteArray) { + return byteArray.map(byte => { + return ('0' + (byte & 0xFF).toString(16)).slice(-2); + }).join(''); + } + if (isMultiByte) { + charCodeArray = [currentByte, nextByte]; + i++; + } else { + charCodeArray = [currentByte]; + } // Use the TextDecoder if one was created for this service + + if (service.textDecoder_ && !isExtended) { + char = service.textDecoder_.decode(new Uint8Array(charCodeArray)); + } else { + // We assume any multi-byte char without a decoder is unicode. + if (isMultiByte) { + const unicode = toHexString(charCodeArray); // Takes a unicode hex string and creates a single character. + + char = String.fromCharCode(parseInt(unicode, 16)); + } else { + char = get708CharFromCode(extended | currentByte); + } + } + if (win.pendingNewLine && !win.isEmpty()) { + win.newLine(this.getPts(i)); + } + win.pendingNewLine = false; + win.addText(char); + return i; + }; + /** + * Handle decoding of multibyte character + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.multiByteCharacter = function (i, service) { + var packetData = this.current708Packet.data; + var firstByte = packetData[i + 1]; + var secondByte = packetData[i + 2]; + if (within708TextBlock(firstByte) && within708TextBlock(secondByte)) { + i = this.handleText(++i, service, { + isMultiByte: true + }); + } + return i; + }; + /** + * Parse and execute the CW# command. + * + * Set the current window. + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.setCurrentWindow = function (i, service) { + var packetData = this.current708Packet.data; + var b = packetData[i]; + var windowNum = b & 0x07; + service.setCurrentWindow(windowNum); + return i; + }; + /** + * Parse and execute the DF# command. + * + * Define a window and set it as the current window. + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.defineWindow = function (i, service) { + var packetData = this.current708Packet.data; + var b = packetData[i]; + var windowNum = b & 0x07; + service.setCurrentWindow(windowNum); + var win = service.currentWindow; + b = packetData[++i]; + win.visible = (b & 0x20) >> 5; // v + + win.rowLock = (b & 0x10) >> 4; // rl + + win.columnLock = (b & 0x08) >> 3; // cl + + win.priority = b & 0x07; // p + + b = packetData[++i]; + win.relativePositioning = (b & 0x80) >> 7; // rp + + win.anchorVertical = b & 0x7f; // av + + b = packetData[++i]; + win.anchorHorizontal = b; // ah + + b = packetData[++i]; + win.anchorPoint = (b & 0xf0) >> 4; // ap + + win.rowCount = b & 0x0f; // rc + + b = packetData[++i]; + win.columnCount = b & 0x3f; // cc + + b = packetData[++i]; + win.windowStyle = (b & 0x38) >> 3; // ws + + win.penStyle = b & 0x07; // ps + // The spec says there are (rowCount+1) "virtual rows" + + win.virtualRowCount = win.rowCount + 1; + return i; + }; + /** + * Parse and execute the SWA command. + * + * Set attributes of the current window. + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.setWindowAttributes = function (i, service) { + var packetData = this.current708Packet.data; + var b = packetData[i]; + var winAttr = service.currentWindow.winAttr; + b = packetData[++i]; + winAttr.fillOpacity = (b & 0xc0) >> 6; // fo + + winAttr.fillRed = (b & 0x30) >> 4; // fr + + winAttr.fillGreen = (b & 0x0c) >> 2; // fg + + winAttr.fillBlue = b & 0x03; // fb + + b = packetData[++i]; + winAttr.borderType = (b & 0xc0) >> 6; // bt + + winAttr.borderRed = (b & 0x30) >> 4; // br + + winAttr.borderGreen = (b & 0x0c) >> 2; // bg + + winAttr.borderBlue = b & 0x03; // bb + + b = packetData[++i]; + winAttr.borderType += (b & 0x80) >> 5; // bt + + winAttr.wordWrap = (b & 0x40) >> 6; // ww + + winAttr.printDirection = (b & 0x30) >> 4; // pd + + winAttr.scrollDirection = (b & 0x0c) >> 2; // sd + + winAttr.justify = b & 0x03; // j + + b = packetData[++i]; + winAttr.effectSpeed = (b & 0xf0) >> 4; // es + + winAttr.effectDirection = (b & 0x0c) >> 2; // ed + + winAttr.displayEffect = b & 0x03; // de + + return i; + }; + /** + * Gather text from all displayed windows and push a caption to output. + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + */ + + Cea708Stream.prototype.flushDisplayed = function (pts, service) { + var displayedText = []; // TODO: Positioning not supported, displaying multiple windows will not necessarily + // display text in the correct order, but sample files so far have not shown any issue. + + for (var winId = 0; winId < 8; winId++) { + if (service.windows[winId].visible && !service.windows[winId].isEmpty()) { + displayedText.push(service.windows[winId].getText()); + } + } + service.endPts = pts; + service.text = displayedText.join('\n\n'); + this.pushCaption(service); + service.startPts = pts; + }; + /** + * Push a caption to output if the caption contains text. + * + * @param {Service} service The service object to be affected + */ + + Cea708Stream.prototype.pushCaption = function (service) { + if (service.text !== '') { + this.trigger('data', { + startPts: service.startPts, + endPts: service.endPts, + text: service.text, + stream: 'cc708_' + service.serviceNum + }); + service.text = ''; + service.startPts = service.endPts; + } + }; + /** + * Parse and execute the DSW command. + * + * Set visible property of windows based on the parsed bitmask. + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.displayWindows = function (i, service) { + var packetData = this.current708Packet.data; + var b = packetData[++i]; + var pts = this.getPts(i); + this.flushDisplayed(pts, service); + for (var winId = 0; winId < 8; winId++) { + if (b & 0x01 << winId) { + service.windows[winId].visible = 1; + } + } + return i; + }; + /** + * Parse and execute the HDW command. + * + * Set visible property of windows based on the parsed bitmask. + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.hideWindows = function (i, service) { + var packetData = this.current708Packet.data; + var b = packetData[++i]; + var pts = this.getPts(i); + this.flushDisplayed(pts, service); + for (var winId = 0; winId < 8; winId++) { + if (b & 0x01 << winId) { + service.windows[winId].visible = 0; + } + } + return i; + }; + /** + * Parse and execute the TGW command. + * + * Set visible property of windows based on the parsed bitmask. + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.toggleWindows = function (i, service) { + var packetData = this.current708Packet.data; + var b = packetData[++i]; + var pts = this.getPts(i); + this.flushDisplayed(pts, service); + for (var winId = 0; winId < 8; winId++) { + if (b & 0x01 << winId) { + service.windows[winId].visible ^= 1; + } + } + return i; + }; + /** + * Parse and execute the CLW command. + * + * Clear text of windows based on the parsed bitmask. + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.clearWindows = function (i, service) { + var packetData = this.current708Packet.data; + var b = packetData[++i]; + var pts = this.getPts(i); + this.flushDisplayed(pts, service); + for (var winId = 0; winId < 8; winId++) { + if (b & 0x01 << winId) { + service.windows[winId].clearText(); + } + } + return i; + }; + /** + * Parse and execute the DLW command. + * + * Re-initialize windows based on the parsed bitmask. + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.deleteWindows = function (i, service) { + var packetData = this.current708Packet.data; + var b = packetData[++i]; + var pts = this.getPts(i); + this.flushDisplayed(pts, service); + for (var winId = 0; winId < 8; winId++) { + if (b & 0x01 << winId) { + service.windows[winId].reset(); + } + } + return i; + }; + /** + * Parse and execute the SPA command. + * + * Set pen attributes of the current window. + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.setPenAttributes = function (i, service) { + var packetData = this.current708Packet.data; + var b = packetData[i]; + var penAttr = service.currentWindow.penAttr; + b = packetData[++i]; + penAttr.textTag = (b & 0xf0) >> 4; // tt + + penAttr.offset = (b & 0x0c) >> 2; // o + + penAttr.penSize = b & 0x03; // s + + b = packetData[++i]; + penAttr.italics = (b & 0x80) >> 7; // i + + penAttr.underline = (b & 0x40) >> 6; // u + + penAttr.edgeType = (b & 0x38) >> 3; // et + + penAttr.fontStyle = b & 0x07; // fs + + return i; + }; + /** + * Parse and execute the SPC command. + * + * Set pen color of the current window. + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.setPenColor = function (i, service) { + var packetData = this.current708Packet.data; + var b = packetData[i]; + var penColor = service.currentWindow.penColor; + b = packetData[++i]; + penColor.fgOpacity = (b & 0xc0) >> 6; // fo + + penColor.fgRed = (b & 0x30) >> 4; // fr + + penColor.fgGreen = (b & 0x0c) >> 2; // fg + + penColor.fgBlue = b & 0x03; // fb + + b = packetData[++i]; + penColor.bgOpacity = (b & 0xc0) >> 6; // bo + + penColor.bgRed = (b & 0x30) >> 4; // br + + penColor.bgGreen = (b & 0x0c) >> 2; // bg + + penColor.bgBlue = b & 0x03; // bb + + b = packetData[++i]; + penColor.edgeRed = (b & 0x30) >> 4; // er + + penColor.edgeGreen = (b & 0x0c) >> 2; // eg + + penColor.edgeBlue = b & 0x03; // eb + + return i; + }; + /** + * Parse and execute the SPL command. + * + * Set pen location of the current window. + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Integer} New index after parsing + */ + + Cea708Stream.prototype.setPenLocation = function (i, service) { + var packetData = this.current708Packet.data; + var b = packetData[i]; + var penLoc = service.currentWindow.penLoc; // Positioning isn't really supported at the moment, so this essentially just inserts a linebreak + + service.currentWindow.pendingNewLine = true; + b = packetData[++i]; + penLoc.row = b & 0x0f; // r + + b = packetData[++i]; + penLoc.column = b & 0x3f; // c + + return i; + }; + /** + * Execute the RST command. + * + * Reset service to a clean slate. Re-initialize. + * + * @param {Integer} i Current index in the 708 packet + * @param {Service} service The service object to be affected + * @return {Service} Re-initialized service + */ + + Cea708Stream.prototype.reset = function (i, service) { + var pts = this.getPts(i); + this.flushDisplayed(pts, service); + return this.initService(service.serviceNum, i); + }; // This hash maps non-ASCII, special, and extended character codes to their + // proper Unicode equivalent. The first keys that are only a single byte + // are the non-standard ASCII characters, which simply map the CEA608 byte + // to the standard ASCII/Unicode. The two-byte keys that follow are the CEA608 + // character codes, but have their MSB bitmasked with 0x03 so that a lookup + // can be performed regardless of the field and data channel on which the + // character code was received. + + var CHARACTER_TRANSLATION = { + 0x2a: 0xe1, + // á + 0x5c: 0xe9, + // é + 0x5e: 0xed, + // í + 0x5f: 0xf3, + // ó + 0x60: 0xfa, + // ú + 0x7b: 0xe7, + // ç + 0x7c: 0xf7, + // ÷ + 0x7d: 0xd1, + // Ñ + 0x7e: 0xf1, + // ñ + 0x7f: 0x2588, + // █ + 0x0130: 0xae, + // ® + 0x0131: 0xb0, + // ° + 0x0132: 0xbd, + // ½ + 0x0133: 0xbf, + // ¿ + 0x0134: 0x2122, + // ™ + 0x0135: 0xa2, + // ¢ + 0x0136: 0xa3, + // £ + 0x0137: 0x266a, + // ♪ + 0x0138: 0xe0, + // à + 0x0139: 0xa0, + // + 0x013a: 0xe8, + // è + 0x013b: 0xe2, + // â + 0x013c: 0xea, + // ê + 0x013d: 0xee, + // î + 0x013e: 0xf4, + // ô + 0x013f: 0xfb, + // û + 0x0220: 0xc1, + // Á + 0x0221: 0xc9, + // É + 0x0222: 0xd3, + // Ó + 0x0223: 0xda, + // Ú + 0x0224: 0xdc, + // Ü + 0x0225: 0xfc, + // ü + 0x0226: 0x2018, + // ‘ + 0x0227: 0xa1, + // ¡ + 0x0228: 0x2a, + // * + 0x0229: 0x27, + // ' + 0x022a: 0x2014, + // — + 0x022b: 0xa9, + // © + 0x022c: 0x2120, + // ℠ + 0x022d: 0x2022, + // • + 0x022e: 0x201c, + // “ + 0x022f: 0x201d, + // ” + 0x0230: 0xc0, + // À + 0x0231: 0xc2, + //  + 0x0232: 0xc7, + // Ç + 0x0233: 0xc8, + // È + 0x0234: 0xca, + // Ê + 0x0235: 0xcb, + // Ë + 0x0236: 0xeb, + // ë + 0x0237: 0xce, + // Î + 0x0238: 0xcf, + // Ï + 0x0239: 0xef, + // ï + 0x023a: 0xd4, + // Ô + 0x023b: 0xd9, + // Ù + 0x023c: 0xf9, + // ù + 0x023d: 0xdb, + // Û + 0x023e: 0xab, + // « + 0x023f: 0xbb, + // » + 0x0320: 0xc3, + // à + 0x0321: 0xe3, + // ã + 0x0322: 0xcd, + // Í + 0x0323: 0xcc, + // Ì + 0x0324: 0xec, + // ì + 0x0325: 0xd2, + // Ò + 0x0326: 0xf2, + // ò + 0x0327: 0xd5, + // Õ + 0x0328: 0xf5, + // õ + 0x0329: 0x7b, + // { + 0x032a: 0x7d, + // } + 0x032b: 0x5c, + // \ + 0x032c: 0x5e, + // ^ + 0x032d: 0x5f, + // _ + 0x032e: 0x7c, + // | + 0x032f: 0x7e, + // ~ + 0x0330: 0xc4, + // Ä + 0x0331: 0xe4, + // ä + 0x0332: 0xd6, + // Ö + 0x0333: 0xf6, + // ö + 0x0334: 0xdf, + // ß + 0x0335: 0xa5, + // ¥ + 0x0336: 0xa4, + // ¤ + 0x0337: 0x2502, + // │ + 0x0338: 0xc5, + // Å + 0x0339: 0xe5, + // å + 0x033a: 0xd8, + // Ø + 0x033b: 0xf8, + // ø + 0x033c: 0x250c, + // ┌ + 0x033d: 0x2510, + // ┐ + 0x033e: 0x2514, + // └ + 0x033f: 0x2518 // ┘ + }; + + var getCharFromCode = function (code) { + if (code === null) { + return ''; + } + code = CHARACTER_TRANSLATION[code] || code; + return String.fromCharCode(code); + }; // the index of the last row in a CEA-608 display buffer + + var BOTTOM_ROW = 14; // This array is used for mapping PACs -> row #, since there's no way of + // getting it through bit logic. + + var ROWS = [0x1100, 0x1120, 0x1200, 0x1220, 0x1500, 0x1520, 0x1600, 0x1620, 0x1700, 0x1720, 0x1000, 0x1300, 0x1320, 0x1400, 0x1420]; // CEA-608 captions are rendered onto a 34x15 matrix of character + // cells. The "bottom" row is the last element in the outer array. + // We keep track of positioning information as we go by storing the + // number of indentations and the tab offset in this buffer. + + var createDisplayBuffer = function () { + var result = [], + i = BOTTOM_ROW + 1; + while (i--) { + result.push({ + text: '', + indent: 0, + offset: 0 + }); + } + return result; + }; + var Cea608Stream = function (field, dataChannel) { + Cea608Stream.prototype.init.call(this); + this.field_ = field || 0; + this.dataChannel_ = dataChannel || 0; + this.name_ = 'CC' + ((this.field_ << 1 | this.dataChannel_) + 1); + this.setConstants(); + this.reset(); + this.push = function (packet) { + var data, swap, char0, char1, text; // remove the parity bits + + data = packet.ccData & 0x7f7f; // ignore duplicate control codes; the spec demands they're sent twice + + if (data === this.lastControlCode_) { + this.lastControlCode_ = null; + return; + } // Store control codes + + if ((data & 0xf000) === 0x1000) { + this.lastControlCode_ = data; + } else if (data !== this.PADDING_) { + this.lastControlCode_ = null; + } + char0 = data >>> 8; + char1 = data & 0xff; + if (data === this.PADDING_) { + return; + } else if (data === this.RESUME_CAPTION_LOADING_) { + this.mode_ = 'popOn'; + } else if (data === this.END_OF_CAPTION_) { + // If an EOC is received while in paint-on mode, the displayed caption + // text should be swapped to non-displayed memory as if it was a pop-on + // caption. Because of that, we should explicitly switch back to pop-on + // mode + this.mode_ = 'popOn'; + this.clearFormatting(packet.pts); // if a caption was being displayed, it's gone now + + this.flushDisplayed(packet.pts); // flip memory + + swap = this.displayed_; + this.displayed_ = this.nonDisplayed_; + this.nonDisplayed_ = swap; // start measuring the time to display the caption + + this.startPts_ = packet.pts; + } else if (data === this.ROLL_UP_2_ROWS_) { + this.rollUpRows_ = 2; + this.setRollUp(packet.pts); + } else if (data === this.ROLL_UP_3_ROWS_) { + this.rollUpRows_ = 3; + this.setRollUp(packet.pts); + } else if (data === this.ROLL_UP_4_ROWS_) { + this.rollUpRows_ = 4; + this.setRollUp(packet.pts); + } else if (data === this.CARRIAGE_RETURN_) { + this.clearFormatting(packet.pts); + this.flushDisplayed(packet.pts); + this.shiftRowsUp_(); + this.startPts_ = packet.pts; + } else if (data === this.BACKSPACE_) { + if (this.mode_ === 'popOn') { + this.nonDisplayed_[this.row_].text = this.nonDisplayed_[this.row_].text.slice(0, -1); + } else { + this.displayed_[this.row_].text = this.displayed_[this.row_].text.slice(0, -1); + } + } else if (data === this.ERASE_DISPLAYED_MEMORY_) { + this.flushDisplayed(packet.pts); + this.displayed_ = createDisplayBuffer(); + } else if (data === this.ERASE_NON_DISPLAYED_MEMORY_) { + this.nonDisplayed_ = createDisplayBuffer(); + } else if (data === this.RESUME_DIRECT_CAPTIONING_) { + if (this.mode_ !== 'paintOn') { + // NOTE: This should be removed when proper caption positioning is + // implemented + this.flushDisplayed(packet.pts); + this.displayed_ = createDisplayBuffer(); + } + this.mode_ = 'paintOn'; + this.startPts_ = packet.pts; // Append special characters to caption text + } else if (this.isSpecialCharacter(char0, char1)) { + // Bitmask char0 so that we can apply character transformations + // regardless of field and data channel. + // Then byte-shift to the left and OR with char1 so we can pass the + // entire character code to `getCharFromCode`. + char0 = (char0 & 0x03) << 8; + text = getCharFromCode(char0 | char1); + this[this.mode_](packet.pts, text); + this.column_++; // Append extended characters to caption text + } else if (this.isExtCharacter(char0, char1)) { + // Extended characters always follow their "non-extended" equivalents. + // IE if a "è" is desired, you'll always receive "eè"; non-compliant + // decoders are supposed to drop the "è", while compliant decoders + // backspace the "e" and insert "è". + // Delete the previous character + if (this.mode_ === 'popOn') { + this.nonDisplayed_[this.row_].text = this.nonDisplayed_[this.row_].text.slice(0, -1); + } else { + this.displayed_[this.row_].text = this.displayed_[this.row_].text.slice(0, -1); + } // Bitmask char0 so that we can apply character transformations + // regardless of field and data channel. + // Then byte-shift to the left and OR with char1 so we can pass the + // entire character code to `getCharFromCode`. + + char0 = (char0 & 0x03) << 8; + text = getCharFromCode(char0 | char1); + this[this.mode_](packet.pts, text); + this.column_++; // Process mid-row codes + } else if (this.isMidRowCode(char0, char1)) { + // Attributes are not additive, so clear all formatting + this.clearFormatting(packet.pts); // According to the standard, mid-row codes + // should be replaced with spaces, so add one now + + this[this.mode_](packet.pts, ' '); + this.column_++; + if ((char1 & 0xe) === 0xe) { + this.addFormatting(packet.pts, ['i']); + } + if ((char1 & 0x1) === 0x1) { + this.addFormatting(packet.pts, ['u']); + } // Detect offset control codes and adjust cursor + } else if (this.isOffsetControlCode(char0, char1)) { + // Cursor position is set by indent PAC (see below) in 4-column + // increments, with an additional offset code of 1-3 to reach any + // of the 32 columns specified by CEA-608. So all we need to do + // here is increment the column cursor by the given offset. + const offset = char1 & 0x03; // For an offest value 1-3, set the offset for that caption + // in the non-displayed array. + + this.nonDisplayed_[this.row_].offset = offset; + this.column_ += offset; // Detect PACs (Preamble Address Codes) + } else if (this.isPAC(char0, char1)) { + // There's no logic for PAC -> row mapping, so we have to just + // find the row code in an array and use its index :( + var row = ROWS.indexOf(data & 0x1f20); // Configure the caption window if we're in roll-up mode + + if (this.mode_ === 'rollUp') { + // This implies that the base row is incorrectly set. + // As per the recommendation in CEA-608(Base Row Implementation), defer to the number + // of roll-up rows set. + if (row - this.rollUpRows_ + 1 < 0) { + row = this.rollUpRows_ - 1; + } + this.setRollUp(packet.pts, row); + } // Ensure the row is between 0 and 14, otherwise use the most + // recent or default row. + + if (row !== this.row_ && row >= 0 && row <= 14) { + // formatting is only persistent for current row + this.clearFormatting(packet.pts); + this.row_ = row; + } // All PACs can apply underline, so detect and apply + // (All odd-numbered second bytes set underline) + + if (char1 & 0x1 && this.formatting_.indexOf('u') === -1) { + this.addFormatting(packet.pts, ['u']); + } + if ((data & 0x10) === 0x10) { + // We've got an indent level code. Each successive even number + // increments the column cursor by 4, so we can get the desired + // column position by bit-shifting to the right (to get n/2) + // and multiplying by 4. + const indentations = (data & 0xe) >> 1; + this.column_ = indentations * 4; // add to the number of indentations for positioning + + this.nonDisplayed_[this.row_].indent += indentations; + } + if (this.isColorPAC(char1)) { + // it's a color code, though we only support white, which + // can be either normal or italicized. white italics can be + // either 0x4e or 0x6e depending on the row, so we just + // bitwise-and with 0xe to see if italics should be turned on + if ((char1 & 0xe) === 0xe) { + this.addFormatting(packet.pts, ['i']); + } + } // We have a normal character in char0, and possibly one in char1 + } else if (this.isNormalChar(char0)) { + if (char1 === 0x00) { + char1 = null; + } + text = getCharFromCode(char0); + text += getCharFromCode(char1); + this[this.mode_](packet.pts, text); + this.column_ += text.length; + } // finish data processing + }; + }; + + Cea608Stream.prototype = new Stream$7(); // Trigger a cue point that captures the current state of the + // display buffer + + Cea608Stream.prototype.flushDisplayed = function (pts) { + const logWarning = index => { + this.trigger('log', { + level: 'warn', + message: 'Skipping a malformed 608 caption at index ' + index + '.' + }); + }; + const content = []; + this.displayed_.forEach((row, i) => { + if (row && row.text && row.text.length) { + try { + // remove spaces from the start and end of the string + row.text = row.text.trim(); + } catch (e) { + // Ordinarily, this shouldn't happen. However, caption + // parsing errors should not throw exceptions and + // break playback. + logWarning(i); + } // See the below link for more details on the following fields: + // https://dvcs.w3.org/hg/text-tracks/raw-file/default/608toVTT/608toVTT.html#positioning-in-cea-608 + + if (row.text.length) { + content.push({ + // The text to be displayed in the caption from this specific row, with whitespace removed. + text: row.text, + // Value between 1 and 15 representing the PAC row used to calculate line height. + line: i + 1, + // A number representing the indent position by percentage (CEA-608 PAC indent code). + // The value will be a number between 10 and 80. Offset is used to add an aditional + // value to the position if necessary. + position: 10 + Math.min(70, row.indent * 10) + row.offset * 2.5 + }); + } + } else if (row === undefined || row === null) { + logWarning(i); + } + }); + if (content.length) { + this.trigger('data', { + startPts: this.startPts_, + endPts: pts, + content, + stream: this.name_ + }); + } + }; + /** + * Zero out the data, used for startup and on seek + */ + + Cea608Stream.prototype.reset = function () { + this.mode_ = 'popOn'; // When in roll-up mode, the index of the last row that will + // actually display captions. If a caption is shifted to a row + // with a lower index than this, it is cleared from the display + // buffer + + this.topRow_ = 0; + this.startPts_ = 0; + this.displayed_ = createDisplayBuffer(); + this.nonDisplayed_ = createDisplayBuffer(); + this.lastControlCode_ = null; // Track row and column for proper line-breaking and spacing + + this.column_ = 0; + this.row_ = BOTTOM_ROW; + this.rollUpRows_ = 2; // This variable holds currently-applied formatting + + this.formatting_ = []; + }; + /** + * Sets up control code and related constants for this instance + */ + + Cea608Stream.prototype.setConstants = function () { + // The following attributes have these uses: + // ext_ : char0 for mid-row codes, and the base for extended + // chars (ext_+0, ext_+1, and ext_+2 are char0s for + // extended codes) + // control_: char0 for control codes, except byte-shifted to the + // left so that we can do this.control_ | CONTROL_CODE + // offset_: char0 for tab offset codes + // + // It's also worth noting that control codes, and _only_ control codes, + // differ between field 1 and field2. Field 2 control codes are always + // their field 1 value plus 1. That's why there's the "| field" on the + // control value. + if (this.dataChannel_ === 0) { + this.BASE_ = 0x10; + this.EXT_ = 0x11; + this.CONTROL_ = (0x14 | this.field_) << 8; + this.OFFSET_ = 0x17; + } else if (this.dataChannel_ === 1) { + this.BASE_ = 0x18; + this.EXT_ = 0x19; + this.CONTROL_ = (0x1c | this.field_) << 8; + this.OFFSET_ = 0x1f; + } // Constants for the LSByte command codes recognized by Cea608Stream. This + // list is not exhaustive. For a more comprehensive listing and semantics see + // http://www.gpo.gov/fdsys/pkg/CFR-2010-title47-vol1/pdf/CFR-2010-title47-vol1-sec15-119.pdf + // Padding + + this.PADDING_ = 0x0000; // Pop-on Mode + + this.RESUME_CAPTION_LOADING_ = this.CONTROL_ | 0x20; + this.END_OF_CAPTION_ = this.CONTROL_ | 0x2f; // Roll-up Mode + + this.ROLL_UP_2_ROWS_ = this.CONTROL_ | 0x25; + this.ROLL_UP_3_ROWS_ = this.CONTROL_ | 0x26; + this.ROLL_UP_4_ROWS_ = this.CONTROL_ | 0x27; + this.CARRIAGE_RETURN_ = this.CONTROL_ | 0x2d; // paint-on mode + + this.RESUME_DIRECT_CAPTIONING_ = this.CONTROL_ | 0x29; // Erasure + + this.BACKSPACE_ = this.CONTROL_ | 0x21; + this.ERASE_DISPLAYED_MEMORY_ = this.CONTROL_ | 0x2c; + this.ERASE_NON_DISPLAYED_MEMORY_ = this.CONTROL_ | 0x2e; + }; + /** + * Detects if the 2-byte packet data is a special character + * + * Special characters have a second byte in the range 0x30 to 0x3f, + * with the first byte being 0x11 (for data channel 1) or 0x19 (for + * data channel 2). + * + * @param {Integer} char0 The first byte + * @param {Integer} char1 The second byte + * @return {Boolean} Whether the 2 bytes are an special character + */ + + Cea608Stream.prototype.isSpecialCharacter = function (char0, char1) { + return char0 === this.EXT_ && char1 >= 0x30 && char1 <= 0x3f; + }; + /** + * Detects if the 2-byte packet data is an extended character + * + * Extended characters have a second byte in the range 0x20 to 0x3f, + * with the first byte being 0x12 or 0x13 (for data channel 1) or + * 0x1a or 0x1b (for data channel 2). + * + * @param {Integer} char0 The first byte + * @param {Integer} char1 The second byte + * @return {Boolean} Whether the 2 bytes are an extended character + */ + + Cea608Stream.prototype.isExtCharacter = function (char0, char1) { + return (char0 === this.EXT_ + 1 || char0 === this.EXT_ + 2) && char1 >= 0x20 && char1 <= 0x3f; + }; + /** + * Detects if the 2-byte packet is a mid-row code + * + * Mid-row codes have a second byte in the range 0x20 to 0x2f, with + * the first byte being 0x11 (for data channel 1) or 0x19 (for data + * channel 2). + * + * @param {Integer} char0 The first byte + * @param {Integer} char1 The second byte + * @return {Boolean} Whether the 2 bytes are a mid-row code + */ + + Cea608Stream.prototype.isMidRowCode = function (char0, char1) { + return char0 === this.EXT_ && char1 >= 0x20 && char1 <= 0x2f; + }; + /** + * Detects if the 2-byte packet is an offset control code + * + * Offset control codes have a second byte in the range 0x21 to 0x23, + * with the first byte being 0x17 (for data channel 1) or 0x1f (for + * data channel 2). + * + * @param {Integer} char0 The first byte + * @param {Integer} char1 The second byte + * @return {Boolean} Whether the 2 bytes are an offset control code + */ + + Cea608Stream.prototype.isOffsetControlCode = function (char0, char1) { + return char0 === this.OFFSET_ && char1 >= 0x21 && char1 <= 0x23; + }; + /** + * Detects if the 2-byte packet is a Preamble Address Code + * + * PACs have a first byte in the range 0x10 to 0x17 (for data channel 1) + * or 0x18 to 0x1f (for data channel 2), with the second byte in the + * range 0x40 to 0x7f. + * + * @param {Integer} char0 The first byte + * @param {Integer} char1 The second byte + * @return {Boolean} Whether the 2 bytes are a PAC + */ + + Cea608Stream.prototype.isPAC = function (char0, char1) { + return char0 >= this.BASE_ && char0 < this.BASE_ + 8 && char1 >= 0x40 && char1 <= 0x7f; + }; + /** + * Detects if a packet's second byte is in the range of a PAC color code + * + * PAC color codes have the second byte be in the range 0x40 to 0x4f, or + * 0x60 to 0x6f. + * + * @param {Integer} char1 The second byte + * @return {Boolean} Whether the byte is a color PAC + */ + + Cea608Stream.prototype.isColorPAC = function (char1) { + return char1 >= 0x40 && char1 <= 0x4f || char1 >= 0x60 && char1 <= 0x7f; + }; + /** + * Detects if a single byte is in the range of a normal character + * + * Normal text bytes are in the range 0x20 to 0x7f. + * + * @param {Integer} char The byte + * @return {Boolean} Whether the byte is a normal character + */ + + Cea608Stream.prototype.isNormalChar = function (char) { + return char >= 0x20 && char <= 0x7f; + }; + /** + * Configures roll-up + * + * @param {Integer} pts Current PTS + * @param {Integer} newBaseRow Used by PACs to slide the current window to + * a new position + */ + + Cea608Stream.prototype.setRollUp = function (pts, newBaseRow) { + // Reset the base row to the bottom row when switching modes + if (this.mode_ !== 'rollUp') { + this.row_ = BOTTOM_ROW; + this.mode_ = 'rollUp'; // Spec says to wipe memories when switching to roll-up + + this.flushDisplayed(pts); + this.nonDisplayed_ = createDisplayBuffer(); + this.displayed_ = createDisplayBuffer(); + } + if (newBaseRow !== undefined && newBaseRow !== this.row_) { + // move currently displayed captions (up or down) to the new base row + for (var i = 0; i < this.rollUpRows_; i++) { + this.displayed_[newBaseRow - i] = this.displayed_[this.row_ - i]; + this.displayed_[this.row_ - i] = { + text: '', + indent: 0, + offset: 0 + }; + } + } + if (newBaseRow === undefined) { + newBaseRow = this.row_; + } + this.topRow_ = newBaseRow - this.rollUpRows_ + 1; + }; // Adds the opening HTML tag for the passed character to the caption text, + // and keeps track of it for later closing + + Cea608Stream.prototype.addFormatting = function (pts, format) { + this.formatting_ = this.formatting_.concat(format); + var text = format.reduce(function (text, format) { + return text + '<' + format + '>'; + }, ''); + this[this.mode_](pts, text); + }; // Adds HTML closing tags for current formatting to caption text and + // clears remembered formatting + + Cea608Stream.prototype.clearFormatting = function (pts) { + if (!this.formatting_.length) { + return; + } + var text = this.formatting_.reverse().reduce(function (text, format) { + return text + '</' + format + '>'; + }, ''); + this.formatting_ = []; + this[this.mode_](pts, text); + }; // Mode Implementations + + Cea608Stream.prototype.popOn = function (pts, text) { + var baseRow = this.nonDisplayed_[this.row_].text; // buffer characters + + baseRow += text; + this.nonDisplayed_[this.row_].text = baseRow; + }; + Cea608Stream.prototype.rollUp = function (pts, text) { + var baseRow = this.displayed_[this.row_].text; + baseRow += text; + this.displayed_[this.row_].text = baseRow; + }; + Cea608Stream.prototype.shiftRowsUp_ = function () { + var i; // clear out inactive rows + + for (i = 0; i < this.topRow_; i++) { + this.displayed_[i] = { + text: '', + indent: 0, + offset: 0 + }; + } + for (i = this.row_ + 1; i < BOTTOM_ROW + 1; i++) { + this.displayed_[i] = { + text: '', + indent: 0, + offset: 0 + }; + } // shift displayed rows up + + for (i = this.topRow_; i < this.row_; i++) { + this.displayed_[i] = this.displayed_[i + 1]; + } // clear out the bottom row + + this.displayed_[this.row_] = { + text: '', + indent: 0, + offset: 0 + }; + }; + Cea608Stream.prototype.paintOn = function (pts, text) { + var baseRow = this.displayed_[this.row_].text; + baseRow += text; + this.displayed_[this.row_].text = baseRow; + }; // exports + + var captionStream = { + CaptionStream: CaptionStream$2, + Cea608Stream: Cea608Stream, + Cea708Stream: Cea708Stream + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + */ + + var streamTypes = { + H264_STREAM_TYPE: 0x1B, + ADTS_STREAM_TYPE: 0x0F, + METADATA_STREAM_TYPE: 0x15 + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * Accepts program elementary stream (PES) data events and corrects + * decode and presentation time stamps to account for a rollover + * of the 33 bit value. + */ + + var Stream$6 = stream; + var MAX_TS = 8589934592; + var RO_THRESH = 4294967296; + var TYPE_SHARED = 'shared'; + var handleRollover$1 = function (value, reference) { + var direction = 1; + if (value > reference) { + // If the current timestamp value is greater than our reference timestamp and we detect a + // timestamp rollover, this means the roll over is happening in the opposite direction. + // Example scenario: Enter a long stream/video just after a rollover occurred. The reference + // point will be set to a small number, e.g. 1. The user then seeks backwards over the + // rollover point. In loading this segment, the timestamp values will be very large, + // e.g. 2^33 - 1. Since this comes before the data we loaded previously, we want to adjust + // the time stamp to be `value - 2^33`. + direction = -1; + } // Note: A seek forwards or back that is greater than the RO_THRESH (2^32, ~13 hours) will + // cause an incorrect adjustment. + + while (Math.abs(reference - value) > RO_THRESH) { + value += direction * MAX_TS; + } + return value; + }; + var TimestampRolloverStream$1 = function (type) { + var lastDTS, referenceDTS; + TimestampRolloverStream$1.prototype.init.call(this); // The "shared" type is used in cases where a stream will contain muxed + // video and audio. We could use `undefined` here, but having a string + // makes debugging a little clearer. + + this.type_ = type || TYPE_SHARED; + this.push = function (data) { + /** + * Rollover stream expects data from elementary stream. + * Elementary stream can push forward 2 types of data + * - Parsed Video/Audio/Timed-metadata PES (packetized elementary stream) packets + * - Tracks metadata from PMT (Program Map Table) + * Rollover stream expects pts/dts info to be available, since it stores lastDTS + * We should ignore non-PES packets since they may override lastDTS to undefined. + * lastDTS is important to signal the next segments + * about rollover from the previous segments. + */ + if (data.type === 'metadata') { + this.trigger('data', data); + return; + } // Any "shared" rollover streams will accept _all_ data. Otherwise, + // streams will only accept data that matches their type. + + if (this.type_ !== TYPE_SHARED && data.type !== this.type_) { + return; + } + if (referenceDTS === undefined) { + referenceDTS = data.dts; + } + data.dts = handleRollover$1(data.dts, referenceDTS); + data.pts = handleRollover$1(data.pts, referenceDTS); + lastDTS = data.dts; + this.trigger('data', data); + }; + this.flush = function () { + referenceDTS = lastDTS; + this.trigger('done'); + }; + this.endTimeline = function () { + this.flush(); + this.trigger('endedtimeline'); + }; + this.discontinuity = function () { + referenceDTS = void 0; + lastDTS = void 0; + }; + this.reset = function () { + this.discontinuity(); + this.trigger('reset'); + }; + }; + TimestampRolloverStream$1.prototype = new Stream$6(); + var timestampRolloverStream = { + TimestampRolloverStream: TimestampRolloverStream$1, + handleRollover: handleRollover$1 + }; // Once IE11 support is dropped, this function should be removed. + + var typedArrayIndexOf$1 = (typedArray, element, fromIndex) => { + if (!typedArray) { + return -1; + } + var currentIndex = fromIndex; + for (; currentIndex < typedArray.length; currentIndex++) { + if (typedArray[currentIndex] === element) { + return currentIndex; + } + } + return -1; + }; + var typedArray = { + typedArrayIndexOf: typedArrayIndexOf$1 + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * Tools for parsing ID3 frame data + * @see http://id3.org/id3v2.3.0 + */ + + var typedArrayIndexOf = typedArray.typedArrayIndexOf, + // Frames that allow different types of text encoding contain a text + // encoding description byte [ID3v2.4.0 section 4.] + textEncodingDescriptionByte = { + Iso88591: 0x00, + // ISO-8859-1, terminated with \0. + Utf16: 0x01, + // UTF-16 encoded Unicode BOM, terminated with \0\0 + Utf16be: 0x02, + // UTF-16BE encoded Unicode, without BOM, terminated with \0\0 + Utf8: 0x03 // UTF-8 encoded Unicode, terminated with \0 + }, + // return a percent-encoded representation of the specified byte range + // @see http://en.wikipedia.org/wiki/Percent-encoding + percentEncode$1 = function (bytes, start, end) { + var i, + result = ''; + for (i = start; i < end; i++) { + result += '%' + ('00' + bytes[i].toString(16)).slice(-2); + } + return result; + }, + // return the string representation of the specified byte range, + // interpreted as UTf-8. + parseUtf8 = function (bytes, start, end) { + return decodeURIComponent(percentEncode$1(bytes, start, end)); + }, + // return the string representation of the specified byte range, + // interpreted as ISO-8859-1. + parseIso88591$1 = function (bytes, start, end) { + return unescape(percentEncode$1(bytes, start, end)); // jshint ignore:line + }, + parseSyncSafeInteger$1 = function (data) { + return data[0] << 21 | data[1] << 14 | data[2] << 7 | data[3]; + }, + frameParsers = { + 'APIC': function (frame) { + var i = 1, + mimeTypeEndIndex, + descriptionEndIndex, + LINK_MIME_TYPE = '-->'; + if (frame.data[0] !== textEncodingDescriptionByte.Utf8) { + // ignore frames with unrecognized character encodings + return; + } // parsing fields [ID3v2.4.0 section 4.14.] + + mimeTypeEndIndex = typedArrayIndexOf(frame.data, 0, i); + if (mimeTypeEndIndex < 0) { + // malformed frame + return; + } // parsing Mime type field (terminated with \0) + + frame.mimeType = parseIso88591$1(frame.data, i, mimeTypeEndIndex); + i = mimeTypeEndIndex + 1; // parsing 1-byte Picture Type field + + frame.pictureType = frame.data[i]; + i++; + descriptionEndIndex = typedArrayIndexOf(frame.data, 0, i); + if (descriptionEndIndex < 0) { + // malformed frame + return; + } // parsing Description field (terminated with \0) + + frame.description = parseUtf8(frame.data, i, descriptionEndIndex); + i = descriptionEndIndex + 1; + if (frame.mimeType === LINK_MIME_TYPE) { + // parsing Picture Data field as URL (always represented as ISO-8859-1 [ID3v2.4.0 section 4.]) + frame.url = parseIso88591$1(frame.data, i, frame.data.length); + } else { + // parsing Picture Data field as binary data + frame.pictureData = frame.data.subarray(i, frame.data.length); + } + }, + 'T*': function (frame) { + if (frame.data[0] !== textEncodingDescriptionByte.Utf8) { + // ignore frames with unrecognized character encodings + return; + } // parse text field, do not include null terminator in the frame value + // frames that allow different types of encoding contain terminated text [ID3v2.4.0 section 4.] + + frame.value = parseUtf8(frame.data, 1, frame.data.length).replace(/\0*$/, ''); // text information frames supports multiple strings, stored as a terminator separated list [ID3v2.4.0 section 4.2.] + + frame.values = frame.value.split('\0'); + }, + 'TXXX': function (frame) { + var descriptionEndIndex; + if (frame.data[0] !== textEncodingDescriptionByte.Utf8) { + // ignore frames with unrecognized character encodings + return; + } + descriptionEndIndex = typedArrayIndexOf(frame.data, 0, 1); + if (descriptionEndIndex === -1) { + return; + } // parse the text fields + + frame.description = parseUtf8(frame.data, 1, descriptionEndIndex); // do not include the null terminator in the tag value + // frames that allow different types of encoding contain terminated text + // [ID3v2.4.0 section 4.] + + frame.value = parseUtf8(frame.data, descriptionEndIndex + 1, frame.data.length).replace(/\0*$/, ''); + frame.data = frame.value; + }, + 'W*': function (frame) { + // parse URL field; URL fields are always represented as ISO-8859-1 [ID3v2.4.0 section 4.] + // if the value is followed by a string termination all the following information should be ignored [ID3v2.4.0 section 4.3] + frame.url = parseIso88591$1(frame.data, 0, frame.data.length).replace(/\0.*$/, ''); + }, + 'WXXX': function (frame) { + var descriptionEndIndex; + if (frame.data[0] !== textEncodingDescriptionByte.Utf8) { + // ignore frames with unrecognized character encodings + return; + } + descriptionEndIndex = typedArrayIndexOf(frame.data, 0, 1); + if (descriptionEndIndex === -1) { + return; + } // parse the description and URL fields + + frame.description = parseUtf8(frame.data, 1, descriptionEndIndex); // URL fields are always represented as ISO-8859-1 [ID3v2.4.0 section 4.] + // if the value is followed by a string termination all the following information + // should be ignored [ID3v2.4.0 section 4.3] + + frame.url = parseIso88591$1(frame.data, descriptionEndIndex + 1, frame.data.length).replace(/\0.*$/, ''); + }, + 'PRIV': function (frame) { + var i; + for (i = 0; i < frame.data.length; i++) { + if (frame.data[i] === 0) { + // parse the description and URL fields + frame.owner = parseIso88591$1(frame.data, 0, i); + break; + } + } + frame.privateData = frame.data.subarray(i + 1); + frame.data = frame.privateData; + } + }; + var parseId3Frames$1 = function (data) { + var frameSize, + frameHeader, + frameStart = 10, + tagSize = 0, + frames = []; // If we don't have enough data for a header, 10 bytes, + // or 'ID3' in the first 3 bytes this is not a valid ID3 tag. + + if (data.length < 10 || data[0] !== 'I'.charCodeAt(0) || data[1] !== 'D'.charCodeAt(0) || data[2] !== '3'.charCodeAt(0)) { + return; + } // the frame size is transmitted as a 28-bit integer in the + // last four bytes of the ID3 header. + // The most significant bit of each byte is dropped and the + // results concatenated to recover the actual value. + + tagSize = parseSyncSafeInteger$1(data.subarray(6, 10)); // ID3 reports the tag size excluding the header but it's more + // convenient for our comparisons to include it + + tagSize += 10; // check bit 6 of byte 5 for the extended header flag. + + var hasExtendedHeader = data[5] & 0x40; + if (hasExtendedHeader) { + // advance the frame start past the extended header + frameStart += 4; // header size field + + frameStart += parseSyncSafeInteger$1(data.subarray(10, 14)); + tagSize -= parseSyncSafeInteger$1(data.subarray(16, 20)); // clip any padding off the end + } // parse one or more ID3 frames + // http://id3.org/id3v2.3.0#ID3v2_frame_overview + + do { + // determine the number of bytes in this frame + frameSize = parseSyncSafeInteger$1(data.subarray(frameStart + 4, frameStart + 8)); + if (frameSize < 1) { + break; + } + frameHeader = String.fromCharCode(data[frameStart], data[frameStart + 1], data[frameStart + 2], data[frameStart + 3]); + var frame = { + id: frameHeader, + data: data.subarray(frameStart + 10, frameStart + frameSize + 10) + }; + frame.key = frame.id; // parse frame values + + if (frameParsers[frame.id]) { + // use frame specific parser + frameParsers[frame.id](frame); + } else if (frame.id[0] === 'T') { + // use text frame generic parser + frameParsers['T*'](frame); + } else if (frame.id[0] === 'W') { + // use URL link frame generic parser + frameParsers['W*'](frame); + } + frames.push(frame); + frameStart += 10; // advance past the frame header + + frameStart += frameSize; // advance past the frame body + } while (frameStart < tagSize); + return frames; + }; + var parseId3 = { + parseId3Frames: parseId3Frames$1, + parseSyncSafeInteger: parseSyncSafeInteger$1, + frameParsers: frameParsers + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * Accepts program elementary stream (PES) data events and parses out + * ID3 metadata from them, if present. + * @see http://id3.org/id3v2.3.0 + */ + + var Stream$5 = stream, + StreamTypes$3 = streamTypes, + id3 = parseId3, + MetadataStream; + MetadataStream = function (options) { + var settings = { + // the bytes of the program-level descriptor field in MP2T + // see ISO/IEC 13818-1:2013 (E), section 2.6 "Program and + // program element descriptors" + descriptor: options && options.descriptor + }, + // the total size in bytes of the ID3 tag being parsed + tagSize = 0, + // tag data that is not complete enough to be parsed + buffer = [], + // the total number of bytes currently in the buffer + bufferSize = 0, + i; + MetadataStream.prototype.init.call(this); // calculate the text track in-band metadata track dispatch type + // https://html.spec.whatwg.org/multipage/embedded-content.html#steps-to-expose-a-media-resource-specific-text-track + + this.dispatchType = StreamTypes$3.METADATA_STREAM_TYPE.toString(16); + if (settings.descriptor) { + for (i = 0; i < settings.descriptor.length; i++) { + this.dispatchType += ('00' + settings.descriptor[i].toString(16)).slice(-2); + } + } + this.push = function (chunk) { + var tag, frameStart, frameSize, frame, i, frameHeader; + if (chunk.type !== 'timed-metadata') { + return; + } // if data_alignment_indicator is set in the PES header, + // we must have the start of a new ID3 tag. Assume anything + // remaining in the buffer was malformed and throw it out + + if (chunk.dataAlignmentIndicator) { + bufferSize = 0; + buffer.length = 0; + } // ignore events that don't look like ID3 data + + if (buffer.length === 0 && (chunk.data.length < 10 || chunk.data[0] !== 'I'.charCodeAt(0) || chunk.data[1] !== 'D'.charCodeAt(0) || chunk.data[2] !== '3'.charCodeAt(0))) { + this.trigger('log', { + level: 'warn', + message: 'Skipping unrecognized metadata packet' + }); + return; + } // add this chunk to the data we've collected so far + + buffer.push(chunk); + bufferSize += chunk.data.byteLength; // grab the size of the entire frame from the ID3 header + + if (buffer.length === 1) { + // the frame size is transmitted as a 28-bit integer in the + // last four bytes of the ID3 header. + // The most significant bit of each byte is dropped and the + // results concatenated to recover the actual value. + tagSize = id3.parseSyncSafeInteger(chunk.data.subarray(6, 10)); // ID3 reports the tag size excluding the header but it's more + // convenient for our comparisons to include it + + tagSize += 10; + } // if the entire frame has not arrived, wait for more data + + if (bufferSize < tagSize) { + return; + } // collect the entire frame so it can be parsed + + tag = { + data: new Uint8Array(tagSize), + frames: [], + pts: buffer[0].pts, + dts: buffer[0].dts + }; + for (i = 0; i < tagSize;) { + tag.data.set(buffer[0].data.subarray(0, tagSize - i), i); + i += buffer[0].data.byteLength; + bufferSize -= buffer[0].data.byteLength; + buffer.shift(); + } // find the start of the first frame and the end of the tag + + frameStart = 10; + if (tag.data[5] & 0x40) { + // advance the frame start past the extended header + frameStart += 4; // header size field + + frameStart += id3.parseSyncSafeInteger(tag.data.subarray(10, 14)); // clip any padding off the end + + tagSize -= id3.parseSyncSafeInteger(tag.data.subarray(16, 20)); + } // parse one or more ID3 frames + // http://id3.org/id3v2.3.0#ID3v2_frame_overview + + do { + // determine the number of bytes in this frame + frameSize = id3.parseSyncSafeInteger(tag.data.subarray(frameStart + 4, frameStart + 8)); + if (frameSize < 1) { + this.trigger('log', { + level: 'warn', + message: 'Malformed ID3 frame encountered. Skipping remaining metadata parsing.' + }); // If the frame is malformed, don't parse any further frames but allow previous valid parsed frames + // to be sent along. + + break; + } + frameHeader = String.fromCharCode(tag.data[frameStart], tag.data[frameStart + 1], tag.data[frameStart + 2], tag.data[frameStart + 3]); + frame = { + id: frameHeader, + data: tag.data.subarray(frameStart + 10, frameStart + frameSize + 10) + }; + frame.key = frame.id; // parse frame values + + if (id3.frameParsers[frame.id]) { + // use frame specific parser + id3.frameParsers[frame.id](frame); + } else if (frame.id[0] === 'T') { + // use text frame generic parser + id3.frameParsers['T*'](frame); + } else if (frame.id[0] === 'W') { + // use URL link frame generic parser + id3.frameParsers['W*'](frame); + } // handle the special PRIV frame used to indicate the start + // time for raw AAC data + + if (frame.owner === 'com.apple.streaming.transportStreamTimestamp') { + var d = frame.data, + size = (d[3] & 0x01) << 30 | d[4] << 22 | d[5] << 14 | d[6] << 6 | d[7] >>> 2; + size *= 4; + size += d[7] & 0x03; + frame.timeStamp = size; // in raw AAC, all subsequent data will be timestamped based + // on the value of this frame + // we couldn't have known the appropriate pts and dts before + // parsing this ID3 tag so set those values now + + if (tag.pts === undefined && tag.dts === undefined) { + tag.pts = frame.timeStamp; + tag.dts = frame.timeStamp; + } + this.trigger('timestamp', frame); + } + tag.frames.push(frame); + frameStart += 10; // advance past the frame header + + frameStart += frameSize; // advance past the frame body + } while (frameStart < tagSize); + this.trigger('data', tag); + }; + }; + MetadataStream.prototype = new Stream$5(); + var metadataStream = MetadataStream; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * A stream-based mp2t to mp4 converter. This utility can be used to + * deliver mp4s to a SourceBuffer on platforms that support native + * Media Source Extensions. + */ + + var Stream$4 = stream, + CaptionStream$1 = captionStream, + StreamTypes$2 = streamTypes, + TimestampRolloverStream = timestampRolloverStream.TimestampRolloverStream; // object types + + var TransportPacketStream, TransportParseStream, ElementaryStream; // constants + + var MP2T_PACKET_LENGTH$1 = 188, + // bytes + SYNC_BYTE$1 = 0x47; + /** + * Splits an incoming stream of binary data into MPEG-2 Transport + * Stream packets. + */ + + TransportPacketStream = function () { + var buffer = new Uint8Array(MP2T_PACKET_LENGTH$1), + bytesInBuffer = 0; + TransportPacketStream.prototype.init.call(this); // Deliver new bytes to the stream. + + /** + * Split a stream of data into M2TS packets + **/ + + this.push = function (bytes) { + var startIndex = 0, + endIndex = MP2T_PACKET_LENGTH$1, + everything; // If there are bytes remaining from the last segment, prepend them to the + // bytes that were pushed in + + if (bytesInBuffer) { + everything = new Uint8Array(bytes.byteLength + bytesInBuffer); + everything.set(buffer.subarray(0, bytesInBuffer)); + everything.set(bytes, bytesInBuffer); + bytesInBuffer = 0; + } else { + everything = bytes; + } // While we have enough data for a packet + + while (endIndex < everything.byteLength) { + // Look for a pair of start and end sync bytes in the data.. + if (everything[startIndex] === SYNC_BYTE$1 && everything[endIndex] === SYNC_BYTE$1) { + // We found a packet so emit it and jump one whole packet forward in + // the stream + this.trigger('data', everything.subarray(startIndex, endIndex)); + startIndex += MP2T_PACKET_LENGTH$1; + endIndex += MP2T_PACKET_LENGTH$1; + continue; + } // If we get here, we have somehow become de-synchronized and we need to step + // forward one byte at a time until we find a pair of sync bytes that denote + // a packet + + startIndex++; + endIndex++; + } // If there was some data left over at the end of the segment that couldn't + // possibly be a whole packet, keep it because it might be the start of a packet + // that continues in the next segment + + if (startIndex < everything.byteLength) { + buffer.set(everything.subarray(startIndex), 0); + bytesInBuffer = everything.byteLength - startIndex; + } + }; + /** + * Passes identified M2TS packets to the TransportParseStream to be parsed + **/ + + this.flush = function () { + // If the buffer contains a whole packet when we are being flushed, emit it + // and empty the buffer. Otherwise hold onto the data because it may be + // important for decoding the next segment + if (bytesInBuffer === MP2T_PACKET_LENGTH$1 && buffer[0] === SYNC_BYTE$1) { + this.trigger('data', buffer); + bytesInBuffer = 0; + } + this.trigger('done'); + }; + this.endTimeline = function () { + this.flush(); + this.trigger('endedtimeline'); + }; + this.reset = function () { + bytesInBuffer = 0; + this.trigger('reset'); + }; + }; + TransportPacketStream.prototype = new Stream$4(); + /** + * Accepts an MP2T TransportPacketStream and emits data events with parsed + * forms of the individual transport stream packets. + */ + + TransportParseStream = function () { + var parsePsi, parsePat, parsePmt, self; + TransportParseStream.prototype.init.call(this); + self = this; + this.packetsWaitingForPmt = []; + this.programMapTable = undefined; + parsePsi = function (payload, psi) { + var offset = 0; // PSI packets may be split into multiple sections and those + // sections may be split into multiple packets. If a PSI + // section starts in this packet, the payload_unit_start_indicator + // will be true and the first byte of the payload will indicate + // the offset from the current position to the start of the + // section. + + if (psi.payloadUnitStartIndicator) { + offset += payload[offset] + 1; + } + if (psi.type === 'pat') { + parsePat(payload.subarray(offset), psi); + } else { + parsePmt(payload.subarray(offset), psi); + } + }; + parsePat = function (payload, pat) { + pat.section_number = payload[7]; // eslint-disable-line camelcase + + pat.last_section_number = payload[8]; // eslint-disable-line camelcase + // skip the PSI header and parse the first PMT entry + + self.pmtPid = (payload[10] & 0x1F) << 8 | payload[11]; + pat.pmtPid = self.pmtPid; + }; + /** + * Parse out the relevant fields of a Program Map Table (PMT). + * @param payload {Uint8Array} the PMT-specific portion of an MP2T + * packet. The first byte in this array should be the table_id + * field. + * @param pmt {object} the object that should be decorated with + * fields parsed from the PMT. + */ + + parsePmt = function (payload, pmt) { + var sectionLength, tableEnd, programInfoLength, offset; // PMTs can be sent ahead of the time when they should actually + // take effect. We don't believe this should ever be the case + // for HLS but we'll ignore "forward" PMT declarations if we see + // them. Future PMT declarations have the current_next_indicator + // set to zero. + + if (!(payload[5] & 0x01)) { + return; + } // overwrite any existing program map table + + self.programMapTable = { + video: null, + audio: null, + 'timed-metadata': {} + }; // the mapping table ends at the end of the current section + + sectionLength = (payload[1] & 0x0f) << 8 | payload[2]; + tableEnd = 3 + sectionLength - 4; // to determine where the table is, we have to figure out how + // long the program info descriptors are + + programInfoLength = (payload[10] & 0x0f) << 8 | payload[11]; // advance the offset to the first entry in the mapping table + + offset = 12 + programInfoLength; + while (offset < tableEnd) { + var streamType = payload[offset]; + var pid = (payload[offset + 1] & 0x1F) << 8 | payload[offset + 2]; // only map a single elementary_pid for audio and video stream types + // TODO: should this be done for metadata too? for now maintain behavior of + // multiple metadata streams + + if (streamType === StreamTypes$2.H264_STREAM_TYPE && self.programMapTable.video === null) { + self.programMapTable.video = pid; + } else if (streamType === StreamTypes$2.ADTS_STREAM_TYPE && self.programMapTable.audio === null) { + self.programMapTable.audio = pid; + } else if (streamType === StreamTypes$2.METADATA_STREAM_TYPE) { + // map pid to stream type for metadata streams + self.programMapTable['timed-metadata'][pid] = streamType; + } // move to the next table entry + // skip past the elementary stream descriptors, if present + + offset += ((payload[offset + 3] & 0x0F) << 8 | payload[offset + 4]) + 5; + } // record the map on the packet as well + + pmt.programMapTable = self.programMapTable; + }; + /** + * Deliver a new MP2T packet to the next stream in the pipeline. + */ + + this.push = function (packet) { + var result = {}, + offset = 4; + result.payloadUnitStartIndicator = !!(packet[1] & 0x40); // pid is a 13-bit field starting at the last bit of packet[1] + + result.pid = packet[1] & 0x1f; + result.pid <<= 8; + result.pid |= packet[2]; // if an adaption field is present, its length is specified by the + // fifth byte of the TS packet header. The adaptation field is + // used to add stuffing to PES packets that don't fill a complete + // TS packet, and to specify some forms of timing and control data + // that we do not currently use. + + if ((packet[3] & 0x30) >>> 4 > 0x01) { + offset += packet[offset] + 1; + } // parse the rest of the packet based on the type + + if (result.pid === 0) { + result.type = 'pat'; + parsePsi(packet.subarray(offset), result); + this.trigger('data', result); + } else if (result.pid === this.pmtPid) { + result.type = 'pmt'; + parsePsi(packet.subarray(offset), result); + this.trigger('data', result); // if there are any packets waiting for a PMT to be found, process them now + + while (this.packetsWaitingForPmt.length) { + this.processPes_.apply(this, this.packetsWaitingForPmt.shift()); + } + } else if (this.programMapTable === undefined) { + // When we have not seen a PMT yet, defer further processing of + // PES packets until one has been parsed + this.packetsWaitingForPmt.push([packet, offset, result]); + } else { + this.processPes_(packet, offset, result); + } + }; + this.processPes_ = function (packet, offset, result) { + // set the appropriate stream type + if (result.pid === this.programMapTable.video) { + result.streamType = StreamTypes$2.H264_STREAM_TYPE; + } else if (result.pid === this.programMapTable.audio) { + result.streamType = StreamTypes$2.ADTS_STREAM_TYPE; + } else { + // if not video or audio, it is timed-metadata or unknown + // if unknown, streamType will be undefined + result.streamType = this.programMapTable['timed-metadata'][result.pid]; + } + result.type = 'pes'; + result.data = packet.subarray(offset); + this.trigger('data', result); + }; + }; + TransportParseStream.prototype = new Stream$4(); + TransportParseStream.STREAM_TYPES = { + h264: 0x1b, + adts: 0x0f + }; + /** + * Reconsistutes program elementary stream (PES) packets from parsed + * transport stream packets. That is, if you pipe an + * mp2t.TransportParseStream into a mp2t.ElementaryStream, the output + * events will be events which capture the bytes for individual PES + * packets plus relevant metadata that has been extracted from the + * container. + */ + + ElementaryStream = function () { + var self = this, + segmentHadPmt = false, + // PES packet fragments + video = { + data: [], + size: 0 + }, + audio = { + data: [], + size: 0 + }, + timedMetadata = { + data: [], + size: 0 + }, + programMapTable, + parsePes = function (payload, pes) { + var ptsDtsFlags; + const startPrefix = payload[0] << 16 | payload[1] << 8 | payload[2]; // default to an empty array + + pes.data = new Uint8Array(); // In certain live streams, the start of a TS fragment has ts packets + // that are frame data that is continuing from the previous fragment. This + // is to check that the pes data is the start of a new pes payload + + if (startPrefix !== 1) { + return; + } // get the packet length, this will be 0 for video + + pes.packetLength = 6 + (payload[4] << 8 | payload[5]); // find out if this packets starts a new keyframe + + pes.dataAlignmentIndicator = (payload[6] & 0x04) !== 0; // PES packets may be annotated with a PTS value, or a PTS value + // and a DTS value. Determine what combination of values is + // available to work with. + + ptsDtsFlags = payload[7]; // PTS and DTS are normally stored as a 33-bit number. Javascript + // performs all bitwise operations on 32-bit integers but javascript + // supports a much greater range (52-bits) of integer using standard + // mathematical operations. + // We construct a 31-bit value using bitwise operators over the 31 + // most significant bits and then multiply by 4 (equal to a left-shift + // of 2) before we add the final 2 least significant bits of the + // timestamp (equal to an OR.) + + if (ptsDtsFlags & 0xC0) { + // the PTS and DTS are not written out directly. For information + // on how they are encoded, see + // http://dvd.sourceforge.net/dvdinfo/pes-hdr.html + pes.pts = (payload[9] & 0x0E) << 27 | (payload[10] & 0xFF) << 20 | (payload[11] & 0xFE) << 12 | (payload[12] & 0xFF) << 5 | (payload[13] & 0xFE) >>> 3; + pes.pts *= 4; // Left shift by 2 + + pes.pts += (payload[13] & 0x06) >>> 1; // OR by the two LSBs + + pes.dts = pes.pts; + if (ptsDtsFlags & 0x40) { + pes.dts = (payload[14] & 0x0E) << 27 | (payload[15] & 0xFF) << 20 | (payload[16] & 0xFE) << 12 | (payload[17] & 0xFF) << 5 | (payload[18] & 0xFE) >>> 3; + pes.dts *= 4; // Left shift by 2 + + pes.dts += (payload[18] & 0x06) >>> 1; // OR by the two LSBs + } + } // the data section starts immediately after the PES header. + // pes_header_data_length specifies the number of header bytes + // that follow the last byte of the field. + + pes.data = payload.subarray(9 + payload[8]); + }, + /** + * Pass completely parsed PES packets to the next stream in the pipeline + **/ + flushStream = function (stream, type, forceFlush) { + var packetData = new Uint8Array(stream.size), + event = { + type: type + }, + i = 0, + offset = 0, + packetFlushable = false, + fragment; // do nothing if there is not enough buffered data for a complete + // PES header + + if (!stream.data.length || stream.size < 9) { + return; + } + event.trackId = stream.data[0].pid; // reassemble the packet + + for (i = 0; i < stream.data.length; i++) { + fragment = stream.data[i]; + packetData.set(fragment.data, offset); + offset += fragment.data.byteLength; + } // parse assembled packet's PES header + + parsePes(packetData, event); // non-video PES packets MUST have a non-zero PES_packet_length + // check that there is enough stream data to fill the packet + + packetFlushable = type === 'video' || event.packetLength <= stream.size; // flush pending packets if the conditions are right + + if (forceFlush || packetFlushable) { + stream.size = 0; + stream.data.length = 0; + } // only emit packets that are complete. this is to avoid assembling + // incomplete PES packets due to poor segmentation + + if (packetFlushable) { + self.trigger('data', event); + } + }; + ElementaryStream.prototype.init.call(this); + /** + * Identifies M2TS packet types and parses PES packets using metadata + * parsed from the PMT + **/ + + this.push = function (data) { + ({ + pat: function () {// we have to wait for the PMT to arrive as well before we + // have any meaningful metadata + }, + pes: function () { + var stream, streamType; + switch (data.streamType) { + case StreamTypes$2.H264_STREAM_TYPE: + stream = video; + streamType = 'video'; + break; + case StreamTypes$2.ADTS_STREAM_TYPE: + stream = audio; + streamType = 'audio'; + break; + case StreamTypes$2.METADATA_STREAM_TYPE: + stream = timedMetadata; + streamType = 'timed-metadata'; + break; + default: + // ignore unknown stream types + return; + } // if a new packet is starting, we can flush the completed + // packet + + if (data.payloadUnitStartIndicator) { + flushStream(stream, streamType, true); + } // buffer this fragment until we are sure we've received the + // complete payload + + stream.data.push(data); + stream.size += data.data.byteLength; + }, + pmt: function () { + var event = { + type: 'metadata', + tracks: [] + }; + programMapTable = data.programMapTable; // translate audio and video streams to tracks + + if (programMapTable.video !== null) { + event.tracks.push({ + timelineStartInfo: { + baseMediaDecodeTime: 0 + }, + id: +programMapTable.video, + codec: 'avc', + type: 'video' + }); + } + if (programMapTable.audio !== null) { + event.tracks.push({ + timelineStartInfo: { + baseMediaDecodeTime: 0 + }, + id: +programMapTable.audio, + codec: 'adts', + type: 'audio' + }); + } + segmentHadPmt = true; + self.trigger('data', event); + } + })[data.type](); + }; + this.reset = function () { + video.size = 0; + video.data.length = 0; + audio.size = 0; + audio.data.length = 0; + this.trigger('reset'); + }; + /** + * Flush any remaining input. Video PES packets may be of variable + * length. Normally, the start of a new video packet can trigger the + * finalization of the previous packet. That is not possible if no + * more video is forthcoming, however. In that case, some other + * mechanism (like the end of the file) has to be employed. When it is + * clear that no additional data is forthcoming, calling this method + * will flush the buffered packets. + */ + + this.flushStreams_ = function () { + // !!THIS ORDER IS IMPORTANT!! + // video first then audio + flushStream(video, 'video'); + flushStream(audio, 'audio'); + flushStream(timedMetadata, 'timed-metadata'); + }; + this.flush = function () { + // if on flush we haven't had a pmt emitted + // and we have a pmt to emit. emit the pmt + // so that we trigger a trackinfo downstream. + if (!segmentHadPmt && programMapTable) { + var pmt = { + type: 'metadata', + tracks: [] + }; // translate audio and video streams to tracks + + if (programMapTable.video !== null) { + pmt.tracks.push({ + timelineStartInfo: { + baseMediaDecodeTime: 0 + }, + id: +programMapTable.video, + codec: 'avc', + type: 'video' + }); + } + if (programMapTable.audio !== null) { + pmt.tracks.push({ + timelineStartInfo: { + baseMediaDecodeTime: 0 + }, + id: +programMapTable.audio, + codec: 'adts', + type: 'audio' + }); + } + self.trigger('data', pmt); + } + segmentHadPmt = false; + this.flushStreams_(); + this.trigger('done'); + }; + }; + ElementaryStream.prototype = new Stream$4(); + var m2ts$1 = { + PAT_PID: 0x0000, + MP2T_PACKET_LENGTH: MP2T_PACKET_LENGTH$1, + TransportPacketStream: TransportPacketStream, + TransportParseStream: TransportParseStream, + ElementaryStream: ElementaryStream, + TimestampRolloverStream: TimestampRolloverStream, + CaptionStream: CaptionStream$1.CaptionStream, + Cea608Stream: CaptionStream$1.Cea608Stream, + Cea708Stream: CaptionStream$1.Cea708Stream, + MetadataStream: metadataStream + }; + for (var type in StreamTypes$2) { + if (StreamTypes$2.hasOwnProperty(type)) { + m2ts$1[type] = StreamTypes$2[type]; + } + } + var m2ts_1 = m2ts$1; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + */ + + var Stream$3 = stream; + var ONE_SECOND_IN_TS$2 = clock$2.ONE_SECOND_IN_TS; + var AdtsStream$1; + var ADTS_SAMPLING_FREQUENCIES$1 = [96000, 88200, 64000, 48000, 44100, 32000, 24000, 22050, 16000, 12000, 11025, 8000, 7350]; + /* + * Accepts a ElementaryStream and emits data events with parsed + * AAC Audio Frames of the individual packets. Input audio in ADTS + * format is unpacked and re-emitted as AAC frames. + * + * @see http://wiki.multimedia.cx/index.php?title=ADTS + * @see http://wiki.multimedia.cx/?title=Understanding_AAC + */ + + AdtsStream$1 = function (handlePartialSegments) { + var buffer, + frameNum = 0; + AdtsStream$1.prototype.init.call(this); + this.skipWarn_ = function (start, end) { + this.trigger('log', { + level: 'warn', + message: `adts skiping bytes ${start} to ${end} in frame ${frameNum} outside syncword` + }); + }; + this.push = function (packet) { + var i = 0, + frameLength, + protectionSkipBytes, + oldBuffer, + sampleCount, + adtsFrameDuration; + if (!handlePartialSegments) { + frameNum = 0; + } + if (packet.type !== 'audio') { + // ignore non-audio data + return; + } // Prepend any data in the buffer to the input data so that we can parse + // aac frames the cross a PES packet boundary + + if (buffer && buffer.length) { + oldBuffer = buffer; + buffer = new Uint8Array(oldBuffer.byteLength + packet.data.byteLength); + buffer.set(oldBuffer); + buffer.set(packet.data, oldBuffer.byteLength); + } else { + buffer = packet.data; + } // unpack any ADTS frames which have been fully received + // for details on the ADTS header, see http://wiki.multimedia.cx/index.php?title=ADTS + + var skip; // We use i + 7 here because we want to be able to parse the entire header. + // If we don't have enough bytes to do that, then we definitely won't have a full frame. + + while (i + 7 < buffer.length) { + // Look for the start of an ADTS header.. + if (buffer[i] !== 0xFF || (buffer[i + 1] & 0xF6) !== 0xF0) { + if (typeof skip !== 'number') { + skip = i; + } // If a valid header was not found, jump one forward and attempt to + // find a valid ADTS header starting at the next byte + + i++; + continue; + } + if (typeof skip === 'number') { + this.skipWarn_(skip, i); + skip = null; + } // The protection skip bit tells us if we have 2 bytes of CRC data at the + // end of the ADTS header + + protectionSkipBytes = (~buffer[i + 1] & 0x01) * 2; // Frame length is a 13 bit integer starting 16 bits from the + // end of the sync sequence + // NOTE: frame length includes the size of the header + + frameLength = (buffer[i + 3] & 0x03) << 11 | buffer[i + 4] << 3 | (buffer[i + 5] & 0xe0) >> 5; + sampleCount = ((buffer[i + 6] & 0x03) + 1) * 1024; + adtsFrameDuration = sampleCount * ONE_SECOND_IN_TS$2 / ADTS_SAMPLING_FREQUENCIES$1[(buffer[i + 2] & 0x3c) >>> 2]; // If we don't have enough data to actually finish this ADTS frame, + // then we have to wait for more data + + if (buffer.byteLength - i < frameLength) { + break; + } // Otherwise, deliver the complete AAC frame + + this.trigger('data', { + pts: packet.pts + frameNum * adtsFrameDuration, + dts: packet.dts + frameNum * adtsFrameDuration, + sampleCount: sampleCount, + audioobjecttype: (buffer[i + 2] >>> 6 & 0x03) + 1, + channelcount: (buffer[i + 2] & 1) << 2 | (buffer[i + 3] & 0xc0) >>> 6, + samplerate: ADTS_SAMPLING_FREQUENCIES$1[(buffer[i + 2] & 0x3c) >>> 2], + samplingfrequencyindex: (buffer[i + 2] & 0x3c) >>> 2, + // assume ISO/IEC 14496-12 AudioSampleEntry default of 16 + samplesize: 16, + // data is the frame without it's header + data: buffer.subarray(i + 7 + protectionSkipBytes, i + frameLength) + }); + frameNum++; + i += frameLength; + } + if (typeof skip === 'number') { + this.skipWarn_(skip, i); + skip = null; + } // remove processed bytes from the buffer. + + buffer = buffer.subarray(i); + }; + this.flush = function () { + frameNum = 0; + this.trigger('done'); + }; + this.reset = function () { + buffer = void 0; + this.trigger('reset'); + }; + this.endTimeline = function () { + buffer = void 0; + this.trigger('endedtimeline'); + }; + }; + AdtsStream$1.prototype = new Stream$3(); + var adts = AdtsStream$1; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + */ + + var ExpGolomb$1; + /** + * Parser for exponential Golomb codes, a variable-bitwidth number encoding + * scheme used by h264. + */ + + ExpGolomb$1 = function (workingData) { + var + // the number of bytes left to examine in workingData + workingBytesAvailable = workingData.byteLength, + // the current word being examined + workingWord = 0, + // :uint + // the number of bits left to examine in the current word + workingBitsAvailable = 0; // :uint; + // ():uint + + this.length = function () { + return 8 * workingBytesAvailable; + }; // ():uint + + this.bitsAvailable = function () { + return 8 * workingBytesAvailable + workingBitsAvailable; + }; // ():void + + this.loadWord = function () { + var position = workingData.byteLength - workingBytesAvailable, + workingBytes = new Uint8Array(4), + availableBytes = Math.min(4, workingBytesAvailable); + if (availableBytes === 0) { + throw new Error('no bytes available'); + } + workingBytes.set(workingData.subarray(position, position + availableBytes)); + workingWord = new DataView(workingBytes.buffer).getUint32(0); // track the amount of workingData that has been processed + + workingBitsAvailable = availableBytes * 8; + workingBytesAvailable -= availableBytes; + }; // (count:int):void + + this.skipBits = function (count) { + var skipBytes; // :int + + if (workingBitsAvailable > count) { + workingWord <<= count; + workingBitsAvailable -= count; + } else { + count -= workingBitsAvailable; + skipBytes = Math.floor(count / 8); + count -= skipBytes * 8; + workingBytesAvailable -= skipBytes; + this.loadWord(); + workingWord <<= count; + workingBitsAvailable -= count; + } + }; // (size:int):uint + + this.readBits = function (size) { + var bits = Math.min(workingBitsAvailable, size), + // :uint + valu = workingWord >>> 32 - bits; // :uint + // if size > 31, handle error + + workingBitsAvailable -= bits; + if (workingBitsAvailable > 0) { + workingWord <<= bits; + } else if (workingBytesAvailable > 0) { + this.loadWord(); + } + bits = size - bits; + if (bits > 0) { + return valu << bits | this.readBits(bits); + } + return valu; + }; // ():uint + + this.skipLeadingZeros = function () { + var leadingZeroCount; // :uint + + for (leadingZeroCount = 0; leadingZeroCount < workingBitsAvailable; ++leadingZeroCount) { + if ((workingWord & 0x80000000 >>> leadingZeroCount) !== 0) { + // the first bit of working word is 1 + workingWord <<= leadingZeroCount; + workingBitsAvailable -= leadingZeroCount; + return leadingZeroCount; + } + } // we exhausted workingWord and still have not found a 1 + + this.loadWord(); + return leadingZeroCount + this.skipLeadingZeros(); + }; // ():void + + this.skipUnsignedExpGolomb = function () { + this.skipBits(1 + this.skipLeadingZeros()); + }; // ():void + + this.skipExpGolomb = function () { + this.skipBits(1 + this.skipLeadingZeros()); + }; // ():uint + + this.readUnsignedExpGolomb = function () { + var clz = this.skipLeadingZeros(); // :uint + + return this.readBits(clz + 1) - 1; + }; // ():int + + this.readExpGolomb = function () { + var valu = this.readUnsignedExpGolomb(); // :int + + if (0x01 & valu) { + // the number is odd if the low order bit is set + return 1 + valu >>> 1; // add 1 to make it even, and divide by 2 + } + + return -1 * (valu >>> 1); // divide by two then make it negative + }; // Some convenience functions + // :Boolean + + this.readBoolean = function () { + return this.readBits(1) === 1; + }; // ():int + + this.readUnsignedByte = function () { + return this.readBits(8); + }; + this.loadWord(); + }; + var expGolomb = ExpGolomb$1; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + */ + + var Stream$2 = stream; + var ExpGolomb = expGolomb; + var H264Stream$1, NalByteStream; + var PROFILES_WITH_OPTIONAL_SPS_DATA; + /** + * Accepts a NAL unit byte stream and unpacks the embedded NAL units. + */ + + NalByteStream = function () { + var syncPoint = 0, + i, + buffer; + NalByteStream.prototype.init.call(this); + /* + * Scans a byte stream and triggers a data event with the NAL units found. + * @param {Object} data Event received from H264Stream + * @param {Uint8Array} data.data The h264 byte stream to be scanned + * + * @see H264Stream.push + */ + + this.push = function (data) { + var swapBuffer; + if (!buffer) { + buffer = data.data; + } else { + swapBuffer = new Uint8Array(buffer.byteLength + data.data.byteLength); + swapBuffer.set(buffer); + swapBuffer.set(data.data, buffer.byteLength); + buffer = swapBuffer; + } + var len = buffer.byteLength; // Rec. ITU-T H.264, Annex B + // scan for NAL unit boundaries + // a match looks like this: + // 0 0 1 .. NAL .. 0 0 1 + // ^ sync point ^ i + // or this: + // 0 0 1 .. NAL .. 0 0 0 + // ^ sync point ^ i + // advance the sync point to a NAL start, if necessary + + for (; syncPoint < len - 3; syncPoint++) { + if (buffer[syncPoint + 2] === 1) { + // the sync point is properly aligned + i = syncPoint + 5; + break; + } + } + while (i < len) { + // look at the current byte to determine if we've hit the end of + // a NAL unit boundary + switch (buffer[i]) { + case 0: + // skip past non-sync sequences + if (buffer[i - 1] !== 0) { + i += 2; + break; + } else if (buffer[i - 2] !== 0) { + i++; + break; + } // deliver the NAL unit if it isn't empty + + if (syncPoint + 3 !== i - 2) { + this.trigger('data', buffer.subarray(syncPoint + 3, i - 2)); + } // drop trailing zeroes + + do { + i++; + } while (buffer[i] !== 1 && i < len); + syncPoint = i - 2; + i += 3; + break; + case 1: + // skip past non-sync sequences + if (buffer[i - 1] !== 0 || buffer[i - 2] !== 0) { + i += 3; + break; + } // deliver the NAL unit + + this.trigger('data', buffer.subarray(syncPoint + 3, i - 2)); + syncPoint = i - 2; + i += 3; + break; + default: + // the current byte isn't a one or zero, so it cannot be part + // of a sync sequence + i += 3; + break; + } + } // filter out the NAL units that were delivered + + buffer = buffer.subarray(syncPoint); + i -= syncPoint; + syncPoint = 0; + }; + this.reset = function () { + buffer = null; + syncPoint = 0; + this.trigger('reset'); + }; + this.flush = function () { + // deliver the last buffered NAL unit + if (buffer && buffer.byteLength > 3) { + this.trigger('data', buffer.subarray(syncPoint + 3)); + } // reset the stream state + + buffer = null; + syncPoint = 0; + this.trigger('done'); + }; + this.endTimeline = function () { + this.flush(); + this.trigger('endedtimeline'); + }; + }; + NalByteStream.prototype = new Stream$2(); // values of profile_idc that indicate additional fields are included in the SPS + // see Recommendation ITU-T H.264 (4/2013), + // 7.3.2.1.1 Sequence parameter set data syntax + + PROFILES_WITH_OPTIONAL_SPS_DATA = { + 100: true, + 110: true, + 122: true, + 244: true, + 44: true, + 83: true, + 86: true, + 118: true, + 128: true, + // TODO: the three profiles below don't + // appear to have sps data in the specificiation anymore? + 138: true, + 139: true, + 134: true + }; + /** + * Accepts input from a ElementaryStream and produces H.264 NAL unit data + * events. + */ + + H264Stream$1 = function () { + var nalByteStream = new NalByteStream(), + self, + trackId, + currentPts, + currentDts, + discardEmulationPreventionBytes, + readSequenceParameterSet, + skipScalingList; + H264Stream$1.prototype.init.call(this); + self = this; + /* + * Pushes a packet from a stream onto the NalByteStream + * + * @param {Object} packet - A packet received from a stream + * @param {Uint8Array} packet.data - The raw bytes of the packet + * @param {Number} packet.dts - Decode timestamp of the packet + * @param {Number} packet.pts - Presentation timestamp of the packet + * @param {Number} packet.trackId - The id of the h264 track this packet came from + * @param {('video'|'audio')} packet.type - The type of packet + * + */ + + this.push = function (packet) { + if (packet.type !== 'video') { + return; + } + trackId = packet.trackId; + currentPts = packet.pts; + currentDts = packet.dts; + nalByteStream.push(packet); + }; + /* + * Identify NAL unit types and pass on the NALU, trackId, presentation and decode timestamps + * for the NALUs to the next stream component. + * Also, preprocess caption and sequence parameter NALUs. + * + * @param {Uint8Array} data - A NAL unit identified by `NalByteStream.push` + * @see NalByteStream.push + */ + + nalByteStream.on('data', function (data) { + var event = { + trackId: trackId, + pts: currentPts, + dts: currentDts, + data: data, + nalUnitTypeCode: data[0] & 0x1f + }; + switch (event.nalUnitTypeCode) { + case 0x05: + event.nalUnitType = 'slice_layer_without_partitioning_rbsp_idr'; + break; + case 0x06: + event.nalUnitType = 'sei_rbsp'; + event.escapedRBSP = discardEmulationPreventionBytes(data.subarray(1)); + break; + case 0x07: + event.nalUnitType = 'seq_parameter_set_rbsp'; + event.escapedRBSP = discardEmulationPreventionBytes(data.subarray(1)); + event.config = readSequenceParameterSet(event.escapedRBSP); + break; + case 0x08: + event.nalUnitType = 'pic_parameter_set_rbsp'; + break; + case 0x09: + event.nalUnitType = 'access_unit_delimiter_rbsp'; + break; + } // This triggers data on the H264Stream + + self.trigger('data', event); + }); + nalByteStream.on('done', function () { + self.trigger('done'); + }); + nalByteStream.on('partialdone', function () { + self.trigger('partialdone'); + }); + nalByteStream.on('reset', function () { + self.trigger('reset'); + }); + nalByteStream.on('endedtimeline', function () { + self.trigger('endedtimeline'); + }); + this.flush = function () { + nalByteStream.flush(); + }; + this.partialFlush = function () { + nalByteStream.partialFlush(); + }; + this.reset = function () { + nalByteStream.reset(); + }; + this.endTimeline = function () { + nalByteStream.endTimeline(); + }; + /** + * Advance the ExpGolomb decoder past a scaling list. The scaling + * list is optionally transmitted as part of a sequence parameter + * set and is not relevant to transmuxing. + * @param count {number} the number of entries in this scaling list + * @param expGolombDecoder {object} an ExpGolomb pointed to the + * start of a scaling list + * @see Recommendation ITU-T H.264, Section 7.3.2.1.1.1 + */ + + skipScalingList = function (count, expGolombDecoder) { + var lastScale = 8, + nextScale = 8, + j, + deltaScale; + for (j = 0; j < count; j++) { + if (nextScale !== 0) { + deltaScale = expGolombDecoder.readExpGolomb(); + nextScale = (lastScale + deltaScale + 256) % 256; + } + lastScale = nextScale === 0 ? lastScale : nextScale; + } + }; + /** + * Expunge any "Emulation Prevention" bytes from a "Raw Byte + * Sequence Payload" + * @param data {Uint8Array} the bytes of a RBSP from a NAL + * unit + * @return {Uint8Array} the RBSP without any Emulation + * Prevention Bytes + */ + + discardEmulationPreventionBytes = function (data) { + var length = data.byteLength, + emulationPreventionBytesPositions = [], + i = 1, + newLength, + newData; // Find all `Emulation Prevention Bytes` + + while (i < length - 2) { + if (data[i] === 0 && data[i + 1] === 0 && data[i + 2] === 0x03) { + emulationPreventionBytesPositions.push(i + 2); + i += 2; + } else { + i++; + } + } // If no Emulation Prevention Bytes were found just return the original + // array + + if (emulationPreventionBytesPositions.length === 0) { + return data; + } // Create a new array to hold the NAL unit data + + newLength = length - emulationPreventionBytesPositions.length; + newData = new Uint8Array(newLength); + var sourceIndex = 0; + for (i = 0; i < newLength; sourceIndex++, i++) { + if (sourceIndex === emulationPreventionBytesPositions[0]) { + // Skip this byte + sourceIndex++; // Remove this position index + + emulationPreventionBytesPositions.shift(); + } + newData[i] = data[sourceIndex]; + } + return newData; + }; + /** + * Read a sequence parameter set and return some interesting video + * properties. A sequence parameter set is the H264 metadata that + * describes the properties of upcoming video frames. + * @param data {Uint8Array} the bytes of a sequence parameter set + * @return {object} an object with configuration parsed from the + * sequence parameter set, including the dimensions of the + * associated video frames. + */ + + readSequenceParameterSet = function (data) { + var frameCropLeftOffset = 0, + frameCropRightOffset = 0, + frameCropTopOffset = 0, + frameCropBottomOffset = 0, + expGolombDecoder, + profileIdc, + levelIdc, + profileCompatibility, + chromaFormatIdc, + picOrderCntType, + numRefFramesInPicOrderCntCycle, + picWidthInMbsMinus1, + picHeightInMapUnitsMinus1, + frameMbsOnlyFlag, + scalingListCount, + sarRatio = [1, 1], + aspectRatioIdc, + i; + expGolombDecoder = new ExpGolomb(data); + profileIdc = expGolombDecoder.readUnsignedByte(); // profile_idc + + profileCompatibility = expGolombDecoder.readUnsignedByte(); // constraint_set[0-5]_flag + + levelIdc = expGolombDecoder.readUnsignedByte(); // level_idc u(8) + + expGolombDecoder.skipUnsignedExpGolomb(); // seq_parameter_set_id + // some profiles have more optional data we don't need + + if (PROFILES_WITH_OPTIONAL_SPS_DATA[profileIdc]) { + chromaFormatIdc = expGolombDecoder.readUnsignedExpGolomb(); + if (chromaFormatIdc === 3) { + expGolombDecoder.skipBits(1); // separate_colour_plane_flag + } + + expGolombDecoder.skipUnsignedExpGolomb(); // bit_depth_luma_minus8 + + expGolombDecoder.skipUnsignedExpGolomb(); // bit_depth_chroma_minus8 + + expGolombDecoder.skipBits(1); // qpprime_y_zero_transform_bypass_flag + + if (expGolombDecoder.readBoolean()) { + // seq_scaling_matrix_present_flag + scalingListCount = chromaFormatIdc !== 3 ? 8 : 12; + for (i = 0; i < scalingListCount; i++) { + if (expGolombDecoder.readBoolean()) { + // seq_scaling_list_present_flag[ i ] + if (i < 6) { + skipScalingList(16, expGolombDecoder); + } else { + skipScalingList(64, expGolombDecoder); + } + } + } + } + } + expGolombDecoder.skipUnsignedExpGolomb(); // log2_max_frame_num_minus4 + + picOrderCntType = expGolombDecoder.readUnsignedExpGolomb(); + if (picOrderCntType === 0) { + expGolombDecoder.readUnsignedExpGolomb(); // log2_max_pic_order_cnt_lsb_minus4 + } else if (picOrderCntType === 1) { + expGolombDecoder.skipBits(1); // delta_pic_order_always_zero_flag + + expGolombDecoder.skipExpGolomb(); // offset_for_non_ref_pic + + expGolombDecoder.skipExpGolomb(); // offset_for_top_to_bottom_field + + numRefFramesInPicOrderCntCycle = expGolombDecoder.readUnsignedExpGolomb(); + for (i = 0; i < numRefFramesInPicOrderCntCycle; i++) { + expGolombDecoder.skipExpGolomb(); // offset_for_ref_frame[ i ] + } + } + + expGolombDecoder.skipUnsignedExpGolomb(); // max_num_ref_frames + + expGolombDecoder.skipBits(1); // gaps_in_frame_num_value_allowed_flag + + picWidthInMbsMinus1 = expGolombDecoder.readUnsignedExpGolomb(); + picHeightInMapUnitsMinus1 = expGolombDecoder.readUnsignedExpGolomb(); + frameMbsOnlyFlag = expGolombDecoder.readBits(1); + if (frameMbsOnlyFlag === 0) { + expGolombDecoder.skipBits(1); // mb_adaptive_frame_field_flag + } + + expGolombDecoder.skipBits(1); // direct_8x8_inference_flag + + if (expGolombDecoder.readBoolean()) { + // frame_cropping_flag + frameCropLeftOffset = expGolombDecoder.readUnsignedExpGolomb(); + frameCropRightOffset = expGolombDecoder.readUnsignedExpGolomb(); + frameCropTopOffset = expGolombDecoder.readUnsignedExpGolomb(); + frameCropBottomOffset = expGolombDecoder.readUnsignedExpGolomb(); + } + if (expGolombDecoder.readBoolean()) { + // vui_parameters_present_flag + if (expGolombDecoder.readBoolean()) { + // aspect_ratio_info_present_flag + aspectRatioIdc = expGolombDecoder.readUnsignedByte(); + switch (aspectRatioIdc) { + case 1: + sarRatio = [1, 1]; + break; + case 2: + sarRatio = [12, 11]; + break; + case 3: + sarRatio = [10, 11]; + break; + case 4: + sarRatio = [16, 11]; + break; + case 5: + sarRatio = [40, 33]; + break; + case 6: + sarRatio = [24, 11]; + break; + case 7: + sarRatio = [20, 11]; + break; + case 8: + sarRatio = [32, 11]; + break; + case 9: + sarRatio = [80, 33]; + break; + case 10: + sarRatio = [18, 11]; + break; + case 11: + sarRatio = [15, 11]; + break; + case 12: + sarRatio = [64, 33]; + break; + case 13: + sarRatio = [160, 99]; + break; + case 14: + sarRatio = [4, 3]; + break; + case 15: + sarRatio = [3, 2]; + break; + case 16: + sarRatio = [2, 1]; + break; + case 255: + { + sarRatio = [expGolombDecoder.readUnsignedByte() << 8 | expGolombDecoder.readUnsignedByte(), expGolombDecoder.readUnsignedByte() << 8 | expGolombDecoder.readUnsignedByte()]; + break; + } + } + if (sarRatio) { + sarRatio[0] / sarRatio[1]; + } + } + } + return { + profileIdc: profileIdc, + levelIdc: levelIdc, + profileCompatibility: profileCompatibility, + width: (picWidthInMbsMinus1 + 1) * 16 - frameCropLeftOffset * 2 - frameCropRightOffset * 2, + height: (2 - frameMbsOnlyFlag) * (picHeightInMapUnitsMinus1 + 1) * 16 - frameCropTopOffset * 2 - frameCropBottomOffset * 2, + // sar is sample aspect ratio + sarRatio: sarRatio + }; + }; + }; + H264Stream$1.prototype = new Stream$2(); + var h264 = { + H264Stream: H264Stream$1, + NalByteStream: NalByteStream + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * Utilities to detect basic properties and metadata about Aac data. + */ + + var ADTS_SAMPLING_FREQUENCIES = [96000, 88200, 64000, 48000, 44100, 32000, 24000, 22050, 16000, 12000, 11025, 8000, 7350]; + var parseId3TagSize = function (header, byteIndex) { + var returnSize = header[byteIndex + 6] << 21 | header[byteIndex + 7] << 14 | header[byteIndex + 8] << 7 | header[byteIndex + 9], + flags = header[byteIndex + 5], + footerPresent = (flags & 16) >> 4; // if we get a negative returnSize clamp it to 0 + + returnSize = returnSize >= 0 ? returnSize : 0; + if (footerPresent) { + return returnSize + 20; + } + return returnSize + 10; + }; + var getId3Offset = function (data, offset) { + if (data.length - offset < 10 || data[offset] !== 'I'.charCodeAt(0) || data[offset + 1] !== 'D'.charCodeAt(0) || data[offset + 2] !== '3'.charCodeAt(0)) { + return offset; + } + offset += parseId3TagSize(data, offset); + return getId3Offset(data, offset); + }; // TODO: use vhs-utils + + var isLikelyAacData$1 = function (data) { + var offset = getId3Offset(data, 0); + return data.length >= offset + 2 && (data[offset] & 0xFF) === 0xFF && (data[offset + 1] & 0xF0) === 0xF0 && + // verify that the 2 layer bits are 0, aka this + // is not mp3 data but aac data. + (data[offset + 1] & 0x16) === 0x10; + }; + var parseSyncSafeInteger = function (data) { + return data[0] << 21 | data[1] << 14 | data[2] << 7 | data[3]; + }; // return a percent-encoded representation of the specified byte range + // @see http://en.wikipedia.org/wiki/Percent-encoding + + var percentEncode = function (bytes, start, end) { + var i, + result = ''; + for (i = start; i < end; i++) { + result += '%' + ('00' + bytes[i].toString(16)).slice(-2); + } + return result; + }; // return the string representation of the specified byte range, + // interpreted as ISO-8859-1. + + var parseIso88591 = function (bytes, start, end) { + return unescape(percentEncode(bytes, start, end)); // jshint ignore:line + }; + + var parseAdtsSize = function (header, byteIndex) { + var lowThree = (header[byteIndex + 5] & 0xE0) >> 5, + middle = header[byteIndex + 4] << 3, + highTwo = header[byteIndex + 3] & 0x3 << 11; + return highTwo | middle | lowThree; + }; + var parseType$4 = function (header, byteIndex) { + if (header[byteIndex] === 'I'.charCodeAt(0) && header[byteIndex + 1] === 'D'.charCodeAt(0) && header[byteIndex + 2] === '3'.charCodeAt(0)) { + return 'timed-metadata'; + } else if (header[byteIndex] & 0xff === 0xff && (header[byteIndex + 1] & 0xf0) === 0xf0) { + return 'audio'; + } + return null; + }; + var parseSampleRate = function (packet) { + var i = 0; + while (i + 5 < packet.length) { + if (packet[i] !== 0xFF || (packet[i + 1] & 0xF6) !== 0xF0) { + // If a valid header was not found, jump one forward and attempt to + // find a valid ADTS header starting at the next byte + i++; + continue; + } + return ADTS_SAMPLING_FREQUENCIES[(packet[i + 2] & 0x3c) >>> 2]; + } + return null; + }; + var parseAacTimestamp = function (packet) { + var frameStart, frameSize, frame, frameHeader; // find the start of the first frame and the end of the tag + + frameStart = 10; + if (packet[5] & 0x40) { + // advance the frame start past the extended header + frameStart += 4; // header size field + + frameStart += parseSyncSafeInteger(packet.subarray(10, 14)); + } // parse one or more ID3 frames + // http://id3.org/id3v2.3.0#ID3v2_frame_overview + + do { + // determine the number of bytes in this frame + frameSize = parseSyncSafeInteger(packet.subarray(frameStart + 4, frameStart + 8)); + if (frameSize < 1) { + return null; + } + frameHeader = String.fromCharCode(packet[frameStart], packet[frameStart + 1], packet[frameStart + 2], packet[frameStart + 3]); + if (frameHeader === 'PRIV') { + frame = packet.subarray(frameStart + 10, frameStart + frameSize + 10); + for (var i = 0; i < frame.byteLength; i++) { + if (frame[i] === 0) { + var owner = parseIso88591(frame, 0, i); + if (owner === 'com.apple.streaming.transportStreamTimestamp') { + var d = frame.subarray(i + 1); + var size = (d[3] & 0x01) << 30 | d[4] << 22 | d[5] << 14 | d[6] << 6 | d[7] >>> 2; + size *= 4; + size += d[7] & 0x03; + return size; + } + break; + } + } + } + frameStart += 10; // advance past the frame header + + frameStart += frameSize; // advance past the frame body + } while (frameStart < packet.byteLength); + return null; + }; + var utils = { + isLikelyAacData: isLikelyAacData$1, + parseId3TagSize: parseId3TagSize, + parseAdtsSize: parseAdtsSize, + parseType: parseType$4, + parseSampleRate: parseSampleRate, + parseAacTimestamp: parseAacTimestamp + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * A stream-based aac to mp4 converter. This utility can be used to + * deliver mp4s to a SourceBuffer on platforms that support native + * Media Source Extensions. + */ + + var Stream$1 = stream; + var aacUtils = utils; // Constants + + var AacStream$1; + /** + * Splits an incoming stream of binary data into ADTS and ID3 Frames. + */ + + AacStream$1 = function () { + var everything = new Uint8Array(), + timeStamp = 0; + AacStream$1.prototype.init.call(this); + this.setTimestamp = function (timestamp) { + timeStamp = timestamp; + }; + this.push = function (bytes) { + var frameSize = 0, + byteIndex = 0, + bytesLeft, + chunk, + packet, + tempLength; // If there are bytes remaining from the last segment, prepend them to the + // bytes that were pushed in + + if (everything.length) { + tempLength = everything.length; + everything = new Uint8Array(bytes.byteLength + tempLength); + everything.set(everything.subarray(0, tempLength)); + everything.set(bytes, tempLength); + } else { + everything = bytes; + } + while (everything.length - byteIndex >= 3) { + if (everything[byteIndex] === 'I'.charCodeAt(0) && everything[byteIndex + 1] === 'D'.charCodeAt(0) && everything[byteIndex + 2] === '3'.charCodeAt(0)) { + // Exit early because we don't have enough to parse + // the ID3 tag header + if (everything.length - byteIndex < 10) { + break; + } // check framesize + + frameSize = aacUtils.parseId3TagSize(everything, byteIndex); // Exit early if we don't have enough in the buffer + // to emit a full packet + // Add to byteIndex to support multiple ID3 tags in sequence + + if (byteIndex + frameSize > everything.length) { + break; + } + chunk = { + type: 'timed-metadata', + data: everything.subarray(byteIndex, byteIndex + frameSize) + }; + this.trigger('data', chunk); + byteIndex += frameSize; + continue; + } else if ((everything[byteIndex] & 0xff) === 0xff && (everything[byteIndex + 1] & 0xf0) === 0xf0) { + // Exit early because we don't have enough to parse + // the ADTS frame header + if (everything.length - byteIndex < 7) { + break; + } + frameSize = aacUtils.parseAdtsSize(everything, byteIndex); // Exit early if we don't have enough in the buffer + // to emit a full packet + + if (byteIndex + frameSize > everything.length) { + break; + } + packet = { + type: 'audio', + data: everything.subarray(byteIndex, byteIndex + frameSize), + pts: timeStamp, + dts: timeStamp + }; + this.trigger('data', packet); + byteIndex += frameSize; + continue; + } + byteIndex++; + } + bytesLeft = everything.length - byteIndex; + if (bytesLeft > 0) { + everything = everything.subarray(byteIndex); + } else { + everything = new Uint8Array(); + } + }; + this.reset = function () { + everything = new Uint8Array(); + this.trigger('reset'); + }; + this.endTimeline = function () { + everything = new Uint8Array(); + this.trigger('endedtimeline'); + }; + }; + AacStream$1.prototype = new Stream$1(); + var aac = AacStream$1; + var AUDIO_PROPERTIES$1 = ['audioobjecttype', 'channelcount', 'samplerate', 'samplingfrequencyindex', 'samplesize']; + var audioProperties = AUDIO_PROPERTIES$1; + var VIDEO_PROPERTIES$1 = ['width', 'height', 'profileIdc', 'levelIdc', 'profileCompatibility', 'sarRatio']; + var videoProperties = VIDEO_PROPERTIES$1; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * A stream-based mp2t to mp4 converter. This utility can be used to + * deliver mp4s to a SourceBuffer on platforms that support native + * Media Source Extensions. + */ + + var Stream = stream; + var mp4 = mp4Generator; + var frameUtils = frameUtils$1; + var audioFrameUtils = audioFrameUtils$1; + var trackDecodeInfo = trackDecodeInfo$1; + var m2ts = m2ts_1; + var clock = clock$2; + var AdtsStream = adts; + var H264Stream = h264.H264Stream; + var AacStream = aac; + var isLikelyAacData = utils.isLikelyAacData; + var ONE_SECOND_IN_TS$1 = clock$2.ONE_SECOND_IN_TS; + var AUDIO_PROPERTIES = audioProperties; + var VIDEO_PROPERTIES = videoProperties; // object types + + var VideoSegmentStream, AudioSegmentStream, Transmuxer, CoalesceStream; + var retriggerForStream = function (key, event) { + event.stream = key; + this.trigger('log', event); + }; + var addPipelineLogRetriggers = function (transmuxer, pipeline) { + var keys = Object.keys(pipeline); + for (var i = 0; i < keys.length; i++) { + var key = keys[i]; // skip non-stream keys and headOfPipeline + // which is just a duplicate + + if (key === 'headOfPipeline' || !pipeline[key].on) { + continue; + } + pipeline[key].on('log', retriggerForStream.bind(transmuxer, key)); + } + }; + /** + * Compare two arrays (even typed) for same-ness + */ + + var arrayEquals = function (a, b) { + var i; + if (a.length !== b.length) { + return false; + } // compare the value of each element in the array + + for (i = 0; i < a.length; i++) { + if (a[i] !== b[i]) { + return false; + } + } + return true; + }; + var generateSegmentTimingInfo = function (baseMediaDecodeTime, startDts, startPts, endDts, endPts, prependedContentDuration) { + var ptsOffsetFromDts = startPts - startDts, + decodeDuration = endDts - startDts, + presentationDuration = endPts - startPts; // The PTS and DTS values are based on the actual stream times from the segment, + // however, the player time values will reflect a start from the baseMediaDecodeTime. + // In order to provide relevant values for the player times, base timing info on the + // baseMediaDecodeTime and the DTS and PTS durations of the segment. + + return { + start: { + dts: baseMediaDecodeTime, + pts: baseMediaDecodeTime + ptsOffsetFromDts + }, + end: { + dts: baseMediaDecodeTime + decodeDuration, + pts: baseMediaDecodeTime + presentationDuration + }, + prependedContentDuration: prependedContentDuration, + baseMediaDecodeTime: baseMediaDecodeTime + }; + }; + /** + * Constructs a single-track, ISO BMFF media segment from AAC data + * events. The output of this stream can be fed to a SourceBuffer + * configured with a suitable initialization segment. + * @param track {object} track metadata configuration + * @param options {object} transmuxer options object + * @param options.keepOriginalTimestamps {boolean} If true, keep the timestamps + * in the source; false to adjust the first segment to start at 0. + */ + + AudioSegmentStream = function (track, options) { + var adtsFrames = [], + sequenceNumber, + earliestAllowedDts = 0, + audioAppendStartTs = 0, + videoBaseMediaDecodeTime = Infinity; + options = options || {}; + sequenceNumber = options.firstSequenceNumber || 0; + AudioSegmentStream.prototype.init.call(this); + this.push = function (data) { + trackDecodeInfo.collectDtsInfo(track, data); + if (track) { + AUDIO_PROPERTIES.forEach(function (prop) { + track[prop] = data[prop]; + }); + } // buffer audio data until end() is called + + adtsFrames.push(data); + }; + this.setEarliestDts = function (earliestDts) { + earliestAllowedDts = earliestDts; + }; + this.setVideoBaseMediaDecodeTime = function (baseMediaDecodeTime) { + videoBaseMediaDecodeTime = baseMediaDecodeTime; + }; + this.setAudioAppendStart = function (timestamp) { + audioAppendStartTs = timestamp; + }; + this.flush = function () { + var frames, moof, mdat, boxes, frameDuration, segmentDuration, videoClockCyclesOfSilencePrefixed; // return early if no audio data has been observed + + if (adtsFrames.length === 0) { + this.trigger('done', 'AudioSegmentStream'); + return; + } + frames = audioFrameUtils.trimAdtsFramesByEarliestDts(adtsFrames, track, earliestAllowedDts); + track.baseMediaDecodeTime = trackDecodeInfo.calculateTrackBaseMediaDecodeTime(track, options.keepOriginalTimestamps); // amount of audio filled but the value is in video clock rather than audio clock + + videoClockCyclesOfSilencePrefixed = audioFrameUtils.prefixWithSilence(track, frames, audioAppendStartTs, videoBaseMediaDecodeTime); // we have to build the index from byte locations to + // samples (that is, adts frames) in the audio data + + track.samples = audioFrameUtils.generateSampleTable(frames); // concatenate the audio data to constuct the mdat + + mdat = mp4.mdat(audioFrameUtils.concatenateFrameData(frames)); + adtsFrames = []; + moof = mp4.moof(sequenceNumber, [track]); + boxes = new Uint8Array(moof.byteLength + mdat.byteLength); // bump the sequence number for next time + + sequenceNumber++; + boxes.set(moof); + boxes.set(mdat, moof.byteLength); + trackDecodeInfo.clearDtsInfo(track); + frameDuration = Math.ceil(ONE_SECOND_IN_TS$1 * 1024 / track.samplerate); // TODO this check was added to maintain backwards compatibility (particularly with + // tests) on adding the timingInfo event. However, it seems unlikely that there's a + // valid use-case where an init segment/data should be triggered without associated + // frames. Leaving for now, but should be looked into. + + if (frames.length) { + segmentDuration = frames.length * frameDuration; + this.trigger('segmentTimingInfo', generateSegmentTimingInfo( + // The audio track's baseMediaDecodeTime is in audio clock cycles, but the + // frame info is in video clock cycles. Convert to match expectation of + // listeners (that all timestamps will be based on video clock cycles). + clock.audioTsToVideoTs(track.baseMediaDecodeTime, track.samplerate), + // frame times are already in video clock, as is segment duration + frames[0].dts, frames[0].pts, frames[0].dts + segmentDuration, frames[0].pts + segmentDuration, videoClockCyclesOfSilencePrefixed || 0)); + this.trigger('timingInfo', { + start: frames[0].pts, + end: frames[0].pts + segmentDuration + }); + } + this.trigger('data', { + track: track, + boxes: boxes + }); + this.trigger('done', 'AudioSegmentStream'); + }; + this.reset = function () { + trackDecodeInfo.clearDtsInfo(track); + adtsFrames = []; + this.trigger('reset'); + }; + }; + AudioSegmentStream.prototype = new Stream(); + /** + * Constructs a single-track, ISO BMFF media segment from H264 data + * events. The output of this stream can be fed to a SourceBuffer + * configured with a suitable initialization segment. + * @param track {object} track metadata configuration + * @param options {object} transmuxer options object + * @param options.alignGopsAtEnd {boolean} If true, start from the end of the + * gopsToAlignWith list when attempting to align gop pts + * @param options.keepOriginalTimestamps {boolean} If true, keep the timestamps + * in the source; false to adjust the first segment to start at 0. + */ + + VideoSegmentStream = function (track, options) { + var sequenceNumber, + nalUnits = [], + gopsToAlignWith = [], + config, + pps; + options = options || {}; + sequenceNumber = options.firstSequenceNumber || 0; + VideoSegmentStream.prototype.init.call(this); + delete track.minPTS; + this.gopCache_ = []; + /** + * Constructs a ISO BMFF segment given H264 nalUnits + * @param {Object} nalUnit A data event representing a nalUnit + * @param {String} nalUnit.nalUnitType + * @param {Object} nalUnit.config Properties for a mp4 track + * @param {Uint8Array} nalUnit.data The nalUnit bytes + * @see lib/codecs/h264.js + **/ + + this.push = function (nalUnit) { + trackDecodeInfo.collectDtsInfo(track, nalUnit); // record the track config + + if (nalUnit.nalUnitType === 'seq_parameter_set_rbsp' && !config) { + config = nalUnit.config; + track.sps = [nalUnit.data]; + VIDEO_PROPERTIES.forEach(function (prop) { + track[prop] = config[prop]; + }, this); + } + if (nalUnit.nalUnitType === 'pic_parameter_set_rbsp' && !pps) { + pps = nalUnit.data; + track.pps = [nalUnit.data]; + } // buffer video until flush() is called + + nalUnits.push(nalUnit); + }; + /** + * Pass constructed ISO BMFF track and boxes on to the + * next stream in the pipeline + **/ + + this.flush = function () { + var frames, + gopForFusion, + gops, + moof, + mdat, + boxes, + prependedContentDuration = 0, + firstGop, + lastGop; // Throw away nalUnits at the start of the byte stream until + // we find the first AUD + + while (nalUnits.length) { + if (nalUnits[0].nalUnitType === 'access_unit_delimiter_rbsp') { + break; + } + nalUnits.shift(); + } // Return early if no video data has been observed + + if (nalUnits.length === 0) { + this.resetStream_(); + this.trigger('done', 'VideoSegmentStream'); + return; + } // Organize the raw nal-units into arrays that represent + // higher-level constructs such as frames and gops + // (group-of-pictures) + + frames = frameUtils.groupNalsIntoFrames(nalUnits); + gops = frameUtils.groupFramesIntoGops(frames); // If the first frame of this fragment is not a keyframe we have + // a problem since MSE (on Chrome) requires a leading keyframe. + // + // We have two approaches to repairing this situation: + // 1) GOP-FUSION: + // This is where we keep track of the GOPS (group-of-pictures) + // from previous fragments and attempt to find one that we can + // prepend to the current fragment in order to create a valid + // fragment. + // 2) KEYFRAME-PULLING: + // Here we search for the first keyframe in the fragment and + // throw away all the frames between the start of the fragment + // and that keyframe. We then extend the duration and pull the + // PTS of the keyframe forward so that it covers the time range + // of the frames that were disposed of. + // + // #1 is far prefereable over #2 which can cause "stuttering" but + // requires more things to be just right. + + if (!gops[0][0].keyFrame) { + // Search for a gop for fusion from our gopCache + gopForFusion = this.getGopForFusion_(nalUnits[0], track); + if (gopForFusion) { + // in order to provide more accurate timing information about the segment, save + // the number of seconds prepended to the original segment due to GOP fusion + prependedContentDuration = gopForFusion.duration; + gops.unshift(gopForFusion); // Adjust Gops' metadata to account for the inclusion of the + // new gop at the beginning + + gops.byteLength += gopForFusion.byteLength; + gops.nalCount += gopForFusion.nalCount; + gops.pts = gopForFusion.pts; + gops.dts = gopForFusion.dts; + gops.duration += gopForFusion.duration; + } else { + // If we didn't find a candidate gop fall back to keyframe-pulling + gops = frameUtils.extendFirstKeyFrame(gops); + } + } // Trim gops to align with gopsToAlignWith + + if (gopsToAlignWith.length) { + var alignedGops; + if (options.alignGopsAtEnd) { + alignedGops = this.alignGopsAtEnd_(gops); + } else { + alignedGops = this.alignGopsAtStart_(gops); + } + if (!alignedGops) { + // save all the nals in the last GOP into the gop cache + this.gopCache_.unshift({ + gop: gops.pop(), + pps: track.pps, + sps: track.sps + }); // Keep a maximum of 6 GOPs in the cache + + this.gopCache_.length = Math.min(6, this.gopCache_.length); // Clear nalUnits + + nalUnits = []; // return early no gops can be aligned with desired gopsToAlignWith + + this.resetStream_(); + this.trigger('done', 'VideoSegmentStream'); + return; + } // Some gops were trimmed. clear dts info so minSegmentDts and pts are correct + // when recalculated before sending off to CoalesceStream + + trackDecodeInfo.clearDtsInfo(track); + gops = alignedGops; + } + trackDecodeInfo.collectDtsInfo(track, gops); // First, we have to build the index from byte locations to + // samples (that is, frames) in the video data + + track.samples = frameUtils.generateSampleTable(gops); // Concatenate the video data and construct the mdat + + mdat = mp4.mdat(frameUtils.concatenateNalData(gops)); + track.baseMediaDecodeTime = trackDecodeInfo.calculateTrackBaseMediaDecodeTime(track, options.keepOriginalTimestamps); + this.trigger('processedGopsInfo', gops.map(function (gop) { + return { + pts: gop.pts, + dts: gop.dts, + byteLength: gop.byteLength + }; + })); + firstGop = gops[0]; + lastGop = gops[gops.length - 1]; + this.trigger('segmentTimingInfo', generateSegmentTimingInfo(track.baseMediaDecodeTime, firstGop.dts, firstGop.pts, lastGop.dts + lastGop.duration, lastGop.pts + lastGop.duration, prependedContentDuration)); + this.trigger('timingInfo', { + start: gops[0].pts, + end: gops[gops.length - 1].pts + gops[gops.length - 1].duration + }); // save all the nals in the last GOP into the gop cache + + this.gopCache_.unshift({ + gop: gops.pop(), + pps: track.pps, + sps: track.sps + }); // Keep a maximum of 6 GOPs in the cache + + this.gopCache_.length = Math.min(6, this.gopCache_.length); // Clear nalUnits + + nalUnits = []; + this.trigger('baseMediaDecodeTime', track.baseMediaDecodeTime); + this.trigger('timelineStartInfo', track.timelineStartInfo); + moof = mp4.moof(sequenceNumber, [track]); // it would be great to allocate this array up front instead of + // throwing away hundreds of media segment fragments + + boxes = new Uint8Array(moof.byteLength + mdat.byteLength); // Bump the sequence number for next time + + sequenceNumber++; + boxes.set(moof); + boxes.set(mdat, moof.byteLength); + this.trigger('data', { + track: track, + boxes: boxes + }); + this.resetStream_(); // Continue with the flush process now + + this.trigger('done', 'VideoSegmentStream'); + }; + this.reset = function () { + this.resetStream_(); + nalUnits = []; + this.gopCache_.length = 0; + gopsToAlignWith.length = 0; + this.trigger('reset'); + }; + this.resetStream_ = function () { + trackDecodeInfo.clearDtsInfo(track); // reset config and pps because they may differ across segments + // for instance, when we are rendition switching + + config = undefined; + pps = undefined; + }; // Search for a candidate Gop for gop-fusion from the gop cache and + // return it or return null if no good candidate was found + + this.getGopForFusion_ = function (nalUnit) { + var halfSecond = 45000, + // Half-a-second in a 90khz clock + allowableOverlap = 10000, + // About 3 frames @ 30fps + nearestDistance = Infinity, + dtsDistance, + nearestGopObj, + currentGop, + currentGopObj, + i; // Search for the GOP nearest to the beginning of this nal unit + + for (i = 0; i < this.gopCache_.length; i++) { + currentGopObj = this.gopCache_[i]; + currentGop = currentGopObj.gop; // Reject Gops with different SPS or PPS + + if (!(track.pps && arrayEquals(track.pps[0], currentGopObj.pps[0])) || !(track.sps && arrayEquals(track.sps[0], currentGopObj.sps[0]))) { + continue; + } // Reject Gops that would require a negative baseMediaDecodeTime + + if (currentGop.dts < track.timelineStartInfo.dts) { + continue; + } // The distance between the end of the gop and the start of the nalUnit + + dtsDistance = nalUnit.dts - currentGop.dts - currentGop.duration; // Only consider GOPS that start before the nal unit and end within + // a half-second of the nal unit + + if (dtsDistance >= -allowableOverlap && dtsDistance <= halfSecond) { + // Always use the closest GOP we found if there is more than + // one candidate + if (!nearestGopObj || nearestDistance > dtsDistance) { + nearestGopObj = currentGopObj; + nearestDistance = dtsDistance; + } + } + } + if (nearestGopObj) { + return nearestGopObj.gop; + } + return null; + }; // trim gop list to the first gop found that has a matching pts with a gop in the list + // of gopsToAlignWith starting from the START of the list + + this.alignGopsAtStart_ = function (gops) { + var alignIndex, gopIndex, align, gop, byteLength, nalCount, duration, alignedGops; + byteLength = gops.byteLength; + nalCount = gops.nalCount; + duration = gops.duration; + alignIndex = gopIndex = 0; + while (alignIndex < gopsToAlignWith.length && gopIndex < gops.length) { + align = gopsToAlignWith[alignIndex]; + gop = gops[gopIndex]; + if (align.pts === gop.pts) { + break; + } + if (gop.pts > align.pts) { + // this current gop starts after the current gop we want to align on, so increment + // align index + alignIndex++; + continue; + } // current gop starts before the current gop we want to align on. so increment gop + // index + + gopIndex++; + byteLength -= gop.byteLength; + nalCount -= gop.nalCount; + duration -= gop.duration; + } + if (gopIndex === 0) { + // no gops to trim + return gops; + } + if (gopIndex === gops.length) { + // all gops trimmed, skip appending all gops + return null; + } + alignedGops = gops.slice(gopIndex); + alignedGops.byteLength = byteLength; + alignedGops.duration = duration; + alignedGops.nalCount = nalCount; + alignedGops.pts = alignedGops[0].pts; + alignedGops.dts = alignedGops[0].dts; + return alignedGops; + }; // trim gop list to the first gop found that has a matching pts with a gop in the list + // of gopsToAlignWith starting from the END of the list + + this.alignGopsAtEnd_ = function (gops) { + var alignIndex, gopIndex, align, gop, alignEndIndex, matchFound; + alignIndex = gopsToAlignWith.length - 1; + gopIndex = gops.length - 1; + alignEndIndex = null; + matchFound = false; + while (alignIndex >= 0 && gopIndex >= 0) { + align = gopsToAlignWith[alignIndex]; + gop = gops[gopIndex]; + if (align.pts === gop.pts) { + matchFound = true; + break; + } + if (align.pts > gop.pts) { + alignIndex--; + continue; + } + if (alignIndex === gopsToAlignWith.length - 1) { + // gop.pts is greater than the last alignment candidate. If no match is found + // by the end of this loop, we still want to append gops that come after this + // point + alignEndIndex = gopIndex; + } + gopIndex--; + } + if (!matchFound && alignEndIndex === null) { + return null; + } + var trimIndex; + if (matchFound) { + trimIndex = gopIndex; + } else { + trimIndex = alignEndIndex; + } + if (trimIndex === 0) { + return gops; + } + var alignedGops = gops.slice(trimIndex); + var metadata = alignedGops.reduce(function (total, gop) { + total.byteLength += gop.byteLength; + total.duration += gop.duration; + total.nalCount += gop.nalCount; + return total; + }, { + byteLength: 0, + duration: 0, + nalCount: 0 + }); + alignedGops.byteLength = metadata.byteLength; + alignedGops.duration = metadata.duration; + alignedGops.nalCount = metadata.nalCount; + alignedGops.pts = alignedGops[0].pts; + alignedGops.dts = alignedGops[0].dts; + return alignedGops; + }; + this.alignGopsWith = function (newGopsToAlignWith) { + gopsToAlignWith = newGopsToAlignWith; + }; + }; + VideoSegmentStream.prototype = new Stream(); + /** + * A Stream that can combine multiple streams (ie. audio & video) + * into a single output segment for MSE. Also supports audio-only + * and video-only streams. + * @param options {object} transmuxer options object + * @param options.keepOriginalTimestamps {boolean} If true, keep the timestamps + * in the source; false to adjust the first segment to start at media timeline start. + */ + + CoalesceStream = function (options, metadataStream) { + // Number of Tracks per output segment + // If greater than 1, we combine multiple + // tracks into a single segment + this.numberOfTracks = 0; + this.metadataStream = metadataStream; + options = options || {}; + if (typeof options.remux !== 'undefined') { + this.remuxTracks = !!options.remux; + } else { + this.remuxTracks = true; + } + if (typeof options.keepOriginalTimestamps === 'boolean') { + this.keepOriginalTimestamps = options.keepOriginalTimestamps; + } else { + this.keepOriginalTimestamps = false; + } + this.pendingTracks = []; + this.videoTrack = null; + this.pendingBoxes = []; + this.pendingCaptions = []; + this.pendingMetadata = []; + this.pendingBytes = 0; + this.emittedTracks = 0; + CoalesceStream.prototype.init.call(this); // Take output from multiple + + this.push = function (output) { + // buffer incoming captions until the associated video segment + // finishes + if (output.content || output.text) { + return this.pendingCaptions.push(output); + } // buffer incoming id3 tags until the final flush + + if (output.frames) { + return this.pendingMetadata.push(output); + } // Add this track to the list of pending tracks and store + // important information required for the construction of + // the final segment + + this.pendingTracks.push(output.track); + this.pendingBytes += output.boxes.byteLength; // TODO: is there an issue for this against chrome? + // We unshift audio and push video because + // as of Chrome 75 when switching from + // one init segment to another if the video + // mdat does not appear after the audio mdat + // only audio will play for the duration of our transmux. + + if (output.track.type === 'video') { + this.videoTrack = output.track; + this.pendingBoxes.push(output.boxes); + } + if (output.track.type === 'audio') { + this.audioTrack = output.track; + this.pendingBoxes.unshift(output.boxes); + } + }; + }; + CoalesceStream.prototype = new Stream(); + CoalesceStream.prototype.flush = function (flushSource) { + var offset = 0, + event = { + captions: [], + captionStreams: {}, + metadata: [], + info: {} + }, + caption, + id3, + initSegment, + timelineStartPts = 0, + i; + if (this.pendingTracks.length < this.numberOfTracks) { + if (flushSource !== 'VideoSegmentStream' && flushSource !== 'AudioSegmentStream') { + // Return because we haven't received a flush from a data-generating + // portion of the segment (meaning that we have only recieved meta-data + // or captions.) + return; + } else if (this.remuxTracks) { + // Return until we have enough tracks from the pipeline to remux (if we + // are remuxing audio and video into a single MP4) + return; + } else if (this.pendingTracks.length === 0) { + // In the case where we receive a flush without any data having been + // received we consider it an emitted track for the purposes of coalescing + // `done` events. + // We do this for the case where there is an audio and video track in the + // segment but no audio data. (seen in several playlists with alternate + // audio tracks and no audio present in the main TS segments.) + this.emittedTracks++; + if (this.emittedTracks >= this.numberOfTracks) { + this.trigger('done'); + this.emittedTracks = 0; + } + return; + } + } + if (this.videoTrack) { + timelineStartPts = this.videoTrack.timelineStartInfo.pts; + VIDEO_PROPERTIES.forEach(function (prop) { + event.info[prop] = this.videoTrack[prop]; + }, this); + } else if (this.audioTrack) { + timelineStartPts = this.audioTrack.timelineStartInfo.pts; + AUDIO_PROPERTIES.forEach(function (prop) { + event.info[prop] = this.audioTrack[prop]; + }, this); + } + if (this.videoTrack || this.audioTrack) { + if (this.pendingTracks.length === 1) { + event.type = this.pendingTracks[0].type; + } else { + event.type = 'combined'; + } + this.emittedTracks += this.pendingTracks.length; + initSegment = mp4.initSegment(this.pendingTracks); // Create a new typed array to hold the init segment + + event.initSegment = new Uint8Array(initSegment.byteLength); // Create an init segment containing a moov + // and track definitions + + event.initSegment.set(initSegment); // Create a new typed array to hold the moof+mdats + + event.data = new Uint8Array(this.pendingBytes); // Append each moof+mdat (one per track) together + + for (i = 0; i < this.pendingBoxes.length; i++) { + event.data.set(this.pendingBoxes[i], offset); + offset += this.pendingBoxes[i].byteLength; + } // Translate caption PTS times into second offsets to match the + // video timeline for the segment, and add track info + + for (i = 0; i < this.pendingCaptions.length; i++) { + caption = this.pendingCaptions[i]; + caption.startTime = clock.metadataTsToSeconds(caption.startPts, timelineStartPts, this.keepOriginalTimestamps); + caption.endTime = clock.metadataTsToSeconds(caption.endPts, timelineStartPts, this.keepOriginalTimestamps); + event.captionStreams[caption.stream] = true; + event.captions.push(caption); + } // Translate ID3 frame PTS times into second offsets to match the + // video timeline for the segment + + for (i = 0; i < this.pendingMetadata.length; i++) { + id3 = this.pendingMetadata[i]; + id3.cueTime = clock.metadataTsToSeconds(id3.pts, timelineStartPts, this.keepOriginalTimestamps); + event.metadata.push(id3); + } // We add this to every single emitted segment even though we only need + // it for the first + + event.metadata.dispatchType = this.metadataStream.dispatchType; // Reset stream state + + this.pendingTracks.length = 0; + this.videoTrack = null; + this.pendingBoxes.length = 0; + this.pendingCaptions.length = 0; + this.pendingBytes = 0; + this.pendingMetadata.length = 0; // Emit the built segment + // We include captions and ID3 tags for backwards compatibility, + // ideally we should send only video and audio in the data event + + this.trigger('data', event); // Emit each caption to the outside world + // Ideally, this would happen immediately on parsing captions, + // but we need to ensure that video data is sent back first + // so that caption timing can be adjusted to match video timing + + for (i = 0; i < event.captions.length; i++) { + caption = event.captions[i]; + this.trigger('caption', caption); + } // Emit each id3 tag to the outside world + // Ideally, this would happen immediately on parsing the tag, + // but we need to ensure that video data is sent back first + // so that ID3 frame timing can be adjusted to match video timing + + for (i = 0; i < event.metadata.length; i++) { + id3 = event.metadata[i]; + this.trigger('id3Frame', id3); + } + } // Only emit `done` if all tracks have been flushed and emitted + + if (this.emittedTracks >= this.numberOfTracks) { + this.trigger('done'); + this.emittedTracks = 0; + } + }; + CoalesceStream.prototype.setRemux = function (val) { + this.remuxTracks = val; + }; + /** + * A Stream that expects MP2T binary data as input and produces + * corresponding media segments, suitable for use with Media Source + * Extension (MSE) implementations that support the ISO BMFF byte + * stream format, like Chrome. + */ + + Transmuxer = function (options) { + var self = this, + hasFlushed = true, + videoTrack, + audioTrack; + Transmuxer.prototype.init.call(this); + options = options || {}; + this.baseMediaDecodeTime = options.baseMediaDecodeTime || 0; + this.transmuxPipeline_ = {}; + this.setupAacPipeline = function () { + var pipeline = {}; + this.transmuxPipeline_ = pipeline; + pipeline.type = 'aac'; + pipeline.metadataStream = new m2ts.MetadataStream(); // set up the parsing pipeline + + pipeline.aacStream = new AacStream(); + pipeline.audioTimestampRolloverStream = new m2ts.TimestampRolloverStream('audio'); + pipeline.timedMetadataTimestampRolloverStream = new m2ts.TimestampRolloverStream('timed-metadata'); + pipeline.adtsStream = new AdtsStream(); + pipeline.coalesceStream = new CoalesceStream(options, pipeline.metadataStream); + pipeline.headOfPipeline = pipeline.aacStream; + pipeline.aacStream.pipe(pipeline.audioTimestampRolloverStream).pipe(pipeline.adtsStream); + pipeline.aacStream.pipe(pipeline.timedMetadataTimestampRolloverStream).pipe(pipeline.metadataStream).pipe(pipeline.coalesceStream); + pipeline.metadataStream.on('timestamp', function (frame) { + pipeline.aacStream.setTimestamp(frame.timeStamp); + }); + pipeline.aacStream.on('data', function (data) { + if (data.type !== 'timed-metadata' && data.type !== 'audio' || pipeline.audioSegmentStream) { + return; + } + audioTrack = audioTrack || { + timelineStartInfo: { + baseMediaDecodeTime: self.baseMediaDecodeTime + }, + codec: 'adts', + type: 'audio' + }; // hook up the audio segment stream to the first track with aac data + + pipeline.coalesceStream.numberOfTracks++; + pipeline.audioSegmentStream = new AudioSegmentStream(audioTrack, options); + pipeline.audioSegmentStream.on('log', self.getLogTrigger_('audioSegmentStream')); + pipeline.audioSegmentStream.on('timingInfo', self.trigger.bind(self, 'audioTimingInfo')); // Set up the final part of the audio pipeline + + pipeline.adtsStream.pipe(pipeline.audioSegmentStream).pipe(pipeline.coalesceStream); // emit pmt info + + self.trigger('trackinfo', { + hasAudio: !!audioTrack, + hasVideo: !!videoTrack + }); + }); // Re-emit any data coming from the coalesce stream to the outside world + + pipeline.coalesceStream.on('data', this.trigger.bind(this, 'data')); // Let the consumer know we have finished flushing the entire pipeline + + pipeline.coalesceStream.on('done', this.trigger.bind(this, 'done')); + addPipelineLogRetriggers(this, pipeline); + }; + this.setupTsPipeline = function () { + var pipeline = {}; + this.transmuxPipeline_ = pipeline; + pipeline.type = 'ts'; + pipeline.metadataStream = new m2ts.MetadataStream(); // set up the parsing pipeline + + pipeline.packetStream = new m2ts.TransportPacketStream(); + pipeline.parseStream = new m2ts.TransportParseStream(); + pipeline.elementaryStream = new m2ts.ElementaryStream(); + pipeline.timestampRolloverStream = new m2ts.TimestampRolloverStream(); + pipeline.adtsStream = new AdtsStream(); + pipeline.h264Stream = new H264Stream(); + pipeline.captionStream = new m2ts.CaptionStream(options); + pipeline.coalesceStream = new CoalesceStream(options, pipeline.metadataStream); + pipeline.headOfPipeline = pipeline.packetStream; // disassemble MPEG2-TS packets into elementary streams + + pipeline.packetStream.pipe(pipeline.parseStream).pipe(pipeline.elementaryStream).pipe(pipeline.timestampRolloverStream); // !!THIS ORDER IS IMPORTANT!! + // demux the streams + + pipeline.timestampRolloverStream.pipe(pipeline.h264Stream); + pipeline.timestampRolloverStream.pipe(pipeline.adtsStream); + pipeline.timestampRolloverStream.pipe(pipeline.metadataStream).pipe(pipeline.coalesceStream); // Hook up CEA-608/708 caption stream + + pipeline.h264Stream.pipe(pipeline.captionStream).pipe(pipeline.coalesceStream); + pipeline.elementaryStream.on('data', function (data) { + var i; + if (data.type === 'metadata') { + i = data.tracks.length; // scan the tracks listed in the metadata + + while (i--) { + if (!videoTrack && data.tracks[i].type === 'video') { + videoTrack = data.tracks[i]; + videoTrack.timelineStartInfo.baseMediaDecodeTime = self.baseMediaDecodeTime; + } else if (!audioTrack && data.tracks[i].type === 'audio') { + audioTrack = data.tracks[i]; + audioTrack.timelineStartInfo.baseMediaDecodeTime = self.baseMediaDecodeTime; + } + } // hook up the video segment stream to the first track with h264 data + + if (videoTrack && !pipeline.videoSegmentStream) { + pipeline.coalesceStream.numberOfTracks++; + pipeline.videoSegmentStream = new VideoSegmentStream(videoTrack, options); + pipeline.videoSegmentStream.on('log', self.getLogTrigger_('videoSegmentStream')); + pipeline.videoSegmentStream.on('timelineStartInfo', function (timelineStartInfo) { + // When video emits timelineStartInfo data after a flush, we forward that + // info to the AudioSegmentStream, if it exists, because video timeline + // data takes precedence. Do not do this if keepOriginalTimestamps is set, + // because this is a particularly subtle form of timestamp alteration. + if (audioTrack && !options.keepOriginalTimestamps) { + audioTrack.timelineStartInfo = timelineStartInfo; // On the first segment we trim AAC frames that exist before the + // very earliest DTS we have seen in video because Chrome will + // interpret any video track with a baseMediaDecodeTime that is + // non-zero as a gap. + + pipeline.audioSegmentStream.setEarliestDts(timelineStartInfo.dts - self.baseMediaDecodeTime); + } + }); + pipeline.videoSegmentStream.on('processedGopsInfo', self.trigger.bind(self, 'gopInfo')); + pipeline.videoSegmentStream.on('segmentTimingInfo', self.trigger.bind(self, 'videoSegmentTimingInfo')); + pipeline.videoSegmentStream.on('baseMediaDecodeTime', function (baseMediaDecodeTime) { + if (audioTrack) { + pipeline.audioSegmentStream.setVideoBaseMediaDecodeTime(baseMediaDecodeTime); + } + }); + pipeline.videoSegmentStream.on('timingInfo', self.trigger.bind(self, 'videoTimingInfo')); // Set up the final part of the video pipeline + + pipeline.h264Stream.pipe(pipeline.videoSegmentStream).pipe(pipeline.coalesceStream); + } + if (audioTrack && !pipeline.audioSegmentStream) { + // hook up the audio segment stream to the first track with aac data + pipeline.coalesceStream.numberOfTracks++; + pipeline.audioSegmentStream = new AudioSegmentStream(audioTrack, options); + pipeline.audioSegmentStream.on('log', self.getLogTrigger_('audioSegmentStream')); + pipeline.audioSegmentStream.on('timingInfo', self.trigger.bind(self, 'audioTimingInfo')); + pipeline.audioSegmentStream.on('segmentTimingInfo', self.trigger.bind(self, 'audioSegmentTimingInfo')); // Set up the final part of the audio pipeline + + pipeline.adtsStream.pipe(pipeline.audioSegmentStream).pipe(pipeline.coalesceStream); + } // emit pmt info + + self.trigger('trackinfo', { + hasAudio: !!audioTrack, + hasVideo: !!videoTrack + }); + } + }); // Re-emit any data coming from the coalesce stream to the outside world + + pipeline.coalesceStream.on('data', this.trigger.bind(this, 'data')); + pipeline.coalesceStream.on('id3Frame', function (id3Frame) { + id3Frame.dispatchType = pipeline.metadataStream.dispatchType; + self.trigger('id3Frame', id3Frame); + }); + pipeline.coalesceStream.on('caption', this.trigger.bind(this, 'caption')); // Let the consumer know we have finished flushing the entire pipeline + + pipeline.coalesceStream.on('done', this.trigger.bind(this, 'done')); + addPipelineLogRetriggers(this, pipeline); + }; // hook up the segment streams once track metadata is delivered + + this.setBaseMediaDecodeTime = function (baseMediaDecodeTime) { + var pipeline = this.transmuxPipeline_; + if (!options.keepOriginalTimestamps) { + this.baseMediaDecodeTime = baseMediaDecodeTime; + } + if (audioTrack) { + audioTrack.timelineStartInfo.dts = undefined; + audioTrack.timelineStartInfo.pts = undefined; + trackDecodeInfo.clearDtsInfo(audioTrack); + if (pipeline.audioTimestampRolloverStream) { + pipeline.audioTimestampRolloverStream.discontinuity(); + } + } + if (videoTrack) { + if (pipeline.videoSegmentStream) { + pipeline.videoSegmentStream.gopCache_ = []; + } + videoTrack.timelineStartInfo.dts = undefined; + videoTrack.timelineStartInfo.pts = undefined; + trackDecodeInfo.clearDtsInfo(videoTrack); + pipeline.captionStream.reset(); + } + if (pipeline.timestampRolloverStream) { + pipeline.timestampRolloverStream.discontinuity(); + } + }; + this.setAudioAppendStart = function (timestamp) { + if (audioTrack) { + this.transmuxPipeline_.audioSegmentStream.setAudioAppendStart(timestamp); + } + }; + this.setRemux = function (val) { + var pipeline = this.transmuxPipeline_; + options.remux = val; + if (pipeline && pipeline.coalesceStream) { + pipeline.coalesceStream.setRemux(val); + } + }; + this.alignGopsWith = function (gopsToAlignWith) { + if (videoTrack && this.transmuxPipeline_.videoSegmentStream) { + this.transmuxPipeline_.videoSegmentStream.alignGopsWith(gopsToAlignWith); + } + }; + this.getLogTrigger_ = function (key) { + var self = this; + return function (event) { + event.stream = key; + self.trigger('log', event); + }; + }; // feed incoming data to the front of the parsing pipeline + + this.push = function (data) { + if (hasFlushed) { + var isAac = isLikelyAacData(data); + if (isAac && this.transmuxPipeline_.type !== 'aac') { + this.setupAacPipeline(); + } else if (!isAac && this.transmuxPipeline_.type !== 'ts') { + this.setupTsPipeline(); + } + hasFlushed = false; + } + this.transmuxPipeline_.headOfPipeline.push(data); + }; // flush any buffered data + + this.flush = function () { + hasFlushed = true; // Start at the top of the pipeline and flush all pending work + + this.transmuxPipeline_.headOfPipeline.flush(); + }; + this.endTimeline = function () { + this.transmuxPipeline_.headOfPipeline.endTimeline(); + }; + this.reset = function () { + if (this.transmuxPipeline_.headOfPipeline) { + this.transmuxPipeline_.headOfPipeline.reset(); + } + }; // Caption data has to be reset when seeking outside buffered range + + this.resetCaptions = function () { + if (this.transmuxPipeline_.captionStream) { + this.transmuxPipeline_.captionStream.reset(); + } + }; + }; + Transmuxer.prototype = new Stream(); + var transmuxer = { + Transmuxer: Transmuxer, + VideoSegmentStream: VideoSegmentStream, + AudioSegmentStream: AudioSegmentStream, + AUDIO_PROPERTIES: AUDIO_PROPERTIES, + VIDEO_PROPERTIES: VIDEO_PROPERTIES, + // exported for testing + generateSegmentTimingInfo: generateSegmentTimingInfo + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + */ + + var toUnsigned$3 = function (value) { + return value >>> 0; + }; + var toHexString$1 = function (value) { + return ('00' + value.toString(16)).slice(-2); + }; + var bin = { + toUnsigned: toUnsigned$3, + toHexString: toHexString$1 + }; + var parseType$3 = function (buffer) { + var result = ''; + result += String.fromCharCode(buffer[0]); + result += String.fromCharCode(buffer[1]); + result += String.fromCharCode(buffer[2]); + result += String.fromCharCode(buffer[3]); + return result; + }; + var parseType_1 = parseType$3; + var toUnsigned$2 = bin.toUnsigned; + var parseType$2 = parseType_1; + var findBox$2 = function (data, path) { + var results = [], + i, + size, + type, + end, + subresults; + if (!path.length) { + // short-circuit the search for empty paths + return null; + } + for (i = 0; i < data.byteLength;) { + size = toUnsigned$2(data[i] << 24 | data[i + 1] << 16 | data[i + 2] << 8 | data[i + 3]); + type = parseType$2(data.subarray(i + 4, i + 8)); + end = size > 1 ? i + size : data.byteLength; + if (type === path[0]) { + if (path.length === 1) { + // this is the end of the path and we've found the box we were + // looking for + results.push(data.subarray(i + 8, end)); + } else { + // recursively search for the next box along the path + subresults = findBox$2(data.subarray(i + 8, end), path.slice(1)); + if (subresults.length) { + results = results.concat(subresults); + } + } + } + i = end; + } // we've finished searching all of data + + return results; + }; + var findBox_1 = findBox$2; + var toUnsigned$1 = bin.toUnsigned; + var getUint64$2 = numbers.getUint64; + var tfdt = function (data) { + var result = { + version: data[0], + flags: new Uint8Array(data.subarray(1, 4)) + }; + if (result.version === 1) { + result.baseMediaDecodeTime = getUint64$2(data.subarray(4)); + } else { + result.baseMediaDecodeTime = toUnsigned$1(data[4] << 24 | data[5] << 16 | data[6] << 8 | data[7]); + } + return result; + }; + var parseTfdt$2 = tfdt; + var parseSampleFlags$1 = function (flags) { + return { + isLeading: (flags[0] & 0x0c) >>> 2, + dependsOn: flags[0] & 0x03, + isDependedOn: (flags[1] & 0xc0) >>> 6, + hasRedundancy: (flags[1] & 0x30) >>> 4, + paddingValue: (flags[1] & 0x0e) >>> 1, + isNonSyncSample: flags[1] & 0x01, + degradationPriority: flags[2] << 8 | flags[3] + }; + }; + var parseSampleFlags_1 = parseSampleFlags$1; + var parseSampleFlags = parseSampleFlags_1; + var trun = function (data) { + var result = { + version: data[0], + flags: new Uint8Array(data.subarray(1, 4)), + samples: [] + }, + view = new DataView(data.buffer, data.byteOffset, data.byteLength), + // Flag interpretation + dataOffsetPresent = result.flags[2] & 0x01, + // compare with 2nd byte of 0x1 + firstSampleFlagsPresent = result.flags[2] & 0x04, + // compare with 2nd byte of 0x4 + sampleDurationPresent = result.flags[1] & 0x01, + // compare with 2nd byte of 0x100 + sampleSizePresent = result.flags[1] & 0x02, + // compare with 2nd byte of 0x200 + sampleFlagsPresent = result.flags[1] & 0x04, + // compare with 2nd byte of 0x400 + sampleCompositionTimeOffsetPresent = result.flags[1] & 0x08, + // compare with 2nd byte of 0x800 + sampleCount = view.getUint32(4), + offset = 8, + sample; + if (dataOffsetPresent) { + // 32 bit signed integer + result.dataOffset = view.getInt32(offset); + offset += 4; + } // Overrides the flags for the first sample only. The order of + // optional values will be: duration, size, compositionTimeOffset + + if (firstSampleFlagsPresent && sampleCount) { + sample = { + flags: parseSampleFlags(data.subarray(offset, offset + 4)) + }; + offset += 4; + if (sampleDurationPresent) { + sample.duration = view.getUint32(offset); + offset += 4; + } + if (sampleSizePresent) { + sample.size = view.getUint32(offset); + offset += 4; + } + if (sampleCompositionTimeOffsetPresent) { + if (result.version === 1) { + sample.compositionTimeOffset = view.getInt32(offset); + } else { + sample.compositionTimeOffset = view.getUint32(offset); + } + offset += 4; + } + result.samples.push(sample); + sampleCount--; + } + while (sampleCount--) { + sample = {}; + if (sampleDurationPresent) { + sample.duration = view.getUint32(offset); + offset += 4; + } + if (sampleSizePresent) { + sample.size = view.getUint32(offset); + offset += 4; + } + if (sampleFlagsPresent) { + sample.flags = parseSampleFlags(data.subarray(offset, offset + 4)); + offset += 4; + } + if (sampleCompositionTimeOffsetPresent) { + if (result.version === 1) { + sample.compositionTimeOffset = view.getInt32(offset); + } else { + sample.compositionTimeOffset = view.getUint32(offset); + } + offset += 4; + } + result.samples.push(sample); + } + return result; + }; + var parseTrun$2 = trun; + var tfhd = function (data) { + var view = new DataView(data.buffer, data.byteOffset, data.byteLength), + result = { + version: data[0], + flags: new Uint8Array(data.subarray(1, 4)), + trackId: view.getUint32(4) + }, + baseDataOffsetPresent = result.flags[2] & 0x01, + sampleDescriptionIndexPresent = result.flags[2] & 0x02, + defaultSampleDurationPresent = result.flags[2] & 0x08, + defaultSampleSizePresent = result.flags[2] & 0x10, + defaultSampleFlagsPresent = result.flags[2] & 0x20, + durationIsEmpty = result.flags[0] & 0x010000, + defaultBaseIsMoof = result.flags[0] & 0x020000, + i; + i = 8; + if (baseDataOffsetPresent) { + i += 4; // truncate top 4 bytes + // FIXME: should we read the full 64 bits? + + result.baseDataOffset = view.getUint32(12); + i += 4; + } + if (sampleDescriptionIndexPresent) { + result.sampleDescriptionIndex = view.getUint32(i); + i += 4; + } + if (defaultSampleDurationPresent) { + result.defaultSampleDuration = view.getUint32(i); + i += 4; + } + if (defaultSampleSizePresent) { + result.defaultSampleSize = view.getUint32(i); + i += 4; + } + if (defaultSampleFlagsPresent) { + result.defaultSampleFlags = view.getUint32(i); + } + if (durationIsEmpty) { + result.durationIsEmpty = true; + } + if (!baseDataOffsetPresent && defaultBaseIsMoof) { + result.baseDataOffsetIsMoof = true; + } + return result; + }; + var parseTfhd$2 = tfhd; + var win; + if (typeof window !== "undefined") { + win = window; + } else if (typeof commonjsGlobal !== "undefined") { + win = commonjsGlobal; + } else if (typeof self !== "undefined") { + win = self; + } else { + win = {}; + } + var window_1 = win; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * Reads in-band CEA-708 captions out of FMP4 segments. + * @see https://en.wikipedia.org/wiki/CEA-708 + */ + + var discardEmulationPreventionBytes = captionPacketParser.discardEmulationPreventionBytes; + var CaptionStream = captionStream.CaptionStream; + var findBox$1 = findBox_1; + var parseTfdt$1 = parseTfdt$2; + var parseTrun$1 = parseTrun$2; + var parseTfhd$1 = parseTfhd$2; + var window$2 = window_1; + /** + * Maps an offset in the mdat to a sample based on the the size of the samples. + * Assumes that `parseSamples` has been called first. + * + * @param {Number} offset - The offset into the mdat + * @param {Object[]} samples - An array of samples, parsed using `parseSamples` + * @return {?Object} The matching sample, or null if no match was found. + * + * @see ISO-BMFF-12/2015, Section 8.8.8 + **/ + + var mapToSample = function (offset, samples) { + var approximateOffset = offset; + for (var i = 0; i < samples.length; i++) { + var sample = samples[i]; + if (approximateOffset < sample.size) { + return sample; + } + approximateOffset -= sample.size; + } + return null; + }; + /** + * Finds SEI nal units contained in a Media Data Box. + * Assumes that `parseSamples` has been called first. + * + * @param {Uint8Array} avcStream - The bytes of the mdat + * @param {Object[]} samples - The samples parsed out by `parseSamples` + * @param {Number} trackId - The trackId of this video track + * @return {Object[]} seiNals - the parsed SEI NALUs found. + * The contents of the seiNal should match what is expected by + * CaptionStream.push (nalUnitType, size, data, escapedRBSP, pts, dts) + * + * @see ISO-BMFF-12/2015, Section 8.1.1 + * @see Rec. ITU-T H.264, 7.3.2.3.1 + **/ + + var findSeiNals = function (avcStream, samples, trackId) { + var avcView = new DataView(avcStream.buffer, avcStream.byteOffset, avcStream.byteLength), + result = { + logs: [], + seiNals: [] + }, + seiNal, + i, + length, + lastMatchedSample; + for (i = 0; i + 4 < avcStream.length; i += length) { + length = avcView.getUint32(i); + i += 4; // Bail if this doesn't appear to be an H264 stream + + if (length <= 0) { + continue; + } + switch (avcStream[i] & 0x1F) { + case 0x06: + var data = avcStream.subarray(i + 1, i + 1 + length); + var matchingSample = mapToSample(i, samples); + seiNal = { + nalUnitType: 'sei_rbsp', + size: length, + data: data, + escapedRBSP: discardEmulationPreventionBytes(data), + trackId: trackId + }; + if (matchingSample) { + seiNal.pts = matchingSample.pts; + seiNal.dts = matchingSample.dts; + lastMatchedSample = matchingSample; + } else if (lastMatchedSample) { + // If a matching sample cannot be found, use the last + // sample's values as they should be as close as possible + seiNal.pts = lastMatchedSample.pts; + seiNal.dts = lastMatchedSample.dts; + } else { + result.logs.push({ + level: 'warn', + message: 'We\'ve encountered a nal unit without data at ' + i + ' for trackId ' + trackId + '. See mux.js#223.' + }); + break; + } + result.seiNals.push(seiNal); + break; + } + } + return result; + }; + /** + * Parses sample information out of Track Run Boxes and calculates + * the absolute presentation and decode timestamps of each sample. + * + * @param {Array<Uint8Array>} truns - The Trun Run boxes to be parsed + * @param {Number|BigInt} baseMediaDecodeTime - base media decode time from tfdt + @see ISO-BMFF-12/2015, Section 8.8.12 + * @param {Object} tfhd - The parsed Track Fragment Header + * @see inspect.parseTfhd + * @return {Object[]} the parsed samples + * + * @see ISO-BMFF-12/2015, Section 8.8.8 + **/ + + var parseSamples = function (truns, baseMediaDecodeTime, tfhd) { + var currentDts = baseMediaDecodeTime; + var defaultSampleDuration = tfhd.defaultSampleDuration || 0; + var defaultSampleSize = tfhd.defaultSampleSize || 0; + var trackId = tfhd.trackId; + var allSamples = []; + truns.forEach(function (trun) { + // Note: We currently do not parse the sample table as well + // as the trun. It's possible some sources will require this. + // moov > trak > mdia > minf > stbl + var trackRun = parseTrun$1(trun); + var samples = trackRun.samples; + samples.forEach(function (sample) { + if (sample.duration === undefined) { + sample.duration = defaultSampleDuration; + } + if (sample.size === undefined) { + sample.size = defaultSampleSize; + } + sample.trackId = trackId; + sample.dts = currentDts; + if (sample.compositionTimeOffset === undefined) { + sample.compositionTimeOffset = 0; + } + if (typeof currentDts === 'bigint') { + sample.pts = currentDts + window$2.BigInt(sample.compositionTimeOffset); + currentDts += window$2.BigInt(sample.duration); + } else { + sample.pts = currentDts + sample.compositionTimeOffset; + currentDts += sample.duration; + } + }); + allSamples = allSamples.concat(samples); + }); + return allSamples; + }; + /** + * Parses out caption nals from an FMP4 segment's video tracks. + * + * @param {Uint8Array} segment - The bytes of a single segment + * @param {Number} videoTrackId - The trackId of a video track in the segment + * @return {Object.<Number, Object[]>} A mapping of video trackId to + * a list of seiNals found in that track + **/ + + var parseCaptionNals = function (segment, videoTrackId) { + // To get the samples + var trafs = findBox$1(segment, ['moof', 'traf']); // To get SEI NAL units + + var mdats = findBox$1(segment, ['mdat']); + var captionNals = {}; + var mdatTrafPairs = []; // Pair up each traf with a mdat as moofs and mdats are in pairs + + mdats.forEach(function (mdat, index) { + var matchingTraf = trafs[index]; + mdatTrafPairs.push({ + mdat: mdat, + traf: matchingTraf + }); + }); + mdatTrafPairs.forEach(function (pair) { + var mdat = pair.mdat; + var traf = pair.traf; + var tfhd = findBox$1(traf, ['tfhd']); // Exactly 1 tfhd per traf + + var headerInfo = parseTfhd$1(tfhd[0]); + var trackId = headerInfo.trackId; + var tfdt = findBox$1(traf, ['tfdt']); // Either 0 or 1 tfdt per traf + + var baseMediaDecodeTime = tfdt.length > 0 ? parseTfdt$1(tfdt[0]).baseMediaDecodeTime : 0; + var truns = findBox$1(traf, ['trun']); + var samples; + var result; // Only parse video data for the chosen video track + + if (videoTrackId === trackId && truns.length > 0) { + samples = parseSamples(truns, baseMediaDecodeTime, headerInfo); + result = findSeiNals(mdat, samples, trackId); + if (!captionNals[trackId]) { + captionNals[trackId] = { + seiNals: [], + logs: [] + }; + } + captionNals[trackId].seiNals = captionNals[trackId].seiNals.concat(result.seiNals); + captionNals[trackId].logs = captionNals[trackId].logs.concat(result.logs); + } + }); + return captionNals; + }; + /** + * Parses out inband captions from an MP4 container and returns + * caption objects that can be used by WebVTT and the TextTrack API. + * @see https://developer.mozilla.org/en-US/docs/Web/API/VTTCue + * @see https://developer.mozilla.org/en-US/docs/Web/API/TextTrack + * Assumes that `probe.getVideoTrackIds` and `probe.timescale` have been called first + * + * @param {Uint8Array} segment - The fmp4 segment containing embedded captions + * @param {Number} trackId - The id of the video track to parse + * @param {Number} timescale - The timescale for the video track from the init segment + * + * @return {?Object[]} parsedCaptions - A list of captions or null if no video tracks + * @return {Number} parsedCaptions[].startTime - The time to show the caption in seconds + * @return {Number} parsedCaptions[].endTime - The time to stop showing the caption in seconds + * @return {Object[]} parsedCaptions[].content - A list of individual caption segments + * @return {String} parsedCaptions[].content.text - The visible content of the caption segment + * @return {Number} parsedCaptions[].content.line - The line height from 1-15 for positioning of the caption segment + * @return {Number} parsedCaptions[].content.position - The column indent percentage for cue positioning from 10-80 + **/ + + var parseEmbeddedCaptions = function (segment, trackId, timescale) { + var captionNals; // the ISO-BMFF spec says that trackId can't be zero, but there's some broken content out there + + if (trackId === null) { + return null; + } + captionNals = parseCaptionNals(segment, trackId); + var trackNals = captionNals[trackId] || {}; + return { + seiNals: trackNals.seiNals, + logs: trackNals.logs, + timescale: timescale + }; + }; + /** + * Converts SEI NALUs into captions that can be used by video.js + **/ + + var CaptionParser = function () { + var isInitialized = false; + var captionStream; // Stores segments seen before trackId and timescale are set + + var segmentCache; // Stores video track ID of the track being parsed + + var trackId; // Stores the timescale of the track being parsed + + var timescale; // Stores captions parsed so far + + var parsedCaptions; // Stores whether we are receiving partial data or not + + var parsingPartial; + /** + * A method to indicate whether a CaptionParser has been initalized + * @returns {Boolean} + **/ + + this.isInitialized = function () { + return isInitialized; + }; + /** + * Initializes the underlying CaptionStream, SEI NAL parsing + * and management, and caption collection + **/ + + this.init = function (options) { + captionStream = new CaptionStream(); + isInitialized = true; + parsingPartial = options ? options.isPartial : false; // Collect dispatched captions + + captionStream.on('data', function (event) { + // Convert to seconds in the source's timescale + event.startTime = event.startPts / timescale; + event.endTime = event.endPts / timescale; + parsedCaptions.captions.push(event); + parsedCaptions.captionStreams[event.stream] = true; + }); + captionStream.on('log', function (log) { + parsedCaptions.logs.push(log); + }); + }; + /** + * Determines if a new video track will be selected + * or if the timescale changed + * @return {Boolean} + **/ + + this.isNewInit = function (videoTrackIds, timescales) { + if (videoTrackIds && videoTrackIds.length === 0 || timescales && typeof timescales === 'object' && Object.keys(timescales).length === 0) { + return false; + } + return trackId !== videoTrackIds[0] || timescale !== timescales[trackId]; + }; + /** + * Parses out SEI captions and interacts with underlying + * CaptionStream to return dispatched captions + * + * @param {Uint8Array} segment - The fmp4 segment containing embedded captions + * @param {Number[]} videoTrackIds - A list of video tracks found in the init segment + * @param {Object.<Number, Number>} timescales - The timescales found in the init segment + * @see parseEmbeddedCaptions + * @see m2ts/caption-stream.js + **/ + + this.parse = function (segment, videoTrackIds, timescales) { + var parsedData; + if (!this.isInitialized()) { + return null; // This is not likely to be a video segment + } else if (!videoTrackIds || !timescales) { + return null; + } else if (this.isNewInit(videoTrackIds, timescales)) { + // Use the first video track only as there is no + // mechanism to switch to other video tracks + trackId = videoTrackIds[0]; + timescale = timescales[trackId]; // If an init segment has not been seen yet, hold onto segment + // data until we have one. + // the ISO-BMFF spec says that trackId can't be zero, but there's some broken content out there + } else if (trackId === null || !timescale) { + segmentCache.push(segment); + return null; + } // Now that a timescale and trackId is set, parse cached segments + + while (segmentCache.length > 0) { + var cachedSegment = segmentCache.shift(); + this.parse(cachedSegment, videoTrackIds, timescales); + } + parsedData = parseEmbeddedCaptions(segment, trackId, timescale); + if (parsedData && parsedData.logs) { + parsedCaptions.logs = parsedCaptions.logs.concat(parsedData.logs); + } + if (parsedData === null || !parsedData.seiNals) { + if (parsedCaptions.logs.length) { + return { + logs: parsedCaptions.logs, + captions: [], + captionStreams: [] + }; + } + return null; + } + this.pushNals(parsedData.seiNals); // Force the parsed captions to be dispatched + + this.flushStream(); + return parsedCaptions; + }; + /** + * Pushes SEI NALUs onto CaptionStream + * @param {Object[]} nals - A list of SEI nals parsed using `parseCaptionNals` + * Assumes that `parseCaptionNals` has been called first + * @see m2ts/caption-stream.js + **/ + + this.pushNals = function (nals) { + if (!this.isInitialized() || !nals || nals.length === 0) { + return null; + } + nals.forEach(function (nal) { + captionStream.push(nal); + }); + }; + /** + * Flushes underlying CaptionStream to dispatch processed, displayable captions + * @see m2ts/caption-stream.js + **/ + + this.flushStream = function () { + if (!this.isInitialized()) { + return null; + } + if (!parsingPartial) { + captionStream.flush(); + } else { + captionStream.partialFlush(); + } + }; + /** + * Reset caption buckets for new data + **/ + + this.clearParsedCaptions = function () { + parsedCaptions.captions = []; + parsedCaptions.captionStreams = {}; + parsedCaptions.logs = []; + }; + /** + * Resets underlying CaptionStream + * @see m2ts/caption-stream.js + **/ + + this.resetCaptionStream = function () { + if (!this.isInitialized()) { + return null; + } + captionStream.reset(); + }; + /** + * Convenience method to clear all captions flushed from the + * CaptionStream and still being parsed + * @see m2ts/caption-stream.js + **/ + + this.clearAllCaptions = function () { + this.clearParsedCaptions(); + this.resetCaptionStream(); + }; + /** + * Reset caption parser + **/ + + this.reset = function () { + segmentCache = []; + trackId = null; + timescale = null; + if (!parsedCaptions) { + parsedCaptions = { + captions: [], + // CC1, CC2, CC3, CC4 + captionStreams: {}, + logs: [] + }; + } else { + this.clearParsedCaptions(); + } + this.resetCaptionStream(); + }; + this.reset(); + }; + var captionParser = CaptionParser; + /** + * Returns the first string in the data array ending with a null char '\0' + * @param {UInt8} data + * @returns the string with the null char + */ + + var uint8ToCString$1 = function (data) { + var index = 0; + var curChar = String.fromCharCode(data[index]); + var retString = ''; + while (curChar !== '\0') { + retString += curChar; + index++; + curChar = String.fromCharCode(data[index]); + } // Add nullChar + + retString += curChar; + return retString; + }; + var string = { + uint8ToCString: uint8ToCString$1 + }; + var uint8ToCString = string.uint8ToCString; + var getUint64$1 = numbers.getUint64; + /** + * Based on: ISO/IEC 23009 Section: 5.10.3.3 + * References: + * https://dashif-documents.azurewebsites.net/Events/master/event.html#emsg-format + * https://aomediacodec.github.io/id3-emsg/ + * + * Takes emsg box data as a uint8 array and returns a emsg box object + * @param {UInt8Array} boxData data from emsg box + * @returns A parsed emsg box object + */ + + var parseEmsgBox = function (boxData) { + // version + flags + var offset = 4; + var version = boxData[0]; + var scheme_id_uri, value, timescale, presentation_time, presentation_time_delta, event_duration, id, message_data; + if (version === 0) { + scheme_id_uri = uint8ToCString(boxData.subarray(offset)); + offset += scheme_id_uri.length; + value = uint8ToCString(boxData.subarray(offset)); + offset += value.length; + var dv = new DataView(boxData.buffer); + timescale = dv.getUint32(offset); + offset += 4; + presentation_time_delta = dv.getUint32(offset); + offset += 4; + event_duration = dv.getUint32(offset); + offset += 4; + id = dv.getUint32(offset); + offset += 4; + } else if (version === 1) { + var dv = new DataView(boxData.buffer); + timescale = dv.getUint32(offset); + offset += 4; + presentation_time = getUint64$1(boxData.subarray(offset)); + offset += 8; + event_duration = dv.getUint32(offset); + offset += 4; + id = dv.getUint32(offset); + offset += 4; + scheme_id_uri = uint8ToCString(boxData.subarray(offset)); + offset += scheme_id_uri.length; + value = uint8ToCString(boxData.subarray(offset)); + offset += value.length; + } + message_data = new Uint8Array(boxData.subarray(offset, boxData.byteLength)); + var emsgBox = { + scheme_id_uri, + value, + // if timescale is undefined or 0 set to 1 + timescale: timescale ? timescale : 1, + presentation_time, + presentation_time_delta, + event_duration, + id, + message_data + }; + return isValidEmsgBox(version, emsgBox) ? emsgBox : undefined; + }; + /** + * Scales a presentation time or time delta with an offset with a provided timescale + * @param {number} presentationTime + * @param {number} timescale + * @param {number} timeDelta + * @param {number} offset + * @returns the scaled time as a number + */ + + var scaleTime = function (presentationTime, timescale, timeDelta, offset) { + return presentationTime || presentationTime === 0 ? presentationTime / timescale : offset + timeDelta / timescale; + }; + /** + * Checks the emsg box data for validity based on the version + * @param {number} version of the emsg box to validate + * @param {Object} emsg the emsg data to validate + * @returns if the box is valid as a boolean + */ + + var isValidEmsgBox = function (version, emsg) { + var hasScheme = emsg.scheme_id_uri !== '\0'; + var isValidV0Box = version === 0 && isDefined(emsg.presentation_time_delta) && hasScheme; + var isValidV1Box = version === 1 && isDefined(emsg.presentation_time) && hasScheme; // Only valid versions of emsg are 0 and 1 + + return !(version > 1) && isValidV0Box || isValidV1Box; + }; // Utility function to check if an object is defined + + var isDefined = function (data) { + return data !== undefined || data !== null; + }; + var emsg$1 = { + parseEmsgBox: parseEmsgBox, + scaleTime: scaleTime + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * Utilities to detect basic properties and metadata about MP4s. + */ + + var toUnsigned = bin.toUnsigned; + var toHexString = bin.toHexString; + var findBox = findBox_1; + var parseType$1 = parseType_1; + var emsg = emsg$1; + var parseTfhd = parseTfhd$2; + var parseTrun = parseTrun$2; + var parseTfdt = parseTfdt$2; + var getUint64 = numbers.getUint64; + var timescale, startTime, compositionStartTime, getVideoTrackIds, getTracks, getTimescaleFromMediaHeader, getEmsgID3; + var window$1 = window_1; + var parseId3Frames = parseId3.parseId3Frames; + /** + * Parses an MP4 initialization segment and extracts the timescale + * values for any declared tracks. Timescale values indicate the + * number of clock ticks per second to assume for time-based values + * elsewhere in the MP4. + * + * To determine the start time of an MP4, you need two pieces of + * information: the timescale unit and the earliest base media decode + * time. Multiple timescales can be specified within an MP4 but the + * base media decode time is always expressed in the timescale from + * the media header box for the track: + * ``` + * moov > trak > mdia > mdhd.timescale + * ``` + * @param init {Uint8Array} the bytes of the init segment + * @return {object} a hash of track ids to timescale values or null if + * the init segment is malformed. + */ + + timescale = function (init) { + var result = {}, + traks = findBox(init, ['moov', 'trak']); // mdhd timescale + + return traks.reduce(function (result, trak) { + var tkhd, version, index, id, mdhd; + tkhd = findBox(trak, ['tkhd'])[0]; + if (!tkhd) { + return null; + } + version = tkhd[0]; + index = version === 0 ? 12 : 20; + id = toUnsigned(tkhd[index] << 24 | tkhd[index + 1] << 16 | tkhd[index + 2] << 8 | tkhd[index + 3]); + mdhd = findBox(trak, ['mdia', 'mdhd'])[0]; + if (!mdhd) { + return null; + } + version = mdhd[0]; + index = version === 0 ? 12 : 20; + result[id] = toUnsigned(mdhd[index] << 24 | mdhd[index + 1] << 16 | mdhd[index + 2] << 8 | mdhd[index + 3]); + return result; + }, result); + }; + /** + * Determine the base media decode start time, in seconds, for an MP4 + * fragment. If multiple fragments are specified, the earliest time is + * returned. + * + * The base media decode time can be parsed from track fragment + * metadata: + * ``` + * moof > traf > tfdt.baseMediaDecodeTime + * ``` + * It requires the timescale value from the mdhd to interpret. + * + * @param timescale {object} a hash of track ids to timescale values. + * @return {number} the earliest base media decode start time for the + * fragment, in seconds + */ + + startTime = function (timescale, fragment) { + var trafs; // we need info from two childrend of each track fragment box + + trafs = findBox(fragment, ['moof', 'traf']); // determine the start times for each track + + var lowestTime = trafs.reduce(function (acc, traf) { + var tfhd = findBox(traf, ['tfhd'])[0]; // get the track id from the tfhd + + var id = toUnsigned(tfhd[4] << 24 | tfhd[5] << 16 | tfhd[6] << 8 | tfhd[7]); // assume a 90kHz clock if no timescale was specified + + var scale = timescale[id] || 90e3; // get the base media decode time from the tfdt + + var tfdt = findBox(traf, ['tfdt'])[0]; + var dv = new DataView(tfdt.buffer, tfdt.byteOffset, tfdt.byteLength); + var baseTime; // version 1 is 64 bit + + if (tfdt[0] === 1) { + baseTime = getUint64(tfdt.subarray(4, 12)); + } else { + baseTime = dv.getUint32(4); + } // convert base time to seconds if it is a valid number. + + let seconds; + if (typeof baseTime === 'bigint') { + seconds = baseTime / window$1.BigInt(scale); + } else if (typeof baseTime === 'number' && !isNaN(baseTime)) { + seconds = baseTime / scale; + } + if (seconds < Number.MAX_SAFE_INTEGER) { + seconds = Number(seconds); + } + if (seconds < acc) { + acc = seconds; + } + return acc; + }, Infinity); + return typeof lowestTime === 'bigint' || isFinite(lowestTime) ? lowestTime : 0; + }; + /** + * Determine the composition start, in seconds, for an MP4 + * fragment. + * + * The composition start time of a fragment can be calculated using the base + * media decode time, composition time offset, and timescale, as follows: + * + * compositionStartTime = (baseMediaDecodeTime + compositionTimeOffset) / timescale + * + * All of the aforementioned information is contained within a media fragment's + * `traf` box, except for timescale info, which comes from the initialization + * segment, so a track id (also contained within a `traf`) is also necessary to + * associate it with a timescale + * + * + * @param timescales {object} - a hash of track ids to timescale values. + * @param fragment {Unit8Array} - the bytes of a media segment + * @return {number} the composition start time for the fragment, in seconds + **/ + + compositionStartTime = function (timescales, fragment) { + var trafBoxes = findBox(fragment, ['moof', 'traf']); + var baseMediaDecodeTime = 0; + var compositionTimeOffset = 0; + var trackId; + if (trafBoxes && trafBoxes.length) { + // The spec states that track run samples contained within a `traf` box are contiguous, but + // it does not explicitly state whether the `traf` boxes themselves are contiguous. + // We will assume that they are, so we only need the first to calculate start time. + var tfhd = findBox(trafBoxes[0], ['tfhd'])[0]; + var trun = findBox(trafBoxes[0], ['trun'])[0]; + var tfdt = findBox(trafBoxes[0], ['tfdt'])[0]; + if (tfhd) { + var parsedTfhd = parseTfhd(tfhd); + trackId = parsedTfhd.trackId; + } + if (tfdt) { + var parsedTfdt = parseTfdt(tfdt); + baseMediaDecodeTime = parsedTfdt.baseMediaDecodeTime; + } + if (trun) { + var parsedTrun = parseTrun(trun); + if (parsedTrun.samples && parsedTrun.samples.length) { + compositionTimeOffset = parsedTrun.samples[0].compositionTimeOffset || 0; + } + } + } // Get timescale for this specific track. Assume a 90kHz clock if no timescale was + // specified. + + var timescale = timescales[trackId] || 90e3; // return the composition start time, in seconds + + if (typeof baseMediaDecodeTime === 'bigint') { + compositionTimeOffset = window$1.BigInt(compositionTimeOffset); + timescale = window$1.BigInt(timescale); + } + var result = (baseMediaDecodeTime + compositionTimeOffset) / timescale; + if (typeof result === 'bigint' && result < Number.MAX_SAFE_INTEGER) { + result = Number(result); + } + return result; + }; + /** + * Find the trackIds of the video tracks in this source. + * Found by parsing the Handler Reference and Track Header Boxes: + * moov > trak > mdia > hdlr + * moov > trak > tkhd + * + * @param {Uint8Array} init - The bytes of the init segment for this source + * @return {Number[]} A list of trackIds + * + * @see ISO-BMFF-12/2015, Section 8.4.3 + **/ + + getVideoTrackIds = function (init) { + var traks = findBox(init, ['moov', 'trak']); + var videoTrackIds = []; + traks.forEach(function (trak) { + var hdlrs = findBox(trak, ['mdia', 'hdlr']); + var tkhds = findBox(trak, ['tkhd']); + hdlrs.forEach(function (hdlr, index) { + var handlerType = parseType$1(hdlr.subarray(8, 12)); + var tkhd = tkhds[index]; + var view; + var version; + var trackId; + if (handlerType === 'vide') { + view = new DataView(tkhd.buffer, tkhd.byteOffset, tkhd.byteLength); + version = view.getUint8(0); + trackId = version === 0 ? view.getUint32(12) : view.getUint32(20); + videoTrackIds.push(trackId); + } + }); + }); + return videoTrackIds; + }; + getTimescaleFromMediaHeader = function (mdhd) { + // mdhd is a FullBox, meaning it will have its own version as the first byte + var version = mdhd[0]; + var index = version === 0 ? 12 : 20; + return toUnsigned(mdhd[index] << 24 | mdhd[index + 1] << 16 | mdhd[index + 2] << 8 | mdhd[index + 3]); + }; + /** + * Get all the video, audio, and hint tracks from a non fragmented + * mp4 segment + */ + + getTracks = function (init) { + var traks = findBox(init, ['moov', 'trak']); + var tracks = []; + traks.forEach(function (trak) { + var track = {}; + var tkhd = findBox(trak, ['tkhd'])[0]; + var view, tkhdVersion; // id + + if (tkhd) { + view = new DataView(tkhd.buffer, tkhd.byteOffset, tkhd.byteLength); + tkhdVersion = view.getUint8(0); + track.id = tkhdVersion === 0 ? view.getUint32(12) : view.getUint32(20); + } + var hdlr = findBox(trak, ['mdia', 'hdlr'])[0]; // type + + if (hdlr) { + var type = parseType$1(hdlr.subarray(8, 12)); + if (type === 'vide') { + track.type = 'video'; + } else if (type === 'soun') { + track.type = 'audio'; + } else { + track.type = type; + } + } // codec + + var stsd = findBox(trak, ['mdia', 'minf', 'stbl', 'stsd'])[0]; + if (stsd) { + var sampleDescriptions = stsd.subarray(8); // gives the codec type string + + track.codec = parseType$1(sampleDescriptions.subarray(4, 8)); + var codecBox = findBox(sampleDescriptions, [track.codec])[0]; + var codecConfig, codecConfigType; + if (codecBox) { + // https://tools.ietf.org/html/rfc6381#section-3.3 + if (/^[asm]vc[1-9]$/i.test(track.codec)) { + // we don't need anything but the "config" parameter of the + // avc1 codecBox + codecConfig = codecBox.subarray(78); + codecConfigType = parseType$1(codecConfig.subarray(4, 8)); + if (codecConfigType === 'avcC' && codecConfig.length > 11) { + track.codec += '.'; // left padded with zeroes for single digit hex + // profile idc + + track.codec += toHexString(codecConfig[9]); // the byte containing the constraint_set flags + + track.codec += toHexString(codecConfig[10]); // level idc + + track.codec += toHexString(codecConfig[11]); + } else { + // TODO: show a warning that we couldn't parse the codec + // and are using the default + track.codec = 'avc1.4d400d'; + } + } else if (/^mp4[a,v]$/i.test(track.codec)) { + // we do not need anything but the streamDescriptor of the mp4a codecBox + codecConfig = codecBox.subarray(28); + codecConfigType = parseType$1(codecConfig.subarray(4, 8)); + if (codecConfigType === 'esds' && codecConfig.length > 20 && codecConfig[19] !== 0) { + track.codec += '.' + toHexString(codecConfig[19]); // this value is only a single digit + + track.codec += '.' + toHexString(codecConfig[20] >>> 2 & 0x3f).replace(/^0/, ''); + } else { + // TODO: show a warning that we couldn't parse the codec + // and are using the default + track.codec = 'mp4a.40.2'; + } + } else { + // flac, opus, etc + track.codec = track.codec.toLowerCase(); + } + } + } + var mdhd = findBox(trak, ['mdia', 'mdhd'])[0]; + if (mdhd) { + track.timescale = getTimescaleFromMediaHeader(mdhd); + } + tracks.push(track); + }); + return tracks; + }; + /** + * Returns an array of emsg ID3 data from the provided segmentData. + * An offset can also be provided as the Latest Arrival Time to calculate + * the Event Start Time of v0 EMSG boxes. + * See: https://dashif-documents.azurewebsites.net/Events/master/event.html#Inband-event-timing + * + * @param {Uint8Array} segmentData the segment byte array. + * @param {number} offset the segment start time or Latest Arrival Time, + * @return {Object[]} an array of ID3 parsed from EMSG boxes + */ + + getEmsgID3 = function (segmentData, offset = 0) { + var emsgBoxes = findBox(segmentData, ['emsg']); + return emsgBoxes.map(data => { + var parsedBox = emsg.parseEmsgBox(new Uint8Array(data)); + var parsedId3Frames = parseId3Frames(parsedBox.message_data); + return { + cueTime: emsg.scaleTime(parsedBox.presentation_time, parsedBox.timescale, parsedBox.presentation_time_delta, offset), + duration: emsg.scaleTime(parsedBox.event_duration, parsedBox.timescale), + frames: parsedId3Frames + }; + }); + }; + var probe$2 = { + // export mp4 inspector's findBox and parseType for backwards compatibility + findBox: findBox, + parseType: parseType$1, + timescale: timescale, + startTime: startTime, + compositionStartTime: compositionStartTime, + videoTrackIds: getVideoTrackIds, + tracks: getTracks, + getTimescaleFromMediaHeader: getTimescaleFromMediaHeader, + getEmsgID3: getEmsgID3 + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * Utilities to detect basic properties and metadata about TS Segments. + */ + + var StreamTypes$1 = streamTypes; + var parsePid = function (packet) { + var pid = packet[1] & 0x1f; + pid <<= 8; + pid |= packet[2]; + return pid; + }; + var parsePayloadUnitStartIndicator = function (packet) { + return !!(packet[1] & 0x40); + }; + var parseAdaptionField = function (packet) { + var offset = 0; // if an adaption field is present, its length is specified by the + // fifth byte of the TS packet header. The adaptation field is + // used to add stuffing to PES packets that don't fill a complete + // TS packet, and to specify some forms of timing and control data + // that we do not currently use. + + if ((packet[3] & 0x30) >>> 4 > 0x01) { + offset += packet[4] + 1; + } + return offset; + }; + var parseType = function (packet, pmtPid) { + var pid = parsePid(packet); + if (pid === 0) { + return 'pat'; + } else if (pid === pmtPid) { + return 'pmt'; + } else if (pmtPid) { + return 'pes'; + } + return null; + }; + var parsePat = function (packet) { + var pusi = parsePayloadUnitStartIndicator(packet); + var offset = 4 + parseAdaptionField(packet); + if (pusi) { + offset += packet[offset] + 1; + } + return (packet[offset + 10] & 0x1f) << 8 | packet[offset + 11]; + }; + var parsePmt = function (packet) { + var programMapTable = {}; + var pusi = parsePayloadUnitStartIndicator(packet); + var payloadOffset = 4 + parseAdaptionField(packet); + if (pusi) { + payloadOffset += packet[payloadOffset] + 1; + } // PMTs can be sent ahead of the time when they should actually + // take effect. We don't believe this should ever be the case + // for HLS but we'll ignore "forward" PMT declarations if we see + // them. Future PMT declarations have the current_next_indicator + // set to zero. + + if (!(packet[payloadOffset + 5] & 0x01)) { + return; + } + var sectionLength, tableEnd, programInfoLength; // the mapping table ends at the end of the current section + + sectionLength = (packet[payloadOffset + 1] & 0x0f) << 8 | packet[payloadOffset + 2]; + tableEnd = 3 + sectionLength - 4; // to determine where the table is, we have to figure out how + // long the program info descriptors are + + programInfoLength = (packet[payloadOffset + 10] & 0x0f) << 8 | packet[payloadOffset + 11]; // advance the offset to the first entry in the mapping table + + var offset = 12 + programInfoLength; + while (offset < tableEnd) { + var i = payloadOffset + offset; // add an entry that maps the elementary_pid to the stream_type + + programMapTable[(packet[i + 1] & 0x1F) << 8 | packet[i + 2]] = packet[i]; // move to the next table entry + // skip past the elementary stream descriptors, if present + + offset += ((packet[i + 3] & 0x0F) << 8 | packet[i + 4]) + 5; + } + return programMapTable; + }; + var parsePesType = function (packet, programMapTable) { + var pid = parsePid(packet); + var type = programMapTable[pid]; + switch (type) { + case StreamTypes$1.H264_STREAM_TYPE: + return 'video'; + case StreamTypes$1.ADTS_STREAM_TYPE: + return 'audio'; + case StreamTypes$1.METADATA_STREAM_TYPE: + return 'timed-metadata'; + default: + return null; + } + }; + var parsePesTime = function (packet) { + var pusi = parsePayloadUnitStartIndicator(packet); + if (!pusi) { + return null; + } + var offset = 4 + parseAdaptionField(packet); + if (offset >= packet.byteLength) { + // From the H 222.0 MPEG-TS spec + // "For transport stream packets carrying PES packets, stuffing is needed when there + // is insufficient PES packet data to completely fill the transport stream packet + // payload bytes. Stuffing is accomplished by defining an adaptation field longer than + // the sum of the lengths of the data elements in it, so that the payload bytes + // remaining after the adaptation field exactly accommodates the available PES packet + // data." + // + // If the offset is >= the length of the packet, then the packet contains no data + // and instead is just adaption field stuffing bytes + return null; + } + var pes = null; + var ptsDtsFlags; // PES packets may be annotated with a PTS value, or a PTS value + // and a DTS value. Determine what combination of values is + // available to work with. + + ptsDtsFlags = packet[offset + 7]; // PTS and DTS are normally stored as a 33-bit number. Javascript + // performs all bitwise operations on 32-bit integers but javascript + // supports a much greater range (52-bits) of integer using standard + // mathematical operations. + // We construct a 31-bit value using bitwise operators over the 31 + // most significant bits and then multiply by 4 (equal to a left-shift + // of 2) before we add the final 2 least significant bits of the + // timestamp (equal to an OR.) + + if (ptsDtsFlags & 0xC0) { + pes = {}; // the PTS and DTS are not written out directly. For information + // on how they are encoded, see + // http://dvd.sourceforge.net/dvdinfo/pes-hdr.html + + pes.pts = (packet[offset + 9] & 0x0E) << 27 | (packet[offset + 10] & 0xFF) << 20 | (packet[offset + 11] & 0xFE) << 12 | (packet[offset + 12] & 0xFF) << 5 | (packet[offset + 13] & 0xFE) >>> 3; + pes.pts *= 4; // Left shift by 2 + + pes.pts += (packet[offset + 13] & 0x06) >>> 1; // OR by the two LSBs + + pes.dts = pes.pts; + if (ptsDtsFlags & 0x40) { + pes.dts = (packet[offset + 14] & 0x0E) << 27 | (packet[offset + 15] & 0xFF) << 20 | (packet[offset + 16] & 0xFE) << 12 | (packet[offset + 17] & 0xFF) << 5 | (packet[offset + 18] & 0xFE) >>> 3; + pes.dts *= 4; // Left shift by 2 + + pes.dts += (packet[offset + 18] & 0x06) >>> 1; // OR by the two LSBs + } + } + + return pes; + }; + var parseNalUnitType = function (type) { + switch (type) { + case 0x05: + return 'slice_layer_without_partitioning_rbsp_idr'; + case 0x06: + return 'sei_rbsp'; + case 0x07: + return 'seq_parameter_set_rbsp'; + case 0x08: + return 'pic_parameter_set_rbsp'; + case 0x09: + return 'access_unit_delimiter_rbsp'; + default: + return null; + } + }; + var videoPacketContainsKeyFrame = function (packet) { + var offset = 4 + parseAdaptionField(packet); + var frameBuffer = packet.subarray(offset); + var frameI = 0; + var frameSyncPoint = 0; + var foundKeyFrame = false; + var nalType; // advance the sync point to a NAL start, if necessary + + for (; frameSyncPoint < frameBuffer.byteLength - 3; frameSyncPoint++) { + if (frameBuffer[frameSyncPoint + 2] === 1) { + // the sync point is properly aligned + frameI = frameSyncPoint + 5; + break; + } + } + while (frameI < frameBuffer.byteLength) { + // look at the current byte to determine if we've hit the end of + // a NAL unit boundary + switch (frameBuffer[frameI]) { + case 0: + // skip past non-sync sequences + if (frameBuffer[frameI - 1] !== 0) { + frameI += 2; + break; + } else if (frameBuffer[frameI - 2] !== 0) { + frameI++; + break; + } + if (frameSyncPoint + 3 !== frameI - 2) { + nalType = parseNalUnitType(frameBuffer[frameSyncPoint + 3] & 0x1f); + if (nalType === 'slice_layer_without_partitioning_rbsp_idr') { + foundKeyFrame = true; + } + } // drop trailing zeroes + + do { + frameI++; + } while (frameBuffer[frameI] !== 1 && frameI < frameBuffer.length); + frameSyncPoint = frameI - 2; + frameI += 3; + break; + case 1: + // skip past non-sync sequences + if (frameBuffer[frameI - 1] !== 0 || frameBuffer[frameI - 2] !== 0) { + frameI += 3; + break; + } + nalType = parseNalUnitType(frameBuffer[frameSyncPoint + 3] & 0x1f); + if (nalType === 'slice_layer_without_partitioning_rbsp_idr') { + foundKeyFrame = true; + } + frameSyncPoint = frameI - 2; + frameI += 3; + break; + default: + // the current byte isn't a one or zero, so it cannot be part + // of a sync sequence + frameI += 3; + break; + } + } + frameBuffer = frameBuffer.subarray(frameSyncPoint); + frameI -= frameSyncPoint; + frameSyncPoint = 0; // parse the final nal + + if (frameBuffer && frameBuffer.byteLength > 3) { + nalType = parseNalUnitType(frameBuffer[frameSyncPoint + 3] & 0x1f); + if (nalType === 'slice_layer_without_partitioning_rbsp_idr') { + foundKeyFrame = true; + } + } + return foundKeyFrame; + }; + var probe$1 = { + parseType: parseType, + parsePat: parsePat, + parsePmt: parsePmt, + parsePayloadUnitStartIndicator: parsePayloadUnitStartIndicator, + parsePesType: parsePesType, + parsePesTime: parsePesTime, + videoPacketContainsKeyFrame: videoPacketContainsKeyFrame + }; + /** + * mux.js + * + * Copyright (c) Brightcove + * Licensed Apache-2.0 https://github.com/videojs/mux.js/blob/master/LICENSE + * + * Parse mpeg2 transport stream packets to extract basic timing information + */ + + var StreamTypes = streamTypes; + var handleRollover = timestampRolloverStream.handleRollover; + var probe = {}; + probe.ts = probe$1; + probe.aac = utils; + var ONE_SECOND_IN_TS = clock$2.ONE_SECOND_IN_TS; + var MP2T_PACKET_LENGTH = 188, + // bytes + SYNC_BYTE = 0x47; + /** + * walks through segment data looking for pat and pmt packets to parse out + * program map table information + */ + + var parsePsi_ = function (bytes, pmt) { + var startIndex = 0, + endIndex = MP2T_PACKET_LENGTH, + packet, + type; + while (endIndex < bytes.byteLength) { + // Look for a pair of start and end sync bytes in the data.. + if (bytes[startIndex] === SYNC_BYTE && bytes[endIndex] === SYNC_BYTE) { + // We found a packet + packet = bytes.subarray(startIndex, endIndex); + type = probe.ts.parseType(packet, pmt.pid); + switch (type) { + case 'pat': + pmt.pid = probe.ts.parsePat(packet); + break; + case 'pmt': + var table = probe.ts.parsePmt(packet); + pmt.table = pmt.table || {}; + Object.keys(table).forEach(function (key) { + pmt.table[key] = table[key]; + }); + break; + } + startIndex += MP2T_PACKET_LENGTH; + endIndex += MP2T_PACKET_LENGTH; + continue; + } // If we get here, we have somehow become de-synchronized and we need to step + // forward one byte at a time until we find a pair of sync bytes that denote + // a packet + + startIndex++; + endIndex++; + } + }; + /** + * walks through the segment data from the start and end to get timing information + * for the first and last audio pes packets + */ + + var parseAudioPes_ = function (bytes, pmt, result) { + var startIndex = 0, + endIndex = MP2T_PACKET_LENGTH, + packet, + type, + pesType, + pusi, + parsed; + var endLoop = false; // Start walking from start of segment to get first audio packet + + while (endIndex <= bytes.byteLength) { + // Look for a pair of start and end sync bytes in the data.. + if (bytes[startIndex] === SYNC_BYTE && (bytes[endIndex] === SYNC_BYTE || endIndex === bytes.byteLength)) { + // We found a packet + packet = bytes.subarray(startIndex, endIndex); + type = probe.ts.parseType(packet, pmt.pid); + switch (type) { + case 'pes': + pesType = probe.ts.parsePesType(packet, pmt.table); + pusi = probe.ts.parsePayloadUnitStartIndicator(packet); + if (pesType === 'audio' && pusi) { + parsed = probe.ts.parsePesTime(packet); + if (parsed) { + parsed.type = 'audio'; + result.audio.push(parsed); + endLoop = true; + } + } + break; + } + if (endLoop) { + break; + } + startIndex += MP2T_PACKET_LENGTH; + endIndex += MP2T_PACKET_LENGTH; + continue; + } // If we get here, we have somehow become de-synchronized and we need to step + // forward one byte at a time until we find a pair of sync bytes that denote + // a packet + + startIndex++; + endIndex++; + } // Start walking from end of segment to get last audio packet + + endIndex = bytes.byteLength; + startIndex = endIndex - MP2T_PACKET_LENGTH; + endLoop = false; + while (startIndex >= 0) { + // Look for a pair of start and end sync bytes in the data.. + if (bytes[startIndex] === SYNC_BYTE && (bytes[endIndex] === SYNC_BYTE || endIndex === bytes.byteLength)) { + // We found a packet + packet = bytes.subarray(startIndex, endIndex); + type = probe.ts.parseType(packet, pmt.pid); + switch (type) { + case 'pes': + pesType = probe.ts.parsePesType(packet, pmt.table); + pusi = probe.ts.parsePayloadUnitStartIndicator(packet); + if (pesType === 'audio' && pusi) { + parsed = probe.ts.parsePesTime(packet); + if (parsed) { + parsed.type = 'audio'; + result.audio.push(parsed); + endLoop = true; + } + } + break; + } + if (endLoop) { + break; + } + startIndex -= MP2T_PACKET_LENGTH; + endIndex -= MP2T_PACKET_LENGTH; + continue; + } // If we get here, we have somehow become de-synchronized and we need to step + // forward one byte at a time until we find a pair of sync bytes that denote + // a packet + + startIndex--; + endIndex--; + } + }; + /** + * walks through the segment data from the start and end to get timing information + * for the first and last video pes packets as well as timing information for the first + * key frame. + */ + + var parseVideoPes_ = function (bytes, pmt, result) { + var startIndex = 0, + endIndex = MP2T_PACKET_LENGTH, + packet, + type, + pesType, + pusi, + parsed, + frame, + i, + pes; + var endLoop = false; + var currentFrame = { + data: [], + size: 0 + }; // Start walking from start of segment to get first video packet + + while (endIndex < bytes.byteLength) { + // Look for a pair of start and end sync bytes in the data.. + if (bytes[startIndex] === SYNC_BYTE && bytes[endIndex] === SYNC_BYTE) { + // We found a packet + packet = bytes.subarray(startIndex, endIndex); + type = probe.ts.parseType(packet, pmt.pid); + switch (type) { + case 'pes': + pesType = probe.ts.parsePesType(packet, pmt.table); + pusi = probe.ts.parsePayloadUnitStartIndicator(packet); + if (pesType === 'video') { + if (pusi && !endLoop) { + parsed = probe.ts.parsePesTime(packet); + if (parsed) { + parsed.type = 'video'; + result.video.push(parsed); + endLoop = true; + } + } + if (!result.firstKeyFrame) { + if (pusi) { + if (currentFrame.size !== 0) { + frame = new Uint8Array(currentFrame.size); + i = 0; + while (currentFrame.data.length) { + pes = currentFrame.data.shift(); + frame.set(pes, i); + i += pes.byteLength; + } + if (probe.ts.videoPacketContainsKeyFrame(frame)) { + var firstKeyFrame = probe.ts.parsePesTime(frame); // PTS/DTS may not be available. Simply *not* setting + // the keyframe seems to work fine with HLS playback + // and definitely preferable to a crash with TypeError... + + if (firstKeyFrame) { + result.firstKeyFrame = firstKeyFrame; + result.firstKeyFrame.type = 'video'; + } else { + // eslint-disable-next-line + console.warn('Failed to extract PTS/DTS from PES at first keyframe. ' + 'This could be an unusual TS segment, or else mux.js did not ' + 'parse your TS segment correctly. If you know your TS ' + 'segments do contain PTS/DTS on keyframes please file a bug ' + 'report! You can try ffprobe to double check for yourself.'); + } + } + currentFrame.size = 0; + } + } + currentFrame.data.push(packet); + currentFrame.size += packet.byteLength; + } + } + break; + } + if (endLoop && result.firstKeyFrame) { + break; + } + startIndex += MP2T_PACKET_LENGTH; + endIndex += MP2T_PACKET_LENGTH; + continue; + } // If we get here, we have somehow become de-synchronized and we need to step + // forward one byte at a time until we find a pair of sync bytes that denote + // a packet + + startIndex++; + endIndex++; + } // Start walking from end of segment to get last video packet + + endIndex = bytes.byteLength; + startIndex = endIndex - MP2T_PACKET_LENGTH; + endLoop = false; + while (startIndex >= 0) { + // Look for a pair of start and end sync bytes in the data.. + if (bytes[startIndex] === SYNC_BYTE && bytes[endIndex] === SYNC_BYTE) { + // We found a packet + packet = bytes.subarray(startIndex, endIndex); + type = probe.ts.parseType(packet, pmt.pid); + switch (type) { + case 'pes': + pesType = probe.ts.parsePesType(packet, pmt.table); + pusi = probe.ts.parsePayloadUnitStartIndicator(packet); + if (pesType === 'video' && pusi) { + parsed = probe.ts.parsePesTime(packet); + if (parsed) { + parsed.type = 'video'; + result.video.push(parsed); + endLoop = true; + } + } + break; + } + if (endLoop) { + break; + } + startIndex -= MP2T_PACKET_LENGTH; + endIndex -= MP2T_PACKET_LENGTH; + continue; + } // If we get here, we have somehow become de-synchronized and we need to step + // forward one byte at a time until we find a pair of sync bytes that denote + // a packet + + startIndex--; + endIndex--; + } + }; + /** + * Adjusts the timestamp information for the segment to account for + * rollover and convert to seconds based on pes packet timescale (90khz clock) + */ + + var adjustTimestamp_ = function (segmentInfo, baseTimestamp) { + if (segmentInfo.audio && segmentInfo.audio.length) { + var audioBaseTimestamp = baseTimestamp; + if (typeof audioBaseTimestamp === 'undefined' || isNaN(audioBaseTimestamp)) { + audioBaseTimestamp = segmentInfo.audio[0].dts; + } + segmentInfo.audio.forEach(function (info) { + info.dts = handleRollover(info.dts, audioBaseTimestamp); + info.pts = handleRollover(info.pts, audioBaseTimestamp); // time in seconds + + info.dtsTime = info.dts / ONE_SECOND_IN_TS; + info.ptsTime = info.pts / ONE_SECOND_IN_TS; + }); + } + if (segmentInfo.video && segmentInfo.video.length) { + var videoBaseTimestamp = baseTimestamp; + if (typeof videoBaseTimestamp === 'undefined' || isNaN(videoBaseTimestamp)) { + videoBaseTimestamp = segmentInfo.video[0].dts; + } + segmentInfo.video.forEach(function (info) { + info.dts = handleRollover(info.dts, videoBaseTimestamp); + info.pts = handleRollover(info.pts, videoBaseTimestamp); // time in seconds + + info.dtsTime = info.dts / ONE_SECOND_IN_TS; + info.ptsTime = info.pts / ONE_SECOND_IN_TS; + }); + if (segmentInfo.firstKeyFrame) { + var frame = segmentInfo.firstKeyFrame; + frame.dts = handleRollover(frame.dts, videoBaseTimestamp); + frame.pts = handleRollover(frame.pts, videoBaseTimestamp); // time in seconds + + frame.dtsTime = frame.dts / ONE_SECOND_IN_TS; + frame.ptsTime = frame.pts / ONE_SECOND_IN_TS; + } + } + }; + /** + * inspects the aac data stream for start and end time information + */ + + var inspectAac_ = function (bytes) { + var endLoop = false, + audioCount = 0, + sampleRate = null, + timestamp = null, + frameSize = 0, + byteIndex = 0, + packet; + while (bytes.length - byteIndex >= 3) { + var type = probe.aac.parseType(bytes, byteIndex); + switch (type) { + case 'timed-metadata': + // Exit early because we don't have enough to parse + // the ID3 tag header + if (bytes.length - byteIndex < 10) { + endLoop = true; + break; + } + frameSize = probe.aac.parseId3TagSize(bytes, byteIndex); // Exit early if we don't have enough in the buffer + // to emit a full packet + + if (frameSize > bytes.length) { + endLoop = true; + break; + } + if (timestamp === null) { + packet = bytes.subarray(byteIndex, byteIndex + frameSize); + timestamp = probe.aac.parseAacTimestamp(packet); + } + byteIndex += frameSize; + break; + case 'audio': + // Exit early because we don't have enough to parse + // the ADTS frame header + if (bytes.length - byteIndex < 7) { + endLoop = true; + break; + } + frameSize = probe.aac.parseAdtsSize(bytes, byteIndex); // Exit early if we don't have enough in the buffer + // to emit a full packet + + if (frameSize > bytes.length) { + endLoop = true; + break; + } + if (sampleRate === null) { + packet = bytes.subarray(byteIndex, byteIndex + frameSize); + sampleRate = probe.aac.parseSampleRate(packet); + } + audioCount++; + byteIndex += frameSize; + break; + default: + byteIndex++; + break; + } + if (endLoop) { + return null; + } + } + if (sampleRate === null || timestamp === null) { + return null; + } + var audioTimescale = ONE_SECOND_IN_TS / sampleRate; + var result = { + audio: [{ + type: 'audio', + dts: timestamp, + pts: timestamp + }, { + type: 'audio', + dts: timestamp + audioCount * 1024 * audioTimescale, + pts: timestamp + audioCount * 1024 * audioTimescale + }] + }; + return result; + }; + /** + * inspects the transport stream segment data for start and end time information + * of the audio and video tracks (when present) as well as the first key frame's + * start time. + */ + + var inspectTs_ = function (bytes) { + var pmt = { + pid: null, + table: null + }; + var result = {}; + parsePsi_(bytes, pmt); + for (var pid in pmt.table) { + if (pmt.table.hasOwnProperty(pid)) { + var type = pmt.table[pid]; + switch (type) { + case StreamTypes.H264_STREAM_TYPE: + result.video = []; + parseVideoPes_(bytes, pmt, result); + if (result.video.length === 0) { + delete result.video; + } + break; + case StreamTypes.ADTS_STREAM_TYPE: + result.audio = []; + parseAudioPes_(bytes, pmt, result); + if (result.audio.length === 0) { + delete result.audio; + } + break; + } + } + } + return result; + }; + /** + * Inspects segment byte data and returns an object with start and end timing information + * + * @param {Uint8Array} bytes The segment byte data + * @param {Number} baseTimestamp Relative reference timestamp used when adjusting frame + * timestamps for rollover. This value must be in 90khz clock. + * @return {Object} Object containing start and end frame timing info of segment. + */ + + var inspect = function (bytes, baseTimestamp) { + var isAacData = probe.aac.isLikelyAacData(bytes); + var result; + if (isAacData) { + result = inspectAac_(bytes); + } else { + result = inspectTs_(bytes); + } + if (!result || !result.audio && !result.video) { + return null; + } + adjustTimestamp_(result, baseTimestamp); + return result; + }; + var tsInspector = { + inspect: inspect, + parseAudioPes_: parseAudioPes_ + }; + /* global self */ + + /** + * Re-emits transmuxer events by converting them into messages to the + * world outside the worker. + * + * @param {Object} transmuxer the transmuxer to wire events on + * @private + */ + + const wireTransmuxerEvents = function (self, transmuxer) { + transmuxer.on('data', function (segment) { + // transfer ownership of the underlying ArrayBuffer + // instead of doing a copy to save memory + // ArrayBuffers are transferable but generic TypedArrays are not + // @link https://developer.mozilla.org/en-US/docs/Web/API/Web_Workers_API/Using_web_workers#Passing_data_by_transferring_ownership_(transferable_objects) + const initArray = segment.initSegment; + segment.initSegment = { + data: initArray.buffer, + byteOffset: initArray.byteOffset, + byteLength: initArray.byteLength + }; + const typedArray = segment.data; + segment.data = typedArray.buffer; + self.postMessage({ + action: 'data', + segment, + byteOffset: typedArray.byteOffset, + byteLength: typedArray.byteLength + }, [segment.data]); + }); + transmuxer.on('done', function (data) { + self.postMessage({ + action: 'done' + }); + }); + transmuxer.on('gopInfo', function (gopInfo) { + self.postMessage({ + action: 'gopInfo', + gopInfo + }); + }); + transmuxer.on('videoSegmentTimingInfo', function (timingInfo) { + const videoSegmentTimingInfo = { + start: { + decode: clock$2.videoTsToSeconds(timingInfo.start.dts), + presentation: clock$2.videoTsToSeconds(timingInfo.start.pts) + }, + end: { + decode: clock$2.videoTsToSeconds(timingInfo.end.dts), + presentation: clock$2.videoTsToSeconds(timingInfo.end.pts) + }, + baseMediaDecodeTime: clock$2.videoTsToSeconds(timingInfo.baseMediaDecodeTime) + }; + if (timingInfo.prependedContentDuration) { + videoSegmentTimingInfo.prependedContentDuration = clock$2.videoTsToSeconds(timingInfo.prependedContentDuration); + } + self.postMessage({ + action: 'videoSegmentTimingInfo', + videoSegmentTimingInfo + }); + }); + transmuxer.on('audioSegmentTimingInfo', function (timingInfo) { + // Note that all times for [audio/video]SegmentTimingInfo events are in video clock + const audioSegmentTimingInfo = { + start: { + decode: clock$2.videoTsToSeconds(timingInfo.start.dts), + presentation: clock$2.videoTsToSeconds(timingInfo.start.pts) + }, + end: { + decode: clock$2.videoTsToSeconds(timingInfo.end.dts), + presentation: clock$2.videoTsToSeconds(timingInfo.end.pts) + }, + baseMediaDecodeTime: clock$2.videoTsToSeconds(timingInfo.baseMediaDecodeTime) + }; + if (timingInfo.prependedContentDuration) { + audioSegmentTimingInfo.prependedContentDuration = clock$2.videoTsToSeconds(timingInfo.prependedContentDuration); + } + self.postMessage({ + action: 'audioSegmentTimingInfo', + audioSegmentTimingInfo + }); + }); + transmuxer.on('id3Frame', function (id3Frame) { + self.postMessage({ + action: 'id3Frame', + id3Frame + }); + }); + transmuxer.on('caption', function (caption) { + self.postMessage({ + action: 'caption', + caption + }); + }); + transmuxer.on('trackinfo', function (trackInfo) { + self.postMessage({ + action: 'trackinfo', + trackInfo + }); + }); + transmuxer.on('audioTimingInfo', function (audioTimingInfo) { + // convert to video TS since we prioritize video time over audio + self.postMessage({ + action: 'audioTimingInfo', + audioTimingInfo: { + start: clock$2.videoTsToSeconds(audioTimingInfo.start), + end: clock$2.videoTsToSeconds(audioTimingInfo.end) + } + }); + }); + transmuxer.on('videoTimingInfo', function (videoTimingInfo) { + self.postMessage({ + action: 'videoTimingInfo', + videoTimingInfo: { + start: clock$2.videoTsToSeconds(videoTimingInfo.start), + end: clock$2.videoTsToSeconds(videoTimingInfo.end) + } + }); + }); + transmuxer.on('log', function (log) { + self.postMessage({ + action: 'log', + log + }); + }); + }; + /** + * All incoming messages route through this hash. If no function exists + * to handle an incoming message, then we ignore the message. + * + * @class MessageHandlers + * @param {Object} options the options to initialize with + */ + + class MessageHandlers { + constructor(self, options) { + this.options = options || {}; + this.self = self; + this.init(); + } + /** + * initialize our web worker and wire all the events. + */ + + init() { + if (this.transmuxer) { + this.transmuxer.dispose(); + } + this.transmuxer = new transmuxer.Transmuxer(this.options); + wireTransmuxerEvents(this.self, this.transmuxer); + } + pushMp4Captions(data) { + if (!this.captionParser) { + this.captionParser = new captionParser(); + this.captionParser.init(); + } + const segment = new Uint8Array(data.data, data.byteOffset, data.byteLength); + const parsed = this.captionParser.parse(segment, data.trackIds, data.timescales); + this.self.postMessage({ + action: 'mp4Captions', + captions: parsed && parsed.captions || [], + logs: parsed && parsed.logs || [], + data: segment.buffer + }, [segment.buffer]); + } + probeMp4StartTime({ + timescales, + data + }) { + const startTime = probe$2.startTime(timescales, data); + this.self.postMessage({ + action: 'probeMp4StartTime', + startTime, + data + }, [data.buffer]); + } + probeMp4Tracks({ + data + }) { + const tracks = probe$2.tracks(data); + this.self.postMessage({ + action: 'probeMp4Tracks', + tracks, + data + }, [data.buffer]); + } + /** + * Probes an mp4 segment for EMSG boxes containing ID3 data. + * https://aomediacodec.github.io/id3-emsg/ + * + * @param {Uint8Array} data segment data + * @param {number} offset segment start time + * @return {Object[]} an array of ID3 frames + */ + + probeEmsgID3({ + data, + offset + }) { + const id3Frames = probe$2.getEmsgID3(data, offset); + this.self.postMessage({ + action: 'probeEmsgID3', + id3Frames, + emsgData: data + }, [data.buffer]); + } + /** + * Probe an mpeg2-ts segment to determine the start time of the segment in it's + * internal "media time," as well as whether it contains video and/or audio. + * + * @private + * @param {Uint8Array} bytes - segment bytes + * @param {number} baseStartTime + * Relative reference timestamp used when adjusting frame timestamps for rollover. + * This value should be in seconds, as it's converted to a 90khz clock within the + * function body. + * @return {Object} The start time of the current segment in "media time" as well as + * whether it contains video and/or audio + */ + + probeTs({ + data, + baseStartTime + }) { + const tsStartTime = typeof baseStartTime === 'number' && !isNaN(baseStartTime) ? baseStartTime * clock$2.ONE_SECOND_IN_TS : void 0; + const timeInfo = tsInspector.inspect(data, tsStartTime); + let result = null; + if (timeInfo) { + result = { + // each type's time info comes back as an array of 2 times, start and end + hasVideo: timeInfo.video && timeInfo.video.length === 2 || false, + hasAudio: timeInfo.audio && timeInfo.audio.length === 2 || false + }; + if (result.hasVideo) { + result.videoStart = timeInfo.video[0].ptsTime; + } + if (result.hasAudio) { + result.audioStart = timeInfo.audio[0].ptsTime; + } + } + this.self.postMessage({ + action: 'probeTs', + result, + data + }, [data.buffer]); + } + clearAllMp4Captions() { + if (this.captionParser) { + this.captionParser.clearAllCaptions(); + } + } + clearParsedMp4Captions() { + if (this.captionParser) { + this.captionParser.clearParsedCaptions(); + } + } + /** + * Adds data (a ts segment) to the start of the transmuxer pipeline for + * processing. + * + * @param {ArrayBuffer} data data to push into the muxer + */ + + push(data) { + // Cast array buffer to correct type for transmuxer + const segment = new Uint8Array(data.data, data.byteOffset, data.byteLength); + this.transmuxer.push(segment); + } + /** + * Recreate the transmuxer so that the next segment added via `push` + * start with a fresh transmuxer. + */ + + reset() { + this.transmuxer.reset(); + } + /** + * Set the value that will be used as the `baseMediaDecodeTime` time for the + * next segment pushed in. Subsequent segments will have their `baseMediaDecodeTime` + * set relative to the first based on the PTS values. + * + * @param {Object} data used to set the timestamp offset in the muxer + */ + + setTimestampOffset(data) { + const timestampOffset = data.timestampOffset || 0; + this.transmuxer.setBaseMediaDecodeTime(Math.round(clock$2.secondsToVideoTs(timestampOffset))); + } + setAudioAppendStart(data) { + this.transmuxer.setAudioAppendStart(Math.ceil(clock$2.secondsToVideoTs(data.appendStart))); + } + setRemux(data) { + this.transmuxer.setRemux(data.remux); + } + /** + * Forces the pipeline to finish processing the last segment and emit it's + * results. + * + * @param {Object} data event data, not really used + */ + + flush(data) { + this.transmuxer.flush(); // transmuxed done action is fired after both audio/video pipelines are flushed + + self.postMessage({ + action: 'done', + type: 'transmuxed' + }); + } + endTimeline() { + this.transmuxer.endTimeline(); // transmuxed endedtimeline action is fired after both audio/video pipelines end their + // timelines + + self.postMessage({ + action: 'endedtimeline', + type: 'transmuxed' + }); + } + alignGopsWith(data) { + this.transmuxer.alignGopsWith(data.gopsToAlignWith.slice()); + } + } + /** + * Our web worker interface so that things can talk to mux.js + * that will be running in a web worker. the scope is passed to this by + * webworkify. + * + * @param {Object} self the scope for the web worker + */ + + self.onmessage = function (event) { + if (event.data.action === 'init' && event.data.options) { + this.messageHandlers = new MessageHandlers(self, event.data.options); + return; + } + if (!this.messageHandlers) { + this.messageHandlers = new MessageHandlers(self); + } + if (event.data && event.data.action && event.data.action !== 'init') { + if (this.messageHandlers[event.data.action]) { + this.messageHandlers[event.data.action](event.data); + } + } + }; +})); +var TransmuxWorker = factory(workerCode$1); +/* rollup-plugin-worker-factory end for worker!/home/runner/work/http-streaming/http-streaming/src/transmuxer-worker.js */ + +const handleData_ = (event, transmuxedData, callback) => { + const { + type, + initSegment, + captions, + captionStreams, + metadata, + videoFrameDtsTime, + videoFramePtsTime + } = event.data.segment; + transmuxedData.buffer.push({ + captions, + captionStreams, + metadata + }); + const boxes = event.data.segment.boxes || { + data: event.data.segment.data + }; + const result = { + type, + // cast ArrayBuffer to TypedArray + data: new Uint8Array(boxes.data, boxes.data.byteOffset, boxes.data.byteLength), + initSegment: new Uint8Array(initSegment.data, initSegment.byteOffset, initSegment.byteLength) + }; + if (typeof videoFrameDtsTime !== 'undefined') { + result.videoFrameDtsTime = videoFrameDtsTime; + } + if (typeof videoFramePtsTime !== 'undefined') { + result.videoFramePtsTime = videoFramePtsTime; + } + callback(result); +}; +const handleDone_ = ({ + transmuxedData, + callback +}) => { + // Previously we only returned data on data events, + // not on done events. Clear out the buffer to keep that consistent. + transmuxedData.buffer = []; // all buffers should have been flushed from the muxer, so start processing anything we + // have received + + callback(transmuxedData); +}; +const handleGopInfo_ = (event, transmuxedData) => { + transmuxedData.gopInfo = event.data.gopInfo; +}; +const processTransmux = options => { + const { + transmuxer, + bytes, + audioAppendStart, + gopsToAlignWith, + remux, + onData, + onTrackInfo, + onAudioTimingInfo, + onVideoTimingInfo, + onVideoSegmentTimingInfo, + onAudioSegmentTimingInfo, + onId3, + onCaptions, + onDone, + onEndedTimeline, + onTransmuxerLog, + isEndOfTimeline + } = options; + const transmuxedData = { + buffer: [] + }; + let waitForEndedTimelineEvent = isEndOfTimeline; + const handleMessage = event => { + if (transmuxer.currentTransmux !== options) { + // disposed + return; + } + if (event.data.action === 'data') { + handleData_(event, transmuxedData, onData); + } + if (event.data.action === 'trackinfo') { + onTrackInfo(event.data.trackInfo); + } + if (event.data.action === 'gopInfo') { + handleGopInfo_(event, transmuxedData); + } + if (event.data.action === 'audioTimingInfo') { + onAudioTimingInfo(event.data.audioTimingInfo); + } + if (event.data.action === 'videoTimingInfo') { + onVideoTimingInfo(event.data.videoTimingInfo); + } + if (event.data.action === 'videoSegmentTimingInfo') { + onVideoSegmentTimingInfo(event.data.videoSegmentTimingInfo); + } + if (event.data.action === 'audioSegmentTimingInfo') { + onAudioSegmentTimingInfo(event.data.audioSegmentTimingInfo); + } + if (event.data.action === 'id3Frame') { + onId3([event.data.id3Frame], event.data.id3Frame.dispatchType); + } + if (event.data.action === 'caption') { + onCaptions(event.data.caption); + } + if (event.data.action === 'endedtimeline') { + waitForEndedTimelineEvent = false; + onEndedTimeline(); + } + if (event.data.action === 'log') { + onTransmuxerLog(event.data.log); + } // wait for the transmuxed event since we may have audio and video + + if (event.data.type !== 'transmuxed') { + return; + } // If the "endedtimeline" event has not yet fired, and this segment represents the end + // of a timeline, that means there may still be data events before the segment + // processing can be considerred complete. In that case, the final event should be + // an "endedtimeline" event with the type "transmuxed." + + if (waitForEndedTimelineEvent) { + return; + } + transmuxer.onmessage = null; + handleDone_({ + transmuxedData, + callback: onDone + }); + /* eslint-disable no-use-before-define */ + + dequeue(transmuxer); + /* eslint-enable */ + }; + + transmuxer.onmessage = handleMessage; + if (audioAppendStart) { + transmuxer.postMessage({ + action: 'setAudioAppendStart', + appendStart: audioAppendStart + }); + } // allow empty arrays to be passed to clear out GOPs + + if (Array.isArray(gopsToAlignWith)) { + transmuxer.postMessage({ + action: 'alignGopsWith', + gopsToAlignWith + }); + } + if (typeof remux !== 'undefined') { + transmuxer.postMessage({ + action: 'setRemux', + remux + }); + } + if (bytes.byteLength) { + const buffer = bytes instanceof ArrayBuffer ? bytes : bytes.buffer; + const byteOffset = bytes instanceof ArrayBuffer ? 0 : bytes.byteOffset; + transmuxer.postMessage({ + action: 'push', + // Send the typed-array of data as an ArrayBuffer so that + // it can be sent as a "Transferable" and avoid the costly + // memory copy + data: buffer, + // To recreate the original typed-array, we need information + // about what portion of the ArrayBuffer it was a view into + byteOffset, + byteLength: bytes.byteLength + }, [buffer]); + } + if (isEndOfTimeline) { + transmuxer.postMessage({ + action: 'endTimeline' + }); + } // even if we didn't push any bytes, we have to make sure we flush in case we reached + // the end of the segment + + transmuxer.postMessage({ + action: 'flush' + }); +}; +const dequeue = transmuxer => { + transmuxer.currentTransmux = null; + if (transmuxer.transmuxQueue.length) { + transmuxer.currentTransmux = transmuxer.transmuxQueue.shift(); + if (typeof transmuxer.currentTransmux === 'function') { + transmuxer.currentTransmux(); + } else { + processTransmux(transmuxer.currentTransmux); + } + } +}; +const processAction = (transmuxer, action) => { + transmuxer.postMessage({ + action + }); + dequeue(transmuxer); +}; +const enqueueAction = (action, transmuxer) => { + if (!transmuxer.currentTransmux) { + transmuxer.currentTransmux = action; + processAction(transmuxer, action); + return; + } + transmuxer.transmuxQueue.push(processAction.bind(null, transmuxer, action)); +}; +const reset = transmuxer => { + enqueueAction('reset', transmuxer); +}; +const endTimeline = transmuxer => { + enqueueAction('endTimeline', transmuxer); +}; +const transmux = options => { + if (!options.transmuxer.currentTransmux) { + options.transmuxer.currentTransmux = options; + processTransmux(options); + return; + } + options.transmuxer.transmuxQueue.push(options); +}; +const createTransmuxer = options => { + const transmuxer = new TransmuxWorker(); + transmuxer.currentTransmux = null; + transmuxer.transmuxQueue = []; + const term = transmuxer.terminate; + transmuxer.terminate = () => { + transmuxer.currentTransmux = null; + transmuxer.transmuxQueue.length = 0; + return term.call(transmuxer); + }; + transmuxer.postMessage({ + action: 'init', + options + }); + return transmuxer; +}; +var segmentTransmuxer = { + reset, + endTimeline, + transmux, + createTransmuxer +}; +const workerCallback = function (options) { + const transmuxer = options.transmuxer; + const endAction = options.endAction || options.action; + const callback = options.callback; + const message = _extends({}, options, { + endAction: null, + transmuxer: null, + callback: null + }); + const listenForEndEvent = event => { + if (event.data.action !== endAction) { + return; + } + transmuxer.removeEventListener('message', listenForEndEvent); // transfer ownership of bytes back to us. + + if (event.data.data) { + event.data.data = new Uint8Array(event.data.data, options.byteOffset || 0, options.byteLength || event.data.data.byteLength); + if (options.data) { + options.data = event.data.data; + } + } + callback(event.data); + }; + transmuxer.addEventListener('message', listenForEndEvent); + if (options.data) { + const isArrayBuffer = options.data instanceof ArrayBuffer; + message.byteOffset = isArrayBuffer ? 0 : options.data.byteOffset; + message.byteLength = options.data.byteLength; + const transfers = [isArrayBuffer ? options.data : options.data.buffer]; + transmuxer.postMessage(message, transfers); + } else { + transmuxer.postMessage(message); + } +}; +const REQUEST_ERRORS = { + FAILURE: 2, + TIMEOUT: -101, + ABORTED: -102 +}; +/** + * Abort all requests + * + * @param {Object} activeXhrs - an object that tracks all XHR requests + */ + +const abortAll = activeXhrs => { + activeXhrs.forEach(xhr => { + xhr.abort(); + }); +}; +/** + * Gather important bandwidth stats once a request has completed + * + * @param {Object} request - the XHR request from which to gather stats + */ + +const getRequestStats = request => { + return { + bandwidth: request.bandwidth, + bytesReceived: request.bytesReceived || 0, + roundTripTime: request.roundTripTime || 0 + }; +}; +/** + * If possible gather bandwidth stats as a request is in + * progress + * + * @param {Event} progressEvent - an event object from an XHR's progress event + */ + +const getProgressStats = progressEvent => { + const request = progressEvent.target; + const roundTripTime = Date.now() - request.requestTime; + const stats = { + bandwidth: Infinity, + bytesReceived: 0, + roundTripTime: roundTripTime || 0 + }; + stats.bytesReceived = progressEvent.loaded; // This can result in Infinity if stats.roundTripTime is 0 but that is ok + // because we should only use bandwidth stats on progress to determine when + // abort a request early due to insufficient bandwidth + + stats.bandwidth = Math.floor(stats.bytesReceived / stats.roundTripTime * 8 * 1000); + return stats; +}; +/** + * Handle all error conditions in one place and return an object + * with all the information + * + * @param {Error|null} error - if non-null signals an error occured with the XHR + * @param {Object} request - the XHR request that possibly generated the error + */ + +const handleErrors = (error, request) => { + if (request.timedout) { + return { + status: request.status, + message: 'HLS request timed-out at URL: ' + request.uri, + code: REQUEST_ERRORS.TIMEOUT, + xhr: request + }; + } + if (request.aborted) { + return { + status: request.status, + message: 'HLS request aborted at URL: ' + request.uri, + code: REQUEST_ERRORS.ABORTED, + xhr: request + }; + } + if (error) { + return { + status: request.status, + message: 'HLS request errored at URL: ' + request.uri, + code: REQUEST_ERRORS.FAILURE, + xhr: request + }; + } + if (request.responseType === 'arraybuffer' && request.response.byteLength === 0) { + return { + status: request.status, + message: 'Empty HLS response at URL: ' + request.uri, + code: REQUEST_ERRORS.FAILURE, + xhr: request + }; + } + return null; +}; +/** + * Handle responses for key data and convert the key data to the correct format + * for the decryption step later + * + * @param {Object} segment - a simplified copy of the segmentInfo object + * from SegmentLoader + * @param {Array} objects - objects to add the key bytes to. + * @param {Function} finishProcessingFn - a callback to execute to continue processing + * this request + */ + +const handleKeyResponse = (segment, objects, finishProcessingFn) => (error, request) => { + const response = request.response; + const errorObj = handleErrors(error, request); + if (errorObj) { + return finishProcessingFn(errorObj, segment); + } + if (response.byteLength !== 16) { + return finishProcessingFn({ + status: request.status, + message: 'Invalid HLS key at URL: ' + request.uri, + code: REQUEST_ERRORS.FAILURE, + xhr: request + }, segment); + } + const view = new DataView(response); + const bytes = new Uint32Array([view.getUint32(0), view.getUint32(4), view.getUint32(8), view.getUint32(12)]); + for (let i = 0; i < objects.length; i++) { + objects[i].bytes = bytes; + } + return finishProcessingFn(null, segment); +}; +const parseInitSegment = (segment, callback) => { + const type = detectContainerForBytes(segment.map.bytes); // TODO: We should also handle ts init segments here, but we + // only know how to parse mp4 init segments at the moment + + if (type !== 'mp4') { + const uri = segment.map.resolvedUri || segment.map.uri; + const mediaType = type || 'unknown'; + return callback({ + internal: true, + message: `Found unsupported ${mediaType} container for initialization segment at URL: ${uri}`, + code: REQUEST_ERRORS.FAILURE, + metadata: { + errorType: videojs.Error.UnsupportedMediaInitialization, + mediaType + } + }); + } + workerCallback({ + action: 'probeMp4Tracks', + data: segment.map.bytes, + transmuxer: segment.transmuxer, + callback: ({ + tracks, + data + }) => { + // transfer bytes back to us + segment.map.bytes = data; + tracks.forEach(function (track) { + segment.map.tracks = segment.map.tracks || {}; // only support one track of each type for now + + if (segment.map.tracks[track.type]) { + return; + } + segment.map.tracks[track.type] = track; + if (typeof track.id === 'number' && track.timescale) { + segment.map.timescales = segment.map.timescales || {}; + segment.map.timescales[track.id] = track.timescale; + } + }); + return callback(null); + } + }); +}; +/** + * Handle init-segment responses + * + * @param {Object} segment - a simplified copy of the segmentInfo object + * from SegmentLoader + * @param {Function} finishProcessingFn - a callback to execute to continue processing + * this request + */ + +const handleInitSegmentResponse = ({ + segment, + finishProcessingFn +}) => (error, request) => { + const errorObj = handleErrors(error, request); + if (errorObj) { + return finishProcessingFn(errorObj, segment); + } + const bytes = new Uint8Array(request.response); // init segment is encypted, we will have to wait + // until the key request is done to decrypt. + + if (segment.map.key) { + segment.map.encryptedBytes = bytes; + return finishProcessingFn(null, segment); + } + segment.map.bytes = bytes; + parseInitSegment(segment, function (parseError) { + if (parseError) { + parseError.xhr = request; + parseError.status = request.status; + return finishProcessingFn(parseError, segment); + } + finishProcessingFn(null, segment); + }); +}; +/** + * Response handler for segment-requests being sure to set the correct + * property depending on whether the segment is encryped or not + * Also records and keeps track of stats that are used for ABR purposes + * + * @param {Object} segment - a simplified copy of the segmentInfo object + * from SegmentLoader + * @param {Function} finishProcessingFn - a callback to execute to continue processing + * this request + */ + +const handleSegmentResponse = ({ + segment, + finishProcessingFn, + responseType +}) => (error, request) => { + const errorObj = handleErrors(error, request); + if (errorObj) { + return finishProcessingFn(errorObj, segment); + } + const newBytes = + // although responseText "should" exist, this guard serves to prevent an error being + // thrown for two primary cases: + // 1. the mime type override stops working, or is not implemented for a specific + // browser + // 2. when using mock XHR libraries like sinon that do not allow the override behavior + responseType === 'arraybuffer' || !request.responseText ? request.response : stringToArrayBuffer(request.responseText.substring(segment.lastReachedChar || 0)); + segment.stats = getRequestStats(request); + if (segment.key) { + segment.encryptedBytes = new Uint8Array(newBytes); + } else { + segment.bytes = new Uint8Array(newBytes); + } + return finishProcessingFn(null, segment); +}; +const transmuxAndNotify = ({ + segment, + bytes, + trackInfoFn, + timingInfoFn, + videoSegmentTimingInfoFn, + audioSegmentTimingInfoFn, + id3Fn, + captionsFn, + isEndOfTimeline, + endedTimelineFn, + dataFn, + doneFn, + onTransmuxerLog +}) => { + const fmp4Tracks = segment.map && segment.map.tracks || {}; + const isMuxed = Boolean(fmp4Tracks.audio && fmp4Tracks.video); // Keep references to each function so we can null them out after we're done with them. + // One reason for this is that in the case of full segments, we want to trust start + // times from the probe, rather than the transmuxer. + + let audioStartFn = timingInfoFn.bind(null, segment, 'audio', 'start'); + const audioEndFn = timingInfoFn.bind(null, segment, 'audio', 'end'); + let videoStartFn = timingInfoFn.bind(null, segment, 'video', 'start'); + const videoEndFn = timingInfoFn.bind(null, segment, 'video', 'end'); + const finish = () => transmux({ + bytes, + transmuxer: segment.transmuxer, + audioAppendStart: segment.audioAppendStart, + gopsToAlignWith: segment.gopsToAlignWith, + remux: isMuxed, + onData: result => { + result.type = result.type === 'combined' ? 'video' : result.type; + dataFn(segment, result); + }, + onTrackInfo: trackInfo => { + if (trackInfoFn) { + if (isMuxed) { + trackInfo.isMuxed = true; + } + trackInfoFn(segment, trackInfo); + } + }, + onAudioTimingInfo: audioTimingInfo => { + // we only want the first start value we encounter + if (audioStartFn && typeof audioTimingInfo.start !== 'undefined') { + audioStartFn(audioTimingInfo.start); + audioStartFn = null; + } // we want to continually update the end time + + if (audioEndFn && typeof audioTimingInfo.end !== 'undefined') { + audioEndFn(audioTimingInfo.end); + } + }, + onVideoTimingInfo: videoTimingInfo => { + // we only want the first start value we encounter + if (videoStartFn && typeof videoTimingInfo.start !== 'undefined') { + videoStartFn(videoTimingInfo.start); + videoStartFn = null; + } // we want to continually update the end time + + if (videoEndFn && typeof videoTimingInfo.end !== 'undefined') { + videoEndFn(videoTimingInfo.end); + } + }, + onVideoSegmentTimingInfo: videoSegmentTimingInfo => { + videoSegmentTimingInfoFn(videoSegmentTimingInfo); + }, + onAudioSegmentTimingInfo: audioSegmentTimingInfo => { + audioSegmentTimingInfoFn(audioSegmentTimingInfo); + }, + onId3: (id3Frames, dispatchType) => { + id3Fn(segment, id3Frames, dispatchType); + }, + onCaptions: captions => { + captionsFn(segment, [captions]); + }, + isEndOfTimeline, + onEndedTimeline: () => { + endedTimelineFn(); + }, + onTransmuxerLog, + onDone: result => { + if (!doneFn) { + return; + } + result.type = result.type === 'combined' ? 'video' : result.type; + doneFn(null, segment, result); + } + }); // In the transmuxer, we don't yet have the ability to extract a "proper" start time. + // Meaning cached frame data may corrupt our notion of where this segment + // really starts. To get around this, probe for the info needed. + + workerCallback({ + action: 'probeTs', + transmuxer: segment.transmuxer, + data: bytes, + baseStartTime: segment.baseStartTime, + callback: data => { + segment.bytes = bytes = data.data; + const probeResult = data.result; + if (probeResult) { + trackInfoFn(segment, { + hasAudio: probeResult.hasAudio, + hasVideo: probeResult.hasVideo, + isMuxed + }); + trackInfoFn = null; + } + finish(); + } + }); +}; +const handleSegmentBytes = ({ + segment, + bytes, + trackInfoFn, + timingInfoFn, + videoSegmentTimingInfoFn, + audioSegmentTimingInfoFn, + id3Fn, + captionsFn, + isEndOfTimeline, + endedTimelineFn, + dataFn, + doneFn, + onTransmuxerLog +}) => { + let bytesAsUint8Array = new Uint8Array(bytes); // TODO: + // We should have a handler that fetches the number of bytes required + // to check if something is fmp4. This will allow us to save bandwidth + // because we can only exclude a playlist and abort requests + // by codec after trackinfo triggers. + + if (isLikelyFmp4MediaSegment(bytesAsUint8Array)) { + segment.isFmp4 = true; + const { + tracks + } = segment.map; + const trackInfo = { + isFmp4: true, + hasVideo: !!tracks.video, + hasAudio: !!tracks.audio + }; // if we have a audio track, with a codec that is not set to + // encrypted audio + + if (tracks.audio && tracks.audio.codec && tracks.audio.codec !== 'enca') { + trackInfo.audioCodec = tracks.audio.codec; + } // if we have a video track, with a codec that is not set to + // encrypted video + + if (tracks.video && tracks.video.codec && tracks.video.codec !== 'encv') { + trackInfo.videoCodec = tracks.video.codec; + } + if (tracks.video && tracks.audio) { + trackInfo.isMuxed = true; + } // since we don't support appending fmp4 data on progress, we know we have the full + // segment here + + trackInfoFn(segment, trackInfo); // The probe doesn't provide the segment end time, so only callback with the start + // time. The end time can be roughly calculated by the receiver using the duration. + // + // Note that the start time returned by the probe reflects the baseMediaDecodeTime, as + // that is the true start of the segment (where the playback engine should begin + // decoding). + + const finishLoading = (captions, id3Frames) => { + // if the track still has audio at this point it is only possible + // for it to be audio only. See `tracks.video && tracks.audio` if statement + // above. + // we make sure to use segment.bytes here as that + dataFn(segment, { + data: bytesAsUint8Array, + type: trackInfo.hasAudio && !trackInfo.isMuxed ? 'audio' : 'video' + }); + if (id3Frames && id3Frames.length) { + id3Fn(segment, id3Frames); + } + if (captions && captions.length) { + captionsFn(segment, captions); + } + doneFn(null, segment, {}); + }; + workerCallback({ + action: 'probeMp4StartTime', + timescales: segment.map.timescales, + data: bytesAsUint8Array, + transmuxer: segment.transmuxer, + callback: ({ + data, + startTime + }) => { + // transfer bytes back to us + bytes = data.buffer; + segment.bytes = bytesAsUint8Array = data; + if (trackInfo.hasAudio && !trackInfo.isMuxed) { + timingInfoFn(segment, 'audio', 'start', startTime); + } + if (trackInfo.hasVideo) { + timingInfoFn(segment, 'video', 'start', startTime); + } + workerCallback({ + action: 'probeEmsgID3', + data: bytesAsUint8Array, + transmuxer: segment.transmuxer, + offset: startTime, + callback: ({ + emsgData, + id3Frames + }) => { + // transfer bytes back to us + bytes = emsgData.buffer; + segment.bytes = bytesAsUint8Array = emsgData; // Run through the CaptionParser in case there are captions. + // Initialize CaptionParser if it hasn't been yet + + if (!tracks.video || !emsgData.byteLength || !segment.transmuxer) { + finishLoading(undefined, id3Frames); + return; + } + workerCallback({ + action: 'pushMp4Captions', + endAction: 'mp4Captions', + transmuxer: segment.transmuxer, + data: bytesAsUint8Array, + timescales: segment.map.timescales, + trackIds: [tracks.video.id], + callback: message => { + // transfer bytes back to us + bytes = message.data.buffer; + segment.bytes = bytesAsUint8Array = message.data; + message.logs.forEach(function (log) { + onTransmuxerLog(merge(log, { + stream: 'mp4CaptionParser' + })); + }); + finishLoading(message.captions, id3Frames); + } + }); + } + }); + } + }); + return; + } // VTT or other segments that don't need processing + + if (!segment.transmuxer) { + doneFn(null, segment, {}); + return; + } + if (typeof segment.container === 'undefined') { + segment.container = detectContainerForBytes(bytesAsUint8Array); + } + if (segment.container !== 'ts' && segment.container !== 'aac') { + trackInfoFn(segment, { + hasAudio: false, + hasVideo: false + }); + doneFn(null, segment, {}); + return; + } // ts or aac + + transmuxAndNotify({ + segment, + bytes, + trackInfoFn, + timingInfoFn, + videoSegmentTimingInfoFn, + audioSegmentTimingInfoFn, + id3Fn, + captionsFn, + isEndOfTimeline, + endedTimelineFn, + dataFn, + doneFn, + onTransmuxerLog + }); +}; +const decrypt = function ({ + id, + key, + encryptedBytes, + decryptionWorker +}, callback) { + const decryptionHandler = event => { + if (event.data.source === id) { + decryptionWorker.removeEventListener('message', decryptionHandler); + const decrypted = event.data.decrypted; + callback(new Uint8Array(decrypted.bytes, decrypted.byteOffset, decrypted.byteLength)); + } + }; + decryptionWorker.addEventListener('message', decryptionHandler); + let keyBytes; + if (key.bytes.slice) { + keyBytes = key.bytes.slice(); + } else { + keyBytes = new Uint32Array(Array.prototype.slice.call(key.bytes)); + } // incrementally decrypt the bytes + + decryptionWorker.postMessage(createTransferableMessage({ + source: id, + encrypted: encryptedBytes, + key: keyBytes, + iv: key.iv + }), [encryptedBytes.buffer, keyBytes.buffer]); +}; +/** + * Decrypt the segment via the decryption web worker + * + * @param {WebWorker} decryptionWorker - a WebWorker interface to AES-128 decryption + * routines + * @param {Object} segment - a simplified copy of the segmentInfo object + * from SegmentLoader + * @param {Function} trackInfoFn - a callback that receives track info + * @param {Function} timingInfoFn - a callback that receives timing info + * @param {Function} videoSegmentTimingInfoFn + * a callback that receives video timing info based on media times and + * any adjustments made by the transmuxer + * @param {Function} audioSegmentTimingInfoFn + * a callback that receives audio timing info based on media times and + * any adjustments made by the transmuxer + * @param {boolean} isEndOfTimeline + * true if this segment represents the last segment in a timeline + * @param {Function} endedTimelineFn + * a callback made when a timeline is ended, will only be called if + * isEndOfTimeline is true + * @param {Function} dataFn - a callback that is executed when segment bytes are available + * and ready to use + * @param {Function} doneFn - a callback that is executed after decryption has completed + */ + +const decryptSegment = ({ + decryptionWorker, + segment, + trackInfoFn, + timingInfoFn, + videoSegmentTimingInfoFn, + audioSegmentTimingInfoFn, + id3Fn, + captionsFn, + isEndOfTimeline, + endedTimelineFn, + dataFn, + doneFn, + onTransmuxerLog +}) => { + decrypt({ + id: segment.requestId, + key: segment.key, + encryptedBytes: segment.encryptedBytes, + decryptionWorker + }, decryptedBytes => { + segment.bytes = decryptedBytes; + handleSegmentBytes({ + segment, + bytes: segment.bytes, + trackInfoFn, + timingInfoFn, + videoSegmentTimingInfoFn, + audioSegmentTimingInfoFn, + id3Fn, + captionsFn, + isEndOfTimeline, + endedTimelineFn, + dataFn, + doneFn, + onTransmuxerLog + }); + }); +}; +/** + * This function waits for all XHRs to finish (with either success or failure) + * before continueing processing via it's callback. The function gathers errors + * from each request into a single errors array so that the error status for + * each request can be examined later. + * + * @param {Object} activeXhrs - an object that tracks all XHR requests + * @param {WebWorker} decryptionWorker - a WebWorker interface to AES-128 decryption + * routines + * @param {Function} trackInfoFn - a callback that receives track info + * @param {Function} timingInfoFn - a callback that receives timing info + * @param {Function} videoSegmentTimingInfoFn + * a callback that receives video timing info based on media times and + * any adjustments made by the transmuxer + * @param {Function} audioSegmentTimingInfoFn + * a callback that receives audio timing info based on media times and + * any adjustments made by the transmuxer + * @param {Function} id3Fn - a callback that receives ID3 metadata + * @param {Function} captionsFn - a callback that receives captions + * @param {boolean} isEndOfTimeline + * true if this segment represents the last segment in a timeline + * @param {Function} endedTimelineFn + * a callback made when a timeline is ended, will only be called if + * isEndOfTimeline is true + * @param {Function} dataFn - a callback that is executed when segment bytes are available + * and ready to use + * @param {Function} doneFn - a callback that is executed after all resources have been + * downloaded and any decryption completed + */ + +const waitForCompletion = ({ + activeXhrs, + decryptionWorker, + trackInfoFn, + timingInfoFn, + videoSegmentTimingInfoFn, + audioSegmentTimingInfoFn, + id3Fn, + captionsFn, + isEndOfTimeline, + endedTimelineFn, + dataFn, + doneFn, + onTransmuxerLog +}) => { + let count = 0; + let didError = false; + return (error, segment) => { + if (didError) { + return; + } + if (error) { + didError = true; // If there are errors, we have to abort any outstanding requests + + abortAll(activeXhrs); // Even though the requests above are aborted, and in theory we could wait until we + // handle the aborted events from those requests, there are some cases where we may + // never get an aborted event. For instance, if the network connection is lost and + // there were two requests, the first may have triggered an error immediately, while + // the second request remains unsent. In that case, the aborted algorithm will not + // trigger an abort: see https://xhr.spec.whatwg.org/#the-abort()-method + // + // We also can't rely on the ready state of the XHR, since the request that + // triggered the connection error may also show as a ready state of 0 (unsent). + // Therefore, we have to finish this group of requests immediately after the first + // seen error. + + return doneFn(error, segment); + } + count += 1; + if (count === activeXhrs.length) { + const segmentFinish = function () { + if (segment.encryptedBytes) { + return decryptSegment({ + decryptionWorker, + segment, + trackInfoFn, + timingInfoFn, + videoSegmentTimingInfoFn, + audioSegmentTimingInfoFn, + id3Fn, + captionsFn, + isEndOfTimeline, + endedTimelineFn, + dataFn, + doneFn, + onTransmuxerLog + }); + } // Otherwise, everything is ready just continue + + handleSegmentBytes({ + segment, + bytes: segment.bytes, + trackInfoFn, + timingInfoFn, + videoSegmentTimingInfoFn, + audioSegmentTimingInfoFn, + id3Fn, + captionsFn, + isEndOfTimeline, + endedTimelineFn, + dataFn, + doneFn, + onTransmuxerLog + }); + }; // Keep track of when *all* of the requests have completed + + segment.endOfAllRequests = Date.now(); + if (segment.map && segment.map.encryptedBytes && !segment.map.bytes) { + return decrypt({ + decryptionWorker, + // add -init to the "id" to differentiate between segment + // and init segment decryption, just in case they happen + // at the same time at some point in the future. + id: segment.requestId + '-init', + encryptedBytes: segment.map.encryptedBytes, + key: segment.map.key + }, decryptedBytes => { + segment.map.bytes = decryptedBytes; + parseInitSegment(segment, parseError => { + if (parseError) { + abortAll(activeXhrs); + return doneFn(parseError, segment); + } + segmentFinish(); + }); + }); + } + segmentFinish(); + } + }; +}; +/** + * Calls the abort callback if any request within the batch was aborted. Will only call + * the callback once per batch of requests, even if multiple were aborted. + * + * @param {Object} loadendState - state to check to see if the abort function was called + * @param {Function} abortFn - callback to call for abort + */ + +const handleLoadEnd = ({ + loadendState, + abortFn +}) => event => { + const request = event.target; + if (request.aborted && abortFn && !loadendState.calledAbortFn) { + abortFn(); + loadendState.calledAbortFn = true; + } +}; +/** + * Simple progress event callback handler that gathers some stats before + * executing a provided callback with the `segment` object + * + * @param {Object} segment - a simplified copy of the segmentInfo object + * from SegmentLoader + * @param {Function} progressFn - a callback that is executed each time a progress event + * is received + * @param {Function} trackInfoFn - a callback that receives track info + * @param {Function} timingInfoFn - a callback that receives timing info + * @param {Function} videoSegmentTimingInfoFn + * a callback that receives video timing info based on media times and + * any adjustments made by the transmuxer + * @param {Function} audioSegmentTimingInfoFn + * a callback that receives audio timing info based on media times and + * any adjustments made by the transmuxer + * @param {boolean} isEndOfTimeline + * true if this segment represents the last segment in a timeline + * @param {Function} endedTimelineFn + * a callback made when a timeline is ended, will only be called if + * isEndOfTimeline is true + * @param {Function} dataFn - a callback that is executed when segment bytes are available + * and ready to use + * @param {Event} event - the progress event object from XMLHttpRequest + */ + +const handleProgress = ({ + segment, + progressFn, + trackInfoFn, + timingInfoFn, + videoSegmentTimingInfoFn, + audioSegmentTimingInfoFn, + id3Fn, + captionsFn, + isEndOfTimeline, + endedTimelineFn, + dataFn +}) => event => { + const request = event.target; + if (request.aborted) { + return; + } + segment.stats = merge(segment.stats, getProgressStats(event)); // record the time that we receive the first byte of data + + if (!segment.stats.firstBytesReceivedAt && segment.stats.bytesReceived) { + segment.stats.firstBytesReceivedAt = Date.now(); + } + return progressFn(event, segment); +}; +/** + * Load all resources and does any processing necessary for a media-segment + * + * Features: + * decrypts the media-segment if it has a key uri and an iv + * aborts *all* requests if *any* one request fails + * + * The segment object, at minimum, has the following format: + * { + * resolvedUri: String, + * [transmuxer]: Object, + * [byterange]: { + * offset: Number, + * length: Number + * }, + * [key]: { + * resolvedUri: String + * [byterange]: { + * offset: Number, + * length: Number + * }, + * iv: { + * bytes: Uint32Array + * } + * }, + * [map]: { + * resolvedUri: String, + * [byterange]: { + * offset: Number, + * length: Number + * }, + * [bytes]: Uint8Array + * } + * } + * ...where [name] denotes optional properties + * + * @param {Function} xhr - an instance of the xhr wrapper in xhr.js + * @param {Object} xhrOptions - the base options to provide to all xhr requests + * @param {WebWorker} decryptionWorker - a WebWorker interface to AES-128 + * decryption routines + * @param {Object} segment - a simplified copy of the segmentInfo object + * from SegmentLoader + * @param {Function} abortFn - a callback called (only once) if any piece of a request was + * aborted + * @param {Function} progressFn - a callback that receives progress events from the main + * segment's xhr request + * @param {Function} trackInfoFn - a callback that receives track info + * @param {Function} timingInfoFn - a callback that receives timing info + * @param {Function} videoSegmentTimingInfoFn + * a callback that receives video timing info based on media times and + * any adjustments made by the transmuxer + * @param {Function} audioSegmentTimingInfoFn + * a callback that receives audio timing info based on media times and + * any adjustments made by the transmuxer + * @param {Function} id3Fn - a callback that receives ID3 metadata + * @param {Function} captionsFn - a callback that receives captions + * @param {boolean} isEndOfTimeline + * true if this segment represents the last segment in a timeline + * @param {Function} endedTimelineFn + * a callback made when a timeline is ended, will only be called if + * isEndOfTimeline is true + * @param {Function} dataFn - a callback that receives data from the main segment's xhr + * request, transmuxed if needed + * @param {Function} doneFn - a callback that is executed only once all requests have + * succeeded or failed + * @return {Function} a function that, when invoked, immediately aborts all + * outstanding requests + */ + +const mediaSegmentRequest = ({ + xhr, + xhrOptions, + decryptionWorker, + segment, + abortFn, + progressFn, + trackInfoFn, + timingInfoFn, + videoSegmentTimingInfoFn, + audioSegmentTimingInfoFn, + id3Fn, + captionsFn, + isEndOfTimeline, + endedTimelineFn, + dataFn, + doneFn, + onTransmuxerLog +}) => { + const activeXhrs = []; + const finishProcessingFn = waitForCompletion({ + activeXhrs, + decryptionWorker, + trackInfoFn, + timingInfoFn, + videoSegmentTimingInfoFn, + audioSegmentTimingInfoFn, + id3Fn, + captionsFn, + isEndOfTimeline, + endedTimelineFn, + dataFn, + doneFn, + onTransmuxerLog + }); // optionally, request the decryption key + + if (segment.key && !segment.key.bytes) { + const objects = [segment.key]; + if (segment.map && !segment.map.bytes && segment.map.key && segment.map.key.resolvedUri === segment.key.resolvedUri) { + objects.push(segment.map.key); + } + const keyRequestOptions = merge(xhrOptions, { + uri: segment.key.resolvedUri, + responseType: 'arraybuffer', + requestType: 'segment-key' + }); + const keyRequestCallback = handleKeyResponse(segment, objects, finishProcessingFn); + const keyXhr = xhr(keyRequestOptions, keyRequestCallback); + activeXhrs.push(keyXhr); + } // optionally, request the associated media init segment + + if (segment.map && !segment.map.bytes) { + const differentMapKey = segment.map.key && (!segment.key || segment.key.resolvedUri !== segment.map.key.resolvedUri); + if (differentMapKey) { + const mapKeyRequestOptions = merge(xhrOptions, { + uri: segment.map.key.resolvedUri, + responseType: 'arraybuffer', + requestType: 'segment-key' + }); + const mapKeyRequestCallback = handleKeyResponse(segment, [segment.map.key], finishProcessingFn); + const mapKeyXhr = xhr(mapKeyRequestOptions, mapKeyRequestCallback); + activeXhrs.push(mapKeyXhr); + } + const initSegmentOptions = merge(xhrOptions, { + uri: segment.map.resolvedUri, + responseType: 'arraybuffer', + headers: segmentXhrHeaders(segment.map), + requestType: 'segment-media-initialization' + }); + const initSegmentRequestCallback = handleInitSegmentResponse({ + segment, + finishProcessingFn + }); + const initSegmentXhr = xhr(initSegmentOptions, initSegmentRequestCallback); + activeXhrs.push(initSegmentXhr); + } + const segmentRequestOptions = merge(xhrOptions, { + uri: segment.part && segment.part.resolvedUri || segment.resolvedUri, + responseType: 'arraybuffer', + headers: segmentXhrHeaders(segment), + requestType: 'segment' + }); + const segmentRequestCallback = handleSegmentResponse({ + segment, + finishProcessingFn, + responseType: segmentRequestOptions.responseType + }); + const segmentXhr = xhr(segmentRequestOptions, segmentRequestCallback); + segmentXhr.addEventListener('progress', handleProgress({ + segment, + progressFn, + trackInfoFn, + timingInfoFn, + videoSegmentTimingInfoFn, + audioSegmentTimingInfoFn, + id3Fn, + captionsFn, + isEndOfTimeline, + endedTimelineFn, + dataFn + })); + activeXhrs.push(segmentXhr); // since all parts of the request must be considered, but should not make callbacks + // multiple times, provide a shared state object + + const loadendState = {}; + activeXhrs.forEach(activeXhr => { + activeXhr.addEventListener('loadend', handleLoadEnd({ + loadendState, + abortFn + })); + }); + return () => abortAll(activeXhrs); +}; + +/** + * @file - codecs.js - Handles tasks regarding codec strings such as translating them to + * codec strings, or translating codec strings into objects that can be examined. + */ +const logFn$1 = logger('CodecUtils'); +/** + * Returns a set of codec strings parsed from the playlist or the default + * codec strings if no codecs were specified in the playlist + * + * @param {Playlist} media the current media playlist + * @return {Object} an object with the video and audio codecs + */ + +const getCodecs = function (media) { + // if the codecs were explicitly specified, use them instead of the + // defaults + const mediaAttributes = media.attributes || {}; + if (mediaAttributes.CODECS) { + return parseCodecs(mediaAttributes.CODECS); + } +}; +const isMaat = (main, media) => { + const mediaAttributes = media.attributes || {}; + return main && main.mediaGroups && main.mediaGroups.AUDIO && mediaAttributes.AUDIO && main.mediaGroups.AUDIO[mediaAttributes.AUDIO]; +}; +const isMuxed = (main, media) => { + if (!isMaat(main, media)) { + return true; + } + const mediaAttributes = media.attributes || {}; + const audioGroup = main.mediaGroups.AUDIO[mediaAttributes.AUDIO]; + for (const groupId in audioGroup) { + // If an audio group has a URI (the case for HLS, as HLS will use external playlists), + // or there are listed playlists (the case for DASH, as the manifest will have already + // provided all of the details necessary to generate the audio playlist, as opposed to + // HLS' externally requested playlists), then the content is demuxed. + if (!audioGroup[groupId].uri && !audioGroup[groupId].playlists) { + return true; + } + } + return false; +}; +const unwrapCodecList = function (codecList) { + const codecs = {}; + codecList.forEach(({ + mediaType, + type, + details + }) => { + codecs[mediaType] = codecs[mediaType] || []; + codecs[mediaType].push(translateLegacyCodec(`${type}${details}`)); + }); + Object.keys(codecs).forEach(function (mediaType) { + if (codecs[mediaType].length > 1) { + logFn$1(`multiple ${mediaType} codecs found as attributes: ${codecs[mediaType].join(', ')}. Setting playlist codecs to null so that we wait for mux.js to probe segments for real codecs.`); + codecs[mediaType] = null; + return; + } + codecs[mediaType] = codecs[mediaType][0]; + }); + return codecs; +}; +const codecCount = function (codecObj) { + let count = 0; + if (codecObj.audio) { + count++; + } + if (codecObj.video) { + count++; + } + return count; +}; +/** + * Calculates the codec strings for a working configuration of + * SourceBuffers to play variant streams in a main playlist. If + * there is no possible working configuration, an empty object will be + * returned. + * + * @param main {Object} the m3u8 object for the main playlist + * @param media {Object} the m3u8 object for the variant playlist + * @return {Object} the codec strings. + * + * @private + */ + +const codecsForPlaylist = function (main, media) { + const mediaAttributes = media.attributes || {}; + const codecInfo = unwrapCodecList(getCodecs(media) || []); // HLS with multiple-audio tracks must always get an audio codec. + // Put another way, there is no way to have a video-only multiple-audio HLS! + + if (isMaat(main, media) && !codecInfo.audio) { + if (!isMuxed(main, media)) { + // It is possible for codecs to be specified on the audio media group playlist but + // not on the rendition playlist. This is mostly the case for DASH, where audio and + // video are always separate (and separately specified). + const defaultCodecs = unwrapCodecList(codecsFromDefault(main, mediaAttributes.AUDIO) || []); + if (defaultCodecs.audio) { + codecInfo.audio = defaultCodecs.audio; + } + } + } + return codecInfo; +}; +const logFn = logger('PlaylistSelector'); +const representationToString = function (representation) { + if (!representation || !representation.playlist) { + return; + } + const playlist = representation.playlist; + return JSON.stringify({ + id: playlist.id, + bandwidth: representation.bandwidth, + width: representation.width, + height: representation.height, + codecs: playlist.attributes && playlist.attributes.CODECS || '' + }); +}; // Utilities + +/** + * Returns the CSS value for the specified property on an element + * using `getComputedStyle`. Firefox has a long-standing issue where + * getComputedStyle() may return null when running in an iframe with + * `display: none`. + * + * @see https://bugzilla.mozilla.org/show_bug.cgi?id=548397 + * @param {HTMLElement} el the htmlelement to work on + * @param {string} the proprety to get the style for + */ + +const safeGetComputedStyle = function (el, property) { + if (!el) { + return ''; + } + const result = window$1.getComputedStyle(el); + if (!result) { + return ''; + } + return result[property]; +}; +/** + * Resuable stable sort function + * + * @param {Playlists} array + * @param {Function} sortFn Different comparators + * @function stableSort + */ + +const stableSort = function (array, sortFn) { + const newArray = array.slice(); + array.sort(function (left, right) { + const cmp = sortFn(left, right); + if (cmp === 0) { + return newArray.indexOf(left) - newArray.indexOf(right); + } + return cmp; + }); +}; +/** + * A comparator function to sort two playlist object by bandwidth. + * + * @param {Object} left a media playlist object + * @param {Object} right a media playlist object + * @return {number} Greater than zero if the bandwidth attribute of + * left is greater than the corresponding attribute of right. Less + * than zero if the bandwidth of right is greater than left and + * exactly zero if the two are equal. + */ + +const comparePlaylistBandwidth = function (left, right) { + let leftBandwidth; + let rightBandwidth; + if (left.attributes.BANDWIDTH) { + leftBandwidth = left.attributes.BANDWIDTH; + } + leftBandwidth = leftBandwidth || window$1.Number.MAX_VALUE; + if (right.attributes.BANDWIDTH) { + rightBandwidth = right.attributes.BANDWIDTH; + } + rightBandwidth = rightBandwidth || window$1.Number.MAX_VALUE; + return leftBandwidth - rightBandwidth; +}; +/** + * A comparator function to sort two playlist object by resolution (width). + * + * @param {Object} left a media playlist object + * @param {Object} right a media playlist object + * @return {number} Greater than zero if the resolution.width attribute of + * left is greater than the corresponding attribute of right. Less + * than zero if the resolution.width of right is greater than left and + * exactly zero if the two are equal. + */ + +const comparePlaylistResolution = function (left, right) { + let leftWidth; + let rightWidth; + if (left.attributes.RESOLUTION && left.attributes.RESOLUTION.width) { + leftWidth = left.attributes.RESOLUTION.width; + } + leftWidth = leftWidth || window$1.Number.MAX_VALUE; + if (right.attributes.RESOLUTION && right.attributes.RESOLUTION.width) { + rightWidth = right.attributes.RESOLUTION.width; + } + rightWidth = rightWidth || window$1.Number.MAX_VALUE; // NOTE - Fallback to bandwidth sort as appropriate in cases where multiple renditions + // have the same media dimensions/ resolution + + if (leftWidth === rightWidth && left.attributes.BANDWIDTH && right.attributes.BANDWIDTH) { + return left.attributes.BANDWIDTH - right.attributes.BANDWIDTH; + } + return leftWidth - rightWidth; +}; +/** + * Chooses the appropriate media playlist based on bandwidth and player size + * + * @param {Object} main + * Object representation of the main manifest + * @param {number} playerBandwidth + * Current calculated bandwidth of the player + * @param {number} playerWidth + * Current width of the player element (should account for the device pixel ratio) + * @param {number} playerHeight + * Current height of the player element (should account for the device pixel ratio) + * @param {boolean} limitRenditionByPlayerDimensions + * True if the player width and height should be used during the selection, false otherwise + * @param {Object} playlistController + * the current playlistController object + * @return {Playlist} the highest bitrate playlist less than the + * currently detected bandwidth, accounting for some amount of + * bandwidth variance + */ + +let simpleSelector = function (main, playerBandwidth, playerWidth, playerHeight, limitRenditionByPlayerDimensions, playlistController) { + // If we end up getting called before `main` is available, exit early + if (!main) { + return; + } + const options = { + bandwidth: playerBandwidth, + width: playerWidth, + height: playerHeight, + limitRenditionByPlayerDimensions + }; + let playlists = main.playlists; // if playlist is audio only, select between currently active audio group playlists. + + if (Playlist.isAudioOnly(main)) { + playlists = playlistController.getAudioTrackPlaylists_(); // add audioOnly to options so that we log audioOnly: true + // at the buttom of this function for debugging. + + options.audioOnly = true; + } // convert the playlists to an intermediary representation to make comparisons easier + + let sortedPlaylistReps = playlists.map(playlist => { + let bandwidth; + const width = playlist.attributes && playlist.attributes.RESOLUTION && playlist.attributes.RESOLUTION.width; + const height = playlist.attributes && playlist.attributes.RESOLUTION && playlist.attributes.RESOLUTION.height; + bandwidth = playlist.attributes && playlist.attributes.BANDWIDTH; + bandwidth = bandwidth || window$1.Number.MAX_VALUE; + return { + bandwidth, + width, + height, + playlist + }; + }); + stableSort(sortedPlaylistReps, (left, right) => left.bandwidth - right.bandwidth); // filter out any playlists that have been excluded due to + // incompatible configurations + + sortedPlaylistReps = sortedPlaylistReps.filter(rep => !Playlist.isIncompatible(rep.playlist)); // filter out any playlists that have been disabled manually through the representations + // api or excluded temporarily due to playback errors. + + let enabledPlaylistReps = sortedPlaylistReps.filter(rep => Playlist.isEnabled(rep.playlist)); + if (!enabledPlaylistReps.length) { + // if there are no enabled playlists, then they have all been excluded or disabled + // by the user through the representations api. In this case, ignore exclusion and + // fallback to what the user wants by using playlists the user has not disabled. + enabledPlaylistReps = sortedPlaylistReps.filter(rep => !Playlist.isDisabled(rep.playlist)); + } // filter out any variant that has greater effective bitrate + // than the current estimated bandwidth + + const bandwidthPlaylistReps = enabledPlaylistReps.filter(rep => rep.bandwidth * Config.BANDWIDTH_VARIANCE < playerBandwidth); + let highestRemainingBandwidthRep = bandwidthPlaylistReps[bandwidthPlaylistReps.length - 1]; // get all of the renditions with the same (highest) bandwidth + // and then taking the very first element + + const bandwidthBestRep = bandwidthPlaylistReps.filter(rep => rep.bandwidth === highestRemainingBandwidthRep.bandwidth)[0]; // if we're not going to limit renditions by player size, make an early decision. + + if (limitRenditionByPlayerDimensions === false) { + const chosenRep = bandwidthBestRep || enabledPlaylistReps[0] || sortedPlaylistReps[0]; + if (chosenRep && chosenRep.playlist) { + let type = 'sortedPlaylistReps'; + if (bandwidthBestRep) { + type = 'bandwidthBestRep'; + } + if (enabledPlaylistReps[0]) { + type = 'enabledPlaylistReps'; + } + logFn(`choosing ${representationToString(chosenRep)} using ${type} with options`, options); + return chosenRep.playlist; + } + logFn('could not choose a playlist with options', options); + return null; + } // filter out playlists without resolution information + + const haveResolution = bandwidthPlaylistReps.filter(rep => rep.width && rep.height); // sort variants by resolution + + stableSort(haveResolution, (left, right) => left.width - right.width); // if we have the exact resolution as the player use it + + const resolutionBestRepList = haveResolution.filter(rep => rep.width === playerWidth && rep.height === playerHeight); + highestRemainingBandwidthRep = resolutionBestRepList[resolutionBestRepList.length - 1]; // ensure that we pick the highest bandwidth variant that have exact resolution + + const resolutionBestRep = resolutionBestRepList.filter(rep => rep.bandwidth === highestRemainingBandwidthRep.bandwidth)[0]; + let resolutionPlusOneList; + let resolutionPlusOneSmallest; + let resolutionPlusOneRep; // find the smallest variant that is larger than the player + // if there is no match of exact resolution + + if (!resolutionBestRep) { + resolutionPlusOneList = haveResolution.filter(rep => rep.width > playerWidth || rep.height > playerHeight); // find all the variants have the same smallest resolution + + resolutionPlusOneSmallest = resolutionPlusOneList.filter(rep => rep.width === resolutionPlusOneList[0].width && rep.height === resolutionPlusOneList[0].height); // ensure that we also pick the highest bandwidth variant that + // is just-larger-than the video player + + highestRemainingBandwidthRep = resolutionPlusOneSmallest[resolutionPlusOneSmallest.length - 1]; + resolutionPlusOneRep = resolutionPlusOneSmallest.filter(rep => rep.bandwidth === highestRemainingBandwidthRep.bandwidth)[0]; + } + let leastPixelDiffRep; // If this selector proves to be better than others, + // resolutionPlusOneRep and resolutionBestRep and all + // the code involving them should be removed. + + if (playlistController.leastPixelDiffSelector) { + // find the variant that is closest to the player's pixel size + const leastPixelDiffList = haveResolution.map(rep => { + rep.pixelDiff = Math.abs(rep.width - playerWidth) + Math.abs(rep.height - playerHeight); + return rep; + }); // get the highest bandwidth, closest resolution playlist + + stableSort(leastPixelDiffList, (left, right) => { + // sort by highest bandwidth if pixelDiff is the same + if (left.pixelDiff === right.pixelDiff) { + return right.bandwidth - left.bandwidth; + } + return left.pixelDiff - right.pixelDiff; + }); + leastPixelDiffRep = leastPixelDiffList[0]; + } // fallback chain of variants + + const chosenRep = leastPixelDiffRep || resolutionPlusOneRep || resolutionBestRep || bandwidthBestRep || enabledPlaylistReps[0] || sortedPlaylistReps[0]; + if (chosenRep && chosenRep.playlist) { + let type = 'sortedPlaylistReps'; + if (leastPixelDiffRep) { + type = 'leastPixelDiffRep'; + } else if (resolutionPlusOneRep) { + type = 'resolutionPlusOneRep'; + } else if (resolutionBestRep) { + type = 'resolutionBestRep'; + } else if (bandwidthBestRep) { + type = 'bandwidthBestRep'; + } else if (enabledPlaylistReps[0]) { + type = 'enabledPlaylistReps'; + } + logFn(`choosing ${representationToString(chosenRep)} using ${type} with options`, options); + return chosenRep.playlist; + } + logFn('could not choose a playlist with options', options); + return null; +}; + +/** + * Chooses the appropriate media playlist based on the most recent + * bandwidth estimate and the player size. + * + * Expects to be called within the context of an instance of VhsHandler + * + * @return {Playlist} the highest bitrate playlist less than the + * currently detected bandwidth, accounting for some amount of + * bandwidth variance + */ + +const lastBandwidthSelector = function () { + let pixelRatio = this.useDevicePixelRatio ? window$1.devicePixelRatio || 1 : 1; + if (!isNaN(this.customPixelRatio)) { + pixelRatio = this.customPixelRatio; + } + return simpleSelector(this.playlists.main, this.systemBandwidth, parseInt(safeGetComputedStyle(this.tech_.el(), 'width'), 10) * pixelRatio, parseInt(safeGetComputedStyle(this.tech_.el(), 'height'), 10) * pixelRatio, this.limitRenditionByPlayerDimensions, this.playlistController_); +}; +/** + * Chooses the appropriate media playlist based on an + * exponential-weighted moving average of the bandwidth after + * filtering for player size. + * + * Expects to be called within the context of an instance of VhsHandler + * + * @param {number} decay - a number between 0 and 1. Higher values of + * this parameter will cause previous bandwidth estimates to lose + * significance more quickly. + * @return {Function} a function which can be invoked to create a new + * playlist selector function. + * @see https://en.wikipedia.org/wiki/Moving_average#Exponential_moving_average + */ + +const movingAverageBandwidthSelector = function (decay) { + let average = -1; + let lastSystemBandwidth = -1; + if (decay < 0 || decay > 1) { + throw new Error('Moving average bandwidth decay must be between 0 and 1.'); + } + return function () { + let pixelRatio = this.useDevicePixelRatio ? window$1.devicePixelRatio || 1 : 1; + if (!isNaN(this.customPixelRatio)) { + pixelRatio = this.customPixelRatio; + } + if (average < 0) { + average = this.systemBandwidth; + lastSystemBandwidth = this.systemBandwidth; + } // stop the average value from decaying for every 250ms + // when the systemBandwidth is constant + // and + // stop average from setting to a very low value when the + // systemBandwidth becomes 0 in case of chunk cancellation + + if (this.systemBandwidth > 0 && this.systemBandwidth !== lastSystemBandwidth) { + average = decay * this.systemBandwidth + (1 - decay) * average; + lastSystemBandwidth = this.systemBandwidth; + } + return simpleSelector(this.playlists.main, average, parseInt(safeGetComputedStyle(this.tech_.el(), 'width'), 10) * pixelRatio, parseInt(safeGetComputedStyle(this.tech_.el(), 'height'), 10) * pixelRatio, this.limitRenditionByPlayerDimensions, this.playlistController_); + }; +}; +/** + * Chooses the appropriate media playlist based on the potential to rebuffer + * + * @param {Object} settings + * Object of information required to use this selector + * @param {Object} settings.main + * Object representation of the main manifest + * @param {number} settings.currentTime + * The current time of the player + * @param {number} settings.bandwidth + * Current measured bandwidth + * @param {number} settings.duration + * Duration of the media + * @param {number} settings.segmentDuration + * Segment duration to be used in round trip time calculations + * @param {number} settings.timeUntilRebuffer + * Time left in seconds until the player has to rebuffer + * @param {number} settings.currentTimeline + * The current timeline segments are being loaded from + * @param {SyncController} settings.syncController + * SyncController for determining if we have a sync point for a given playlist + * @return {Object|null} + * {Object} return.playlist + * The highest bandwidth playlist with the least amount of rebuffering + * {Number} return.rebufferingImpact + * The amount of time in seconds switching to this playlist will rebuffer. A + * negative value means that switching will cause zero rebuffering. + */ + +const minRebufferMaxBandwidthSelector = function (settings) { + const { + main, + currentTime, + bandwidth, + duration, + segmentDuration, + timeUntilRebuffer, + currentTimeline, + syncController + } = settings; // filter out any playlists that have been excluded due to + // incompatible configurations + + const compatiblePlaylists = main.playlists.filter(playlist => !Playlist.isIncompatible(playlist)); // filter out any playlists that have been disabled manually through the representations + // api or excluded temporarily due to playback errors. + + let enabledPlaylists = compatiblePlaylists.filter(Playlist.isEnabled); + if (!enabledPlaylists.length) { + // if there are no enabled playlists, then they have all been excluded or disabled + // by the user through the representations api. In this case, ignore exclusion and + // fallback to what the user wants by using playlists the user has not disabled. + enabledPlaylists = compatiblePlaylists.filter(playlist => !Playlist.isDisabled(playlist)); + } + const bandwidthPlaylists = enabledPlaylists.filter(Playlist.hasAttribute.bind(null, 'BANDWIDTH')); + const rebufferingEstimates = bandwidthPlaylists.map(playlist => { + const syncPoint = syncController.getSyncPoint(playlist, duration, currentTimeline, currentTime); // If there is no sync point for this playlist, switching to it will require a + // sync request first. This will double the request time + + const numRequests = syncPoint ? 1 : 2; + const requestTimeEstimate = Playlist.estimateSegmentRequestTime(segmentDuration, bandwidth, playlist); + const rebufferingImpact = requestTimeEstimate * numRequests - timeUntilRebuffer; + return { + playlist, + rebufferingImpact + }; + }); + const noRebufferingPlaylists = rebufferingEstimates.filter(estimate => estimate.rebufferingImpact <= 0); // Sort by bandwidth DESC + + stableSort(noRebufferingPlaylists, (a, b) => comparePlaylistBandwidth(b.playlist, a.playlist)); + if (noRebufferingPlaylists.length) { + return noRebufferingPlaylists[0]; + } + stableSort(rebufferingEstimates, (a, b) => a.rebufferingImpact - b.rebufferingImpact); + return rebufferingEstimates[0] || null; +}; +/** + * Chooses the appropriate media playlist, which in this case is the lowest bitrate + * one with video. If no renditions with video exist, return the lowest audio rendition. + * + * Expects to be called within the context of an instance of VhsHandler + * + * @return {Object|null} + * {Object} return.playlist + * The lowest bitrate playlist that contains a video codec. If no such rendition + * exists pick the lowest audio rendition. + */ + +const lowestBitrateCompatibleVariantSelector = function () { + // filter out any playlists that have been excluded due to + // incompatible configurations or playback errors + const playlists = this.playlists.main.playlists.filter(Playlist.isEnabled); // Sort ascending by bitrate + + stableSort(playlists, (a, b) => comparePlaylistBandwidth(a, b)); // Parse and assume that playlists with no video codec have no video + // (this is not necessarily true, although it is generally true). + // + // If an entire manifest has no valid videos everything will get filtered + // out. + + const playlistsWithVideo = playlists.filter(playlist => !!codecsForPlaylist(this.playlists.main, playlist).video); + return playlistsWithVideo[0] || null; +}; + +/** + * Combine all segments into a single Uint8Array + * + * @param {Object} segmentObj + * @return {Uint8Array} concatenated bytes + * @private + */ +const concatSegments = segmentObj => { + let offset = 0; + let tempBuffer; + if (segmentObj.bytes) { + tempBuffer = new Uint8Array(segmentObj.bytes); // combine the individual segments into one large typed-array + + segmentObj.segments.forEach(segment => { + tempBuffer.set(segment, offset); + offset += segment.byteLength; + }); + } + return tempBuffer; +}; +/** + * Example: + * https://host.com/path1/path2/path3/segment.ts?arg1=val1 + * --> + * path3/segment.ts + * + * @param resolvedUri + * @return {string} + */ + +function compactSegmentUrlDescription(resolvedUri) { + try { + return new URL(resolvedUri).pathname.split('/').slice(-2).join('/'); + } catch (e) { + return ''; + } +} + +/** + * @file text-tracks.js + */ +/** + * Create captions text tracks on video.js if they do not exist + * + * @param {Object} inbandTextTracks a reference to current inbandTextTracks + * @param {Object} tech the video.js tech + * @param {Object} captionStream the caption stream to create + * @private + */ + +const createCaptionsTrackIfNotExists = function (inbandTextTracks, tech, captionStream) { + if (!inbandTextTracks[captionStream]) { + tech.trigger({ + type: 'usage', + name: 'vhs-608' + }); + let instreamId = captionStream; // we need to translate SERVICEn for 708 to how mux.js currently labels them + + if (/^cc708_/.test(captionStream)) { + instreamId = 'SERVICE' + captionStream.split('_')[1]; + } + const track = tech.textTracks().getTrackById(instreamId); + if (track) { + // Resuse an existing track with a CC# id because this was + // very likely created by videojs-contrib-hls from information + // in the m3u8 for us to use + inbandTextTracks[captionStream] = track; + } else { + // This section gets called when we have caption services that aren't specified in the manifest. + // Manifest level caption services are handled in media-groups.js under CLOSED-CAPTIONS. + const captionServices = tech.options_.vhs && tech.options_.vhs.captionServices || {}; + let label = captionStream; + let language = captionStream; + let def = false; + const captionService = captionServices[instreamId]; + if (captionService) { + label = captionService.label; + language = captionService.language; + def = captionService.default; + } // Otherwise, create a track with the default `CC#` label and + // without a language + + inbandTextTracks[captionStream] = tech.addRemoteTextTrack({ + kind: 'captions', + id: instreamId, + // TODO: investigate why this doesn't seem to turn the caption on by default + default: def, + label, + language + }, false).track; + } + } +}; +/** + * Add caption text track data to a source handler given an array of captions + * + * @param {Object} + * @param {Object} inbandTextTracks the inband text tracks + * @param {number} timestampOffset the timestamp offset of the source buffer + * @param {Array} captionArray an array of caption data + * @private + */ + +const addCaptionData = function ({ + inbandTextTracks, + captionArray, + timestampOffset +}) { + if (!captionArray) { + return; + } + const Cue = window$1.WebKitDataCue || window$1.VTTCue; + captionArray.forEach(caption => { + const track = caption.stream; // in CEA 608 captions, video.js/mux.js sends a content array + // with positioning data + + if (caption.content) { + caption.content.forEach(value => { + const cue = new Cue(caption.startTime + timestampOffset, caption.endTime + timestampOffset, value.text); + cue.line = value.line; + cue.align = 'left'; + cue.position = value.position; + cue.positionAlign = 'line-left'; + inbandTextTracks[track].addCue(cue); + }); + } else { + // otherwise, a text value with combined captions is sent + inbandTextTracks[track].addCue(new Cue(caption.startTime + timestampOffset, caption.endTime + timestampOffset, caption.text)); + } + }); +}; +/** + * Define properties on a cue for backwards compatability, + * but warn the user that the way that they are using it + * is depricated and will be removed at a later date. + * + * @param {Cue} cue the cue to add the properties on + * @private + */ + +const deprecateOldCue = function (cue) { + Object.defineProperties(cue.frame, { + id: { + get() { + videojs.log.warn('cue.frame.id is deprecated. Use cue.value.key instead.'); + return cue.value.key; + } + }, + value: { + get() { + videojs.log.warn('cue.frame.value is deprecated. Use cue.value.data instead.'); + return cue.value.data; + } + }, + privateData: { + get() { + videojs.log.warn('cue.frame.privateData is deprecated. Use cue.value.data instead.'); + return cue.value.data; + } + } + }); +}; +/** + * Add metadata text track data to a source handler given an array of metadata + * + * @param {Object} + * @param {Object} inbandTextTracks the inband text tracks + * @param {Array} metadataArray an array of meta data + * @param {number} timestampOffset the timestamp offset of the source buffer + * @param {number} videoDuration the duration of the video + * @private + */ + +const addMetadata = ({ + inbandTextTracks, + metadataArray, + timestampOffset, + videoDuration +}) => { + if (!metadataArray) { + return; + } + const Cue = window$1.WebKitDataCue || window$1.VTTCue; + const metadataTrack = inbandTextTracks.metadataTrack_; + if (!metadataTrack) { + return; + } + metadataArray.forEach(metadata => { + const time = metadata.cueTime + timestampOffset; // if time isn't a finite number between 0 and Infinity, like NaN, + // ignore this bit of metadata. + // This likely occurs when you have an non-timed ID3 tag like TIT2, + // which is the "Title/Songname/Content description" frame + + if (typeof time !== 'number' || window$1.isNaN(time) || time < 0 || !(time < Infinity)) { + return; + } // If we have no frames, we can't create a cue. + + if (!metadata.frames || !metadata.frames.length) { + return; + } + metadata.frames.forEach(frame => { + const cue = new Cue(time, time, frame.value || frame.url || frame.data || ''); + cue.frame = frame; + cue.value = frame; + deprecateOldCue(cue); + metadataTrack.addCue(cue); + }); + }); + if (!metadataTrack.cues || !metadataTrack.cues.length) { + return; + } // Updating the metadeta cues so that + // the endTime of each cue is the startTime of the next cue + // the endTime of last cue is the duration of the video + + const cues = metadataTrack.cues; + const cuesArray = []; // Create a copy of the TextTrackCueList... + // ...disregarding cues with a falsey value + + for (let i = 0; i < cues.length; i++) { + if (cues[i]) { + cuesArray.push(cues[i]); + } + } // Group cues by their startTime value + + const cuesGroupedByStartTime = cuesArray.reduce((obj, cue) => { + const timeSlot = obj[cue.startTime] || []; + timeSlot.push(cue); + obj[cue.startTime] = timeSlot; + return obj; + }, {}); // Sort startTimes by ascending order + + const sortedStartTimes = Object.keys(cuesGroupedByStartTime).sort((a, b) => Number(a) - Number(b)); // Map each cue group's endTime to the next group's startTime + + sortedStartTimes.forEach((startTime, idx) => { + const cueGroup = cuesGroupedByStartTime[startTime]; + const finiteDuration = isFinite(videoDuration) ? videoDuration : startTime; + const nextTime = Number(sortedStartTimes[idx + 1]) || finiteDuration; // Map each cue's endTime the next group's startTime + + cueGroup.forEach(cue => { + cue.endTime = nextTime; + }); + }); +}; // object for mapping daterange attributes + +const dateRangeAttr = { + id: 'ID', + class: 'CLASS', + startDate: 'START-DATE', + duration: 'DURATION', + endDate: 'END-DATE', + endOnNext: 'END-ON-NEXT', + plannedDuration: 'PLANNED-DURATION', + scte35Out: 'SCTE35-OUT', + scte35In: 'SCTE35-IN' +}; +const dateRangeKeysToOmit = new Set(['id', 'class', 'startDate', 'duration', 'endDate', 'endOnNext', 'startTime', 'endTime', 'processDateRange']); +/** + * Add DateRange metadata text track to a source handler given an array of metadata + * + * @param {Object} + * @param {Object} inbandTextTracks the inband text tracks + * @param {Array} dateRanges parsed media playlist + * @private + */ + +const addDateRangeMetadata = ({ + inbandTextTracks, + dateRanges +}) => { + const metadataTrack = inbandTextTracks.metadataTrack_; + if (!metadataTrack) { + return; + } + const Cue = window$1.WebKitDataCue || window$1.VTTCue; + dateRanges.forEach(dateRange => { + // we generate multiple cues for each date range with different attributes + for (const key of Object.keys(dateRange)) { + if (dateRangeKeysToOmit.has(key)) { + continue; + } + const cue = new Cue(dateRange.startTime, dateRange.endTime, ''); + cue.id = dateRange.id; + cue.type = 'com.apple.quicktime.HLS'; + cue.value = { + key: dateRangeAttr[key], + data: dateRange[key] + }; + if (key === 'scte35Out' || key === 'scte35In') { + cue.value.data = new Uint8Array(cue.value.data.match(/[\da-f]{2}/gi)).buffer; + } + metadataTrack.addCue(cue); + } + dateRange.processDateRange(); + }); +}; +/** + * Create metadata text track on video.js if it does not exist + * + * @param {Object} inbandTextTracks a reference to current inbandTextTracks + * @param {string} dispatchType the inband metadata track dispatch type + * @param {Object} tech the video.js tech + * @private + */ + +const createMetadataTrackIfNotExists = (inbandTextTracks, dispatchType, tech) => { + if (inbandTextTracks.metadataTrack_) { + return; + } + inbandTextTracks.metadataTrack_ = tech.addRemoteTextTrack({ + kind: 'metadata', + label: 'Timed Metadata' + }, false).track; + if (!videojs.browser.IS_ANY_SAFARI) { + inbandTextTracks.metadataTrack_.inBandMetadataTrackDispatchType = dispatchType; + } +}; +/** + * Remove cues from a track on video.js. + * + * @param {Double} start start of where we should remove the cue + * @param {Double} end end of where the we should remove the cue + * @param {Object} track the text track to remove the cues from + * @private + */ + +const removeCuesFromTrack = function (start, end, track) { + let i; + let cue; + if (!track) { + return; + } + if (!track.cues) { + return; + } + i = track.cues.length; + while (i--) { + cue = track.cues[i]; // Remove any cue within the provided start and end time + + if (cue.startTime >= start && cue.endTime <= end) { + track.removeCue(cue); + } + } +}; +/** + * Remove duplicate cues from a track on video.js (a cue is considered a + * duplicate if it has the same time interval and text as another) + * + * @param {Object} track the text track to remove the duplicate cues from + * @private + */ + +const removeDuplicateCuesFromTrack = function (track) { + const cues = track.cues; + if (!cues) { + return; + } + const uniqueCues = {}; + for (let i = cues.length - 1; i >= 0; i--) { + const cue = cues[i]; + const cueKey = `${cue.startTime}-${cue.endTime}-${cue.text}`; + if (uniqueCues[cueKey]) { + track.removeCue(cue); + } else { + uniqueCues[cueKey] = cue; + } + } +}; + +/** + * Returns a list of gops in the buffer that have a pts value of 3 seconds or more in + * front of current time. + * + * @param {Array} buffer + * The current buffer of gop information + * @param {number} currentTime + * The current time + * @param {Double} mapping + * Offset to map display time to stream presentation time + * @return {Array} + * List of gops considered safe to append over + */ + +const gopsSafeToAlignWith = (buffer, currentTime, mapping) => { + if (typeof currentTime === 'undefined' || currentTime === null || !buffer.length) { + return []; + } // pts value for current time + 3 seconds to give a bit more wiggle room + + const currentTimePts = Math.ceil((currentTime - mapping + 3) * ONE_SECOND_IN_TS); + let i; + for (i = 0; i < buffer.length; i++) { + if (buffer[i].pts > currentTimePts) { + break; + } + } + return buffer.slice(i); +}; +/** + * Appends gop information (timing and byteLength) received by the transmuxer for the + * gops appended in the last call to appendBuffer + * + * @param {Array} buffer + * The current buffer of gop information + * @param {Array} gops + * List of new gop information + * @param {boolean} replace + * If true, replace the buffer with the new gop information. If false, append the + * new gop information to the buffer in the right location of time. + * @return {Array} + * Updated list of gop information + */ + +const updateGopBuffer = (buffer, gops, replace) => { + if (!gops.length) { + return buffer; + } + if (replace) { + // If we are in safe append mode, then completely overwrite the gop buffer + // with the most recent appeneded data. This will make sure that when appending + // future segments, we only try to align with gops that are both ahead of current + // time and in the last segment appended. + return gops.slice(); + } + const start = gops[0].pts; + let i = 0; + for (i; i < buffer.length; i++) { + if (buffer[i].pts >= start) { + break; + } + } + return buffer.slice(0, i).concat(gops); +}; +/** + * Removes gop information in buffer that overlaps with provided start and end + * + * @param {Array} buffer + * The current buffer of gop information + * @param {Double} start + * position to start the remove at + * @param {Double} end + * position to end the remove at + * @param {Double} mapping + * Offset to map display time to stream presentation time + */ + +const removeGopBuffer = (buffer, start, end, mapping) => { + const startPts = Math.ceil((start - mapping) * ONE_SECOND_IN_TS); + const endPts = Math.ceil((end - mapping) * ONE_SECOND_IN_TS); + const updatedBuffer = buffer.slice(); + let i = buffer.length; + while (i--) { + if (buffer[i].pts <= endPts) { + break; + } + } + if (i === -1) { + // no removal because end of remove range is before start of buffer + return updatedBuffer; + } + let j = i + 1; + while (j--) { + if (buffer[j].pts <= startPts) { + break; + } + } // clamp remove range start to 0 index + + j = Math.max(j, 0); + updatedBuffer.splice(j, i - j + 1); + return updatedBuffer; +}; +const shallowEqual = function (a, b) { + // if both are undefined + // or one or the other is undefined + // they are not equal + if (!a && !b || !a && b || a && !b) { + return false; + } // they are the same object and thus, equal + + if (a === b) { + return true; + } // sort keys so we can make sure they have + // all the same keys later. + + const akeys = Object.keys(a).sort(); + const bkeys = Object.keys(b).sort(); // different number of keys, not equal + + if (akeys.length !== bkeys.length) { + return false; + } + for (let i = 0; i < akeys.length; i++) { + const key = akeys[i]; // different sorted keys, not equal + + if (key !== bkeys[i]) { + return false; + } // different values, not equal + + if (a[key] !== b[key]) { + return false; + } + } + return true; +}; + +// https://www.w3.org/TR/WebIDL-1/#quotaexceedederror +const QUOTA_EXCEEDED_ERR = 22; + +/** + * The segment loader has no recourse except to fetch a segment in the + * current playlist and use the internal timestamps in that segment to + * generate a syncPoint. This function returns a good candidate index + * for that process. + * + * @param {Array} segments - the segments array from a playlist. + * @return {number} An index of a segment from the playlist to load + */ + +const getSyncSegmentCandidate = function (currentTimeline, segments, targetTime) { + segments = segments || []; + const timelineSegments = []; + let time = 0; + for (let i = 0; i < segments.length; i++) { + const segment = segments[i]; + if (currentTimeline === segment.timeline) { + timelineSegments.push(i); + time += segment.duration; + if (time > targetTime) { + return i; + } + } + } + if (timelineSegments.length === 0) { + return 0; + } // default to the last timeline segment + + return timelineSegments[timelineSegments.length - 1]; +}; // In the event of a quota exceeded error, keep at least one second of back buffer. This +// number was arbitrarily chosen and may be updated in the future, but seemed reasonable +// as a start to prevent any potential issues with removing content too close to the +// playhead. + +const MIN_BACK_BUFFER = 1; // in ms + +const CHECK_BUFFER_DELAY = 500; +const finite = num => typeof num === 'number' && isFinite(num); // With most content hovering around 30fps, if a segment has a duration less than a half +// frame at 30fps or one frame at 60fps, the bandwidth and throughput calculations will +// not accurately reflect the rest of the content. + +const MIN_SEGMENT_DURATION_TO_SAVE_STATS = 1 / 60; +const illegalMediaSwitch = (loaderType, startingMedia, trackInfo) => { + // Although these checks should most likely cover non 'main' types, for now it narrows + // the scope of our checks. + if (loaderType !== 'main' || !startingMedia || !trackInfo) { + return null; + } + if (!trackInfo.hasAudio && !trackInfo.hasVideo) { + return 'Neither audio nor video found in segment.'; + } + if (startingMedia.hasVideo && !trackInfo.hasVideo) { + return 'Only audio found in segment when we expected video.' + ' We can\'t switch to audio only from a stream that had video.' + ' To get rid of this message, please add codec information to the manifest.'; + } + if (!startingMedia.hasVideo && trackInfo.hasVideo) { + return 'Video found in segment when we expected only audio.' + ' We can\'t switch to a stream with video from an audio only stream.' + ' To get rid of this message, please add codec information to the manifest.'; + } + return null; +}; +/** + * Calculates a time value that is safe to remove from the back buffer without interrupting + * playback. + * + * @param {TimeRange} seekable + * The current seekable range + * @param {number} currentTime + * The current time of the player + * @param {number} targetDuration + * The target duration of the current playlist + * @return {number} + * Time that is safe to remove from the back buffer without interrupting playback + */ + +const safeBackBufferTrimTime = (seekable, currentTime, targetDuration) => { + // 30 seconds before the playhead provides a safe default for trimming. + // + // Choosing a reasonable default is particularly important for high bitrate content and + // VOD videos/live streams with large windows, as the buffer may end up overfilled and + // throw an APPEND_BUFFER_ERR. + let trimTime = currentTime - Config.BACK_BUFFER_LENGTH; + if (seekable.length) { + // Some live playlists may have a shorter window of content than the full allowed back + // buffer. For these playlists, don't save content that's no longer within the window. + trimTime = Math.max(trimTime, seekable.start(0)); + } // Don't remove within target duration of the current time to avoid the possibility of + // removing the GOP currently being played, as removing it can cause playback stalls. + + const maxTrimTime = currentTime - targetDuration; + return Math.min(maxTrimTime, trimTime); +}; +const segmentInfoString = segmentInfo => { + const { + startOfSegment, + duration, + segment, + part, + playlist: { + mediaSequence: seq, + id, + segments = [] + }, + mediaIndex: index, + partIndex, + timeline + } = segmentInfo; + const segmentLen = segments.length - 1; + let selection = 'mediaIndex/partIndex increment'; + if (segmentInfo.getMediaInfoForTime) { + selection = `getMediaInfoForTime (${segmentInfo.getMediaInfoForTime})`; + } else if (segmentInfo.isSyncRequest) { + selection = 'getSyncSegmentCandidate (isSyncRequest)'; + } + if (segmentInfo.independent) { + selection += ` with independent ${segmentInfo.independent}`; + } + const hasPartIndex = typeof partIndex === 'number'; + const name = segmentInfo.segment.uri ? 'segment' : 'pre-segment'; + const zeroBasedPartCount = hasPartIndex ? getKnownPartCount({ + preloadSegment: segment + }) - 1 : 0; + return `${name} [${seq + index}/${seq + segmentLen}]` + (hasPartIndex ? ` part [${partIndex}/${zeroBasedPartCount}]` : '') + ` segment start/end [${segment.start} => ${segment.end}]` + (hasPartIndex ? ` part start/end [${part.start} => ${part.end}]` : '') + ` startOfSegment [${startOfSegment}]` + ` duration [${duration}]` + ` timeline [${timeline}]` + ` selected by [${selection}]` + ` playlist [${id}]`; +}; +const timingInfoPropertyForMedia = mediaType => `${mediaType}TimingInfo`; +/** + * Returns the timestamp offset to use for the segment. + * + * @param {number} segmentTimeline + * The timeline of the segment + * @param {number} currentTimeline + * The timeline currently being followed by the loader + * @param {number} startOfSegment + * The estimated segment start + * @param {TimeRange[]} buffered + * The loader's buffer + * @param {boolean} overrideCheck + * If true, no checks are made to see if the timestamp offset value should be set, + * but sets it directly to a value. + * + * @return {number|null} + * Either a number representing a new timestamp offset, or null if the segment is + * part of the same timeline + */ + +const timestampOffsetForSegment = ({ + segmentTimeline, + currentTimeline, + startOfSegment, + buffered, + overrideCheck +}) => { + // Check to see if we are crossing a discontinuity to see if we need to set the + // timestamp offset on the transmuxer and source buffer. + // + // Previously, we changed the timestampOffset if the start of this segment was less than + // the currently set timestampOffset, but this isn't desirable as it can produce bad + // behavior, especially around long running live streams. + if (!overrideCheck && segmentTimeline === currentTimeline) { + return null; + } // When changing renditions, it's possible to request a segment on an older timeline. For + // instance, given two renditions with the following: + // + // #EXTINF:10 + // segment1 + // #EXT-X-DISCONTINUITY + // #EXTINF:10 + // segment2 + // #EXTINF:10 + // segment3 + // + // And the current player state: + // + // current time: 8 + // buffer: 0 => 20 + // + // The next segment on the current rendition would be segment3, filling the buffer from + // 20s onwards. However, if a rendition switch happens after segment2 was requested, + // then the next segment to be requested will be segment1 from the new rendition in + // order to fill time 8 and onwards. Using the buffered end would result in repeated + // content (since it would position segment1 of the new rendition starting at 20s). This + // case can be identified when the new segment's timeline is a prior value. Instead of + // using the buffered end, the startOfSegment can be used, which, hopefully, will be + // more accurate to the actual start time of the segment. + + if (segmentTimeline < currentTimeline) { + return startOfSegment; + } // segmentInfo.startOfSegment used to be used as the timestamp offset, however, that + // value uses the end of the last segment if it is available. While this value + // should often be correct, it's better to rely on the buffered end, as the new + // content post discontinuity should line up with the buffered end as if it were + // time 0 for the new content. + + return buffered.length ? buffered.end(buffered.length - 1) : startOfSegment; +}; +/** + * Returns whether or not the loader should wait for a timeline change from the timeline + * change controller before processing the segment. + * + * Primary timing in VHS goes by video. This is different from most media players, as + * audio is more often used as the primary timing source. For the foreseeable future, VHS + * will continue to use video as the primary timing source, due to the current logic and + * expectations built around it. + + * Since the timing follows video, in order to maintain sync, the video loader is + * responsible for setting both audio and video source buffer timestamp offsets. + * + * Setting different values for audio and video source buffers could lead to + * desyncing. The following examples demonstrate some of the situations where this + * distinction is important. Note that all of these cases involve demuxed content. When + * content is muxed, the audio and video are packaged together, therefore syncing + * separate media playlists is not an issue. + * + * CASE 1: Audio prepares to load a new timeline before video: + * + * Timeline: 0 1 + * Audio Segments: 0 1 2 3 4 5 DISCO 6 7 8 9 + * Audio Loader: ^ + * Video Segments: 0 1 2 3 4 5 DISCO 6 7 8 9 + * Video Loader ^ + * + * In the above example, the audio loader is preparing to load the 6th segment, the first + * after a discontinuity, while the video loader is still loading the 5th segment, before + * the discontinuity. + * + * If the audio loader goes ahead and loads and appends the 6th segment before the video + * loader crosses the discontinuity, then when appended, the 6th audio segment will use + * the timestamp offset from timeline 0. This will likely lead to desyncing. In addition, + * the audio loader must provide the audioAppendStart value to trim the content in the + * transmuxer, and that value relies on the audio timestamp offset. Since the audio + * timestamp offset is set by the video (main) loader, the audio loader shouldn't load the + * segment until that value is provided. + * + * CASE 2: Video prepares to load a new timeline before audio: + * + * Timeline: 0 1 + * Audio Segments: 0 1 2 3 4 5 DISCO 6 7 8 9 + * Audio Loader: ^ + * Video Segments: 0 1 2 3 4 5 DISCO 6 7 8 9 + * Video Loader ^ + * + * In the above example, the video loader is preparing to load the 6th segment, the first + * after a discontinuity, while the audio loader is still loading the 5th segment, before + * the discontinuity. + * + * If the video loader goes ahead and loads and appends the 6th segment, then once the + * segment is loaded and processed, both the video and audio timestamp offsets will be + * set, since video is used as the primary timing source. This is to ensure content lines + * up appropriately, as any modifications to the video timing are reflected by audio when + * the video loader sets the audio and video timestamp offsets to the same value. However, + * setting the timestamp offset for audio before audio has had a chance to change + * timelines will likely lead to desyncing, as the audio loader will append segment 5 with + * a timestamp intended to apply to segments from timeline 1 rather than timeline 0. + * + * CASE 3: When seeking, audio prepares to load a new timeline before video + * + * Timeline: 0 1 + * Audio Segments: 0 1 2 3 4 5 DISCO 6 7 8 9 + * Audio Loader: ^ + * Video Segments: 0 1 2 3 4 5 DISCO 6 7 8 9 + * Video Loader ^ + * + * In the above example, both audio and video loaders are loading segments from timeline + * 0, but imagine that the seek originated from timeline 1. + * + * When seeking to a new timeline, the timestamp offset will be set based on the expected + * segment start of the loaded video segment. In order to maintain sync, the audio loader + * must wait for the video loader to load its segment and update both the audio and video + * timestamp offsets before it may load and append its own segment. This is the case + * whether the seek results in a mismatched segment request (e.g., the audio loader + * chooses to load segment 3 and the video loader chooses to load segment 4) or the + * loaders choose to load the same segment index from each playlist, as the segments may + * not be aligned perfectly, even for matching segment indexes. + * + * @param {Object} timelinechangeController + * @param {number} currentTimeline + * The timeline currently being followed by the loader + * @param {number} segmentTimeline + * The timeline of the segment being loaded + * @param {('main'|'audio')} loaderType + * The loader type + * @param {boolean} audioDisabled + * Whether the audio is disabled for the loader. This should only be true when the + * loader may have muxed audio in its segment, but should not append it, e.g., for + * the main loader when an alternate audio playlist is active. + * + * @return {boolean} + * Whether the loader should wait for a timeline change from the timeline change + * controller before processing the segment + */ + +const shouldWaitForTimelineChange = ({ + timelineChangeController, + currentTimeline, + segmentTimeline, + loaderType, + audioDisabled +}) => { + if (currentTimeline === segmentTimeline) { + return false; + } + if (loaderType === 'audio') { + const lastMainTimelineChange = timelineChangeController.lastTimelineChange({ + type: 'main' + }); // Audio loader should wait if: + // + // * main hasn't had a timeline change yet (thus has not loaded its first segment) + // * main hasn't yet changed to the timeline audio is looking to load + + return !lastMainTimelineChange || lastMainTimelineChange.to !== segmentTimeline; + } // The main loader only needs to wait for timeline changes if there's demuxed audio. + // Otherwise, there's nothing to wait for, since audio would be muxed into the main + // loader's segments (or the content is audio/video only and handled by the main + // loader). + + if (loaderType === 'main' && audioDisabled) { + const pendingAudioTimelineChange = timelineChangeController.pendingTimelineChange({ + type: 'audio' + }); // Main loader should wait for the audio loader if audio is not pending a timeline + // change to the current timeline. + // + // Since the main loader is responsible for setting the timestamp offset for both + // audio and video, the main loader must wait for audio to be about to change to its + // timeline before setting the offset, otherwise, if audio is behind in loading, + // segments from the previous timeline would be adjusted by the new timestamp offset. + // + // This requirement means that video will not cross a timeline until the audio is + // about to cross to it, so that way audio and video will always cross the timeline + // together. + // + // In addition to normal timeline changes, these rules also apply to the start of a + // stream (going from a non-existent timeline, -1, to timeline 0). It's important + // that these rules apply to the first timeline change because if they did not, it's + // possible that the main loader will cross two timelines before the audio loader has + // crossed one. Logic may be implemented to handle the startup as a special case, but + // it's easier to simply treat all timeline changes the same. + + if (pendingAudioTimelineChange && pendingAudioTimelineChange.to === segmentTimeline) { + return false; + } + return true; + } + return false; +}; +const mediaDuration = timingInfos => { + let maxDuration = 0; + ['video', 'audio'].forEach(function (type) { + const typeTimingInfo = timingInfos[`${type}TimingInfo`]; + if (!typeTimingInfo) { + return; + } + const { + start, + end + } = typeTimingInfo; + let duration; + if (typeof start === 'bigint' || typeof end === 'bigint') { + duration = window$1.BigInt(end) - window$1.BigInt(start); + } else if (typeof start === 'number' && typeof end === 'number') { + duration = end - start; + } + if (typeof duration !== 'undefined' && duration > maxDuration) { + maxDuration = duration; + } + }); // convert back to a number if it is lower than MAX_SAFE_INTEGER + // as we only need BigInt when we are above that. + + if (typeof maxDuration === 'bigint' && maxDuration < Number.MAX_SAFE_INTEGER) { + maxDuration = Number(maxDuration); + } + return maxDuration; +}; +const segmentTooLong = ({ + segmentDuration, + maxDuration +}) => { + // 0 duration segments are most likely due to metadata only segments or a lack of + // information. + if (!segmentDuration) { + return false; + } // For HLS: + // + // https://tools.ietf.org/html/draft-pantos-http-live-streaming-23#section-4.3.3.1 + // The EXTINF duration of each Media Segment in the Playlist + // file, when rounded to the nearest integer, MUST be less than or equal + // to the target duration; longer segments can trigger playback stalls + // or other errors. + // + // For DASH, the mpd-parser uses the largest reported segment duration as the target + // duration. Although that reported duration is occasionally approximate (i.e., not + // exact), a strict check may report that a segment is too long more often in DASH. + + return Math.round(segmentDuration) > maxDuration + TIME_FUDGE_FACTOR; +}; +const getTroublesomeSegmentDurationMessage = (segmentInfo, sourceType) => { + // Right now we aren't following DASH's timing model exactly, so only perform + // this check for HLS content. + if (sourceType !== 'hls') { + return null; + } + const segmentDuration = mediaDuration({ + audioTimingInfo: segmentInfo.audioTimingInfo, + videoTimingInfo: segmentInfo.videoTimingInfo + }); // Don't report if we lack information. + // + // If the segment has a duration of 0 it is either a lack of information or a + // metadata only segment and shouldn't be reported here. + + if (!segmentDuration) { + return null; + } + const targetDuration = segmentInfo.playlist.targetDuration; + const isSegmentWayTooLong = segmentTooLong({ + segmentDuration, + maxDuration: targetDuration * 2 + }); + const isSegmentSlightlyTooLong = segmentTooLong({ + segmentDuration, + maxDuration: targetDuration + }); + const segmentTooLongMessage = `Segment with index ${segmentInfo.mediaIndex} ` + `from playlist ${segmentInfo.playlist.id} ` + `has a duration of ${segmentDuration} ` + `when the reported duration is ${segmentInfo.duration} ` + `and the target duration is ${targetDuration}. ` + 'For HLS content, a duration in excess of the target duration may result in ' + 'playback issues. See the HLS specification section on EXT-X-TARGETDURATION for ' + 'more details: ' + 'https://tools.ietf.org/html/draft-pantos-http-live-streaming-23#section-4.3.3.1'; + if (isSegmentWayTooLong || isSegmentSlightlyTooLong) { + return { + severity: isSegmentWayTooLong ? 'warn' : 'info', + message: segmentTooLongMessage + }; + } + return null; +}; +/** + * An object that manages segment loading and appending. + * + * @class SegmentLoader + * @param {Object} options required and optional options + * @extends videojs.EventTarget + */ + +class SegmentLoader extends videojs.EventTarget { + constructor(settings, options = {}) { + super(); // check pre-conditions + + if (!settings) { + throw new TypeError('Initialization settings are required'); + } + if (typeof settings.currentTime !== 'function') { + throw new TypeError('No currentTime getter specified'); + } + if (!settings.mediaSource) { + throw new TypeError('No MediaSource specified'); + } // public properties + + this.bandwidth = settings.bandwidth; + this.throughput = { + rate: 0, + count: 0 + }; + this.roundTrip = NaN; + this.resetStats_(); + this.mediaIndex = null; + this.partIndex = null; // private settings + + this.hasPlayed_ = settings.hasPlayed; + this.currentTime_ = settings.currentTime; + this.seekable_ = settings.seekable; + this.seeking_ = settings.seeking; + this.duration_ = settings.duration; + this.mediaSource_ = settings.mediaSource; + this.vhs_ = settings.vhs; + this.loaderType_ = settings.loaderType; + this.currentMediaInfo_ = void 0; + this.startingMediaInfo_ = void 0; + this.segmentMetadataTrack_ = settings.segmentMetadataTrack; + this.goalBufferLength_ = settings.goalBufferLength; + this.sourceType_ = settings.sourceType; + this.sourceUpdater_ = settings.sourceUpdater; + this.inbandTextTracks_ = settings.inbandTextTracks; + this.state_ = 'INIT'; + this.timelineChangeController_ = settings.timelineChangeController; + this.shouldSaveSegmentTimingInfo_ = true; + this.parse708captions_ = settings.parse708captions; + this.useDtsForTimestampOffset_ = settings.useDtsForTimestampOffset; + this.captionServices_ = settings.captionServices; + this.exactManifestTimings = settings.exactManifestTimings; + this.addMetadataToTextTrack = settings.addMetadataToTextTrack; // private instance variables + + this.checkBufferTimeout_ = null; + this.error_ = void 0; + this.currentTimeline_ = -1; + this.shouldForceTimestampOffsetAfterResync_ = false; + this.pendingSegment_ = null; + this.xhrOptions_ = null; + this.pendingSegments_ = []; + this.audioDisabled_ = false; + this.isPendingTimestampOffset_ = false; // TODO possibly move gopBuffer and timeMapping info to a separate controller + + this.gopBuffer_ = []; + this.timeMapping_ = 0; + this.safeAppend_ = false; + this.appendInitSegment_ = { + audio: true, + video: true + }; + this.playlistOfLastInitSegment_ = { + audio: null, + video: null + }; + this.callQueue_ = []; // If the segment loader prepares to load a segment, but does not have enough + // information yet to start the loading process (e.g., if the audio loader wants to + // load a segment from the next timeline but the main loader hasn't yet crossed that + // timeline), then the load call will be added to the queue until it is ready to be + // processed. + + this.loadQueue_ = []; + this.metadataQueue_ = { + id3: [], + caption: [] + }; + this.waitingOnRemove_ = false; + this.quotaExceededErrorRetryTimeout_ = null; // Fragmented mp4 playback + + this.activeInitSegmentId_ = null; + this.initSegments_ = {}; // HLSe playback + + this.cacheEncryptionKeys_ = settings.cacheEncryptionKeys; + this.keyCache_ = {}; + this.decrypter_ = settings.decrypter; // Manages the tracking and generation of sync-points, mappings + // between a time in the display time and a segment index within + // a playlist + + this.syncController_ = settings.syncController; + this.syncPoint_ = { + segmentIndex: 0, + time: 0 + }; + this.transmuxer_ = this.createTransmuxer_(); + this.triggerSyncInfoUpdate_ = () => this.trigger('syncinfoupdate'); + this.syncController_.on('syncinfoupdate', this.triggerSyncInfoUpdate_); + this.mediaSource_.addEventListener('sourceopen', () => { + if (!this.isEndOfStream_()) { + this.ended_ = false; + } + }); // ...for determining the fetch location + + this.fetchAtBuffer_ = false; + this.logger_ = logger(`SegmentLoader[${this.loaderType_}]`); + Object.defineProperty(this, 'state', { + get() { + return this.state_; + }, + set(newState) { + if (newState !== this.state_) { + this.logger_(`${this.state_} -> ${newState}`); + this.state_ = newState; + this.trigger('statechange'); + } + } + }); + this.sourceUpdater_.on('ready', () => { + if (this.hasEnoughInfoToAppend_()) { + this.processCallQueue_(); + } + }); // Only the main loader needs to listen for pending timeline changes, as the main + // loader should wait for audio to be ready to change its timeline so that both main + // and audio timelines change together. For more details, see the + // shouldWaitForTimelineChange function. + + if (this.loaderType_ === 'main') { + this.timelineChangeController_.on('pendingtimelinechange', () => { + if (this.hasEnoughInfoToAppend_()) { + this.processCallQueue_(); + } + }); + } // The main loader only listens on pending timeline changes, but the audio loader, + // since its loads follow main, needs to listen on timeline changes. For more details, + // see the shouldWaitForTimelineChange function. + + if (this.loaderType_ === 'audio') { + this.timelineChangeController_.on('timelinechange', () => { + if (this.hasEnoughInfoToLoad_()) { + this.processLoadQueue_(); + } + if (this.hasEnoughInfoToAppend_()) { + this.processCallQueue_(); + } + }); + } + } + /** + * TODO: Current sync controller consists of many hls-specific strategies + * media sequence sync is also hls-specific, and we would like to be protocol-agnostic on this level + * this should be a part of the sync-controller and sync controller should expect different strategy list based on the protocol. + * + * @return {MediaSequenceSync|null} + * @private + */ + + get mediaSequenceSync_() { + return this.syncController_.getMediaSequenceSync(this.loaderType_); + } + createTransmuxer_() { + return segmentTransmuxer.createTransmuxer({ + remux: false, + alignGopsAtEnd: this.safeAppend_, + keepOriginalTimestamps: true, + parse708captions: this.parse708captions_, + captionServices: this.captionServices_ + }); + } + /** + * reset all of our media stats + * + * @private + */ + + resetStats_() { + this.mediaBytesTransferred = 0; + this.mediaRequests = 0; + this.mediaRequestsAborted = 0; + this.mediaRequestsTimedout = 0; + this.mediaRequestsErrored = 0; + this.mediaTransferDuration = 0; + this.mediaSecondsLoaded = 0; + this.mediaAppends = 0; + } + /** + * dispose of the SegmentLoader and reset to the default state + */ + + dispose() { + this.trigger('dispose'); + this.state = 'DISPOSED'; + this.pause(); + this.abort_(); + if (this.transmuxer_) { + this.transmuxer_.terminate(); + } + this.resetStats_(); + if (this.checkBufferTimeout_) { + window$1.clearTimeout(this.checkBufferTimeout_); + } + if (this.syncController_ && this.triggerSyncInfoUpdate_) { + this.syncController_.off('syncinfoupdate', this.triggerSyncInfoUpdate_); + } + this.off(); + } + setAudio(enable) { + this.audioDisabled_ = !enable; + if (enable) { + this.appendInitSegment_.audio = true; + } else { + // remove current track audio if it gets disabled + this.sourceUpdater_.removeAudio(0, this.duration_()); + } + } + /** + * abort anything that is currently doing on with the SegmentLoader + * and reset to a default state + */ + + abort() { + if (this.state !== 'WAITING') { + if (this.pendingSegment_) { + this.pendingSegment_ = null; + } + return; + } + this.abort_(); // We aborted the requests we were waiting on, so reset the loader's state to READY + // since we are no longer "waiting" on any requests. XHR callback is not always run + // when the request is aborted. This will prevent the loader from being stuck in the + // WAITING state indefinitely. + + this.state = 'READY'; // don't wait for buffer check timeouts to begin fetching the + // next segment + + if (!this.paused()) { + this.monitorBuffer_(); + } + } + /** + * abort all pending xhr requests and null any pending segements + * + * @private + */ + + abort_() { + if (this.pendingSegment_ && this.pendingSegment_.abortRequests) { + this.pendingSegment_.abortRequests(); + } // clear out the segment being processed + + this.pendingSegment_ = null; + this.callQueue_ = []; + this.loadQueue_ = []; + this.metadataQueue_.id3 = []; + this.metadataQueue_.caption = []; + this.timelineChangeController_.clearPendingTimelineChange(this.loaderType_); + this.waitingOnRemove_ = false; + window$1.clearTimeout(this.quotaExceededErrorRetryTimeout_); + this.quotaExceededErrorRetryTimeout_ = null; + } + checkForAbort_(requestId) { + // If the state is APPENDING, then aborts will not modify the state, meaning the first + // callback that happens should reset the state to READY so that loading can continue. + if (this.state === 'APPENDING' && !this.pendingSegment_) { + this.state = 'READY'; + return true; + } + if (!this.pendingSegment_ || this.pendingSegment_.requestId !== requestId) { + return true; + } + return false; + } + /** + * set an error on the segment loader and null out any pending segements + * + * @param {Error} error the error to set on the SegmentLoader + * @return {Error} the error that was set or that is currently set + */ + + error(error) { + if (typeof error !== 'undefined') { + this.logger_('error occurred:', error); + this.error_ = error; + } + this.pendingSegment_ = null; + return this.error_; + } + endOfStream() { + this.ended_ = true; + if (this.transmuxer_) { + // need to clear out any cached data to prepare for the new segment + segmentTransmuxer.reset(this.transmuxer_); + } + this.gopBuffer_.length = 0; + this.pause(); + this.trigger('ended'); + } + /** + * Indicates which time ranges are buffered + * + * @return {TimeRange} + * TimeRange object representing the current buffered ranges + */ + + buffered_() { + const trackInfo = this.getMediaInfo_(); + if (!this.sourceUpdater_ || !trackInfo) { + return createTimeRanges(); + } + if (this.loaderType_ === 'main') { + const { + hasAudio, + hasVideo, + isMuxed + } = trackInfo; + if (hasVideo && hasAudio && !this.audioDisabled_ && !isMuxed) { + return this.sourceUpdater_.buffered(); + } + if (hasVideo) { + return this.sourceUpdater_.videoBuffered(); + } + } // One case that can be ignored for now is audio only with alt audio, + // as we don't yet have proper support for that. + + return this.sourceUpdater_.audioBuffered(); + } + /** + * Gets and sets init segment for the provided map + * + * @param {Object} map + * The map object representing the init segment to get or set + * @param {boolean=} set + * If true, the init segment for the provided map should be saved + * @return {Object} + * map object for desired init segment + */ + + initSegmentForMap(map, set = false) { + if (!map) { + return null; + } + const id = initSegmentId(map); + let storedMap = this.initSegments_[id]; + if (set && !storedMap && map.bytes) { + this.initSegments_[id] = storedMap = { + resolvedUri: map.resolvedUri, + byterange: map.byterange, + bytes: map.bytes, + tracks: map.tracks, + timescales: map.timescales + }; + } + return storedMap || map; + } + /** + * Gets and sets key for the provided key + * + * @param {Object} key + * The key object representing the key to get or set + * @param {boolean=} set + * If true, the key for the provided key should be saved + * @return {Object} + * Key object for desired key + */ + + segmentKey(key, set = false) { + if (!key) { + return null; + } + const id = segmentKeyId(key); + let storedKey = this.keyCache_[id]; // TODO: We should use the HTTP Expires header to invalidate our cache per + // https://tools.ietf.org/html/draft-pantos-http-live-streaming-23#section-6.2.3 + + if (this.cacheEncryptionKeys_ && set && !storedKey && key.bytes) { + this.keyCache_[id] = storedKey = { + resolvedUri: key.resolvedUri, + bytes: key.bytes + }; + } + const result = { + resolvedUri: (storedKey || key).resolvedUri + }; + if (storedKey) { + result.bytes = storedKey.bytes; + } + return result; + } + /** + * Returns true if all configuration required for loading is present, otherwise false. + * + * @return {boolean} True if the all configuration is ready for loading + * @private + */ + + couldBeginLoading_() { + return this.playlist_ && !this.paused(); + } + /** + * load a playlist and start to fill the buffer + */ + + load() { + // un-pause + this.monitorBuffer_(); // if we don't have a playlist yet, keep waiting for one to be + // specified + + if (!this.playlist_) { + return; + } // if all the configuration is ready, initialize and begin loading + + if (this.state === 'INIT' && this.couldBeginLoading_()) { + return this.init_(); + } // if we're in the middle of processing a segment already, don't + // kick off an additional segment request + + if (!this.couldBeginLoading_() || this.state !== 'READY' && this.state !== 'INIT') { + return; + } + this.state = 'READY'; + } + /** + * Once all the starting parameters have been specified, begin + * operation. This method should only be invoked from the INIT + * state. + * + * @private + */ + + init_() { + this.state = 'READY'; // if this is the audio segment loader, and it hasn't been inited before, then any old + // audio data from the muxed content should be removed + + this.resetEverything(); + return this.monitorBuffer_(); + } + /** + * set a playlist on the segment loader + * + * @param {PlaylistLoader} media the playlist to set on the segment loader + */ + + playlist(newPlaylist, options = {}) { + if (!newPlaylist) { + return; + } + const oldPlaylist = this.playlist_; + const segmentInfo = this.pendingSegment_; + this.playlist_ = newPlaylist; + this.xhrOptions_ = options; // when we haven't started playing yet, the start of a live playlist + // is always our zero-time so force a sync update each time the playlist + // is refreshed from the server + // + // Use the INIT state to determine if playback has started, as the playlist sync info + // should be fixed once requests begin (as sync points are generated based on sync + // info), but not before then. + + if (this.state === 'INIT') { + newPlaylist.syncInfo = { + mediaSequence: newPlaylist.mediaSequence, + time: 0 + }; // Setting the date time mapping means mapping the program date time (if available) + // to time 0 on the player's timeline. The playlist's syncInfo serves a similar + // purpose, mapping the initial mediaSequence to time zero. Since the syncInfo can + // be updated as the playlist is refreshed before the loader starts loading, the + // program date time mapping needs to be updated as well. + // + // This mapping is only done for the main loader because a program date time should + // map equivalently between playlists. + + if (this.loaderType_ === 'main') { + this.syncController_.setDateTimeMappingForStart(newPlaylist); + } + } + let oldId = null; + if (oldPlaylist) { + if (oldPlaylist.id) { + oldId = oldPlaylist.id; + } else if (oldPlaylist.uri) { + oldId = oldPlaylist.uri; + } + } + this.logger_(`playlist update [${oldId} => ${newPlaylist.id || newPlaylist.uri}]`); + if (this.mediaSequenceSync_) { + this.mediaSequenceSync_.update(newPlaylist, this.currentTime_()); + this.logger_(`Playlist update: +currentTime: ${this.currentTime_()} +bufferedEnd: ${lastBufferedEnd(this.buffered_())} +`, this.mediaSequenceSync_.diagnostics); + } // in VOD, this is always a rendition switch (or we updated our syncInfo above) + // in LIVE, we always want to update with new playlists (including refreshes) + + this.trigger('syncinfoupdate'); // if we were unpaused but waiting for a playlist, start + // buffering now + + if (this.state === 'INIT' && this.couldBeginLoading_()) { + return this.init_(); + } + if (!oldPlaylist || oldPlaylist.uri !== newPlaylist.uri) { + if (this.mediaIndex !== null) { + // we must reset/resync the segment loader when we switch renditions and + // the segment loader is already synced to the previous rendition + // We only want to reset the loader here for LLHLS playback, as resetLoader sets fetchAtBuffer_ + // to false, resulting in fetching segments at currentTime and causing repeated + // same-segment requests on playlist change. This erroneously drives up the playback watcher + // stalled segment count, as re-requesting segments at the currentTime or browser cached segments + // will not change the buffer. + // Reference for LLHLS fixes: https://github.com/videojs/http-streaming/pull/1201 + const isLLHLS = !newPlaylist.endList && typeof newPlaylist.partTargetDuration === 'number'; + if (isLLHLS) { + this.resetLoader(); + } else { + this.resyncLoader(); + } + } + this.currentMediaInfo_ = void 0; + this.trigger('playlistupdate'); // the rest of this function depends on `oldPlaylist` being defined + + return; + } // we reloaded the same playlist so we are in a live scenario + // and we will likely need to adjust the mediaIndex + + const mediaSequenceDiff = newPlaylist.mediaSequence - oldPlaylist.mediaSequence; + this.logger_(`live window shift [${mediaSequenceDiff}]`); // update the mediaIndex on the SegmentLoader + // this is important because we can abort a request and this value must be + // equal to the last appended mediaIndex + + if (this.mediaIndex !== null) { + this.mediaIndex -= mediaSequenceDiff; // this can happen if we are going to load the first segment, but get a playlist + // update during that. mediaIndex would go from 0 to -1 if mediaSequence in the + // new playlist was incremented by 1. + + if (this.mediaIndex < 0) { + this.mediaIndex = null; + this.partIndex = null; + } else { + const segment = this.playlist_.segments[this.mediaIndex]; // partIndex should remain the same for the same segment + // unless parts fell off of the playlist for this segment. + // In that case we need to reset partIndex and resync + + if (this.partIndex && (!segment.parts || !segment.parts.length || !segment.parts[this.partIndex])) { + const mediaIndex = this.mediaIndex; + this.logger_(`currently processing part (index ${this.partIndex}) no longer exists.`); + this.resetLoader(); // We want to throw away the partIndex and the data associated with it, + // as the part was dropped from our current playlists segment. + // The mediaIndex will still be valid so keep that around. + + this.mediaIndex = mediaIndex; + } + } + } // update the mediaIndex on the SegmentInfo object + // this is important because we will update this.mediaIndex with this value + // in `handleAppendsDone_` after the segment has been successfully appended + + if (segmentInfo) { + segmentInfo.mediaIndex -= mediaSequenceDiff; + if (segmentInfo.mediaIndex < 0) { + segmentInfo.mediaIndex = null; + segmentInfo.partIndex = null; + } else { + // we need to update the referenced segment so that timing information is + // saved for the new playlist's segment, however, if the segment fell off the + // playlist, we can leave the old reference and just lose the timing info + if (segmentInfo.mediaIndex >= 0) { + segmentInfo.segment = newPlaylist.segments[segmentInfo.mediaIndex]; + } + if (segmentInfo.partIndex >= 0 && segmentInfo.segment.parts) { + segmentInfo.part = segmentInfo.segment.parts[segmentInfo.partIndex]; + } + } + } + this.syncController_.saveExpiredSegmentInfo(oldPlaylist, newPlaylist); + } + /** + * Prevent the loader from fetching additional segments. If there + * is a segment request outstanding, it will finish processing + * before the loader halts. A segment loader can be unpaused by + * calling load(). + */ + + pause() { + if (this.checkBufferTimeout_) { + window$1.clearTimeout(this.checkBufferTimeout_); + this.checkBufferTimeout_ = null; + } + } + /** + * Returns whether the segment loader is fetching additional + * segments when given the opportunity. This property can be + * modified through calls to pause() and load(). + */ + + paused() { + return this.checkBufferTimeout_ === null; + } + /** + * Delete all the buffered data and reset the SegmentLoader + * + * @param {Function} [done] an optional callback to be executed when the remove + * operation is complete + */ + + resetEverything(done) { + this.ended_ = false; + this.activeInitSegmentId_ = null; + this.appendInitSegment_ = { + audio: true, + video: true + }; + this.resetLoader(); // remove from 0, the earliest point, to Infinity, to signify removal of everything. + // VTT Segment Loader doesn't need to do anything but in the regular SegmentLoader, + // we then clamp the value to duration if necessary. + + this.remove(0, Infinity, done); // clears fmp4 captions + + if (this.transmuxer_) { + this.transmuxer_.postMessage({ + action: 'clearAllMp4Captions' + }); // reset the cache in the transmuxer + + this.transmuxer_.postMessage({ + action: 'reset' + }); + } + } + /** + * Force the SegmentLoader to resync and start loading around the currentTime instead + * of starting at the end of the buffer + * + * Useful for fast quality changes + */ + + resetLoader() { + this.fetchAtBuffer_ = false; + if (this.mediaSequenceSync_) { + this.mediaSequenceSync_.resetAppendedStatus(); + } + this.resyncLoader(); + } + /** + * Force the SegmentLoader to restart synchronization and make a conservative guess + * before returning to the simple walk-forward method + */ + + resyncLoader() { + if (this.transmuxer_) { + // need to clear out any cached data to prepare for the new segment + segmentTransmuxer.reset(this.transmuxer_); + } + this.mediaIndex = null; + this.partIndex = null; + this.syncPoint_ = null; + this.isPendingTimestampOffset_ = false; // this is mainly to sync timing-info when switching between renditions with and without timestamp-rollover, + // so we don't want it for DASH or fragmented mp4 segments. + + const isFmp4 = this.currentMediaInfo_ && this.currentMediaInfo_.isFmp4; + const isHlsTs = this.sourceType_ === 'hls' && !isFmp4; + if (isHlsTs) { + this.shouldForceTimestampOffsetAfterResync_ = true; + } + this.callQueue_ = []; + this.loadQueue_ = []; + this.metadataQueue_.id3 = []; + this.metadataQueue_.caption = []; + this.abort(); + if (this.transmuxer_) { + this.transmuxer_.postMessage({ + action: 'clearParsedMp4Captions' + }); + } + } + /** + * Remove any data in the source buffer between start and end times + * + * @param {number} start - the start time of the region to remove from the buffer + * @param {number} end - the end time of the region to remove from the buffer + * @param {Function} [done] - an optional callback to be executed when the remove + * @param {boolean} force - force all remove operations to happen + * operation is complete + */ + + remove(start, end, done = () => {}, force = false) { + // clamp end to duration if we need to remove everything. + // This is due to a browser bug that causes issues if we remove to Infinity. + // videojs/videojs-contrib-hls#1225 + if (end === Infinity) { + end = this.duration_(); + } // skip removes that would throw an error + // commonly happens during a rendition switch at the start of a video + // from start 0 to end 0 + + if (end <= start) { + this.logger_('skipping remove because end ${end} is <= start ${start}'); + return; + } + if (!this.sourceUpdater_ || !this.getMediaInfo_()) { + this.logger_('skipping remove because no source updater or starting media info'); // nothing to remove if we haven't processed any media + + return; + } // set it to one to complete this function's removes + + let removesRemaining = 1; + const removeFinished = () => { + removesRemaining--; + if (removesRemaining === 0) { + done(); + } + }; + if (force || !this.audioDisabled_) { + removesRemaining++; + this.sourceUpdater_.removeAudio(start, end, removeFinished); + } // While it would be better to only remove video if the main loader has video, this + // should be safe with audio only as removeVideo will call back even if there's no + // video buffer. + // + // In theory we can check to see if there's video before calling the remove, but in + // the event that we're switching between renditions and from video to audio only + // (when we add support for that), we may need to clear the video contents despite + // what the new media will contain. + + if (force || this.loaderType_ === 'main') { + this.gopBuffer_ = removeGopBuffer(this.gopBuffer_, start, end, this.timeMapping_); + removesRemaining++; + this.sourceUpdater_.removeVideo(start, end, removeFinished); + } // remove any captions and ID3 tags + + for (const track in this.inbandTextTracks_) { + removeCuesFromTrack(start, end, this.inbandTextTracks_[track]); + } + removeCuesFromTrack(start, end, this.segmentMetadataTrack_); // finished this function's removes + + removeFinished(); + } + /** + * (re-)schedule monitorBufferTick_ to run as soon as possible + * + * @private + */ + + monitorBuffer_() { + if (this.checkBufferTimeout_) { + window$1.clearTimeout(this.checkBufferTimeout_); + } + this.checkBufferTimeout_ = window$1.setTimeout(this.monitorBufferTick_.bind(this), 1); + } + /** + * As long as the SegmentLoader is in the READY state, periodically + * invoke fillBuffer_(). + * + * @private + */ + + monitorBufferTick_() { + if (this.state === 'READY') { + this.fillBuffer_(); + } + if (this.checkBufferTimeout_) { + window$1.clearTimeout(this.checkBufferTimeout_); + } + this.checkBufferTimeout_ = window$1.setTimeout(this.monitorBufferTick_.bind(this), CHECK_BUFFER_DELAY); + } + /** + * fill the buffer with segements unless the sourceBuffers are + * currently updating + * + * Note: this function should only ever be called by monitorBuffer_ + * and never directly + * + * @private + */ + + fillBuffer_() { + // TODO since the source buffer maintains a queue, and we shouldn't call this function + // except when we're ready for the next segment, this check can most likely be removed + if (this.sourceUpdater_.updating()) { + return; + } // see if we need to begin loading immediately + + const segmentInfo = this.chooseNextRequest_(); + if (!segmentInfo) { + return; + } + if (typeof segmentInfo.timestampOffset === 'number') { + this.isPendingTimestampOffset_ = false; + this.timelineChangeController_.pendingTimelineChange({ + type: this.loaderType_, + from: this.currentTimeline_, + to: segmentInfo.timeline + }); + } + this.loadSegment_(segmentInfo); + } + /** + * Determines if we should call endOfStream on the media source based + * on the state of the buffer or if appened segment was the final + * segment in the playlist. + * + * @param {number} [mediaIndex] the media index of segment we last appended + * @param {Object} [playlist] a media playlist object + * @return {boolean} do we need to call endOfStream on the MediaSource + */ + + isEndOfStream_(mediaIndex = this.mediaIndex, playlist = this.playlist_, partIndex = this.partIndex) { + if (!playlist || !this.mediaSource_) { + return false; + } + const segment = typeof mediaIndex === 'number' && playlist.segments[mediaIndex]; // mediaIndex is zero based but length is 1 based + + const appendedLastSegment = mediaIndex + 1 === playlist.segments.length; // true if there are no parts, or this is the last part. + + const appendedLastPart = !segment || !segment.parts || partIndex + 1 === segment.parts.length; // if we've buffered to the end of the video, we need to call endOfStream + // so that MediaSources can trigger the `ended` event when it runs out of + // buffered data instead of waiting for me + + return playlist.endList && this.mediaSource_.readyState === 'open' && appendedLastSegment && appendedLastPart; + } + /** + * Determines what request should be made given current segment loader state. + * + * @return {Object} a request object that describes the segment/part to load + */ + + chooseNextRequest_() { + const buffered = this.buffered_(); + const bufferedEnd = lastBufferedEnd(buffered) || 0; + const bufferedTime = timeAheadOf(buffered, this.currentTime_()); + const preloaded = !this.hasPlayed_() && bufferedTime >= 1; + const haveEnoughBuffer = bufferedTime >= this.goalBufferLength_(); + const segments = this.playlist_.segments; // return no segment if: + // 1. we don't have segments + // 2. The video has not yet played and we already downloaded a segment + // 3. we already have enough buffered time + + if (!segments.length || preloaded || haveEnoughBuffer) { + return null; + } + this.syncPoint_ = this.syncPoint_ || this.syncController_.getSyncPoint(this.playlist_, this.duration_(), this.currentTimeline_, this.currentTime_(), this.loaderType_); + const next = { + partIndex: null, + mediaIndex: null, + startOfSegment: null, + playlist: this.playlist_, + isSyncRequest: Boolean(!this.syncPoint_) + }; + if (next.isSyncRequest) { + next.mediaIndex = getSyncSegmentCandidate(this.currentTimeline_, segments, bufferedEnd); + this.logger_(`choose next request. Can not find sync point. Fallback to media Index: ${next.mediaIndex}`); + } else if (this.mediaIndex !== null) { + const segment = segments[this.mediaIndex]; + const partIndex = typeof this.partIndex === 'number' ? this.partIndex : -1; + next.startOfSegment = segment.end ? segment.end : bufferedEnd; + if (segment.parts && segment.parts[partIndex + 1]) { + next.mediaIndex = this.mediaIndex; + next.partIndex = partIndex + 1; + } else { + next.mediaIndex = this.mediaIndex + 1; + } + } else { + let segmentIndex; + let partIndex; + let startTime; + const targetTime = this.fetchAtBuffer_ ? bufferedEnd : this.currentTime_(); + if (this.mediaSequenceSync_) { + this.logger_(`chooseNextRequest_ request after Quality Switch: +For TargetTime: ${targetTime}. +CurrentTime: ${this.currentTime_()} +BufferedEnd: ${bufferedEnd} +Fetch At Buffer: ${this.fetchAtBuffer_} +`, this.mediaSequenceSync_.diagnostics); + } + if (this.mediaSequenceSync_ && this.mediaSequenceSync_.isReliable) { + const syncInfo = this.getSyncInfoFromMediaSequenceSync_(targetTime); + if (!syncInfo) { + this.logger_('chooseNextRequest_ - no sync info found using media sequence sync'); // no match + + return null; + } + this.logger_(`chooseNextRequest_ mediaSequence syncInfo (${syncInfo.start} --> ${syncInfo.end})`); + segmentIndex = syncInfo.segmentIndex; + partIndex = syncInfo.partIndex; + startTime = syncInfo.start; + } else { + this.logger_('chooseNextRequest_ - fallback to a regular segment selection algorithm, based on a syncPoint.'); // fallback + + const mediaInfoForTime = Playlist.getMediaInfoForTime({ + exactManifestTimings: this.exactManifestTimings, + playlist: this.playlist_, + currentTime: targetTime, + startingPartIndex: this.syncPoint_.partIndex, + startingSegmentIndex: this.syncPoint_.segmentIndex, + startTime: this.syncPoint_.time + }); + segmentIndex = mediaInfoForTime.segmentIndex; + partIndex = mediaInfoForTime.partIndex; + startTime = mediaInfoForTime.startTime; + } + next.getMediaInfoForTime = this.fetchAtBuffer_ ? `bufferedEnd ${targetTime}` : `currentTime ${targetTime}`; + next.mediaIndex = segmentIndex; + next.startOfSegment = startTime; + next.partIndex = partIndex; + this.logger_(`choose next request. Playlist switched and we have a sync point. Media Index: ${next.mediaIndex} `); + } + const nextSegment = segments[next.mediaIndex]; + let nextPart = nextSegment && typeof next.partIndex === 'number' && nextSegment.parts && nextSegment.parts[next.partIndex]; // if the next segment index is invalid or + // the next partIndex is invalid do not choose a next segment. + + if (!nextSegment || typeof next.partIndex === 'number' && !nextPart) { + return null; + } // if the next segment has parts, and we don't have a partIndex. + // Set partIndex to 0 + + if (typeof next.partIndex !== 'number' && nextSegment.parts) { + next.partIndex = 0; + nextPart = nextSegment.parts[0]; + } // independentSegments applies to every segment in a playlist. If independentSegments appears in a main playlist, + // it applies to each segment in each media playlist. + // https://datatracker.ietf.org/doc/html/draft-pantos-http-live-streaming-23#section-4.3.5.1 + + const hasIndependentSegments = this.vhs_.playlists && this.vhs_.playlists.main && this.vhs_.playlists.main.independentSegments || this.playlist_.independentSegments; // if we have no buffered data then we need to make sure + // that the next part we append is "independent" if possible. + // So we check if the previous part is independent, and request + // it if it is. + + if (!bufferedTime && nextPart && !hasIndependentSegments && !nextPart.independent) { + if (next.partIndex === 0) { + const lastSegment = segments[next.mediaIndex - 1]; + const lastSegmentLastPart = lastSegment.parts && lastSegment.parts.length && lastSegment.parts[lastSegment.parts.length - 1]; + if (lastSegmentLastPart && lastSegmentLastPart.independent) { + next.mediaIndex -= 1; + next.partIndex = lastSegment.parts.length - 1; + next.independent = 'previous segment'; + } + } else if (nextSegment.parts[next.partIndex - 1].independent) { + next.partIndex -= 1; + next.independent = 'previous part'; + } + } + const ended = this.mediaSource_ && this.mediaSource_.readyState === 'ended'; // do not choose a next segment if all of the following: + // 1. this is the last segment in the playlist + // 2. end of stream has been called on the media source already + // 3. the player is not seeking + + if (next.mediaIndex >= segments.length - 1 && ended && !this.seeking_()) { + return null; + } + if (this.shouldForceTimestampOffsetAfterResync_) { + this.shouldForceTimestampOffsetAfterResync_ = false; + next.forceTimestampOffset = true; + this.logger_('choose next request. Force timestamp offset after loader resync'); + } + return this.generateSegmentInfo_(next); + } + getSyncInfoFromMediaSequenceSync_(targetTime) { + if (!this.mediaSequenceSync_) { + return null; + } // we should pull the target time to the least available time if we drop out of sync for any reason + + const finalTargetTime = Math.max(targetTime, this.mediaSequenceSync_.start); + if (targetTime !== finalTargetTime) { + this.logger_(`getSyncInfoFromMediaSequenceSync_. Pulled target time from ${targetTime} to ${finalTargetTime}`); + } + const mediaSequenceSyncInfo = this.mediaSequenceSync_.getSyncInfoForTime(finalTargetTime); + if (!mediaSequenceSyncInfo) { + // no match at all + return null; + } + if (!mediaSequenceSyncInfo.isAppended) { + // has a perfect match + return mediaSequenceSyncInfo; + } // has match, but segment was already appended. + // attempt to auto-advance to the nearest next segment: + + const nextMediaSequenceSyncInfo = this.mediaSequenceSync_.getSyncInfoForTime(mediaSequenceSyncInfo.end); + if (!nextMediaSequenceSyncInfo) { + // no match at all + return null; + } + if (nextMediaSequenceSyncInfo.isAppended) { + this.logger_('getSyncInfoFromMediaSequenceSync_: We encounter unexpected scenario where next media sequence sync info is also appended!'); + } // got match with the nearest next segment + + return nextMediaSequenceSyncInfo; + } + generateSegmentInfo_(options) { + const { + independent, + playlist, + mediaIndex, + startOfSegment, + isSyncRequest, + partIndex, + forceTimestampOffset, + getMediaInfoForTime + } = options; + const segment = playlist.segments[mediaIndex]; + const part = typeof partIndex === 'number' && segment.parts[partIndex]; + const segmentInfo = { + requestId: 'segment-loader-' + Math.random(), + // resolve the segment URL relative to the playlist + uri: part && part.resolvedUri || segment.resolvedUri, + // the segment's mediaIndex at the time it was requested + mediaIndex, + partIndex: part ? partIndex : null, + // whether or not to update the SegmentLoader's state with this + // segment's mediaIndex + isSyncRequest, + startOfSegment, + // the segment's playlist + playlist, + // unencrypted bytes of the segment + bytes: null, + // when a key is defined for this segment, the encrypted bytes + encryptedBytes: null, + // The target timestampOffset for this segment when we append it + // to the source buffer + timestampOffset: null, + // The timeline that the segment is in + timeline: segment.timeline, + // The expected duration of the segment in seconds + duration: part && part.duration || segment.duration, + // retain the segment in case the playlist updates while doing an async process + segment, + part, + byteLength: 0, + transmuxer: this.transmuxer_, + // type of getMediaInfoForTime that was used to get this segment + getMediaInfoForTime, + independent + }; + const overrideCheck = typeof forceTimestampOffset !== 'undefined' ? forceTimestampOffset : this.isPendingTimestampOffset_; + segmentInfo.timestampOffset = this.timestampOffsetForSegment_({ + segmentTimeline: segment.timeline, + currentTimeline: this.currentTimeline_, + startOfSegment, + buffered: this.buffered_(), + overrideCheck + }); + const audioBufferedEnd = lastBufferedEnd(this.sourceUpdater_.audioBuffered()); + if (typeof audioBufferedEnd === 'number') { + // since the transmuxer is using the actual timing values, but the buffer is + // adjusted by the timestamp offset, we must adjust the value here + segmentInfo.audioAppendStart = audioBufferedEnd - this.sourceUpdater_.audioTimestampOffset(); + } + if (this.sourceUpdater_.videoBuffered().length) { + segmentInfo.gopsToAlignWith = gopsSafeToAlignWith(this.gopBuffer_, + // since the transmuxer is using the actual timing values, but the time is + // adjusted by the timestmap offset, we must adjust the value here + this.currentTime_() - this.sourceUpdater_.videoTimestampOffset(), this.timeMapping_); + } + return segmentInfo; + } // get the timestampoffset for a segment, + // added so that vtt segment loader can override and prevent + // adding timestamp offsets. + + timestampOffsetForSegment_(options) { + return timestampOffsetForSegment(options); + } + /** + * Determines if the network has enough bandwidth to complete the current segment + * request in a timely manner. If not, the request will be aborted early and bandwidth + * updated to trigger a playlist switch. + * + * @param {Object} stats + * Object containing stats about the request timing and size + * @private + */ + + earlyAbortWhenNeeded_(stats) { + if (this.vhs_.tech_.paused() || + // Don't abort if the current playlist is on the lowestEnabledRendition + // TODO: Replace using timeout with a boolean indicating whether this playlist is + // the lowestEnabledRendition. + !this.xhrOptions_.timeout || + // Don't abort if we have no bandwidth information to estimate segment sizes + !this.playlist_.attributes.BANDWIDTH) { + return; + } // Wait at least 1 second since the first byte of data has been received before + // using the calculated bandwidth from the progress event to allow the bitrate + // to stabilize + + if (Date.now() - (stats.firstBytesReceivedAt || Date.now()) < 1000) { + return; + } + const currentTime = this.currentTime_(); + const measuredBandwidth = stats.bandwidth; + const segmentDuration = this.pendingSegment_.duration; + const requestTimeRemaining = Playlist.estimateSegmentRequestTime(segmentDuration, measuredBandwidth, this.playlist_, stats.bytesReceived); // Subtract 1 from the timeUntilRebuffer so we still consider an early abort + // if we are only left with less than 1 second when the request completes. + // A negative timeUntilRebuffering indicates we are already rebuffering + + const timeUntilRebuffer$1 = timeUntilRebuffer(this.buffered_(), currentTime, this.vhs_.tech_.playbackRate()) - 1; // Only consider aborting early if the estimated time to finish the download + // is larger than the estimated time until the player runs out of forward buffer + + if (requestTimeRemaining <= timeUntilRebuffer$1) { + return; + } + const switchCandidate = minRebufferMaxBandwidthSelector({ + main: this.vhs_.playlists.main, + currentTime, + bandwidth: measuredBandwidth, + duration: this.duration_(), + segmentDuration, + timeUntilRebuffer: timeUntilRebuffer$1, + currentTimeline: this.currentTimeline_, + syncController: this.syncController_ + }); + if (!switchCandidate) { + return; + } + const rebufferingImpact = requestTimeRemaining - timeUntilRebuffer$1; + const timeSavedBySwitching = rebufferingImpact - switchCandidate.rebufferingImpact; + let minimumTimeSaving = 0.5; // If we are already rebuffering, increase the amount of variance we add to the + // potential round trip time of the new request so that we are not too aggressive + // with switching to a playlist that might save us a fraction of a second. + + if (timeUntilRebuffer$1 <= TIME_FUDGE_FACTOR) { + minimumTimeSaving = 1; + } + if (!switchCandidate.playlist || switchCandidate.playlist.uri === this.playlist_.uri || timeSavedBySwitching < minimumTimeSaving) { + return; + } // set the bandwidth to that of the desired playlist being sure to scale by + // BANDWIDTH_VARIANCE and add one so the playlist selector does not exclude it + // don't trigger a bandwidthupdate as the bandwidth is artifial + + this.bandwidth = switchCandidate.playlist.attributes.BANDWIDTH * Config.BANDWIDTH_VARIANCE + 1; + this.trigger('earlyabort'); + } + handleAbort_(segmentInfo) { + this.logger_(`Aborting ${segmentInfoString(segmentInfo)}`); + this.mediaRequestsAborted += 1; + } + /** + * XHR `progress` event handler + * + * @param {Event} + * The XHR `progress` event + * @param {Object} simpleSegment + * A simplified segment object copy + * @private + */ + + handleProgress_(event, simpleSegment) { + this.earlyAbortWhenNeeded_(simpleSegment.stats); + if (this.checkForAbort_(simpleSegment.requestId)) { + return; + } + this.trigger('progress'); + } + handleTrackInfo_(simpleSegment, trackInfo) { + this.earlyAbortWhenNeeded_(simpleSegment.stats); + if (this.checkForAbort_(simpleSegment.requestId)) { + return; + } + if (this.checkForIllegalMediaSwitch(trackInfo)) { + return; + } + trackInfo = trackInfo || {}; // When we have track info, determine what media types this loader is dealing with. + // Guard against cases where we're not getting track info at all until we are + // certain that all streams will provide it. + + if (!shallowEqual(this.currentMediaInfo_, trackInfo)) { + this.appendInitSegment_ = { + audio: true, + video: true + }; + this.startingMediaInfo_ = trackInfo; + this.currentMediaInfo_ = trackInfo; + this.logger_('trackinfo update', trackInfo); + this.trigger('trackinfo'); + } // trackinfo may cause an abort if the trackinfo + // causes a codec change to an unsupported codec. + + if (this.checkForAbort_(simpleSegment.requestId)) { + return; + } // set trackinfo on the pending segment so that + // it can append. + + this.pendingSegment_.trackInfo = trackInfo; // check if any calls were waiting on the track info + + if (this.hasEnoughInfoToAppend_()) { + this.processCallQueue_(); + } + } + handleTimingInfo_(simpleSegment, mediaType, timeType, time) { + this.earlyAbortWhenNeeded_(simpleSegment.stats); + if (this.checkForAbort_(simpleSegment.requestId)) { + return; + } + const segmentInfo = this.pendingSegment_; + const timingInfoProperty = timingInfoPropertyForMedia(mediaType); + segmentInfo[timingInfoProperty] = segmentInfo[timingInfoProperty] || {}; + segmentInfo[timingInfoProperty][timeType] = time; + this.logger_(`timinginfo: ${mediaType} - ${timeType} - ${time}`); // check if any calls were waiting on the timing info + + if (this.hasEnoughInfoToAppend_()) { + this.processCallQueue_(); + } + } + handleCaptions_(simpleSegment, captionData) { + this.earlyAbortWhenNeeded_(simpleSegment.stats); + if (this.checkForAbort_(simpleSegment.requestId)) { + return; + } // This could only happen with fmp4 segments, but + // should still not happen in general + + if (captionData.length === 0) { + this.logger_('SegmentLoader received no captions from a caption event'); + return; + } + const segmentInfo = this.pendingSegment_; // Wait until we have some video data so that caption timing + // can be adjusted by the timestamp offset + + if (!segmentInfo.hasAppendedData_) { + this.metadataQueue_.caption.push(this.handleCaptions_.bind(this, simpleSegment, captionData)); + return; + } + const timestampOffset = this.sourceUpdater_.videoTimestampOffset() === null ? this.sourceUpdater_.audioTimestampOffset() : this.sourceUpdater_.videoTimestampOffset(); + const captionTracks = {}; // get total start/end and captions for each track/stream + + captionData.forEach(caption => { + // caption.stream is actually a track name... + // set to the existing values in tracks or default values + captionTracks[caption.stream] = captionTracks[caption.stream] || { + // Infinity, as any other value will be less than this + startTime: Infinity, + captions: [], + // 0 as an other value will be more than this + endTime: 0 + }; + const captionTrack = captionTracks[caption.stream]; + captionTrack.startTime = Math.min(captionTrack.startTime, caption.startTime + timestampOffset); + captionTrack.endTime = Math.max(captionTrack.endTime, caption.endTime + timestampOffset); + captionTrack.captions.push(caption); + }); + Object.keys(captionTracks).forEach(trackName => { + const { + startTime, + endTime, + captions + } = captionTracks[trackName]; + const inbandTextTracks = this.inbandTextTracks_; + this.logger_(`adding cues from ${startTime} -> ${endTime} for ${trackName}`); + createCaptionsTrackIfNotExists(inbandTextTracks, this.vhs_.tech_, trackName); // clear out any cues that start and end at the same time period for the same track. + // We do this because a rendition change that also changes the timescale for captions + // will result in captions being re-parsed for certain segments. If we add them again + // without clearing we will have two of the same captions visible. + + removeCuesFromTrack(startTime, endTime, inbandTextTracks[trackName]); + addCaptionData({ + captionArray: captions, + inbandTextTracks, + timestampOffset + }); + }); // Reset stored captions since we added parsed + // captions to a text track at this point + + if (this.transmuxer_) { + this.transmuxer_.postMessage({ + action: 'clearParsedMp4Captions' + }); + } + } + handleId3_(simpleSegment, id3Frames, dispatchType) { + this.earlyAbortWhenNeeded_(simpleSegment.stats); + if (this.checkForAbort_(simpleSegment.requestId)) { + return; + } + const segmentInfo = this.pendingSegment_; // we need to have appended data in order for the timestamp offset to be set + + if (!segmentInfo.hasAppendedData_) { + this.metadataQueue_.id3.push(this.handleId3_.bind(this, simpleSegment, id3Frames, dispatchType)); + return; + } + this.addMetadataToTextTrack(dispatchType, id3Frames, this.duration_()); + } + processMetadataQueue_() { + this.metadataQueue_.id3.forEach(fn => fn()); + this.metadataQueue_.caption.forEach(fn => fn()); + this.metadataQueue_.id3 = []; + this.metadataQueue_.caption = []; + } + processCallQueue_() { + const callQueue = this.callQueue_; // Clear out the queue before the queued functions are run, since some of the + // functions may check the length of the load queue and default to pushing themselves + // back onto the queue. + + this.callQueue_ = []; + callQueue.forEach(fun => fun()); + } + processLoadQueue_() { + const loadQueue = this.loadQueue_; // Clear out the queue before the queued functions are run, since some of the + // functions may check the length of the load queue and default to pushing themselves + // back onto the queue. + + this.loadQueue_ = []; + loadQueue.forEach(fun => fun()); + } + /** + * Determines whether the loader has enough info to load the next segment. + * + * @return {boolean} + * Whether or not the loader has enough info to load the next segment + */ + + hasEnoughInfoToLoad_() { + // Since primary timing goes by video, only the audio loader potentially needs to wait + // to load. + if (this.loaderType_ !== 'audio') { + return true; + } + const segmentInfo = this.pendingSegment_; // A fill buffer must have already run to establish a pending segment before there's + // enough info to load. + + if (!segmentInfo) { + return false; + } // The first segment can and should be loaded immediately so that source buffers are + // created together (before appending). Source buffer creation uses the presence of + // audio and video data to determine whether to create audio/video source buffers, and + // uses processed (transmuxed or parsed) media to determine the types required. + + if (!this.getCurrentMediaInfo_()) { + return true; + } + if ( + // Technically, instead of waiting to load a segment on timeline changes, a segment + // can be requested and downloaded and only wait before it is transmuxed or parsed. + // But in practice, there are a few reasons why it is better to wait until a loader + // is ready to append that segment before requesting and downloading: + // + // 1. Because audio and main loaders cross discontinuities together, if this loader + // is waiting for the other to catch up, then instead of requesting another + // segment and using up more bandwidth, by not yet loading, more bandwidth is + // allotted to the loader currently behind. + // 2. media-segment-request doesn't have to have logic to consider whether a segment + // is ready to be processed or not, isolating the queueing behavior to the loader. + // 3. The audio loader bases some of its segment properties on timing information + // provided by the main loader, meaning that, if the logic for waiting on + // processing was in media-segment-request, then it would also need to know how + // to re-generate the segment information after the main loader caught up. + shouldWaitForTimelineChange({ + timelineChangeController: this.timelineChangeController_, + currentTimeline: this.currentTimeline_, + segmentTimeline: segmentInfo.timeline, + loaderType: this.loaderType_, + audioDisabled: this.audioDisabled_ + })) { + return false; + } + return true; + } + getCurrentMediaInfo_(segmentInfo = this.pendingSegment_) { + return segmentInfo && segmentInfo.trackInfo || this.currentMediaInfo_; + } + getMediaInfo_(segmentInfo = this.pendingSegment_) { + return this.getCurrentMediaInfo_(segmentInfo) || this.startingMediaInfo_; + } + getPendingSegmentPlaylist() { + return this.pendingSegment_ ? this.pendingSegment_.playlist : null; + } + hasEnoughInfoToAppend_() { + if (!this.sourceUpdater_.ready()) { + return false; + } // If content needs to be removed or the loader is waiting on an append reattempt, + // then no additional content should be appended until the prior append is resolved. + + if (this.waitingOnRemove_ || this.quotaExceededErrorRetryTimeout_) { + return false; + } + const segmentInfo = this.pendingSegment_; + const trackInfo = this.getCurrentMediaInfo_(); // no segment to append any data for or + // we do not have information on this specific + // segment yet + + if (!segmentInfo || !trackInfo) { + return false; + } + const { + hasAudio, + hasVideo, + isMuxed + } = trackInfo; + if (hasVideo && !segmentInfo.videoTimingInfo) { + return false; + } // muxed content only relies on video timing information for now. + + if (hasAudio && !this.audioDisabled_ && !isMuxed && !segmentInfo.audioTimingInfo) { + return false; + } + if (shouldWaitForTimelineChange({ + timelineChangeController: this.timelineChangeController_, + currentTimeline: this.currentTimeline_, + segmentTimeline: segmentInfo.timeline, + loaderType: this.loaderType_, + audioDisabled: this.audioDisabled_ + })) { + return false; + } + return true; + } + handleData_(simpleSegment, result) { + this.earlyAbortWhenNeeded_(simpleSegment.stats); + if (this.checkForAbort_(simpleSegment.requestId)) { + return; + } // If there's anything in the call queue, then this data came later and should be + // executed after the calls currently queued. + + if (this.callQueue_.length || !this.hasEnoughInfoToAppend_()) { + this.callQueue_.push(this.handleData_.bind(this, simpleSegment, result)); + return; + } + const segmentInfo = this.pendingSegment_; // update the time mapping so we can translate from display time to media time + + this.setTimeMapping_(segmentInfo.timeline); // for tracking overall stats + + this.updateMediaSecondsLoaded_(segmentInfo.part || segmentInfo.segment); // Note that the state isn't changed from loading to appending. This is because abort + // logic may change behavior depending on the state, and changing state too early may + // inflate our estimates of bandwidth. In the future this should be re-examined to + // note more granular states. + // don't process and append data if the mediaSource is closed + + if (this.mediaSource_.readyState === 'closed') { + return; + } // if this request included an initialization segment, save that data + // to the initSegment cache + + if (simpleSegment.map) { + simpleSegment.map = this.initSegmentForMap(simpleSegment.map, true); // move over init segment properties to media request + + segmentInfo.segment.map = simpleSegment.map; + } // if this request included a segment key, save that data in the cache + + if (simpleSegment.key) { + this.segmentKey(simpleSegment.key, true); + } + segmentInfo.isFmp4 = simpleSegment.isFmp4; + segmentInfo.timingInfo = segmentInfo.timingInfo || {}; + if (segmentInfo.isFmp4) { + this.trigger('fmp4'); + segmentInfo.timingInfo.start = segmentInfo[timingInfoPropertyForMedia(result.type)].start; + } else { + const trackInfo = this.getCurrentMediaInfo_(); + const useVideoTimingInfo = this.loaderType_ === 'main' && trackInfo && trackInfo.hasVideo; + let firstVideoFrameTimeForData; + if (useVideoTimingInfo) { + firstVideoFrameTimeForData = segmentInfo.videoTimingInfo.start; + } // Segment loader knows more about segment timing than the transmuxer (in certain + // aspects), so make any changes required for a more accurate start time. + // Don't set the end time yet, as the segment may not be finished processing. + + segmentInfo.timingInfo.start = this.trueSegmentStart_({ + currentStart: segmentInfo.timingInfo.start, + playlist: segmentInfo.playlist, + mediaIndex: segmentInfo.mediaIndex, + currentVideoTimestampOffset: this.sourceUpdater_.videoTimestampOffset(), + useVideoTimingInfo, + firstVideoFrameTimeForData, + videoTimingInfo: segmentInfo.videoTimingInfo, + audioTimingInfo: segmentInfo.audioTimingInfo + }); + } // Init segments for audio and video only need to be appended in certain cases. Now + // that data is about to be appended, we can check the final cases to determine + // whether we should append an init segment. + + this.updateAppendInitSegmentStatus(segmentInfo, result.type); // Timestamp offset should be updated once we get new data and have its timing info, + // as we use the start of the segment to offset the best guess (playlist provided) + // timestamp offset. + + this.updateSourceBufferTimestampOffset_(segmentInfo); // if this is a sync request we need to determine whether it should + // be appended or not. + + if (segmentInfo.isSyncRequest) { + // first save/update our timing info for this segment. + // this is what allows us to choose an accurate segment + // and the main reason we make a sync request. + this.updateTimingInfoEnd_(segmentInfo); + this.syncController_.saveSegmentTimingInfo({ + segmentInfo, + shouldSaveTimelineMapping: this.loaderType_ === 'main' + }); + const next = this.chooseNextRequest_(); // If the sync request isn't the segment that would be requested next + // after taking into account its timing info, do not append it. + + if (next.mediaIndex !== segmentInfo.mediaIndex || next.partIndex !== segmentInfo.partIndex) { + this.logger_('sync segment was incorrect, not appending'); + return; + } // otherwise append it like any other segment as our guess was correct. + + this.logger_('sync segment was correct, appending'); + } // Save some state so that in the future anything waiting on first append (and/or + // timestamp offset(s)) can process immediately. While the extra state isn't optimal, + // we need some notion of whether the timestamp offset or other relevant information + // has had a chance to be set. + + segmentInfo.hasAppendedData_ = true; // Now that the timestamp offset should be set, we can append any waiting ID3 tags. + + this.processMetadataQueue_(); + this.appendData_(segmentInfo, result); + } + updateAppendInitSegmentStatus(segmentInfo, type) { + // alt audio doesn't manage timestamp offset + if (this.loaderType_ === 'main' && typeof segmentInfo.timestampOffset === 'number' && + // in the case that we're handling partial data, we don't want to append an init + // segment for each chunk + !segmentInfo.changedTimestampOffset) { + // if the timestamp offset changed, the timeline may have changed, so we have to re- + // append init segments + this.appendInitSegment_ = { + audio: true, + video: true + }; + } + if (this.playlistOfLastInitSegment_[type] !== segmentInfo.playlist) { + // make sure we append init segment on playlist changes, in case the media config + // changed + this.appendInitSegment_[type] = true; + } + } + getInitSegmentAndUpdateState_({ + type, + initSegment, + map, + playlist + }) { + // "The EXT-X-MAP tag specifies how to obtain the Media Initialization Section + // (Section 3) required to parse the applicable Media Segments. It applies to every + // Media Segment that appears after it in the Playlist until the next EXT-X-MAP tag + // or until the end of the playlist." + // https://tools.ietf.org/html/draft-pantos-http-live-streaming-23#section-4.3.2.5 + if (map) { + const id = initSegmentId(map); + if (this.activeInitSegmentId_ === id) { + // don't need to re-append the init segment if the ID matches + return null; + } // a map-specified init segment takes priority over any transmuxed (or otherwise + // obtained) init segment + // + // this also caches the init segment for later use + + initSegment = this.initSegmentForMap(map, true).bytes; + this.activeInitSegmentId_ = id; + } // We used to always prepend init segments for video, however, that shouldn't be + // necessary. Instead, we should only append on changes, similar to what we've always + // done for audio. This is more important (though may not be that important) for + // frame-by-frame appending for LHLS, simply because of the increased quantity of + // appends. + + if (initSegment && this.appendInitSegment_[type]) { + // Make sure we track the playlist that we last used for the init segment, so that + // we can re-append the init segment in the event that we get data from a new + // playlist. Discontinuities and track changes are handled in other sections. + this.playlistOfLastInitSegment_[type] = playlist; // Disable future init segment appends for this type. Until a change is necessary. + + this.appendInitSegment_[type] = false; // we need to clear out the fmp4 active init segment id, since + // we are appending the muxer init segment + + this.activeInitSegmentId_ = null; + return initSegment; + } + return null; + } + handleQuotaExceededError_({ + segmentInfo, + type, + bytes + }, error) { + const audioBuffered = this.sourceUpdater_.audioBuffered(); + const videoBuffered = this.sourceUpdater_.videoBuffered(); // For now we're ignoring any notion of gaps in the buffer, but they, in theory, + // should be cleared out during the buffer removals. However, log in case it helps + // debug. + + if (audioBuffered.length > 1) { + this.logger_('On QUOTA_EXCEEDED_ERR, found gaps in the audio buffer: ' + timeRangesToArray(audioBuffered).join(', ')); + } + if (videoBuffered.length > 1) { + this.logger_('On QUOTA_EXCEEDED_ERR, found gaps in the video buffer: ' + timeRangesToArray(videoBuffered).join(', ')); + } + const audioBufferStart = audioBuffered.length ? audioBuffered.start(0) : 0; + const audioBufferEnd = audioBuffered.length ? audioBuffered.end(audioBuffered.length - 1) : 0; + const videoBufferStart = videoBuffered.length ? videoBuffered.start(0) : 0; + const videoBufferEnd = videoBuffered.length ? videoBuffered.end(videoBuffered.length - 1) : 0; + if (audioBufferEnd - audioBufferStart <= MIN_BACK_BUFFER && videoBufferEnd - videoBufferStart <= MIN_BACK_BUFFER) { + // Can't remove enough buffer to make room for new segment (or the browser doesn't + // allow for appends of segments this size). In the future, it may be possible to + // split up the segment and append in pieces, but for now, error out this playlist + // in an attempt to switch to a more manageable rendition. + this.logger_('On QUOTA_EXCEEDED_ERR, single segment too large to append to ' + 'buffer, triggering an error. ' + `Appended byte length: ${bytes.byteLength}, ` + `audio buffer: ${timeRangesToArray(audioBuffered).join(', ')}, ` + `video buffer: ${timeRangesToArray(videoBuffered).join(', ')}, `); + this.error({ + message: 'Quota exceeded error with append of a single segment of content', + excludeUntil: Infinity, + metadata: { + errorType: videojs.Error.SegmentExceedsSourceBufferQuota + } + }); + this.trigger('error'); + return; + } // To try to resolve the quota exceeded error, clear back buffer and retry. This means + // that the segment-loader should block on future events until this one is handled, so + // that it doesn't keep moving onto further segments. Adding the call to the call + // queue will prevent further appends until waitingOnRemove_ and + // quotaExceededErrorRetryTimeout_ are cleared. + // + // Note that this will only block the current loader. In the case of demuxed content, + // the other load may keep filling as fast as possible. In practice, this should be + // OK, as it is a rare case when either audio has a high enough bitrate to fill up a + // source buffer, or video fills without enough room for audio to append (and without + // the availability of clearing out seconds of back buffer to make room for audio). + // But it might still be good to handle this case in the future as a TODO. + + this.waitingOnRemove_ = true; + this.callQueue_.push(this.appendToSourceBuffer_.bind(this, { + segmentInfo, + type, + bytes + })); + const currentTime = this.currentTime_(); // Try to remove as much audio and video as possible to make room for new content + // before retrying. + + const timeToRemoveUntil = currentTime - MIN_BACK_BUFFER; + this.logger_(`On QUOTA_EXCEEDED_ERR, removing audio/video from 0 to ${timeToRemoveUntil}`); + this.remove(0, timeToRemoveUntil, () => { + this.logger_(`On QUOTA_EXCEEDED_ERR, retrying append in ${MIN_BACK_BUFFER}s`); + this.waitingOnRemove_ = false; // wait the length of time alotted in the back buffer to prevent wasted + // attempts (since we can't clear less than the minimum) + + this.quotaExceededErrorRetryTimeout_ = window$1.setTimeout(() => { + this.logger_('On QUOTA_EXCEEDED_ERR, re-processing call queue'); + this.quotaExceededErrorRetryTimeout_ = null; + this.processCallQueue_(); + }, MIN_BACK_BUFFER * 1000); + }, true); + } + handleAppendError_({ + segmentInfo, + type, + bytes + }, error) { + // if there's no error, nothing to do + if (!error) { + return; + } + if (error.code === QUOTA_EXCEEDED_ERR) { + this.handleQuotaExceededError_({ + segmentInfo, + type, + bytes + }); // A quota exceeded error should be recoverable with a future re-append, so no need + // to trigger an append error. + + return; + } + this.logger_('Received non QUOTA_EXCEEDED_ERR on append', error); // If an append errors, we often can't recover. + // (see https://w3c.github.io/media-source/#sourcebuffer-append-error). + // + // Trigger a special error so that it can be handled separately from normal, + // recoverable errors. + + this.error({ + message: `${type} append of ${bytes.length}b failed for segment ` + `#${segmentInfo.mediaIndex} in playlist ${segmentInfo.playlist.id}`, + metadata: { + errorType: videojs.Error.SegmentAppendError + } + }); + this.trigger('appenderror'); + } + appendToSourceBuffer_({ + segmentInfo, + type, + initSegment, + data, + bytes + }) { + // If this is a re-append, bytes were already created and don't need to be recreated + if (!bytes) { + const segments = [data]; + let byteLength = data.byteLength; + if (initSegment) { + // if the media initialization segment is changing, append it before the content + // segment + segments.unshift(initSegment); + byteLength += initSegment.byteLength; + } // Technically we should be OK appending the init segment separately, however, we + // haven't yet tested that, and prepending is how we have always done things. + + bytes = concatSegments({ + bytes: byteLength, + segments + }); + } + this.sourceUpdater_.appendBuffer({ + segmentInfo, + type, + bytes + }, this.handleAppendError_.bind(this, { + segmentInfo, + type, + bytes + })); + } + handleSegmentTimingInfo_(type, requestId, segmentTimingInfo) { + if (!this.pendingSegment_ || requestId !== this.pendingSegment_.requestId) { + return; + } + const segment = this.pendingSegment_.segment; + const timingInfoProperty = `${type}TimingInfo`; + if (!segment[timingInfoProperty]) { + segment[timingInfoProperty] = {}; + } + segment[timingInfoProperty].transmuxerPrependedSeconds = segmentTimingInfo.prependedContentDuration || 0; + segment[timingInfoProperty].transmuxedPresentationStart = segmentTimingInfo.start.presentation; + segment[timingInfoProperty].transmuxedDecodeStart = segmentTimingInfo.start.decode; + segment[timingInfoProperty].transmuxedPresentationEnd = segmentTimingInfo.end.presentation; + segment[timingInfoProperty].transmuxedDecodeEnd = segmentTimingInfo.end.decode; // mainly used as a reference for debugging + + segment[timingInfoProperty].baseMediaDecodeTime = segmentTimingInfo.baseMediaDecodeTime; + } + appendData_(segmentInfo, result) { + const { + type, + data + } = result; + if (!data || !data.byteLength) { + return; + } + if (type === 'audio' && this.audioDisabled_) { + return; + } + const initSegment = this.getInitSegmentAndUpdateState_({ + type, + initSegment: result.initSegment, + playlist: segmentInfo.playlist, + map: segmentInfo.isFmp4 ? segmentInfo.segment.map : null + }); + this.appendToSourceBuffer_({ + segmentInfo, + type, + initSegment, + data + }); + } + /** + * load a specific segment from a request into the buffer + * + * @private + */ + + loadSegment_(segmentInfo) { + this.state = 'WAITING'; + this.pendingSegment_ = segmentInfo; + this.trimBackBuffer_(segmentInfo); + if (typeof segmentInfo.timestampOffset === 'number') { + if (this.transmuxer_) { + this.transmuxer_.postMessage({ + action: 'clearAllMp4Captions' + }); + } + } + if (!this.hasEnoughInfoToLoad_()) { + this.loadQueue_.push(() => { + // regenerate the audioAppendStart, timestampOffset, etc as they + // may have changed since this function was added to the queue. + const options = _extends({}, segmentInfo, { + forceTimestampOffset: true + }); + _extends(segmentInfo, this.generateSegmentInfo_(options)); + this.isPendingTimestampOffset_ = false; + this.updateTransmuxerAndRequestSegment_(segmentInfo); + }); + return; + } + this.updateTransmuxerAndRequestSegment_(segmentInfo); + } + updateTransmuxerAndRequestSegment_(segmentInfo) { + // We'll update the source buffer's timestamp offset once we have transmuxed data, but + // the transmuxer still needs to be updated before then. + // + // Even though keepOriginalTimestamps is set to true for the transmuxer, timestamp + // offset must be passed to the transmuxer for stream correcting adjustments. + if (this.shouldUpdateTransmuxerTimestampOffset_(segmentInfo.timestampOffset)) { + this.gopBuffer_.length = 0; // gopsToAlignWith was set before the GOP buffer was cleared + + segmentInfo.gopsToAlignWith = []; + this.timeMapping_ = 0; // reset values in the transmuxer since a discontinuity should start fresh + + this.transmuxer_.postMessage({ + action: 'reset' + }); + this.transmuxer_.postMessage({ + action: 'setTimestampOffset', + timestampOffset: segmentInfo.timestampOffset + }); + } + const simpleSegment = this.createSimplifiedSegmentObj_(segmentInfo); + const isEndOfStream = this.isEndOfStream_(segmentInfo.mediaIndex, segmentInfo.playlist, segmentInfo.partIndex); + const isWalkingForward = this.mediaIndex !== null; + const isDiscontinuity = segmentInfo.timeline !== this.currentTimeline_ && + // currentTimeline starts at -1, so we shouldn't end the timeline switching to 0, + // the first timeline + segmentInfo.timeline > 0; + const isEndOfTimeline = isEndOfStream || isWalkingForward && isDiscontinuity; + this.logger_(`Requesting +${compactSegmentUrlDescription(segmentInfo.uri)} +${segmentInfoString(segmentInfo)}`); // If there's an init segment associated with this segment, but it is not cached (identified by a lack of bytes), + // then this init segment has never been seen before and should be appended. + // + // At this point the content type (audio/video or both) is not yet known, but it should be safe to set + // both to true and leave the decision of whether to append the init segment to append time. + + if (simpleSegment.map && !simpleSegment.map.bytes) { + this.logger_('going to request init segment.'); + this.appendInitSegment_ = { + video: true, + audio: true + }; + } + segmentInfo.abortRequests = mediaSegmentRequest({ + xhr: this.vhs_.xhr, + xhrOptions: this.xhrOptions_, + decryptionWorker: this.decrypter_, + segment: simpleSegment, + abortFn: this.handleAbort_.bind(this, segmentInfo), + progressFn: this.handleProgress_.bind(this), + trackInfoFn: this.handleTrackInfo_.bind(this), + timingInfoFn: this.handleTimingInfo_.bind(this), + videoSegmentTimingInfoFn: this.handleSegmentTimingInfo_.bind(this, 'video', segmentInfo.requestId), + audioSegmentTimingInfoFn: this.handleSegmentTimingInfo_.bind(this, 'audio', segmentInfo.requestId), + captionsFn: this.handleCaptions_.bind(this), + isEndOfTimeline, + endedTimelineFn: () => { + this.logger_('received endedtimeline callback'); + }, + id3Fn: this.handleId3_.bind(this), + dataFn: this.handleData_.bind(this), + doneFn: this.segmentRequestFinished_.bind(this), + onTransmuxerLog: ({ + message, + level, + stream + }) => { + this.logger_(`${segmentInfoString(segmentInfo)} logged from transmuxer stream ${stream} as a ${level}: ${message}`); + } + }); + } + /** + * trim the back buffer so that we don't have too much data + * in the source buffer + * + * @private + * + * @param {Object} segmentInfo - the current segment + */ + + trimBackBuffer_(segmentInfo) { + const removeToTime = safeBackBufferTrimTime(this.seekable_(), this.currentTime_(), this.playlist_.targetDuration || 10); // Chrome has a hard limit of 150MB of + // buffer and a very conservative "garbage collector" + // We manually clear out the old buffer to ensure + // we don't trigger the QuotaExceeded error + // on the source buffer during subsequent appends + + if (removeToTime > 0) { + this.remove(0, removeToTime); + } + } + /** + * created a simplified copy of the segment object with just the + * information necessary to perform the XHR and decryption + * + * @private + * + * @param {Object} segmentInfo - the current segment + * @return {Object} a simplified segment object copy + */ + + createSimplifiedSegmentObj_(segmentInfo) { + const segment = segmentInfo.segment; + const part = segmentInfo.part; + const simpleSegment = { + resolvedUri: part ? part.resolvedUri : segment.resolvedUri, + byterange: part ? part.byterange : segment.byterange, + requestId: segmentInfo.requestId, + transmuxer: segmentInfo.transmuxer, + audioAppendStart: segmentInfo.audioAppendStart, + gopsToAlignWith: segmentInfo.gopsToAlignWith, + part: segmentInfo.part + }; + const previousSegment = segmentInfo.playlist.segments[segmentInfo.mediaIndex - 1]; + if (previousSegment && previousSegment.timeline === segment.timeline) { + // The baseStartTime of a segment is used to handle rollover when probing the TS + // segment to retrieve timing information. Since the probe only looks at the media's + // times (e.g., PTS and DTS values of the segment), and doesn't consider the + // player's time (e.g., player.currentTime()), baseStartTime should reflect the + // media time as well. transmuxedDecodeEnd represents the end time of a segment, in + // seconds of media time, so should be used here. The previous segment is used since + // the end of the previous segment should represent the beginning of the current + // segment, so long as they are on the same timeline. + if (previousSegment.videoTimingInfo) { + simpleSegment.baseStartTime = previousSegment.videoTimingInfo.transmuxedDecodeEnd; + } else if (previousSegment.audioTimingInfo) { + simpleSegment.baseStartTime = previousSegment.audioTimingInfo.transmuxedDecodeEnd; + } + } + if (segment.key) { + // if the media sequence is greater than 2^32, the IV will be incorrect + // assuming 10s segments, that would be about 1300 years + const iv = segment.key.iv || new Uint32Array([0, 0, 0, segmentInfo.mediaIndex + segmentInfo.playlist.mediaSequence]); + simpleSegment.key = this.segmentKey(segment.key); + simpleSegment.key.iv = iv; + } + if (segment.map) { + simpleSegment.map = this.initSegmentForMap(segment.map); + } + return simpleSegment; + } + saveTransferStats_(stats) { + // every request counts as a media request even if it has been aborted + // or canceled due to a timeout + this.mediaRequests += 1; + if (stats) { + this.mediaBytesTransferred += stats.bytesReceived; + this.mediaTransferDuration += stats.roundTripTime; + } + } + saveBandwidthRelatedStats_(duration, stats) { + // byteLength will be used for throughput, and should be based on bytes receieved, + // which we only know at the end of the request and should reflect total bytes + // downloaded rather than just bytes processed from components of the segment + this.pendingSegment_.byteLength = stats.bytesReceived; + if (duration < MIN_SEGMENT_DURATION_TO_SAVE_STATS) { + this.logger_(`Ignoring segment's bandwidth because its duration of ${duration}` + ` is less than the min to record ${MIN_SEGMENT_DURATION_TO_SAVE_STATS}`); + return; + } + this.bandwidth = stats.bandwidth; + this.roundTrip = stats.roundTripTime; + } + handleTimeout_() { + // although the VTT segment loader bandwidth isn't really used, it's good to + // maintain functinality between segment loaders + this.mediaRequestsTimedout += 1; + this.bandwidth = 1; + this.roundTrip = NaN; + this.trigger('bandwidthupdate'); + this.trigger('timeout'); + } + /** + * Handle the callback from the segmentRequest function and set the + * associated SegmentLoader state and errors if necessary + * + * @private + */ + + segmentRequestFinished_(error, simpleSegment, result) { + // TODO handle special cases, e.g., muxed audio/video but only audio in the segment + // check the call queue directly since this function doesn't need to deal with any + // data, and can continue even if the source buffers are not set up and we didn't get + // any data from the segment + if (this.callQueue_.length) { + this.callQueue_.push(this.segmentRequestFinished_.bind(this, error, simpleSegment, result)); + return; + } + this.saveTransferStats_(simpleSegment.stats); // The request was aborted and the SegmentLoader has already been reset + + if (!this.pendingSegment_) { + return; + } // the request was aborted and the SegmentLoader has already started + // another request. this can happen when the timeout for an aborted + // request triggers due to a limitation in the XHR library + // do not count this as any sort of request or we risk double-counting + + if (simpleSegment.requestId !== this.pendingSegment_.requestId) { + return; + } // an error occurred from the active pendingSegment_ so reset everything + + if (error) { + this.pendingSegment_ = null; + this.state = 'READY'; // aborts are not a true error condition and nothing corrective needs to be done + + if (error.code === REQUEST_ERRORS.ABORTED) { + return; + } + this.pause(); // the error is really just that at least one of the requests timed-out + // set the bandwidth to a very low value and trigger an ABR switch to + // take emergency action + + if (error.code === REQUEST_ERRORS.TIMEOUT) { + this.handleTimeout_(); + return; + } // if control-flow has arrived here, then the error is real + // emit an error event to exclude the current playlist + + this.mediaRequestsErrored += 1; + this.error(error); + this.trigger('error'); + return; + } + const segmentInfo = this.pendingSegment_; // the response was a success so set any bandwidth stats the request + // generated for ABR purposes + + this.saveBandwidthRelatedStats_(segmentInfo.duration, simpleSegment.stats); + segmentInfo.endOfAllRequests = simpleSegment.endOfAllRequests; + if (result.gopInfo) { + this.gopBuffer_ = updateGopBuffer(this.gopBuffer_, result.gopInfo, this.safeAppend_); + } // Although we may have already started appending on progress, we shouldn't switch the + // state away from loading until we are officially done loading the segment data. + + this.state = 'APPENDING'; // used for testing + + this.trigger('appending'); + this.waitForAppendsToComplete_(segmentInfo); + } + setTimeMapping_(timeline) { + const timelineMapping = this.syncController_.mappingForTimeline(timeline); + if (timelineMapping !== null) { + this.timeMapping_ = timelineMapping; + } + } + updateMediaSecondsLoaded_(segment) { + if (typeof segment.start === 'number' && typeof segment.end === 'number') { + this.mediaSecondsLoaded += segment.end - segment.start; + } else { + this.mediaSecondsLoaded += segment.duration; + } + } + shouldUpdateTransmuxerTimestampOffset_(timestampOffset) { + if (timestampOffset === null) { + return false; + } // note that we're potentially using the same timestamp offset for both video and + // audio + + if (this.loaderType_ === 'main' && timestampOffset !== this.sourceUpdater_.videoTimestampOffset()) { + return true; + } + if (!this.audioDisabled_ && timestampOffset !== this.sourceUpdater_.audioTimestampOffset()) { + return true; + } + return false; + } + trueSegmentStart_({ + currentStart, + playlist, + mediaIndex, + firstVideoFrameTimeForData, + currentVideoTimestampOffset, + useVideoTimingInfo, + videoTimingInfo, + audioTimingInfo + }) { + if (typeof currentStart !== 'undefined') { + // if start was set once, keep using it + return currentStart; + } + if (!useVideoTimingInfo) { + return audioTimingInfo.start; + } + const previousSegment = playlist.segments[mediaIndex - 1]; // The start of a segment should be the start of the first full frame contained + // within that segment. Since the transmuxer maintains a cache of incomplete data + // from and/or the last frame seen, the start time may reflect a frame that starts + // in the previous segment. Check for that case and ensure the start time is + // accurate for the segment. + + if (mediaIndex === 0 || !previousSegment || typeof previousSegment.start === 'undefined' || previousSegment.end !== firstVideoFrameTimeForData + currentVideoTimestampOffset) { + return firstVideoFrameTimeForData; + } + return videoTimingInfo.start; + } + waitForAppendsToComplete_(segmentInfo) { + const trackInfo = this.getCurrentMediaInfo_(segmentInfo); + if (!trackInfo) { + this.error({ + message: 'No starting media returned, likely due to an unsupported media format.', + playlistExclusionDuration: Infinity, + metadata: { + errorType: videojs.Error.SegmentUnsupportedMediaFormat + } + }); + this.trigger('error'); + return; + } // Although transmuxing is done, appends may not yet be finished. Throw a marker + // on each queue this loader is responsible for to ensure that the appends are + // complete. + + const { + hasAudio, + hasVideo, + isMuxed + } = trackInfo; + const waitForVideo = this.loaderType_ === 'main' && hasVideo; + const waitForAudio = !this.audioDisabled_ && hasAudio && !isMuxed; + segmentInfo.waitingOnAppends = 0; // segments with no data + + if (!segmentInfo.hasAppendedData_) { + if (!segmentInfo.timingInfo && typeof segmentInfo.timestampOffset === 'number') { + // When there's no audio or video data in the segment, there's no audio or video + // timing information. + // + // If there's no audio or video timing information, then the timestamp offset + // can't be adjusted to the appropriate value for the transmuxer and source + // buffers. + // + // Therefore, the next segment should be used to set the timestamp offset. + this.isPendingTimestampOffset_ = true; + } // override settings for metadata only segments + + segmentInfo.timingInfo = { + start: 0 + }; + segmentInfo.waitingOnAppends++; + if (!this.isPendingTimestampOffset_) { + // update the timestampoffset + this.updateSourceBufferTimestampOffset_(segmentInfo); // make sure the metadata queue is processed even though we have + // no video/audio data. + + this.processMetadataQueue_(); + } // append is "done" instantly with no data. + + this.checkAppendsDone_(segmentInfo); + return; + } // Since source updater could call back synchronously, do the increments first. + + if (waitForVideo) { + segmentInfo.waitingOnAppends++; + } + if (waitForAudio) { + segmentInfo.waitingOnAppends++; + } + if (waitForVideo) { + this.sourceUpdater_.videoQueueCallback(this.checkAppendsDone_.bind(this, segmentInfo)); + } + if (waitForAudio) { + this.sourceUpdater_.audioQueueCallback(this.checkAppendsDone_.bind(this, segmentInfo)); + } + } + checkAppendsDone_(segmentInfo) { + if (this.checkForAbort_(segmentInfo.requestId)) { + return; + } + segmentInfo.waitingOnAppends--; + if (segmentInfo.waitingOnAppends === 0) { + this.handleAppendsDone_(); + } + } + checkForIllegalMediaSwitch(trackInfo) { + const illegalMediaSwitchError = illegalMediaSwitch(this.loaderType_, this.getCurrentMediaInfo_(), trackInfo); + if (illegalMediaSwitchError) { + this.error({ + message: illegalMediaSwitchError, + playlistExclusionDuration: Infinity, + metadata: { + errorType: videojs.Error.SegmentSwitchError + } + }); + this.trigger('error'); + return true; + } + return false; + } + updateSourceBufferTimestampOffset_(segmentInfo) { + if (segmentInfo.timestampOffset === null || + // we don't yet have the start for whatever media type (video or audio) has + // priority, timing-wise, so we must wait + typeof segmentInfo.timingInfo.start !== 'number' || + // already updated the timestamp offset for this segment + segmentInfo.changedTimestampOffset || + // the alt audio loader should not be responsible for setting the timestamp offset + this.loaderType_ !== 'main') { + return; + } + let didChange = false; // Primary timing goes by video, and audio is trimmed in the transmuxer, meaning that + // the timing info here comes from video. In the event that the audio is longer than + // the video, this will trim the start of the audio. + // This also trims any offset from 0 at the beginning of the media + + segmentInfo.timestampOffset -= this.getSegmentStartTimeForTimestampOffsetCalculation_({ + videoTimingInfo: segmentInfo.segment.videoTimingInfo, + audioTimingInfo: segmentInfo.segment.audioTimingInfo, + timingInfo: segmentInfo.timingInfo + }); // In the event that there are part segment downloads, each will try to update the + // timestamp offset. Retaining this bit of state prevents us from updating in the + // future (within the same segment), however, there may be a better way to handle it. + + segmentInfo.changedTimestampOffset = true; + if (segmentInfo.timestampOffset !== this.sourceUpdater_.videoTimestampOffset()) { + this.sourceUpdater_.videoTimestampOffset(segmentInfo.timestampOffset); + didChange = true; + } + if (segmentInfo.timestampOffset !== this.sourceUpdater_.audioTimestampOffset()) { + this.sourceUpdater_.audioTimestampOffset(segmentInfo.timestampOffset); + didChange = true; + } + if (didChange) { + this.trigger('timestampoffset'); + } + } + getSegmentStartTimeForTimestampOffsetCalculation_({ + videoTimingInfo, + audioTimingInfo, + timingInfo + }) { + if (!this.useDtsForTimestampOffset_) { + return timingInfo.start; + } + if (videoTimingInfo && typeof videoTimingInfo.transmuxedDecodeStart === 'number') { + return videoTimingInfo.transmuxedDecodeStart; + } // handle audio only + + if (audioTimingInfo && typeof audioTimingInfo.transmuxedDecodeStart === 'number') { + return audioTimingInfo.transmuxedDecodeStart; + } // handle content not transmuxed (e.g., MP4) + + return timingInfo.start; + } + updateTimingInfoEnd_(segmentInfo) { + segmentInfo.timingInfo = segmentInfo.timingInfo || {}; + const trackInfo = this.getMediaInfo_(); + const useVideoTimingInfo = this.loaderType_ === 'main' && trackInfo && trackInfo.hasVideo; + const prioritizedTimingInfo = useVideoTimingInfo && segmentInfo.videoTimingInfo ? segmentInfo.videoTimingInfo : segmentInfo.audioTimingInfo; + if (!prioritizedTimingInfo) { + return; + } + segmentInfo.timingInfo.end = typeof prioritizedTimingInfo.end === 'number' ? + // End time may not exist in a case where we aren't parsing the full segment (one + // current example is the case of fmp4), so use the rough duration to calculate an + // end time. + prioritizedTimingInfo.end : prioritizedTimingInfo.start + segmentInfo.duration; + } + /** + * callback to run when appendBuffer is finished. detects if we are + * in a good state to do things with the data we got, or if we need + * to wait for more + * + * @private + */ + + handleAppendsDone_() { + // appendsdone can cause an abort + if (this.pendingSegment_) { + this.trigger('appendsdone'); + } + if (!this.pendingSegment_) { + this.state = 'READY'; // TODO should this move into this.checkForAbort to speed up requests post abort in + // all appending cases? + + if (!this.paused()) { + this.monitorBuffer_(); + } + return; + } + const segmentInfo = this.pendingSegment_; + if (segmentInfo.part && segmentInfo.part.syncInfo) { + // low-latency flow + segmentInfo.part.syncInfo.markAppended(); + } else if (segmentInfo.segment.syncInfo) { + // normal flow + segmentInfo.segment.syncInfo.markAppended(); + } // Now that the end of the segment has been reached, we can set the end time. It's + // best to wait until all appends are done so we're sure that the primary media is + // finished (and we have its end time). + + this.updateTimingInfoEnd_(segmentInfo); + if (this.shouldSaveSegmentTimingInfo_) { + // Timeline mappings should only be saved for the main loader. This is for multiple + // reasons: + // + // 1) Only one mapping is saved per timeline, meaning that if both the audio loader + // and the main loader try to save the timeline mapping, whichever comes later + // will overwrite the first. In theory this is OK, as the mappings should be the + // same, however, it breaks for (2) + // 2) In the event of a live stream, the initial live point will make for a somewhat + // arbitrary mapping. If audio and video streams are not perfectly in-sync, then + // the mapping will be off for one of the streams, dependent on which one was + // first saved (see (1)). + // 3) Primary timing goes by video in VHS, so the mapping should be video. + // + // Since the audio loader will wait for the main loader to load the first segment, + // the main loader will save the first timeline mapping, and ensure that there won't + // be a case where audio loads two segments without saving a mapping (thus leading + // to missing segment timing info). + this.syncController_.saveSegmentTimingInfo({ + segmentInfo, + shouldSaveTimelineMapping: this.loaderType_ === 'main' + }); + } + const segmentDurationMessage = getTroublesomeSegmentDurationMessage(segmentInfo, this.sourceType_); + if (segmentDurationMessage) { + if (segmentDurationMessage.severity === 'warn') { + videojs.log.warn(segmentDurationMessage.message); + } else { + this.logger_(segmentDurationMessage.message); + } + } + this.recordThroughput_(segmentInfo); + this.pendingSegment_ = null; + this.state = 'READY'; + if (segmentInfo.isSyncRequest) { + this.trigger('syncinfoupdate'); // if the sync request was not appended + // then it was not the correct segment. + // throw it away and use the data it gave us + // to get the correct one. + + if (!segmentInfo.hasAppendedData_) { + this.logger_(`Throwing away un-appended sync request ${segmentInfoString(segmentInfo)}`); + return; + } + } + this.logger_(`Appended ${segmentInfoString(segmentInfo)}`); + this.addSegmentMetadataCue_(segmentInfo); + this.fetchAtBuffer_ = true; + if (this.currentTimeline_ !== segmentInfo.timeline) { + this.timelineChangeController_.lastTimelineChange({ + type: this.loaderType_, + from: this.currentTimeline_, + to: segmentInfo.timeline + }); // If audio is not disabled, the main segment loader is responsible for updating + // the audio timeline as well. If the content is video only, this won't have any + // impact. + + if (this.loaderType_ === 'main' && !this.audioDisabled_) { + this.timelineChangeController_.lastTimelineChange({ + type: 'audio', + from: this.currentTimeline_, + to: segmentInfo.timeline + }); + } + } + this.currentTimeline_ = segmentInfo.timeline; // We must update the syncinfo to recalculate the seekable range before + // the following conditional otherwise it may consider this a bad "guess" + // and attempt to resync when the post-update seekable window and live + // point would mean that this was the perfect segment to fetch + + this.trigger('syncinfoupdate'); + const segment = segmentInfo.segment; + const part = segmentInfo.part; + const badSegmentGuess = segment.end && this.currentTime_() - segment.end > segmentInfo.playlist.targetDuration * 3; + const badPartGuess = part && part.end && this.currentTime_() - part.end > segmentInfo.playlist.partTargetDuration * 3; // If we previously appended a segment/part that ends more than 3 part/targetDurations before + // the currentTime_ that means that our conservative guess was too conservative. + // In that case, reset the loader state so that we try to use any information gained + // from the previous request to create a new, more accurate, sync-point. + + if (badSegmentGuess || badPartGuess) { + this.logger_(`bad ${badSegmentGuess ? 'segment' : 'part'} ${segmentInfoString(segmentInfo)}`); + this.resetEverything(); + return; + } + const isWalkingForward = this.mediaIndex !== null; // Don't do a rendition switch unless we have enough time to get a sync segment + // and conservatively guess + + if (isWalkingForward) { + this.trigger('bandwidthupdate'); + } + this.trigger('progress'); + this.mediaIndex = segmentInfo.mediaIndex; + this.partIndex = segmentInfo.partIndex; // any time an update finishes and the last segment is in the + // buffer, end the stream. this ensures the "ended" event will + // fire if playback reaches that point. + + if (this.isEndOfStream_(segmentInfo.mediaIndex, segmentInfo.playlist, segmentInfo.partIndex)) { + this.endOfStream(); + } // used for testing + + this.trigger('appended'); + if (segmentInfo.hasAppendedData_) { + this.mediaAppends++; + } + if (!this.paused()) { + this.monitorBuffer_(); + } + } + /** + * Records the current throughput of the decrypt, transmux, and append + * portion of the semgment pipeline. `throughput.rate` is a the cumulative + * moving average of the throughput. `throughput.count` is the number of + * data points in the average. + * + * @private + * @param {Object} segmentInfo the object returned by loadSegment + */ + + recordThroughput_(segmentInfo) { + if (segmentInfo.duration < MIN_SEGMENT_DURATION_TO_SAVE_STATS) { + this.logger_(`Ignoring segment's throughput because its duration of ${segmentInfo.duration}` + ` is less than the min to record ${MIN_SEGMENT_DURATION_TO_SAVE_STATS}`); + return; + } + const rate = this.throughput.rate; // Add one to the time to ensure that we don't accidentally attempt to divide + // by zero in the case where the throughput is ridiculously high + + const segmentProcessingTime = Date.now() - segmentInfo.endOfAllRequests + 1; // Multiply by 8000 to convert from bytes/millisecond to bits/second + + const segmentProcessingThroughput = Math.floor(segmentInfo.byteLength / segmentProcessingTime * 8 * 1000); // This is just a cumulative moving average calculation: + // newAvg = oldAvg + (sample - oldAvg) / (sampleCount + 1) + + this.throughput.rate += (segmentProcessingThroughput - rate) / ++this.throughput.count; + } + /** + * Adds a cue to the segment-metadata track with some metadata information about the + * segment + * + * @private + * @param {Object} segmentInfo + * the object returned by loadSegment + * @method addSegmentMetadataCue_ + */ + + addSegmentMetadataCue_(segmentInfo) { + if (!this.segmentMetadataTrack_) { + return; + } + const segment = segmentInfo.segment; + const start = segment.start; + const end = segment.end; // Do not try adding the cue if the start and end times are invalid. + + if (!finite(start) || !finite(end)) { + return; + } + removeCuesFromTrack(start, end, this.segmentMetadataTrack_); + const Cue = window$1.WebKitDataCue || window$1.VTTCue; + const value = { + custom: segment.custom, + dateTimeObject: segment.dateTimeObject, + dateTimeString: segment.dateTimeString, + programDateTime: segment.programDateTime, + bandwidth: segmentInfo.playlist.attributes.BANDWIDTH, + resolution: segmentInfo.playlist.attributes.RESOLUTION, + codecs: segmentInfo.playlist.attributes.CODECS, + byteLength: segmentInfo.byteLength, + uri: segmentInfo.uri, + timeline: segmentInfo.timeline, + playlist: segmentInfo.playlist.id, + start, + end + }; + const data = JSON.stringify(value); + const cue = new Cue(start, end, data); // Attach the metadata to the value property of the cue to keep consistency between + // the differences of WebKitDataCue in safari and VTTCue in other browsers + + cue.value = value; + this.segmentMetadataTrack_.addCue(cue); + } +} +function noop() {} +const toTitleCase = function (string) { + if (typeof string !== 'string') { + return string; + } + return string.replace(/./, w => w.toUpperCase()); +}; + +/** + * @file source-updater.js + */ +const bufferTypes = ['video', 'audio']; +const updating = (type, sourceUpdater) => { + const sourceBuffer = sourceUpdater[`${type}Buffer`]; + return sourceBuffer && sourceBuffer.updating || sourceUpdater.queuePending[type]; +}; +const nextQueueIndexOfType = (type, queue) => { + for (let i = 0; i < queue.length; i++) { + const queueEntry = queue[i]; + if (queueEntry.type === 'mediaSource') { + // If the next entry is a media source entry (uses multiple source buffers), block + // processing to allow it to go through first. + return null; + } + if (queueEntry.type === type) { + return i; + } + } + return null; +}; +const shiftQueue = (type, sourceUpdater) => { + if (sourceUpdater.queue.length === 0) { + return; + } + let queueIndex = 0; + let queueEntry = sourceUpdater.queue[queueIndex]; + if (queueEntry.type === 'mediaSource') { + if (!sourceUpdater.updating() && sourceUpdater.mediaSource.readyState !== 'closed') { + sourceUpdater.queue.shift(); + queueEntry.action(sourceUpdater); + if (queueEntry.doneFn) { + queueEntry.doneFn(); + } // Only specific source buffer actions must wait for async updateend events. Media + // Source actions process synchronously. Therefore, both audio and video source + // buffers are now clear to process the next queue entries. + + shiftQueue('audio', sourceUpdater); + shiftQueue('video', sourceUpdater); + } // Media Source actions require both source buffers, so if the media source action + // couldn't process yet (because one or both source buffers are busy), block other + // queue actions until both are available and the media source action can process. + + return; + } + if (type === 'mediaSource') { + // If the queue was shifted by a media source action (this happens when pushing a + // media source action onto the queue), then it wasn't from an updateend event from an + // audio or video source buffer, so there's no change from previous state, and no + // processing should be done. + return; + } // Media source queue entries don't need to consider whether the source updater is + // started (i.e., source buffers are created) as they don't need the source buffers, but + // source buffer queue entries do. + + if (!sourceUpdater.ready() || sourceUpdater.mediaSource.readyState === 'closed' || updating(type, sourceUpdater)) { + return; + } + if (queueEntry.type !== type) { + queueIndex = nextQueueIndexOfType(type, sourceUpdater.queue); + if (queueIndex === null) { + // Either there's no queue entry that uses this source buffer type in the queue, or + // there's a media source queue entry before the next entry of this type, in which + // case wait for that action to process first. + return; + } + queueEntry = sourceUpdater.queue[queueIndex]; + } + sourceUpdater.queue.splice(queueIndex, 1); // Keep a record that this source buffer type is in use. + // + // The queue pending operation must be set before the action is performed in the event + // that the action results in a synchronous event that is acted upon. For instance, if + // an exception is thrown that can be handled, it's possible that new actions will be + // appended to an empty queue and immediately executed, but would not have the correct + // pending information if this property was set after the action was performed. + + sourceUpdater.queuePending[type] = queueEntry; + queueEntry.action(type, sourceUpdater); + if (!queueEntry.doneFn) { + // synchronous operation, process next entry + sourceUpdater.queuePending[type] = null; + shiftQueue(type, sourceUpdater); + return; + } +}; +const cleanupBuffer = (type, sourceUpdater) => { + const buffer = sourceUpdater[`${type}Buffer`]; + const titleType = toTitleCase(type); + if (!buffer) { + return; + } + buffer.removeEventListener('updateend', sourceUpdater[`on${titleType}UpdateEnd_`]); + buffer.removeEventListener('error', sourceUpdater[`on${titleType}Error_`]); + sourceUpdater.codecs[type] = null; + sourceUpdater[`${type}Buffer`] = null; +}; +const inSourceBuffers = (mediaSource, sourceBuffer) => mediaSource && sourceBuffer && Array.prototype.indexOf.call(mediaSource.sourceBuffers, sourceBuffer) !== -1; +const actions = { + appendBuffer: (bytes, segmentInfo, onError) => (type, sourceUpdater) => { + const sourceBuffer = sourceUpdater[`${type}Buffer`]; // can't do anything if the media source / source buffer is null + // or the media source does not contain this source buffer. + + if (!inSourceBuffers(sourceUpdater.mediaSource, sourceBuffer)) { + return; + } + sourceUpdater.logger_(`Appending segment ${segmentInfo.mediaIndex}'s ${bytes.length} bytes to ${type}Buffer`); + try { + sourceBuffer.appendBuffer(bytes); + } catch (e) { + sourceUpdater.logger_(`Error with code ${e.code} ` + (e.code === QUOTA_EXCEEDED_ERR ? '(QUOTA_EXCEEDED_ERR) ' : '') + `when appending segment ${segmentInfo.mediaIndex} to ${type}Buffer`); + sourceUpdater.queuePending[type] = null; + onError(e); + } + }, + remove: (start, end) => (type, sourceUpdater) => { + const sourceBuffer = sourceUpdater[`${type}Buffer`]; // can't do anything if the media source / source buffer is null + // or the media source does not contain this source buffer. + + if (!inSourceBuffers(sourceUpdater.mediaSource, sourceBuffer)) { + return; + } + sourceUpdater.logger_(`Removing ${start} to ${end} from ${type}Buffer`); + try { + sourceBuffer.remove(start, end); + } catch (e) { + sourceUpdater.logger_(`Remove ${start} to ${end} from ${type}Buffer failed`); + } + }, + timestampOffset: offset => (type, sourceUpdater) => { + const sourceBuffer = sourceUpdater[`${type}Buffer`]; // can't do anything if the media source / source buffer is null + // or the media source does not contain this source buffer. + + if (!inSourceBuffers(sourceUpdater.mediaSource, sourceBuffer)) { + return; + } + sourceUpdater.logger_(`Setting ${type}timestampOffset to ${offset}`); + sourceBuffer.timestampOffset = offset; + }, + callback: callback => (type, sourceUpdater) => { + callback(); + }, + endOfStream: error => sourceUpdater => { + if (sourceUpdater.mediaSource.readyState !== 'open') { + return; + } + sourceUpdater.logger_(`Calling mediaSource endOfStream(${error || ''})`); + try { + sourceUpdater.mediaSource.endOfStream(error); + } catch (e) { + videojs.log.warn('Failed to call media source endOfStream', e); + } + }, + duration: duration => sourceUpdater => { + sourceUpdater.logger_(`Setting mediaSource duration to ${duration}`); + try { + sourceUpdater.mediaSource.duration = duration; + } catch (e) { + videojs.log.warn('Failed to set media source duration', e); + } + }, + abort: () => (type, sourceUpdater) => { + if (sourceUpdater.mediaSource.readyState !== 'open') { + return; + } + const sourceBuffer = sourceUpdater[`${type}Buffer`]; // can't do anything if the media source / source buffer is null + // or the media source does not contain this source buffer. + + if (!inSourceBuffers(sourceUpdater.mediaSource, sourceBuffer)) { + return; + } + sourceUpdater.logger_(`calling abort on ${type}Buffer`); + try { + sourceBuffer.abort(); + } catch (e) { + videojs.log.warn(`Failed to abort on ${type}Buffer`, e); + } + }, + addSourceBuffer: (type, codec) => sourceUpdater => { + const titleType = toTitleCase(type); + const mime = getMimeForCodec(codec); + sourceUpdater.logger_(`Adding ${type}Buffer with codec ${codec} to mediaSource`); + const sourceBuffer = sourceUpdater.mediaSource.addSourceBuffer(mime); + sourceBuffer.addEventListener('updateend', sourceUpdater[`on${titleType}UpdateEnd_`]); + sourceBuffer.addEventListener('error', sourceUpdater[`on${titleType}Error_`]); + sourceUpdater.codecs[type] = codec; + sourceUpdater[`${type}Buffer`] = sourceBuffer; + }, + removeSourceBuffer: type => sourceUpdater => { + const sourceBuffer = sourceUpdater[`${type}Buffer`]; + cleanupBuffer(type, sourceUpdater); // can't do anything if the media source / source buffer is null + // or the media source does not contain this source buffer. + + if (!inSourceBuffers(sourceUpdater.mediaSource, sourceBuffer)) { + return; + } + sourceUpdater.logger_(`Removing ${type}Buffer with codec ${sourceUpdater.codecs[type]} from mediaSource`); + try { + sourceUpdater.mediaSource.removeSourceBuffer(sourceBuffer); + } catch (e) { + videojs.log.warn(`Failed to removeSourceBuffer ${type}Buffer`, e); + } + }, + changeType: codec => (type, sourceUpdater) => { + const sourceBuffer = sourceUpdater[`${type}Buffer`]; + const mime = getMimeForCodec(codec); // can't do anything if the media source / source buffer is null + // or the media source does not contain this source buffer. + + if (!inSourceBuffers(sourceUpdater.mediaSource, sourceBuffer)) { + return; + } // do not update codec if we don't need to. + // Only update if we change the codec base. + // For example, going from avc1.640028 to avc1.64001f does not require a changeType call. + + const newCodecBase = codec.substring(0, codec.indexOf('.')); + const oldCodec = sourceUpdater.codecs[type]; + const oldCodecBase = oldCodec.substring(0, oldCodec.indexOf('.')); + if (oldCodecBase === newCodecBase) { + return; + } + sourceUpdater.logger_(`changing ${type}Buffer codec from ${sourceUpdater.codecs[type]} to ${codec}`); // check if change to the provided type is supported + + try { + sourceBuffer.changeType(mime); + sourceUpdater.codecs[type] = codec; + } catch (e) { + videojs.log.warn(`Failed to changeType on ${type}Buffer`, e); + } + } +}; +const pushQueue = ({ + type, + sourceUpdater, + action, + doneFn, + name +}) => { + sourceUpdater.queue.push({ + type, + action, + doneFn, + name + }); + shiftQueue(type, sourceUpdater); +}; +const onUpdateend = (type, sourceUpdater) => e => { + // Although there should, in theory, be a pending action for any updateend receieved, + // there are some actions that may trigger updateend events without set definitions in + // the w3c spec. For instance, setting the duration on the media source may trigger + // updateend events on source buffers. This does not appear to be in the spec. As such, + // if we encounter an updateend without a corresponding pending action from our queue + // for that source buffer type, process the next action. + const bufferedRangesForType = sourceUpdater[`${type}Buffered`](); + const descriptiveString = bufferedRangesToString(bufferedRangesForType); + sourceUpdater.logger_(`received "updateend" event for ${type} Source Buffer: `, descriptiveString); + if (sourceUpdater.queuePending[type]) { + const doneFn = sourceUpdater.queuePending[type].doneFn; + sourceUpdater.queuePending[type] = null; + if (doneFn) { + // if there's an error, report it + doneFn(sourceUpdater[`${type}Error_`]); + } + } + shiftQueue(type, sourceUpdater); +}; +/** + * A queue of callbacks to be serialized and applied when a + * MediaSource and its associated SourceBuffers are not in the + * updating state. It is used by the segment loader to update the + * underlying SourceBuffers when new data is loaded, for instance. + * + * @class SourceUpdater + * @param {MediaSource} mediaSource the MediaSource to create the SourceBuffer from + * @param {string} mimeType the desired MIME type of the underlying SourceBuffer + */ + +class SourceUpdater extends videojs.EventTarget { + constructor(mediaSource) { + super(); + this.mediaSource = mediaSource; + this.sourceopenListener_ = () => shiftQueue('mediaSource', this); + this.mediaSource.addEventListener('sourceopen', this.sourceopenListener_); + this.logger_ = logger('SourceUpdater'); // initial timestamp offset is 0 + + this.audioTimestampOffset_ = 0; + this.videoTimestampOffset_ = 0; + this.queue = []; + this.queuePending = { + audio: null, + video: null + }; + this.delayedAudioAppendQueue_ = []; + this.videoAppendQueued_ = false; + this.codecs = {}; + this.onVideoUpdateEnd_ = onUpdateend('video', this); + this.onAudioUpdateEnd_ = onUpdateend('audio', this); + this.onVideoError_ = e => { + // used for debugging + this.videoError_ = e; + }; + this.onAudioError_ = e => { + // used for debugging + this.audioError_ = e; + }; + this.createdSourceBuffers_ = false; + this.initializedEme_ = false; + this.triggeredReady_ = false; + } + initializedEme() { + this.initializedEme_ = true; + this.triggerReady(); + } + hasCreatedSourceBuffers() { + // if false, likely waiting on one of the segment loaders to get enough data to create + // source buffers + return this.createdSourceBuffers_; + } + hasInitializedAnyEme() { + return this.initializedEme_; + } + ready() { + return this.hasCreatedSourceBuffers() && this.hasInitializedAnyEme(); + } + createSourceBuffers(codecs) { + if (this.hasCreatedSourceBuffers()) { + // already created them before + return; + } // the intial addOrChangeSourceBuffers will always be + // two add buffers. + + this.addOrChangeSourceBuffers(codecs); + this.createdSourceBuffers_ = true; + this.trigger('createdsourcebuffers'); + this.triggerReady(); + } + triggerReady() { + // only allow ready to be triggered once, this prevents the case + // where: + // 1. we trigger createdsourcebuffers + // 2. ie 11 synchronously initializates eme + // 3. the synchronous initialization causes us to trigger ready + // 4. We go back to the ready check in createSourceBuffers and ready is triggered again. + if (this.ready() && !this.triggeredReady_) { + this.triggeredReady_ = true; + this.trigger('ready'); + } + } + /** + * Add a type of source buffer to the media source. + * + * @param {string} type + * The type of source buffer to add. + * + * @param {string} codec + * The codec to add the source buffer with. + */ + + addSourceBuffer(type, codec) { + pushQueue({ + type: 'mediaSource', + sourceUpdater: this, + action: actions.addSourceBuffer(type, codec), + name: 'addSourceBuffer' + }); + } + /** + * call abort on a source buffer. + * + * @param {string} type + * The type of source buffer to call abort on. + */ + + abort(type) { + pushQueue({ + type, + sourceUpdater: this, + action: actions.abort(type), + name: 'abort' + }); + } + /** + * Call removeSourceBuffer and remove a specific type + * of source buffer on the mediaSource. + * + * @param {string} type + * The type of source buffer to remove. + */ + + removeSourceBuffer(type) { + if (!this.canRemoveSourceBuffer()) { + videojs.log.error('removeSourceBuffer is not supported!'); + return; + } + pushQueue({ + type: 'mediaSource', + sourceUpdater: this, + action: actions.removeSourceBuffer(type), + name: 'removeSourceBuffer' + }); + } + /** + * Whether or not the removeSourceBuffer function is supported + * on the mediaSource. + * + * @return {boolean} + * if removeSourceBuffer can be called. + */ + + canRemoveSourceBuffer() { + // As of Firefox 83 removeSourceBuffer + // throws errors, so we report that it does not support this. + return !videojs.browser.IS_FIREFOX && window$1.MediaSource && window$1.MediaSource.prototype && typeof window$1.MediaSource.prototype.removeSourceBuffer === 'function'; + } + /** + * Whether or not the changeType function is supported + * on our SourceBuffers. + * + * @return {boolean} + * if changeType can be called. + */ + + static canChangeType() { + return window$1.SourceBuffer && window$1.SourceBuffer.prototype && typeof window$1.SourceBuffer.prototype.changeType === 'function'; + } + /** + * Whether or not the changeType function is supported + * on our SourceBuffers. + * + * @return {boolean} + * if changeType can be called. + */ + + canChangeType() { + return this.constructor.canChangeType(); + } + /** + * Call the changeType function on a source buffer, given the code and type. + * + * @param {string} type + * The type of source buffer to call changeType on. + * + * @param {string} codec + * The codec string to change type with on the source buffer. + */ + + changeType(type, codec) { + if (!this.canChangeType()) { + videojs.log.error('changeType is not supported!'); + return; + } + pushQueue({ + type, + sourceUpdater: this, + action: actions.changeType(codec), + name: 'changeType' + }); + } + /** + * Add source buffers with a codec or, if they are already created, + * call changeType on source buffers using changeType. + * + * @param {Object} codecs + * Codecs to switch to + */ + + addOrChangeSourceBuffers(codecs) { + if (!codecs || typeof codecs !== 'object' || Object.keys(codecs).length === 0) { + throw new Error('Cannot addOrChangeSourceBuffers to undefined codecs'); + } + Object.keys(codecs).forEach(type => { + const codec = codecs[type]; + if (!this.hasCreatedSourceBuffers()) { + return this.addSourceBuffer(type, codec); + } + if (this.canChangeType()) { + this.changeType(type, codec); + } + }); + } + /** + * Queue an update to append an ArrayBuffer. + * + * @param {MediaObject} object containing audioBytes and/or videoBytes + * @param {Function} done the function to call when done + * @see http://www.w3.org/TR/media-source/#widl-SourceBuffer-appendBuffer-void-ArrayBuffer-data + */ + + appendBuffer(options, doneFn) { + const { + segmentInfo, + type, + bytes + } = options; + this.processedAppend_ = true; + if (type === 'audio' && this.videoBuffer && !this.videoAppendQueued_) { + this.delayedAudioAppendQueue_.push([options, doneFn]); + this.logger_(`delayed audio append of ${bytes.length} until video append`); + return; + } // In the case of certain errors, for instance, QUOTA_EXCEEDED_ERR, updateend will + // not be fired. This means that the queue will be blocked until the next action + // taken by the segment-loader. Provide a mechanism for segment-loader to handle + // these errors by calling the doneFn with the specific error. + + const onError = doneFn; + pushQueue({ + type, + sourceUpdater: this, + action: actions.appendBuffer(bytes, segmentInfo || { + mediaIndex: -1 + }, onError), + doneFn, + name: 'appendBuffer' + }); + if (type === 'video') { + this.videoAppendQueued_ = true; + if (!this.delayedAudioAppendQueue_.length) { + return; + } + const queue = this.delayedAudioAppendQueue_.slice(); + this.logger_(`queuing delayed audio ${queue.length} appendBuffers`); + this.delayedAudioAppendQueue_.length = 0; + queue.forEach(que => { + this.appendBuffer.apply(this, que); + }); + } + } + /** + * Get the audio buffer's buffered timerange. + * + * @return {TimeRange} + * The audio buffer's buffered time range + */ + + audioBuffered() { + // no media source/source buffer or it isn't in the media sources + // source buffer list + if (!inSourceBuffers(this.mediaSource, this.audioBuffer)) { + return createTimeRanges(); + } + return this.audioBuffer.buffered ? this.audioBuffer.buffered : createTimeRanges(); + } + /** + * Get the video buffer's buffered timerange. + * + * @return {TimeRange} + * The video buffer's buffered time range + */ + + videoBuffered() { + // no media source/source buffer or it isn't in the media sources + // source buffer list + if (!inSourceBuffers(this.mediaSource, this.videoBuffer)) { + return createTimeRanges(); + } + return this.videoBuffer.buffered ? this.videoBuffer.buffered : createTimeRanges(); + } + /** + * Get a combined video/audio buffer's buffered timerange. + * + * @return {TimeRange} + * the combined time range + */ + + buffered() { + const video = inSourceBuffers(this.mediaSource, this.videoBuffer) ? this.videoBuffer : null; + const audio = inSourceBuffers(this.mediaSource, this.audioBuffer) ? this.audioBuffer : null; + if (audio && !video) { + return this.audioBuffered(); + } + if (video && !audio) { + return this.videoBuffered(); + } + return bufferIntersection(this.audioBuffered(), this.videoBuffered()); + } + /** + * Add a callback to the queue that will set duration on the mediaSource. + * + * @param {number} duration + * The duration to set + * + * @param {Function} [doneFn] + * function to run after duration has been set. + */ + + setDuration(duration, doneFn = noop) { + // In order to set the duration on the media source, it's necessary to wait for all + // source buffers to no longer be updating. "If the updating attribute equals true on + // any SourceBuffer in sourceBuffers, then throw an InvalidStateError exception and + // abort these steps." (source: https://www.w3.org/TR/media-source/#attributes). + pushQueue({ + type: 'mediaSource', + sourceUpdater: this, + action: actions.duration(duration), + name: 'duration', + doneFn + }); + } + /** + * Add a mediaSource endOfStream call to the queue + * + * @param {Error} [error] + * Call endOfStream with an error + * + * @param {Function} [doneFn] + * A function that should be called when the + * endOfStream call has finished. + */ + + endOfStream(error = null, doneFn = noop) { + if (typeof error !== 'string') { + error = undefined; + } // In order to set the duration on the media source, it's necessary to wait for all + // source buffers to no longer be updating. "If the updating attribute equals true on + // any SourceBuffer in sourceBuffers, then throw an InvalidStateError exception and + // abort these steps." (source: https://www.w3.org/TR/media-source/#attributes). + + pushQueue({ + type: 'mediaSource', + sourceUpdater: this, + action: actions.endOfStream(error), + name: 'endOfStream', + doneFn + }); + } + /** + * Queue an update to remove a time range from the buffer. + * + * @param {number} start where to start the removal + * @param {number} end where to end the removal + * @param {Function} [done=noop] optional callback to be executed when the remove + * operation is complete + * @see http://www.w3.org/TR/media-source/#widl-SourceBuffer-remove-void-double-start-unrestricted-double-end + */ + + removeAudio(start, end, done = noop) { + if (!this.audioBuffered().length || this.audioBuffered().end(0) === 0) { + done(); + return; + } + pushQueue({ + type: 'audio', + sourceUpdater: this, + action: actions.remove(start, end), + doneFn: done, + name: 'remove' + }); + } + /** + * Queue an update to remove a time range from the buffer. + * + * @param {number} start where to start the removal + * @param {number} end where to end the removal + * @param {Function} [done=noop] optional callback to be executed when the remove + * operation is complete + * @see http://www.w3.org/TR/media-source/#widl-SourceBuffer-remove-void-double-start-unrestricted-double-end + */ + + removeVideo(start, end, done = noop) { + if (!this.videoBuffered().length || this.videoBuffered().end(0) === 0) { + done(); + return; + } + pushQueue({ + type: 'video', + sourceUpdater: this, + action: actions.remove(start, end), + doneFn: done, + name: 'remove' + }); + } + /** + * Whether the underlying sourceBuffer is updating or not + * + * @return {boolean} the updating status of the SourceBuffer + */ + + updating() { + // the audio/video source buffer is updating + if (updating('audio', this) || updating('video', this)) { + return true; + } + return false; + } + /** + * Set/get the timestampoffset on the audio SourceBuffer + * + * @return {number} the timestamp offset + */ + + audioTimestampOffset(offset) { + if (typeof offset !== 'undefined' && this.audioBuffer && + // no point in updating if it's the same + this.audioTimestampOffset_ !== offset) { + pushQueue({ + type: 'audio', + sourceUpdater: this, + action: actions.timestampOffset(offset), + name: 'timestampOffset' + }); + this.audioTimestampOffset_ = offset; + } + return this.audioTimestampOffset_; + } + /** + * Set/get the timestampoffset on the video SourceBuffer + * + * @return {number} the timestamp offset + */ + + videoTimestampOffset(offset) { + if (typeof offset !== 'undefined' && this.videoBuffer && + // no point in updating if it's the same + this.videoTimestampOffset !== offset) { + pushQueue({ + type: 'video', + sourceUpdater: this, + action: actions.timestampOffset(offset), + name: 'timestampOffset' + }); + this.videoTimestampOffset_ = offset; + } + return this.videoTimestampOffset_; + } + /** + * Add a function to the queue that will be called + * when it is its turn to run in the audio queue. + * + * @param {Function} callback + * The callback to queue. + */ + + audioQueueCallback(callback) { + if (!this.audioBuffer) { + return; + } + pushQueue({ + type: 'audio', + sourceUpdater: this, + action: actions.callback(callback), + name: 'callback' + }); + } + /** + * Add a function to the queue that will be called + * when it is its turn to run in the video queue. + * + * @param {Function} callback + * The callback to queue. + */ + + videoQueueCallback(callback) { + if (!this.videoBuffer) { + return; + } + pushQueue({ + type: 'video', + sourceUpdater: this, + action: actions.callback(callback), + name: 'callback' + }); + } + /** + * dispose of the source updater and the underlying sourceBuffer + */ + + dispose() { + this.trigger('dispose'); + bufferTypes.forEach(type => { + this.abort(type); + if (this.canRemoveSourceBuffer()) { + this.removeSourceBuffer(type); + } else { + this[`${type}QueueCallback`](() => cleanupBuffer(type, this)); + } + }); + this.videoAppendQueued_ = false; + this.delayedAudioAppendQueue_.length = 0; + if (this.sourceopenListener_) { + this.mediaSource.removeEventListener('sourceopen', this.sourceopenListener_); + } + this.off(); + } +} +const uint8ToUtf8 = uintArray => decodeURIComponent(escape(String.fromCharCode.apply(null, uintArray))); +const bufferToHexString = buffer => { + const uInt8Buffer = new Uint8Array(buffer); + return Array.from(uInt8Buffer).map(byte => byte.toString(16).padStart(2, '0')).join(''); +}; + +/** + * @file vtt-segment-loader.js + */ +const VTT_LINE_TERMINATORS = new Uint8Array('\n\n'.split('').map(char => char.charCodeAt(0))); +class NoVttJsError extends Error { + constructor() { + super('Trying to parse received VTT cues, but there is no WebVTT. Make sure vtt.js is loaded.'); + } +} +/** + * An object that manages segment loading and appending. + * + * @class VTTSegmentLoader + * @param {Object} options required and optional options + * @extends videojs.EventTarget + */ + +class VTTSegmentLoader extends SegmentLoader { + constructor(settings, options = {}) { + super(settings, options); // SegmentLoader requires a MediaSource be specified or it will throw an error; + // however, VTTSegmentLoader has no need of a media source, so delete the reference + + this.mediaSource_ = null; + this.subtitlesTrack_ = null; + this.featuresNativeTextTracks_ = settings.featuresNativeTextTracks; + this.loadVttJs = settings.loadVttJs; // The VTT segment will have its own time mappings. Saving VTT segment timing info in + // the sync controller leads to improper behavior. + + this.shouldSaveSegmentTimingInfo_ = false; + } + createTransmuxer_() { + // don't need to transmux any subtitles + return null; + } + /** + * Indicates which time ranges are buffered + * + * @return {TimeRange} + * TimeRange object representing the current buffered ranges + */ + + buffered_() { + if (!this.subtitlesTrack_ || !this.subtitlesTrack_.cues || !this.subtitlesTrack_.cues.length) { + return createTimeRanges(); + } + const cues = this.subtitlesTrack_.cues; + const start = cues[0].startTime; + const end = cues[cues.length - 1].startTime; + return createTimeRanges([[start, end]]); + } + /** + * Gets and sets init segment for the provided map + * + * @param {Object} map + * The map object representing the init segment to get or set + * @param {boolean=} set + * If true, the init segment for the provided map should be saved + * @return {Object} + * map object for desired init segment + */ + + initSegmentForMap(map, set = false) { + if (!map) { + return null; + } + const id = initSegmentId(map); + let storedMap = this.initSegments_[id]; + if (set && !storedMap && map.bytes) { + // append WebVTT line terminators to the media initialization segment if it exists + // to follow the WebVTT spec (https://w3c.github.io/webvtt/#file-structure) that + // requires two or more WebVTT line terminators between the WebVTT header and the + // rest of the file + const combinedByteLength = VTT_LINE_TERMINATORS.byteLength + map.bytes.byteLength; + const combinedSegment = new Uint8Array(combinedByteLength); + combinedSegment.set(map.bytes); + combinedSegment.set(VTT_LINE_TERMINATORS, map.bytes.byteLength); + this.initSegments_[id] = storedMap = { + resolvedUri: map.resolvedUri, + byterange: map.byterange, + bytes: combinedSegment + }; + } + return storedMap || map; + } + /** + * Returns true if all configuration required for loading is present, otherwise false. + * + * @return {boolean} True if the all configuration is ready for loading + * @private + */ + + couldBeginLoading_() { + return this.playlist_ && this.subtitlesTrack_ && !this.paused(); + } + /** + * Once all the starting parameters have been specified, begin + * operation. This method should only be invoked from the INIT + * state. + * + * @private + */ + + init_() { + this.state = 'READY'; + this.resetEverything(); + return this.monitorBuffer_(); + } + /** + * Set a subtitle track on the segment loader to add subtitles to + * + * @param {TextTrack=} track + * The text track to add loaded subtitles to + * @return {TextTrack} + * Returns the subtitles track + */ + + track(track) { + if (typeof track === 'undefined') { + return this.subtitlesTrack_; + } + this.subtitlesTrack_ = track; // if we were unpaused but waiting for a sourceUpdater, start + // buffering now + + if (this.state === 'INIT' && this.couldBeginLoading_()) { + this.init_(); + } + return this.subtitlesTrack_; + } + /** + * Remove any data in the source buffer between start and end times + * + * @param {number} start - the start time of the region to remove from the buffer + * @param {number} end - the end time of the region to remove from the buffer + */ + + remove(start, end) { + removeCuesFromTrack(start, end, this.subtitlesTrack_); + } + /** + * fill the buffer with segements unless the sourceBuffers are + * currently updating + * + * Note: this function should only ever be called by monitorBuffer_ + * and never directly + * + * @private + */ + + fillBuffer_() { + // see if we need to begin loading immediately + const segmentInfo = this.chooseNextRequest_(); + if (!segmentInfo) { + return; + } + if (this.syncController_.timestampOffsetForTimeline(segmentInfo.timeline) === null) { + // We don't have the timestamp offset that we need to sync subtitles. + // Rerun on a timestamp offset or user interaction. + const checkTimestampOffset = () => { + this.state = 'READY'; + if (!this.paused()) { + // if not paused, queue a buffer check as soon as possible + this.monitorBuffer_(); + } + }; + this.syncController_.one('timestampoffset', checkTimestampOffset); + this.state = 'WAITING_ON_TIMELINE'; + return; + } + this.loadSegment_(segmentInfo); + } // never set a timestamp offset for vtt segments. + + timestampOffsetForSegment_() { + return null; + } + chooseNextRequest_() { + return this.skipEmptySegments_(super.chooseNextRequest_()); + } + /** + * Prevents the segment loader from requesting segments we know contain no subtitles + * by walking forward until we find the next segment that we don't know whether it is + * empty or not. + * + * @param {Object} segmentInfo + * a segment info object that describes the current segment + * @return {Object} + * a segment info object that describes the current segment + */ + + skipEmptySegments_(segmentInfo) { + while (segmentInfo && segmentInfo.segment.empty) { + // stop at the last possible segmentInfo + if (segmentInfo.mediaIndex + 1 >= segmentInfo.playlist.segments.length) { + segmentInfo = null; + break; + } + segmentInfo = this.generateSegmentInfo_({ + playlist: segmentInfo.playlist, + mediaIndex: segmentInfo.mediaIndex + 1, + startOfSegment: segmentInfo.startOfSegment + segmentInfo.duration, + isSyncRequest: segmentInfo.isSyncRequest + }); + } + return segmentInfo; + } + stopForError(error) { + this.error(error); + this.state = 'READY'; + this.pause(); + this.trigger('error'); + } + /** + * append a decrypted segement to the SourceBuffer through a SourceUpdater + * + * @private + */ + + segmentRequestFinished_(error, simpleSegment, result) { + if (!this.subtitlesTrack_) { + this.state = 'READY'; + return; + } + this.saveTransferStats_(simpleSegment.stats); // the request was aborted + + if (!this.pendingSegment_) { + this.state = 'READY'; + this.mediaRequestsAborted += 1; + return; + } + if (error) { + if (error.code === REQUEST_ERRORS.TIMEOUT) { + this.handleTimeout_(); + } + if (error.code === REQUEST_ERRORS.ABORTED) { + this.mediaRequestsAborted += 1; + } else { + this.mediaRequestsErrored += 1; + } + this.stopForError(error); + return; + } + const segmentInfo = this.pendingSegment_; // although the VTT segment loader bandwidth isn't really used, it's good to + // maintain functionality between segment loaders + + this.saveBandwidthRelatedStats_(segmentInfo.duration, simpleSegment.stats); // if this request included a segment key, save that data in the cache + + if (simpleSegment.key) { + this.segmentKey(simpleSegment.key, true); + } + this.state = 'APPENDING'; // used for tests + + this.trigger('appending'); + const segment = segmentInfo.segment; + if (segment.map) { + segment.map.bytes = simpleSegment.map.bytes; + } + segmentInfo.bytes = simpleSegment.bytes; // Make sure that vttjs has loaded, otherwise, load it and wait till it finished loading + + if (typeof window$1.WebVTT !== 'function' && typeof this.loadVttJs === 'function') { + this.state = 'WAITING_ON_VTTJS'; // should be fine to call multiple times + // script will be loaded once but multiple listeners will be added to the queue, which is expected. + + this.loadVttJs().then(() => this.segmentRequestFinished_(error, simpleSegment, result), () => this.stopForError({ + message: 'Error loading vtt.js', + metadata: { + errorType: videojs.Error.VttLoadError + } + })); + return; + } + segment.requested = true; + try { + this.parseVTTCues_(segmentInfo); + } catch (e) { + this.stopForError({ + message: e.message, + metadata: { + errorType: videojs.Error.VttCueParsingError + } + }); + return; + } + this.updateTimeMapping_(segmentInfo, this.syncController_.timelines[segmentInfo.timeline], this.playlist_); + if (segmentInfo.cues.length) { + segmentInfo.timingInfo = { + start: segmentInfo.cues[0].startTime, + end: segmentInfo.cues[segmentInfo.cues.length - 1].endTime + }; + } else { + segmentInfo.timingInfo = { + start: segmentInfo.startOfSegment, + end: segmentInfo.startOfSegment + segmentInfo.duration + }; + } + if (segmentInfo.isSyncRequest) { + this.trigger('syncinfoupdate'); + this.pendingSegment_ = null; + this.state = 'READY'; + return; + } + segmentInfo.byteLength = segmentInfo.bytes.byteLength; + this.mediaSecondsLoaded += segment.duration; // Create VTTCue instances for each cue in the new segment and add them to + // the subtitle track + + segmentInfo.cues.forEach(cue => { + this.subtitlesTrack_.addCue(this.featuresNativeTextTracks_ ? new window$1.VTTCue(cue.startTime, cue.endTime, cue.text) : cue); + }); // Remove any duplicate cues from the subtitle track. The WebVTT spec allows + // cues to have identical time-intervals, but if the text is also identical + // we can safely assume it is a duplicate that can be removed (ex. when a cue + // "overlaps" VTT segments) + + removeDuplicateCuesFromTrack(this.subtitlesTrack_); + this.handleAppendsDone_(); + } + handleData_() {// noop as we shouldn't be getting video/audio data captions + // that we do not support here. + } + updateTimingInfoEnd_() {// noop + } + /** + * Uses the WebVTT parser to parse the segment response + * + * @throws NoVttJsError + * + * @param {Object} segmentInfo + * a segment info object that describes the current segment + * @private + */ + + parseVTTCues_(segmentInfo) { + let decoder; + let decodeBytesToString = false; + if (typeof window$1.WebVTT !== 'function') { + // caller is responsible for exception handling. + throw new NoVttJsError(); + } + if (typeof window$1.TextDecoder === 'function') { + decoder = new window$1.TextDecoder('utf8'); + } else { + decoder = window$1.WebVTT.StringDecoder(); + decodeBytesToString = true; + } + const parser = new window$1.WebVTT.Parser(window$1, window$1.vttjs, decoder); + segmentInfo.cues = []; + segmentInfo.timestampmap = { + MPEGTS: 0, + LOCAL: 0 + }; + parser.oncue = segmentInfo.cues.push.bind(segmentInfo.cues); + parser.ontimestampmap = map => { + segmentInfo.timestampmap = map; + }; + parser.onparsingerror = error => { + videojs.log.warn('Error encountered when parsing cues: ' + error.message); + }; + if (segmentInfo.segment.map) { + let mapData = segmentInfo.segment.map.bytes; + if (decodeBytesToString) { + mapData = uint8ToUtf8(mapData); + } + parser.parse(mapData); + } + let segmentData = segmentInfo.bytes; + if (decodeBytesToString) { + segmentData = uint8ToUtf8(segmentData); + } + parser.parse(segmentData); + parser.flush(); + } + /** + * Updates the start and end times of any cues parsed by the WebVTT parser using + * the information parsed from the X-TIMESTAMP-MAP header and a TS to media time mapping + * from the SyncController + * + * @param {Object} segmentInfo + * a segment info object that describes the current segment + * @param {Object} mappingObj + * object containing a mapping from TS to media time + * @param {Object} playlist + * the playlist object containing the segment + * @private + */ + + updateTimeMapping_(segmentInfo, mappingObj, playlist) { + const segment = segmentInfo.segment; + if (!mappingObj) { + // If the sync controller does not have a mapping of TS to Media Time for the + // timeline, then we don't have enough information to update the cue + // start/end times + return; + } + if (!segmentInfo.cues.length) { + // If there are no cues, we also do not have enough information to figure out + // segment timing. Mark that the segment contains no cues so we don't re-request + // an empty segment. + segment.empty = true; + return; + } + const { + MPEGTS, + LOCAL + } = segmentInfo.timestampmap; + /** + * From the spec: + * The MPEGTS media timestamp MUST use a 90KHz timescale, + * even when non-WebVTT Media Segments use a different timescale. + */ + + const mpegTsInSeconds = MPEGTS / ONE_SECOND_IN_TS; + const diff = mpegTsInSeconds - LOCAL + mappingObj.mapping; + segmentInfo.cues.forEach(cue => { + const duration = cue.endTime - cue.startTime; + const startTime = MPEGTS === 0 ? cue.startTime + diff : this.handleRollover_(cue.startTime + diff, mappingObj.time); + cue.startTime = Math.max(startTime, 0); + cue.endTime = Math.max(startTime + duration, 0); + }); + if (!playlist.syncInfo) { + const firstStart = segmentInfo.cues[0].startTime; + const lastStart = segmentInfo.cues[segmentInfo.cues.length - 1].startTime; + playlist.syncInfo = { + mediaSequence: playlist.mediaSequence + segmentInfo.mediaIndex, + time: Math.min(firstStart, lastStart - segment.duration) + }; + } + } + /** + * MPEG-TS PES timestamps are limited to 2^33. + * Once they reach 2^33, they roll over to 0. + * mux.js handles PES timestamp rollover for the following scenarios: + * [forward rollover(right)] -> + * PES timestamps monotonically increase, and once they reach 2^33, they roll over to 0 + * [backward rollover(left)] --> + * we seek back to position before rollover. + * + * According to the HLS SPEC: + * When synchronizing WebVTT with PES timestamps, clients SHOULD account + * for cases where the 33-bit PES timestamps have wrapped and the WebVTT + * cue times have not. When the PES timestamp wraps, the WebVTT Segment + * SHOULD have a X-TIMESTAMP-MAP header that maps the current WebVTT + * time to the new (low valued) PES timestamp. + * + * So we want to handle rollover here and align VTT Cue start/end time to the player's time. + */ + + handleRollover_(value, reference) { + if (reference === null) { + return value; + } + let valueIn90khz = value * ONE_SECOND_IN_TS; + const referenceIn90khz = reference * ONE_SECOND_IN_TS; + let offset; + if (referenceIn90khz < valueIn90khz) { + // - 2^33 + offset = -8589934592; + } else { + // + 2^33 + offset = 8589934592; + } // distance(value - reference) > 2^32 + + while (Math.abs(valueIn90khz - referenceIn90khz) > 4294967296) { + valueIn90khz += offset; + } + return valueIn90khz / ONE_SECOND_IN_TS; + } +} + +/** + * @file ad-cue-tags.js + */ +/** + * Searches for an ad cue that overlaps with the given mediaTime + * + * @param {Object} track + * the track to find the cue for + * + * @param {number} mediaTime + * the time to find the cue at + * + * @return {Object|null} + * the found cue or null + */ + +const findAdCue = function (track, mediaTime) { + const cues = track.cues; + for (let i = 0; i < cues.length; i++) { + const cue = cues[i]; + if (mediaTime >= cue.adStartTime && mediaTime <= cue.adEndTime) { + return cue; + } + } + return null; +}; +const updateAdCues = function (media, track, offset = 0) { + if (!media.segments) { + return; + } + let mediaTime = offset; + let cue; + for (let i = 0; i < media.segments.length; i++) { + const segment = media.segments[i]; + if (!cue) { + // Since the cues will span for at least the segment duration, adding a fudge + // factor of half segment duration will prevent duplicate cues from being + // created when timing info is not exact (e.g. cue start time initialized + // at 10.006677, but next call mediaTime is 10.003332 ) + cue = findAdCue(track, mediaTime + segment.duration / 2); + } + if (cue) { + if ('cueIn' in segment) { + // Found a CUE-IN so end the cue + cue.endTime = mediaTime; + cue.adEndTime = mediaTime; + mediaTime += segment.duration; + cue = null; + continue; + } + if (mediaTime < cue.endTime) { + // Already processed this mediaTime for this cue + mediaTime += segment.duration; + continue; + } // otherwise extend cue until a CUE-IN is found + + cue.endTime += segment.duration; + } else { + if ('cueOut' in segment) { + cue = new window$1.VTTCue(mediaTime, mediaTime + segment.duration, segment.cueOut); + cue.adStartTime = mediaTime; // Assumes tag format to be + // #EXT-X-CUE-OUT:30 + + cue.adEndTime = mediaTime + parseFloat(segment.cueOut); + track.addCue(cue); + } + if ('cueOutCont' in segment) { + // Entered into the middle of an ad cue + // Assumes tag formate to be + // #EXT-X-CUE-OUT-CONT:10/30 + const [adOffset, adTotal] = segment.cueOutCont.split('/').map(parseFloat); + cue = new window$1.VTTCue(mediaTime, mediaTime + segment.duration, ''); + cue.adStartTime = mediaTime - adOffset; + cue.adEndTime = cue.adStartTime + adTotal; + track.addCue(cue); + } + } + mediaTime += segment.duration; + } +}; +class SyncInfo { + /** + * @param {number} start - media sequence start + * @param {number} end - media sequence end + * @param {number} segmentIndex - index for associated segment + * @param {number|null} [partIndex] - index for associated part + * @param {boolean} [appended] - appended indicator + * + */ + constructor({ + start, + end, + segmentIndex, + partIndex = null, + appended = false + }) { + this.start_ = start; + this.end_ = end; + this.segmentIndex_ = segmentIndex; + this.partIndex_ = partIndex; + this.appended_ = appended; + } + isInRange(targetTime) { + return targetTime >= this.start && targetTime < this.end; + } + markAppended() { + this.appended_ = true; + } + resetAppendedStatus() { + this.appended_ = false; + } + get isAppended() { + return this.appended_; + } + get start() { + return this.start_; + } + get end() { + return this.end_; + } + get segmentIndex() { + return this.segmentIndex_; + } + get partIndex() { + return this.partIndex_; + } +} +class SyncInfoData { + /** + * + * @param {SyncInfo} segmentSyncInfo - sync info for a given segment + * @param {Array<SyncInfo>} [partsSyncInfo] - sync infos for a list of parts for a given segment + */ + constructor(segmentSyncInfo, partsSyncInfo = []) { + this.segmentSyncInfo_ = segmentSyncInfo; + this.partsSyncInfo_ = partsSyncInfo; + } + get segmentSyncInfo() { + return this.segmentSyncInfo_; + } + get partsSyncInfo() { + return this.partsSyncInfo_; + } + get hasPartsSyncInfo() { + return this.partsSyncInfo_.length > 0; + } + resetAppendStatus() { + this.segmentSyncInfo_.resetAppendedStatus(); + this.partsSyncInfo_.forEach(partSyncInfo => partSyncInfo.resetAppendedStatus()); + } +} +class MediaSequenceSync { + constructor() { + /** + * @type {Map<number, SyncInfoData>} + * @private + */ + this.storage_ = new Map(); + this.diagnostics_ = ''; + this.isReliable_ = false; + this.start_ = -Infinity; + this.end_ = Infinity; + } + get start() { + return this.start_; + } + get end() { + return this.end_; + } + get diagnostics() { + return this.diagnostics_; + } + get isReliable() { + return this.isReliable_; + } + resetAppendedStatus() { + this.storage_.forEach(syncInfoData => syncInfoData.resetAppendStatus()); + } + /** + * update sync storage + * + * @param {Object} playlist + * @param {number} currentTime + * + * @return {void} + */ + + update(playlist, currentTime) { + const { + mediaSequence, + segments + } = playlist; + this.isReliable_ = this.isReliablePlaylist_(mediaSequence, segments); + if (!this.isReliable_) { + return; + } + return this.updateStorage_(segments, mediaSequence, this.calculateBaseTime_(mediaSequence, currentTime)); + } + /** + * @param {number} targetTime + * @return {SyncInfo|null} + */ + + getSyncInfoForTime(targetTime) { + for (const { + segmentSyncInfo, + partsSyncInfo + } of this.storage_.values()) { + // Normal segment flow: + if (!partsSyncInfo.length) { + if (segmentSyncInfo.isInRange(targetTime)) { + return segmentSyncInfo; + } + } else { + // Low latency flow: + for (const partSyncInfo of partsSyncInfo) { + if (partSyncInfo.isInRange(targetTime)) { + return partSyncInfo; + } + } + } + } + return null; + } + updateStorage_(segments, startingMediaSequence, startingTime) { + const newStorage = new Map(); + let newDiagnostics = '\n'; + let currentStart = startingTime; + let currentMediaSequence = startingMediaSequence; + this.start_ = currentStart; + segments.forEach((segment, segmentIndex) => { + const prevSyncInfoData = this.storage_.get(currentMediaSequence); + const segmentStart = currentStart; + const segmentEnd = segmentStart + segment.duration; + const segmentIsAppended = Boolean(prevSyncInfoData && prevSyncInfoData.segmentSyncInfo && prevSyncInfoData.segmentSyncInfo.isAppended); + const segmentSyncInfo = new SyncInfo({ + start: segmentStart, + end: segmentEnd, + appended: segmentIsAppended, + segmentIndex + }); + segment.syncInfo = segmentSyncInfo; + let currentPartStart = currentStart; + const partsSyncInfo = (segment.parts || []).map((part, partIndex) => { + const partStart = currentPartStart; + const partEnd = currentPartStart + part.duration; + const partIsAppended = Boolean(prevSyncInfoData && prevSyncInfoData.partsSyncInfo && prevSyncInfoData.partsSyncInfo[partIndex] && prevSyncInfoData.partsSyncInfo[partIndex].isAppended); + const partSyncInfo = new SyncInfo({ + start: partStart, + end: partEnd, + appended: partIsAppended, + segmentIndex, + partIndex + }); + currentPartStart = partEnd; + newDiagnostics += `Media Sequence: ${currentMediaSequence}.${partIndex} | Range: ${partStart} --> ${partEnd} | Appended: ${partIsAppended}\n`; + part.syncInfo = partSyncInfo; + return partSyncInfo; + }); + newStorage.set(currentMediaSequence, new SyncInfoData(segmentSyncInfo, partsSyncInfo)); + newDiagnostics += `${compactSegmentUrlDescription(segment.resolvedUri)} | Media Sequence: ${currentMediaSequence} | Range: ${segmentStart} --> ${segmentEnd} | Appended: ${segmentIsAppended}\n`; + currentMediaSequence++; + currentStart = segmentEnd; + }); + this.end_ = currentStart; + this.storage_ = newStorage; + this.diagnostics_ = newDiagnostics; + } + calculateBaseTime_(mediaSequence, fallback) { + if (!this.storage_.size) { + // Initial setup flow. + return 0; + } + if (this.storage_.has(mediaSequence)) { + // Normal flow. + return this.storage_.get(mediaSequence).segmentSyncInfo.start; + } // Fallback flow. + // There is a gap between last recorded playlist and a new one received. + + return fallback; + } + isReliablePlaylist_(mediaSequence, segments) { + return mediaSequence !== undefined && mediaSequence !== null && Array.isArray(segments) && segments.length; + } +} + +/** + * @file sync-controller.js + */ +// synchronize expired playlist segments. +// the max media sequence diff is 48 hours of live stream +// content with two second segments. Anything larger than that +// will likely be invalid. + +const MAX_MEDIA_SEQUENCE_DIFF_FOR_SYNC = 86400; +const syncPointStrategies = [ +// Stategy "VOD": Handle the VOD-case where the sync-point is *always* +// the equivalence display-time 0 === segment-index 0 +{ + name: 'VOD', + run: (syncController, playlist, duration, currentTimeline, currentTime) => { + if (duration !== Infinity) { + const syncPoint = { + time: 0, + segmentIndex: 0, + partIndex: null + }; + return syncPoint; + } + return null; + } +}, { + name: 'MediaSequence', + /** + * run media sequence strategy + * + * @param {SyncController} syncController + * @param {Object} playlist + * @param {number} duration + * @param {number} currentTimeline + * @param {number} currentTime + * @param {string} type + */ + run: (syncController, playlist, duration, currentTimeline, currentTime, type) => { + const mediaSequenceSync = syncController.getMediaSequenceSync(type); + if (!mediaSequenceSync) { + return null; + } + if (!mediaSequenceSync.isReliable) { + return null; + } + const syncInfo = mediaSequenceSync.getSyncInfoForTime(currentTime); + if (!syncInfo) { + return null; + } + return { + time: syncInfo.start, + partIndex: syncInfo.partIndex, + segmentIndex: syncInfo.segmentIndex + }; + } +}, +// Stategy "ProgramDateTime": We have a program-date-time tag in this playlist +{ + name: 'ProgramDateTime', + run: (syncController, playlist, duration, currentTimeline, currentTime) => { + if (!Object.keys(syncController.timelineToDatetimeMappings).length) { + return null; + } + let syncPoint = null; + let lastDistance = null; + const partsAndSegments = getPartsAndSegments(playlist); + currentTime = currentTime || 0; + for (let i = 0; i < partsAndSegments.length; i++) { + // start from the end and loop backwards for live + // or start from the front and loop forwards for non-live + const index = playlist.endList || currentTime === 0 ? i : partsAndSegments.length - (i + 1); + const partAndSegment = partsAndSegments[index]; + const segment = partAndSegment.segment; + const datetimeMapping = syncController.timelineToDatetimeMappings[segment.timeline]; + if (!datetimeMapping || !segment.dateTimeObject) { + continue; + } + const segmentTime = segment.dateTimeObject.getTime() / 1000; + let start = segmentTime + datetimeMapping; // take part duration into account. + + if (segment.parts && typeof partAndSegment.partIndex === 'number') { + for (let z = 0; z < partAndSegment.partIndex; z++) { + start += segment.parts[z].duration; + } + } + const distance = Math.abs(currentTime - start); // Once the distance begins to increase, or if distance is 0, we have passed + // currentTime and can stop looking for better candidates + + if (lastDistance !== null && (distance === 0 || lastDistance < distance)) { + break; + } + lastDistance = distance; + syncPoint = { + time: start, + segmentIndex: partAndSegment.segmentIndex, + partIndex: partAndSegment.partIndex + }; + } + return syncPoint; + } +}, +// Stategy "Segment": We have a known time mapping for a timeline and a +// segment in the current timeline with timing data +{ + name: 'Segment', + run: (syncController, playlist, duration, currentTimeline, currentTime) => { + let syncPoint = null; + let lastDistance = null; + currentTime = currentTime || 0; + const partsAndSegments = getPartsAndSegments(playlist); + for (let i = 0; i < partsAndSegments.length; i++) { + // start from the end and loop backwards for live + // or start from the front and loop forwards for non-live + const index = playlist.endList || currentTime === 0 ? i : partsAndSegments.length - (i + 1); + const partAndSegment = partsAndSegments[index]; + const segment = partAndSegment.segment; + const start = partAndSegment.part && partAndSegment.part.start || segment && segment.start; + if (segment.timeline === currentTimeline && typeof start !== 'undefined') { + const distance = Math.abs(currentTime - start); // Once the distance begins to increase, we have passed + // currentTime and can stop looking for better candidates + + if (lastDistance !== null && lastDistance < distance) { + break; + } + if (!syncPoint || lastDistance === null || lastDistance >= distance) { + lastDistance = distance; + syncPoint = { + time: start, + segmentIndex: partAndSegment.segmentIndex, + partIndex: partAndSegment.partIndex + }; + } + } + } + return syncPoint; + } +}, +// Stategy "Discontinuity": We have a discontinuity with a known +// display-time +{ + name: 'Discontinuity', + run: (syncController, playlist, duration, currentTimeline, currentTime) => { + let syncPoint = null; + currentTime = currentTime || 0; + if (playlist.discontinuityStarts && playlist.discontinuityStarts.length) { + let lastDistance = null; + for (let i = 0; i < playlist.discontinuityStarts.length; i++) { + const segmentIndex = playlist.discontinuityStarts[i]; + const discontinuity = playlist.discontinuitySequence + i + 1; + const discontinuitySync = syncController.discontinuities[discontinuity]; + if (discontinuitySync) { + const distance = Math.abs(currentTime - discontinuitySync.time); // Once the distance begins to increase, we have passed + // currentTime and can stop looking for better candidates + + if (lastDistance !== null && lastDistance < distance) { + break; + } + if (!syncPoint || lastDistance === null || lastDistance >= distance) { + lastDistance = distance; + syncPoint = { + time: discontinuitySync.time, + segmentIndex, + partIndex: null + }; + } + } + } + } + return syncPoint; + } +}, +// Stategy "Playlist": We have a playlist with a known mapping of +// segment index to display time +{ + name: 'Playlist', + run: (syncController, playlist, duration, currentTimeline, currentTime) => { + if (playlist.syncInfo) { + const syncPoint = { + time: playlist.syncInfo.time, + segmentIndex: playlist.syncInfo.mediaSequence - playlist.mediaSequence, + partIndex: null + }; + return syncPoint; + } + return null; + } +}]; +class SyncController extends videojs.EventTarget { + constructor(options = {}) { + super(); // ...for synching across variants + + this.timelines = []; + this.discontinuities = []; + this.timelineToDatetimeMappings = {}; // TODO: this map should be only available for HLS. Since only HLS has MediaSequence. + // For some reason this map helps with syncing between quality switch for MPEG-DASH as well. + // Moreover if we disable this map for MPEG-DASH - quality switch will be broken. + // MPEG-DASH should have its own separate sync strategy + + this.mediaSequenceStorage_ = { + main: new MediaSequenceSync(), + audio: new MediaSequenceSync(), + vtt: new MediaSequenceSync() + }; + this.logger_ = logger('SyncController'); + } + /** + * + * @param {string} loaderType + * @return {MediaSequenceSync|null} + */ + + getMediaSequenceSync(loaderType) { + return this.mediaSequenceStorage_[loaderType] || null; + } + /** + * Find a sync-point for the playlist specified + * + * A sync-point is defined as a known mapping from display-time to + * a segment-index in the current playlist. + * + * @param {Playlist} playlist + * The playlist that needs a sync-point + * @param {number} duration + * Duration of the MediaSource (Infinite if playing a live source) + * @param {number} currentTimeline + * The last timeline from which a segment was loaded + * @param {number} currentTime + * Current player's time + * @param {string} type + * Segment loader type + * @return {Object} + * A sync-point object + */ + + getSyncPoint(playlist, duration, currentTimeline, currentTime, type) { + // Always use VOD sync point for VOD + if (duration !== Infinity) { + const vodSyncPointStrategy = syncPointStrategies.find(({ + name + }) => name === 'VOD'); + return vodSyncPointStrategy.run(this, playlist, duration); + } + const syncPoints = this.runStrategies_(playlist, duration, currentTimeline, currentTime, type); + if (!syncPoints.length) { + // Signal that we need to attempt to get a sync-point manually + // by fetching a segment in the playlist and constructing + // a sync-point from that information + return null; + } // If we have exact match just return it instead of finding the nearest distance + + for (const syncPointInfo of syncPoints) { + const { + syncPoint, + strategy + } = syncPointInfo; + const { + segmentIndex, + time + } = syncPoint; + if (segmentIndex < 0) { + continue; + } + const selectedSegment = playlist.segments[segmentIndex]; + const start = time; + const end = start + selectedSegment.duration; + this.logger_(`Strategy: ${strategy}. Current time: ${currentTime}. selected segment: ${segmentIndex}. Time: [${start} -> ${end}]}`); + if (currentTime >= start && currentTime < end) { + this.logger_('Found sync point with exact match: ', syncPoint); + return syncPoint; + } + } // Now find the sync-point that is closest to the currentTime because + // that should result in the most accurate guess about which segment + // to fetch + + return this.selectSyncPoint_(syncPoints, { + key: 'time', + value: currentTime + }); + } + /** + * Calculate the amount of time that has expired off the playlist during playback + * + * @param {Playlist} playlist + * Playlist object to calculate expired from + * @param {number} duration + * Duration of the MediaSource (Infinity if playling a live source) + * @return {number|null} + * The amount of time that has expired off the playlist during playback. Null + * if no sync-points for the playlist can be found. + */ + + getExpiredTime(playlist, duration) { + if (!playlist || !playlist.segments) { + return null; + } + const syncPoints = this.runStrategies_(playlist, duration, playlist.discontinuitySequence, 0); // Without sync-points, there is not enough information to determine the expired time + + if (!syncPoints.length) { + return null; + } + const syncPoint = this.selectSyncPoint_(syncPoints, { + key: 'segmentIndex', + value: 0 + }); // If the sync-point is beyond the start of the playlist, we want to subtract the + // duration from index 0 to syncPoint.segmentIndex instead of adding. + + if (syncPoint.segmentIndex > 0) { + syncPoint.time *= -1; + } + return Math.abs(syncPoint.time + sumDurations({ + defaultDuration: playlist.targetDuration, + durationList: playlist.segments, + startIndex: syncPoint.segmentIndex, + endIndex: 0 + })); + } + /** + * Runs each sync-point strategy and returns a list of sync-points returned by the + * strategies + * + * @private + * @param {Playlist} playlist + * The playlist that needs a sync-point + * @param {number} duration + * Duration of the MediaSource (Infinity if playing a live source) + * @param {number} currentTimeline + * The last timeline from which a segment was loaded + * @param {number} currentTime + * Current player's time + * @param {string} type + * Segment loader type + * @return {Array} + * A list of sync-point objects + */ + + runStrategies_(playlist, duration, currentTimeline, currentTime, type) { + const syncPoints = []; // Try to find a sync-point in by utilizing various strategies... + + for (let i = 0; i < syncPointStrategies.length; i++) { + const strategy = syncPointStrategies[i]; + const syncPoint = strategy.run(this, playlist, duration, currentTimeline, currentTime, type); + if (syncPoint) { + syncPoint.strategy = strategy.name; + syncPoints.push({ + strategy: strategy.name, + syncPoint + }); + } + } + return syncPoints; + } + /** + * Selects the sync-point nearest the specified target + * + * @private + * @param {Array} syncPoints + * List of sync-points to select from + * @param {Object} target + * Object specifying the property and value we are targeting + * @param {string} target.key + * Specifies the property to target. Must be either 'time' or 'segmentIndex' + * @param {number} target.value + * The value to target for the specified key. + * @return {Object} + * The sync-point nearest the target + */ + + selectSyncPoint_(syncPoints, target) { + let bestSyncPoint = syncPoints[0].syncPoint; + let bestDistance = Math.abs(syncPoints[0].syncPoint[target.key] - target.value); + let bestStrategy = syncPoints[0].strategy; + for (let i = 1; i < syncPoints.length; i++) { + const newDistance = Math.abs(syncPoints[i].syncPoint[target.key] - target.value); + if (newDistance < bestDistance) { + bestDistance = newDistance; + bestSyncPoint = syncPoints[i].syncPoint; + bestStrategy = syncPoints[i].strategy; + } + } + this.logger_(`syncPoint for [${target.key}: ${target.value}] chosen with strategy` + ` [${bestStrategy}]: [time:${bestSyncPoint.time},` + ` segmentIndex:${bestSyncPoint.segmentIndex}` + (typeof bestSyncPoint.partIndex === 'number' ? `,partIndex:${bestSyncPoint.partIndex}` : '') + ']'); + return bestSyncPoint; + } + /** + * Save any meta-data present on the segments when segments leave + * the live window to the playlist to allow for synchronization at the + * playlist level later. + * + * @param {Playlist} oldPlaylist - The previous active playlist + * @param {Playlist} newPlaylist - The updated and most current playlist + */ + + saveExpiredSegmentInfo(oldPlaylist, newPlaylist) { + const mediaSequenceDiff = newPlaylist.mediaSequence - oldPlaylist.mediaSequence; // Ignore large media sequence gaps + + if (mediaSequenceDiff > MAX_MEDIA_SEQUENCE_DIFF_FOR_SYNC) { + videojs.log.warn(`Not saving expired segment info. Media sequence gap ${mediaSequenceDiff} is too large.`); + return; + } // When a segment expires from the playlist and it has a start time + // save that information as a possible sync-point reference in future + + for (let i = mediaSequenceDiff - 1; i >= 0; i--) { + const lastRemovedSegment = oldPlaylist.segments[i]; + if (lastRemovedSegment && typeof lastRemovedSegment.start !== 'undefined') { + newPlaylist.syncInfo = { + mediaSequence: oldPlaylist.mediaSequence + i, + time: lastRemovedSegment.start + }; + this.logger_(`playlist refresh sync: [time:${newPlaylist.syncInfo.time},` + ` mediaSequence: ${newPlaylist.syncInfo.mediaSequence}]`); + this.trigger('syncinfoupdate'); + break; + } + } + } + /** + * Save the mapping from playlist's ProgramDateTime to display. This should only happen + * before segments start to load. + * + * @param {Playlist} playlist - The currently active playlist + */ + + setDateTimeMappingForStart(playlist) { + // It's possible for the playlist to be updated before playback starts, meaning time + // zero is not yet set. If, during these playlist refreshes, a discontinuity is + // crossed, then the old time zero mapping (for the prior timeline) would be retained + // unless the mappings are cleared. + this.timelineToDatetimeMappings = {}; + if (playlist.segments && playlist.segments.length && playlist.segments[0].dateTimeObject) { + const firstSegment = playlist.segments[0]; + const playlistTimestamp = firstSegment.dateTimeObject.getTime() / 1000; + this.timelineToDatetimeMappings[firstSegment.timeline] = -playlistTimestamp; + } + } + /** + * Calculates and saves timeline mappings, playlist sync info, and segment timing values + * based on the latest timing information. + * + * @param {Object} options + * Options object + * @param {SegmentInfo} options.segmentInfo + * The current active request information + * @param {boolean} options.shouldSaveTimelineMapping + * If there's a timeline change, determines if the timeline mapping should be + * saved for timeline mapping and program date time mappings. + */ + + saveSegmentTimingInfo({ + segmentInfo, + shouldSaveTimelineMapping + }) { + const didCalculateSegmentTimeMapping = this.calculateSegmentTimeMapping_(segmentInfo, segmentInfo.timingInfo, shouldSaveTimelineMapping); + const segment = segmentInfo.segment; + if (didCalculateSegmentTimeMapping) { + this.saveDiscontinuitySyncInfo_(segmentInfo); // If the playlist does not have sync information yet, record that information + // now with segment timing information + + if (!segmentInfo.playlist.syncInfo) { + segmentInfo.playlist.syncInfo = { + mediaSequence: segmentInfo.playlist.mediaSequence + segmentInfo.mediaIndex, + time: segment.start + }; + } + } + const dateTime = segment.dateTimeObject; + if (segment.discontinuity && shouldSaveTimelineMapping && dateTime) { + this.timelineToDatetimeMappings[segment.timeline] = -(dateTime.getTime() / 1000); + } + } + timestampOffsetForTimeline(timeline) { + if (typeof this.timelines[timeline] === 'undefined') { + return null; + } + return this.timelines[timeline].time; + } + mappingForTimeline(timeline) { + if (typeof this.timelines[timeline] === 'undefined') { + return null; + } + return this.timelines[timeline].mapping; + } + /** + * Use the "media time" for a segment to generate a mapping to "display time" and + * save that display time to the segment. + * + * @private + * @param {SegmentInfo} segmentInfo + * The current active request information + * @param {Object} timingInfo + * The start and end time of the current segment in "media time" + * @param {boolean} shouldSaveTimelineMapping + * If there's a timeline change, determines if the timeline mapping should be + * saved in timelines. + * @return {boolean} + * Returns false if segment time mapping could not be calculated + */ + + calculateSegmentTimeMapping_(segmentInfo, timingInfo, shouldSaveTimelineMapping) { + // TODO: remove side effects + const segment = segmentInfo.segment; + const part = segmentInfo.part; + let mappingObj = this.timelines[segmentInfo.timeline]; + let start; + let end; + if (typeof segmentInfo.timestampOffset === 'number') { + mappingObj = { + time: segmentInfo.startOfSegment, + mapping: segmentInfo.startOfSegment - timingInfo.start + }; + if (shouldSaveTimelineMapping) { + this.timelines[segmentInfo.timeline] = mappingObj; + this.trigger('timestampoffset'); + this.logger_(`time mapping for timeline ${segmentInfo.timeline}: ` + `[time: ${mappingObj.time}] [mapping: ${mappingObj.mapping}]`); + } + start = segmentInfo.startOfSegment; + end = timingInfo.end + mappingObj.mapping; + } else if (mappingObj) { + start = timingInfo.start + mappingObj.mapping; + end = timingInfo.end + mappingObj.mapping; + } else { + return false; + } + if (part) { + part.start = start; + part.end = end; + } // If we don't have a segment start yet or the start value we got + // is less than our current segment.start value, save a new start value. + // We have to do this because parts will have segment timing info saved + // multiple times and we want segment start to be the earliest part start + // value for that segment. + + if (!segment.start || start < segment.start) { + segment.start = start; + } + segment.end = end; + return true; + } + /** + * Each time we have discontinuity in the playlist, attempt to calculate the location + * in display of the start of the discontinuity and save that. We also save an accuracy + * value so that we save values with the most accuracy (closest to 0.) + * + * @private + * @param {SegmentInfo} segmentInfo - The current active request information + */ + + saveDiscontinuitySyncInfo_(segmentInfo) { + const playlist = segmentInfo.playlist; + const segment = segmentInfo.segment; // If the current segment is a discontinuity then we know exactly where + // the start of the range and it's accuracy is 0 (greater accuracy values + // mean more approximation) + + if (segment.discontinuity) { + this.discontinuities[segment.timeline] = { + time: segment.start, + accuracy: 0 + }; + } else if (playlist.discontinuityStarts && playlist.discontinuityStarts.length) { + // Search for future discontinuities that we can provide better timing + // information for and save that information for sync purposes + for (let i = 0; i < playlist.discontinuityStarts.length; i++) { + const segmentIndex = playlist.discontinuityStarts[i]; + const discontinuity = playlist.discontinuitySequence + i + 1; + const mediaIndexDiff = segmentIndex - segmentInfo.mediaIndex; + const accuracy = Math.abs(mediaIndexDiff); + if (!this.discontinuities[discontinuity] || this.discontinuities[discontinuity].accuracy > accuracy) { + let time; + if (mediaIndexDiff < 0) { + time = segment.start - sumDurations({ + defaultDuration: playlist.targetDuration, + durationList: playlist.segments, + startIndex: segmentInfo.mediaIndex, + endIndex: segmentIndex + }); + } else { + time = segment.end + sumDurations({ + defaultDuration: playlist.targetDuration, + durationList: playlist.segments, + startIndex: segmentInfo.mediaIndex + 1, + endIndex: segmentIndex + }); + } + this.discontinuities[discontinuity] = { + time, + accuracy + }; + } + } + } + } + dispose() { + this.trigger('dispose'); + this.off(); + } +} + +/** + * The TimelineChangeController acts as a source for segment loaders to listen for and + * keep track of latest and pending timeline changes. This is useful to ensure proper + * sync, as each loader may need to make a consideration for what timeline the other + * loader is on before making changes which could impact the other loader's media. + * + * @class TimelineChangeController + * @extends videojs.EventTarget + */ + +class TimelineChangeController extends videojs.EventTarget { + constructor() { + super(); + this.pendingTimelineChanges_ = {}; + this.lastTimelineChanges_ = {}; + } + clearPendingTimelineChange(type) { + this.pendingTimelineChanges_[type] = null; + this.trigger('pendingtimelinechange'); + } + pendingTimelineChange({ + type, + from, + to + }) { + if (typeof from === 'number' && typeof to === 'number') { + this.pendingTimelineChanges_[type] = { + type, + from, + to + }; + this.trigger('pendingtimelinechange'); + } + return this.pendingTimelineChanges_[type]; + } + lastTimelineChange({ + type, + from, + to + }) { + if (typeof from === 'number' && typeof to === 'number') { + this.lastTimelineChanges_[type] = { + type, + from, + to + }; + delete this.pendingTimelineChanges_[type]; + this.trigger('timelinechange'); + } + return this.lastTimelineChanges_[type]; + } + dispose() { + this.trigger('dispose'); + this.pendingTimelineChanges_ = {}; + this.lastTimelineChanges_ = {}; + this.off(); + } +} + +/* rollup-plugin-worker-factory start for worker!/home/runner/work/http-streaming/http-streaming/src/decrypter-worker.js */ +const workerCode = transform(getWorkerString(function () { + /** + * @file stream.js + */ + + /** + * A lightweight readable stream implemention that handles event dispatching. + * + * @class Stream + */ + + var Stream = /*#__PURE__*/function () { + function Stream() { + this.listeners = {}; + } + /** + * Add a listener for a specified event type. + * + * @param {string} type the event name + * @param {Function} listener the callback to be invoked when an event of + * the specified type occurs + */ + + var _proto = Stream.prototype; + _proto.on = function on(type, listener) { + if (!this.listeners[type]) { + this.listeners[type] = []; + } + this.listeners[type].push(listener); + } + /** + * Remove a listener for a specified event type. + * + * @param {string} type the event name + * @param {Function} listener a function previously registered for this + * type of event through `on` + * @return {boolean} if we could turn it off or not + */; + + _proto.off = function off(type, listener) { + if (!this.listeners[type]) { + return false; + } + var index = this.listeners[type].indexOf(listener); // TODO: which is better? + // In Video.js we slice listener functions + // on trigger so that it does not mess up the order + // while we loop through. + // + // Here we slice on off so that the loop in trigger + // can continue using it's old reference to loop without + // messing up the order. + + this.listeners[type] = this.listeners[type].slice(0); + this.listeners[type].splice(index, 1); + return index > -1; + } + /** + * Trigger an event of the specified type on this stream. Any additional + * arguments to this function are passed as parameters to event listeners. + * + * @param {string} type the event name + */; + + _proto.trigger = function trigger(type) { + var callbacks = this.listeners[type]; + if (!callbacks) { + return; + } // Slicing the arguments on every invocation of this method + // can add a significant amount of overhead. Avoid the + // intermediate object creation for the common case of a + // single callback argument + + if (arguments.length === 2) { + var length = callbacks.length; + for (var i = 0; i < length; ++i) { + callbacks[i].call(this, arguments[1]); + } + } else { + var args = Array.prototype.slice.call(arguments, 1); + var _length = callbacks.length; + for (var _i = 0; _i < _length; ++_i) { + callbacks[_i].apply(this, args); + } + } + } + /** + * Destroys the stream and cleans up. + */; + + _proto.dispose = function dispose() { + this.listeners = {}; + } + /** + * Forwards all `data` events on this stream to the destination stream. The + * destination stream should provide a method `push` to receive the data + * events as they arrive. + * + * @param {Stream} destination the stream that will receive all `data` events + * @see http://nodejs.org/api/stream.html#stream_readable_pipe_destination_options + */; + + _proto.pipe = function pipe(destination) { + this.on('data', function (data) { + destination.push(data); + }); + }; + return Stream; + }(); + /*! @name pkcs7 @version 1.0.4 @license Apache-2.0 */ + + /** + * Returns the subarray of a Uint8Array without PKCS#7 padding. + * + * @param padded {Uint8Array} unencrypted bytes that have been padded + * @return {Uint8Array} the unpadded bytes + * @see http://tools.ietf.org/html/rfc5652 + */ + + function unpad(padded) { + return padded.subarray(0, padded.byteLength - padded[padded.byteLength - 1]); + } + /*! @name aes-decrypter @version 4.0.1 @license Apache-2.0 */ + + /** + * @file aes.js + * + * This file contains an adaptation of the AES decryption algorithm + * from the Standford Javascript Cryptography Library. That work is + * covered by the following copyright and permissions notice: + * + * Copyright 2009-2010 Emily Stark, Mike Hamburg, Dan Boneh. + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are + * met: + * + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * + * 2. Redistributions in binary form must reproduce the above + * copyright notice, this list of conditions and the following + * disclaimer in the documentation and/or other materials provided + * with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE AUTHORS ``AS IS'' AND ANY EXPRESS OR + * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED + * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE + * DISCLAIMED. IN NO EVENT SHALL <COPYRIGHT HOLDER> OR CONTRIBUTORS BE + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR + * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, + * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE + * OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN + * IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + * + * The views and conclusions contained in the software and documentation + * are those of the authors and should not be interpreted as representing + * official policies, either expressed or implied, of the authors. + */ + + /** + * Expand the S-box tables. + * + * @private + */ + + const precompute = function () { + const tables = [[[], [], [], [], []], [[], [], [], [], []]]; + const encTable = tables[0]; + const decTable = tables[1]; + const sbox = encTable[4]; + const sboxInv = decTable[4]; + let i; + let x; + let xInv; + const d = []; + const th = []; + let x2; + let x4; + let x8; + let s; + let tEnc; + let tDec; // Compute double and third tables + + for (i = 0; i < 256; i++) { + th[(d[i] = i << 1 ^ (i >> 7) * 283) ^ i] = i; + } + for (x = xInv = 0; !sbox[x]; x ^= x2 || 1, xInv = th[xInv] || 1) { + // Compute sbox + s = xInv ^ xInv << 1 ^ xInv << 2 ^ xInv << 3 ^ xInv << 4; + s = s >> 8 ^ s & 255 ^ 99; + sbox[x] = s; + sboxInv[s] = x; // Compute MixColumns + + x8 = d[x4 = d[x2 = d[x]]]; + tDec = x8 * 0x1010101 ^ x4 * 0x10001 ^ x2 * 0x101 ^ x * 0x1010100; + tEnc = d[s] * 0x101 ^ s * 0x1010100; + for (i = 0; i < 4; i++) { + encTable[i][x] = tEnc = tEnc << 24 ^ tEnc >>> 8; + decTable[i][s] = tDec = tDec << 24 ^ tDec >>> 8; + } + } // Compactify. Considerable speedup on Firefox. + + for (i = 0; i < 5; i++) { + encTable[i] = encTable[i].slice(0); + decTable[i] = decTable[i].slice(0); + } + return tables; + }; + let aesTables = null; + /** + * Schedule out an AES key for both encryption and decryption. This + * is a low-level class. Use a cipher mode to do bulk encryption. + * + * @class AES + * @param key {Array} The key as an array of 4, 6 or 8 words. + */ + + class AES { + constructor(key) { + /** + * The expanded S-box and inverse S-box tables. These will be computed + * on the client so that we don't have to send them down the wire. + * + * There are two tables, _tables[0] is for encryption and + * _tables[1] is for decryption. + * + * The first 4 sub-tables are the expanded S-box with MixColumns. The + * last (_tables[01][4]) is the S-box itself. + * + * @private + */ + // if we have yet to precompute the S-box tables + // do so now + if (!aesTables) { + aesTables = precompute(); + } // then make a copy of that object for use + + this._tables = [[aesTables[0][0].slice(), aesTables[0][1].slice(), aesTables[0][2].slice(), aesTables[0][3].slice(), aesTables[0][4].slice()], [aesTables[1][0].slice(), aesTables[1][1].slice(), aesTables[1][2].slice(), aesTables[1][3].slice(), aesTables[1][4].slice()]]; + let i; + let j; + let tmp; + const sbox = this._tables[0][4]; + const decTable = this._tables[1]; + const keyLen = key.length; + let rcon = 1; + if (keyLen !== 4 && keyLen !== 6 && keyLen !== 8) { + throw new Error('Invalid aes key size'); + } + const encKey = key.slice(0); + const decKey = []; + this._key = [encKey, decKey]; // schedule encryption keys + + for (i = keyLen; i < 4 * keyLen + 28; i++) { + tmp = encKey[i - 1]; // apply sbox + + if (i % keyLen === 0 || keyLen === 8 && i % keyLen === 4) { + tmp = sbox[tmp >>> 24] << 24 ^ sbox[tmp >> 16 & 255] << 16 ^ sbox[tmp >> 8 & 255] << 8 ^ sbox[tmp & 255]; // shift rows and add rcon + + if (i % keyLen === 0) { + tmp = tmp << 8 ^ tmp >>> 24 ^ rcon << 24; + rcon = rcon << 1 ^ (rcon >> 7) * 283; + } + } + encKey[i] = encKey[i - keyLen] ^ tmp; + } // schedule decryption keys + + for (j = 0; i; j++, i--) { + tmp = encKey[j & 3 ? i : i - 4]; + if (i <= 4 || j < 4) { + decKey[j] = tmp; + } else { + decKey[j] = decTable[0][sbox[tmp >>> 24]] ^ decTable[1][sbox[tmp >> 16 & 255]] ^ decTable[2][sbox[tmp >> 8 & 255]] ^ decTable[3][sbox[tmp & 255]]; + } + } + } + /** + * Decrypt 16 bytes, specified as four 32-bit words. + * + * @param {number} encrypted0 the first word to decrypt + * @param {number} encrypted1 the second word to decrypt + * @param {number} encrypted2 the third word to decrypt + * @param {number} encrypted3 the fourth word to decrypt + * @param {Int32Array} out the array to write the decrypted words + * into + * @param {number} offset the offset into the output array to start + * writing results + * @return {Array} The plaintext. + */ + + decrypt(encrypted0, encrypted1, encrypted2, encrypted3, out, offset) { + const key = this._key[1]; // state variables a,b,c,d are loaded with pre-whitened data + + let a = encrypted0 ^ key[0]; + let b = encrypted3 ^ key[1]; + let c = encrypted2 ^ key[2]; + let d = encrypted1 ^ key[3]; + let a2; + let b2; + let c2; // key.length === 2 ? + + const nInnerRounds = key.length / 4 - 2; + let i; + let kIndex = 4; + const table = this._tables[1]; // load up the tables + + const table0 = table[0]; + const table1 = table[1]; + const table2 = table[2]; + const table3 = table[3]; + const sbox = table[4]; // Inner rounds. Cribbed from OpenSSL. + + for (i = 0; i < nInnerRounds; i++) { + a2 = table0[a >>> 24] ^ table1[b >> 16 & 255] ^ table2[c >> 8 & 255] ^ table3[d & 255] ^ key[kIndex]; + b2 = table0[b >>> 24] ^ table1[c >> 16 & 255] ^ table2[d >> 8 & 255] ^ table3[a & 255] ^ key[kIndex + 1]; + c2 = table0[c >>> 24] ^ table1[d >> 16 & 255] ^ table2[a >> 8 & 255] ^ table3[b & 255] ^ key[kIndex + 2]; + d = table0[d >>> 24] ^ table1[a >> 16 & 255] ^ table2[b >> 8 & 255] ^ table3[c & 255] ^ key[kIndex + 3]; + kIndex += 4; + a = a2; + b = b2; + c = c2; + } // Last round. + + for (i = 0; i < 4; i++) { + out[(3 & -i) + offset] = sbox[a >>> 24] << 24 ^ sbox[b >> 16 & 255] << 16 ^ sbox[c >> 8 & 255] << 8 ^ sbox[d & 255] ^ key[kIndex++]; + a2 = a; + a = b; + b = c; + c = d; + d = a2; + } + } + } + /** + * @file async-stream.js + */ + + /** + * A wrapper around the Stream class to use setTimeout + * and run stream "jobs" Asynchronously + * + * @class AsyncStream + * @extends Stream + */ + + class AsyncStream extends Stream { + constructor() { + super(Stream); + this.jobs = []; + this.delay = 1; + this.timeout_ = null; + } + /** + * process an async job + * + * @private + */ + + processJob_() { + this.jobs.shift()(); + if (this.jobs.length) { + this.timeout_ = setTimeout(this.processJob_.bind(this), this.delay); + } else { + this.timeout_ = null; + } + } + /** + * push a job into the stream + * + * @param {Function} job the job to push into the stream + */ + + push(job) { + this.jobs.push(job); + if (!this.timeout_) { + this.timeout_ = setTimeout(this.processJob_.bind(this), this.delay); + } + } + } + /** + * @file decrypter.js + * + * An asynchronous implementation of AES-128 CBC decryption with + * PKCS#7 padding. + */ + + /** + * Convert network-order (big-endian) bytes into their little-endian + * representation. + */ + + const ntoh = function (word) { + return word << 24 | (word & 0xff00) << 8 | (word & 0xff0000) >> 8 | word >>> 24; + }; + /** + * Decrypt bytes using AES-128 with CBC and PKCS#7 padding. + * + * @param {Uint8Array} encrypted the encrypted bytes + * @param {Uint32Array} key the bytes of the decryption key + * @param {Uint32Array} initVector the initialization vector (IV) to + * use for the first round of CBC. + * @return {Uint8Array} the decrypted bytes + * + * @see http://en.wikipedia.org/wiki/Advanced_Encryption_Standard + * @see http://en.wikipedia.org/wiki/Block_cipher_mode_of_operation#Cipher_Block_Chaining_.28CBC.29 + * @see https://tools.ietf.org/html/rfc2315 + */ + + const decrypt = function (encrypted, key, initVector) { + // word-level access to the encrypted bytes + const encrypted32 = new Int32Array(encrypted.buffer, encrypted.byteOffset, encrypted.byteLength >> 2); + const decipher = new AES(Array.prototype.slice.call(key)); // byte and word-level access for the decrypted output + + const decrypted = new Uint8Array(encrypted.byteLength); + const decrypted32 = new Int32Array(decrypted.buffer); // temporary variables for working with the IV, encrypted, and + // decrypted data + + let init0; + let init1; + let init2; + let init3; + let encrypted0; + let encrypted1; + let encrypted2; + let encrypted3; // iteration variable + + let wordIx; // pull out the words of the IV to ensure we don't modify the + // passed-in reference and easier access + + init0 = initVector[0]; + init1 = initVector[1]; + init2 = initVector[2]; + init3 = initVector[3]; // decrypt four word sequences, applying cipher-block chaining (CBC) + // to each decrypted block + + for (wordIx = 0; wordIx < encrypted32.length; wordIx += 4) { + // convert big-endian (network order) words into little-endian + // (javascript order) + encrypted0 = ntoh(encrypted32[wordIx]); + encrypted1 = ntoh(encrypted32[wordIx + 1]); + encrypted2 = ntoh(encrypted32[wordIx + 2]); + encrypted3 = ntoh(encrypted32[wordIx + 3]); // decrypt the block + + decipher.decrypt(encrypted0, encrypted1, encrypted2, encrypted3, decrypted32, wordIx); // XOR with the IV, and restore network byte-order to obtain the + // plaintext + + decrypted32[wordIx] = ntoh(decrypted32[wordIx] ^ init0); + decrypted32[wordIx + 1] = ntoh(decrypted32[wordIx + 1] ^ init1); + decrypted32[wordIx + 2] = ntoh(decrypted32[wordIx + 2] ^ init2); + decrypted32[wordIx + 3] = ntoh(decrypted32[wordIx + 3] ^ init3); // setup the IV for the next round + + init0 = encrypted0; + init1 = encrypted1; + init2 = encrypted2; + init3 = encrypted3; + } + return decrypted; + }; + /** + * The `Decrypter` class that manages decryption of AES + * data through `AsyncStream` objects and the `decrypt` + * function + * + * @param {Uint8Array} encrypted the encrypted bytes + * @param {Uint32Array} key the bytes of the decryption key + * @param {Uint32Array} initVector the initialization vector (IV) to + * @param {Function} done the function to run when done + * @class Decrypter + */ + + class Decrypter { + constructor(encrypted, key, initVector, done) { + const step = Decrypter.STEP; + const encrypted32 = new Int32Array(encrypted.buffer); + const decrypted = new Uint8Array(encrypted.byteLength); + let i = 0; + this.asyncStream_ = new AsyncStream(); // split up the encryption job and do the individual chunks asynchronously + + this.asyncStream_.push(this.decryptChunk_(encrypted32.subarray(i, i + step), key, initVector, decrypted)); + for (i = step; i < encrypted32.length; i += step) { + initVector = new Uint32Array([ntoh(encrypted32[i - 4]), ntoh(encrypted32[i - 3]), ntoh(encrypted32[i - 2]), ntoh(encrypted32[i - 1])]); + this.asyncStream_.push(this.decryptChunk_(encrypted32.subarray(i, i + step), key, initVector, decrypted)); + } // invoke the done() callback when everything is finished + + this.asyncStream_.push(function () { + // remove pkcs#7 padding from the decrypted bytes + done(null, unpad(decrypted)); + }); + } + /** + * a getter for step the maximum number of bytes to process at one time + * + * @return {number} the value of step 32000 + */ + + static get STEP() { + // 4 * 8000; + return 32000; + } + /** + * @private + */ + + decryptChunk_(encrypted, key, initVector, decrypted) { + return function () { + const bytes = decrypt(encrypted, key, initVector); + decrypted.set(bytes, encrypted.byteOffset); + }; + } + } + var commonjsGlobal = typeof globalThis !== 'undefined' ? globalThis : typeof window !== 'undefined' ? window : typeof global !== 'undefined' ? global : typeof self !== 'undefined' ? self : {}; + var win; + if (typeof window !== "undefined") { + win = window; + } else if (typeof commonjsGlobal !== "undefined") { + win = commonjsGlobal; + } else if (typeof self !== "undefined") { + win = self; + } else { + win = {}; + } + var window_1 = win; + var isArrayBufferView = function isArrayBufferView(obj) { + if (ArrayBuffer.isView === 'function') { + return ArrayBuffer.isView(obj); + } + return obj && obj.buffer instanceof ArrayBuffer; + }; + var BigInt = window_1.BigInt || Number; + [BigInt('0x1'), BigInt('0x100'), BigInt('0x10000'), BigInt('0x1000000'), BigInt('0x100000000'), BigInt('0x10000000000'), BigInt('0x1000000000000'), BigInt('0x100000000000000'), BigInt('0x10000000000000000')]; + (function () { + var a = new Uint16Array([0xFFCC]); + var b = new Uint8Array(a.buffer, a.byteOffset, a.byteLength); + if (b[0] === 0xFF) { + return 'big'; + } + if (b[0] === 0xCC) { + return 'little'; + } + return 'unknown'; + })(); + /** + * Creates an object for sending to a web worker modifying properties that are TypedArrays + * into a new object with seperated properties for the buffer, byteOffset, and byteLength. + * + * @param {Object} message + * Object of properties and values to send to the web worker + * @return {Object} + * Modified message with TypedArray values expanded + * @function createTransferableMessage + */ + + const createTransferableMessage = function (message) { + const transferable = {}; + Object.keys(message).forEach(key => { + const value = message[key]; + if (isArrayBufferView(value)) { + transferable[key] = { + bytes: value.buffer, + byteOffset: value.byteOffset, + byteLength: value.byteLength + }; + } else { + transferable[key] = value; + } + }); + return transferable; + }; + /* global self */ + + /** + * Our web worker interface so that things can talk to aes-decrypter + * that will be running in a web worker. the scope is passed to this by + * webworkify. + */ + + self.onmessage = function (event) { + const data = event.data; + const encrypted = new Uint8Array(data.encrypted.bytes, data.encrypted.byteOffset, data.encrypted.byteLength); + const key = new Uint32Array(data.key.bytes, data.key.byteOffset, data.key.byteLength / 4); + const iv = new Uint32Array(data.iv.bytes, data.iv.byteOffset, data.iv.byteLength / 4); + /* eslint-disable no-new, handle-callback-err */ + + new Decrypter(encrypted, key, iv, function (err, bytes) { + self.postMessage(createTransferableMessage({ + source: data.source, + decrypted: bytes + }), [bytes.buffer]); + }); + /* eslint-enable */ + }; +})); + +var Decrypter = factory(workerCode); +/* rollup-plugin-worker-factory end for worker!/home/runner/work/http-streaming/http-streaming/src/decrypter-worker.js */ + +/** + * Convert the properties of an HLS track into an audioTrackKind. + * + * @private + */ + +const audioTrackKind_ = properties => { + let kind = properties.default ? 'main' : 'alternative'; + if (properties.characteristics && properties.characteristics.indexOf('public.accessibility.describes-video') >= 0) { + kind = 'main-desc'; + } + return kind; +}; +/** + * Pause provided segment loader and playlist loader if active + * + * @param {SegmentLoader} segmentLoader + * SegmentLoader to pause + * @param {Object} mediaType + * Active media type + * @function stopLoaders + */ + +const stopLoaders = (segmentLoader, mediaType) => { + segmentLoader.abort(); + segmentLoader.pause(); + if (mediaType && mediaType.activePlaylistLoader) { + mediaType.activePlaylistLoader.pause(); + mediaType.activePlaylistLoader = null; + } +}; +/** + * Start loading provided segment loader and playlist loader + * + * @param {PlaylistLoader} playlistLoader + * PlaylistLoader to start loading + * @param {Object} mediaType + * Active media type + * @function startLoaders + */ + +const startLoaders = (playlistLoader, mediaType) => { + // Segment loader will be started after `loadedmetadata` or `loadedplaylist` from the + // playlist loader + mediaType.activePlaylistLoader = playlistLoader; + playlistLoader.load(); +}; +/** + * Returns a function to be called when the media group changes. It performs a + * non-destructive (preserve the buffer) resync of the SegmentLoader. This is because a + * change of group is merely a rendition switch of the same content at another encoding, + * rather than a change of content, such as switching audio from English to Spanish. + * + * @param {string} type + * MediaGroup type + * @param {Object} settings + * Object containing required information for media groups + * @return {Function} + * Handler for a non-destructive resync of SegmentLoader when the active media + * group changes. + * @function onGroupChanged + */ + +const onGroupChanged = (type, settings) => () => { + const { + segmentLoaders: { + [type]: segmentLoader, + main: mainSegmentLoader + }, + mediaTypes: { + [type]: mediaType + } + } = settings; + const activeTrack = mediaType.activeTrack(); + const activeGroup = mediaType.getActiveGroup(); + const previousActiveLoader = mediaType.activePlaylistLoader; + const lastGroup = mediaType.lastGroup_; // the group did not change do nothing + + if (activeGroup && lastGroup && activeGroup.id === lastGroup.id) { + return; + } + mediaType.lastGroup_ = activeGroup; + mediaType.lastTrack_ = activeTrack; + stopLoaders(segmentLoader, mediaType); + if (!activeGroup || activeGroup.isMainPlaylist) { + // there is no group active or active group is a main playlist and won't change + return; + } + if (!activeGroup.playlistLoader) { + if (previousActiveLoader) { + // The previous group had a playlist loader but the new active group does not + // this means we are switching from demuxed to muxed audio. In this case we want to + // do a destructive reset of the main segment loader and not restart the audio + // loaders. + mainSegmentLoader.resetEverything(); + } + return; + } // Non-destructive resync + + segmentLoader.resyncLoader(); + startLoaders(activeGroup.playlistLoader, mediaType); +}; +const onGroupChanging = (type, settings) => () => { + const { + segmentLoaders: { + [type]: segmentLoader + }, + mediaTypes: { + [type]: mediaType + } + } = settings; + mediaType.lastGroup_ = null; + segmentLoader.abort(); + segmentLoader.pause(); +}; +/** + * Returns a function to be called when the media track changes. It performs a + * destructive reset of the SegmentLoader to ensure we start loading as close to + * currentTime as possible. + * + * @param {string} type + * MediaGroup type + * @param {Object} settings + * Object containing required information for media groups + * @return {Function} + * Handler for a destructive reset of SegmentLoader when the active media + * track changes. + * @function onTrackChanged + */ + +const onTrackChanged = (type, settings) => () => { + const { + mainPlaylistLoader, + segmentLoaders: { + [type]: segmentLoader, + main: mainSegmentLoader + }, + mediaTypes: { + [type]: mediaType + } + } = settings; + const activeTrack = mediaType.activeTrack(); + const activeGroup = mediaType.getActiveGroup(); + const previousActiveLoader = mediaType.activePlaylistLoader; + const lastTrack = mediaType.lastTrack_; // track did not change, do nothing + + if (lastTrack && activeTrack && lastTrack.id === activeTrack.id) { + return; + } + mediaType.lastGroup_ = activeGroup; + mediaType.lastTrack_ = activeTrack; + stopLoaders(segmentLoader, mediaType); + if (!activeGroup) { + // there is no group active so we do not want to restart loaders + return; + } + if (activeGroup.isMainPlaylist) { + // track did not change, do nothing + if (!activeTrack || !lastTrack || activeTrack.id === lastTrack.id) { + return; + } + const pc = settings.vhs.playlistController_; + const newPlaylist = pc.selectPlaylist(); // media will not change do nothing + + if (pc.media() === newPlaylist) { + return; + } + mediaType.logger_(`track change. Switching main audio from ${lastTrack.id} to ${activeTrack.id}`); + mainPlaylistLoader.pause(); + mainSegmentLoader.resetEverything(); + pc.fastQualityChange_(newPlaylist); + return; + } + if (type === 'AUDIO') { + if (!activeGroup.playlistLoader) { + // when switching from demuxed audio/video to muxed audio/video (noted by no + // playlist loader for the audio group), we want to do a destructive reset of the + // main segment loader and not restart the audio loaders + mainSegmentLoader.setAudio(true); // don't have to worry about disabling the audio of the audio segment loader since + // it should be stopped + + mainSegmentLoader.resetEverything(); + return; + } // although the segment loader is an audio segment loader, call the setAudio + // function to ensure it is prepared to re-append the init segment (or handle other + // config changes) + + segmentLoader.setAudio(true); + mainSegmentLoader.setAudio(false); + } + if (previousActiveLoader === activeGroup.playlistLoader) { + // Nothing has actually changed. This can happen because track change events can fire + // multiple times for a "single" change. One for enabling the new active track, and + // one for disabling the track that was active + startLoaders(activeGroup.playlistLoader, mediaType); + return; + } + if (segmentLoader.track) { + // For WebVTT, set the new text track in the segmentloader + segmentLoader.track(activeTrack); + } // destructive reset + + segmentLoader.resetEverything(); + startLoaders(activeGroup.playlistLoader, mediaType); +}; +const onError = { + /** + * Returns a function to be called when a SegmentLoader or PlaylistLoader encounters + * an error. + * + * @param {string} type + * MediaGroup type + * @param {Object} settings + * Object containing required information for media groups + * @return {Function} + * Error handler. Logs warning (or error if the playlist is excluded) to + * console and switches back to default audio track. + * @function onError.AUDIO + */ + AUDIO: (type, settings) => () => { + const { + mediaTypes: { + [type]: mediaType + }, + excludePlaylist + } = settings; // switch back to default audio track + + const activeTrack = mediaType.activeTrack(); + const activeGroup = mediaType.activeGroup(); + const id = (activeGroup.filter(group => group.default)[0] || activeGroup[0]).id; + const defaultTrack = mediaType.tracks[id]; + if (activeTrack === defaultTrack) { + // Default track encountered an error. All we can do now is exclude the current + // rendition and hope another will switch audio groups + excludePlaylist({ + error: { + message: 'Problem encountered loading the default audio track.' + } + }); + return; + } + videojs.log.warn('Problem encountered loading the alternate audio track.' + 'Switching back to default.'); + for (const trackId in mediaType.tracks) { + mediaType.tracks[trackId].enabled = mediaType.tracks[trackId] === defaultTrack; + } + mediaType.onTrackChanged(); + }, + /** + * Returns a function to be called when a SegmentLoader or PlaylistLoader encounters + * an error. + * + * @param {string} type + * MediaGroup type + * @param {Object} settings + * Object containing required information for media groups + * @return {Function} + * Error handler. Logs warning to console and disables the active subtitle track + * @function onError.SUBTITLES + */ + SUBTITLES: (type, settings) => () => { + const { + mediaTypes: { + [type]: mediaType + } + } = settings; + videojs.log.warn('Problem encountered loading the subtitle track.' + 'Disabling subtitle track.'); + const track = mediaType.activeTrack(); + if (track) { + track.mode = 'disabled'; + } + mediaType.onTrackChanged(); + } +}; +const setupListeners = { + /** + * Setup event listeners for audio playlist loader + * + * @param {string} type + * MediaGroup type + * @param {PlaylistLoader|null} playlistLoader + * PlaylistLoader to register listeners on + * @param {Object} settings + * Object containing required information for media groups + * @function setupListeners.AUDIO + */ + AUDIO: (type, playlistLoader, settings) => { + if (!playlistLoader) { + // no playlist loader means audio will be muxed with the video + return; + } + const { + tech, + requestOptions, + segmentLoaders: { + [type]: segmentLoader + } + } = settings; + playlistLoader.on('loadedmetadata', () => { + const media = playlistLoader.media(); + segmentLoader.playlist(media, requestOptions); // if the video is already playing, or if this isn't a live video and preload + // permits, start downloading segments + + if (!tech.paused() || media.endList && tech.preload() !== 'none') { + segmentLoader.load(); + } + }); + playlistLoader.on('loadedplaylist', () => { + segmentLoader.playlist(playlistLoader.media(), requestOptions); // If the player isn't paused, ensure that the segment loader is running + + if (!tech.paused()) { + segmentLoader.load(); + } + }); + playlistLoader.on('error', onError[type](type, settings)); + }, + /** + * Setup event listeners for subtitle playlist loader + * + * @param {string} type + * MediaGroup type + * @param {PlaylistLoader|null} playlistLoader + * PlaylistLoader to register listeners on + * @param {Object} settings + * Object containing required information for media groups + * @function setupListeners.SUBTITLES + */ + SUBTITLES: (type, playlistLoader, settings) => { + const { + tech, + requestOptions, + segmentLoaders: { + [type]: segmentLoader + }, + mediaTypes: { + [type]: mediaType + } + } = settings; + playlistLoader.on('loadedmetadata', () => { + const media = playlistLoader.media(); + segmentLoader.playlist(media, requestOptions); + segmentLoader.track(mediaType.activeTrack()); // if the video is already playing, or if this isn't a live video and preload + // permits, start downloading segments + + if (!tech.paused() || media.endList && tech.preload() !== 'none') { + segmentLoader.load(); + } + }); + playlistLoader.on('loadedplaylist', () => { + segmentLoader.playlist(playlistLoader.media(), requestOptions); // If the player isn't paused, ensure that the segment loader is running + + if (!tech.paused()) { + segmentLoader.load(); + } + }); + playlistLoader.on('error', onError[type](type, settings)); + } +}; +const initialize = { + /** + * Setup PlaylistLoaders and AudioTracks for the audio groups + * + * @param {string} type + * MediaGroup type + * @param {Object} settings + * Object containing required information for media groups + * @function initialize.AUDIO + */ + 'AUDIO': (type, settings) => { + const { + vhs, + sourceType, + segmentLoaders: { + [type]: segmentLoader + }, + requestOptions, + main: { + mediaGroups + }, + mediaTypes: { + [type]: { + groups, + tracks, + logger_ + } + }, + mainPlaylistLoader + } = settings; + const audioOnlyMain = isAudioOnly(mainPlaylistLoader.main); // force a default if we have none + + if (!mediaGroups[type] || Object.keys(mediaGroups[type]).length === 0) { + mediaGroups[type] = { + main: { + default: { + default: true + } + } + }; + if (audioOnlyMain) { + mediaGroups[type].main.default.playlists = mainPlaylistLoader.main.playlists; + } + } + for (const groupId in mediaGroups[type]) { + if (!groups[groupId]) { + groups[groupId] = []; + } + for (const variantLabel in mediaGroups[type][groupId]) { + let properties = mediaGroups[type][groupId][variantLabel]; + let playlistLoader; + if (audioOnlyMain) { + logger_(`AUDIO group '${groupId}' label '${variantLabel}' is a main playlist`); + properties.isMainPlaylist = true; + playlistLoader = null; // if vhs-json was provided as the source, and the media playlist was resolved, + // use the resolved media playlist object + } else if (sourceType === 'vhs-json' && properties.playlists) { + playlistLoader = new PlaylistLoader(properties.playlists[0], vhs, requestOptions); + } else if (properties.resolvedUri) { + playlistLoader = new PlaylistLoader(properties.resolvedUri, vhs, requestOptions); // TODO: dash isn't the only type with properties.playlists + // should we even have properties.playlists in this check. + } else if (properties.playlists && sourceType === 'dash') { + playlistLoader = new DashPlaylistLoader(properties.playlists[0], vhs, requestOptions, mainPlaylistLoader); + } else { + // no resolvedUri means the audio is muxed with the video when using this + // audio track + playlistLoader = null; + } + properties = merge({ + id: variantLabel, + playlistLoader + }, properties); + setupListeners[type](type, properties.playlistLoader, settings); + groups[groupId].push(properties); + if (typeof tracks[variantLabel] === 'undefined') { + const track = new videojs.AudioTrack({ + id: variantLabel, + kind: audioTrackKind_(properties), + enabled: false, + language: properties.language, + default: properties.default, + label: variantLabel + }); + tracks[variantLabel] = track; + } + } + } // setup single error event handler for the segment loader + + segmentLoader.on('error', onError[type](type, settings)); + }, + /** + * Setup PlaylistLoaders and TextTracks for the subtitle groups + * + * @param {string} type + * MediaGroup type + * @param {Object} settings + * Object containing required information for media groups + * @function initialize.SUBTITLES + */ + 'SUBTITLES': (type, settings) => { + const { + tech, + vhs, + sourceType, + segmentLoaders: { + [type]: segmentLoader + }, + requestOptions, + main: { + mediaGroups + }, + mediaTypes: { + [type]: { + groups, + tracks + } + }, + mainPlaylistLoader + } = settings; + for (const groupId in mediaGroups[type]) { + if (!groups[groupId]) { + groups[groupId] = []; + } + for (const variantLabel in mediaGroups[type][groupId]) { + if (!vhs.options_.useForcedSubtitles && mediaGroups[type][groupId][variantLabel].forced) { + // Subtitle playlists with the forced attribute are not selectable in Safari. + // According to Apple's HLS Authoring Specification: + // If content has forced subtitles and regular subtitles in a given language, + // the regular subtitles track in that language MUST contain both the forced + // subtitles and the regular subtitles for that language. + // Because of this requirement and that Safari does not add forced subtitles, + // forced subtitles are skipped here to maintain consistent experience across + // all platforms + continue; + } + let properties = mediaGroups[type][groupId][variantLabel]; + let playlistLoader; + if (sourceType === 'hls') { + playlistLoader = new PlaylistLoader(properties.resolvedUri, vhs, requestOptions); + } else if (sourceType === 'dash') { + const playlists = properties.playlists.filter(p => p.excludeUntil !== Infinity); + if (!playlists.length) { + return; + } + playlistLoader = new DashPlaylistLoader(properties.playlists[0], vhs, requestOptions, mainPlaylistLoader); + } else if (sourceType === 'vhs-json') { + playlistLoader = new PlaylistLoader( + // if the vhs-json object included the media playlist, use the media playlist + // as provided, otherwise use the resolved URI to load the playlist + properties.playlists ? properties.playlists[0] : properties.resolvedUri, vhs, requestOptions); + } + properties = merge({ + id: variantLabel, + playlistLoader + }, properties); + setupListeners[type](type, properties.playlistLoader, settings); + groups[groupId].push(properties); + if (typeof tracks[variantLabel] === 'undefined') { + const track = tech.addRemoteTextTrack({ + id: variantLabel, + kind: 'subtitles', + default: properties.default && properties.autoselect, + language: properties.language, + label: variantLabel + }, false).track; + tracks[variantLabel] = track; + } + } + } // setup single error event handler for the segment loader + + segmentLoader.on('error', onError[type](type, settings)); + }, + /** + * Setup TextTracks for the closed-caption groups + * + * @param {String} type + * MediaGroup type + * @param {Object} settings + * Object containing required information for media groups + * @function initialize['CLOSED-CAPTIONS'] + */ + 'CLOSED-CAPTIONS': (type, settings) => { + const { + tech, + main: { + mediaGroups + }, + mediaTypes: { + [type]: { + groups, + tracks + } + } + } = settings; + for (const groupId in mediaGroups[type]) { + if (!groups[groupId]) { + groups[groupId] = []; + } + for (const variantLabel in mediaGroups[type][groupId]) { + const properties = mediaGroups[type][groupId][variantLabel]; // Look for either 608 (CCn) or 708 (SERVICEn) caption services + + if (!/^(?:CC|SERVICE)/.test(properties.instreamId)) { + continue; + } + const captionServices = tech.options_.vhs && tech.options_.vhs.captionServices || {}; + let newProps = { + label: variantLabel, + language: properties.language, + instreamId: properties.instreamId, + default: properties.default && properties.autoselect + }; + if (captionServices[newProps.instreamId]) { + newProps = merge(newProps, captionServices[newProps.instreamId]); + } + if (newProps.default === undefined) { + delete newProps.default; + } // No PlaylistLoader is required for Closed-Captions because the captions are + // embedded within the video stream + + groups[groupId].push(merge({ + id: variantLabel + }, properties)); + if (typeof tracks[variantLabel] === 'undefined') { + const track = tech.addRemoteTextTrack({ + id: newProps.instreamId, + kind: 'captions', + default: newProps.default, + language: newProps.language, + label: newProps.label + }, false).track; + tracks[variantLabel] = track; + } + } + } + } +}; +const groupMatch = (list, media) => { + for (let i = 0; i < list.length; i++) { + if (playlistMatch(media, list[i])) { + return true; + } + if (list[i].playlists && groupMatch(list[i].playlists, media)) { + return true; + } + } + return false; +}; +/** + * Returns a function used to get the active group of the provided type + * + * @param {string} type + * MediaGroup type + * @param {Object} settings + * Object containing required information for media groups + * @return {Function} + * Function that returns the active media group for the provided type. Takes an + * optional parameter {TextTrack} track. If no track is provided, a list of all + * variants in the group, otherwise the variant corresponding to the provided + * track is returned. + * @function activeGroup + */ + +const activeGroup = (type, settings) => track => { + const { + mainPlaylistLoader, + mediaTypes: { + [type]: { + groups + } + } + } = settings; + const media = mainPlaylistLoader.media(); + if (!media) { + return null; + } + let variants = null; // set to variants to main media active group + + if (media.attributes[type]) { + variants = groups[media.attributes[type]]; + } + const groupKeys = Object.keys(groups); + if (!variants) { + // find the mainPlaylistLoader media + // that is in a media group if we are dealing + // with audio only + if (type === 'AUDIO' && groupKeys.length > 1 && isAudioOnly(settings.main)) { + for (let i = 0; i < groupKeys.length; i++) { + const groupPropertyList = groups[groupKeys[i]]; + if (groupMatch(groupPropertyList, media)) { + variants = groupPropertyList; + break; + } + } // use the main group if it exists + } else if (groups.main) { + variants = groups.main; // only one group, use that one + } else if (groupKeys.length === 1) { + variants = groups[groupKeys[0]]; + } + } + if (typeof track === 'undefined') { + return variants; + } + if (track === null || !variants) { + // An active track was specified so a corresponding group is expected. track === null + // means no track is currently active so there is no corresponding group + return null; + } + return variants.filter(props => props.id === track.id)[0] || null; +}; +const activeTrack = { + /** + * Returns a function used to get the active track of type provided + * + * @param {string} type + * MediaGroup type + * @param {Object} settings + * Object containing required information for media groups + * @return {Function} + * Function that returns the active media track for the provided type. Returns + * null if no track is active + * @function activeTrack.AUDIO + */ + AUDIO: (type, settings) => () => { + const { + mediaTypes: { + [type]: { + tracks + } + } + } = settings; + for (const id in tracks) { + if (tracks[id].enabled) { + return tracks[id]; + } + } + return null; + }, + /** + * Returns a function used to get the active track of type provided + * + * @param {string} type + * MediaGroup type + * @param {Object} settings + * Object containing required information for media groups + * @return {Function} + * Function that returns the active media track for the provided type. Returns + * null if no track is active + * @function activeTrack.SUBTITLES + */ + SUBTITLES: (type, settings) => () => { + const { + mediaTypes: { + [type]: { + tracks + } + } + } = settings; + for (const id in tracks) { + if (tracks[id].mode === 'showing' || tracks[id].mode === 'hidden') { + return tracks[id]; + } + } + return null; + } +}; +const getActiveGroup = (type, { + mediaTypes +}) => () => { + const activeTrack_ = mediaTypes[type].activeTrack(); + if (!activeTrack_) { + return null; + } + return mediaTypes[type].activeGroup(activeTrack_); +}; +/** + * Setup PlaylistLoaders and Tracks for media groups (Audio, Subtitles, + * Closed-Captions) specified in the main manifest. + * + * @param {Object} settings + * Object containing required information for setting up the media groups + * @param {Tech} settings.tech + * The tech of the player + * @param {Object} settings.requestOptions + * XHR request options used by the segment loaders + * @param {PlaylistLoader} settings.mainPlaylistLoader + * PlaylistLoader for the main source + * @param {VhsHandler} settings.vhs + * VHS SourceHandler + * @param {Object} settings.main + * The parsed main manifest + * @param {Object} settings.mediaTypes + * Object to store the loaders, tracks, and utility methods for each media type + * @param {Function} settings.excludePlaylist + * Excludes the current rendition and forces a rendition switch. + * @function setupMediaGroups + */ + +const setupMediaGroups = settings => { + ['AUDIO', 'SUBTITLES', 'CLOSED-CAPTIONS'].forEach(type => { + initialize[type](type, settings); + }); + const { + mediaTypes, + mainPlaylistLoader, + tech, + vhs, + segmentLoaders: { + ['AUDIO']: audioSegmentLoader, + main: mainSegmentLoader + } + } = settings; // setup active group and track getters and change event handlers + + ['AUDIO', 'SUBTITLES'].forEach(type => { + mediaTypes[type].activeGroup = activeGroup(type, settings); + mediaTypes[type].activeTrack = activeTrack[type](type, settings); + mediaTypes[type].onGroupChanged = onGroupChanged(type, settings); + mediaTypes[type].onGroupChanging = onGroupChanging(type, settings); + mediaTypes[type].onTrackChanged = onTrackChanged(type, settings); + mediaTypes[type].getActiveGroup = getActiveGroup(type, settings); + }); // DO NOT enable the default subtitle or caption track. + // DO enable the default audio track + + const audioGroup = mediaTypes.AUDIO.activeGroup(); + if (audioGroup) { + const groupId = (audioGroup.filter(group => group.default)[0] || audioGroup[0]).id; + mediaTypes.AUDIO.tracks[groupId].enabled = true; + mediaTypes.AUDIO.onGroupChanged(); + mediaTypes.AUDIO.onTrackChanged(); + const activeAudioGroup = mediaTypes.AUDIO.getActiveGroup(); // a similar check for handling setAudio on each loader is run again each time the + // track is changed, but needs to be handled here since the track may not be considered + // changed on the first call to onTrackChanged + + if (!activeAudioGroup.playlistLoader) { + // either audio is muxed with video or the stream is audio only + mainSegmentLoader.setAudio(true); + } else { + // audio is demuxed + mainSegmentLoader.setAudio(false); + audioSegmentLoader.setAudio(true); + } + } + mainPlaylistLoader.on('mediachange', () => { + ['AUDIO', 'SUBTITLES'].forEach(type => mediaTypes[type].onGroupChanged()); + }); + mainPlaylistLoader.on('mediachanging', () => { + ['AUDIO', 'SUBTITLES'].forEach(type => mediaTypes[type].onGroupChanging()); + }); // custom audio track change event handler for usage event + + const onAudioTrackChanged = () => { + mediaTypes.AUDIO.onTrackChanged(); + tech.trigger({ + type: 'usage', + name: 'vhs-audio-change' + }); + }; + tech.audioTracks().addEventListener('change', onAudioTrackChanged); + tech.remoteTextTracks().addEventListener('change', mediaTypes.SUBTITLES.onTrackChanged); + vhs.on('dispose', () => { + tech.audioTracks().removeEventListener('change', onAudioTrackChanged); + tech.remoteTextTracks().removeEventListener('change', mediaTypes.SUBTITLES.onTrackChanged); + }); // clear existing audio tracks and add the ones we just created + + tech.clearTracks('audio'); + for (const id in mediaTypes.AUDIO.tracks) { + tech.audioTracks().addTrack(mediaTypes.AUDIO.tracks[id]); + } +}; +/** + * Creates skeleton object used to store the loaders, tracks, and utility methods for each + * media type + * + * @return {Object} + * Object to store the loaders, tracks, and utility methods for each media type + * @function createMediaTypes + */ + +const createMediaTypes = () => { + const mediaTypes = {}; + ['AUDIO', 'SUBTITLES', 'CLOSED-CAPTIONS'].forEach(type => { + mediaTypes[type] = { + groups: {}, + tracks: {}, + activePlaylistLoader: null, + activeGroup: noop, + activeTrack: noop, + getActiveGroup: noop, + onGroupChanged: noop, + onTrackChanged: noop, + lastTrack_: null, + logger_: logger(`MediaGroups[${type}]`) + }; + }); + return mediaTypes; +}; + +/** + * A utility class for setting properties and maintaining the state of the content steering manifest. + * + * Content Steering manifest format: + * VERSION: number (required) currently only version 1 is supported. + * TTL: number in seconds (optional) until the next content steering manifest reload. + * RELOAD-URI: string (optional) uri to fetch the next content steering manifest. + * SERVICE-LOCATION-PRIORITY or PATHWAY-PRIORITY a non empty array of unique string values. + * PATHWAY-CLONES: array (optional) (HLS only) pathway clone objects to copy from other playlists. + */ + +class SteeringManifest { + constructor() { + this.priority_ = []; + this.pathwayClones_ = new Map(); + } + set version(number) { + // Only version 1 is currently supported for both DASH and HLS. + if (number === 1) { + this.version_ = number; + } + } + set ttl(seconds) { + // TTL = time-to-live, default = 300 seconds. + this.ttl_ = seconds || 300; + } + set reloadUri(uri) { + if (uri) { + // reload URI can be relative to the previous reloadUri. + this.reloadUri_ = resolveUrl(this.reloadUri_, uri); + } + } + set priority(array) { + // priority must be non-empty and unique values. + if (array && array.length) { + this.priority_ = array; + } + } + set pathwayClones(array) { + // pathwayClones must be non-empty. + if (array && array.length) { + this.pathwayClones_ = new Map(array.map(clone => [clone.ID, clone])); + } + } + get version() { + return this.version_; + } + get ttl() { + return this.ttl_; + } + get reloadUri() { + return this.reloadUri_; + } + get priority() { + return this.priority_; + } + get pathwayClones() { + return this.pathwayClones_; + } +} +/** + * This class represents a content steering manifest and associated state. See both HLS and DASH specifications. + * HLS: https://developer.apple.com/streaming/HLSContentSteeringSpecification.pdf and + * https://datatracker.ietf.org/doc/draft-pantos-hls-rfc8216bis/ section 4.4.6.6. + * DASH: https://dashif.org/docs/DASH-IF-CTS-00XX-Content-Steering-Community-Review.pdf + * + * @param {function} xhr for making a network request from the browser. + * @param {function} bandwidth for fetching the current bandwidth from the main segment loader. + */ + +class ContentSteeringController extends videojs.EventTarget { + constructor(xhr, bandwidth) { + super(); + this.currentPathway = null; + this.defaultPathway = null; + this.queryBeforeStart = false; + this.availablePathways_ = new Set(); + this.steeringManifest = new SteeringManifest(); + this.proxyServerUrl_ = null; + this.manifestType_ = null; + this.ttlTimeout_ = null; + this.request_ = null; + this.currentPathwayClones = new Map(); + this.nextPathwayClones = new Map(); + this.excludedSteeringManifestURLs = new Set(); + this.logger_ = logger('Content Steering'); + this.xhr_ = xhr; + this.getBandwidth_ = bandwidth; + } + /** + * Assigns the content steering tag properties to the steering controller + * + * @param {string} baseUrl the baseURL from the main manifest for resolving the steering manifest url + * @param {Object} steeringTag the content steering tag from the main manifest + */ + + assignTagProperties(baseUrl, steeringTag) { + this.manifestType_ = steeringTag.serverUri ? 'HLS' : 'DASH'; // serverUri is HLS serverURL is DASH + + const steeringUri = steeringTag.serverUri || steeringTag.serverURL; + if (!steeringUri) { + this.logger_(`steering manifest URL is ${steeringUri}, cannot request steering manifest.`); + this.trigger('error'); + return; + } // Content steering manifests can be encoded as a data URI. We can decode, parse and return early if that's the case. + + if (steeringUri.startsWith('data:')) { + this.decodeDataUriManifest_(steeringUri.substring(steeringUri.indexOf(',') + 1)); + return; + } // reloadUri is the resolution of the main manifest URL and steering URL. + + this.steeringManifest.reloadUri = resolveUrl(baseUrl, steeringUri); // pathwayId is HLS defaultServiceLocation is DASH + + this.defaultPathway = steeringTag.pathwayId || steeringTag.defaultServiceLocation; // currently only DASH supports the following properties on <ContentSteering> tags. + + this.queryBeforeStart = steeringTag.queryBeforeStart; + this.proxyServerUrl_ = steeringTag.proxyServerURL; // trigger a steering event if we have a pathway from the content steering tag. + // this tells VHS which segment pathway to start with. + // If queryBeforeStart is true we need to wait for the steering manifest response. + + if (this.defaultPathway && !this.queryBeforeStart) { + this.trigger('content-steering'); + } + } + /** + * Requests the content steering manifest and parse the response. This should only be called after + * assignTagProperties was called with a content steering tag. + * + * @param {string} initialUri The optional uri to make the request with. + * If set, the request should be made with exactly what is passed in this variable. + * This scenario should only happen once on initalization. + */ + + requestSteeringManifest(initial) { + const reloadUri = this.steeringManifest.reloadUri; + if (!reloadUri) { + return; + } // We currently don't support passing MPD query parameters directly to the content steering URL as this requires + // ExtUrlQueryInfo tag support. See the DASH content steering spec section 8.1. + // This request URI accounts for manifest URIs that have been excluded. + + const uri = initial ? reloadUri : this.getRequestURI(reloadUri); // If there are no valid manifest URIs, we should stop content steering. + + if (!uri) { + this.logger_('No valid content steering manifest URIs. Stopping content steering.'); + this.trigger('error'); + this.dispose(); + return; + } + this.request_ = this.xhr_({ + uri, + requestType: 'content-steering-manifest' + }, (error, errorInfo) => { + if (error) { + // If the client receives HTTP 410 Gone in response to a manifest request, + // it MUST NOT issue another request for that URI for the remainder of the + // playback session. It MAY continue to use the most-recently obtained set + // of Pathways. + if (errorInfo.status === 410) { + this.logger_(`manifest request 410 ${error}.`); + this.logger_(`There will be no more content steering requests to ${uri} this session.`); + this.excludedSteeringManifestURLs.add(uri); + return; + } // If the client receives HTTP 429 Too Many Requests with a Retry-After + // header in response to a manifest request, it SHOULD wait until the time + // specified by the Retry-After header to reissue the request. + + if (errorInfo.status === 429) { + const retrySeconds = errorInfo.responseHeaders['retry-after']; + this.logger_(`manifest request 429 ${error}.`); + this.logger_(`content steering will retry in ${retrySeconds} seconds.`); + this.startTTLTimeout_(parseInt(retrySeconds, 10)); + return; + } // If the Steering Manifest cannot be loaded and parsed correctly, the + // client SHOULD continue to use the previous values and attempt to reload + // it after waiting for the previously-specified TTL (or 5 minutes if + // none). + + this.logger_(`manifest failed to load ${error}.`); + this.startTTLTimeout_(); + return; + } + const steeringManifestJson = JSON.parse(this.request_.responseText); + this.assignSteeringProperties_(steeringManifestJson); + this.startTTLTimeout_(); + }); + } + /** + * Set the proxy server URL and add the steering manifest url as a URI encoded parameter. + * + * @param {string} steeringUrl the steering manifest url + * @return the steering manifest url to a proxy server with all parameters set + */ + + setProxyServerUrl_(steeringUrl) { + const steeringUrlObject = new window$1.URL(steeringUrl); + const proxyServerUrlObject = new window$1.URL(this.proxyServerUrl_); + proxyServerUrlObject.searchParams.set('url', encodeURI(steeringUrlObject.toString())); + return this.setSteeringParams_(proxyServerUrlObject.toString()); + } + /** + * Decodes and parses the data uri encoded steering manifest + * + * @param {string} dataUri the data uri to be decoded and parsed. + */ + + decodeDataUriManifest_(dataUri) { + const steeringManifestJson = JSON.parse(window$1.atob(dataUri)); + this.assignSteeringProperties_(steeringManifestJson); + } + /** + * Set the HLS or DASH content steering manifest request query parameters. For example: + * _HLS_pathway="<CURRENT-PATHWAY-ID>" and _HLS_throughput=<THROUGHPUT> + * _DASH_pathway and _DASH_throughput + * + * @param {string} uri to add content steering server parameters to. + * @return a new uri as a string with the added steering query parameters. + */ + + setSteeringParams_(url) { + const urlObject = new window$1.URL(url); + const path = this.getPathway(); + const networkThroughput = this.getBandwidth_(); + if (path) { + const pathwayKey = `_${this.manifestType_}_pathway`; + urlObject.searchParams.set(pathwayKey, path); + } + if (networkThroughput) { + const throughputKey = `_${this.manifestType_}_throughput`; + urlObject.searchParams.set(throughputKey, networkThroughput); + } + return urlObject.toString(); + } + /** + * Assigns the current steering manifest properties and to the SteeringManifest object + * + * @param {Object} steeringJson the raw JSON steering manifest + */ + + assignSteeringProperties_(steeringJson) { + this.steeringManifest.version = steeringJson.VERSION; + if (!this.steeringManifest.version) { + this.logger_(`manifest version is ${steeringJson.VERSION}, which is not supported.`); + this.trigger('error'); + return; + } + this.steeringManifest.ttl = steeringJson.TTL; + this.steeringManifest.reloadUri = steeringJson['RELOAD-URI']; // HLS = PATHWAY-PRIORITY required. DASH = SERVICE-LOCATION-PRIORITY optional + + this.steeringManifest.priority = steeringJson['PATHWAY-PRIORITY'] || steeringJson['SERVICE-LOCATION-PRIORITY']; // Pathway clones to be created/updated in HLS. + // See section 7.2 https://datatracker.ietf.org/doc/draft-pantos-hls-rfc8216bis/ + + this.steeringManifest.pathwayClones = steeringJson['PATHWAY-CLONES']; + this.nextPathwayClones = this.steeringManifest.pathwayClones; // 1. apply first pathway from the array. + // 2. if first pathway doesn't exist in manifest, try next pathway. + // a. if all pathways are exhausted, ignore the steering manifest priority. + // 3. if segments fail from an established pathway, try all variants/renditions, then exclude the failed pathway. + // a. exclude a pathway for a minimum of the last TTL duration. Meaning, from the next steering response, + // the excluded pathway will be ignored. + // See excludePathway usage in excludePlaylist(). + // If there are no available pathways, we need to stop content steering. + + if (!this.availablePathways_.size) { + this.logger_('There are no available pathways for content steering. Ending content steering.'); + this.trigger('error'); + this.dispose(); + } + const chooseNextPathway = pathwaysByPriority => { + for (const path of pathwaysByPriority) { + if (this.availablePathways_.has(path)) { + return path; + } + } // If no pathway matches, ignore the manifest and choose the first available. + + return [...this.availablePathways_][0]; + }; + const nextPathway = chooseNextPathway(this.steeringManifest.priority); + if (this.currentPathway !== nextPathway) { + this.currentPathway = nextPathway; + this.trigger('content-steering'); + } + } + /** + * Returns the pathway to use for steering decisions + * + * @return {string} returns the current pathway or the default + */ + + getPathway() { + return this.currentPathway || this.defaultPathway; + } + /** + * Chooses the manifest request URI based on proxy URIs and server URLs. + * Also accounts for exclusion on certain manifest URIs. + * + * @param {string} reloadUri the base uri before parameters + * + * @return {string} the final URI for the request to the manifest server. + */ + + getRequestURI(reloadUri) { + if (!reloadUri) { + return null; + } + const isExcluded = uri => this.excludedSteeringManifestURLs.has(uri); + if (this.proxyServerUrl_) { + const proxyURI = this.setProxyServerUrl_(reloadUri); + if (!isExcluded(proxyURI)) { + return proxyURI; + } + } + const steeringURI = this.setSteeringParams_(reloadUri); + if (!isExcluded(steeringURI)) { + return steeringURI; + } // Return nothing if all valid manifest URIs are excluded. + + return null; + } + /** + * Start the timeout for re-requesting the steering manifest at the TTL interval. + * + * @param {number} ttl time in seconds of the timeout. Defaults to the + * ttl interval in the steering manifest + */ + + startTTLTimeout_(ttl = this.steeringManifest.ttl) { + // 300 (5 minutes) is the default value. + const ttlMS = ttl * 1000; + this.ttlTimeout_ = window$1.setTimeout(() => { + this.requestSteeringManifest(); + }, ttlMS); + } + /** + * Clear the TTL timeout if necessary. + */ + + clearTTLTimeout_() { + window$1.clearTimeout(this.ttlTimeout_); + this.ttlTimeout_ = null; + } + /** + * aborts any current steering xhr and sets the current request object to null + */ + + abort() { + if (this.request_) { + this.request_.abort(); + } + this.request_ = null; + } + /** + * aborts steering requests clears the ttl timeout and resets all properties. + */ + + dispose() { + this.off('content-steering'); + this.off('error'); + this.abort(); + this.clearTTLTimeout_(); + this.currentPathway = null; + this.defaultPathway = null; + this.queryBeforeStart = null; + this.proxyServerUrl_ = null; + this.manifestType_ = null; + this.ttlTimeout_ = null; + this.request_ = null; + this.excludedSteeringManifestURLs = new Set(); + this.availablePathways_ = new Set(); + this.steeringManifest = new SteeringManifest(); + } + /** + * adds a pathway to the available pathways set + * + * @param {string} pathway the pathway string to add + */ + + addAvailablePathway(pathway) { + if (pathway) { + this.availablePathways_.add(pathway); + } + } + /** + * Clears all pathways from the available pathways set + */ + + clearAvailablePathways() { + this.availablePathways_.clear(); + } + /** + * Removes a pathway from the available pathways set. + */ + + excludePathway(pathway) { + return this.availablePathways_.delete(pathway); + } + /** + * Checks the refreshed DASH manifest content steering tag for changes. + * + * @param {string} baseURL new steering tag on DASH manifest refresh + * @param {Object} newTag the new tag to check for changes + * @return a true or false whether the new tag has different values + */ + + didDASHTagChange(baseURL, newTag) { + return !newTag && this.steeringManifest.reloadUri || newTag && (resolveUrl(baseURL, newTag.serverURL) !== this.steeringManifest.reloadUri || newTag.defaultServiceLocation !== this.defaultPathway || newTag.queryBeforeStart !== this.queryBeforeStart || newTag.proxyServerURL !== this.proxyServerUrl_); + } + getAvailablePathways() { + return this.availablePathways_; + } +} + +/** + * @file playlist-controller.js + */ +const ABORT_EARLY_EXCLUSION_SECONDS = 10; +let Vhs$1; // SegmentLoader stats that need to have each loader's +// values summed to calculate the final value + +const loaderStats = ['mediaRequests', 'mediaRequestsAborted', 'mediaRequestsTimedout', 'mediaRequestsErrored', 'mediaTransferDuration', 'mediaBytesTransferred', 'mediaAppends']; +const sumLoaderStat = function (stat) { + return this.audioSegmentLoader_[stat] + this.mainSegmentLoader_[stat]; +}; +const shouldSwitchToMedia = function ({ + currentPlaylist, + buffered, + currentTime, + nextPlaylist, + bufferLowWaterLine, + bufferHighWaterLine, + duration, + bufferBasedABR, + log +}) { + // we have no other playlist to switch to + if (!nextPlaylist) { + videojs.log.warn('We received no playlist to switch to. Please check your stream.'); + return false; + } + const sharedLogLine = `allowing switch ${currentPlaylist && currentPlaylist.id || 'null'} -> ${nextPlaylist.id}`; + if (!currentPlaylist) { + log(`${sharedLogLine} as current playlist is not set`); + return true; + } // no need to switch if playlist is the same + + if (nextPlaylist.id === currentPlaylist.id) { + return false; + } // determine if current time is in a buffered range. + + const isBuffered = Boolean(findRange(buffered, currentTime).length); // If the playlist is live, then we want to not take low water line into account. + // This is because in LIVE, the player plays 3 segments from the end of the + // playlist, and if `BUFFER_LOW_WATER_LINE` is greater than the duration availble + // in those segments, a viewer will never experience a rendition upswitch. + + if (!currentPlaylist.endList) { + // For LLHLS live streams, don't switch renditions before playback has started, as it almost + // doubles the time to first playback. + if (!isBuffered && typeof currentPlaylist.partTargetDuration === 'number') { + log(`not ${sharedLogLine} as current playlist is live llhls, but currentTime isn't in buffered.`); + return false; + } + log(`${sharedLogLine} as current playlist is live`); + return true; + } + const forwardBuffer = timeAheadOf(buffered, currentTime); + const maxBufferLowWaterLine = bufferBasedABR ? Config.EXPERIMENTAL_MAX_BUFFER_LOW_WATER_LINE : Config.MAX_BUFFER_LOW_WATER_LINE; // For the same reason as LIVE, we ignore the low water line when the VOD + // duration is below the max potential low water line + + if (duration < maxBufferLowWaterLine) { + log(`${sharedLogLine} as duration < max low water line (${duration} < ${maxBufferLowWaterLine})`); + return true; + } + const nextBandwidth = nextPlaylist.attributes.BANDWIDTH; + const currBandwidth = currentPlaylist.attributes.BANDWIDTH; // when switching down, if our buffer is lower than the high water line, + // we can switch down + + if (nextBandwidth < currBandwidth && (!bufferBasedABR || forwardBuffer < bufferHighWaterLine)) { + let logLine = `${sharedLogLine} as next bandwidth < current bandwidth (${nextBandwidth} < ${currBandwidth})`; + if (bufferBasedABR) { + logLine += ` and forwardBuffer < bufferHighWaterLine (${forwardBuffer} < ${bufferHighWaterLine})`; + } + log(logLine); + return true; + } // and if our buffer is higher than the low water line, + // we can switch up + + if ((!bufferBasedABR || nextBandwidth > currBandwidth) && forwardBuffer >= bufferLowWaterLine) { + let logLine = `${sharedLogLine} as forwardBuffer >= bufferLowWaterLine (${forwardBuffer} >= ${bufferLowWaterLine})`; + if (bufferBasedABR) { + logLine += ` and next bandwidth > current bandwidth (${nextBandwidth} > ${currBandwidth})`; + } + log(logLine); + return true; + } + log(`not ${sharedLogLine} as no switching criteria met`); + return false; +}; +/** + * the main playlist controller controller all interactons + * between playlists and segmentloaders. At this time this mainly + * involves a main playlist and a series of audio playlists + * if they are available + * + * @class PlaylistController + * @extends videojs.EventTarget + */ + +class PlaylistController extends videojs.EventTarget { + constructor(options) { + super(); + const { + src, + withCredentials, + tech, + bandwidth, + externVhs, + useCueTags, + playlistExclusionDuration, + enableLowInitialPlaylist, + sourceType, + cacheEncryptionKeys, + bufferBasedABR, + leastPixelDiffSelector, + captionServices + } = options; + if (!src) { + throw new Error('A non-empty playlist URL or JSON manifest string is required'); + } + let { + maxPlaylistRetries + } = options; + if (maxPlaylistRetries === null || typeof maxPlaylistRetries === 'undefined') { + maxPlaylistRetries = Infinity; + } + Vhs$1 = externVhs; + this.bufferBasedABR = Boolean(bufferBasedABR); + this.leastPixelDiffSelector = Boolean(leastPixelDiffSelector); + this.withCredentials = withCredentials; + this.tech_ = tech; + this.vhs_ = tech.vhs; + this.sourceType_ = sourceType; + this.useCueTags_ = useCueTags; + this.playlistExclusionDuration = playlistExclusionDuration; + this.maxPlaylistRetries = maxPlaylistRetries; + this.enableLowInitialPlaylist = enableLowInitialPlaylist; + if (this.useCueTags_) { + this.cueTagsTrack_ = this.tech_.addTextTrack('metadata', 'ad-cues'); + this.cueTagsTrack_.inBandMetadataTrackDispatchType = ''; + } + this.requestOptions_ = { + withCredentials, + maxPlaylistRetries, + timeout: null + }; + this.on('error', this.pauseLoading); + this.mediaTypes_ = createMediaTypes(); + this.mediaSource = new window$1.MediaSource(); + this.handleDurationChange_ = this.handleDurationChange_.bind(this); + this.handleSourceOpen_ = this.handleSourceOpen_.bind(this); + this.handleSourceEnded_ = this.handleSourceEnded_.bind(this); + this.mediaSource.addEventListener('durationchange', this.handleDurationChange_); // load the media source into the player + + this.mediaSource.addEventListener('sourceopen', this.handleSourceOpen_); + this.mediaSource.addEventListener('sourceended', this.handleSourceEnded_); // we don't have to handle sourceclose since dispose will handle termination of + // everything, and the MediaSource should not be detached without a proper disposal + + this.seekable_ = createTimeRanges(); + this.hasPlayed_ = false; + this.syncController_ = new SyncController(options); + this.segmentMetadataTrack_ = tech.addRemoteTextTrack({ + kind: 'metadata', + label: 'segment-metadata' + }, false).track; + this.decrypter_ = new Decrypter(); + this.sourceUpdater_ = new SourceUpdater(this.mediaSource); + this.inbandTextTracks_ = {}; + this.timelineChangeController_ = new TimelineChangeController(); + this.keyStatusMap_ = new Map(); + const segmentLoaderSettings = { + vhs: this.vhs_, + parse708captions: options.parse708captions, + useDtsForTimestampOffset: options.useDtsForTimestampOffset, + captionServices, + mediaSource: this.mediaSource, + currentTime: this.tech_.currentTime.bind(this.tech_), + seekable: () => this.seekable(), + seeking: () => this.tech_.seeking(), + duration: () => this.duration(), + hasPlayed: () => this.hasPlayed_, + goalBufferLength: () => this.goalBufferLength(), + bandwidth, + syncController: this.syncController_, + decrypter: this.decrypter_, + sourceType: this.sourceType_, + inbandTextTracks: this.inbandTextTracks_, + cacheEncryptionKeys, + sourceUpdater: this.sourceUpdater_, + timelineChangeController: this.timelineChangeController_, + exactManifestTimings: options.exactManifestTimings, + addMetadataToTextTrack: this.addMetadataToTextTrack.bind(this) + }; // The source type check not only determines whether a special DASH playlist loader + // should be used, but also covers the case where the provided src is a vhs-json + // manifest object (instead of a URL). In the case of vhs-json, the default + // PlaylistLoader should be used. + + this.mainPlaylistLoader_ = this.sourceType_ === 'dash' ? new DashPlaylistLoader(src, this.vhs_, merge(this.requestOptions_, { + addMetadataToTextTrack: this.addMetadataToTextTrack.bind(this) + })) : new PlaylistLoader(src, this.vhs_, merge(this.requestOptions_, { + addDateRangesToTextTrack: this.addDateRangesToTextTrack_.bind(this) + })); + this.setupMainPlaylistLoaderListeners_(); // setup segment loaders + // combined audio/video or just video when alternate audio track is selected + + this.mainSegmentLoader_ = new SegmentLoader(merge(segmentLoaderSettings, { + segmentMetadataTrack: this.segmentMetadataTrack_, + loaderType: 'main' + }), options); // alternate audio track + + this.audioSegmentLoader_ = new SegmentLoader(merge(segmentLoaderSettings, { + loaderType: 'audio' + }), options); + this.subtitleSegmentLoader_ = new VTTSegmentLoader(merge(segmentLoaderSettings, { + loaderType: 'vtt', + featuresNativeTextTracks: this.tech_.featuresNativeTextTracks, + loadVttJs: () => new Promise((resolve, reject) => { + function onLoad() { + tech.off('vttjserror', onError); + resolve(); + } + function onError() { + tech.off('vttjsloaded', onLoad); + reject(); + } + tech.one('vttjsloaded', onLoad); + tech.one('vttjserror', onError); // safe to call multiple times, script will be loaded only once: + + tech.addWebVttScript_(); + }) + }), options); + const getBandwidth = () => { + return this.mainSegmentLoader_.bandwidth; + }; + this.contentSteeringController_ = new ContentSteeringController(this.vhs_.xhr, getBandwidth); + this.setupSegmentLoaderListeners_(); + if (this.bufferBasedABR) { + this.mainPlaylistLoader_.one('loadedplaylist', () => this.startABRTimer_()); + this.tech_.on('pause', () => this.stopABRTimer_()); + this.tech_.on('play', () => this.startABRTimer_()); + } // Create SegmentLoader stat-getters + // mediaRequests_ + // mediaRequestsAborted_ + // mediaRequestsTimedout_ + // mediaRequestsErrored_ + // mediaTransferDuration_ + // mediaBytesTransferred_ + // mediaAppends_ + + loaderStats.forEach(stat => { + this[stat + '_'] = sumLoaderStat.bind(this, stat); + }); + this.logger_ = logger('pc'); + this.triggeredFmp4Usage = false; + if (this.tech_.preload() === 'none') { + this.loadOnPlay_ = () => { + this.loadOnPlay_ = null; + this.mainPlaylistLoader_.load(); + }; + this.tech_.one('play', this.loadOnPlay_); + } else { + this.mainPlaylistLoader_.load(); + } + this.timeToLoadedData__ = -1; + this.mainAppendsToLoadedData__ = -1; + this.audioAppendsToLoadedData__ = -1; + const event = this.tech_.preload() === 'none' ? 'play' : 'loadstart'; // start the first frame timer on loadstart or play (for preload none) + + this.tech_.one(event, () => { + const timeToLoadedDataStart = Date.now(); + this.tech_.one('loadeddata', () => { + this.timeToLoadedData__ = Date.now() - timeToLoadedDataStart; + this.mainAppendsToLoadedData__ = this.mainSegmentLoader_.mediaAppends; + this.audioAppendsToLoadedData__ = this.audioSegmentLoader_.mediaAppends; + }); + }); + } + mainAppendsToLoadedData_() { + return this.mainAppendsToLoadedData__; + } + audioAppendsToLoadedData_() { + return this.audioAppendsToLoadedData__; + } + appendsToLoadedData_() { + const main = this.mainAppendsToLoadedData_(); + const audio = this.audioAppendsToLoadedData_(); + if (main === -1 || audio === -1) { + return -1; + } + return main + audio; + } + timeToLoadedData_() { + return this.timeToLoadedData__; + } + /** + * Run selectPlaylist and switch to the new playlist if we should + * + * @param {string} [reason=abr] a reason for why the ABR check is made + * @private + */ + + checkABR_(reason = 'abr') { + const nextPlaylist = this.selectPlaylist(); + if (nextPlaylist && this.shouldSwitchToMedia_(nextPlaylist)) { + this.switchMedia_(nextPlaylist, reason); + } + } + switchMedia_(playlist, cause, delay) { + const oldMedia = this.media(); + const oldId = oldMedia && (oldMedia.id || oldMedia.uri); + const newId = playlist && (playlist.id || playlist.uri); + if (oldId && oldId !== newId) { + this.logger_(`switch media ${oldId} -> ${newId} from ${cause}`); + this.tech_.trigger({ + type: 'usage', + name: `vhs-rendition-change-${cause}` + }); + } + this.mainPlaylistLoader_.media(playlist, delay); + } + /** + * A function that ensures we switch our playlists inside of `mediaTypes` + * to match the current `serviceLocation` provided by the contentSteering controller. + * We want to check media types of `AUDIO`, `SUBTITLES`, and `CLOSED-CAPTIONS`. + * + * This should only be called on a DASH playback scenario while using content steering. + * This is necessary due to differences in how media in HLS manifests are generally tied to + * a video playlist, where in DASH that is not always the case. + */ + + switchMediaForDASHContentSteering_() { + ['AUDIO', 'SUBTITLES', 'CLOSED-CAPTIONS'].forEach(type => { + const mediaType = this.mediaTypes_[type]; + const activeGroup = mediaType ? mediaType.activeGroup() : null; + const pathway = this.contentSteeringController_.getPathway(); + if (activeGroup && pathway) { + // activeGroup can be an array or a single group + const mediaPlaylists = activeGroup.length ? activeGroup[0].playlists : activeGroup.playlists; + const dashMediaPlaylists = mediaPlaylists.filter(p => p.attributes.serviceLocation === pathway); // Switch the current active playlist to the correct CDN + + if (dashMediaPlaylists.length) { + this.mediaTypes_[type].activePlaylistLoader.media(dashMediaPlaylists[0]); + } + } + }); + } + /** + * Start a timer that periodically calls checkABR_ + * + * @private + */ + + startABRTimer_() { + this.stopABRTimer_(); + this.abrTimer_ = window$1.setInterval(() => this.checkABR_(), 250); + } + /** + * Stop the timer that periodically calls checkABR_ + * + * @private + */ + + stopABRTimer_() { + // if we're scrubbing, we don't need to pause. + // This getter will be added to Video.js in version 7.11. + if (this.tech_.scrubbing && this.tech_.scrubbing()) { + return; + } + window$1.clearInterval(this.abrTimer_); + this.abrTimer_ = null; + } + /** + * Get a list of playlists for the currently selected audio playlist + * + * @return {Array} the array of audio playlists + */ + + getAudioTrackPlaylists_() { + const main = this.main(); + const defaultPlaylists = main && main.playlists || []; // if we don't have any audio groups then we can only + // assume that the audio tracks are contained in main + // playlist array, use that or an empty array. + + if (!main || !main.mediaGroups || !main.mediaGroups.AUDIO) { + return defaultPlaylists; + } + const AUDIO = main.mediaGroups.AUDIO; + const groupKeys = Object.keys(AUDIO); + let track; // get the current active track + + if (Object.keys(this.mediaTypes_.AUDIO.groups).length) { + track = this.mediaTypes_.AUDIO.activeTrack(); // or get the default track from main if mediaTypes_ isn't setup yet + } else { + // default group is `main` or just the first group. + const defaultGroup = AUDIO.main || groupKeys.length && AUDIO[groupKeys[0]]; + for (const label in defaultGroup) { + if (defaultGroup[label].default) { + track = { + label + }; + break; + } + } + } // no active track no playlists. + + if (!track) { + return defaultPlaylists; + } + const playlists = []; // get all of the playlists that are possible for the + // active track. + + for (const group in AUDIO) { + if (AUDIO[group][track.label]) { + const properties = AUDIO[group][track.label]; + if (properties.playlists && properties.playlists.length) { + playlists.push.apply(playlists, properties.playlists); + } else if (properties.uri) { + playlists.push(properties); + } else if (main.playlists.length) { + // if an audio group does not have a uri + // see if we have main playlists that use it as a group. + // if we do then add those to the playlists list. + for (let i = 0; i < main.playlists.length; i++) { + const playlist = main.playlists[i]; + if (playlist.attributes && playlist.attributes.AUDIO && playlist.attributes.AUDIO === group) { + playlists.push(playlist); + } + } + } + } + } + if (!playlists.length) { + return defaultPlaylists; + } + return playlists; + } + /** + * Register event handlers on the main playlist loader. A helper + * function for construction time. + * + * @private + */ + + setupMainPlaylistLoaderListeners_() { + this.mainPlaylistLoader_.on('loadedmetadata', () => { + const media = this.mainPlaylistLoader_.media(); + const requestTimeout = media.targetDuration * 1.5 * 1000; // If we don't have any more available playlists, we don't want to + // timeout the request. + + if (isLowestEnabledRendition(this.mainPlaylistLoader_.main, this.mainPlaylistLoader_.media())) { + this.requestOptions_.timeout = 0; + } else { + this.requestOptions_.timeout = requestTimeout; + } // if this isn't a live video and preload permits, start + // downloading segments + + if (media.endList && this.tech_.preload() !== 'none') { + this.mainSegmentLoader_.playlist(media, this.requestOptions_); + this.mainSegmentLoader_.load(); + } + setupMediaGroups({ + sourceType: this.sourceType_, + segmentLoaders: { + AUDIO: this.audioSegmentLoader_, + SUBTITLES: this.subtitleSegmentLoader_, + main: this.mainSegmentLoader_ + }, + tech: this.tech_, + requestOptions: this.requestOptions_, + mainPlaylistLoader: this.mainPlaylistLoader_, + vhs: this.vhs_, + main: this.main(), + mediaTypes: this.mediaTypes_, + excludePlaylist: this.excludePlaylist.bind(this) + }); + this.triggerPresenceUsage_(this.main(), media); + this.setupFirstPlay(); + if (!this.mediaTypes_.AUDIO.activePlaylistLoader || this.mediaTypes_.AUDIO.activePlaylistLoader.media()) { + this.trigger('selectedinitialmedia'); + } else { + // We must wait for the active audio playlist loader to + // finish setting up before triggering this event so the + // representations API and EME setup is correct + this.mediaTypes_.AUDIO.activePlaylistLoader.one('loadedmetadata', () => { + this.trigger('selectedinitialmedia'); + }); + } + }); + this.mainPlaylistLoader_.on('loadedplaylist', () => { + if (this.loadOnPlay_) { + this.tech_.off('play', this.loadOnPlay_); + } + let updatedPlaylist = this.mainPlaylistLoader_.media(); + if (!updatedPlaylist) { + // Add content steering listeners on first load and init. + this.attachContentSteeringListeners_(); + this.initContentSteeringController_(); // exclude any variants that are not supported by the browser before selecting + // an initial media as the playlist selectors do not consider browser support + + this.excludeUnsupportedVariants_(); + let selectedMedia; + if (this.enableLowInitialPlaylist) { + selectedMedia = this.selectInitialPlaylist(); + } + if (!selectedMedia) { + selectedMedia = this.selectPlaylist(); + } + if (!selectedMedia || !this.shouldSwitchToMedia_(selectedMedia)) { + return; + } + this.initialMedia_ = selectedMedia; + this.switchMedia_(this.initialMedia_, 'initial'); // Under the standard case where a source URL is provided, loadedplaylist will + // fire again since the playlist will be requested. In the case of vhs-json + // (where the manifest object is provided as the source), when the media + // playlist's `segments` list is already available, a media playlist won't be + // requested, and loadedplaylist won't fire again, so the playlist handler must be + // called on its own here. + + const haveJsonSource = this.sourceType_ === 'vhs-json' && this.initialMedia_.segments; + if (!haveJsonSource) { + return; + } + updatedPlaylist = this.initialMedia_; + } + this.handleUpdatedMediaPlaylist(updatedPlaylist); + }); + this.mainPlaylistLoader_.on('error', () => { + const error = this.mainPlaylistLoader_.error; + this.excludePlaylist({ + playlistToExclude: error.playlist, + error + }); + }); + this.mainPlaylistLoader_.on('mediachanging', () => { + this.mainSegmentLoader_.abort(); + this.mainSegmentLoader_.pause(); + }); + this.mainPlaylistLoader_.on('mediachange', () => { + const media = this.mainPlaylistLoader_.media(); + const requestTimeout = media.targetDuration * 1.5 * 1000; // If we don't have any more available playlists, we don't want to + // timeout the request. + + if (isLowestEnabledRendition(this.mainPlaylistLoader_.main, this.mainPlaylistLoader_.media())) { + this.requestOptions_.timeout = 0; + } else { + this.requestOptions_.timeout = requestTimeout; + } + if (this.sourceType_ === 'dash') { + // we don't want to re-request the same hls playlist right after it was changed + this.mainPlaylistLoader_.load(); + } // TODO: Create a new event on the PlaylistLoader that signals + // that the segments have changed in some way and use that to + // update the SegmentLoader instead of doing it twice here and + // on `loadedplaylist` + + this.mainSegmentLoader_.pause(); + this.mainSegmentLoader_.playlist(media, this.requestOptions_); + if (this.waitingForFastQualityPlaylistReceived_) { + this.runFastQualitySwitch_(); + } else { + this.mainSegmentLoader_.load(); + } + this.tech_.trigger({ + type: 'mediachange', + bubbles: true + }); + }); + this.mainPlaylistLoader_.on('playlistunchanged', () => { + const updatedPlaylist = this.mainPlaylistLoader_.media(); // ignore unchanged playlists that have already been + // excluded for not-changing. We likely just have a really slowly updating + // playlist. + + if (updatedPlaylist.lastExcludeReason_ === 'playlist-unchanged') { + return; + } + const playlistOutdated = this.stuckAtPlaylistEnd_(updatedPlaylist); + if (playlistOutdated) { + // Playlist has stopped updating and we're stuck at its end. Try to + // exclude it and switch to another playlist in the hope that that + // one is updating (and give the player a chance to re-adjust to the + // safe live point). + this.excludePlaylist({ + error: { + message: 'Playlist no longer updating.', + reason: 'playlist-unchanged' + } + }); // useful for monitoring QoS + + this.tech_.trigger('playliststuck'); + } + }); + this.mainPlaylistLoader_.on('renditiondisabled', () => { + this.tech_.trigger({ + type: 'usage', + name: 'vhs-rendition-disabled' + }); + }); + this.mainPlaylistLoader_.on('renditionenabled', () => { + this.tech_.trigger({ + type: 'usage', + name: 'vhs-rendition-enabled' + }); + }); + } + /** + * Given an updated media playlist (whether it was loaded for the first time, or + * refreshed for live playlists), update any relevant properties and state to reflect + * changes in the media that should be accounted for (e.g., cues and duration). + * + * @param {Object} updatedPlaylist the updated media playlist object + * + * @private + */ + + handleUpdatedMediaPlaylist(updatedPlaylist) { + if (this.useCueTags_) { + this.updateAdCues_(updatedPlaylist); + } // TODO: Create a new event on the PlaylistLoader that signals + // that the segments have changed in some way and use that to + // update the SegmentLoader instead of doing it twice here and + // on `mediachange` + + this.mainSegmentLoader_.pause(); + this.mainSegmentLoader_.playlist(updatedPlaylist, this.requestOptions_); + if (this.waitingForFastQualityPlaylistReceived_) { + this.runFastQualitySwitch_(); + } + this.updateDuration(!updatedPlaylist.endList); // If the player isn't paused, ensure that the segment loader is running, + // as it is possible that it was temporarily stopped while waiting for + // a playlist (e.g., in case the playlist errored and we re-requested it). + + if (!this.tech_.paused()) { + this.mainSegmentLoader_.load(); + if (this.audioSegmentLoader_) { + this.audioSegmentLoader_.load(); + } + } + } + /** + * A helper function for triggerring presence usage events once per source + * + * @private + */ + + triggerPresenceUsage_(main, media) { + const mediaGroups = main.mediaGroups || {}; + let defaultDemuxed = true; + const audioGroupKeys = Object.keys(mediaGroups.AUDIO); + for (const mediaGroup in mediaGroups.AUDIO) { + for (const label in mediaGroups.AUDIO[mediaGroup]) { + const properties = mediaGroups.AUDIO[mediaGroup][label]; + if (!properties.uri) { + defaultDemuxed = false; + } + } + } + if (defaultDemuxed) { + this.tech_.trigger({ + type: 'usage', + name: 'vhs-demuxed' + }); + } + if (Object.keys(mediaGroups.SUBTITLES).length) { + this.tech_.trigger({ + type: 'usage', + name: 'vhs-webvtt' + }); + } + if (Vhs$1.Playlist.isAes(media)) { + this.tech_.trigger({ + type: 'usage', + name: 'vhs-aes' + }); + } + if (audioGroupKeys.length && Object.keys(mediaGroups.AUDIO[audioGroupKeys[0]]).length > 1) { + this.tech_.trigger({ + type: 'usage', + name: 'vhs-alternate-audio' + }); + } + if (this.useCueTags_) { + this.tech_.trigger({ + type: 'usage', + name: 'vhs-playlist-cue-tags' + }); + } + } + shouldSwitchToMedia_(nextPlaylist) { + const currentPlaylist = this.mainPlaylistLoader_.media() || this.mainPlaylistLoader_.pendingMedia_; + const currentTime = this.tech_.currentTime(); + const bufferLowWaterLine = this.bufferLowWaterLine(); + const bufferHighWaterLine = this.bufferHighWaterLine(); + const buffered = this.tech_.buffered(); + return shouldSwitchToMedia({ + buffered, + currentTime, + currentPlaylist, + nextPlaylist, + bufferLowWaterLine, + bufferHighWaterLine, + duration: this.duration(), + bufferBasedABR: this.bufferBasedABR, + log: this.logger_ + }); + } + /** + * Register event handlers on the segment loaders. A helper function + * for construction time. + * + * @private + */ + + setupSegmentLoaderListeners_() { + this.mainSegmentLoader_.on('bandwidthupdate', () => { + // Whether or not buffer based ABR or another ABR is used, on a bandwidth change it's + // useful to check to see if a rendition switch should be made. + this.checkABR_('bandwidthupdate'); + this.tech_.trigger('bandwidthupdate'); + }); + this.mainSegmentLoader_.on('timeout', () => { + if (this.bufferBasedABR) { + // If a rendition change is needed, then it would've be done on `bandwidthupdate`. + // Here the only consideration is that for buffer based ABR there's no guarantee + // of an immediate switch (since the bandwidth is averaged with a timeout + // bandwidth value of 1), so force a load on the segment loader to keep it going. + this.mainSegmentLoader_.load(); + } + }); // `progress` events are not reliable enough of a bandwidth measure to trigger buffer + // based ABR. + + if (!this.bufferBasedABR) { + this.mainSegmentLoader_.on('progress', () => { + this.trigger('progress'); + }); + } + this.mainSegmentLoader_.on('error', () => { + const error = this.mainSegmentLoader_.error(); + this.excludePlaylist({ + playlistToExclude: error.playlist, + error + }); + }); + this.mainSegmentLoader_.on('appenderror', () => { + this.error = this.mainSegmentLoader_.error_; + this.trigger('error'); + }); + this.mainSegmentLoader_.on('syncinfoupdate', () => { + this.onSyncInfoUpdate_(); + }); + this.mainSegmentLoader_.on('timestampoffset', () => { + this.tech_.trigger({ + type: 'usage', + name: 'vhs-timestamp-offset' + }); + }); + this.audioSegmentLoader_.on('syncinfoupdate', () => { + this.onSyncInfoUpdate_(); + }); + this.audioSegmentLoader_.on('appenderror', () => { + this.error = this.audioSegmentLoader_.error_; + this.trigger('error'); + }); + this.mainSegmentLoader_.on('ended', () => { + this.logger_('main segment loader ended'); + this.onEndOfStream(); + }); + this.mainSegmentLoader_.on('earlyabort', event => { + // never try to early abort with the new ABR algorithm + if (this.bufferBasedABR) { + return; + } + this.delegateLoaders_('all', ['abort']); + this.excludePlaylist({ + error: { + message: 'Aborted early because there isn\'t enough bandwidth to complete ' + 'the request without rebuffering.' + }, + playlistExclusionDuration: ABORT_EARLY_EXCLUSION_SECONDS + }); + }); + const updateCodecs = () => { + if (!this.sourceUpdater_.hasCreatedSourceBuffers()) { + return this.tryToCreateSourceBuffers_(); + } + const codecs = this.getCodecsOrExclude_(); // no codecs means that the playlist was excluded + + if (!codecs) { + return; + } + this.sourceUpdater_.addOrChangeSourceBuffers(codecs); + }; + this.mainSegmentLoader_.on('trackinfo', updateCodecs); + this.audioSegmentLoader_.on('trackinfo', updateCodecs); + this.mainSegmentLoader_.on('fmp4', () => { + if (!this.triggeredFmp4Usage) { + this.tech_.trigger({ + type: 'usage', + name: 'vhs-fmp4' + }); + this.triggeredFmp4Usage = true; + } + }); + this.audioSegmentLoader_.on('fmp4', () => { + if (!this.triggeredFmp4Usage) { + this.tech_.trigger({ + type: 'usage', + name: 'vhs-fmp4' + }); + this.triggeredFmp4Usage = true; + } + }); + this.audioSegmentLoader_.on('ended', () => { + this.logger_('audioSegmentLoader ended'); + this.onEndOfStream(); + }); + } + mediaSecondsLoaded_() { + return Math.max(this.audioSegmentLoader_.mediaSecondsLoaded + this.mainSegmentLoader_.mediaSecondsLoaded); + } + /** + * Call load on our SegmentLoaders + */ + + load() { + this.mainSegmentLoader_.load(); + if (this.mediaTypes_.AUDIO.activePlaylistLoader) { + this.audioSegmentLoader_.load(); + } + if (this.mediaTypes_.SUBTITLES.activePlaylistLoader) { + this.subtitleSegmentLoader_.load(); + } + } + /** + * Re-tune playback quality level for the current player + * conditions. This method will perform destructive actions like removing + * already buffered content in order to readjust the currently active + * playlist quickly. This is good for manual quality changes + * + * @private + */ + + fastQualityChange_(media = this.selectPlaylist()) { + if (media && media === this.mainPlaylistLoader_.media()) { + this.logger_('skipping fastQualityChange because new media is same as old'); + return; + } + this.switchMedia_(media, 'fast-quality'); // we would like to avoid race condition when we call fastQuality, + // reset everything and start loading segments from prev segments instead of new because new playlist is not received yet + + this.waitingForFastQualityPlaylistReceived_ = true; + } + runFastQualitySwitch_() { + this.waitingForFastQualityPlaylistReceived_ = false; // Delete all buffered data to allow an immediate quality switch, then seek to give + // the browser a kick to remove any cached frames from the previous rendtion (.04 seconds + // ahead was roughly the minimum that will accomplish this across a variety of content + // in IE and Edge, but seeking in place is sufficient on all other browsers) + // Edge/IE bug: https://developer.microsoft.com/en-us/microsoft-edge/platform/issues/14600375/ + // Chrome bug: https://bugs.chromium.org/p/chromium/issues/detail?id=651904 + + this.mainSegmentLoader_.pause(); + this.mainSegmentLoader_.resetEverything(() => { + this.tech_.setCurrentTime(this.tech_.currentTime()); + }); // don't need to reset audio as it is reset when media changes + } + /** + * Begin playback. + */ + + play() { + if (this.setupFirstPlay()) { + return; + } + if (this.tech_.ended()) { + this.tech_.setCurrentTime(0); + } + if (this.hasPlayed_) { + this.load(); + } + const seekable = this.tech_.seekable(); // if the viewer has paused and we fell out of the live window, + // seek forward to the live point + + if (this.tech_.duration() === Infinity) { + if (this.tech_.currentTime() < seekable.start(0)) { + return this.tech_.setCurrentTime(seekable.end(seekable.length - 1)); + } + } + } + /** + * Seek to the latest media position if this is a live video and the + * player and video are loaded and initialized. + */ + + setupFirstPlay() { + const media = this.mainPlaylistLoader_.media(); // Check that everything is ready to begin buffering for the first call to play + // If 1) there is no active media + // 2) the player is paused + // 3) the first play has already been setup + // then exit early + + if (!media || this.tech_.paused() || this.hasPlayed_) { + return false; + } // when the video is a live stream and/or has a start time + + if (!media.endList || media.start) { + const seekable = this.seekable(); + if (!seekable.length) { + // without a seekable range, the player cannot seek to begin buffering at the + // live or start point + return false; + } + const seekableEnd = seekable.end(0); + let startPoint = seekableEnd; + if (media.start) { + const offset = media.start.timeOffset; + if (offset < 0) { + startPoint = Math.max(seekableEnd + offset, seekable.start(0)); + } else { + startPoint = Math.min(seekableEnd, offset); + } + } // trigger firstplay to inform the source handler to ignore the next seek event + + this.trigger('firstplay'); // seek to the live point + + this.tech_.setCurrentTime(startPoint); + } + this.hasPlayed_ = true; // we can begin loading now that everything is ready + + this.load(); + return true; + } + /** + * handle the sourceopen event on the MediaSource + * + * @private + */ + + handleSourceOpen_() { + // Only attempt to create the source buffer if none already exist. + // handleSourceOpen is also called when we are "re-opening" a source buffer + // after `endOfStream` has been called (in response to a seek for instance) + this.tryToCreateSourceBuffers_(); // if autoplay is enabled, begin playback. This is duplicative of + // code in video.js but is required because play() must be invoked + // *after* the media source has opened. + + if (this.tech_.autoplay()) { + const playPromise = this.tech_.play(); // Catch/silence error when a pause interrupts a play request + // on browsers which return a promise + + if (typeof playPromise !== 'undefined' && typeof playPromise.then === 'function') { + playPromise.then(null, e => {}); + } + } + this.trigger('sourceopen'); + } + /** + * handle the sourceended event on the MediaSource + * + * @private + */ + + handleSourceEnded_() { + if (!this.inbandTextTracks_.metadataTrack_) { + return; + } + const cues = this.inbandTextTracks_.metadataTrack_.cues; + if (!cues || !cues.length) { + return; + } + const duration = this.duration(); + cues[cues.length - 1].endTime = isNaN(duration) || Math.abs(duration) === Infinity ? Number.MAX_VALUE : duration; + } + /** + * handle the durationchange event on the MediaSource + * + * @private + */ + + handleDurationChange_() { + this.tech_.trigger('durationchange'); + } + /** + * Calls endOfStream on the media source when all active stream types have called + * endOfStream + * + * @param {string} streamType + * Stream type of the segment loader that called endOfStream + * @private + */ + + onEndOfStream() { + let isEndOfStream = this.mainSegmentLoader_.ended_; + if (this.mediaTypes_.AUDIO.activePlaylistLoader) { + const mainMediaInfo = this.mainSegmentLoader_.getCurrentMediaInfo_(); // if the audio playlist loader exists, then alternate audio is active + + if (!mainMediaInfo || mainMediaInfo.hasVideo) { + // if we do not know if the main segment loader contains video yet or if we + // definitively know the main segment loader contains video, then we need to wait + // for both main and audio segment loaders to call endOfStream + isEndOfStream = isEndOfStream && this.audioSegmentLoader_.ended_; + } else { + // otherwise just rely on the audio loader + isEndOfStream = this.audioSegmentLoader_.ended_; + } + } + if (!isEndOfStream) { + return; + } + this.stopABRTimer_(); + this.sourceUpdater_.endOfStream(); + } + /** + * Check if a playlist has stopped being updated + * + * @param {Object} playlist the media playlist object + * @return {boolean} whether the playlist has stopped being updated or not + */ + + stuckAtPlaylistEnd_(playlist) { + const seekable = this.seekable(); + if (!seekable.length) { + // playlist doesn't have enough information to determine whether we are stuck + return false; + } + const expired = this.syncController_.getExpiredTime(playlist, this.duration()); + if (expired === null) { + return false; + } // does not use the safe live end to calculate playlist end, since we + // don't want to say we are stuck while there is still content + + const absolutePlaylistEnd = Vhs$1.Playlist.playlistEnd(playlist, expired); + const currentTime = this.tech_.currentTime(); + const buffered = this.tech_.buffered(); + if (!buffered.length) { + // return true if the playhead reached the absolute end of the playlist + return absolutePlaylistEnd - currentTime <= SAFE_TIME_DELTA; + } + const bufferedEnd = buffered.end(buffered.length - 1); // return true if there is too little buffer left and buffer has reached absolute + // end of playlist + + return bufferedEnd - currentTime <= SAFE_TIME_DELTA && absolutePlaylistEnd - bufferedEnd <= SAFE_TIME_DELTA; + } + /** + * Exclude a playlist for a set amount of time, making it unavailable for selection by + * the rendition selection algorithm, then force a new playlist (rendition) selection. + * + * @param {Object=} playlistToExclude + * the playlist to exclude, defaults to the currently selected playlist + * @param {Object=} error + * an optional error + * @param {number=} playlistExclusionDuration + * an optional number of seconds to exclude the playlist + */ + + excludePlaylist({ + playlistToExclude = this.mainPlaylistLoader_.media(), + error = {}, + playlistExclusionDuration + }) { + // If the `error` was generated by the playlist loader, it will contain + // the playlist we were trying to load (but failed) and that should be + // excluded instead of the currently selected playlist which is likely + // out-of-date in this scenario + playlistToExclude = playlistToExclude || this.mainPlaylistLoader_.media(); + playlistExclusionDuration = playlistExclusionDuration || error.playlistExclusionDuration || this.playlistExclusionDuration; // If there is no current playlist, then an error occurred while we were + // trying to load the main OR while we were disposing of the tech + + if (!playlistToExclude) { + this.error = error; + if (this.mediaSource.readyState !== 'open') { + this.trigger('error'); + } else { + this.sourceUpdater_.endOfStream('network'); + } + return; + } + playlistToExclude.playlistErrors_++; + const playlists = this.mainPlaylistLoader_.main.playlists; + const enabledPlaylists = playlists.filter(isEnabled); + const isFinalRendition = enabledPlaylists.length === 1 && enabledPlaylists[0] === playlistToExclude; // Don't exclude the only playlist unless it was excluded + // forever + + if (playlists.length === 1 && playlistExclusionDuration !== Infinity) { + videojs.log.warn(`Problem encountered with playlist ${playlistToExclude.id}. ` + 'Trying again since it is the only playlist.'); + this.tech_.trigger('retryplaylist'); // if this is a final rendition, we should delay + + return this.mainPlaylistLoader_.load(isFinalRendition); + } + if (isFinalRendition) { + // If we're content steering, try other pathways. + if (this.main().contentSteering) { + const pathway = this.pathwayAttribute_(playlistToExclude); // Ignore at least 1 steering manifest refresh. + + const reIncludeDelay = this.contentSteeringController_.steeringManifest.ttl * 1000; + this.contentSteeringController_.excludePathway(pathway); + this.excludeThenChangePathway_(); + setTimeout(() => { + this.contentSteeringController_.addAvailablePathway(pathway); + }, reIncludeDelay); + return; + } // Since we're on the final non-excluded playlist, and we're about to exclude + // it, instead of erring the player or retrying this playlist, clear out the current + // exclusion list. This allows other playlists to be attempted in case any have been + // fixed. + + let reincluded = false; + playlists.forEach(playlist => { + // skip current playlist which is about to be excluded + if (playlist === playlistToExclude) { + return; + } + const excludeUntil = playlist.excludeUntil; // a playlist cannot be reincluded if it wasn't excluded to begin with. + + if (typeof excludeUntil !== 'undefined' && excludeUntil !== Infinity) { + reincluded = true; + delete playlist.excludeUntil; + } + }); + if (reincluded) { + videojs.log.warn('Removing other playlists from the exclusion list because the last ' + 'rendition is about to be excluded.'); // Technically we are retrying a playlist, in that we are simply retrying a previous + // playlist. This is needed for users relying on the retryplaylist event to catch a + // case where the player might be stuck and looping through "dead" playlists. + + this.tech_.trigger('retryplaylist'); + } + } // Exclude this playlist + + let excludeUntil; + if (playlistToExclude.playlistErrors_ > this.maxPlaylistRetries) { + excludeUntil = Infinity; + } else { + excludeUntil = Date.now() + playlistExclusionDuration * 1000; + } + playlistToExclude.excludeUntil = excludeUntil; + if (error.reason) { + playlistToExclude.lastExcludeReason_ = error.reason; + } + this.tech_.trigger('excludeplaylist'); + this.tech_.trigger({ + type: 'usage', + name: 'vhs-rendition-excluded' + }); // TODO: only load a new playlist if we're excluding the current playlist + // If this function was called with a playlist that's not the current active playlist + // (e.g., media().id !== playlistToExclude.id), + // then a new playlist should not be selected and loaded, as there's nothing wrong with the current playlist. + + const nextPlaylist = this.selectPlaylist(); + if (!nextPlaylist) { + this.error = 'Playback cannot continue. No available working or supported playlists.'; + this.trigger('error'); + return; + } + const logFn = error.internal ? this.logger_ : videojs.log.warn; + const errorMessage = error.message ? ' ' + error.message : ''; + logFn(`${error.internal ? 'Internal problem' : 'Problem'} encountered with playlist ${playlistToExclude.id}.` + `${errorMessage} Switching to playlist ${nextPlaylist.id}.`); // if audio group changed reset audio loaders + + if (nextPlaylist.attributes.AUDIO !== playlistToExclude.attributes.AUDIO) { + this.delegateLoaders_('audio', ['abort', 'pause']); + } // if subtitle group changed reset subtitle loaders + + if (nextPlaylist.attributes.SUBTITLES !== playlistToExclude.attributes.SUBTITLES) { + this.delegateLoaders_('subtitle', ['abort', 'pause']); + } + this.delegateLoaders_('main', ['abort', 'pause']); + const delayDuration = nextPlaylist.targetDuration / 2 * 1000 || 5 * 1000; + const shouldDelay = typeof nextPlaylist.lastRequest === 'number' && Date.now() - nextPlaylist.lastRequest <= delayDuration; // delay if it's a final rendition or if the last refresh is sooner than half targetDuration + + return this.switchMedia_(nextPlaylist, 'exclude', isFinalRendition || shouldDelay); + } + /** + * Pause all segment/playlist loaders + */ + + pauseLoading() { + this.delegateLoaders_('all', ['abort', 'pause']); + this.stopABRTimer_(); + } + /** + * Call a set of functions in order on playlist loaders, segment loaders, + * or both types of loaders. + * + * @param {string} filter + * Filter loaders that should call fnNames using a string. Can be: + * * all - run on all loaders + * * audio - run on all audio loaders + * * subtitle - run on all subtitle loaders + * * main - run on the main loaders + * + * @param {Array|string} fnNames + * A string or array of function names to call. + */ + + delegateLoaders_(filter, fnNames) { + const loaders = []; + const dontFilterPlaylist = filter === 'all'; + if (dontFilterPlaylist || filter === 'main') { + loaders.push(this.mainPlaylistLoader_); + } + const mediaTypes = []; + if (dontFilterPlaylist || filter === 'audio') { + mediaTypes.push('AUDIO'); + } + if (dontFilterPlaylist || filter === 'subtitle') { + mediaTypes.push('CLOSED-CAPTIONS'); + mediaTypes.push('SUBTITLES'); + } + mediaTypes.forEach(mediaType => { + const loader = this.mediaTypes_[mediaType] && this.mediaTypes_[mediaType].activePlaylistLoader; + if (loader) { + loaders.push(loader); + } + }); + ['main', 'audio', 'subtitle'].forEach(name => { + const loader = this[`${name}SegmentLoader_`]; + if (loader && (filter === name || filter === 'all')) { + loaders.push(loader); + } + }); + loaders.forEach(loader => fnNames.forEach(fnName => { + if (typeof loader[fnName] === 'function') { + loader[fnName](); + } + })); + } + /** + * set the current time on all segment loaders + * + * @param {TimeRange} currentTime the current time to set + * @return {TimeRange} the current time + */ + + setCurrentTime(currentTime) { + const buffered = findRange(this.tech_.buffered(), currentTime); + if (!(this.mainPlaylistLoader_ && this.mainPlaylistLoader_.media())) { + // return immediately if the metadata is not ready yet + return 0; + } // it's clearly an edge-case but don't thrown an error if asked to + // seek within an empty playlist + + if (!this.mainPlaylistLoader_.media().segments) { + return 0; + } // if the seek location is already buffered, continue buffering as usual + + if (buffered && buffered.length) { + return currentTime; + } // cancel outstanding requests so we begin buffering at the new + // location + + this.mainSegmentLoader_.pause(); + this.mainSegmentLoader_.resetEverything(); + if (this.mediaTypes_.AUDIO.activePlaylistLoader) { + this.audioSegmentLoader_.pause(); + this.audioSegmentLoader_.resetEverything(); + } + if (this.mediaTypes_.SUBTITLES.activePlaylistLoader) { + this.subtitleSegmentLoader_.pause(); + this.subtitleSegmentLoader_.resetEverything(); + } // start segment loader loading in case they are paused + + this.load(); + } + /** + * get the current duration + * + * @return {TimeRange} the duration + */ + + duration() { + if (!this.mainPlaylistLoader_) { + return 0; + } + const media = this.mainPlaylistLoader_.media(); + if (!media) { + // no playlists loaded yet, so can't determine a duration + return 0; + } // Don't rely on the media source for duration in the case of a live playlist since + // setting the native MediaSource's duration to infinity ends up with consequences to + // seekable behavior. See https://github.com/w3c/media-source/issues/5 for details. + // + // This is resolved in the spec by https://github.com/w3c/media-source/pull/92, + // however, few browsers have support for setLiveSeekableRange() + // https://developer.mozilla.org/en-US/docs/Web/API/MediaSource/setLiveSeekableRange + // + // Until a time when the duration of the media source can be set to infinity, and a + // seekable range specified across browsers, just return Infinity. + + if (!media.endList) { + return Infinity; + } // Since this is a VOD video, it is safe to rely on the media source's duration (if + // available). If it's not available, fall back to a playlist-calculated estimate. + + if (this.mediaSource) { + return this.mediaSource.duration; + } + return Vhs$1.Playlist.duration(media); + } + /** + * check the seekable range + * + * @return {TimeRange} the seekable range + */ + + seekable() { + return this.seekable_; + } + onSyncInfoUpdate_() { + let audioSeekable; // TODO check for creation of both source buffers before updating seekable + // + // A fix was made to this function where a check for + // this.sourceUpdater_.hasCreatedSourceBuffers + // was added to ensure that both source buffers were created before seekable was + // updated. However, it originally had a bug where it was checking for a true and + // returning early instead of checking for false. Setting it to check for false to + // return early though created other issues. A call to play() would check for seekable + // end without verifying that a seekable range was present. In addition, even checking + // for that didn't solve some issues, as handleFirstPlay is sometimes worked around + // due to a media update calling load on the segment loaders, skipping a seek to live, + // thereby starting live streams at the beginning of the stream rather than at the end. + // + // This conditional should be fixed to wait for the creation of two source buffers at + // the same time as the other sections of code are fixed to properly seek to live and + // not throw an error due to checking for a seekable end when no seekable range exists. + // + // For now, fall back to the older behavior, with the understanding that the seekable + // range may not be completely correct, leading to a suboptimal initial live point. + + if (!this.mainPlaylistLoader_) { + return; + } + let media = this.mainPlaylistLoader_.media(); + if (!media) { + return; + } + let expired = this.syncController_.getExpiredTime(media, this.duration()); + if (expired === null) { + // not enough information to update seekable + return; + } + const main = this.mainPlaylistLoader_.main; + const mainSeekable = Vhs$1.Playlist.seekable(media, expired, Vhs$1.Playlist.liveEdgeDelay(main, media)); + if (mainSeekable.length === 0) { + return; + } + if (this.mediaTypes_.AUDIO.activePlaylistLoader) { + media = this.mediaTypes_.AUDIO.activePlaylistLoader.media(); + expired = this.syncController_.getExpiredTime(media, this.duration()); + if (expired === null) { + return; + } + audioSeekable = Vhs$1.Playlist.seekable(media, expired, Vhs$1.Playlist.liveEdgeDelay(main, media)); + if (audioSeekable.length === 0) { + return; + } + } + let oldEnd; + let oldStart; + if (this.seekable_ && this.seekable_.length) { + oldEnd = this.seekable_.end(0); + oldStart = this.seekable_.start(0); + } + if (!audioSeekable) { + // seekable has been calculated based on buffering video data so it + // can be returned directly + this.seekable_ = mainSeekable; + } else if (audioSeekable.start(0) > mainSeekable.end(0) || mainSeekable.start(0) > audioSeekable.end(0)) { + // seekables are pretty far off, rely on main + this.seekable_ = mainSeekable; + } else { + this.seekable_ = createTimeRanges([[audioSeekable.start(0) > mainSeekable.start(0) ? audioSeekable.start(0) : mainSeekable.start(0), audioSeekable.end(0) < mainSeekable.end(0) ? audioSeekable.end(0) : mainSeekable.end(0)]]); + } // seekable is the same as last time + + if (this.seekable_ && this.seekable_.length) { + if (this.seekable_.end(0) === oldEnd && this.seekable_.start(0) === oldStart) { + return; + } + } + this.logger_(`seekable updated [${printableRange(this.seekable_)}]`); + this.tech_.trigger('seekablechanged'); + } + /** + * Update the player duration + */ + + updateDuration(isLive) { + if (this.updateDuration_) { + this.mediaSource.removeEventListener('sourceopen', this.updateDuration_); + this.updateDuration_ = null; + } + if (this.mediaSource.readyState !== 'open') { + this.updateDuration_ = this.updateDuration.bind(this, isLive); + this.mediaSource.addEventListener('sourceopen', this.updateDuration_); + return; + } + if (isLive) { + const seekable = this.seekable(); + if (!seekable.length) { + return; + } // Even in the case of a live playlist, the native MediaSource's duration should not + // be set to Infinity (even though this would be expected for a live playlist), since + // setting the native MediaSource's duration to infinity ends up with consequences to + // seekable behavior. See https://github.com/w3c/media-source/issues/5 for details. + // + // This is resolved in the spec by https://github.com/w3c/media-source/pull/92, + // however, few browsers have support for setLiveSeekableRange() + // https://developer.mozilla.org/en-US/docs/Web/API/MediaSource/setLiveSeekableRange + // + // Until a time when the duration of the media source can be set to infinity, and a + // seekable range specified across browsers, the duration should be greater than or + // equal to the last possible seekable value. + // MediaSource duration starts as NaN + // It is possible (and probable) that this case will never be reached for many + // sources, since the MediaSource reports duration as the highest value without + // accounting for timestamp offset. For example, if the timestamp offset is -100 and + // we buffered times 0 to 100 with real times of 100 to 200, even though current + // time will be between 0 and 100, the native media source may report the duration + // as 200. However, since we report duration separate from the media source (as + // Infinity), and as long as the native media source duration value is greater than + // our reported seekable range, seeks will work as expected. The large number as + // duration for live is actually a strategy used by some players to work around the + // issue of live seekable ranges cited above. + + if (isNaN(this.mediaSource.duration) || this.mediaSource.duration < seekable.end(seekable.length - 1)) { + this.sourceUpdater_.setDuration(seekable.end(seekable.length - 1)); + } + return; + } + const buffered = this.tech_.buffered(); + let duration = Vhs$1.Playlist.duration(this.mainPlaylistLoader_.media()); + if (buffered.length > 0) { + duration = Math.max(duration, buffered.end(buffered.length - 1)); + } + if (this.mediaSource.duration !== duration) { + this.sourceUpdater_.setDuration(duration); + } + } + /** + * dispose of the PlaylistController and everything + * that it controls + */ + + dispose() { + this.trigger('dispose'); + this.decrypter_.terminate(); + this.mainPlaylistLoader_.dispose(); + this.mainSegmentLoader_.dispose(); + this.contentSteeringController_.dispose(); + this.keyStatusMap_.clear(); + if (this.loadOnPlay_) { + this.tech_.off('play', this.loadOnPlay_); + } + ['AUDIO', 'SUBTITLES'].forEach(type => { + const groups = this.mediaTypes_[type].groups; + for (const id in groups) { + groups[id].forEach(group => { + if (group.playlistLoader) { + group.playlistLoader.dispose(); + } + }); + } + }); + this.audioSegmentLoader_.dispose(); + this.subtitleSegmentLoader_.dispose(); + this.sourceUpdater_.dispose(); + this.timelineChangeController_.dispose(); + this.stopABRTimer_(); + if (this.updateDuration_) { + this.mediaSource.removeEventListener('sourceopen', this.updateDuration_); + } + this.mediaSource.removeEventListener('durationchange', this.handleDurationChange_); // load the media source into the player + + this.mediaSource.removeEventListener('sourceopen', this.handleSourceOpen_); + this.mediaSource.removeEventListener('sourceended', this.handleSourceEnded_); + this.off(); + } + /** + * return the main playlist object if we have one + * + * @return {Object} the main playlist object that we parsed + */ + + main() { + return this.mainPlaylistLoader_.main; + } + /** + * return the currently selected playlist + * + * @return {Object} the currently selected playlist object that we parsed + */ + + media() { + // playlist loader will not return media if it has not been fully loaded + return this.mainPlaylistLoader_.media() || this.initialMedia_; + } + areMediaTypesKnown_() { + const usingAudioLoader = !!this.mediaTypes_.AUDIO.activePlaylistLoader; + const hasMainMediaInfo = !!this.mainSegmentLoader_.getCurrentMediaInfo_(); // if we are not using an audio loader, then we have audio media info + // otherwise check on the segment loader. + + const hasAudioMediaInfo = !usingAudioLoader ? true : !!this.audioSegmentLoader_.getCurrentMediaInfo_(); // one or both loaders has not loaded sufficently to get codecs + + if (!hasMainMediaInfo || !hasAudioMediaInfo) { + return false; + } + return true; + } + getCodecsOrExclude_() { + const media = { + main: this.mainSegmentLoader_.getCurrentMediaInfo_() || {}, + audio: this.audioSegmentLoader_.getCurrentMediaInfo_() || {} + }; + const playlist = this.mainSegmentLoader_.getPendingSegmentPlaylist() || this.media(); // set "main" media equal to video + + media.video = media.main; + const playlistCodecs = codecsForPlaylist(this.main(), playlist); + const codecs = {}; + const usingAudioLoader = !!this.mediaTypes_.AUDIO.activePlaylistLoader; + if (media.main.hasVideo) { + codecs.video = playlistCodecs.video || media.main.videoCodec || DEFAULT_VIDEO_CODEC; + } + if (media.main.isMuxed) { + codecs.video += `,${playlistCodecs.audio || media.main.audioCodec || DEFAULT_AUDIO_CODEC}`; + } + if (media.main.hasAudio && !media.main.isMuxed || media.audio.hasAudio || usingAudioLoader) { + codecs.audio = playlistCodecs.audio || media.main.audioCodec || media.audio.audioCodec || DEFAULT_AUDIO_CODEC; // set audio isFmp4 so we use the correct "supports" function below + + media.audio.isFmp4 = media.main.hasAudio && !media.main.isMuxed ? media.main.isFmp4 : media.audio.isFmp4; + } // no codecs, no playback. + + if (!codecs.audio && !codecs.video) { + this.excludePlaylist({ + playlistToExclude: playlist, + error: { + message: 'Could not determine codecs for playlist.' + }, + playlistExclusionDuration: Infinity + }); + return; + } // fmp4 relies on browser support, while ts relies on muxer support + + const supportFunction = (isFmp4, codec) => isFmp4 ? browserSupportsCodec(codec) : muxerSupportsCodec(codec); + const unsupportedCodecs = {}; + let unsupportedAudio; + ['video', 'audio'].forEach(function (type) { + if (codecs.hasOwnProperty(type) && !supportFunction(media[type].isFmp4, codecs[type])) { + const supporter = media[type].isFmp4 ? 'browser' : 'muxer'; + unsupportedCodecs[supporter] = unsupportedCodecs[supporter] || []; + unsupportedCodecs[supporter].push(codecs[type]); + if (type === 'audio') { + unsupportedAudio = supporter; + } + } + }); + if (usingAudioLoader && unsupportedAudio && playlist.attributes.AUDIO) { + const audioGroup = playlist.attributes.AUDIO; + this.main().playlists.forEach(variant => { + const variantAudioGroup = variant.attributes && variant.attributes.AUDIO; + if (variantAudioGroup === audioGroup && variant !== playlist) { + variant.excludeUntil = Infinity; + } + }); + this.logger_(`excluding audio group ${audioGroup} as ${unsupportedAudio} does not support codec(s): "${codecs.audio}"`); + } // if we have any unsupported codecs exclude this playlist. + + if (Object.keys(unsupportedCodecs).length) { + const message = Object.keys(unsupportedCodecs).reduce((acc, supporter) => { + if (acc) { + acc += ', '; + } + acc += `${supporter} does not support codec(s): "${unsupportedCodecs[supporter].join(',')}"`; + return acc; + }, '') + '.'; + this.excludePlaylist({ + playlistToExclude: playlist, + error: { + internal: true, + message + }, + playlistExclusionDuration: Infinity + }); + return; + } // check if codec switching is happening + + if (this.sourceUpdater_.hasCreatedSourceBuffers() && !this.sourceUpdater_.canChangeType()) { + const switchMessages = []; + ['video', 'audio'].forEach(type => { + const newCodec = (parseCodecs(this.sourceUpdater_.codecs[type] || '')[0] || {}).type; + const oldCodec = (parseCodecs(codecs[type] || '')[0] || {}).type; + if (newCodec && oldCodec && newCodec.toLowerCase() !== oldCodec.toLowerCase()) { + switchMessages.push(`"${this.sourceUpdater_.codecs[type]}" -> "${codecs[type]}"`); + } + }); + if (switchMessages.length) { + this.excludePlaylist({ + playlistToExclude: playlist, + error: { + message: `Codec switching not supported: ${switchMessages.join(', ')}.`, + internal: true + }, + playlistExclusionDuration: Infinity + }); + return; + } + } // TODO: when using the muxer shouldn't we just return + // the codecs that the muxer outputs? + + return codecs; + } + /** + * Create source buffers and exlude any incompatible renditions. + * + * @private + */ + + tryToCreateSourceBuffers_() { + // media source is not ready yet or sourceBuffers are already + // created. + if (this.mediaSource.readyState !== 'open' || this.sourceUpdater_.hasCreatedSourceBuffers()) { + return; + } + if (!this.areMediaTypesKnown_()) { + return; + } + const codecs = this.getCodecsOrExclude_(); // no codecs means that the playlist was excluded + + if (!codecs) { + return; + } + this.sourceUpdater_.createSourceBuffers(codecs); + const codecString = [codecs.video, codecs.audio].filter(Boolean).join(','); + this.excludeIncompatibleVariants_(codecString); + } + /** + * Excludes playlists with codecs that are unsupported by the muxer and browser. + */ + + excludeUnsupportedVariants_() { + const playlists = this.main().playlists; + const ids = []; // TODO: why don't we have a property to loop through all + // playlist? Why did we ever mix indexes and keys? + + Object.keys(playlists).forEach(key => { + const variant = playlists[key]; // check if we already processed this playlist. + + if (ids.indexOf(variant.id) !== -1) { + return; + } + ids.push(variant.id); + const codecs = codecsForPlaylist(this.main, variant); + const unsupported = []; + if (codecs.audio && !muxerSupportsCodec(codecs.audio) && !browserSupportsCodec(codecs.audio)) { + unsupported.push(`audio codec ${codecs.audio}`); + } + if (codecs.video && !muxerSupportsCodec(codecs.video) && !browserSupportsCodec(codecs.video)) { + unsupported.push(`video codec ${codecs.video}`); + } + if (codecs.text && codecs.text === 'stpp.ttml.im1t') { + unsupported.push(`text codec ${codecs.text}`); + } + if (unsupported.length) { + variant.excludeUntil = Infinity; + this.logger_(`excluding ${variant.id} for unsupported: ${unsupported.join(', ')}`); + } + }); + } + /** + * Exclude playlists that are known to be codec or + * stream-incompatible with the SourceBuffer configuration. For + * instance, Media Source Extensions would cause the video element to + * stall waiting for video data if you switched from a variant with + * video and audio to an audio-only one. + * + * @param {Object} media a media playlist compatible with the current + * set of SourceBuffers. Variants in the current main playlist that + * do not appear to have compatible codec or stream configurations + * will be excluded from the default playlist selection algorithm + * indefinitely. + * @private + */ + + excludeIncompatibleVariants_(codecString) { + const ids = []; + const playlists = this.main().playlists; + const codecs = unwrapCodecList(parseCodecs(codecString)); + const codecCount_ = codecCount(codecs); + const videoDetails = codecs.video && parseCodecs(codecs.video)[0] || null; + const audioDetails = codecs.audio && parseCodecs(codecs.audio)[0] || null; + Object.keys(playlists).forEach(key => { + const variant = playlists[key]; // check if we already processed this playlist. + // or it if it is already excluded forever. + + if (ids.indexOf(variant.id) !== -1 || variant.excludeUntil === Infinity) { + return; + } + ids.push(variant.id); + const exclusionReasons = []; // get codecs from the playlist for this variant + + const variantCodecs = codecsForPlaylist(this.mainPlaylistLoader_.main, variant); + const variantCodecCount = codecCount(variantCodecs); // if no codecs are listed, we cannot determine that this + // variant is incompatible. Wait for mux.js to probe + + if (!variantCodecs.audio && !variantCodecs.video) { + return; + } // TODO: we can support this by removing the + // old media source and creating a new one, but it will take some work. + // The number of streams cannot change + + if (variantCodecCount !== codecCount_) { + exclusionReasons.push(`codec count "${variantCodecCount}" !== "${codecCount_}"`); + } // only exclude playlists by codec change, if codecs cannot switch + // during playback. + + if (!this.sourceUpdater_.canChangeType()) { + const variantVideoDetails = variantCodecs.video && parseCodecs(variantCodecs.video)[0] || null; + const variantAudioDetails = variantCodecs.audio && parseCodecs(variantCodecs.audio)[0] || null; // the video codec cannot change + + if (variantVideoDetails && videoDetails && variantVideoDetails.type.toLowerCase() !== videoDetails.type.toLowerCase()) { + exclusionReasons.push(`video codec "${variantVideoDetails.type}" !== "${videoDetails.type}"`); + } // the audio codec cannot change + + if (variantAudioDetails && audioDetails && variantAudioDetails.type.toLowerCase() !== audioDetails.type.toLowerCase()) { + exclusionReasons.push(`audio codec "${variantAudioDetails.type}" !== "${audioDetails.type}"`); + } + } + if (exclusionReasons.length) { + variant.excludeUntil = Infinity; + this.logger_(`excluding ${variant.id}: ${exclusionReasons.join(' && ')}`); + } + }); + } + updateAdCues_(media) { + let offset = 0; + const seekable = this.seekable(); + if (seekable.length) { + offset = seekable.start(0); + } + updateAdCues(media, this.cueTagsTrack_, offset); + } + /** + * Calculates the desired forward buffer length based on current time + * + * @return {number} Desired forward buffer length in seconds + */ + + goalBufferLength() { + const currentTime = this.tech_.currentTime(); + const initial = Config.GOAL_BUFFER_LENGTH; + const rate = Config.GOAL_BUFFER_LENGTH_RATE; + const max = Math.max(initial, Config.MAX_GOAL_BUFFER_LENGTH); + return Math.min(initial + currentTime * rate, max); + } + /** + * Calculates the desired buffer low water line based on current time + * + * @return {number} Desired buffer low water line in seconds + */ + + bufferLowWaterLine() { + const currentTime = this.tech_.currentTime(); + const initial = Config.BUFFER_LOW_WATER_LINE; + const rate = Config.BUFFER_LOW_WATER_LINE_RATE; + const max = Math.max(initial, Config.MAX_BUFFER_LOW_WATER_LINE); + const newMax = Math.max(initial, Config.EXPERIMENTAL_MAX_BUFFER_LOW_WATER_LINE); + return Math.min(initial + currentTime * rate, this.bufferBasedABR ? newMax : max); + } + bufferHighWaterLine() { + return Config.BUFFER_HIGH_WATER_LINE; + } + addDateRangesToTextTrack_(dateRanges) { + createMetadataTrackIfNotExists(this.inbandTextTracks_, 'com.apple.streaming', this.tech_); + addDateRangeMetadata({ + inbandTextTracks: this.inbandTextTracks_, + dateRanges + }); + } + addMetadataToTextTrack(dispatchType, metadataArray, videoDuration) { + const timestampOffset = this.sourceUpdater_.videoBuffer ? this.sourceUpdater_.videoTimestampOffset() : this.sourceUpdater_.audioTimestampOffset(); // There's potentially an issue where we could double add metadata if there's a muxed + // audio/video source with a metadata track, and an alt audio with a metadata track. + // However, this probably won't happen, and if it does it can be handled then. + + createMetadataTrackIfNotExists(this.inbandTextTracks_, dispatchType, this.tech_); + addMetadata({ + inbandTextTracks: this.inbandTextTracks_, + metadataArray, + timestampOffset, + videoDuration + }); + } + /** + * Utility for getting the pathway or service location from an HLS or DASH playlist. + * + * @param {Object} playlist for getting pathway from. + * @return the pathway attribute of a playlist + */ + + pathwayAttribute_(playlist) { + return playlist.attributes['PATHWAY-ID'] || playlist.attributes.serviceLocation; + } + /** + * Initialize available pathways and apply the tag properties. + */ + + initContentSteeringController_() { + const main = this.main(); + if (!main.contentSteering) { + return; + } + for (const playlist of main.playlists) { + this.contentSteeringController_.addAvailablePathway(this.pathwayAttribute_(playlist)); + } + this.contentSteeringController_.assignTagProperties(main.uri, main.contentSteering); // request the steering manifest immediately if queryBeforeStart is set. + + if (this.contentSteeringController_.queryBeforeStart) { + // When queryBeforeStart is true, initial request should omit steering parameters. + this.contentSteeringController_.requestSteeringManifest(true); + return; + } // otherwise start content steering after playback starts + + this.tech_.one('canplay', () => { + this.contentSteeringController_.requestSteeringManifest(); + }); + } + /** + * Reset the content steering controller and re-init. + */ + + resetContentSteeringController_() { + this.contentSteeringController_.clearAvailablePathways(); + this.contentSteeringController_.dispose(); + this.initContentSteeringController_(); + } + /** + * Attaches the listeners for content steering. + */ + + attachContentSteeringListeners_() { + this.contentSteeringController_.on('content-steering', this.excludeThenChangePathway_.bind(this)); + if (this.sourceType_ === 'dash') { + this.mainPlaylistLoader_.on('loadedplaylist', () => { + const main = this.main(); // check if steering tag or pathways changed. + + const didDashTagChange = this.contentSteeringController_.didDASHTagChange(main.uri, main.contentSteering); + const didPathwaysChange = () => { + const availablePathways = this.contentSteeringController_.getAvailablePathways(); + const newPathways = []; + for (const playlist of main.playlists) { + const serviceLocation = playlist.attributes.serviceLocation; + if (serviceLocation) { + newPathways.push(serviceLocation); + if (!availablePathways.has(serviceLocation)) { + return true; + } + } + } // If we have no new serviceLocations and previously had availablePathways + + if (!newPathways.length && availablePathways.size) { + return true; + } + return false; + }; + if (didDashTagChange || didPathwaysChange()) { + this.resetContentSteeringController_(); + } + }); + } + } + /** + * Simple exclude and change playlist logic for content steering. + */ + + excludeThenChangePathway_() { + const currentPathway = this.contentSteeringController_.getPathway(); + if (!currentPathway) { + return; + } + this.handlePathwayClones_(); + const main = this.main(); + const playlists = main.playlists; + const ids = new Set(); + let didEnablePlaylists = false; + Object.keys(playlists).forEach(key => { + const variant = playlists[key]; + const pathwayId = this.pathwayAttribute_(variant); + const differentPathwayId = pathwayId && currentPathway !== pathwayId; + const steeringExclusion = variant.excludeUntil === Infinity && variant.lastExcludeReason_ === 'content-steering'; + if (steeringExclusion && !differentPathwayId) { + delete variant.excludeUntil; + delete variant.lastExcludeReason_; + didEnablePlaylists = true; + } + const noExcludeUntil = !variant.excludeUntil && variant.excludeUntil !== Infinity; + const shouldExclude = !ids.has(variant.id) && differentPathwayId && noExcludeUntil; + if (!shouldExclude) { + return; + } + ids.add(variant.id); + variant.excludeUntil = Infinity; + variant.lastExcludeReason_ = 'content-steering'; // TODO: kind of spammy, maybe move this. + + this.logger_(`excluding ${variant.id} for ${variant.lastExcludeReason_}`); + }); + if (this.contentSteeringController_.manifestType_ === 'DASH') { + Object.keys(this.mediaTypes_).forEach(key => { + const type = this.mediaTypes_[key]; + if (type.activePlaylistLoader) { + const currentPlaylist = type.activePlaylistLoader.media_; // Check if the current media playlist matches the current CDN + + if (currentPlaylist && currentPlaylist.attributes.serviceLocation !== currentPathway) { + didEnablePlaylists = true; + } + } + }); + } + if (didEnablePlaylists) { + this.changeSegmentPathway_(); + } + } + /** + * Add, update, or delete playlists and media groups for + * the pathway clones for HLS Content Steering. + * + * See https://datatracker.ietf.org/doc/draft-pantos-hls-rfc8216bis/ + * + * NOTE: Pathway cloning does not currently support the `PER_VARIANT_URIS` and + * `PER_RENDITION_URIS` as we do not handle `STABLE-VARIANT-ID` or + * `STABLE-RENDITION-ID` values. + */ + + handlePathwayClones_() { + const main = this.main(); + const playlists = main.playlists; + const currentPathwayClones = this.contentSteeringController_.currentPathwayClones; + const nextPathwayClones = this.contentSteeringController_.nextPathwayClones; + const hasClones = currentPathwayClones && currentPathwayClones.size || nextPathwayClones && nextPathwayClones.size; + if (!hasClones) { + return; + } + for (const [id, clone] of currentPathwayClones.entries()) { + const newClone = nextPathwayClones.get(id); // Delete the old pathway clone. + + if (!newClone) { + this.mainPlaylistLoader_.updateOrDeleteClone(clone); + this.contentSteeringController_.excludePathway(id); + } + } + for (const [id, clone] of nextPathwayClones.entries()) { + const oldClone = currentPathwayClones.get(id); // Create a new pathway if it is a new pathway clone object. + + if (!oldClone) { + const playlistsToClone = playlists.filter(p => { + return p.attributes['PATHWAY-ID'] === clone['BASE-ID']; + }); + playlistsToClone.forEach(p => { + this.mainPlaylistLoader_.addClonePathway(clone, p); + }); + this.contentSteeringController_.addAvailablePathway(id); + continue; + } // There have not been changes to the pathway clone object, so skip. + + if (this.equalPathwayClones_(oldClone, clone)) { + continue; + } // Update a preexisting cloned pathway. + // True is set for the update flag. + + this.mainPlaylistLoader_.updateOrDeleteClone(clone, true); + this.contentSteeringController_.addAvailablePathway(id); + } // Deep copy contents of next to current pathways. + + this.contentSteeringController_.currentPathwayClones = new Map(JSON.parse(JSON.stringify([...nextPathwayClones]))); + } + /** + * Determines whether two pathway clone objects are equivalent. + * + * @param {Object} a The first pathway clone object. + * @param {Object} b The second pathway clone object. + * @return {boolean} True if the pathway clone objects are equal, false otherwise. + */ + + equalPathwayClones_(a, b) { + if (a['BASE-ID'] !== b['BASE-ID'] || a.ID !== b.ID || a['URI-REPLACEMENT'].HOST !== b['URI-REPLACEMENT'].HOST) { + return false; + } + const aParams = a['URI-REPLACEMENT'].PARAMS; + const bParams = b['URI-REPLACEMENT'].PARAMS; // We need to iterate through both lists of params because one could be + // missing a parameter that the other has. + + for (const p in aParams) { + if (aParams[p] !== bParams[p]) { + return false; + } + } + for (const p in bParams) { + if (aParams[p] !== bParams[p]) { + return false; + } + } + return true; + } + /** + * Changes the current playlists for audio, video and subtitles after a new pathway + * is chosen from content steering. + */ + + changeSegmentPathway_() { + const nextPlaylist = this.selectPlaylist(); + this.pauseLoading(); // Switch audio and text track playlists if necessary in DASH + + if (this.contentSteeringController_.manifestType_ === 'DASH') { + this.switchMediaForDASHContentSteering_(); + } + this.switchMedia_(nextPlaylist, 'content-steering'); + } + /** + * Iterates through playlists and check their keyId set and compare with the + * keyStatusMap, only enable playlists that have a usable key. If the playlist + * has no keyId leave it enabled by default. + */ + + excludeNonUsablePlaylistsByKeyId_() { + if (!this.mainPlaylistLoader_ || !this.mainPlaylistLoader_.main) { + return; + } + let nonUsableKeyStatusCount = 0; + const NON_USABLE = 'non-usable'; + this.mainPlaylistLoader_.main.playlists.forEach(playlist => { + const keyIdSet = this.mainPlaylistLoader_.getKeyIdSet(playlist); // If the playlist doesn't have keyIDs lets not exclude it. + + if (!keyIdSet || !keyIdSet.size) { + return; + } + keyIdSet.forEach(key => { + const USABLE = 'usable'; + const hasUsableKeyStatus = this.keyStatusMap_.has(key) && this.keyStatusMap_.get(key) === USABLE; + const nonUsableExclusion = playlist.lastExcludeReason_ === NON_USABLE && playlist.excludeUntil === Infinity; + if (!hasUsableKeyStatus) { + // Only exclude playlists that haven't already been excluded as non-usable. + if (playlist.excludeUntil !== Infinity && playlist.lastExcludeReason_ !== NON_USABLE) { + playlist.excludeUntil = Infinity; + playlist.lastExcludeReason_ = NON_USABLE; + this.logger_(`excluding playlist ${playlist.id} because the key ID ${key} doesn't exist in the keyStatusMap or is not ${USABLE}`); + } // count all nonUsableKeyStatus + + nonUsableKeyStatusCount++; + } else if (hasUsableKeyStatus && nonUsableExclusion) { + delete playlist.excludeUntil; + delete playlist.lastExcludeReason_; + this.logger_(`enabling playlist ${playlist.id} because key ID ${key} is ${USABLE}`); + } + }); + }); // If for whatever reason every playlist has a non usable key status. Lets try re-including the SD renditions as a failsafe. + + if (nonUsableKeyStatusCount >= this.mainPlaylistLoader_.main.playlists.length) { + this.mainPlaylistLoader_.main.playlists.forEach(playlist => { + const isNonHD = playlist && playlist.attributes && playlist.attributes.RESOLUTION && playlist.attributes.RESOLUTION.height < 720; + const excludedForNonUsableKey = playlist.excludeUntil === Infinity && playlist.lastExcludeReason_ === NON_USABLE; + if (isNonHD && excludedForNonUsableKey) { + // Only delete the excludeUntil so we don't try and re-exclude these playlists. + delete playlist.excludeUntil; + videojs.log.warn(`enabling non-HD playlist ${playlist.id} because all playlists were excluded due to ${NON_USABLE} key IDs`); + } + }); + } + } + /** + * Adds a keystatus to the keystatus map, tries to convert to string if necessary. + * + * @param {any} keyId the keyId to add a status for + * @param {string} status the status of the keyId + */ + + addKeyStatus_(keyId, status) { + const isString = typeof keyId === 'string'; + const keyIdHexString = isString ? keyId : bufferToHexString(keyId); + const formattedKeyIdString = keyIdHexString.slice(0, 32).toLowerCase(); + this.logger_(`KeyStatus '${status}' with key ID ${formattedKeyIdString} added to the keyStatusMap`); + this.keyStatusMap_.set(formattedKeyIdString, status); + } + /** + * Utility function for adding key status to the keyStatusMap and filtering usable encrypted playlists. + * + * @param {any} keyId the keyId from the keystatuschange event + * @param {string} status the key status string + */ + + updatePlaylistByKeyStatus(keyId, status) { + this.addKeyStatus_(keyId, status); + if (!this.waitingForFastQualityPlaylistReceived_) { + this.excludeNonUsableThenChangePlaylist_(); + } // Listen to loadedplaylist with a single listener and check for new contentProtection elements when a playlist is updated. + + this.mainPlaylistLoader_.off('loadedplaylist', this.excludeNonUsableThenChangePlaylist_.bind(this)); + this.mainPlaylistLoader_.on('loadedplaylist', this.excludeNonUsableThenChangePlaylist_.bind(this)); + } + excludeNonUsableThenChangePlaylist_() { + this.excludeNonUsablePlaylistsByKeyId_(); + this.fastQualityChange_(); + } +} + +/** + * Returns a function that acts as the Enable/disable playlist function. + * + * @param {PlaylistLoader} loader - The main playlist loader + * @param {string} playlistID - id of the playlist + * @param {Function} changePlaylistFn - A function to be called after a + * playlist's enabled-state has been changed. Will NOT be called if a + * playlist's enabled-state is unchanged + * @param {boolean=} enable - Value to set the playlist enabled-state to + * or if undefined returns the current enabled-state for the playlist + * @return {Function} Function for setting/getting enabled + */ + +const enableFunction = (loader, playlistID, changePlaylistFn) => enable => { + const playlist = loader.main.playlists[playlistID]; + const incompatible = isIncompatible(playlist); + const currentlyEnabled = isEnabled(playlist); + if (typeof enable === 'undefined') { + return currentlyEnabled; + } + if (enable) { + delete playlist.disabled; + } else { + playlist.disabled = true; + } + if (enable !== currentlyEnabled && !incompatible) { + // Ensure the outside world knows about our changes + changePlaylistFn(playlist); + if (enable) { + loader.trigger('renditionenabled'); + } else { + loader.trigger('renditiondisabled'); + } + } + return enable; +}; +/** + * The representation object encapsulates the publicly visible information + * in a media playlist along with a setter/getter-type function (enabled) + * for changing the enabled-state of a particular playlist entry + * + * @class Representation + */ + +class Representation { + constructor(vhsHandler, playlist, id) { + const { + playlistController_: pc + } = vhsHandler; + const qualityChangeFunction = pc.fastQualityChange_.bind(pc); // some playlist attributes are optional + + if (playlist.attributes) { + const resolution = playlist.attributes.RESOLUTION; + this.width = resolution && resolution.width; + this.height = resolution && resolution.height; + this.bandwidth = playlist.attributes.BANDWIDTH; + this.frameRate = playlist.attributes['FRAME-RATE']; + } + this.codecs = codecsForPlaylist(pc.main(), playlist); + this.playlist = playlist; // The id is simply the ordinality of the media playlist + // within the main playlist + + this.id = id; // Partially-apply the enableFunction to create a playlist- + // specific variant + + this.enabled = enableFunction(vhsHandler.playlists, playlist.id, qualityChangeFunction); + } +} +/** + * A mixin function that adds the `representations` api to an instance + * of the VhsHandler class + * + * @param {VhsHandler} vhsHandler - An instance of VhsHandler to add the + * representation API into + */ + +const renditionSelectionMixin = function (vhsHandler) { + // Add a single API-specific function to the VhsHandler instance + vhsHandler.representations = () => { + const main = vhsHandler.playlistController_.main(); + const playlists = isAudioOnly(main) ? vhsHandler.playlistController_.getAudioTrackPlaylists_() : main.playlists; + if (!playlists) { + return []; + } + return playlists.filter(media => !isIncompatible(media)).map((e, i) => new Representation(vhsHandler, e, e.id)); + }; +}; + +/** + * @file playback-watcher.js + * + * Playback starts, and now my watch begins. It shall not end until my death. I shall + * take no wait, hold no uncleared timeouts, father no bad seeks. I shall wear no crowns + * and win no glory. I shall live and die at my post. I am the corrector of the underflow. + * I am the watcher of gaps. I am the shield that guards the realms of seekable. I pledge + * my life and honor to the Playback Watch, for this Player and all the Players to come. + */ + +const timerCancelEvents = ['seeking', 'seeked', 'pause', 'playing', 'error']; +/** + * @class PlaybackWatcher + */ + +class PlaybackWatcher { + /** + * Represents an PlaybackWatcher object. + * + * @class + * @param {Object} options an object that includes the tech and settings + */ + constructor(options) { + this.playlistController_ = options.playlistController; + this.tech_ = options.tech; + this.seekable = options.seekable; + this.allowSeeksWithinUnsafeLiveWindow = options.allowSeeksWithinUnsafeLiveWindow; + this.liveRangeSafeTimeDelta = options.liveRangeSafeTimeDelta; + this.media = options.media; + this.consecutiveUpdates = 0; + this.lastRecordedTime = null; + this.checkCurrentTimeTimeout_ = null; + this.logger_ = logger('PlaybackWatcher'); + this.logger_('initialize'); + const playHandler = () => this.monitorCurrentTime_(); + const canPlayHandler = () => this.monitorCurrentTime_(); + const waitingHandler = () => this.techWaiting_(); + const cancelTimerHandler = () => this.resetTimeUpdate_(); + const pc = this.playlistController_; + const loaderTypes = ['main', 'subtitle', 'audio']; + const loaderChecks = {}; + loaderTypes.forEach(type => { + loaderChecks[type] = { + reset: () => this.resetSegmentDownloads_(type), + updateend: () => this.checkSegmentDownloads_(type) + }; + pc[`${type}SegmentLoader_`].on('appendsdone', loaderChecks[type].updateend); // If a rendition switch happens during a playback stall where the buffer + // isn't changing we want to reset. We cannot assume that the new rendition + // will also be stalled, until after new appends. + + pc[`${type}SegmentLoader_`].on('playlistupdate', loaderChecks[type].reset); // Playback stalls should not be detected right after seeking. + // This prevents one segment playlists (single vtt or single segment content) + // from being detected as stalling. As the buffer will not change in those cases, since + // the buffer is the entire video duration. + + this.tech_.on(['seeked', 'seeking'], loaderChecks[type].reset); + }); + /** + * We check if a seek was into a gap through the following steps: + * 1. We get a seeking event and we do not get a seeked event. This means that + * a seek was attempted but not completed. + * 2. We run `fixesBadSeeks_` on segment loader appends. This means that we already + * removed everything from our buffer and appended a segment, and should be ready + * to check for gaps. + */ + + const setSeekingHandlers = fn => { + ['main', 'audio'].forEach(type => { + pc[`${type}SegmentLoader_`][fn]('appended', this.seekingAppendCheck_); + }); + }; + this.seekingAppendCheck_ = () => { + if (this.fixesBadSeeks_()) { + this.consecutiveUpdates = 0; + this.lastRecordedTime = this.tech_.currentTime(); + setSeekingHandlers('off'); + } + }; + this.clearSeekingAppendCheck_ = () => setSeekingHandlers('off'); + this.watchForBadSeeking_ = () => { + this.clearSeekingAppendCheck_(); + setSeekingHandlers('on'); + }; + this.tech_.on('seeked', this.clearSeekingAppendCheck_); + this.tech_.on('seeking', this.watchForBadSeeking_); + this.tech_.on('waiting', waitingHandler); + this.tech_.on(timerCancelEvents, cancelTimerHandler); + this.tech_.on('canplay', canPlayHandler); + /* + An edge case exists that results in gaps not being skipped when they exist at the beginning of a stream. This case + is surfaced in one of two ways: + 1) The `waiting` event is fired before the player has buffered content, making it impossible + to find or skip the gap. The `waiting` event is followed by a `play` event. On first play + we can check if playback is stalled due to a gap, and skip the gap if necessary. + 2) A source with a gap at the beginning of the stream is loaded programatically while the player + is in a playing state. To catch this case, it's important that our one-time play listener is setup + even if the player is in a playing state + */ + + this.tech_.one('play', playHandler); // Define the dispose function to clean up our events + + this.dispose = () => { + this.clearSeekingAppendCheck_(); + this.logger_('dispose'); + this.tech_.off('waiting', waitingHandler); + this.tech_.off(timerCancelEvents, cancelTimerHandler); + this.tech_.off('canplay', canPlayHandler); + this.tech_.off('play', playHandler); + this.tech_.off('seeking', this.watchForBadSeeking_); + this.tech_.off('seeked', this.clearSeekingAppendCheck_); + loaderTypes.forEach(type => { + pc[`${type}SegmentLoader_`].off('appendsdone', loaderChecks[type].updateend); + pc[`${type}SegmentLoader_`].off('playlistupdate', loaderChecks[type].reset); + this.tech_.off(['seeked', 'seeking'], loaderChecks[type].reset); + }); + if (this.checkCurrentTimeTimeout_) { + window$1.clearTimeout(this.checkCurrentTimeTimeout_); + } + this.resetTimeUpdate_(); + }; + } + /** + * Periodically check current time to see if playback stopped + * + * @private + */ + + monitorCurrentTime_() { + this.checkCurrentTime_(); + if (this.checkCurrentTimeTimeout_) { + window$1.clearTimeout(this.checkCurrentTimeTimeout_); + } // 42 = 24 fps // 250 is what Webkit uses // FF uses 15 + + this.checkCurrentTimeTimeout_ = window$1.setTimeout(this.monitorCurrentTime_.bind(this), 250); + } + /** + * Reset stalled download stats for a specific type of loader + * + * @param {string} type + * The segment loader type to check. + * + * @listens SegmentLoader#playlistupdate + * @listens Tech#seeking + * @listens Tech#seeked + */ + + resetSegmentDownloads_(type) { + const loader = this.playlistController_[`${type}SegmentLoader_`]; + if (this[`${type}StalledDownloads_`] > 0) { + this.logger_(`resetting possible stalled download count for ${type} loader`); + } + this[`${type}StalledDownloads_`] = 0; + this[`${type}Buffered_`] = loader.buffered_(); + } + /** + * Checks on every segment `appendsdone` to see + * if segment appends are making progress. If they are not + * and we are still downloading bytes. We exclude the playlist. + * + * @param {string} type + * The segment loader type to check. + * + * @listens SegmentLoader#appendsdone + */ + + checkSegmentDownloads_(type) { + const pc = this.playlistController_; + const loader = pc[`${type}SegmentLoader_`]; + const buffered = loader.buffered_(); + const isBufferedDifferent = isRangeDifferent(this[`${type}Buffered_`], buffered); + this[`${type}Buffered_`] = buffered; // if another watcher is going to fix the issue or + // the buffered value for this loader changed + // appends are working + + if (isBufferedDifferent) { + this.resetSegmentDownloads_(type); + return; + } + this[`${type}StalledDownloads_`]++; + this.logger_(`found #${this[`${type}StalledDownloads_`]} ${type} appends that did not increase buffer (possible stalled download)`, { + playlistId: loader.playlist_ && loader.playlist_.id, + buffered: timeRangesToArray(buffered) + }); // after 10 possibly stalled appends with no reset, exclude + + if (this[`${type}StalledDownloads_`] < 10) { + return; + } + this.logger_(`${type} loader stalled download exclusion`); + this.resetSegmentDownloads_(type); + this.tech_.trigger({ + type: 'usage', + name: `vhs-${type}-download-exclusion` + }); + if (type === 'subtitle') { + return; + } // TODO: should we exclude audio tracks rather than main tracks + // when type is audio? + + pc.excludePlaylist({ + error: { + message: `Excessive ${type} segment downloading detected.` + }, + playlistExclusionDuration: Infinity + }); + } + /** + * The purpose of this function is to emulate the "waiting" event on + * browsers that do not emit it when they are waiting for more + * data to continue playback + * + * @private + */ + + checkCurrentTime_() { + if (this.tech_.paused() || this.tech_.seeking()) { + return; + } + const currentTime = this.tech_.currentTime(); + const buffered = this.tech_.buffered(); + if (this.lastRecordedTime === currentTime && (!buffered.length || currentTime + SAFE_TIME_DELTA >= buffered.end(buffered.length - 1))) { + // If current time is at the end of the final buffered region, then any playback + // stall is most likely caused by buffering in a low bandwidth environment. The tech + // should fire a `waiting` event in this scenario, but due to browser and tech + // inconsistencies. Calling `techWaiting_` here allows us to simulate + // responding to a native `waiting` event when the tech fails to emit one. + return this.techWaiting_(); + } + if (this.consecutiveUpdates >= 5 && currentTime === this.lastRecordedTime) { + this.consecutiveUpdates++; + this.waiting_(); + } else if (currentTime === this.lastRecordedTime) { + this.consecutiveUpdates++; + } else { + this.consecutiveUpdates = 0; + this.lastRecordedTime = currentTime; + } + } + /** + * Resets the 'timeupdate' mechanism designed to detect that we are stalled + * + * @private + */ + + resetTimeUpdate_() { + this.consecutiveUpdates = 0; + } + /** + * Fixes situations where there's a bad seek + * + * @return {boolean} whether an action was taken to fix the seek + * @private + */ + + fixesBadSeeks_() { + const seeking = this.tech_.seeking(); + if (!seeking) { + return false; + } // TODO: It's possible that these seekable checks should be moved out of this function + // and into a function that runs on seekablechange. It's also possible that we only need + // afterSeekableWindow as the buffered check at the bottom is good enough to handle before + // seekable range. + + const seekable = this.seekable(); + const currentTime = this.tech_.currentTime(); + const isAfterSeekableRange = this.afterSeekableWindow_(seekable, currentTime, this.media(), this.allowSeeksWithinUnsafeLiveWindow); + let seekTo; + if (isAfterSeekableRange) { + const seekableEnd = seekable.end(seekable.length - 1); // sync to live point (if VOD, our seekable was updated and we're simply adjusting) + + seekTo = seekableEnd; + } + if (this.beforeSeekableWindow_(seekable, currentTime)) { + const seekableStart = seekable.start(0); // sync to the beginning of the live window + // provide a buffer of .1 seconds to handle rounding/imprecise numbers + + seekTo = seekableStart + ( + // if the playlist is too short and the seekable range is an exact time (can + // happen in live with a 3 segment playlist), then don't use a time delta + seekableStart === seekable.end(0) ? 0 : SAFE_TIME_DELTA); + } + if (typeof seekTo !== 'undefined') { + this.logger_(`Trying to seek outside of seekable at time ${currentTime} with ` + `seekable range ${printableRange(seekable)}. Seeking to ` + `${seekTo}.`); + this.tech_.setCurrentTime(seekTo); + return true; + } + const sourceUpdater = this.playlistController_.sourceUpdater_; + const buffered = this.tech_.buffered(); + const audioBuffered = sourceUpdater.audioBuffer ? sourceUpdater.audioBuffered() : null; + const videoBuffered = sourceUpdater.videoBuffer ? sourceUpdater.videoBuffered() : null; + const media = this.media(); // verify that at least two segment durations or one part duration have been + // appended before checking for a gap. + + const minAppendedDuration = media.partTargetDuration ? media.partTargetDuration : (media.targetDuration - TIME_FUDGE_FACTOR) * 2; // verify that at least two segment durations have been + // appended before checking for a gap. + + const bufferedToCheck = [audioBuffered, videoBuffered]; + for (let i = 0; i < bufferedToCheck.length; i++) { + // skip null buffered + if (!bufferedToCheck[i]) { + continue; + } + const timeAhead = timeAheadOf(bufferedToCheck[i], currentTime); // if we are less than two video/audio segment durations or one part + // duration behind we haven't appended enough to call this a bad seek. + + if (timeAhead < minAppendedDuration) { + return false; + } + } + const nextRange = findNextRange(buffered, currentTime); // we have appended enough content, but we don't have anything buffered + // to seek over the gap + + if (nextRange.length === 0) { + return false; + } + seekTo = nextRange.start(0) + SAFE_TIME_DELTA; + this.logger_(`Buffered region starts (${nextRange.start(0)}) ` + ` just beyond seek point (${currentTime}). Seeking to ${seekTo}.`); + this.tech_.setCurrentTime(seekTo); + return true; + } + /** + * Handler for situations when we determine the player is waiting. + * + * @private + */ + + waiting_() { + if (this.techWaiting_()) { + return; + } // All tech waiting checks failed. Use last resort correction + + const currentTime = this.tech_.currentTime(); + const buffered = this.tech_.buffered(); + const currentRange = findRange(buffered, currentTime); // Sometimes the player can stall for unknown reasons within a contiguous buffered + // region with no indication that anything is amiss (seen in Firefox). Seeking to + // currentTime is usually enough to kickstart the player. This checks that the player + // is currently within a buffered region before attempting a corrective seek. + // Chrome does not appear to continue `timeupdate` events after a `waiting` event + // until there is ~ 3 seconds of forward buffer available. PlaybackWatcher should also + // make sure there is ~3 seconds of forward buffer before taking any corrective action + // to avoid triggering an `unknownwaiting` event when the network is slow. + + if (currentRange.length && currentTime + 3 <= currentRange.end(0)) { + this.resetTimeUpdate_(); + this.tech_.setCurrentTime(currentTime); + this.logger_(`Stopped at ${currentTime} while inside a buffered region ` + `[${currentRange.start(0)} -> ${currentRange.end(0)}]. Attempting to resume ` + 'playback by seeking to the current time.'); // unknown waiting corrections may be useful for monitoring QoS + + this.tech_.trigger({ + type: 'usage', + name: 'vhs-unknown-waiting' + }); + return; + } + } + /** + * Handler for situations when the tech fires a `waiting` event + * + * @return {boolean} + * True if an action (or none) was needed to correct the waiting. False if no + * checks passed + * @private + */ + + techWaiting_() { + const seekable = this.seekable(); + const currentTime = this.tech_.currentTime(); + if (this.tech_.seeking()) { + // Tech is seeking or already waiting on another action, no action needed + return true; + } + if (this.beforeSeekableWindow_(seekable, currentTime)) { + const livePoint = seekable.end(seekable.length - 1); + this.logger_(`Fell out of live window at time ${currentTime}. Seeking to ` + `live point (seekable end) ${livePoint}`); + this.resetTimeUpdate_(); + this.tech_.setCurrentTime(livePoint); // live window resyncs may be useful for monitoring QoS + + this.tech_.trigger({ + type: 'usage', + name: 'vhs-live-resync' + }); + return true; + } + const sourceUpdater = this.tech_.vhs.playlistController_.sourceUpdater_; + const buffered = this.tech_.buffered(); + const videoUnderflow = this.videoUnderflow_({ + audioBuffered: sourceUpdater.audioBuffered(), + videoBuffered: sourceUpdater.videoBuffered(), + currentTime + }); + if (videoUnderflow) { + // Even though the video underflowed and was stuck in a gap, the audio overplayed + // the gap, leading currentTime into a buffered range. Seeking to currentTime + // allows the video to catch up to the audio position without losing any audio + // (only suffering ~3 seconds of frozen video and a pause in audio playback). + this.resetTimeUpdate_(); + this.tech_.setCurrentTime(currentTime); // video underflow may be useful for monitoring QoS + + this.tech_.trigger({ + type: 'usage', + name: 'vhs-video-underflow' + }); + return true; + } + const nextRange = findNextRange(buffered, currentTime); // check for gap + + if (nextRange.length > 0) { + this.logger_(`Stopped at ${currentTime} and seeking to ${nextRange.start(0)}`); + this.resetTimeUpdate_(); + this.skipTheGap_(currentTime); + return true; + } // All checks failed. Returning false to indicate failure to correct waiting + + return false; + } + afterSeekableWindow_(seekable, currentTime, playlist, allowSeeksWithinUnsafeLiveWindow = false) { + if (!seekable.length) { + // we can't make a solid case if there's no seekable, default to false + return false; + } + let allowedEnd = seekable.end(seekable.length - 1) + SAFE_TIME_DELTA; + const isLive = !playlist.endList; + const isLLHLS = typeof playlist.partTargetDuration === 'number'; + if (isLive && (isLLHLS || allowSeeksWithinUnsafeLiveWindow)) { + allowedEnd = seekable.end(seekable.length - 1) + playlist.targetDuration * 3; + } + if (currentTime > allowedEnd) { + return true; + } + return false; + } + beforeSeekableWindow_(seekable, currentTime) { + if (seekable.length && + // can't fall before 0 and 0 seekable start identifies VOD stream + seekable.start(0) > 0 && currentTime < seekable.start(0) - this.liveRangeSafeTimeDelta) { + return true; + } + return false; + } + videoUnderflow_({ + videoBuffered, + audioBuffered, + currentTime + }) { + // audio only content will not have video underflow :) + if (!videoBuffered) { + return; + } + let gap; // find a gap in demuxed content. + + if (videoBuffered.length && audioBuffered.length) { + // in Chrome audio will continue to play for ~3s when we run out of video + // so we have to check that the video buffer did have some buffer in the + // past. + const lastVideoRange = findRange(videoBuffered, currentTime - 3); + const videoRange = findRange(videoBuffered, currentTime); + const audioRange = findRange(audioBuffered, currentTime); + if (audioRange.length && !videoRange.length && lastVideoRange.length) { + gap = { + start: lastVideoRange.end(0), + end: audioRange.end(0) + }; + } // find a gap in muxed content. + } else { + const nextRange = findNextRange(videoBuffered, currentTime); // Even if there is no available next range, there is still a possibility we are + // stuck in a gap due to video underflow. + + if (!nextRange.length) { + gap = this.gapFromVideoUnderflow_(videoBuffered, currentTime); + } + } + if (gap) { + this.logger_(`Encountered a gap in video from ${gap.start} to ${gap.end}. ` + `Seeking to current time ${currentTime}`); + return true; + } + return false; + } + /** + * Timer callback. If playback still has not proceeded, then we seek + * to the start of the next buffered region. + * + * @private + */ + + skipTheGap_(scheduledCurrentTime) { + const buffered = this.tech_.buffered(); + const currentTime = this.tech_.currentTime(); + const nextRange = findNextRange(buffered, currentTime); + this.resetTimeUpdate_(); + if (nextRange.length === 0 || currentTime !== scheduledCurrentTime) { + return; + } + this.logger_('skipTheGap_:', 'currentTime:', currentTime, 'scheduled currentTime:', scheduledCurrentTime, 'nextRange start:', nextRange.start(0)); // only seek if we still have not played + + this.tech_.setCurrentTime(nextRange.start(0) + TIME_FUDGE_FACTOR); + this.tech_.trigger({ + type: 'usage', + name: 'vhs-gap-skip' + }); + } + gapFromVideoUnderflow_(buffered, currentTime) { + // At least in Chrome, if there is a gap in the video buffer, the audio will continue + // playing for ~3 seconds after the video gap starts. This is done to account for + // video buffer underflow/underrun (note that this is not done when there is audio + // buffer underflow/underrun -- in that case the video will stop as soon as it + // encounters the gap, as audio stalls are more noticeable/jarring to a user than + // video stalls). The player's time will reflect the playthrough of audio, so the + // time will appear as if we are in a buffered region, even if we are stuck in a + // "gap." + // + // Example: + // video buffer: 0 => 10.1, 10.2 => 20 + // audio buffer: 0 => 20 + // overall buffer: 0 => 10.1, 10.2 => 20 + // current time: 13 + // + // Chrome's video froze at 10 seconds, where the video buffer encountered the gap, + // however, the audio continued playing until it reached ~3 seconds past the gap + // (13 seconds), at which point it stops as well. Since current time is past the + // gap, findNextRange will return no ranges. + // + // To check for this issue, we see if there is a gap that starts somewhere within + // a 3 second range (3 seconds +/- 1 second) back from our current time. + const gaps = findGaps(buffered); + for (let i = 0; i < gaps.length; i++) { + const start = gaps.start(i); + const end = gaps.end(i); // gap is starts no more than 4 seconds back + + if (currentTime - start < 4 && currentTime - start > 2) { + return { + start, + end + }; + } + } + return null; + } +} +const defaultOptions = { + errorInterval: 30, + getSource(next) { + const tech = this.tech({ + IWillNotUseThisInPlugins: true + }); + const sourceObj = tech.currentSource_ || this.currentSource(); + return next(sourceObj); + } +}; +/** + * Main entry point for the plugin + * + * @param {Player} player a reference to a videojs Player instance + * @param {Object} [options] an object with plugin options + * @private + */ + +const initPlugin = function (player, options) { + let lastCalled = 0; + let seekTo = 0; + const localOptions = merge(defaultOptions, options); + player.ready(() => { + player.trigger({ + type: 'usage', + name: 'vhs-error-reload-initialized' + }); + }); + /** + * Player modifications to perform that must wait until `loadedmetadata` + * has been triggered + * + * @private + */ + + const loadedMetadataHandler = function () { + if (seekTo) { + player.currentTime(seekTo); + } + }; + /** + * Set the source on the player element, play, and seek if necessary + * + * @param {Object} sourceObj An object specifying the source url and mime-type to play + * @private + */ + + const setSource = function (sourceObj) { + if (sourceObj === null || sourceObj === undefined) { + return; + } + seekTo = player.duration() !== Infinity && player.currentTime() || 0; + player.one('loadedmetadata', loadedMetadataHandler); + player.src(sourceObj); + player.trigger({ + type: 'usage', + name: 'vhs-error-reload' + }); + player.play(); + }; + /** + * Attempt to get a source from either the built-in getSource function + * or a custom function provided via the options + * + * @private + */ + + const errorHandler = function () { + // Do not attempt to reload the source if a source-reload occurred before + // 'errorInterval' time has elapsed since the last source-reload + if (Date.now() - lastCalled < localOptions.errorInterval * 1000) { + player.trigger({ + type: 'usage', + name: 'vhs-error-reload-canceled' + }); + return; + } + if (!localOptions.getSource || typeof localOptions.getSource !== 'function') { + videojs.log.error('ERROR: reloadSourceOnError - The option getSource must be a function!'); + return; + } + lastCalled = Date.now(); + return localOptions.getSource.call(player, setSource); + }; + /** + * Unbind any event handlers that were bound by the plugin + * + * @private + */ + + const cleanupEvents = function () { + player.off('loadedmetadata', loadedMetadataHandler); + player.off('error', errorHandler); + player.off('dispose', cleanupEvents); + }; + /** + * Cleanup before re-initializing the plugin + * + * @param {Object} [newOptions] an object with plugin options + * @private + */ + + const reinitPlugin = function (newOptions) { + cleanupEvents(); + initPlugin(player, newOptions); + }; + player.on('error', errorHandler); + player.on('dispose', cleanupEvents); // Overwrite the plugin function so that we can correctly cleanup before + // initializing the plugin + + player.reloadSourceOnError = reinitPlugin; +}; +/** + * Reload the source when an error is detected as long as there + * wasn't an error previously within the last 30 seconds + * + * @param {Object} [options] an object with plugin options + */ + +const reloadSourceOnError = function (options) { + initPlugin(this, options); +}; +var version$4 = "3.12.1"; +var version$3 = "7.0.3"; +var version$2 = "1.3.0"; +var version$1 = "7.1.0"; +var version = "4.0.1"; + +/** + * @file videojs-http-streaming.js + * + * The main file for the VHS project. + * License: https://github.com/videojs/videojs-http-streaming/blob/main/LICENSE + */ +const Vhs = { + PlaylistLoader, + Playlist, + utils, + STANDARD_PLAYLIST_SELECTOR: lastBandwidthSelector, + INITIAL_PLAYLIST_SELECTOR: lowestBitrateCompatibleVariantSelector, + lastBandwidthSelector, + movingAverageBandwidthSelector, + comparePlaylistBandwidth, + comparePlaylistResolution, + xhr: xhrFactory() +}; // Define getter/setters for config properties + +Object.keys(Config).forEach(prop => { + Object.defineProperty(Vhs, prop, { + get() { + videojs.log.warn(`using Vhs.${prop} is UNSAFE be sure you know what you are doing`); + return Config[prop]; + }, + set(value) { + videojs.log.warn(`using Vhs.${prop} is UNSAFE be sure you know what you are doing`); + if (typeof value !== 'number' || value < 0) { + videojs.log.warn(`value of Vhs.${prop} must be greater than or equal to 0`); + return; + } + Config[prop] = value; + } + }); +}); +const LOCAL_STORAGE_KEY = 'videojs-vhs'; +/** + * Updates the selectedIndex of the QualityLevelList when a mediachange happens in vhs. + * + * @param {QualityLevelList} qualityLevels The QualityLevelList to update. + * @param {PlaylistLoader} playlistLoader PlaylistLoader containing the new media info. + * @function handleVhsMediaChange + */ + +const handleVhsMediaChange = function (qualityLevels, playlistLoader) { + const newPlaylist = playlistLoader.media(); + let selectedIndex = -1; + for (let i = 0; i < qualityLevels.length; i++) { + if (qualityLevels[i].id === newPlaylist.id) { + selectedIndex = i; + break; + } + } + qualityLevels.selectedIndex_ = selectedIndex; + qualityLevels.trigger({ + selectedIndex, + type: 'change' + }); +}; +/** + * Adds quality levels to list once playlist metadata is available + * + * @param {QualityLevelList} qualityLevels The QualityLevelList to attach events to. + * @param {Object} vhs Vhs object to listen to for media events. + * @function handleVhsLoadedMetadata + */ + +const handleVhsLoadedMetadata = function (qualityLevels, vhs) { + vhs.representations().forEach(rep => { + qualityLevels.addQualityLevel(rep); + }); + handleVhsMediaChange(qualityLevels, vhs.playlists); +}; // VHS is a source handler, not a tech. Make sure attempts to use it +// as one do not cause exceptions. + +Vhs.canPlaySource = function () { + return videojs.log.warn('VHS is no longer a tech. Please remove it from ' + 'your player\'s techOrder.'); +}; +const emeKeySystems = (keySystemOptions, mainPlaylist, audioPlaylist) => { + if (!keySystemOptions) { + return keySystemOptions; + } + let codecs = {}; + if (mainPlaylist && mainPlaylist.attributes && mainPlaylist.attributes.CODECS) { + codecs = unwrapCodecList(parseCodecs(mainPlaylist.attributes.CODECS)); + } + if (audioPlaylist && audioPlaylist.attributes && audioPlaylist.attributes.CODECS) { + codecs.audio = audioPlaylist.attributes.CODECS; + } + const videoContentType = getMimeForCodec(codecs.video); + const audioContentType = getMimeForCodec(codecs.audio); // upsert the content types based on the selected playlist + + const keySystemContentTypes = {}; + for (const keySystem in keySystemOptions) { + keySystemContentTypes[keySystem] = {}; + if (audioContentType) { + keySystemContentTypes[keySystem].audioContentType = audioContentType; + } + if (videoContentType) { + keySystemContentTypes[keySystem].videoContentType = videoContentType; + } // Default to using the video playlist's PSSH even though they may be different, as + // videojs-contrib-eme will only accept one in the options. + // + // This shouldn't be an issue for most cases as early intialization will handle all + // unique PSSH values, and if they aren't, then encrypted events should have the + // specific information needed for the unique license. + + if (mainPlaylist.contentProtection && mainPlaylist.contentProtection[keySystem] && mainPlaylist.contentProtection[keySystem].pssh) { + keySystemContentTypes[keySystem].pssh = mainPlaylist.contentProtection[keySystem].pssh; + } // videojs-contrib-eme accepts the option of specifying: 'com.some.cdm': 'url' + // so we need to prevent overwriting the URL entirely + + if (typeof keySystemOptions[keySystem] === 'string') { + keySystemContentTypes[keySystem].url = keySystemOptions[keySystem]; + } + } + return merge(keySystemOptions, keySystemContentTypes); +}; +/** + * @typedef {Object} KeySystems + * + * keySystems configuration for https://github.com/videojs/videojs-contrib-eme + * Note: not all options are listed here. + * + * @property {Uint8Array} [pssh] + * Protection System Specific Header + */ + +/** + * Goes through all the playlists and collects an array of KeySystems options objects + * containing each playlist's keySystems and their pssh values, if available. + * + * @param {Object[]} playlists + * The playlists to look through + * @param {string[]} keySystems + * The keySystems to collect pssh values for + * + * @return {KeySystems[]} + * An array of KeySystems objects containing available key systems and their + * pssh values + */ + +const getAllPsshKeySystemsOptions = (playlists, keySystems) => { + return playlists.reduce((keySystemsArr, playlist) => { + if (!playlist.contentProtection) { + return keySystemsArr; + } + const keySystemsOptions = keySystems.reduce((keySystemsObj, keySystem) => { + const keySystemOptions = playlist.contentProtection[keySystem]; + if (keySystemOptions && keySystemOptions.pssh) { + keySystemsObj[keySystem] = { + pssh: keySystemOptions.pssh + }; + } + return keySystemsObj; + }, {}); + if (Object.keys(keySystemsOptions).length) { + keySystemsArr.push(keySystemsOptions); + } + return keySystemsArr; + }, []); +}; +/** + * Returns a promise that waits for the + * [eme plugin](https://github.com/videojs/videojs-contrib-eme) to create a key session. + * + * Works around https://bugs.chromium.org/p/chromium/issues/detail?id=895449 in non-IE11 + * browsers. + * + * As per the above ticket, this is particularly important for Chrome, where, if + * unencrypted content is appended before encrypted content and the key session has not + * been created, a MEDIA_ERR_DECODE will be thrown once the encrypted content is reached + * during playback. + * + * @param {Object} player + * The player instance + * @param {Object[]} sourceKeySystems + * The key systems options from the player source + * @param {Object} [audioMedia] + * The active audio media playlist (optional) + * @param {Object[]} mainPlaylists + * The playlists found on the main playlist object + * + * @return {Object} + * Promise that resolves when the key session has been created + */ + +const waitForKeySessionCreation = ({ + player, + sourceKeySystems, + audioMedia, + mainPlaylists +}) => { + if (!player.eme.initializeMediaKeys) { + return Promise.resolve(); + } // TODO should all audio PSSH values be initialized for DRM? + // + // All unique video rendition pssh values are initialized for DRM, but here only + // the initial audio playlist license is initialized. In theory, an encrypted + // event should be fired if the user switches to an alternative audio playlist + // where a license is required, but this case hasn't yet been tested. In addition, there + // may be many alternate audio playlists unlikely to be used (e.g., multiple different + // languages). + + const playlists = audioMedia ? mainPlaylists.concat([audioMedia]) : mainPlaylists; + const keySystemsOptionsArr = getAllPsshKeySystemsOptions(playlists, Object.keys(sourceKeySystems)); + const initializationFinishedPromises = []; + const keySessionCreatedPromises = []; // Since PSSH values are interpreted as initData, EME will dedupe any duplicates. The + // only place where it should not be deduped is for ms-prefixed APIs, but + // the existence of modern EME APIs in addition to + // ms-prefixed APIs on Edge should prevent this from being a concern. + // initializeMediaKeys also won't use the webkit-prefixed APIs. + + keySystemsOptionsArr.forEach(keySystemsOptions => { + keySessionCreatedPromises.push(new Promise((resolve, reject) => { + player.tech_.one('keysessioncreated', resolve); + })); + initializationFinishedPromises.push(new Promise((resolve, reject) => { + player.eme.initializeMediaKeys({ + keySystems: keySystemsOptions + }, err => { + if (err) { + reject(err); + return; + } + resolve(); + }); + })); + }); // The reasons Promise.race is chosen over Promise.any: + // + // * Promise.any is only available in Safari 14+. + // * None of these promises are expected to reject. If they do reject, it might be + // better here for the race to surface the rejection, rather than mask it by using + // Promise.any. + + return Promise.race([ + // If a session was previously created, these will all finish resolving without + // creating a new session, otherwise it will take until the end of all license + // requests, which is why the key session check is used (to make setup much faster). + Promise.all(initializationFinishedPromises), + // Once a single session is created, the browser knows DRM will be used. + Promise.race(keySessionCreatedPromises)]); +}; +/** + * If the [eme](https://github.com/videojs/videojs-contrib-eme) plugin is available, and + * there are keySystems on the source, sets up source options to prepare the source for + * eme. + * + * @param {Object} player + * The player instance + * @param {Object[]} sourceKeySystems + * The key systems options from the player source + * @param {Object} media + * The active media playlist + * @param {Object} [audioMedia] + * The active audio media playlist (optional) + * + * @return {boolean} + * Whether or not options were configured and EME is available + */ + +const setupEmeOptions = ({ + player, + sourceKeySystems, + media, + audioMedia +}) => { + const sourceOptions = emeKeySystems(sourceKeySystems, media, audioMedia); + if (!sourceOptions) { + return false; + } + player.currentSource().keySystems = sourceOptions; // eme handles the rest of the setup, so if it is missing + // do nothing. + + if (sourceOptions && !player.eme) { + videojs.log.warn('DRM encrypted source cannot be decrypted without a DRM plugin'); + return false; + } + return true; +}; +const getVhsLocalStorage = () => { + if (!window$1.localStorage) { + return null; + } + const storedObject = window$1.localStorage.getItem(LOCAL_STORAGE_KEY); + if (!storedObject) { + return null; + } + try { + return JSON.parse(storedObject); + } catch (e) { + // someone may have tampered with the value + return null; + } +}; +const updateVhsLocalStorage = options => { + if (!window$1.localStorage) { + return false; + } + let objectToStore = getVhsLocalStorage(); + objectToStore = objectToStore ? merge(objectToStore, options) : options; + try { + window$1.localStorage.setItem(LOCAL_STORAGE_KEY, JSON.stringify(objectToStore)); + } catch (e) { + // Throws if storage is full (e.g., always on iOS 5+ Safari private mode, where + // storage is set to 0). + // https://developer.mozilla.org/en-US/docs/Web/API/Storage/setItem#Exceptions + // No need to perform any operation. + return false; + } + return objectToStore; +}; +/** + * Parses VHS-supported media types from data URIs. See + * https://developer.mozilla.org/en-US/docs/Web/HTTP/Basics_of_HTTP/Data_URIs + * for information on data URIs. + * + * @param {string} dataUri + * The data URI + * + * @return {string|Object} + * The parsed object/string, or the original string if no supported media type + * was found + */ + +const expandDataUri = dataUri => { + if (dataUri.toLowerCase().indexOf('data:application/vnd.videojs.vhs+json,') === 0) { + return JSON.parse(dataUri.substring(dataUri.indexOf(',') + 1)); + } // no known case for this data URI, return the string as-is + + return dataUri; +}; +/** + * Adds a request hook to an xhr object + * + * @param {Object} xhr object to add the onRequest hook to + * @param {function} callback hook function for an xhr request + */ + +const addOnRequestHook = (xhr, callback) => { + if (!xhr._requestCallbackSet) { + xhr._requestCallbackSet = new Set(); + } + xhr._requestCallbackSet.add(callback); +}; +/** + * Adds a response hook to an xhr object + * + * @param {Object} xhr object to add the onResponse hook to + * @param {function} callback hook function for an xhr response + */ + +const addOnResponseHook = (xhr, callback) => { + if (!xhr._responseCallbackSet) { + xhr._responseCallbackSet = new Set(); + } + xhr._responseCallbackSet.add(callback); +}; +/** + * Removes a request hook on an xhr object, deletes the onRequest set if empty. + * + * @param {Object} xhr object to remove the onRequest hook from + * @param {function} callback hook function to remove + */ + +const removeOnRequestHook = (xhr, callback) => { + if (!xhr._requestCallbackSet) { + return; + } + xhr._requestCallbackSet.delete(callback); + if (!xhr._requestCallbackSet.size) { + delete xhr._requestCallbackSet; + } +}; +/** + * Removes a response hook on an xhr object, deletes the onResponse set if empty. + * + * @param {Object} xhr object to remove the onResponse hook from + * @param {function} callback hook function to remove + */ + +const removeOnResponseHook = (xhr, callback) => { + if (!xhr._responseCallbackSet) { + return; + } + xhr._responseCallbackSet.delete(callback); + if (!xhr._responseCallbackSet.size) { + delete xhr._responseCallbackSet; + } +}; +/** + * Whether the browser has built-in HLS support. + */ + +Vhs.supportsNativeHls = function () { + if (!document || !document.createElement) { + return false; + } + const video = document.createElement('video'); // native HLS is definitely not supported if HTML5 video isn't + + if (!videojs.getTech('Html5').isSupported()) { + return false; + } // HLS manifests can go by many mime-types + + const canPlay = [ + // Apple santioned + 'application/vnd.apple.mpegurl', + // Apple sanctioned for backwards compatibility + 'audio/mpegurl', + // Very common + 'audio/x-mpegurl', + // Very common + 'application/x-mpegurl', + // Included for completeness + 'video/x-mpegurl', 'video/mpegurl', 'application/mpegurl']; + return canPlay.some(function (canItPlay) { + return /maybe|probably/i.test(video.canPlayType(canItPlay)); + }); +}(); +Vhs.supportsNativeDash = function () { + if (!document || !document.createElement || !videojs.getTech('Html5').isSupported()) { + return false; + } + return /maybe|probably/i.test(document.createElement('video').canPlayType('application/dash+xml')); +}(); +Vhs.supportsTypeNatively = type => { + if (type === 'hls') { + return Vhs.supportsNativeHls; + } + if (type === 'dash') { + return Vhs.supportsNativeDash; + } + return false; +}; +/** + * VHS is a source handler, not a tech. Make sure attempts to use it + * as one do not cause exceptions. + */ + +Vhs.isSupported = function () { + return videojs.log.warn('VHS is no longer a tech. Please remove it from ' + 'your player\'s techOrder.'); +}; +/** + * A global function for setting an onRequest hook + * + * @param {function} callback for request modifiction + */ + +Vhs.xhr.onRequest = function (callback) { + addOnRequestHook(Vhs.xhr, callback); +}; +/** + * A global function for setting an onResponse hook + * + * @param {callback} callback for response data retrieval + */ + +Vhs.xhr.onResponse = function (callback) { + addOnResponseHook(Vhs.xhr, callback); +}; +/** + * Deletes a global onRequest callback if it exists + * + * @param {function} callback to delete from the global set + */ + +Vhs.xhr.offRequest = function (callback) { + removeOnRequestHook(Vhs.xhr, callback); +}; +/** + * Deletes a global onResponse callback if it exists + * + * @param {function} callback to delete from the global set + */ + +Vhs.xhr.offResponse = function (callback) { + removeOnResponseHook(Vhs.xhr, callback); +}; +const Component = videojs.getComponent('Component'); +/** + * The Vhs Handler object, where we orchestrate all of the parts + * of VHS to interact with video.js + * + * @class VhsHandler + * @extends videojs.Component + * @param {Object} source the soruce object + * @param {Tech} tech the parent tech object + * @param {Object} options optional and required options + */ + +class VhsHandler extends Component { + constructor(source, tech, options) { + super(tech, options.vhs); // if a tech level `initialBandwidth` option was passed + // use that over the VHS level `bandwidth` option + + if (typeof options.initialBandwidth === 'number') { + this.options_.bandwidth = options.initialBandwidth; + } + this.logger_ = logger('VhsHandler'); // we need access to the player in some cases, + // so, get it from Video.js via the `playerId` + + if (tech.options_ && tech.options_.playerId) { + const _player = videojs.getPlayer(tech.options_.playerId); + this.player_ = _player; + } + this.tech_ = tech; + this.source_ = source; + this.stats = {}; + this.ignoreNextSeekingEvent_ = false; + this.setOptions_(); + if (this.options_.overrideNative && tech.overrideNativeAudioTracks && tech.overrideNativeVideoTracks) { + tech.overrideNativeAudioTracks(true); + tech.overrideNativeVideoTracks(true); + } else if (this.options_.overrideNative && (tech.featuresNativeVideoTracks || tech.featuresNativeAudioTracks)) { + // overriding native VHS only works if audio tracks have been emulated + // error early if we're misconfigured + throw new Error('Overriding native VHS requires emulated tracks. ' + 'See https://git.io/vMpjB'); + } // listen for fullscreenchange events for this player so that we + // can adjust our quality selection quickly + + this.on(document, ['fullscreenchange', 'webkitfullscreenchange', 'mozfullscreenchange', 'MSFullscreenChange'], event => { + const fullscreenElement = document.fullscreenElement || document.webkitFullscreenElement || document.mozFullScreenElement || document.msFullscreenElement; + if (fullscreenElement && fullscreenElement.contains(this.tech_.el())) { + this.playlistController_.fastQualityChange_(); + } else { + // When leaving fullscreen, since the in page pixel dimensions should be smaller + // than full screen, see if there should be a rendition switch down to preserve + // bandwidth. + this.playlistController_.checkABR_(); + } + }); + this.on(this.tech_, 'seeking', function () { + if (this.ignoreNextSeekingEvent_) { + this.ignoreNextSeekingEvent_ = false; + return; + } + this.setCurrentTime(this.tech_.currentTime()); + }); + this.on(this.tech_, 'error', function () { + // verify that the error was real and we are loaded + // enough to have pc loaded. + if (this.tech_.error() && this.playlistController_) { + this.playlistController_.pauseLoading(); + } + }); + this.on(this.tech_, 'play', this.play); + } + /** + * Set VHS options based on options from configuration, as well as partial + * options to be passed at a later time. + * + * @param {Object} options A partial chunk of config options + */ + + setOptions_(options = {}) { + this.options_ = merge(this.options_, options); // defaults + + this.options_.withCredentials = this.options_.withCredentials || false; + this.options_.limitRenditionByPlayerDimensions = this.options_.limitRenditionByPlayerDimensions === false ? false : true; + this.options_.useDevicePixelRatio = this.options_.useDevicePixelRatio || false; + this.options_.useBandwidthFromLocalStorage = typeof this.source_.useBandwidthFromLocalStorage !== 'undefined' ? this.source_.useBandwidthFromLocalStorage : this.options_.useBandwidthFromLocalStorage || false; + this.options_.useForcedSubtitles = this.options_.useForcedSubtitles || false; + this.options_.useNetworkInformationApi = this.options_.useNetworkInformationApi || false; + this.options_.useDtsForTimestampOffset = this.options_.useDtsForTimestampOffset || false; + this.options_.customTagParsers = this.options_.customTagParsers || []; + this.options_.customTagMappers = this.options_.customTagMappers || []; + this.options_.cacheEncryptionKeys = this.options_.cacheEncryptionKeys || false; + this.options_.llhls = this.options_.llhls === false ? false : true; + this.options_.bufferBasedABR = this.options_.bufferBasedABR || false; + if (typeof this.options_.playlistExclusionDuration !== 'number') { + this.options_.playlistExclusionDuration = 60; + } + if (typeof this.options_.bandwidth !== 'number') { + if (this.options_.useBandwidthFromLocalStorage) { + const storedObject = getVhsLocalStorage(); + if (storedObject && storedObject.bandwidth) { + this.options_.bandwidth = storedObject.bandwidth; + this.tech_.trigger({ + type: 'usage', + name: 'vhs-bandwidth-from-local-storage' + }); + } + if (storedObject && storedObject.throughput) { + this.options_.throughput = storedObject.throughput; + this.tech_.trigger({ + type: 'usage', + name: 'vhs-throughput-from-local-storage' + }); + } + } + } // if bandwidth was not set by options or pulled from local storage, start playlist + // selection at a reasonable bandwidth + + if (typeof this.options_.bandwidth !== 'number') { + this.options_.bandwidth = Config.INITIAL_BANDWIDTH; + } // If the bandwidth number is unchanged from the initial setting + // then this takes precedence over the enableLowInitialPlaylist option + + this.options_.enableLowInitialPlaylist = this.options_.enableLowInitialPlaylist && this.options_.bandwidth === Config.INITIAL_BANDWIDTH; // grab options passed to player.src + + ['withCredentials', 'useDevicePixelRatio', 'customPixelRatio', 'limitRenditionByPlayerDimensions', 'bandwidth', 'customTagParsers', 'customTagMappers', 'cacheEncryptionKeys', 'playlistSelector', 'initialPlaylistSelector', 'bufferBasedABR', 'liveRangeSafeTimeDelta', 'llhls', 'useForcedSubtitles', 'useNetworkInformationApi', 'useDtsForTimestampOffset', 'exactManifestTimings', 'leastPixelDiffSelector'].forEach(option => { + if (typeof this.source_[option] !== 'undefined') { + this.options_[option] = this.source_[option]; + } + }); + this.limitRenditionByPlayerDimensions = this.options_.limitRenditionByPlayerDimensions; + this.useDevicePixelRatio = this.options_.useDevicePixelRatio; + const customPixelRatio = this.options_.customPixelRatio; // Ensure the custom pixel ratio is a number greater than or equal to 0 + + if (typeof customPixelRatio === 'number' && customPixelRatio >= 0) { + this.customPixelRatio = customPixelRatio; + } + } // alias for public method to set options + + setOptions(options = {}) { + this.setOptions_(options); + } + /** + * called when player.src gets called, handle a new source + * + * @param {Object} src the source object to handle + */ + + src(src, type) { + // do nothing if the src is falsey + if (!src) { + return; + } + this.setOptions_(); // add main playlist controller options + + this.options_.src = expandDataUri(this.source_.src); + this.options_.tech = this.tech_; + this.options_.externVhs = Vhs; + this.options_.sourceType = simpleTypeFromSourceType(type); // Whenever we seek internally, we should update the tech + + this.options_.seekTo = time => { + this.tech_.setCurrentTime(time); + }; + this.playlistController_ = new PlaylistController(this.options_); + const playbackWatcherOptions = merge({ + liveRangeSafeTimeDelta: SAFE_TIME_DELTA + }, this.options_, { + seekable: () => this.seekable(), + media: () => this.playlistController_.media(), + playlistController: this.playlistController_ + }); + this.playbackWatcher_ = new PlaybackWatcher(playbackWatcherOptions); + this.playlistController_.on('error', () => { + const player = videojs.players[this.tech_.options_.playerId]; + let error = this.playlistController_.error; + if (typeof error === 'object' && !error.code) { + error.code = 3; + } else if (typeof error === 'string') { + error = { + message: error, + code: 3 + }; + } + player.error(error); + }); + const defaultSelector = this.options_.bufferBasedABR ? Vhs.movingAverageBandwidthSelector(0.55) : Vhs.STANDARD_PLAYLIST_SELECTOR; // `this` in selectPlaylist should be the VhsHandler for backwards + // compatibility with < v2 + + this.playlistController_.selectPlaylist = this.selectPlaylist ? this.selectPlaylist.bind(this) : defaultSelector.bind(this); + this.playlistController_.selectInitialPlaylist = Vhs.INITIAL_PLAYLIST_SELECTOR.bind(this); // re-expose some internal objects for backwards compatibility with < v2 + + this.playlists = this.playlistController_.mainPlaylistLoader_; + this.mediaSource = this.playlistController_.mediaSource; // Proxy assignment of some properties to the main playlist + // controller. Using a custom property for backwards compatibility + // with < v2 + + Object.defineProperties(this, { + selectPlaylist: { + get() { + return this.playlistController_.selectPlaylist; + }, + set(selectPlaylist) { + this.playlistController_.selectPlaylist = selectPlaylist.bind(this); + } + }, + throughput: { + get() { + return this.playlistController_.mainSegmentLoader_.throughput.rate; + }, + set(throughput) { + this.playlistController_.mainSegmentLoader_.throughput.rate = throughput; // By setting `count` to 1 the throughput value becomes the starting value + // for the cumulative average + + this.playlistController_.mainSegmentLoader_.throughput.count = 1; + } + }, + bandwidth: { + get() { + let playerBandwidthEst = this.playlistController_.mainSegmentLoader_.bandwidth; + const networkInformation = window$1.navigator.connection || window$1.navigator.mozConnection || window$1.navigator.webkitConnection; + const tenMbpsAsBitsPerSecond = 10e6; + if (this.options_.useNetworkInformationApi && networkInformation) { + // downlink returns Mbps + // https://developer.mozilla.org/en-US/docs/Web/API/NetworkInformation/downlink + const networkInfoBandwidthEstBitsPerSec = networkInformation.downlink * 1000 * 1000; // downlink maxes out at 10 Mbps. In the event that both networkInformationApi and the player + // estimate a bandwidth greater than 10 Mbps, use the larger of the two estimates to ensure that + // high quality streams are not filtered out. + + if (networkInfoBandwidthEstBitsPerSec >= tenMbpsAsBitsPerSecond && playerBandwidthEst >= tenMbpsAsBitsPerSecond) { + playerBandwidthEst = Math.max(playerBandwidthEst, networkInfoBandwidthEstBitsPerSec); + } else { + playerBandwidthEst = networkInfoBandwidthEstBitsPerSec; + } + } + return playerBandwidthEst; + }, + set(bandwidth) { + this.playlistController_.mainSegmentLoader_.bandwidth = bandwidth; // setting the bandwidth manually resets the throughput counter + // `count` is set to zero that current value of `rate` isn't included + // in the cumulative average + + this.playlistController_.mainSegmentLoader_.throughput = { + rate: 0, + count: 0 + }; + } + }, + /** + * `systemBandwidth` is a combination of two serial processes bit-rates. The first + * is the network bitrate provided by `bandwidth` and the second is the bitrate of + * the entire process after that - decryption, transmuxing, and appending - provided + * by `throughput`. + * + * Since the two process are serial, the overall system bandwidth is given by: + * sysBandwidth = 1 / (1 / bandwidth + 1 / throughput) + */ + systemBandwidth: { + get() { + const invBandwidth = 1 / (this.bandwidth || 1); + let invThroughput; + if (this.throughput > 0) { + invThroughput = 1 / this.throughput; + } else { + invThroughput = 0; + } + const systemBitrate = Math.floor(1 / (invBandwidth + invThroughput)); + return systemBitrate; + }, + set() { + videojs.log.error('The "systemBandwidth" property is read-only'); + } + } + }); + if (this.options_.bandwidth) { + this.bandwidth = this.options_.bandwidth; + } + if (this.options_.throughput) { + this.throughput = this.options_.throughput; + } + Object.defineProperties(this.stats, { + bandwidth: { + get: () => this.bandwidth || 0, + enumerable: true + }, + mediaRequests: { + get: () => this.playlistController_.mediaRequests_() || 0, + enumerable: true + }, + mediaRequestsAborted: { + get: () => this.playlistController_.mediaRequestsAborted_() || 0, + enumerable: true + }, + mediaRequestsTimedout: { + get: () => this.playlistController_.mediaRequestsTimedout_() || 0, + enumerable: true + }, + mediaRequestsErrored: { + get: () => this.playlistController_.mediaRequestsErrored_() || 0, + enumerable: true + }, + mediaTransferDuration: { + get: () => this.playlistController_.mediaTransferDuration_() || 0, + enumerable: true + }, + mediaBytesTransferred: { + get: () => this.playlistController_.mediaBytesTransferred_() || 0, + enumerable: true + }, + mediaSecondsLoaded: { + get: () => this.playlistController_.mediaSecondsLoaded_() || 0, + enumerable: true + }, + mediaAppends: { + get: () => this.playlistController_.mediaAppends_() || 0, + enumerable: true + }, + mainAppendsToLoadedData: { + get: () => this.playlistController_.mainAppendsToLoadedData_() || 0, + enumerable: true + }, + audioAppendsToLoadedData: { + get: () => this.playlistController_.audioAppendsToLoadedData_() || 0, + enumerable: true + }, + appendsToLoadedData: { + get: () => this.playlistController_.appendsToLoadedData_() || 0, + enumerable: true + }, + timeToLoadedData: { + get: () => this.playlistController_.timeToLoadedData_() || 0, + enumerable: true + }, + buffered: { + get: () => timeRangesToArray(this.tech_.buffered()), + enumerable: true + }, + currentTime: { + get: () => this.tech_.currentTime(), + enumerable: true + }, + currentSource: { + get: () => this.tech_.currentSource_, + enumerable: true + }, + currentTech: { + get: () => this.tech_.name_, + enumerable: true + }, + duration: { + get: () => this.tech_.duration(), + enumerable: true + }, + main: { + get: () => this.playlists.main, + enumerable: true + }, + playerDimensions: { + get: () => this.tech_.currentDimensions(), + enumerable: true + }, + seekable: { + get: () => timeRangesToArray(this.tech_.seekable()), + enumerable: true + }, + timestamp: { + get: () => Date.now(), + enumerable: true + }, + videoPlaybackQuality: { + get: () => this.tech_.getVideoPlaybackQuality(), + enumerable: true + } + }); + this.tech_.one('canplay', this.playlistController_.setupFirstPlay.bind(this.playlistController_)); + this.tech_.on('bandwidthupdate', () => { + if (this.options_.useBandwidthFromLocalStorage) { + updateVhsLocalStorage({ + bandwidth: this.bandwidth, + throughput: Math.round(this.throughput) + }); + } + }); + this.playlistController_.on('selectedinitialmedia', () => { + // Add the manual rendition mix-in to VhsHandler + renditionSelectionMixin(this); + }); + this.playlistController_.sourceUpdater_.on('createdsourcebuffers', () => { + this.setupEme_(); + }); // the bandwidth of the primary segment loader is our best + // estimate of overall bandwidth + + this.on(this.playlistController_, 'progress', function () { + this.tech_.trigger('progress'); + }); // In the live case, we need to ignore the very first `seeking` event since + // that will be the result of the seek-to-live behavior + + this.on(this.playlistController_, 'firstplay', function () { + this.ignoreNextSeekingEvent_ = true; + }); + this.setupQualityLevels_(); // do nothing if the tech has been disposed already + // this can occur if someone sets the src in player.ready(), for instance + + if (!this.tech_.el()) { + return; + } + this.mediaSourceUrl_ = window$1.URL.createObjectURL(this.playlistController_.mediaSource); + this.tech_.src(this.mediaSourceUrl_); + } + createKeySessions_() { + const audioPlaylistLoader = this.playlistController_.mediaTypes_.AUDIO.activePlaylistLoader; + this.logger_('waiting for EME key session creation'); + waitForKeySessionCreation({ + player: this.player_, + sourceKeySystems: this.source_.keySystems, + audioMedia: audioPlaylistLoader && audioPlaylistLoader.media(), + mainPlaylists: this.playlists.main.playlists + }).then(() => { + this.logger_('created EME key session'); + this.playlistController_.sourceUpdater_.initializedEme(); + }).catch(err => { + this.logger_('error while creating EME key session', err); + this.player_.error({ + message: 'Failed to initialize media keys for EME', + code: 3, + metadata: { + errorType: videojs.Error.EMEKeySessionCreationError + } + }); + }); + } + handleWaitingForKey_() { + // If waitingforkey is fired, it's possible that the data that's necessary to retrieve + // the key is in the manifest. While this should've happened on initial source load, it + // may happen again in live streams where the keys change, and the manifest info + // reflects the update. + // + // Because videojs-contrib-eme compares the PSSH data we send to that of PSSH data it's + // already requested keys for, we don't have to worry about this generating extraneous + // requests. + this.logger_('waitingforkey fired, attempting to create any new key sessions'); + this.createKeySessions_(); + } + /** + * If necessary and EME is available, sets up EME options and waits for key session + * creation. + * + * This function also updates the source updater so taht it can be used, as for some + * browsers, EME must be configured before content is appended (if appending unencrypted + * content before encrypted content). + */ + + setupEme_() { + const audioPlaylistLoader = this.playlistController_.mediaTypes_.AUDIO.activePlaylistLoader; + const didSetupEmeOptions = setupEmeOptions({ + player: this.player_, + sourceKeySystems: this.source_.keySystems, + media: this.playlists.media(), + audioMedia: audioPlaylistLoader && audioPlaylistLoader.media() + }); + this.player_.tech_.on('keystatuschange', e => { + this.playlistController_.updatePlaylistByKeyStatus(e.keyId, e.status); + }); + this.handleWaitingForKey_ = this.handleWaitingForKey_.bind(this); + this.player_.tech_.on('waitingforkey', this.handleWaitingForKey_); + if (!didSetupEmeOptions) { + // If EME options were not set up, we've done all we could to initialize EME. + this.playlistController_.sourceUpdater_.initializedEme(); + return; + } + this.createKeySessions_(); + } + /** + * Initializes the quality levels and sets listeners to update them. + * + * @method setupQualityLevels_ + * @private + */ + + setupQualityLevels_() { + const player = videojs.players[this.tech_.options_.playerId]; // if there isn't a player or there isn't a qualityLevels plugin + // or qualityLevels_ listeners have already been setup, do nothing. + + if (!player || !player.qualityLevels || this.qualityLevels_) { + return; + } + this.qualityLevels_ = player.qualityLevels(); + this.playlistController_.on('selectedinitialmedia', () => { + handleVhsLoadedMetadata(this.qualityLevels_, this); + }); + this.playlists.on('mediachange', () => { + handleVhsMediaChange(this.qualityLevels_, this.playlists); + }); + } + /** + * return the version + */ + + static version() { + return { + '@videojs/http-streaming': version$4, + 'mux.js': version$3, + 'mpd-parser': version$2, + 'm3u8-parser': version$1, + 'aes-decrypter': version + }; + } + /** + * return the version + */ + + version() { + return this.constructor.version(); + } + canChangeType() { + return SourceUpdater.canChangeType(); + } + /** + * Begin playing the video. + */ + + play() { + this.playlistController_.play(); + } + /** + * a wrapper around the function in PlaylistController + */ + + setCurrentTime(currentTime) { + this.playlistController_.setCurrentTime(currentTime); + } + /** + * a wrapper around the function in PlaylistController + */ + + duration() { + return this.playlistController_.duration(); + } + /** + * a wrapper around the function in PlaylistController + */ + + seekable() { + return this.playlistController_.seekable(); + } + /** + * Abort all outstanding work and cleanup. + */ + + dispose() { + if (this.playbackWatcher_) { + this.playbackWatcher_.dispose(); + } + if (this.playlistController_) { + this.playlistController_.dispose(); + } + if (this.qualityLevels_) { + this.qualityLevels_.dispose(); + } + if (this.tech_ && this.tech_.vhs) { + delete this.tech_.vhs; + } + if (this.mediaSourceUrl_ && window$1.URL.revokeObjectURL) { + window$1.URL.revokeObjectURL(this.mediaSourceUrl_); + this.mediaSourceUrl_ = null; + } + if (this.tech_) { + this.tech_.off('waitingforkey', this.handleWaitingForKey_); + } + super.dispose(); + } + convertToProgramTime(time, callback) { + return getProgramTime({ + playlist: this.playlistController_.media(), + time, + callback + }); + } // the player must be playing before calling this + + seekToProgramTime(programTime, callback, pauseAfterSeek = true, retryCount = 2) { + return seekToProgramTime({ + programTime, + playlist: this.playlistController_.media(), + retryCount, + pauseAfterSeek, + seekTo: this.options_.seekTo, + tech: this.options_.tech, + callback + }); + } + /** + * Adds the onRequest, onResponse, offRequest and offResponse functions + * to the VhsHandler xhr Object. + */ + + setupXhrHooks_() { + /** + * A player function for setting an onRequest hook + * + * @param {function} callback for request modifiction + */ + this.xhr.onRequest = callback => { + addOnRequestHook(this.xhr, callback); + }; + /** + * A player function for setting an onResponse hook + * + * @param {callback} callback for response data retrieval + */ + + this.xhr.onResponse = callback => { + addOnResponseHook(this.xhr, callback); + }; + /** + * Deletes a player onRequest callback if it exists + * + * @param {function} callback to delete from the player set + */ + + this.xhr.offRequest = callback => { + removeOnRequestHook(this.xhr, callback); + }; + /** + * Deletes a player onResponse callback if it exists + * + * @param {function} callback to delete from the player set + */ + + this.xhr.offResponse = callback => { + removeOnResponseHook(this.xhr, callback); + }; // Trigger an event on the player to notify the user that vhs is ready to set xhr hooks. + // This allows hooks to be set before the source is set to vhs when handleSource is called. + + this.player_.trigger('xhr-hooks-ready'); + } +} +/** + * The Source Handler object, which informs video.js what additional + * MIME types are supported and sets up playback. It is registered + * automatically to the appropriate tech based on the capabilities of + * the browser it is running in. It is not necessary to use or modify + * this object in normal usage. + */ + +const VhsSourceHandler = { + name: 'videojs-http-streaming', + VERSION: version$4, + canHandleSource(srcObj, options = {}) { + const localOptions = merge(videojs.options, options); + return VhsSourceHandler.canPlayType(srcObj.type, localOptions); + }, + handleSource(source, tech, options = {}) { + const localOptions = merge(videojs.options, options); + tech.vhs = new VhsHandler(source, tech, localOptions); + tech.vhs.xhr = xhrFactory(); + tech.vhs.setupXhrHooks_(); + tech.vhs.src(source.src, source.type); + return tech.vhs; + }, + canPlayType(type, options) { + const simpleType = simpleTypeFromSourceType(type); + if (!simpleType) { + return ''; + } + const overrideNative = VhsSourceHandler.getOverrideNative(options); + const supportsTypeNatively = Vhs.supportsTypeNatively(simpleType); + const canUseMsePlayback = !supportsTypeNatively || overrideNative; + return canUseMsePlayback ? 'maybe' : ''; + }, + getOverrideNative(options = {}) { + const { + vhs = {} + } = options; + const defaultOverrideNative = !(videojs.browser.IS_ANY_SAFARI || videojs.browser.IS_IOS); + const { + overrideNative = defaultOverrideNative + } = vhs; + return overrideNative; + } +}; +/** + * Check to see if the native MediaSource object exists and supports + * an MP4 container with both H.264 video and AAC-LC audio. + * + * @return {boolean} if native media sources are supported + */ + +const supportsNativeMediaSources = () => { + return browserSupportsCodec('avc1.4d400d,mp4a.40.2'); +}; // register source handlers with the appropriate techs + +if (supportsNativeMediaSources()) { + videojs.getTech('Html5').registerSourceHandler(VhsSourceHandler, 0); +} +videojs.VhsHandler = VhsHandler; +videojs.VhsSourceHandler = VhsSourceHandler; +videojs.Vhs = Vhs; +if (!videojs.use) { + videojs.registerComponent('Vhs', Vhs); +} +videojs.options.vhs = videojs.options.vhs || {}; +if (!videojs.getPlugin || !videojs.getPlugin('reloadSourceOnError')) { + videojs.registerPlugin('reloadSourceOnError', reloadSourceOnError); +} + +export { videojs as default }; |