prompt
large_stringlengths
70
991k
completion
large_stringlengths
0
1.02k
<|file_name|>CpuSpyApp.java<|end_file_name|><|fim▁begin|>//----------------------------------------------------------------------------- // // (C) Brandon Valosek, 2011 <[email protected]> // //----------------------------------------------------------------------------- // Modified by Willi Ye to work with big.LITTLE package com.bvalosek.cpuspy; import android.app.Application; import android.content.Context; import com.swapnil133609.zeuscontrols.utils.Utils; import java.util.HashMap; import java.util.Map; /** * main application class */ public class CpuSpyApp extends Application { private final String PREF_OFFSETS; /** * the long-living object used to monitor the system frequency states */ private final CpuStateMonitor _monitor; public CpuSpyApp(int core) { PREF_OFFSETS = "offsets" + core; _monitor = new CpuStateMonitor(core); } /** * On application start, load the saved offsets and stash the current kernel * version string */<|fim▁hole|> } /** * @return the internal CpuStateMonitor object */ public CpuStateMonitor getCpuStateMonitor() { return _monitor; } /** * Load the saved string of offsets from preferences and put it into the * state monitor */ public void loadOffsets(Context context) { String prefs = Utils.getString(PREF_OFFSETS, "", context); if (prefs.length() < 1) return; // split the string by peroids and then the info by commas and load Map<Integer, Long> offsets = new HashMap<>(); String[] sOffsets = prefs.split(","); for (String offset : sOffsets) { String[] parts = offset.split(" "); offsets.put(Utils.stringToInt(parts[0]), Utils.stringToLong(parts[1])); } _monitor.setOffsets(offsets); } /** * Save the state-time offsets as a string e.g. "100 24, 200 251, 500 124 * etc */ public void saveOffsets(Context context) { // build the string by iterating over the freq->duration map String str = ""; for (Map.Entry<Integer, Long> entry : _monitor.getOffsets().entrySet()) str += entry.getKey() + " " + entry.getValue() + ","; Utils.saveString(PREF_OFFSETS, str, context); } }<|fim▁end|>
@Override public void onCreate() { super.onCreate(); loadOffsets(getApplicationContext());
<|file_name|>test_edit_contact.py<|end_file_name|><|fim▁begin|>from model.contact import Contact from random import randrange def test_edit_contact(app, db, check_ui): if len(db.get_contact_list()) == 0: app.contact.create(Contact(first_name ="Sabina", last_name="test", company="Pewex", address="osiedle", phone_home="123456789", e_mail="[email protected]", year="2016",)) old_contact = db.get_contact_list() index = randrange(len(old_contact)) contact = Contact(first_name='Kasia', last_name='Bober')<|fim▁hole|> assert len(old_contact) == app.contact.count() new_contact = db.get_contact_list() old_contact[index] = contact assert old_contact == new_contact if check_ui: assert sorted(new_contact, key=Contact.id_or_max) == sorted( app.group.get_contact_list(), key=Contact.id_or_max )<|fim▁end|>
contact.id = old_contact[index].id app.contact.edit_contact_by_index(index, contact)
<|file_name|>pipe.rs<|end_file_name|><|fim▁begin|>//! //! Desync pipes provide a way to generate and process streams via a `Desync` object //! //! Pipes are an excellent way to interface `Desync` objects and the futures library. Piping //! a stream into a `Desync` object is equivalent to spawning it with an executor, except //! without the need to dedicate a thread to running it. //! //! There are two kinds of pipe. The `pipe_in` function creates a pipe that processes each //! value made available from a stream on a desync object as they arrive, producing no //! results. This is useful for cases where a `Desync` object is being used as the endpoint //! for some data processing (for example, to insert the results of an operation into an //! asynchronous database object). //! //! The `pipe` function pipes data through an object. For every input value, it produces //! an output value. This is good for creating streams that perform some kind of asynchronous //! processing operation or that need to access data from inside a `Desync` object. //! //! Here's an example of using `pipe_in` to store data in a `HashSet`: //! //! ``` //! # extern crate futures; //! # extern crate desync; //! # use std::collections::HashSet; //! # use std::sync::*; //! # //! use futures::future; //! use futures::channel::mpsc; //! use futures::executor; //! use futures::prelude::*; //! # use ::desync::*; //! //! executor::block_on(async { //! let desync_hashset = Arc::new(Desync::new(HashSet::new())); //! let (mut sender, receiver) = mpsc::channel(5); //! //! pipe_in(Arc::clone(&desync_hashset), receiver, |hashset, value| { hashset.insert(value); future::ready(()).boxed() }); //! //! sender.send("Test".to_string()).await.unwrap(); //! sender.send("Another value".to_string()).await.unwrap(); //! # //! # assert!(desync_hashset.sync(|hashset| hashset.contains(&("Test".to_string())))) //! }); //! ``` //! use super::desync::*; use futures::*; use futures::future::{BoxFuture}; use futures::stream::{Stream}; use futures::task; use futures::task::{Poll, Context}; use std::sync::*; use std::pin::{Pin}; use std::collections::VecDeque; lazy_static! { /// Desync for disposing of references used in pipes (if a pipe is closed with pending data, this avoids clearing it in the same context as the pipe monitor) static ref REFERENCE_CHUTE: Desync<()> = Desync::new(()); } /// The maximum number of items to queue on a pipe stream before we stop accepting new input const PIPE_BACKPRESSURE_COUNT: usize = 5; /// /// Futures notifier used to wake up a pipe when a stream or future notifies /// struct PipeContext<Core, PollFn> where Core: Send+Unpin { /// The desync target that will be woken when the stream notifies that it needs to be polled /// We keep a weak reference so that if the stream/future is all that's left referencing the /// desync, it's thrown away target: Weak<Desync<Core>>, /// The function that should be called to poll this stream. Returns false if we should no longer keep polling the stream poll_fn: Arc<Mutex<Option<PollFn>>> } impl<Core, PollFn> PipeContext<Core, PollFn> where Core: 'static+Send+Unpin, PollFn: 'static+Send+for<'a> FnMut(&'a mut Core, Context<'a>) -> BoxFuture<'a, bool> { /// /// Creates a new pipe context, ready to poll /// fn new(target: &Arc<Desync<Core>>, poll_fn: PollFn) -> Arc<PipeContext<Core, PollFn>> { let context = PipeContext { target: Arc::downgrade(target), poll_fn: Arc::new(Mutex::new(Some(poll_fn))) }; Arc::new(context) } /// /// Triggers the poll function in the context of the target desync /// fn poll(arc_self: Arc<Self>) { // If the desync is stil live... if let Some(target) = arc_self.target.upgrade() { // Grab the poll function let maybe_poll_fn = Arc::clone(&arc_self.poll_fn); // Schedule a polling operation on the desync let _ = target.future_desync(move |core| { async move { // Create a futures context from the context reference let waker = task::waker_ref(&arc_self); let context = Context::from_waker(&waker); // Pass in to the poll function let future_poll = { let mut maybe_poll_fn = maybe_poll_fn.lock().unwrap(); let future_poll = maybe_poll_fn.as_mut().map(|poll_fn| (poll_fn)(core, context)); future_poll }; if let Some(future_poll) = future_poll { let keep_polling = future_poll.await; if !keep_polling { // Deallocate the function when it's time to stop polling altogether (*arc_self.poll_fn.lock().unwrap()) = None; } } }.boxed() }); } else { // Stream has woken up but the desync is no longer listening (*arc_self.poll_fn.lock().unwrap()) = None; } } } impl<Core, PollFn> task::ArcWake for PipeContext<Core, PollFn> where Core: 'static+Send+Unpin, PollFn: 'static+Send+for<'a> FnMut(&'a mut Core, Context<'a>) -> BoxFuture<'a, bool> { fn wake_by_ref(arc_self: &Arc<Self>) { Self::poll(Arc::clone(arc_self)); } fn wake(self: Arc<Self>) { Self::poll(self); } } /// /// Pipes a stream into a desync object. Whenever an item becomes available on the stream, the /// processing function is called asynchronously with the item that was received. /// /// This takes a weak reference to the passed in `Desync` object, so the pipe will stop if it's /// the only thing referencing this object. /// /// Piping a stream to a `Desync` like this will cause it to start executing: ie, this is /// similar to calling `executor::spawn(stream)`, except that the stream will immediately /// start draining into the `Desync` object. /// pub fn pipe_in<Core, S, ProcessFn>(desync: Arc<Desync<Core>>, stream: S, process: ProcessFn) where Core: 'static+Send+Unpin, S: 'static+Send+Unpin+Stream, S::Item: Send, ProcessFn: 'static+Send+for<'a> FnMut(&'a mut Core, S::Item) -> BoxFuture<'a, ()> { let stream = Arc::new(Mutex::new(stream)); let process = Arc::new(Mutex::new(process)); // The context is used to trigger polling of the stream let context = PipeContext::new(&desync, move |core, context| { let process = Arc::clone(&process); let stream = Arc::clone(&stream); async move { let mut context = context; loop { // Poll the stream let next = stream.lock().unwrap().poll_next_unpin(&mut context); match next { // Wait for notification when the stream goes pending Poll::Pending => return true, // Stop polling when the stream stops generating new events Poll::Ready(None) => return false, // Invoke the callback when there's some data on the stream Poll::Ready(Some(next)) => { let process_future = (&mut *process.lock().unwrap())(core, next); process_future.await; } } } }.boxed() }); // Trigger the initial poll PipeContext::poll(context); } /// /// Pipes a stream into this object. Whenever an item becomes available on the stream, the /// processing function is called asynchronously with the item that was received. The /// return value is placed onto the output stream. /// /// Unlike `pipe_in`, this keeps a strong reference to the `Desync` object so the processing /// will continue so long as the input stream has data and the output stream is not dropped. /// /// The input stream will start executing and reading values immediately when this is called. /// Dropping the output stream will cause the pipe to be closed (the input stream will be /// dropped and no further processing will occur). /// /// This example demonstrates how to create a simple demonstration pipe that takes hashset values /// and returns a stream indicating whether or not they were already included: /// /// ``` /// # extern crate futures; /// # extern crate desync; /// # use std::collections::HashSet; /// # use std::sync::*; /// # /// use futures::prelude::*; /// use futures::future; /// use futures::channel::mpsc; /// use futures::executor; /// # use ::desync::*; /// /// executor::block_on(async { /// let desync_hashset = Arc::new(Desync::new(HashSet::new())); /// let (mut sender, receiver) = mpsc::channel::<String>(5); /// /// let mut value_inserted = pipe(Arc::clone(&desync_hashset), receiver, /// |hashset, value| { future::ready((value.clone(), hashset.insert(value))).boxed() }); /// /// sender.send("Test".to_string()).await.unwrap(); /// sender.send("Another value".to_string()).await.unwrap(); /// sender.send("Test".to_string()).await.unwrap(); /// /// assert!(value_inserted.next().await == Some(("Test".to_string(), true))); /// assert!(value_inserted.next().await == Some(("Another value".to_string(), true))); /// assert!(value_inserted.next().await == Some(("Test".to_string(), false))); /// }); /// ``` /// pub fn pipe<Core, S, Output, ProcessFn>(desync: Arc<Desync<Core>>, stream: S, process: ProcessFn) -> PipeStream<Output> where Core: 'static+Send+Unpin, S: 'static+Send+Unpin+Stream, S::Item: Send, Output: 'static+Send, ProcessFn: 'static+Send+for <'a> FnMut(&'a mut Core, S::Item) -> BoxFuture<'a, Output> { // Prepare the streams let input_stream = Arc::new(Mutex::new(stream)); let process = Arc::new(Mutex::new(process)); let mut output_desync = Some(Arc::clone(&desync)); let output_stream = PipeStream::<Output>::new(move || { output_desync.take(); }); // Get the core from the output stream let stream_core = Arc::clone(&output_stream.core); let stream_core = Arc::downgrade(&stream_core); // Create the read context let context = PipeContext::new(&desync, move |core, context| { let stream_core = stream_core.upgrade(); let mut context = context; let input_stream = Arc::clone(&input_stream); let process = Arc::clone(&process); async move { if let Some(stream_core) = stream_core { // Defer processing if the stream core is full let is_closed = { // Fetch the core let mut stream_core = stream_core.lock().unwrap(); // If the pending queue is full, then stop processing events if stream_core.pending.len() >= stream_core.max_pipe_depth { // Wake when the stream accepts some input stream_core.backpressure_release_notify = Some(context.waker().clone()); // Go back to sleep without reading from the stream return true; } // If the core is closed, finish up stream_core.closed }; // Wake up the stream and stop reading from the stream if the core is closed if is_closed { let notify = { stream_core.lock().unwrap().notify.take() }; notify.map(|notify| notify.wake()); return false; } loop { // Disable the notifier if there was one left over stream_core.lock().unwrap().notify_stream_closed = None; // Poll the stream let next = input_stream.lock().unwrap().poll_next_unpin(&mut context);<|fim▁hole|> match next { // Wait for notification when the stream goes pending Poll::Pending => { stream_core.lock().unwrap().notify_stream_closed = Some(context.waker().clone()); return true }, // Stop polling when the stream stops generating new events Poll::Ready(None) => { // Mark the stream as closed, and wake it up let notify = { let mut stream_core = stream_core.lock().unwrap(); stream_core.closed = true; stream_core.notify.take() }; notify.map(|notify| notify.wake()); return false; }, // Invoke the callback when there's some data on the stream Poll::Ready(Some(next)) => { // Pipe the next item through let next_item = (&mut *process.lock().unwrap())(core, next); let next_item = next_item.await; // Send to the pipe stream, and wake it up let notify = { let mut stream_core = stream_core.lock().unwrap(); stream_core.pending.push_back(next_item); stream_core.notify.take() }; notify.map(|notify| notify.wake()); } } } } else { // The stream core has been released return false; } }.boxed() }); // Poll the context to start the stream running PipeContext::poll(context); output_stream } /// /// The shared data for a pipe stream /// struct PipeStreamCore<Item> { /// The maximum number of items we allow to be queued in this stream before producing backpressure max_pipe_depth: usize, /// The pending data for this stream pending: VecDeque<Item>, /// True if the input stream has closed (the stream is closed once this is true and there are no more pending items) closed: bool, /// The task to notify when the stream changes notify: Option<task::Waker>, /// The task to notify when the stream changes notify_stream_closed: Option<task::Waker>, /// The task to notify when we reduce the amount of pending data backpressure_release_notify: Option<task::Waker> } /// /// A stream generated by a pipe /// pub struct PipeStream<Item> { /// The core contains the stream state core: Arc<Mutex<PipeStreamCore<Item>>>, /// Called when dropping the output stream on_drop: Option<Box<dyn Send+Sync+FnMut() -> ()>> } impl<Item> PipeStream<Item> { /// /// Creates a new, empty, pipestream /// fn new<FnOnDrop: 'static+Send+Sync+FnMut() -> ()>(on_drop: FnOnDrop) -> PipeStream<Item> { PipeStream { core: Arc::new(Mutex::new(PipeStreamCore { max_pipe_depth: PIPE_BACKPRESSURE_COUNT, pending: VecDeque::new(), closed: false, notify: None, notify_stream_closed: None, backpressure_release_notify: None })), on_drop: Some(Box::new(on_drop)) } } /// /// Sets the number of items that this pipe stream will buffer before producing backpressure /// /// If this call is not made, this will be set to 5. /// pub fn set_backpressure_depth(&mut self, max_depth: usize) { self.core.lock().unwrap().max_pipe_depth = max_depth; } } impl<Item> Drop for PipeStream<Item> { fn drop(&mut self) { let mut core = self.core.lock().unwrap(); // Flush the pending queue core.pending = VecDeque::new(); // Mark the core as closed to stop it from reading from the stream core.closed = true; // Wake the stream to finish closing it core.notify_stream_closed.take().map(|notify_stream_closed| notify_stream_closed.wake()); // Run the drop function self.on_drop.take().map(|mut on_drop| { REFERENCE_CHUTE.desync(move |_| { (on_drop)() }) }); } } impl<Item> Stream for PipeStream<Item> { type Item = Item; fn poll_next(self: Pin<&mut Self>, context: &mut Context) -> Poll<Option<Item>> { let (result, notify) = { // Fetch the state from the core (locking the state) let mut core = self.core.lock().unwrap(); if let Some(item) = core.pending.pop_front() { // Value waiting at the start of the stream let notify_backpressure = core.backpressure_release_notify.take(); (Poll::Ready(Some(item)), notify_backpressure) } else if core.closed { // No more data will be returned from this stream (Poll::Ready(None), None) } else { // Stream not ready let notify_backpressure = core.backpressure_release_notify.take(); core.notify = Some(context.waker().clone()); (Poll::Pending, notify_backpressure) } }; // If anything needs notifying, do so outside of the lock notify.map(|notify| notify.wake()); result } }<|fim▁end|>
<|file_name|>__init__.py<|end_file_name|><|fim▁begin|># -*- coding: utf-8 -*-<|fim▁hole|><|fim▁end|>
# License AGPL-3 - See http://www.gnu.org/licenses/agpl-3.0.html from . import test_account_payment_transfer_reconcile_batch
<|file_name|>__init__.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python<|fim▁hole|>.. moduleauthor:: Pat Daburu <[email protected]> Provide a brief description of the module. """<|fim▁end|>
# -*- coding: utf-8 -*- """ .. currentmodule:: __init__.py
<|file_name|>resource.py<|end_file_name|><|fim▁begin|>import hashlib import inspect import json import logging import re from xml.sax.saxutils import escape as xml_escape import webob from wsgiservice import xmlserializer from wsgiservice.decorators import mount from wsgiservice.exceptions import (MultiValidationException, ResponseException, ValidationException) from wsgiservice.status import * logger = logging.getLogger(__name__) class Resource(object): """Base class for all WsgiService resources. A resource is a unique REST endpoint which accepts different methods for different actions. For each HTTP call the corresponding method (equal to the HTTP method) will be called. """ #: The root tag for generated XML output. Used by :func:`to_text_xml`. #: (Default: 'response') XML_ROOT_TAG = 'response' #: List of the known HTTP methods. Used by :func:`get_method` to handle #: methods that are not implemented. (Default: All methods defined by the #: HTTP 1.1 standard :rfc:`2616`) KNOWN_METHODS = ['OPTIONS', 'GET', 'HEAD', 'POST', 'PUT', 'DELETE', 'TRACE', 'CONNECT'] #: List of tuples mapping file extensions to MIME types. The first item of #: the tuple is the extension and the second is the associated MIME type. #: Used by :func:`get_content_type` to determine the requested MIME type. #: (Default: '.xml' and '.json'). EXTENSION_MAP = [ ('.xml', 'text/xml'), ('.json', 'application/json'), ] #: A tuple of exceptions that should be treated as 404. An ideal candidate #: is KeyError if you do dictionary accesses. Used by :func:`call` which #: calls :func:`handle_exception_404` whenever an exception from this #: tuple occurs. (Default: Empty tuple) NOT_FOUND = () #: A tuple of absolute paths that should return a 404. By default this is #: used to ignored requests for favicon.ico and robots.txt so that #: browsers don't cause too many exceptions. IGNORED_PATHS = ('/favicon.ico', '/robots.txt') #: Whether the input parameters from GET and POST should be decoded #: according to the encoding specified by the request. This should only be #: changed to False if the input is supposed to be byte values. (Default: #: True) DECODE_PARAMS = True #: Object representing the current request. Set by the constructor. request = None #: Object representing the current response. Set by the constructor. response = None #: Dictionary with the path parameters. Set by the constructor. path_params = None #: String with the current path. Same as request.path except the extension #: is removed. So instead of `/movies.json` it is just `/movies`. Set by #: the constructor. request_path = None #: Reference to the application. Set by the constructor. application = None #: Charset to output in the Content-Type headers. Set to None to avoid #: sending this. charset = 'UTF-8' # Cache for the `data` property _data = None def __init__(self, request, response, path_params, application=None): """Constructor. Order of the parameters is not guaranteed, always used named parameters. :param request: Object representing the current request. :type request: :class:`webob.Request` :param response: Object representing the response to be sent. :type response: :class:`webob.Response` :param path_params: Dictionary of all parameters passed in via the path. This is the return value of :func:`Router.__call__`. :type path_params: dict :param application: Reference to the application which is calling this resource. Can be used to reference other resources or properties of the application itself. :type path_params: :class:`wsgiservice.Application` """ self.request = request self.response = response self.path_params = path_params self.application = application self.request_path = '' if request: self.request_path = request.path if path_params and path_params.get('_extension'): ext = path_params['_extension'] if self.request_path.endswith(ext): self.request_path = self.request_path[0:-len(ext)] def OPTIONS(self): """Default implementation of the OPTIONS verb. Outputs a list of allowed methods on this resource in the ``Allow`` response header. """ self.response.headers['Allow'] = self.get_allowed_methods() def __call__(self): """Main entry point for calling this resource. Handles the method dispatching, response conversion, etc. for this resource. Catches all exceptions: - :class:`webob.exceptions.ResponseException`: Replaces the instance's response attribute with the one from the exception. - For all exceptions in the :attr:`NOT_FOUND` tuple :func:`handle_exception_404` is called. - :class:`wsgiservice.exceptions.ValidationException`: :func:`handle_exception_400` is called. - For all other exceptions deriving from the :class:`Exception` base class, the :func:`handle_exception` method is called. """ self.type = self.get_content_type() try: self.method = self.get_method() self.handle_ignored_resources() self.assert_conditions() self.response.body_raw = self.call_method(self.method) except ResponseException, e: # a response was raised, catch it self.response = e.response r = e.response if r.status_int == 404 and not r.body and not hasattr(r, 'body_raw'): self.handle_exception_404(e) except self.NOT_FOUND, e: self.handle_exception_404(e) except ValidationException, e: self.handle_exception_400(e) except Exception, e: self.handle_exception(e) self.convert_response() self.set_response_headers() return self.response @property def data(self): """Returns the request data as a dictionary. Merges the path parameters, GET parameters and POST parameters (form-encoded or JSON dictionary). If a key is present in multiple of these, the first one defined is used. """ if self._data: return self._data retval = {} data = self.get_request_data() for subdata in data: for key, value in subdata.iteritems(): if not key in retval: retval[key] = value self._data = retval return retval def get_resource(self, resource, **kwargs): """Returns a new instance of the resource class passed in as resource. This is a helper to make future-compatibility easier when new arguments get added to the constructor. :param resource: Resource class to instantiate. Gets called with the named arguments as required for the constructor. :type resource: :class:`Resource` :param kwargs: Additional named arguments to pass to the constructor function. :type kwargs: dict """ return resource(request=self.request, response=self.response, path_params=self.path_params, application=self.application, **kwargs) def get_method(self, method=None): """Returns the method to call on this instance as a string. Raises a HTTP exception if no method can be found. Aborts with a 405 status code for known methods (based on the :attr:`KNOWN_METHODS` list) and a 501 status code for all other methods. :param method: Name of the method to return. Must be all-uppercase. :type method: str :raises: :class:`webob.exceptions.ResponseException` of status 405 or 501 if the method is not implemented on this resource. """ if method is None: method = self.request.method if hasattr(self, method) and callable(getattr(self, method)): return method elif method == 'HEAD': return self.get_method('GET') # Error: did not find any method, raise a 405 or 501 exception elif method in self.KNOWN_METHODS: # Known HTTP methods => 405 Method Not Allowed raise_405(self) else: # Unknown HTTP methods => 501 Not Implemented raise_501(self) def get_content_type(self): """Returns the Content Type to serve from either the extension or the Accept headers. Uses the :attr:`EXTENSION_MAP` list for all the configured MIME types. """ extension = self.path_params.get('_extension') for ext, mime in self.EXTENSION_MAP: if ext == extension: return mime # Else: use the Accept headers if self.response.vary is None: self.response.vary = ['Accept'] else: self.response.vary.append('Accept') types = [mime for ext, mime in self.EXTENSION_MAP] ct = self.request.accept.best_match(types) # No best match found. The specification allows us to either return a # 406 or just use another format in this case. # We pick the default format, though that may become a configurable # behavior in the future. if not ct: ct = types[0] return ct def handle_ignored_resources(self): """Ignore robots.txt and favicon.ico GET requests based on a list of absolute paths in :attr:`IGNORED_PATHS`. Aborts the request with a 404 status code. This is mostly a usability issue to avoid extra log entries for resources we are not interested in. :raises: :class:`webob.exceptions.ResponseException` of status 404 if the resource is ignored. """ if (self.method in ('GET', 'HEAD') and self.request.path_qs in self.IGNORED_PATHS): raise_404(self) def assert_conditions(self): """Handles various HTTP conditions and raises HTTP exceptions to abort the request. - Content-MD5 request header must match the MD5 hash of the full input (:func:`assert_condition_md5`). - If-Match and If-None-Match etags are checked against the ETag of this resource (:func:`assert_condition_etag`). - If-Modified-Since and If-Unmodified-Since are checked against the modification date of this resource (:func:`assert_condition_last_modified`). .. todo:: Return a 501 exception when any Content-* headers have been set in the request. (See :rfc:`2616`, section 9.6) """ self.assert_condition_md5() etag = self.clean_etag(self.call_method('get_etag')) self.response.last_modified = self.call_method('get_last_modified') self.assert_condition_etag() self.assert_condition_last_modified() def assert_condition_md5(self): """If the ``Content-MD5`` request header is present in the request it's verified against the MD5 hash of the request body. If they don't match, a 400 HTTP response is returned. :raises: :class:`webob.exceptions.ResponseException` of status 400 if the MD5 hash does not match the body. """ if 'Content-MD5' in self.request.headers: body_md5 = hashlib.md5(self.request.body).hexdigest() if body_md5 != self.request.headers['Content-MD5']: raise_400(self, msg='Invalid Content-MD5 request header.') def assert_condition_etag(self): """If the resource has an ETag (see :func:`get_etag`) the request headers ``If-Match`` and ``If-None-Match`` are verified. May abort the request with 304 or 412 response codes. :raises: - :class:`webob.exceptions.ResponseException` of status 304 if the ETag matches the ``If-None-Match`` request header (GET/HEAD requests only). - :class:`webob.exceptions.ResponseException` of status 412 if the ETag matches the ``If-None-Match`` request header (for requests other than GET/HEAD) or the ETag does not match the ``If-Match`` header. """ if self.response.etag: etag = self.response.etag.replace('"', '') if not etag in self.request.if_match: raise_412(self, 'If-Match request header does not the resource ETag.') if etag in self.request.if_none_match: if self.request.method in ('GET', 'HEAD'): raise_304(self) else: raise_412(self, 'If-None-Match request header matches resource ETag.') def assert_condition_last_modified(self): """If the resource has a last modified date (see :func:`get_last_modified`) the request headers ``If-Modified-Since`` and ``If-Unmodified-Since`` are verified. May abort the request with 304 or 412 response codes. :raises: - :class:`webob.exceptions.ResponseException` of status 304 if the ``If-Modified-Since`` is later than the last modified date. - :class:`webob.exceptions.ResponseException` of status 412 if the last modified date is later than the ``If-Unmodified-Since`` header. """ rq = self.request rs = self.response if rs.last_modified: rsl = rs.last_modified if rq.if_modified_since and rsl <= rq.if_modified_since: raise_304(self) if rq.if_unmodified_since and rsl > rq.if_unmodified_since: raise_412(self, 'Resource is newer than the ' 'If-Unmodified-Since request header.') def get_etag(self): """Returns a string to be used as the ETag for this resource. Used to set the ``ETag`` response headers and for conditional requests using the ``If-Match`` and ``If-None-Match`` request headers. """ return None def clean_etag(self, etag): """Cleans the ETag as returned by :func:`get_etag`. Will wrap it in quotes and append the extension for the current MIME type. """ if etag: etag = etag.replace('"', '') extension = None for ext, mime in self.EXTENSION_MAP: if mime == self.type: extension = ext[1:] break if extension: etag += '_' + extension self.response.etag = etag def get_last_modified(self): """Return a :class:`datetime.datetime` object of the when the resource was last modified. Used to set the ``Last-Modified`` response header and for conditional requests using the ``If-Modified-Since`` and ``If-Unmodified-Since`` request headers. :rtype: :class:`datetime.datetime` """ return None def get_allowed_methods(self): """Returns a coma-separated list of method names that are allowed on this instance. Useful to set the ``Allowed`` response header. """ return ", ".join([method for method in dir(self) if method.upper() == method and callable(getattr(self, method))]) def call_method(self, method_name): """Call an instance method filling in all the method parameters based on their names. The parameters are filled in from the following locations (in that order of precedence): 1. Path parameters from routing 2. GET parameters 3. POST parameters All values are validated using the method :func:`validate_param`. The return value of the method is returned unaltered. :param method_name: Name of the method on the current instance to call. :type method_name: str """ params = [] method = getattr(self, method_name) method_params, varargs, varkw, defaults = inspect.getargspec(method) if method_params and len(method_params) > 1: method_params.pop(0) # pop the self off data = self._merge_defaults(self.data, method_params, defaults) params = self._get_validated_params(method, method_params, data) return self._call_method(method, params, method_params) def _call_method(self, method, params, method_params): """ Override this method to add additional validation. :param method: Method to be called. :param params: Validated parameter values for calling the method. :param method_params: The list of parameters in the method's signature. :return: The result of calling the method. """ return method(*params) def _get_validated_params(self, method, method_params, data): """ Returns a list of params to call the method with. The parameters are returned in the order as they are expected by the method. All parameters are validated and converted to the requested datatype, based on the configuration from the `@validate` decorator. :param method_params: The introspected parameters the method needs. :type method_params: List of parameter names. :param data: All key/value pairs passed in with the request. :type data: dict """ errors = {} params = [] for param in method_params: value = data.get(param) try: self.validate_param(method, param, value) value = self.convert_param(method, param, value) except ValidationException, e: # Collect all errors, so we can raise them as one consolidated # validation exception. errors[param] = e else: params.append(value) if not errors: return params # Raise consolidated exceptions. The message of the first error is used # as message of the new consolidated exception. raise MultiValidationException(errors, unicode(errors.values()[0]).encode('utf-8')) def validate_param(self, method, param, value): """Validates the parameter according to the configurations in the _validations dictionary of either the method or the instance. This dictionaries are written by the decorator :func:`wsgiservice.decorators.validate`. .. todo:: Allow validation by type (e.g. header, post, query, etc.) :param method: A function to get the validation information from (done using :func:`_get_validation`). :type method: Python function :param param: Name of the parameter to validate the value for. :type param: str :param value: Value passed in for the given parameter. :type value: Any valid Python value :raises: :class:`wsgiservice.exceptions.ValidationException` if the value is invalid for the given method and parameter. """ rules = self._get_validation(method, param) if not rules: return if rules.get('mandatory') and ( value is None or (isinstance(value, basestring) and len(value) == 0)): raise ValidationException("Missing value for {0}.".format(param)) elif rules.get('re') and (rules.get('mandatory') or value is not None): if not re.search('^' + rules['re'] + '$', value): raise ValidationException("Invalid value for {0}.".format(param)) def convert_param(self, method, param, value): """Converts the parameter using the function 'convert' function of the validation rules. Same parameters as the `validate_param` method, so it might have just been added there. But lumping together the two functionalities would make overwriting harder. :param method: A function to get the validation information from (done using :func:`_get_validation`). :type method: Python function :param param: Name of the parameter to validate the value for. :type param: str :param value: Value passed in for the given parameter. :type value: Any valid Python value :raises: :class:`wsgiservice.exceptions.ValidationException` if the value is invalid for the given method and parameter. """ rules = self._get_validation(method, param) has_convert = rules and rules.get('convert') if not has_convert or (value is None and not rules['mandatory']): return value try: return rules['convert'](value) except ValueError: raise ValidationException("Invalid value for {0}.".format(param)) def _get_validation(self, method, param): """Return the correct validations dictionary for this parameter. First checks the method itself and then its class. If no validation is defined for this parameter, None is returned. :param method: A function to get the validation information from. :type method: Python function :param param: Name of the parameter to get validation information for. :type param: str """ if hasattr(method, '_validations') and param in method._validations: return method._validations[param] elif (hasattr(method.im_class, '_validations') and param in method.im_class._validations): return method.im_class._validations[param] else: return None def convert_response(self): """Finish filling the instance's response object so it's ready to be served to the client. This includes converting the body_raw property to the content type requested by the user if necessary. """ if hasattr(self.response, 'body_raw'): if self.response.body_raw is not None: to_type = re.sub('[^a-zA-Z_]', '_', self.type) to_type_method = 'to_' + to_type if hasattr(self, to_type_method): self.response.body = getattr(self, to_type_method)( self.response.body_raw) del self.response.body_raw def to_application_json(self, raw): """Returns the JSON version of the given raw Python object. :param raw: The return value of the resource method. :type raw: Any valid Python value :rtype: string """ return json.dumps(raw) def to_text_xml(self, raw): """Returns the XML string version of the given raw Python object. Uses :func:`_get_xml_value` which applies some heuristics for converting data to XML. The default root tag is 'response', but that can be overwritten by changing the :attr:`XML_ROOT_TAG` instance variable. Uses :func:`wsgiservice.xmlserializer.dumps()` for the actual work. :param raw: The return value of the resource method. :type raw: Any valid Python value :rtype: string """ return xmlserializer.dumps(raw, self.XML_ROOT_TAG) def handle_exception(self, e, status=500): """Handle the given exception. Log, sets the response code and output the exception message as an error message. :param e: Exception which is being handled. :type e: :class:`Exception` :param status: Status code to set. :type status: int """ logger.exception( "An exception occurred while handling the request: %s", e) self.response.body_raw = {'error': unicode(e).encode('utf-8')} self.response.status = status def handle_exception_400(self, e): """Handle the given validation exception. Log, sets the response code to 400 and output all errors as a dictionary. :param e: Exception which is being handled. :type e: :class:`ValidationException`, usually its subclass :class:`MultiValidationException`. """ if isinstance(e, MultiValidationException): errors = dict([ (param, unicode(exc).encode('utf-8')) for param, exc in e.errors.iteritems() ]) logger.info("A 400 Bad Request exception occurred while " "handling the request", exc_info=True, extra={'errors': errors, 'e': e}) first_error = unicode(e.errors.values()[0]).encode('utf-8') self.response.body_raw = {'errors': errors, 'error': first_error} self.response.status = 400 else: # Use normal `handle_exception` as fallback. This point is probably # reached by legacy code only. self.handle_exception(e, status=400) def handle_exception_404(self, e): """Handle the given exception. Log, sets the response code to 404 and output the exception message as an error message. :param e: Exception which is being handled. :type e: :class:`Exception` """ logger.debug("A 404 Not Found exception occurred while handling " "the request.") self.response.body_raw = {'error': 'Not Found'} self.response.status = 404 def set_response_headers(self): """Sets all the calculated response headers.""" self.set_response_content_type() self.set_response_content_md5() def set_response_content_type(self): """Set the Content-Type in the response. Uses the :attr:`type` instance attribute which was set by :func:`get_content_type`. Also declares a UTF-8 charset. """ if self.response.body: ct = self.type if self.charset: ct += '; charset=' + self.charset self.response.headers['Content-Type'] = ct elif 'Content-Type' in self.response.headers: del self.response.headers['Content-Type'] def set_response_content_md5(self): """Set the Content-MD5 response header. Calculated from the the response body by creating the MD5 hash from it. """ self.response.content_md5 = hashlib.md5(self.response.body).hexdigest() def get_request_data(self): """ Read the input values. Returns a list of dictionaries. These will be used to automatically pass them into the method. Additionally a combined dictionary is written to `self.data`. In the case of JSON input, that element in this list will be the parsed JSON value. That may not be a dictionary. """ request_data = [self.path_params, self.request.GET] content_type = self.request.headers.get('Content-Type', '')\ .split(';')[0].strip() if content_type == 'application/json' and self.request.body: try: post = json.loads(self.request.body) except ValueError: raise_400(self, msg='Invalid JSON content data') if isinstance(post, dict): request_data.append(post) else: request_data.append(self.request.POST) return request_data def _merge_defaults(self, data, method_params, defaults): """Helper method for adding default values to the data dictionary. The `defaults` are the default values inspected from the method that will be called. For any values that are not present in the incoming data, the default value is added. """ if defaults: optional_args = method_params[-len(defaults):] for key, value in zip(optional_args, defaults): if not key in data: data[key] = value return data @mount('/_internal/help') class Help(Resource): """Provides documentation for all resources of the current application. .. todo:: Allow documentation of output. .. todo:: Use first sentence of docstring for summary, add bigger version at the bottom. """ EXTENSION_MAP = [('.html', 'text/html')] + Resource.EXTENSION_MAP XML_ROOT_TAG = 'help' def GET(self): """Returns documentation for the application.""" retval = [] for res in self.application._resources: retval.append({ 'name': res.__name__, 'desc': self._get_doc(res), 'properties': { 'XML_ROOT_TAG': res.XML_ROOT_TAG, 'KNOWN_METHODS': res.KNOWN_METHODS, 'EXTENSION_MAP': dict((key[1:], value) for key, value in res.EXTENSION_MAP), 'NOT_FOUND': [ex.__name__ for ex in res.NOT_FOUND], }, 'methods': self._get_methods(res), 'path': self.request.script_name + res._path, }) # Sort by name retval = [(r['name'], r) for r in retval] retval.sort() retval = [r[1] for r in retval] return retval def _get_methods(self, res): """Return a dictionary of method descriptions for the given resource. :param res: Resource class to get all HTTP methods from. :type res: :class:`webob.resource.Resource` """ retval = {} inst = res(request=webob.Request.blank('/'), response=webob.Response(), path_params={}) methods = [m.strip() for m in inst.get_allowed_methods().split(',')] for method_name in methods: method = getattr(res, method_name) retval[method_name] = { 'desc': self._get_doc(method), 'parameters': self._get_parameters(res, method)} return retval def _get_doc(self, obj): """Returns a slightly modified (stripped) docstring for the given Python object. Returns an empty string if the object doesn't have any documentation. :param obj: Python object to get the docstring from. :type obj: A method or class. """ doc = obj.__doc__ if doc: return doc.strip() else: return '' def _get_parameters(self, res, method): """Return a parameters dictionary for the given resource/method. :param res: Resource class to get all HTTP methods from. :type res: :class:`webob.resource.Resource` :param method: The method to get parameters from. :type method: Python function """ method_params, varargs, varkw, defaults = inspect.getargspec(method) if method_params: method_params.pop(0) # pop the self off self._add_path_parameters(method_params, res) retval = {} for param in method_params: is_path_param = '{' + param + '}' in res._path validation = self._get_validation(method, param) retval[param] = { 'path_param': is_path_param, 'mandatory': is_path_param or validation, 'validate_re': None, 'desc': '', } if validation: retval[param]['validate_re'] = validation['re'] retval[param]['desc'] = validation['doc'] or '' return retval def _add_path_parameters(self, method_params, res): """Extract all path parameters as they are always required even though some methods may not have them in their definition. :param method_params: Current list of parameters from the method. :type method_params: Ordered list of method parameter names. :param res: Resource class to get the path from. :type res: :class:`webob.resource.Resource` """ for param in re.findall('{([^}]+)}', res._path): if param not in method_params: method_params.append(param) def _get_xml_value(self, value): """Overwritten _get_xml_value which uses the tag 'resource' for list children. Calls :func:`Resource._get_xml_value` for all non-list values. :param value: The value to convert to HTML. :type raw: Any valid Python value """ if isinstance(value, list): retval = [] for key, value in enumerate(value): retval.append('<resource>') retval.append(self._get_xml_value(value)) retval.append('</resource>') return "".join(retval) else: return Resource._get_xml_value(self, value) def to_text_html(self, raw): """Returns the HTML string version of the given raw Python object. Hard-coded to return a nicely-presented service information document. :param raw: The return value of the resource method. :type raw: Any valid Python object :rtype: string .. todo:: Treat paragraphs and/or newlines better in output. """ retval = ["""<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"> <html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en" lang="en"> <head> <meta http-equiv="Content-Type" content="text/html; charset=utf-8"/> <title>Help Example</title> <style> /* YUI reset.css */ html{color:#000;background:#FFF;}body,div,dl,dt,dd,ul,ol,li,h1,h2,h3,h4,h5,h6,pre,code,form,fieldset,legend,input,button,textarea,p,blockquote,th,td{margin:0;padding:0;}table{border-collapse:collapse;border-spacing:0;}fieldset,img{border:0;}address,caption,cite,code,dfn,em,strong,th,var,optgroup{font-style:inherit;font-weight:inherit;}del,ins{text-decoration:none;}li{list-style:none;}caption,th{text-align:left;}h1,h2,h3,h4,h5,h6{font-size:100%;font-weight:normal;}q:before,q:after{content:'';}abbr,acronym{border:0;font-variant:normal;}sup{vertical-align:baseline;}sub{vertical-align:baseline;}legend{color:#000;}input,button,textarea,select,optgroup,option{font-family:inherit;font-size:inherit;font-style:inherit;font-weight:inherit;}input,button,textarea,select{*font-size:100%;} /* YUI fonts.css */ body{font:13px/1.231 arial,helvetica,clean,sans-serif;*font-size:small;*font:x-small;}select,input,button,textarea,button{font:99% arial,helvetica,clean,sans-serif;}table{font-size:inherit;font:100%;}pre,code,kbd,samp,tt{font-family:monospace;*font-size:108%;line-height:100%;} /* YUI base.css */ body{margin:10px;}h1{font-size:138.5%;}h2{font-size:123.1%;}h3{font-size:108%;}h1,h2,h3{margin:1em 0;}h1,h2,h3,h4,h5,h6,strong,dt{font-weight:bold;}optgroup{font-weight:normal;}abbr,acronym{border-bottom:1px dotted #000;cursor:help;}em{font-style:italic;}del{text-decoration:line-through;}blockquote,ul,ol,dl{margin:1em;}ol,ul,dl{margin-left:2em;}ol li{list-style:decimal outside;}ul li{list-style:disc outside;}dl dd{margin-left:1em;}th,td{border:1px solid #000;padding:.5em;}th{font-weight:bold;text-align:center;}caption{margin-bottom:.5em;text-align:center;}sup{vertical-align:super;}sub{vertical-align:sub;}p,fieldset,table,pre{margin-bottom:1em;}button,input[type="checkbox"],input[type="radio"],input[type="reset"],input[type="submit"]{padding:1px;} h2 {margin-top: 0;} .resource_details {padding-top: 2em;border-top: 1px dotted #ccc;margin-top: 2em;} .method_details {margin-left: 2em;} /* JS form */ form { padding: 1em; border: 1px solid #ccc; } input.error { background: #FCECEC; color: red; border: 1px solid red; } label { font-weight: bold; float: left; width: 10em; } p.form_element, input.submit { clear: left; } div.result { margin-top: 1em; } h4 { margin-bottom: 0.5em; } a.add_input { margin: 0.5em; } .hidden { display: none !important; } a.toggle_details { margin-bottom: 1em; display: block; } </style> <script> /** * Adds a resource's method - a form to the current location with the ability * to submit a request to the service filling in all the parameters. */ function add_resource_method(target, resource, method_name, method) { new ResourceMethodForm(target, resource, method_name, method); } function ResourceMethodForm(target, resource, method_name, method) { this.targetName = target; this.target = document.getElementById(target); this.resource = resource; this.method_name = method_name; this.method = method; this.init(); } var pr = ResourceMethodForm.prototype; pr.init = function() { var fragment = document.createDocumentFragment(); var form = this.create_form(fragment); var input_container = document.createElement('div'); this.input_container = input_container; form.appendChild(input_container); this.create_form_params(input_container); this.create_form_buttons(form); this.create_result_field(form); this.target.appendChild(fragment); }; pr.create_form = function(parent) { var form = document.createElement('form'); form.action = ''; form.target = '_blank'; var that = this; form.onsubmit = function() { return that.on_submit(); }; var h4 = document.createElement('h4'); h4.innerHTML = 'Debug form'; form.appendChild(h4); parent.appendChild(form); return form; }; pr.create_form_params = function(parent) { for (param in this.method.parameters) { this.create_form_field(parent, 'param', 'text', param); } // Accept header var mimes = [];<|fim▁hole|> } this.create_form_field(parent, 'header', 'select', 'Accept', mimes); }; pr.create_form_field = function(parent, type, field_type, name, options) { var id = type + '_' + this.resource['name'] + '_' + this.method_name + '_' + name; var d = document.createElement('div'); d.className = type; d.id = id; var input_id = id + '_input'; var lbl = document.createElement('label'); lbl.innerHTML = name; if (type == 'header') { lbl.innerHTML += ' (Header)'; } lbl.setAttribute('for', input_id); d.appendChild(lbl); var field = null; if (field_type == 'select') { field = document.createElement('select'); for (var i = 0; i < options.length; i++) { var option = document.createElement('option'); option.value = options[i]; option.innerHTML = this.format(options[i]); field.appendChild(option); } } else { field = document.createElement('input'); field.type = 'text'; } field.id = input_id; field.name = name; d.appendChild(field); parent.appendChild(d); }; pr.create_form_buttons = function(parent) { var subm = document.createElement('input'); subm.type = 'submit'; subm.value = 'Execute request (' + this.method_name + ')'; subm.className = 'submit'; parent.appendChild(subm); var that = this; var create_field = document.createElement('a'); create_field.href = '#'; create_field.className = 'add_input'; create_field.innerHTML = 'Add parameter'; create_field.onclick = function() { var name = prompt("Enter a field name:"); if (name !== null) { that.create_form_field(that.input_container, 'param', 'text', name); } return false; }; parent.appendChild(create_field); var create_header = document.createElement('a'); create_header.href = '#'; create_header.className = 'add_input'; create_header.innerHTML = 'Add header'; create_header.onclick = function() { var name = prompt("Enter a header name:"); if (name !== null) { that.create_form_field(that.input_container, 'header', 'text', name); } return false; }; parent.appendChild(create_header); }; pr.create_result_field = function(parent) { this.result_node = document.createElement('div'); this.result_node.className = 'result'; parent.appendChild(this.result_node); }; pr.on_submit = function() { var xhr = null; var that = this; this.result_node.innerHTML = 'Executing...'; if (window.XMLHttpRequest) { xhr = new XMLHttpRequest(); } else if (window.ActiveXObject) { xhr = new ActiveXObject("Microsoft.XMLHTTP"); } xhr.onreadystatechange = function() { if (xhr.readyState == 4) { // Received var data = [ '<h5>Status</h5>', xhr.status, ' ', that.get_status(xhr.status, xhr.statusText), '<h5>Response Headers</h5>', that.format(xhr.getAllResponseHeaders()), '<h5>Response Body</h5>', that.format(xhr.responseText), '(' + xhr.responseText.length + ' bytes)' ]; that.result_node.innerHTML = ''; for (var i = 0; i < data.length; i++) { that.result_node.innerHTML += data[i]; } } }; // Get parameters and fill them into path, query string and POST data var input = this.get_parameters(); if (input['__error__']) { this.result_node.innerHTML = 'ERROR: Missing data.'; return false; } var path = this.resource.path; var data = ''; // Request parameters for (param_name in this.method.parameters) { var param = this.method.parameters[param_name]; if (param['path_param']) { path = path.replace('{' + param_name + '}', input['params'][param_name]); } else { data += escape(param_name) + '=' + escape(input['params'][param_name]) + '&'; } } if (data === '') { data = null; } else if (this.method_name == 'GET' || this.method_name == 'HEAD') { // Convert data to query string path += '?' + data; data = null; } xhr.open(this.method_name, path, true); if (data !== null) { xhr.setRequestHeader('Content-Type', 'application/x-www-form-urlencoded'); } // Request headers for (header in input['headers']) { xhr.setRequestHeader(header, input['headers'][header]); } xhr.send(data); return false; }; pr.format = function(str) { var str = str.replace(/&/g, "&amp;").replace(/</g, "&lt;"). replace(/>/g, "&gt;"); // Linkify HTTP URLs str = str.replace(/(http:\/\/[^ ]+)/g, '<a href="$1" target="_blank">$1</a>'); str = '<pre>' + str + '</pre>'; return str; }; pr.get_status = function(status, statusText) { /* Need to get the status text manually as statusText is broken on Safari. */ if (statusText == 'OK') { // Safari always uses OK, replace it manually var STATI = { 200: 'OK', 201: 'Created', 202: 'Accepted', 203: 'Non-Authoritative Information', 204: 'No Content', 205: 'Reset Content', 206: 'Partial Content', 300: 'Multiple Choices', 301: 'Moved Permanently', 302: 'Found', 303: 'See Other', 304: 'Not Modified', 305: 'Use Proxy', 307: 'Temporary Redirect', 400: 'Bad Request', 401: 'Unauthorized', 402: 'Payment Required', 403: 'Forbidden', 404: 'Not Found', 405: 'Method Not Allowed', 406: 'Not Acceptable', 407: 'Proxy Authentication Required', 408: 'Request Timeout', 409: 'Conflict', 410: 'Gone', 411: 'Length Required', 412: 'Precondition Failed', 413: 'Request Entity Too Large', 414: 'Request-URI Too Long', 415: 'Unsupported Media Type', 416: 'Requested Range Not Satisfiable', 417: 'Expectation Failed', 500: 'Internal Server Error', 501: 'Not Implemented', 502: 'Bad Gateway', 503: 'Service Unavailable', 504: 'Gateway Timeout', 505: 'HTTP Version Not Supported' }; if (typeof(STATI[status]) !== 'undefined') { return STATI[status]; } } return statusText; }; pr.get_parameters = function() { var params = {'__error__': false, 'headers': {}, 'params': {}}; var fields = []; // Get all fields var inputs = this.target.getElementsByTagName('input'); for (var i = 0; i < inputs.length; i++) { fields.push(inputs[i]); } var selects = this.target.getElementsByTagName('select'); for (var i = 0; i < selects.length; i++) { fields.push(selects[i]); } var inp = null; for (var i = 0; i < fields.length; i++) { inp = fields[i]; var type = inp.parentNode.className; if (type == 'header') { params['headers'][inp.name] = inp.value; } else if (type == 'param') { inp.className = ''; params['params'][inp.name] = inp.value; // Validate input if (this.method.parameters[inp.name]['mandatory'] && inp.value === '') { inp.className = 'error'; params['__error__'] = true; } } } return params; }; /* Hides all .resource_details elements and inserts a toggle link at their place. */ function toggle_visibility() { var divs = document.getElementsByTagName('div'); var len = divs.length; for (var i = 0; i < len; i++) { var div = divs[i]; if (div.className == 'method_details') { toggle_visibility_div(div); } } } function toggle_visibility_div(div) { div.className += ' hidden'; var link = document.createElement('a'); link.innerHTML = 'Show details'; link.href = '#'; link.className = 'toggle_details'; link.onclick = function() { console.debug(div); if (link.innerHTML == 'Show details') { div.className = div.className.replace(' hidden', ''); link.innerHTML = 'Hide details'; } else { div.className += ' hidden'; link.innerHTML = 'Show details'; } return false; }; div.parentNode.insertBefore(link, div); } </script> </head> <body> <h1>WsgiService help</h1> """] self.to_text_html_overview(retval, raw) self.to_text_html_resources(retval, raw) retval.append('<script>toggle_visibility();</script>') retval.append('</body></html>') return re.compile('^ *', re.MULTILINE).sub('', "".join(retval)) def to_text_html_overview(self, retval, raw): """Add the overview table to the HTML output. :param retval: The list of strings which is used to collect the HTML response. :type retval: list :param raw: The original return value of this resources :func:`GET` method. :type raw: Dictionary """ retval.append('<table id="overview">') retval.append('<tr><th>Resource</th><th>Path</th><th>Description</th></tr>') for resource in raw: retval.append('<tr><td><a href="#{0}">{0}</a></td><td>{1}</td><td>{2}</td></tr>'.format( xml_escape(resource['name']), xml_escape(resource['path']), xml_escape(resource['desc']))) retval.append('</table>') def to_text_html_resources(self, retval, raw): """Add the resources details to the HTML output. :param retval: The list of strings which is used to collect the HTML response. :type retval: list :param raw: The original return value of this resources :func:`GET` method. :type raw: Dictionary """ for resource in raw: retval.append('<div class="resource_details">') retval.append('<h2 id="{0}">{0}</h2>'.format( xml_escape(resource['name']))) if resource['desc']: retval.append('<p class="desc">{0}</p>'.format(xml_escape(resource['desc']))) retval.append('<table class="config">') retval.append('<tr><th>Path</th><td>{0}</td>'.format(xml_escape( resource['path']))) representations = [value + ' (.' + key + ')' for key, value in resource['properties']['EXTENSION_MAP'].iteritems()] retval.append('<tr><th>Representations</th><td>{0}</td>'.format( xml_escape(', '.join(representations)))) retval.append('</table>') self.to_text_html_methods(retval, resource) retval.append('</div>') def to_text_html_methods(self, retval, resource): """Add the methods of this resource to the HTML output. :param retval: The list of strings which is used to collect the HTML response. :type retval: list :param resource: The documentation of one resource. :type resource: Dictionary """ for method_name, method in resource['methods'].iteritems(): retval.append('<h3 id="{0}_{1}">{1}</h3>'.format( xml_escape(resource['name']), xml_escape(method_name))) retval.append('<div class="method_details" id="{0}_{1}_container">'.format( xml_escape(resource['name']), xml_escape(method_name))) if method['desc']: retval.append('<p class="desc">{0}</p>'.format(xml_escape(method['desc']))) if method['parameters']: retval.append('<table class="parameters">') retval.append('<tr><th>Name</th><th>Mandatory</th><th>Description</th><th>Validation</th>') for param_name, param in method['parameters'].iteritems(): # filter out any parameters that can't be written as html. # can contain stuff in other encodings than ascii, # so convert it to ascii mandatory = '-' description = param['desc'].encode('ascii', 'replace') validation = '' if param['mandatory']: mandatory = 'Yes' # the 'convert' key is a <type 'function'> that can't # be written as a string, so convert it to one. if type(param['mandatory']) is dict and \ 'convert' in param['mandatory']: param['mandatory']['convert'] = \ str(param['mandatory']['convert']) if param['path_param']: mandatory += ' (Path parameter)' if param['validate_re']: validation = 'Regular expression: <tt>' + \ xml_escape(param['validate_re'].encode('ascii', 'replace')) + \ '</tt>' retval.append('<tr><td>{0}</td><td>{1}</td><td>{2}</td>' '<td>{3}</td>'.format(xml_escape(param_name), xml_escape(mandatory), xml_escape(description), validation)) retval.append('</table>') retval.append('</div>') retval.append('<script>add_resource_method({0},{1},{2},{3});</script>'.format( xml_escape(json.dumps(resource['name']+'_'+method_name+'_container')), xml_escape(json.dumps(resource)), xml_escape(json.dumps(method_name)), xml_escape(json.dumps(method)))) class NotFoundResource(Resource): EXTENSION_MAP = [('.html', 'text/html')] + Resource.EXTENSION_MAP def GET(self): self.response.status = 404 return {'error': 'The requested resource does not exist.'} def get_method(self, method=None): return 'GET' def handle_ignored_resources(self): return def to_text_html(self, raw): return "".join([ '<html>', '<head><title>404 Not Found</title></head>', '<body>', '<center><h1>404 Not Found</h1></center>', '<center>The requested resource does not exist.</center>', '</body></html>' ])<|fim▁end|>
var emap = this.resource.properties.EXTENSION_MAP; for (extension in emap) { mimes.push(emap[extension]);
<|file_name|>classifier.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python # -*- coding: utf-8 -*- # # Author: Enrique Henestroza Anguiano # """ Classifier for liblinear, svmlin. Agnostic about features or labels. Uses 'ranking', or dynamic classes. """ import sys import os import codecs import svmlight from libsvm import svmutil, svm # Optional LIBLinear try: from liblinear import liblinearutil, liblinear liblinear_found = True except ImportError: liblinear_found = False from ctypes import * from dtbutils import * from perceptron import KernelLBRankPerceptron, polynomial_kernel import numpy as np import cPickle class Classifier(object): def __init__(self, param={'model':'', 'pref':'def', 'verbose':False,\ 'classtype':'classifier'}): self._param = param # Model self._classtype = param['classtype'] self._modelname = param['model'] self._pref = param['pref'] self._model = None self._numex_train = 0 self._numex_dev = 0 self._max_feat = 0 self._labs = {} self._svmclassp = "-s 0 -t 1 -d 3 -g 0.1 -r 0.0 -e 1.0 -c 1.0 -q" self._svmstrucp = "-z p -t 1 -d 3 -s 0.1 -r 0.0 -e 1.0 -c 0.05 -b 0" # # Make a decision using a model on an example, A feature vector is: # ((feat1,val1), (feat3,val3), ...) # Where each index corresponds to a feature in the model alphabet. Output # a list of tuples (class/idx, score) sorted by decending score. # def score(self, feats): m = self._model if self._classtype == "classifier": x,_ = svm.gen_svm_nodearray(dict(feats)) return int(svm.libsvm.svm_predict(m, x)) elif self._classtype == "structured": maxscore = -sys.maxint maxidx = None for idx in range(len(feats)): dec_val = svmlight.classify(m, [(0, feats[idx])]) if dec_val > maxscore: maxscore = dec_val maxidx = idx return maxidx elif self._classtype == "percrank": X = [None]*len(feats) Xisd = [0]*len(feats) Xisd[0] = 1 for idx in range(len(feats)): X[idx] = set([f for f,v in feats[idx]]) dec_vals = m.project(X, Xisd) return dec_vals.index(max(dec_vals)) # # Reads a ranking problem. # def read_rank_problem(self, ef): efile = codecs.open(ef, 'r', 'ascii') qid = None allex = [] rex = [] print >> sys.stderr, "Reading ranking problem..." for line in efile: fields = line.rstrip().split(' ') glab = int(fields.pop(0)) cqid = int(fields.pop(0).split(":")[1]) feats = [] for field in fields: f,v = field.split(":") #feats.append((int(f),float(v))) feats.append(int(f)) feats = set(feats) if qid == None: qid = cqid rex = [(glab, feats)] elif qid == cqid: rex.append((glab, feats)) else: allex.append(rex) qid = cqid rex = [(glab, feats)] allex.append(rex) efile.close() # Only supports a one-vs-all ranking (highest glab over rest) print >> sys.stderr, "Generating ranking constraints...", X1 = [] X2 = [] X2cnt = 0 Xidx = [] X1isdef = [] X2isdef = [] bline = 0 for rex in allex: glabs = [glab for glab,_ in rex] gidx = glabs.index(max(glabs)) cidx = [] for i in range(len(rex)): glab,feats = rex[i] if i == 0 and glab == 1: bline += 1 if i == gidx: X1.append(feats) if i == 0: X1isdef.append(1) else: X1isdef.append(0) else: cidx.append(X2cnt) X2.append(feats) if i == 0: X2isdef.append(1) else: X2isdef.append(0) X2cnt += 1 Xidx.append(tuple(cidx)) print >> sys.stderr, X2cnt return X1, X1isdef, X2, X2isdef, Xidx, bline # # Append stream of examples to file. Feature vectors are as follows: # [(feat1, val1), (feat3, val3), ..., (featn, valn)] # def write_examples(self, examples, mode="train"): exstream = codecs.open(self._modelname+"/"+self._pref+"."+mode,\ 'a', 'ascii') # Classification examples over a single line. Label and feature vector: # 2 0:1 2:1 5:1 # 5 1:1 2:1 4:1 if self._classtype == "classifier": for glab,feats in examples: if mode == 'train': self._numex_train += 1 self._max_feat = max(self._max_feat, feats[-1][0]) self._labs[glab] = True else: self._numex_dev += 1 print >> exstream, glab, \ " ".join([str(f)+":"+str(v) for f,v in feats]) # Structured binary examples.<|fim▁hole|> if mode == 'train': self._numex_train += 1 qid = self._numex_train else: self._numex_dev += 1 qid = self._numex_dev for idx in range(len(ex)): feats = ex[idx] if mode == 'train': self._max_feat = max(self._max_feat, feats[-1][0]) if idxg == idx: glab = 1 else: glab = 0 print >> exstream, glab, 'qid:'+str(qid),\ " ".join([str(f)+":"+str(v) \ for f,v in feats]) exstream.close() # # Train model. # def train_model(self): if self._classtype in ["structured", "percrank"]: self._labs = {1:True} print >> sys.stderr, "Training model with",\ self._numex_train,"examples,", self._max_feat+1, "features and",\ len(self._labs), "labels." if self._numex_dev: print >> sys.stderr, "Also with", self._numex_dev,"dev examples." ef = self._modelname+"/"+self._pref+".train" df = self._modelname+"/"+self._pref+".dev" mf = self._modelname+"/"+self._pref+".model" if self._classtype == "classifier": os.system("$LIBSVM/svm-train "+self._svmclassp+" "+ef+" "+mf) elif self._classtype == "structured": os.system("$SVMLIGHT/svm_learn "+self._svmstrucp+" "+ef+" "+mf) elif self._classtype == "percrank": X1,X1isdef,X2,X2isdef,Xidx,bline = self.read_rank_problem(ef) X1dev,X1devisdef,X2dev,X2devisdef,Xdevidx,devbline = \ self.read_rank_problem(df) m = KernelLBRankPerceptron(kernel=polynomial_kernel, T=10, B=0) m.fit(X1, X1isdef, X2, X2isdef, Xidx, X1dev, X1devisdef, X2dev,\ X2devisdef, Xdevidx, gm=False, bl=devbline) mfile = open(mf, 'wb') cPickle.dump([m.sv_a,m.sv_1,m.sv_2,m.bias], mfile, -1) mfile.close() # # Load model. # def load_model(self): if not os.path.isfile(self._modelname+"/"+self._pref+".model"): return False if self._classtype == "classifier": self._model = svmutil.svm_load_model(self._modelname+\ "/"+self._pref+".model") elif self._classtype == "structured": self._model = svmlight.read_model(self._modelname+\ "/"+self._pref+".model") elif self._classtype == "percrank": m = KernelLBRankPerceptron(kernel=polynomial_kernel) mfile = open(self._modelname+"/"+self._pref+".model", 'rb') m.sv_a,m.sv_1,m.sv_2,m.bias = cPickle.load(mfile) mfile.close() self._model = m return True<|fim▁end|>
# 1 qid:1 1:1 2:-1 5:-1 # 0 qid:1 1:-1 2:1 4:-1 elif self._classtype in ["structured", "percrank"]: for idxg,ex in examples:
<|file_name|>omni.py<|end_file_name|><|fim▁begin|>""" Methods for importing data from the OMNI. """ from heliopy.data import cdasrest def _docstring(identifier, description): return cdasrest._docstring(identifier, 'O', description) def _omni(starttime, endtime, identifier, intervals='monthly', warn_missing_units=True): """ Generic method for downloading OMNI data. """ dl = cdasrest.CDASDwonloader('omni', identifier, 'omni', warn_missing_units=warn_missing_units) # Override intervals if intervals == 'daily': dl.intervals = dl.intervals_daily elif intervals == 'monthly': dl.intervals = dl.intervals_monthly elif intervals == 'yearly': dl.intervals = dl.intervals_yearly return dl.load(starttime, endtime) # Actual download functions start here def h0_mrg1hr(starttime, endtime): identifier = 'OMNI2_H0_MRG1HR' return _omni(starttime, endtime, identifier, warn_missing_units=False, intervals='yearly') h0_mrg1hr.__doc__ = _docstring( 'OMNI2_H0_MRG1HR', 'Hourly averaged definitive multi-spacecraft ' 'interplanetary parameters.') def hro2_1min(starttime, endtime): identifier = 'OMNI_HRO2_1MIN' return _omni(starttime, endtime, identifier, warn_missing_units=False,<|fim▁hole|>hro2_1min.__doc__ = _docstring( 'OMNI_HRO2_1MIN', '1 minute averaged definitive multi-spacecraft ' 'interplanetary parameters.') def hro2_5min(starttime, endtime): identifier = 'OMNI_HRO2_5MIN' return _omni(starttime, endtime, identifier, warn_missing_units=False, intervals='monthly') hro2_1min.__doc__ = _docstring( 'OMNI_HRO2_5MIN', '5 minute averaged definitive multi-spacecraft ' 'interplanetary parameters.')<|fim▁end|>
intervals='monthly')
<|file_name|>hls.js<|end_file_name|><|fim▁begin|>typeof window !== "undefined" && (function webpackUniversalModuleDefinition(root, factory) { if(typeof exports === 'object' && typeof module === 'object') module.exports = factory(); else if(typeof define === 'function' && define.amd) define([], factory); else if(typeof exports === 'object') exports["Hls"] = factory(); else root["Hls"] = factory(); })(this, function() { return /******/ (function(modules) { // webpackBootstrap /******/ // The module cache /******/ var installedModules = {}; /******/ /******/ // The require function /******/ function __webpack_require__(moduleId) { /******/ /******/ // Check if module is in cache /******/ if(installedModules[moduleId]) { /******/ return installedModules[moduleId].exports; /******/ } /******/ // Create a new module (and put it into the cache) /******/ var module = installedModules[moduleId] = { /******/ i: moduleId, /******/ l: false, /******/ exports: {} /******/ }; /******/ /******/ // Execute the module function /******/ modules[moduleId].call(module.exports, module, module.exports, __webpack_require__); /******/ /******/ // Flag the module as loaded /******/ module.l = true; /******/ /******/ // Return the exports of the module /******/ return module.exports; /******/ } /******/ /******/ /******/ // expose the modules object (__webpack_modules__) /******/ __webpack_require__.m = modules; /******/ /******/ // expose the module cache /******/ __webpack_require__.c = installedModules; /******/ /******/ // define getter function for harmony exports /******/ __webpack_require__.d = function(exports, name, getter) { /******/ if(!__webpack_require__.o(exports, name)) { /******/ Object.defineProperty(exports, name, { enumerable: true, get: getter }); /******/ } /******/ }; /******/ /******/ // define __esModule on exports /******/ __webpack_require__.r = function(exports) { /******/ if(typeof Symbol !== 'undefined' && Symbol.toStringTag) { /******/ Object.defineProperty(exports, Symbol.toStringTag, { value: 'Module' }); /******/ } /******/ Object.defineProperty(exports, '__esModule', { value: true }); /******/ }; /******/ /******/ // create a fake namespace object /******/ // mode & 1: value is a module id, require it /******/ // mode & 2: merge all properties of value into the ns /******/ // mode & 4: return value when already ns object /******/ // mode & 8|1: behave like require /******/ __webpack_require__.t = function(value, mode) { /******/ if(mode & 1) value = __webpack_require__(value); /******/ if(mode & 8) return value; /******/ if((mode & 4) && typeof value === 'object' && value && value.__esModule) return value; /******/ var ns = Object.create(null); /******/ __webpack_require__.r(ns); /******/ Object.defineProperty(ns, 'default', { enumerable: true, value: value }); /******/ if(mode & 2 && typeof value != 'string') for(var key in value) __webpack_require__.d(ns, key, function(key) { return value[key]; }.bind(null, key)); /******/ return ns; /******/ }; /******/ /******/ // getDefaultExport function for compatibility with non-harmony modules /******/ __webpack_require__.n = function(module) { /******/ var getter = module && module.__esModule ? /******/ function getDefault() { return module['default']; } : /******/ function getModuleExports() { return module; }; /******/ __webpack_require__.d(getter, 'a', getter); /******/ return getter; /******/ }; /******/ /******/ // Object.prototype.hasOwnProperty.call /******/ __webpack_require__.o = function(object, property) { return Object.prototype.hasOwnProperty.call(object, property); }; /******/ /******/ // __webpack_public_path__ /******/ __webpack_require__.p = "/dist/"; /******/ /******/ /******/ // Load entry module and return exports /******/ return __webpack_require__(__webpack_require__.s = "./src/hls.ts"); /******/ }) /************************************************************************/ /******/ ({ /***/ "./node_modules/eventemitter3/index.js": /*!*********************************************!*\ !*** ./node_modules/eventemitter3/index.js ***! \*********************************************/ /*! no static exports found */ /*! ModuleConcatenation bailout: Module is not an ECMAScript module */ /***/ (function(module, exports, __webpack_require__) { "use strict"; var has = Object.prototype.hasOwnProperty , prefix = '~'; /** * Constructor to create a storage for our `EE` objects. * An `Events` instance is a plain object whose properties are event names. * * @constructor * @private */ function Events() {} // // We try to not inherit from `Object.prototype`. In some engines creating an // instance in this way is faster than calling `Object.create(null)` directly. // If `Object.create(null)` is not supported we prefix the event names with a // character to make sure that the built-in object properties are not // overridden or used as an attack vector. // if (Object.create) { Events.prototype = Object.create(null); // // This hack is needed because the `__proto__` property is still inherited in // some old browsers like Android 4, iPhone 5.1, Opera 11 and Safari 5. // if (!new Events().__proto__) prefix = false; } /** * Representation of a single event listener. * * @param {Function} fn The listener function. * @param {*} context The context to invoke the listener with. * @param {Boolean} [once=false] Specify if the listener is a one-time listener. * @constructor * @private */ function EE(fn, context, once) { this.fn = fn; this.context = context; this.once = once || false; } /** * Add a listener for a given event. * * @param {EventEmitter} emitter Reference to the `EventEmitter` instance. * @param {(String|Symbol)} event The event name. * @param {Function} fn The listener function. * @param {*} context The context to invoke the listener with. * @param {Boolean} once Specify if the listener is a one-time listener. * @returns {EventEmitter} * @private */ function addListener(emitter, event, fn, context, once) { if (typeof fn !== 'function') { throw new TypeError('The listener must be a function'); } var listener = new EE(fn, context || emitter, once) , evt = prefix ? prefix + event : event; if (!emitter._events[evt]) emitter._events[evt] = listener, emitter._eventsCount++; else if (!emitter._events[evt].fn) emitter._events[evt].push(listener); else emitter._events[evt] = [emitter._events[evt], listener]; return emitter; } /** * Clear event by name. * * @param {EventEmitter} emitter Reference to the `EventEmitter` instance. * @param {(String|Symbol)} evt The Event name. * @private */ function clearEvent(emitter, evt) { if (--emitter._eventsCount === 0) emitter._events = new Events(); else delete emitter._events[evt]; } /** * Minimal `EventEmitter` interface that is molded against the Node.js * `EventEmitter` interface. * * @constructor * @public */ function EventEmitter() { this._events = new Events(); this._eventsCount = 0; } /** * Return an array listing the events for which the emitter has registered * listeners. * * @returns {Array} * @public */ EventEmitter.prototype.eventNames = function eventNames() { var names = [] , events , name; if (this._eventsCount === 0) return names; for (name in (events = this._events)) { if (has.call(events, name)) names.push(prefix ? name.slice(1) : name); } if (Object.getOwnPropertySymbols) { return names.concat(Object.getOwnPropertySymbols(events)); } return names; }; /** * Return the listeners registered for a given event. * * @param {(String|Symbol)} event The event name. * @returns {Array} The registered listeners. * @public */ EventEmitter.prototype.listeners = function listeners(event) { var evt = prefix ? prefix + event : event , handlers = this._events[evt]; if (!handlers) return []; if (handlers.fn) return [handlers.fn]; for (var i = 0, l = handlers.length, ee = new Array(l); i < l; i++) { ee[i] = handlers[i].fn; } return ee; }; /** * Return the number of listeners listening to a given event. * * @param {(String|Symbol)} event The event name. * @returns {Number} The number of listeners. * @public */ EventEmitter.prototype.listenerCount = function listenerCount(event) { var evt = prefix ? prefix + event : event , listeners = this._events[evt]; if (!listeners) return 0; if (listeners.fn) return 1; return listeners.length; }; /** * Calls each of the listeners registered for a given event. * * @param {(String|Symbol)} event The event name. * @returns {Boolean} `true` if the event had listeners, else `false`. * @public */ EventEmitter.prototype.emit = function emit(event, a1, a2, a3, a4, a5) { var evt = prefix ? prefix + event : event; if (!this._events[evt]) return false; var listeners = this._events[evt] , len = arguments.length , args , i; if (listeners.fn) { if (listeners.once) this.removeListener(event, listeners.fn, undefined, true); switch (len) { case 1: return listeners.fn.call(listeners.context), true; case 2: return listeners.fn.call(listeners.context, a1), true; case 3: return listeners.fn.call(listeners.context, a1, a2), true; case 4: return listeners.fn.call(listeners.context, a1, a2, a3), true; case 5: return listeners.fn.call(listeners.context, a1, a2, a3, a4), true; case 6: return listeners.fn.call(listeners.context, a1, a2, a3, a4, a5), true; } for (i = 1, args = new Array(len -1); i < len; i++) { args[i - 1] = arguments[i]; } listeners.fn.apply(listeners.context, args); } else { var length = listeners.length , j; for (i = 0; i < length; i++) { if (listeners[i].once) this.removeListener(event, listeners[i].fn, undefined, true); switch (len) { case 1: listeners[i].fn.call(listeners[i].context); break; case 2: listeners[i].fn.call(listeners[i].context, a1); break; case 3: listeners[i].fn.call(listeners[i].context, a1, a2); break; case 4: listeners[i].fn.call(listeners[i].context, a1, a2, a3); break; default: if (!args) for (j = 1, args = new Array(len -1); j < len; j++) { args[j - 1] = arguments[j]; } listeners[i].fn.apply(listeners[i].context, args); } } } return true; }; /** * Add a listener for a given event. * * @param {(String|Symbol)} event The event name. * @param {Function} fn The listener function. * @param {*} [context=this] The context to invoke the listener with. * @returns {EventEmitter} `this`. * @public */ EventEmitter.prototype.on = function on(event, fn, context) { return addListener(this, event, fn, context, false); }; /** * Add a one-time listener for a given event. * * @param {(String|Symbol)} event The event name. * @param {Function} fn The listener function. * @param {*} [context=this] The context to invoke the listener with. * @returns {EventEmitter} `this`. * @public */ EventEmitter.prototype.once = function once(event, fn, context) { return addListener(this, event, fn, context, true); }; /** * Remove the listeners of a given event. * * @param {(String|Symbol)} event The event name. * @param {Function} fn Only remove the listeners that match this function. * @param {*} context Only remove the listeners that have this context. * @param {Boolean} once Only remove one-time listeners. * @returns {EventEmitter} `this`. * @public */ EventEmitter.prototype.removeListener = function removeListener(event, fn, context, once) { var evt = prefix ? prefix + event : event; if (!this._events[evt]) return this; if (!fn) { clearEvent(this, evt); return this; } var listeners = this._events[evt]; if (listeners.fn) { if ( listeners.fn === fn && (!once || listeners.once) && (!context || listeners.context === context) ) { clearEvent(this, evt); } } else { for (var i = 0, events = [], length = listeners.length; i < length; i++) { if ( listeners[i].fn !== fn || (once && !listeners[i].once) || (context && listeners[i].context !== context) ) { events.push(listeners[i]); } } // // Reset the array, or remove it completely if we have no more listeners. // if (events.length) this._events[evt] = events.length === 1 ? events[0] : events; else clearEvent(this, evt); } return this; }; /** * Remove all listeners, or those of the specified event. * * @param {(String|Symbol)} [event] The event name. * @returns {EventEmitter} `this`. * @public */ EventEmitter.prototype.removeAllListeners = function removeAllListeners(event) { var evt; if (event) { evt = prefix ? prefix + event : event; if (this._events[evt]) clearEvent(this, evt); } else { this._events = new Events(); this._eventsCount = 0; } return this; }; // // Alias methods names because people roll like that. // EventEmitter.prototype.off = EventEmitter.prototype.removeListener; EventEmitter.prototype.addListener = EventEmitter.prototype.on; // // Expose the prefix. // EventEmitter.prefixed = prefix; // // Allow `EventEmitter` to be imported as module namespace. // EventEmitter.EventEmitter = EventEmitter; // // Expose the module. // if (true) { module.exports = EventEmitter; } /***/ }), /***/ "./node_modules/url-toolkit/src/url-toolkit.js": /*!*****************************************************!*\ !*** ./node_modules/url-toolkit/src/url-toolkit.js ***! \*****************************************************/ /*! no static exports found */ /*! ModuleConcatenation bailout: Module is not an ECMAScript module */ /***/ (function(module, exports, __webpack_require__) { // see https://tools.ietf.org/html/rfc1808 (function (root) { var URL_REGEX = /^((?:[a-zA-Z0-9+\-.]+:)?)(\/\/[^\/?#]*)?((?:[^\/?#]*\/)*[^;?#]*)?(;[^?#]*)?(\?[^#]*)?(#.*)?$/; var FIRST_SEGMENT_REGEX = /^([^\/?#]*)(.*)$/; var SLASH_DOT_REGEX = /(?:\/|^)\.(?=\/)/g; var SLASH_DOT_DOT_REGEX = /(?:\/|^)\.\.\/(?!\.\.\/)[^\/]*(?=\/)/g; var URLToolkit = { // If opts.alwaysNormalize is true then the path will always be normalized even when it starts with / or // // E.g // With opts.alwaysNormalize = false (default, spec compliant) // http://a.com/b/cd + /e/f/../g => http://a.com/e/f/../g // With opts.alwaysNormalize = true (not spec compliant) // http://a.com/b/cd + /e/f/../g => http://a.com/e/g buildAbsoluteURL: function (baseURL, relativeURL, opts) { opts = opts || {}; // remove any remaining space and CRLF baseURL = baseURL.trim(); relativeURL = relativeURL.trim(); if (!relativeURL) { // 2a) If the embedded URL is entirely empty, it inherits the // entire base URL (i.e., is set equal to the base URL) // and we are done. if (!opts.alwaysNormalize) { return baseURL; } var basePartsForNormalise = URLToolkit.parseURL(baseURL); if (!basePartsForNormalise) { throw new Error('Error trying to parse base URL.'); } basePartsForNormalise.path = URLToolkit.normalizePath( basePartsForNormalise.path ); return URLToolkit.buildURLFromParts(basePartsForNormalise); } var relativeParts = URLToolkit.parseURL(relativeURL); if (!relativeParts) { throw new Error('Error trying to parse relative URL.'); } if (relativeParts.scheme) { // 2b) If the embedded URL starts with a scheme name, it is // interpreted as an absolute URL and we are done. if (!opts.alwaysNormalize) { return relativeURL; } relativeParts.path = URLToolkit.normalizePath(relativeParts.path); return URLToolkit.buildURLFromParts(relativeParts); } var baseParts = URLToolkit.parseURL(baseURL); if (!baseParts) { throw new Error('Error trying to parse base URL.'); } if (!baseParts.netLoc && baseParts.path && baseParts.path[0] !== '/') { // If netLoc missing and path doesn't start with '/', assume everthing before the first '/' is the netLoc // This causes 'example.com/a' to be handled as '//example.com/a' instead of '/example.com/a' var pathParts = FIRST_SEGMENT_REGEX.exec(baseParts.path); baseParts.netLoc = pathParts[1]; baseParts.path = pathParts[2]; } if (baseParts.netLoc && !baseParts.path) { baseParts.path = '/'; } var builtParts = { // 2c) Otherwise, the embedded URL inherits the scheme of // the base URL. scheme: baseParts.scheme, netLoc: relativeParts.netLoc, path: null, params: relativeParts.params, query: relativeParts.query, fragment: relativeParts.fragment, }; if (!relativeParts.netLoc) { // 3) If the embedded URL's <net_loc> is non-empty, we skip to // Step 7. Otherwise, the embedded URL inherits the <net_loc> // (if any) of the base URL. builtParts.netLoc = baseParts.netLoc; // 4) If the embedded URL path is preceded by a slash "/", the // path is not relative and we skip to Step 7. if (relativeParts.path[0] !== '/') { if (!relativeParts.path) { // 5) If the embedded URL path is empty (and not preceded by a // slash), then the embedded URL inherits the base URL path builtParts.path = baseParts.path; // 5a) if the embedded URL's <params> is non-empty, we skip to // step 7; otherwise, it inherits the <params> of the base // URL (if any) and if (!relativeParts.params) { builtParts.params = baseParts.params; // 5b) if the embedded URL's <query> is non-empty, we skip to // step 7; otherwise, it inherits the <query> of the base // URL (if any) and we skip to step 7. if (!relativeParts.query) { builtParts.query = baseParts.query; } } } else { // 6) The last segment of the base URL's path (anything // following the rightmost slash "/", or the entire path if no // slash is present) is removed and the embedded URL's path is // appended in its place. var baseURLPath = baseParts.path; var newPath = baseURLPath.substring(0, baseURLPath.lastIndexOf('/') + 1) + relativeParts.path; builtParts.path = URLToolkit.normalizePath(newPath); } } } if (builtParts.path === null) { builtParts.path = opts.alwaysNormalize ? URLToolkit.normalizePath(relativeParts.path) : relativeParts.path; } return URLToolkit.buildURLFromParts(builtParts); }, parseURL: function (url) { var parts = URL_REGEX.exec(url); if (!parts) { return null; } return { scheme: parts[1] || '', netLoc: parts[2] || '', path: parts[3] || '', params: parts[4] || '', query: parts[5] || '', fragment: parts[6] || '', }; }, normalizePath: function (path) { // The following operations are // then applied, in order, to the new path: // 6a) All occurrences of "./", where "." is a complete path // segment, are removed. // 6b) If the path ends with "." as a complete path segment, // that "." is removed. path = path.split('').reverse().join('').replace(SLASH_DOT_REGEX, ''); // 6c) All occurrences of "<segment>/../", where <segment> is a // complete path segment not equal to "..", are removed. // Removal of these path segments is performed iteratively, // removing the leftmost matching pattern on each iteration, // until no matching pattern remains. // 6d) If the path ends with "<segment>/..", where <segment> is a // complete path segment not equal to "..", that // "<segment>/.." is removed. while ( path.length !== (path = path.replace(SLASH_DOT_DOT_REGEX, '')).length ) {} return path.split('').reverse().join(''); }, buildURLFromParts: function (parts) { return ( parts.scheme + parts.netLoc + parts.path + parts.params + parts.query + parts.fragment ); }, }; if (true) module.exports = URLToolkit; else {} })(this); /***/ }), /***/ "./node_modules/webworkify-webpack/index.js": /*!**************************************************!*\ !*** ./node_modules/webworkify-webpack/index.js ***! \**************************************************/ /*! no static exports found */ /*! ModuleConcatenation bailout: Module is not an ECMAScript module */ /***/ (function(module, exports, __webpack_require__) { function webpackBootstrapFunc (modules) { /******/ // The module cache /******/ var installedModules = {}; /******/ // The require function /******/ function __webpack_require__(moduleId) { /******/ // Check if module is in cache /******/ if(installedModules[moduleId]) /******/ return installedModules[moduleId].exports; /******/ // Create a new module (and put it into the cache) /******/ var module = installedModules[moduleId] = { /******/ i: moduleId, /******/ l: false, /******/ exports: {} /******/ }; /******/ // Execute the module function /******/ modules[moduleId].call(module.exports, module, module.exports, __webpack_require__); /******/ // Flag the module as loaded /******/ module.l = true; /******/ // Return the exports of the module /******/ return module.exports; /******/ } /******/ // expose the modules object (__webpack_modules__) /******/ __webpack_require__.m = modules; /******/ // expose the module cache /******/ __webpack_require__.c = installedModules; /******/ // identity function for calling harmony imports with the correct context /******/ __webpack_require__.i = function(value) { return value; }; /******/ // define getter function for harmony exports /******/ __webpack_require__.d = function(exports, name, getter) { /******/ if(!__webpack_require__.o(exports, name)) { /******/ Object.defineProperty(exports, name, { /******/ configurable: false, /******/ enumerable: true, /******/ get: getter /******/ }); /******/ } /******/ }; /******/ // define __esModule on exports /******/ __webpack_require__.r = function(exports) { /******/ Object.defineProperty(exports, '__esModule', { value: true }); /******/ }; /******/ // getDefaultExport function for compatibility with non-harmony modules /******/ __webpack_require__.n = function(module) { /******/ var getter = module && module.__esModule ? /******/ function getDefault() { return module['default']; } : /******/ function getModuleExports() { return module; }; /******/ __webpack_require__.d(getter, 'a', getter); /******/ return getter; /******/ }; /******/ // Object.prototype.hasOwnProperty.call /******/ __webpack_require__.o = function(object, property) { return Object.prototype.hasOwnProperty.call(object, property); }; /******/ // __webpack_public_path__ /******/ __webpack_require__.p = "/"; /******/ // on error function for async loading /******/ __webpack_require__.oe = function(err) { console.error(err); throw err; }; var f = __webpack_require__(__webpack_require__.s = ENTRY_MODULE) return f.default || f // try to call default if defined to also support babel esmodule exports } var moduleNameReqExp = '[\\.|\\-|\\+|\\w|\/|@]+' var dependencyRegExp = '\\(\\s*(\/\\*.*?\\*\/)?\\s*.*?(' + moduleNameReqExp + ').*?\\)' // additional chars when output.pathinfo is true // http://stackoverflow.com/a/2593661/130442 function quoteRegExp (str) { return (str + '').replace(/[.?*+^$[\]\\(){}|-]/g, '\\$&') } function isNumeric(n) { return !isNaN(1 * n); // 1 * n converts integers, integers as string ("123"), 1e3 and "1e3" to integers and strings to NaN } function getModuleDependencies (sources, module, queueName) { var retval = {} retval[queueName] = [] var fnString = module.toString() var wrapperSignature = fnString.match(/^function\s?\w*\(\w+,\s*\w+,\s*(\w+)\)/) if (!wrapperSignature) return retval var webpackRequireName = wrapperSignature[1] // main bundle deps var re = new RegExp('(\\\\n|\\W)' + quoteRegExp(webpackRequireName) + dependencyRegExp, 'g') var match while ((match = re.exec(fnString))) { if (match[3] === 'dll-reference') continue retval[queueName].push(match[3]) } // dll deps re = new RegExp('\\(' + quoteRegExp(webpackRequireName) + '\\("(dll-reference\\s(' + moduleNameReqExp + '))"\\)\\)' + dependencyRegExp, 'g') while ((match = re.exec(fnString))) { if (!sources[match[2]]) { retval[queueName].push(match[1]) sources[match[2]] = __webpack_require__(match[1]).m } retval[match[2]] = retval[match[2]] || [] retval[match[2]].push(match[4]) } // convert 1e3 back to 1000 - this can be important after uglify-js converted 1000 to 1e3 var keys = Object.keys(retval); for (var i = 0; i < keys.length; i++) { for (var j = 0; j < retval[keys[i]].length; j++) { if (isNumeric(retval[keys[i]][j])) { retval[keys[i]][j] = 1 * retval[keys[i]][j]; } } } return retval } function hasValuesInQueues (queues) { var keys = Object.keys(queues) return keys.reduce(function (hasValues, key) { return hasValues || queues[key].length > 0 }, false) } function getRequiredModules (sources, moduleId) { var modulesQueue = { main: [moduleId] } var requiredModules = { main: [] } var seenModules = { main: {} } while (hasValuesInQueues(modulesQueue)) { var queues = Object.keys(modulesQueue) for (var i = 0; i < queues.length; i++) { var queueName = queues[i] var queue = modulesQueue[queueName] var moduleToCheck = queue.pop() seenModules[queueName] = seenModules[queueName] || {} if (seenModules[queueName][moduleToCheck] || !sources[queueName][moduleToCheck]) continue seenModules[queueName][moduleToCheck] = true requiredModules[queueName] = requiredModules[queueName] || [] requiredModules[queueName].push(moduleToCheck) var newModules = getModuleDependencies(sources, sources[queueName][moduleToCheck], queueName) var newModulesKeys = Object.keys(newModules) for (var j = 0; j < newModulesKeys.length; j++) { modulesQueue[newModulesKeys[j]] = modulesQueue[newModulesKeys[j]] || [] modulesQueue[newModulesKeys[j]] = modulesQueue[newModulesKeys[j]].concat(newModules[newModulesKeys[j]]) } } } return requiredModules } module.exports = function (moduleId, options) { options = options || {} var sources = { main: __webpack_require__.m } var requiredModules = options.all ? { main: Object.keys(sources.main) } : getRequiredModules(sources, moduleId) var src = '' Object.keys(requiredModules).filter(function (m) { return m !== 'main' }).forEach(function (module) { var entryModule = 0 while (requiredModules[module][entryModule]) { entryModule++ } requiredModules[module].push(entryModule) sources[module][entryModule] = '(function(module, exports, __webpack_require__) { module.exports = __webpack_require__; })' src = src + 'var ' + module + ' = (' + webpackBootstrapFunc.toString().replace('ENTRY_MODULE', JSON.stringify(entryModule)) + ')({' + requiredModules[module].map(function (id) { return '' + JSON.stringify(id) + ': ' + sources[module][id].toString() }).join(',') + '});\n' }) src = src + 'new ((' + webpackBootstrapFunc.toString().replace('ENTRY_MODULE', JSON.stringify(moduleId)) + ')({' + requiredModules.main.map(function (id) { return '' + JSON.stringify(id) + ': ' + sources.main[id].toString() }).join(',') + '}))(self);' var blob = new window.Blob([src], { type: 'text/javascript' }) if (options.bare) { return blob } var URL = window.URL || window.webkitURL || window.mozURL || window.msURL var workerUrl = URL.createObjectURL(blob) var worker = new window.Worker(workerUrl) worker.objectURL = workerUrl return worker } /***/ }), /***/ "./src/crypt/decrypter.js": /*!********************************************!*\ !*** ./src/crypt/decrypter.js + 3 modules ***! \********************************************/ /*! exports provided: default */ /*! ModuleConcatenation bailout: Cannot concat with ./src/errors.ts because of ./src/hls.ts */ /*! ModuleConcatenation bailout: Cannot concat with ./src/events.js because of ./src/hls.ts */ /*! ModuleConcatenation bailout: Cannot concat with ./src/utils/get-self-scope.js because of ./src/hls.ts */ /*! ModuleConcatenation bailout: Cannot concat with ./src/utils/logger.js because of ./src/hls.ts */ /***/ (function(module, __webpack_exports__, __webpack_require__) { "use strict"; // ESM COMPAT FLAG __webpack_require__.r(__webpack_exports__); // CONCATENATED MODULE: ./src/crypt/aes-crypto.js var AESCrypto = /*#__PURE__*/function () { function AESCrypto(subtle, iv) { this.subtle = subtle; this.aesIV = iv; } var _proto = AESCrypto.prototype; _proto.decrypt = function decrypt(data, key) { return this.subtle.decrypt({ name: 'AES-CBC', iv: this.aesIV }, key, data); }; return AESCrypto; }(); // CONCATENATED MODULE: ./src/crypt/fast-aes-key.js var FastAESKey = /*#__PURE__*/function () { function FastAESKey(subtle, key) { this.subtle = subtle; this.key = key; } var _proto = FastAESKey.prototype; _proto.expandKey = function expandKey() { return this.subtle.importKey('raw', this.key, { name: 'AES-CBC' }, false, ['encrypt', 'decrypt']); }; return FastAESKey; }(); /* harmony default export */ var fast_aes_key = (FastAESKey); // CONCATENATED MODULE: ./src/crypt/aes-decryptor.js // PKCS7 function removePadding(buffer) { var outputBytes = buffer.byteLength; var paddingBytes = outputBytes && new DataView(buffer).getUint8(outputBytes - 1); if (paddingBytes) { return buffer.slice(0, outputBytes - paddingBytes); } else { return buffer; } } var AESDecryptor = /*#__PURE__*/function () { function AESDecryptor() { // Static after running initTable this.rcon = [0x0, 0x1, 0x2, 0x4, 0x8, 0x10, 0x20, 0x40, 0x80, 0x1b, 0x36]; this.subMix = [new Uint32Array(256), new Uint32Array(256), new Uint32Array(256), new Uint32Array(256)]; this.invSubMix = [new Uint32Array(256), new Uint32Array(256), new Uint32Array(256), new Uint32Array(256)]; this.sBox = new Uint32Array(256); this.invSBox = new Uint32Array(256); // Changes during runtime this.key = new Uint32Array(0); this.initTable(); } // Using view.getUint32() also swaps the byte order. var _proto = AESDecryptor.prototype; _proto.uint8ArrayToUint32Array_ = function uint8ArrayToUint32Array_(arrayBuffer) { var view = new DataView(arrayBuffer); var newArray = new Uint32Array(4); for (var i = 0; i < 4; i++) { newArray[i] = view.getUint32(i * 4); } return newArray; }; _proto.initTable = function initTable() { var sBox = this.sBox; var invSBox = this.invSBox; var subMix = this.subMix; var subMix0 = subMix[0]; var subMix1 = subMix[1]; var subMix2 = subMix[2]; var subMix3 = subMix[3]; var invSubMix = this.invSubMix; var invSubMix0 = invSubMix[0]; var invSubMix1 = invSubMix[1]; var invSubMix2 = invSubMix[2]; var invSubMix3 = invSubMix[3]; var d = new Uint32Array(256); var x = 0; var xi = 0; var i = 0; for (i = 0; i < 256; i++) { if (i < 128) { d[i] = i << 1; } else { d[i] = i << 1 ^ 0x11b; } } for (i = 0; i < 256; i++) { var sx = xi ^ xi << 1 ^ xi << 2 ^ xi << 3 ^ xi << 4; sx = sx >>> 8 ^ sx & 0xff ^ 0x63; sBox[x] = sx; invSBox[sx] = x; // Compute multiplication var x2 = d[x]; var x4 = d[x2]; var x8 = d[x4]; // Compute sub/invSub bytes, mix columns tables var t = d[sx] * 0x101 ^ sx * 0x1010100; subMix0[x] = t << 24 | t >>> 8; subMix1[x] = t << 16 | t >>> 16; subMix2[x] = t << 8 | t >>> 24; subMix3[x] = t; // Compute inv sub bytes, inv mix columns tables t = x8 * 0x1010101 ^ x4 * 0x10001 ^ x2 * 0x101 ^ x * 0x1010100; invSubMix0[sx] = t << 24 | t >>> 8; invSubMix1[sx] = t << 16 | t >>> 16; invSubMix2[sx] = t << 8 | t >>> 24; invSubMix3[sx] = t; // Compute next counter if (!x) { x = xi = 1; } else { x = x2 ^ d[d[d[x8 ^ x2]]]; xi ^= d[d[xi]]; } } }; _proto.expandKey = function expandKey(keyBuffer) { // convert keyBuffer to Uint32Array var key = this.uint8ArrayToUint32Array_(keyBuffer); var sameKey = true; var offset = 0; while (offset < key.length && sameKey) { sameKey = key[offset] === this.key[offset]; offset++; } if (sameKey) { return; } this.key = key; var keySize = this.keySize = key.length; if (keySize !== 4 && keySize !== 6 && keySize !== 8) { throw new Error('Invalid aes key size=' + keySize); } var ksRows = this.ksRows = (keySize + 6 + 1) * 4; var ksRow; var invKsRow; var keySchedule = this.keySchedule = new Uint32Array(ksRows); var invKeySchedule = this.invKeySchedule = new Uint32Array(ksRows); var sbox = this.sBox; var rcon = this.rcon; var invSubMix = this.invSubMix; var invSubMix0 = invSubMix[0]; var invSubMix1 = invSubMix[1]; var invSubMix2 = invSubMix[2]; var invSubMix3 = invSubMix[3]; var prev; var t; for (ksRow = 0; ksRow < ksRows; ksRow++) { if (ksRow < keySize) { prev = keySchedule[ksRow] = key[ksRow]; continue; } t = prev; if (ksRow % keySize === 0) { // Rot word t = t << 8 | t >>> 24; // Sub word t = sbox[t >>> 24] << 24 | sbox[t >>> 16 & 0xff] << 16 | sbox[t >>> 8 & 0xff] << 8 | sbox[t & 0xff]; // Mix Rcon t ^= rcon[ksRow / keySize | 0] << 24; } else if (keySize > 6 && ksRow % keySize === 4) { // Sub word t = sbox[t >>> 24] << 24 | sbox[t >>> 16 & 0xff] << 16 | sbox[t >>> 8 & 0xff] << 8 | sbox[t & 0xff]; } keySchedule[ksRow] = prev = (keySchedule[ksRow - keySize] ^ t) >>> 0; } for (invKsRow = 0; invKsRow < ksRows; invKsRow++) { ksRow = ksRows - invKsRow; if (invKsRow & 3) { t = keySchedule[ksRow]; } else { t = keySchedule[ksRow - 4]; } if (invKsRow < 4 || ksRow <= 4) { invKeySchedule[invKsRow] = t; } else { invKeySchedule[invKsRow] = invSubMix0[sbox[t >>> 24]] ^ invSubMix1[sbox[t >>> 16 & 0xff]] ^ invSubMix2[sbox[t >>> 8 & 0xff]] ^ invSubMix3[sbox[t & 0xff]]; } invKeySchedule[invKsRow] = invKeySchedule[invKsRow] >>> 0; } } // Adding this as a method greatly improves performance. ; _proto.networkToHostOrderSwap = function networkToHostOrderSwap(word) { return word << 24 | (word & 0xff00) << 8 | (word & 0xff0000) >> 8 | word >>> 24; }; _proto.decrypt = function decrypt(inputArrayBuffer, offset, aesIV, removePKCS7Padding) { var nRounds = this.keySize + 6; var invKeySchedule = this.invKeySchedule; var invSBOX = this.invSBox; var invSubMix = this.invSubMix; var invSubMix0 = invSubMix[0]; var invSubMix1 = invSubMix[1]; var invSubMix2 = invSubMix[2]; var invSubMix3 = invSubMix[3]; var initVector = this.uint8ArrayToUint32Array_(aesIV); var initVector0 = initVector[0]; var initVector1 = initVector[1]; var initVector2 = initVector[2]; var initVector3 = initVector[3]; var inputInt32 = new Int32Array(inputArrayBuffer); var outputInt32 = new Int32Array(inputInt32.length); var t0, t1, t2, t3; var s0, s1, s2, s3; var inputWords0, inputWords1, inputWords2, inputWords3; var ksRow, i; var swapWord = this.networkToHostOrderSwap; while (offset < inputInt32.length) { inputWords0 = swapWord(inputInt32[offset]); inputWords1 = swapWord(inputInt32[offset + 1]); inputWords2 = swapWord(inputInt32[offset + 2]); inputWords3 = swapWord(inputInt32[offset + 3]); s0 = inputWords0 ^ invKeySchedule[0]; s1 = inputWords3 ^ invKeySchedule[1]; s2 = inputWords2 ^ invKeySchedule[2]; s3 = inputWords1 ^ invKeySchedule[3]; ksRow = 4; // Iterate through the rounds of decryption for (i = 1; i < nRounds; i++) { t0 = invSubMix0[s0 >>> 24] ^ invSubMix1[s1 >> 16 & 0xff] ^ invSubMix2[s2 >> 8 & 0xff] ^ invSubMix3[s3 & 0xff] ^ invKeySchedule[ksRow]; t1 = invSubMix0[s1 >>> 24] ^ invSubMix1[s2 >> 16 & 0xff] ^ invSubMix2[s3 >> 8 & 0xff] ^ invSubMix3[s0 & 0xff] ^ invKeySchedule[ksRow + 1]; t2 = invSubMix0[s2 >>> 24] ^ invSubMix1[s3 >> 16 & 0xff] ^ invSubMix2[s0 >> 8 & 0xff] ^ invSubMix3[s1 & 0xff] ^ invKeySchedule[ksRow + 2]; t3 = invSubMix0[s3 >>> 24] ^ invSubMix1[s0 >> 16 & 0xff] ^ invSubMix2[s1 >> 8 & 0xff] ^ invSubMix3[s2 & 0xff] ^ invKeySchedule[ksRow + 3]; // Update state s0 = t0; s1 = t1; s2 = t2; s3 = t3; ksRow = ksRow + 4; } // Shift rows, sub bytes, add round key t0 = invSBOX[s0 >>> 24] << 24 ^ invSBOX[s1 >> 16 & 0xff] << 16 ^ invSBOX[s2 >> 8 & 0xff] << 8 ^ invSBOX[s3 & 0xff] ^ invKeySchedule[ksRow]; t1 = invSBOX[s1 >>> 24] << 24 ^ invSBOX[s2 >> 16 & 0xff] << 16 ^ invSBOX[s3 >> 8 & 0xff] << 8 ^ invSBOX[s0 & 0xff] ^ invKeySchedule[ksRow + 1]; t2 = invSBOX[s2 >>> 24] << 24 ^ invSBOX[s3 >> 16 & 0xff] << 16 ^ invSBOX[s0 >> 8 & 0xff] << 8 ^ invSBOX[s1 & 0xff] ^ invKeySchedule[ksRow + 2]; t3 = invSBOX[s3 >>> 24] << 24 ^ invSBOX[s0 >> 16 & 0xff] << 16 ^ invSBOX[s1 >> 8 & 0xff] << 8 ^ invSBOX[s2 & 0xff] ^ invKeySchedule[ksRow + 3]; ksRow = ksRow + 3; // Write outputInt32[offset] = swapWord(t0 ^ initVector0); outputInt32[offset + 1] = swapWord(t3 ^ initVector1); outputInt32[offset + 2] = swapWord(t2 ^ initVector2); outputInt32[offset + 3] = swapWord(t1 ^ initVector3); // reset initVector to last 4 unsigned int initVector0 = inputWords0; initVector1 = inputWords1; initVector2 = inputWords2; initVector3 = inputWords3; offset = offset + 4; } return removePKCS7Padding ? removePadding(outputInt32.buffer) : outputInt32.buffer; }; _proto.destroy = function destroy() { this.key = undefined; this.keySize = undefined; this.ksRows = undefined; this.sBox = undefined; this.invSBox = undefined; this.subMix = undefined; this.invSubMix = undefined; this.keySchedule = undefined; this.invKeySchedule = undefined; this.rcon = undefined; }; return AESDecryptor; }(); /* harmony default export */ var aes_decryptor = (AESDecryptor); // EXTERNAL MODULE: ./src/errors.ts var errors = __webpack_require__("./src/errors.ts"); // EXTERNAL MODULE: ./src/utils/logger.js var logger = __webpack_require__("./src/utils/logger.js"); // EXTERNAL MODULE: ./src/events.js var events = __webpack_require__("./src/events.js"); // EXTERNAL MODULE: ./src/utils/get-self-scope.js var get_self_scope = __webpack_require__("./src/utils/get-self-scope.js"); // CONCATENATED MODULE: ./src/crypt/decrypter.js // see https://stackoverflow.com/a/11237259/589493 var global = Object(get_self_scope["getSelfScope"])(); // safeguard for code that might run both on worker and main thread var decrypter_Decrypter = /*#__PURE__*/function () { function Decrypter(observer, config, _temp) { var _ref = _temp === void 0 ? {} : _temp, _ref$removePKCS7Paddi = _ref.removePKCS7Padding, removePKCS7Padding = _ref$removePKCS7Paddi === void 0 ? true : _ref$removePKCS7Paddi; this.logEnabled = true; this.observer = observer; this.config = config; this.removePKCS7Padding = removePKCS7Padding; // built in decryptor expects PKCS7 padding if (removePKCS7Padding) { try { var browserCrypto = global.crypto; if (browserCrypto) { this.subtle = browserCrypto.subtle || browserCrypto.webkitSubtle; } } catch (e) {} } this.disableWebCrypto = !this.subtle; } var _proto = Decrypter.prototype; _proto.isSync = function isSync() { return this.disableWebCrypto && this.config.enableSoftwareAES; }; _proto.decrypt = function decrypt(data, key, iv, callback) { var _this = this; if (this.disableWebCrypto && this.config.enableSoftwareAES) { if (this.logEnabled) { logger["logger"].log('JS AES decrypt'); this.logEnabled = false; } var decryptor = this.decryptor; if (!decryptor) { this.decryptor = decryptor = new aes_decryptor(); } decryptor.expandKey(key); callback(decryptor.decrypt(data, 0, iv, this.removePKCS7Padding)); } else { if (this.logEnabled) { logger["logger"].log('WebCrypto AES decrypt'); this.logEnabled = false; } var subtle = this.subtle; if (this.key !== key) { this.key = key; this.fastAesKey = new fast_aes_key(subtle, key); } this.fastAesKey.expandKey().then(function (aesKey) { // decrypt using web crypto var crypto = new AESCrypto(subtle, iv); crypto.decrypt(data, aesKey).catch(function (err) { _this.onWebCryptoError(err, data, key, iv, callback); }).then(function (result) { callback(result); }); }).catch(function (err) { _this.onWebCryptoError(err, data, key, iv, callback); }); } }; _proto.onWebCryptoError = function onWebCryptoError(err, data, key, iv, callback) { if (this.config.enableSoftwareAES) { logger["logger"].log('WebCrypto Error, disable WebCrypto API'); this.disableWebCrypto = true; this.logEnabled = true; this.decrypt(data, key, iv, callback); } else { logger["logger"].error("decrypting error : " + err.message); this.observer.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].FRAG_DECRYPT_ERROR, fatal: true, reason: err.message }); } }; _proto.destroy = function destroy() { var decryptor = this.decryptor; if (decryptor) { decryptor.destroy(); this.decryptor = undefined; } }; return Decrypter; }(); /* harmony default export */ var decrypter = __webpack_exports__["default"] = (decrypter_Decrypter); /***/ }), /***/ "./src/demux/demuxer-inline.js": /*!**************************************************!*\ !*** ./src/demux/demuxer-inline.js + 12 modules ***! \**************************************************/ /*! exports provided: default */ /*! ModuleConcatenation bailout: Cannot concat with ./src/crypt/decrypter.js because of ./src/hls.ts */ /*! ModuleConcatenation bailout: Cannot concat with ./src/demux/id3.js because of ./src/hls.ts */ /*! ModuleConcatenation bailout: Cannot concat with ./src/demux/mp4demuxer.js because of ./src/hls.ts */ /*! ModuleConcatenation bailout: Cannot concat with ./src/errors.ts because of ./src/hls.ts */ /*! ModuleConcatenation bailout: Cannot concat with ./src/events.js because of ./src/hls.ts */ /*! ModuleConcatenation bailout: Cannot concat with ./src/polyfills/number.js because of ./src/hls.ts */ /*! ModuleConcatenation bailout: Cannot concat with ./src/utils/get-self-scope.js because of ./src/hls.ts */ /*! ModuleConcatenation bailout: Cannot concat with ./src/utils/logger.js because of ./src/hls.ts */ /***/ (function(module, __webpack_exports__, __webpack_require__) { "use strict"; // ESM COMPAT FLAG __webpack_require__.r(__webpack_exports__); // EXTERNAL MODULE: ./src/events.js var events = __webpack_require__("./src/events.js"); // EXTERNAL MODULE: ./src/errors.ts var errors = __webpack_require__("./src/errors.ts"); // EXTERNAL MODULE: ./src/crypt/decrypter.js + 3 modules var crypt_decrypter = __webpack_require__("./src/crypt/decrypter.js"); // EXTERNAL MODULE: ./src/polyfills/number.js var number = __webpack_require__("./src/polyfills/number.js"); // EXTERNAL MODULE: ./src/utils/logger.js var logger = __webpack_require__("./src/utils/logger.js"); // EXTERNAL MODULE: ./src/utils/get-self-scope.js var get_self_scope = __webpack_require__("./src/utils/get-self-scope.js"); // CONCATENATED MODULE: ./src/demux/adts.js /** * ADTS parser helper * @link https://wiki.multimedia.cx/index.php?title=ADTS */ function getAudioConfig(observer, data, offset, audioCodec) { var adtsObjectType, // :int adtsSampleingIndex, // :int adtsExtensionSampleingIndex, // :int adtsChanelConfig, // :int config, userAgent = navigator.userAgent.toLowerCase(), manifestCodec = audioCodec, adtsSampleingRates = [96000, 88200, 64000, 48000, 44100, 32000, 24000, 22050, 16000, 12000, 11025, 8000, 7350]; // byte 2 adtsObjectType = ((data[offset + 2] & 0xC0) >>> 6) + 1; adtsSampleingIndex = (data[offset + 2] & 0x3C) >>> 2; if (adtsSampleingIndex > adtsSampleingRates.length - 1) { observer.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].FRAG_PARSING_ERROR, fatal: true, reason: "invalid ADTS sampling index:" + adtsSampleingIndex }); return; } adtsChanelConfig = (data[offset + 2] & 0x01) << 2; // byte 3 adtsChanelConfig |= (data[offset + 3] & 0xC0) >>> 6; logger["logger"].log("manifest codec:" + audioCodec + ",ADTS data:type:" + adtsObjectType + ",sampleingIndex:" + adtsSampleingIndex + "[" + adtsSampleingRates[adtsSampleingIndex] + "Hz],channelConfig:" + adtsChanelConfig); // firefox: freq less than 24kHz = AAC SBR (HE-AAC) if (/firefox/i.test(userAgent)) { if (adtsSampleingIndex >= 6) { adtsObjectType = 5; config = new Array(4); // HE-AAC uses SBR (Spectral Band Replication) , high frequencies are constructed from low frequencies // there is a factor 2 between frame sample rate and output sample rate // multiply frequency by 2 (see table below, equivalent to substract 3) adtsExtensionSampleingIndex = adtsSampleingIndex - 3; } else { adtsObjectType = 2; config = new Array(2); adtsExtensionSampleingIndex = adtsSampleingIndex; } // Android : always use AAC } else if (userAgent.indexOf('android') !== -1) { adtsObjectType = 2; config = new Array(2); adtsExtensionSampleingIndex = adtsSampleingIndex; } else { /* for other browsers (Chrome/Vivaldi/Opera ...) always force audio type to be HE-AAC SBR, as some browsers do not support audio codec switch properly (like Chrome ...) */ adtsObjectType = 5; config = new Array(4); // if (manifest codec is HE-AAC or HE-AACv2) OR (manifest codec not specified AND frequency less than 24kHz) if (audioCodec && (audioCodec.indexOf('mp4a.40.29') !== -1 || audioCodec.indexOf('mp4a.40.5') !== -1) || !audioCodec && adtsSampleingIndex >= 6) { // HE-AAC uses SBR (Spectral Band Replication) , high frequencies are constructed from low frequencies // there is a factor 2 between frame sample rate and output sample rate // multiply frequency by 2 (see table below, equivalent to substract 3) adtsExtensionSampleingIndex = adtsSampleingIndex - 3; } else { // if (manifest codec is AAC) AND (frequency less than 24kHz AND nb channel is 1) OR (manifest codec not specified and mono audio) // Chrome fails to play back with low frequency AAC LC mono when initialized with HE-AAC. This is not a problem with stereo. if (audioCodec && audioCodec.indexOf('mp4a.40.2') !== -1 && (adtsSampleingIndex >= 6 && adtsChanelConfig === 1 || /vivaldi/i.test(userAgent)) || !audioCodec && adtsChanelConfig === 1) { adtsObjectType = 2; config = new Array(2); } adtsExtensionSampleingIndex = adtsSampleingIndex; } } /* refer to http://wiki.multimedia.cx/index.php?title=MPEG-4_Audio#Audio_Specific_Config ISO 14496-3 (AAC).pdf - Table 1.13 — Syntax of AudioSpecificConfig() Audio Profile / Audio Object Type 0: Null 1: AAC Main 2: AAC LC (Low Complexity) 3: AAC SSR (Scalable Sample Rate) 4: AAC LTP (Long Term Prediction) 5: SBR (Spectral Band Replication) 6: AAC Scalable sampling freq 0: 96000 Hz 1: 88200 Hz 2: 64000 Hz 3: 48000 Hz 4: 44100 Hz 5: 32000 Hz 6: 24000 Hz 7: 22050 Hz 8: 16000 Hz 9: 12000 Hz 10: 11025 Hz 11: 8000 Hz 12: 7350 Hz 13: Reserved 14: Reserved 15: frequency is written explictly Channel Configurations These are the channel configurations: 0: Defined in AOT Specifc Config 1: 1 channel: front-center 2: 2 channels: front-left, front-right */ // audioObjectType = profile => profile, the MPEG-4 Audio Object Type minus 1 config[0] = adtsObjectType << 3; // samplingFrequencyIndex config[0] |= (adtsSampleingIndex & 0x0E) >> 1; config[1] |= (adtsSampleingIndex & 0x01) << 7; // channelConfiguration config[1] |= adtsChanelConfig << 3; if (adtsObjectType === 5) { // adtsExtensionSampleingIndex config[1] |= (adtsExtensionSampleingIndex & 0x0E) >> 1; config[2] = (adtsExtensionSampleingIndex & 0x01) << 7; // adtsObjectType (force to 2, chrome is checking that object type is less than 5 ??? // https://chromium.googlesource.com/chromium/src.git/+/master/media/formats/mp4/aac.cc config[2] |= 2 << 2; config[3] = 0; } return { config: config, samplerate: adtsSampleingRates[adtsSampleingIndex], channelCount: adtsChanelConfig, codec: 'mp4a.40.' + adtsObjectType, manifestCodec: manifestCodec }; } function isHeaderPattern(data, offset) { return data[offset] === 0xff && (data[offset + 1] & 0xf6) === 0xf0; } function getHeaderLength(data, offset) { return data[offset + 1] & 0x01 ? 7 : 9; } function getFullFrameLength(data, offset) { return (data[offset + 3] & 0x03) << 11 | data[offset + 4] << 3 | (data[offset + 5] & 0xE0) >>> 5; } function isHeader(data, offset) { // Look for ADTS header | 1111 1111 | 1111 X00X | where X can be either 0 or 1 // Layer bits (position 14 and 15) in header should be always 0 for ADTS // More info https://wiki.multimedia.cx/index.php?title=ADTS if (offset + 1 < data.length && isHeaderPattern(data, offset)) { return true; } return false; } function adts_probe(data, offset) { // same as isHeader but we also check that ADTS frame follows last ADTS frame // or end of data is reached if (isHeader(data, offset)) { // ADTS header Length var headerLength = getHeaderLength(data, offset); if (offset + headerLength >= data.length) { return false; } // ADTS frame Length var frameLength = getFullFrameLength(data, offset); if (frameLength <= headerLength) { return false; } var newOffset = offset + frameLength; if (newOffset === data.length || newOffset + 1 < data.length && isHeaderPattern(data, newOffset)) { return true; } } return false; } function initTrackConfig(track, observer, data, offset, audioCodec) { if (!track.samplerate) { var config = getAudioConfig(observer, data, offset, audioCodec); track.config = config.config; track.samplerate = config.samplerate; track.channelCount = config.channelCount; track.codec = config.codec; track.manifestCodec = config.manifestCodec; logger["logger"].log("parsed codec:" + track.codec + ",rate:" + config.samplerate + ",nb channel:" + config.channelCount); } } function getFrameDuration(samplerate) { return 1024 * 90000 / samplerate; } function parseFrameHeader(data, offset, pts, frameIndex, frameDuration) { var headerLength, frameLength, stamp; var length = data.length; // The protection skip bit tells us if we have 2 bytes of CRC data at the end of the ADTS header headerLength = getHeaderLength(data, offset); // retrieve frame size frameLength = getFullFrameLength(data, offset); frameLength -= headerLength; if (frameLength > 0 && offset + headerLength + frameLength <= length) { stamp = pts + frameIndex * frameDuration; // logger.log(`AAC frame, offset/length/total/pts:${offset+headerLength}/${frameLength}/${data.byteLength}/${(stamp/90).toFixed(0)}`); return { headerLength: headerLength, frameLength: frameLength, stamp: stamp }; } return undefined; } function appendFrame(track, data, offset, pts, frameIndex) { var frameDuration = getFrameDuration(track.samplerate); var header = parseFrameHeader(data, offset, pts, frameIndex, frameDuration); if (header) { var stamp = header.stamp; var headerLength = header.headerLength; var frameLength = header.frameLength; // logger.log(`AAC frame, offset/length/total/pts:${offset+headerLength}/${frameLength}/${data.byteLength}/${(stamp/90).toFixed(0)}`); var aacSample = { unit: data.subarray(offset + headerLength, offset + headerLength + frameLength), pts: stamp, dts: stamp }; track.samples.push(aacSample); return { sample: aacSample, length: frameLength + headerLength }; } return undefined; } // EXTERNAL MODULE: ./src/demux/id3.js var id3 = __webpack_require__("./src/demux/id3.js"); // CONCATENATED MODULE: ./src/demux/aacdemuxer.js /** * AAC demuxer */ var aacdemuxer_AACDemuxer = /*#__PURE__*/function () { function AACDemuxer(observer, remuxer, config) { this.observer = observer; this.config = config; this.remuxer = remuxer; } var _proto = AACDemuxer.prototype; _proto.resetInitSegment = function resetInitSegment(initSegment, audioCodec, videoCodec, duration) { this._audioTrack = { container: 'audio/adts', type: 'audio', id: 0, sequenceNumber: 0, isAAC: true, samples: [], len: 0, manifestCodec: audioCodec, duration: duration, inputTimeScale: 90000 }; }; _proto.resetTimeStamp = function resetTimeStamp() {}; AACDemuxer.probe = function probe(data) { if (!data) { return false; } // Check for the ADTS sync word // Look for ADTS header | 1111 1111 | 1111 X00X | where X can be either 0 or 1 // Layer bits (position 14 and 15) in header should be always 0 for ADTS // More info https://wiki.multimedia.cx/index.php?title=ADTS var id3Data = id3["default"].getID3Data(data, 0) || []; var offset = id3Data.length; for (var length = data.length; offset < length; offset++) { if (adts_probe(data, offset)) { logger["logger"].log('ADTS sync word found !'); return true; } } return false; } // feed incoming data to the front of the parsing pipeline ; _proto.append = function append(data, timeOffset, contiguous, accurateTimeOffset) { var track = this._audioTrack; var id3Data = id3["default"].getID3Data(data, 0) || []; var timestamp = id3["default"].getTimeStamp(id3Data); var pts = Object(number["isFiniteNumber"])(timestamp) ? timestamp * 90 : timeOffset * 90000; var frameIndex = 0; var stamp = pts; var length = data.length; var offset = id3Data.length; var id3Samples = [{ pts: stamp, dts: stamp, data: id3Data }]; while (offset < length - 1) { if (isHeader(data, offset) && offset + 5 < length) { initTrackConfig(track, this.observer, data, offset, track.manifestCodec); var frame = appendFrame(track, data, offset, pts, frameIndex); if (frame) { offset += frame.length; stamp = frame.sample.pts; frameIndex++; } else { logger["logger"].log('Unable to parse AAC frame'); break; } } else if (id3["default"].isHeader(data, offset)) { id3Data = id3["default"].getID3Data(data, offset); id3Samples.push({ pts: stamp, dts: stamp, data: id3Data }); offset += id3Data.length; } else { // nothing found, keep looking offset++; } } this.remuxer.remux(track, { samples: [] }, { samples: id3Samples, inputTimeScale: 90000 }, { samples: [] }, timeOffset, contiguous, accurateTimeOffset); }; _proto.destroy = function destroy() {}; return AACDemuxer; }(); /* harmony default export */ var aacdemuxer = (aacdemuxer_AACDemuxer); // EXTERNAL MODULE: ./src/demux/mp4demuxer.js var mp4demuxer = __webpack_require__("./src/demux/mp4demuxer.js"); // CONCATENATED MODULE: ./src/demux/mpegaudio.js /** * MPEG parser helper */ var MpegAudio = { BitratesMap: [32, 64, 96, 128, 160, 192, 224, 256, 288, 320, 352, 384, 416, 448, 32, 48, 56, 64, 80, 96, 112, 128, 160, 192, 224, 256, 320, 384, 32, 40, 48, 56, 64, 80, 96, 112, 128, 160, 192, 224, 256, 320, 32, 48, 56, 64, 80, 96, 112, 128, 144, 160, 176, 192, 224, 256, 8, 16, 24, 32, 40, 48, 56, 64, 80, 96, 112, 128, 144, 160], SamplingRateMap: [44100, 48000, 32000, 22050, 24000, 16000, 11025, 12000, 8000], SamplesCoefficients: [// MPEG 2.5 [0, // Reserved 72, // Layer3 144, // Layer2 12 // Layer1 ], // Reserved [0, // Reserved 0, // Layer3 0, // Layer2 0 // Layer1 ], // MPEG 2 [0, // Reserved 72, // Layer3 144, // Layer2 12 // Layer1 ], // MPEG 1 [0, // Reserved 144, // Layer3 144, // Layer2 12 // Layer1 ]], BytesInSlot: [0, // Reserved 1, // Layer3 1, // Layer2 4 // Layer1 ], appendFrame: function appendFrame(track, data, offset, pts, frameIndex) { // Using http://www.datavoyage.com/mpgscript/mpeghdr.htm as a reference if (offset + 24 > data.length) { return undefined; } var header = this.parseHeader(data, offset); if (header && offset + header.frameLength <= data.length) { var frameDuration = header.samplesPerFrame * 90000 / header.sampleRate; var stamp = pts + frameIndex * frameDuration; var sample = { unit: data.subarray(offset, offset + header.frameLength), pts: stamp, dts: stamp }; track.config = []; track.channelCount = header.channelCount; track.samplerate = header.sampleRate; track.samples.push(sample); return { sample: sample, length: header.frameLength }; } return undefined; }, parseHeader: function parseHeader(data, offset) { var headerB = data[offset + 1] >> 3 & 3; var headerC = data[offset + 1] >> 1 & 3; var headerE = data[offset + 2] >> 4 & 15; var headerF = data[offset + 2] >> 2 & 3; var headerG = data[offset + 2] >> 1 & 1; if (headerB !== 1 && headerE !== 0 && headerE !== 15 && headerF !== 3) { var columnInBitrates = headerB === 3 ? 3 - headerC : headerC === 3 ? 3 : 4; var bitRate = MpegAudio.BitratesMap[columnInBitrates * 14 + headerE - 1] * 1000; var columnInSampleRates = headerB === 3 ? 0 : headerB === 2 ? 1 : 2; var sampleRate = MpegAudio.SamplingRateMap[columnInSampleRates * 3 + headerF]; var channelCount = data[offset + 3] >> 6 === 3 ? 1 : 2; // If bits of channel mode are `11` then it is a single channel (Mono) var sampleCoefficient = MpegAudio.SamplesCoefficients[headerB][headerC]; var bytesInSlot = MpegAudio.BytesInSlot[headerC]; var samplesPerFrame = sampleCoefficient * 8 * bytesInSlot; var frameLength = parseInt(sampleCoefficient * bitRate / sampleRate + headerG, 10) * bytesInSlot; return { sampleRate: sampleRate, channelCount: channelCount, frameLength: frameLength, samplesPerFrame: samplesPerFrame }; } return undefined; }, isHeaderPattern: function isHeaderPattern(data, offset) { return data[offset] === 0xff && (data[offset + 1] & 0xe0) === 0xe0 && (data[offset + 1] & 0x06) !== 0x00; }, isHeader: function isHeader(data, offset) { // Look for MPEG header | 1111 1111 | 111X XYZX | where X can be either 0 or 1 and Y or Z should be 1 // Layer bits (position 14 and 15) in header should be always different from 0 (Layer I or Layer II or Layer III) // More info http://www.mp3-tech.org/programmer/frame_header.html if (offset + 1 < data.length && this.isHeaderPattern(data, offset)) { return true; } return false; }, probe: function probe(data, offset) { // same as isHeader but we also check that MPEG frame follows last MPEG frame // or end of data is reached if (offset + 1 < data.length && this.isHeaderPattern(data, offset)) { // MPEG header Length var headerLength = 4; // MPEG frame Length var header = this.parseHeader(data, offset); var frameLength = headerLength; if (header && header.frameLength) { frameLength = header.frameLength; } var newOffset = offset + frameLength; if (newOffset === data.length || newOffset + 1 < data.length && this.isHeaderPattern(data, newOffset)) { return true; } } return false; } }; /* harmony default export */ var mpegaudio = (MpegAudio); // CONCATENATED MODULE: ./src/demux/exp-golomb.js /** * Parser for exponential Golomb codes, a variable-bitwidth number encoding scheme used by h264. */ var exp_golomb_ExpGolomb = /*#__PURE__*/function () { function ExpGolomb(data) { this.data = data; // the number of bytes left to examine in this.data this.bytesAvailable = data.byteLength; // the current word being examined this.word = 0; // :uint // the number of bits left to examine in the current word this.bitsAvailable = 0; // :uint } // ():void var _proto = ExpGolomb.prototype; _proto.loadWord = function loadWord() { var data = this.data, bytesAvailable = this.bytesAvailable, position = data.byteLength - bytesAvailable, workingBytes = new Uint8Array(4), availableBytes = Math.min(4, bytesAvailable); if (availableBytes === 0) { throw new Error('no bytes available'); } workingBytes.set(data.subarray(position, position + availableBytes)); this.word = new DataView(workingBytes.buffer).getUint32(0); // track the amount of this.data that has been processed this.bitsAvailable = availableBytes * 8; this.bytesAvailable -= availableBytes; } // (count:int):void ; _proto.skipBits = function skipBits(count) { var skipBytes; // :int if (this.bitsAvailable > count) { this.word <<= count; this.bitsAvailable -= count; } else { count -= this.bitsAvailable; skipBytes = count >> 3; count -= skipBytes >> 3; this.bytesAvailable -= skipBytes; this.loadWord(); this.word <<= count; this.bitsAvailable -= count; } } // (size:int):uint ; _proto.readBits = function readBits(size) { var bits = Math.min(this.bitsAvailable, size), // :uint valu = this.word >>> 32 - bits; // :uint if (size > 32) { logger["logger"].error('Cannot read more than 32 bits at a time'); } this.bitsAvailable -= bits; if (this.bitsAvailable > 0) { this.word <<= bits; } else if (this.bytesAvailable > 0) { this.loadWord(); } bits = size - bits; if (bits > 0 && this.bitsAvailable) { return valu << bits | this.readBits(bits); } else { return valu; } } // ():uint ; _proto.skipLZ = function skipLZ() { var leadingZeroCount; // :uint for (leadingZeroCount = 0; leadingZeroCount < this.bitsAvailable; ++leadingZeroCount) { if ((this.word & 0x80000000 >>> leadingZeroCount) !== 0) { // the first bit of working word is 1 this.word <<= leadingZeroCount; this.bitsAvailable -= leadingZeroCount; return leadingZeroCount; } } // we exhausted word and still have not found a 1 this.loadWord(); return leadingZeroCount + this.skipLZ(); } // ():void ; _proto.skipUEG = function skipUEG() { this.skipBits(1 + this.skipLZ()); } // ():void ; _proto.skipEG = function skipEG() { this.skipBits(1 + this.skipLZ()); } // ():uint ; _proto.readUEG = function readUEG() { var clz = this.skipLZ(); // :uint return this.readBits(clz + 1) - 1; } // ():int ; _proto.readEG = function readEG() { var valu = this.readUEG(); // :int if (0x01 & valu) { // the number is odd if the low order bit is set return 1 + valu >>> 1; // add 1 to make it even, and divide by 2 } else { return -1 * (valu >>> 1); // divide by two then make it negative } } // Some convenience functions // :Boolean ; _proto.readBoolean = function readBoolean() { return this.readBits(1) === 1; } // ():int ; _proto.readUByte = function readUByte() { return this.readBits(8); } // ():int ; _proto.readUShort = function readUShort() { return this.readBits(16); } // ():int ; _proto.readUInt = function readUInt() { return this.readBits(32); } /** * Advance the ExpGolomb decoder past a scaling list. The scaling * list is optionally transmitted as part of a sequence parameter * set and is not relevant to transmuxing. * @param count {number} the number of entries in this scaling list * @see Recommendation ITU-T H.264, Section 7.3.2.1.1.1 */ ; _proto.skipScalingList = function skipScalingList(count) { var lastScale = 8, nextScale = 8, j, deltaScale; for (j = 0; j < count; j++) { if (nextScale !== 0) { deltaScale = this.readEG(); nextScale = (lastScale + deltaScale + 256) % 256; } lastScale = nextScale === 0 ? lastScale : nextScale; } } /** * Read a sequence parameter set and return some interesting video * properties. A sequence parameter set is the H264 metadata that * describes the properties of upcoming video frames. * @param data {Uint8Array} the bytes of a sequence parameter set * @return {object} an object with configuration parsed from the * sequence parameter set, including the dimensions of the * associated video frames. */ ; _proto.readSPS = function readSPS() { var frameCropLeftOffset = 0, frameCropRightOffset = 0, frameCropTopOffset = 0, frameCropBottomOffset = 0, profileIdc, profileCompat, levelIdc, numRefFramesInPicOrderCntCycle, picWidthInMbsMinus1, picHeightInMapUnitsMinus1, frameMbsOnlyFlag, scalingListCount, i, readUByte = this.readUByte.bind(this), readBits = this.readBits.bind(this), readUEG = this.readUEG.bind(this), readBoolean = this.readBoolean.bind(this), skipBits = this.skipBits.bind(this), skipEG = this.skipEG.bind(this), skipUEG = this.skipUEG.bind(this), skipScalingList = this.skipScalingList.bind(this); readUByte(); profileIdc = readUByte(); // profile_idc profileCompat = readBits(5); // constraint_set[0-4]_flag, u(5) skipBits(3); // reserved_zero_3bits u(3), levelIdc = readUByte(); // level_idc u(8) skipUEG(); // seq_parameter_set_id // some profiles have more optional data we don't need if (profileIdc === 100 || profileIdc === 110 || profileIdc === 122 || profileIdc === 244 || profileIdc === 44 || profileIdc === 83 || profileIdc === 86 || profileIdc === 118 || profileIdc === 128) { var chromaFormatIdc = readUEG(); if (chromaFormatIdc === 3) { skipBits(1); } // separate_colour_plane_flag skipUEG(); // bit_depth_luma_minus8 skipUEG(); // bit_depth_chroma_minus8 skipBits(1); // qpprime_y_zero_transform_bypass_flag if (readBoolean()) { // seq_scaling_matrix_present_flag scalingListCount = chromaFormatIdc !== 3 ? 8 : 12; for (i = 0; i < scalingListCount; i++) { if (readBoolean()) { // seq_scaling_list_present_flag[ i ] if (i < 6) { skipScalingList(16); } else { skipScalingList(64); } } } } } skipUEG(); // log2_max_frame_num_minus4 var picOrderCntType = readUEG(); if (picOrderCntType === 0) { readUEG(); // log2_max_pic_order_cnt_lsb_minus4 } else if (picOrderCntType === 1) { skipBits(1); // delta_pic_order_always_zero_flag skipEG(); // offset_for_non_ref_pic skipEG(); // offset_for_top_to_bottom_field numRefFramesInPicOrderCntCycle = readUEG(); for (i = 0; i < numRefFramesInPicOrderCntCycle; i++) { skipEG(); } // offset_for_ref_frame[ i ] } skipUEG(); // max_num_ref_frames skipBits(1); // gaps_in_frame_num_value_allowed_flag picWidthInMbsMinus1 = readUEG(); picHeightInMapUnitsMinus1 = readUEG(); frameMbsOnlyFlag = readBits(1); if (frameMbsOnlyFlag === 0) { skipBits(1); } // mb_adaptive_frame_field_flag skipBits(1); // direct_8x8_inference_flag if (readBoolean()) { // frame_cropping_flag frameCropLeftOffset = readUEG(); frameCropRightOffset = readUEG(); frameCropTopOffset = readUEG(); frameCropBottomOffset = readUEG(); } var pixelRatio = [1, 1]; if (readBoolean()) { // vui_parameters_present_flag if (readBoolean()) { // aspect_ratio_info_present_flag var aspectRatioIdc = readUByte(); switch (aspectRatioIdc) { case 1: pixelRatio = [1, 1]; break; case 2: pixelRatio = [12, 11]; break; case 3: pixelRatio = [10, 11]; break; case 4: pixelRatio = [16, 11]; break; case 5: pixelRatio = [40, 33]; break; case 6: pixelRatio = [24, 11]; break; case 7: pixelRatio = [20, 11]; break; case 8: pixelRatio = [32, 11]; break; case 9: pixelRatio = [80, 33]; break; case 10: pixelRatio = [18, 11]; break; case 11: pixelRatio = [15, 11]; break; case 12: pixelRatio = [64, 33]; break; case 13: pixelRatio = [160, 99]; break; case 14: pixelRatio = [4, 3]; break; case 15: pixelRatio = [3, 2]; break; case 16: pixelRatio = [2, 1]; break; case 255: { pixelRatio = [readUByte() << 8 | readUByte(), readUByte() << 8 | readUByte()]; break; } } } } return { width: Math.ceil((picWidthInMbsMinus1 + 1) * 16 - frameCropLeftOffset * 2 - frameCropRightOffset * 2), height: (2 - frameMbsOnlyFlag) * (picHeightInMapUnitsMinus1 + 1) * 16 - (frameMbsOnlyFlag ? 2 : 4) * (frameCropTopOffset + frameCropBottomOffset), pixelRatio: pixelRatio }; }; _proto.readSliceType = function readSliceType() { // skip NALu type this.readUByte(); // discard first_mb_in_slice this.readUEG(); // return slice_type return this.readUEG(); }; return ExpGolomb; }(); /* harmony default export */ var exp_golomb = (exp_golomb_ExpGolomb); // CONCATENATED MODULE: ./src/demux/sample-aes.js /** * SAMPLE-AES decrypter */ var sample_aes_SampleAesDecrypter = /*#__PURE__*/function () { function SampleAesDecrypter(observer, config, decryptdata, discardEPB) { this.decryptdata = decryptdata; this.discardEPB = discardEPB; this.decrypter = new crypt_decrypter["default"](observer, config, { removePKCS7Padding: false }); } var _proto = SampleAesDecrypter.prototype; _proto.decryptBuffer = function decryptBuffer(encryptedData, callback) { this.decrypter.decrypt(encryptedData, this.decryptdata.key.buffer, this.decryptdata.iv.buffer, callback); } // AAC - encrypt all full 16 bytes blocks starting from offset 16 ; _proto.decryptAacSample = function decryptAacSample(samples, sampleIndex, callback, sync) { var curUnit = samples[sampleIndex].unit; var encryptedData = curUnit.subarray(16, curUnit.length - curUnit.length % 16); var encryptedBuffer = encryptedData.buffer.slice(encryptedData.byteOffset, encryptedData.byteOffset + encryptedData.length); var localthis = this; this.decryptBuffer(encryptedBuffer, function (decryptedData) { decryptedData = new Uint8Array(decryptedData); curUnit.set(decryptedData, 16); if (!sync) { localthis.decryptAacSamples(samples, sampleIndex + 1, callback); } }); }; _proto.decryptAacSamples = function decryptAacSamples(samples, sampleIndex, callback) { for (;; sampleIndex++) { if (sampleIndex >= samples.length) { callback(); return; } if (samples[sampleIndex].unit.length < 32) { continue; } var sync = this.decrypter.isSync(); this.decryptAacSample(samples, sampleIndex, callback, sync); if (!sync) { return; } } } // AVC - encrypt one 16 bytes block out of ten, starting from offset 32 ; _proto.getAvcEncryptedData = function getAvcEncryptedData(decodedData) { var encryptedDataLen = Math.floor((decodedData.length - 48) / 160) * 16 + 16; var encryptedData = new Int8Array(encryptedDataLen); var outputPos = 0; for (var inputPos = 32; inputPos <= decodedData.length - 16; inputPos += 160, outputPos += 16) { encryptedData.set(decodedData.subarray(inputPos, inputPos + 16), outputPos); } return encryptedData; }; _proto.getAvcDecryptedUnit = function getAvcDecryptedUnit(decodedData, decryptedData) { decryptedData = new Uint8Array(decryptedData); var inputPos = 0; for (var outputPos = 32; outputPos <= decodedData.length - 16; outputPos += 160, inputPos += 16) { decodedData.set(decryptedData.subarray(inputPos, inputPos + 16), outputPos); } return decodedData; }; _proto.decryptAvcSample = function decryptAvcSample(samples, sampleIndex, unitIndex, callback, curUnit, sync) { var decodedData = this.discardEPB(curUnit.data); var encryptedData = this.getAvcEncryptedData(decodedData); var localthis = this; this.decryptBuffer(encryptedData.buffer, function (decryptedData) { curUnit.data = localthis.getAvcDecryptedUnit(decodedData, decryptedData); if (!sync) { localthis.decryptAvcSamples(samples, sampleIndex, unitIndex + 1, callback); } }); }; _proto.decryptAvcSamples = function decryptAvcSamples(samples, sampleIndex, unitIndex, callback) { for (;; sampleIndex++, unitIndex = 0) { if (sampleIndex >= samples.length) { callback(); return; } var curUnits = samples[sampleIndex].units; for (;; unitIndex++) { if (unitIndex >= curUnits.length) { break; } var curUnit = curUnits[unitIndex]; if (curUnit.length <= 48 || curUnit.type !== 1 && curUnit.type !== 5) { continue; } var sync = this.decrypter.isSync(); this.decryptAvcSample(samples, sampleIndex, unitIndex, callback, curUnit, sync); if (!sync) { return; } } } }; return SampleAesDecrypter; }(); /* harmony default export */ var sample_aes = (sample_aes_SampleAesDecrypter); // CONCATENATED MODULE: ./src/demux/tsdemuxer.js /** * highly optimized TS demuxer: * parse PAT, PMT * extract PES packet from audio and video PIDs * extract AVC/H264 NAL units and AAC/ADTS samples from PES packet * trigger the remuxer upon parsing completion * it also tries to workaround as best as it can audio codec switch (HE-AAC to AAC and vice versa), without having to restart the MediaSource. * it also controls the remuxing process : * upon discontinuity or level switch detection, it will also notifies the remuxer so that it can reset its state. */ // import Hex from '../utils/hex'; // We are using fixed track IDs for driving the MP4 remuxer // instead of following the TS PIDs. // There is no reason not to do this and some browsers/SourceBuffer-demuxers // may not like if there are TrackID "switches" // See https://github.com/video-dev/hls.js/issues/1331 // Here we are mapping our internal track types to constant MP4 track IDs // With MSE currently one can only have one track of each, and we are muxing // whatever video/audio rendition in them. var RemuxerTrackIdConfig = { video: 1, audio: 2, id3: 3, text: 4 }; var tsdemuxer_TSDemuxer = /*#__PURE__*/function () { function TSDemuxer(observer, remuxer, config, typeSupported) { this.observer = observer; this.config = config; this.typeSupported = typeSupported; this.remuxer = remuxer; this.sampleAes = null; this.pmtUnknownTypes = {}; } var _proto = TSDemuxer.prototype; _proto.setDecryptData = function setDecryptData(decryptdata) { if (decryptdata != null && decryptdata.key != null && decryptdata.method === 'SAMPLE-AES') { this.sampleAes = new sample_aes(this.observer, this.config, decryptdata, this.discardEPB); } else { this.sampleAes = null; } }; TSDemuxer.probe = function probe(data) { var syncOffset = TSDemuxer._syncOffset(data); if (syncOffset < 0) { return false; } else { if (syncOffset) { logger["logger"].warn("MPEG2-TS detected but first sync word found @ offset " + syncOffset + ", junk ahead ?"); } return true; } }; TSDemuxer._syncOffset = function _syncOffset(data) { // scan 1000 first bytes var scanwindow = Math.min(1000, data.length - 3 * 188); var i = 0; while (i < scanwindow) { // a TS fragment should contain at least 3 TS packets, a PAT, a PMT, and one PID, each starting with 0x47 if (data[i] === 0x47 && data[i + 188] === 0x47 && data[i + 2 * 188] === 0x47) { return i; } else { i++; } } return -1; } /** * Creates a track model internal to demuxer used to drive remuxing input * * @param {string} type 'audio' | 'video' | 'id3' | 'text' * @param {number} duration * @return {object} TSDemuxer's internal track model */ ; TSDemuxer.createTrack = function createTrack(type, duration) { return { container: type === 'video' || type === 'audio' ? 'video/mp2t' : undefined, type: type, id: RemuxerTrackIdConfig[type], pid: -1, inputTimeScale: 90000, sequenceNumber: 0, samples: [], dropped: type === 'video' ? 0 : undefined, isAAC: type === 'audio' ? true : undefined, duration: type === 'audio' ? duration : undefined }; } /** * Initializes a new init segment on the demuxer/remuxer interface. Needed for discontinuities/track-switches (or at stream start) * Resets all internal track instances of the demuxer. * * @override Implements generic demuxing/remuxing interface (see DemuxerInline) * @param {object} initSegment * @param {string} audioCodec * @param {string} videoCodec * @param {number} duration (in TS timescale = 90kHz) */ ; _proto.resetInitSegment = function resetInitSegment(initSegment, audioCodec, videoCodec, duration) { this.pmtParsed = false; this._pmtId = -1; this.pmtUnknownTypes = {}; this._avcTrack = TSDemuxer.createTrack('video', duration); this._audioTrack = TSDemuxer.createTrack('audio', duration); this._id3Track = TSDemuxer.createTrack('id3', duration); this._txtTrack = TSDemuxer.createTrack('text', duration); // flush any partial content this.aacOverFlow = null; this.aacLastPTS = null; this.avcSample = null; this.audioCodec = audioCodec; this.videoCodec = videoCodec; this._duration = duration; } /** * * @override */ ; _proto.resetTimeStamp = function resetTimeStamp() {} // feed incoming data to the front of the parsing pipeline ; _proto.append = function append(data, timeOffset, contiguous, accurateTimeOffset) { var start, len = data.length, stt, pid, atf, offset, pes, unknownPIDs = false; this.pmtUnknownTypes = {}; this.contiguous = contiguous; var pmtParsed = this.pmtParsed, avcTrack = this._avcTrack, audioTrack = this._audioTrack, id3Track = this._id3Track, avcId = avcTrack.pid, audioId = audioTrack.pid, id3Id = id3Track.pid, pmtId = this._pmtId, avcData = avcTrack.pesData, audioData = audioTrack.pesData, id3Data = id3Track.pesData, parsePAT = this._parsePAT, parsePMT = this._parsePMT.bind(this), parsePES = this._parsePES, parseAVCPES = this._parseAVCPES.bind(this), parseAACPES = this._parseAACPES.bind(this), parseMPEGPES = this._parseMPEGPES.bind(this), parseID3PES = this._parseID3PES.bind(this); var syncOffset = TSDemuxer._syncOffset(data); // don't parse last TS packet if incomplete len -= (len + syncOffset) % 188; // loop through TS packets for (start = syncOffset; start < len; start += 188) { if (data[start] === 0x47) { stt = !!(data[start + 1] & 0x40); // pid is a 13-bit field starting at the last bit of TS[1] pid = ((data[start + 1] & 0x1f) << 8) + data[start + 2]; atf = (data[start + 3] & 0x30) >> 4; // if an adaption field is present, its length is specified by the fifth byte of the TS packet header. if (atf > 1) { offset = start + 5 + data[start + 4]; // continue if there is only adaptation field if (offset === start + 188) { continue; } } else { offset = start + 4; } switch (pid) { case avcId: if (stt) { if (avcData && (pes = parsePES(avcData))) { parseAVCPES(pes, false); } avcData = { data: [], size: 0 }; } if (avcData) { avcData.data.push(data.subarray(offset, start + 188)); avcData.size += start + 188 - offset; } break; case audioId: if (stt) { if (audioData && (pes = parsePES(audioData))) { if (audioTrack.isAAC) { parseAACPES(pes); } else { parseMPEGPES(pes); } } audioData = { data: [], size: 0 }; } if (audioData) { audioData.data.push(data.subarray(offset, start + 188)); audioData.size += start + 188 - offset; } break; case id3Id: if (stt) { if (id3Data && (pes = parsePES(id3Data))) { parseID3PES(pes); } id3Data = { data: [], size: 0 }; } if (id3Data) { id3Data.data.push(data.subarray(offset, start + 188)); id3Data.size += start + 188 - offset; } break; case 0: if (stt) { offset += data[offset] + 1; } pmtId = this._pmtId = parsePAT(data, offset); break; case pmtId: if (stt) { offset += data[offset] + 1; } var parsedPIDs = parsePMT(data, offset, this.typeSupported.mpeg === true || this.typeSupported.mp3 === true, this.sampleAes != null); // only update track id if track PID found while parsing PMT // this is to avoid resetting the PID to -1 in case // track PID transiently disappears from the stream // this could happen in case of transient missing audio samples for example // NOTE this is only the PID of the track as found in TS, // but we are not using this for MP4 track IDs. avcId = parsedPIDs.avc; if (avcId > 0) { avcTrack.pid = avcId; } audioId = parsedPIDs.audio; if (audioId > 0) { audioTrack.pid = audioId; audioTrack.isAAC = parsedPIDs.isAAC; } id3Id = parsedPIDs.id3; if (id3Id > 0) { id3Track.pid = id3Id; } if (unknownPIDs && !pmtParsed) { logger["logger"].log('reparse from beginning'); unknownPIDs = false; // we set it to -188, the += 188 in the for loop will reset start to 0 start = syncOffset - 188; } pmtParsed = this.pmtParsed = true; break; case 17: case 0x1fff: break; default: unknownPIDs = true; break; } } else { this.observer.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].FRAG_PARSING_ERROR, fatal: false, reason: 'TS packet did not start with 0x47' }); } } // try to parse last PES packets if (avcData && (pes = parsePES(avcData))) { parseAVCPES(pes, true); avcTrack.pesData = null; } else { // either avcData null or PES truncated, keep it for next frag parsing avcTrack.pesData = avcData; } if (audioData && (pes = parsePES(audioData))) { if (audioTrack.isAAC) { parseAACPES(pes); } else { parseMPEGPES(pes); } audioTrack.pesData = null; } else { if (audioData && audioData.size) { logger["logger"].log('last AAC PES packet truncated,might overlap between fragments'); } // either audioData null or PES truncated, keep it for next frag parsing audioTrack.pesData = audioData; } if (id3Data && (pes = parsePES(id3Data))) { parseID3PES(pes); id3Track.pesData = null; } else { // either id3Data null or PES truncated, keep it for next frag parsing id3Track.pesData = id3Data; } if (this.sampleAes == null) { this.remuxer.remux(audioTrack, avcTrack, id3Track, this._txtTrack, timeOffset, contiguous, accurateTimeOffset); } else { this.decryptAndRemux(audioTrack, avcTrack, id3Track, this._txtTrack, timeOffset, contiguous, accurateTimeOffset); } }; _proto.decryptAndRemux = function decryptAndRemux(audioTrack, videoTrack, id3Track, textTrack, timeOffset, contiguous, accurateTimeOffset) { if (audioTrack.samples && audioTrack.isAAC) { var localthis = this; this.sampleAes.decryptAacSamples(audioTrack.samples, 0, function () { localthis.decryptAndRemuxAvc(audioTrack, videoTrack, id3Track, textTrack, timeOffset, contiguous, accurateTimeOffset); }); } else { this.decryptAndRemuxAvc(audioTrack, videoTrack, id3Track, textTrack, timeOffset, contiguous, accurateTimeOffset); } }; _proto.decryptAndRemuxAvc = function decryptAndRemuxAvc(audioTrack, videoTrack, id3Track, textTrack, timeOffset, contiguous, accurateTimeOffset) { if (videoTrack.samples) { var localthis = this; this.sampleAes.decryptAvcSamples(videoTrack.samples, 0, 0, function () { localthis.remuxer.remux(audioTrack, videoTrack, id3Track, textTrack, timeOffset, contiguous, accurateTimeOffset); }); } else { this.remuxer.remux(audioTrack, videoTrack, id3Track, textTrack, timeOffset, contiguous, accurateTimeOffset); } }; _proto.destroy = function destroy() { this._initPTS = this._initDTS = undefined; this._duration = 0; }; _proto._parsePAT = function _parsePAT(data, offset) { // skip the PSI header and parse the first PMT entry return (data[offset + 10] & 0x1F) << 8 | data[offset + 11]; // logger.log('PMT PID:' + this._pmtId); }; _proto._trackUnknownPmt = function _trackUnknownPmt(type, logLevel, message) { // Only log unknown and unsupported stream types once per append or stream (by resetting this.pmtUnknownTypes) // For more information on elementary stream types see: // https://en.wikipedia.org/wiki/Program-specific_information#Elementary_stream_types var result = this.pmtUnknownTypes[type] || 0; if (result === 0) { this.pmtUnknownTypes[type] = 0; logLevel.call(logger["logger"], message); } this.pmtUnknownTypes[type]++; return result; }; _proto._parsePMT = function _parsePMT(data, offset, mpegSupported, isSampleAes) { var sectionLength, tableEnd, programInfoLength, pid, result = { audio: -1, avc: -1, id3: -1, isAAC: true }; sectionLength = (data[offset + 1] & 0x0f) << 8 | data[offset + 2]; tableEnd = offset + 3 + sectionLength - 4; // to determine where the table is, we have to figure out how // long the program info descriptors are programInfoLength = (data[offset + 10] & 0x0f) << 8 | data[offset + 11]; // advance the offset to the first entry in the mapping table offset += 12 + programInfoLength; while (offset < tableEnd) { pid = (data[offset + 1] & 0x1F) << 8 | data[offset + 2]; switch (data[offset]) { case 0xcf: // SAMPLE-AES AAC if (!isSampleAes) { this._trackUnknownPmt(data[offset], logger["logger"].warn, 'ADTS AAC with AES-128-CBC frame encryption found in unencrypted stream'); break; } /* falls through */ // ISO/IEC 13818-7 ADTS AAC (MPEG-2 lower bit-rate audio) case 0x0f: // logger.log('AAC PID:' + pid); if (result.audio === -1) { result.audio = pid; } break; // Packetized metadata (ID3) case 0x15: // logger.log('ID3 PID:' + pid); if (result.id3 === -1) { result.id3 = pid; } break; case 0xdb: // SAMPLE-AES AVC if (!isSampleAes) { this._trackUnknownPmt(data[offset], logger["logger"].warn, 'H.264 with AES-128-CBC slice encryption found in unencrypted stream'); break; } /* falls through */ // ITU-T Rec. H.264 and ISO/IEC 14496-10 (lower bit-rate video) case 0x1b: // logger.log('AVC PID:' + pid); if (result.avc === -1) { result.avc = pid; } break; // ISO/IEC 11172-3 (MPEG-1 audio) // or ISO/IEC 13818-3 (MPEG-2 halved sample rate audio) case 0x03: case 0x04: // logger.log('MPEG PID:' + pid); if (!mpegSupported) { this._trackUnknownPmt(data[offset], logger["logger"].warn, 'MPEG audio found, not supported in this browser'); } else if (result.audio === -1) { result.audio = pid; result.isAAC = false; } break; case 0x24: this._trackUnknownPmt(data[offset], logger["logger"].warn, 'Unsupported HEVC stream type found'); break; default: this._trackUnknownPmt(data[offset], logger["logger"].log, 'Unknown stream type:' + data[offset]); break; } // move to the next table entry // skip past the elementary stream descriptors, if present offset += ((data[offset + 3] & 0x0F) << 8 | data[offset + 4]) + 5; } return result; }; _proto._parsePES = function _parsePES(stream) { var i = 0, frag, pesFlags, pesPrefix, pesLen, pesHdrLen, pesData, pesPts, pesDts, payloadStartOffset, data = stream.data; // safety check if (!stream || stream.size === 0) { return null; } // we might need up to 19 bytes to read PES header // if first chunk of data is less than 19 bytes, let's merge it with following ones until we get 19 bytes // usually only one merge is needed (and this is rare ...) while (data[0].length < 19 && data.length > 1) { var newData = new Uint8Array(data[0].length + data[1].length); newData.set(data[0]); newData.set(data[1], data[0].length); data[0] = newData; data.splice(1, 1); } // retrieve PTS/DTS from first fragment frag = data[0]; pesPrefix = (frag[0] << 16) + (frag[1] << 8) + frag[2]; if (pesPrefix === 1) { pesLen = (frag[4] << 8) + frag[5]; // if PES parsed length is not zero and greater than total received length, stop parsing. PES might be truncated // minus 6 : PES header size if (pesLen && pesLen > stream.size - 6) { return null; } pesFlags = frag[7]; if (pesFlags & 0xC0) { /* PES header described here : http://dvd.sourceforge.net/dvdinfo/pes-hdr.html as PTS / DTS is 33 bit we cannot use bitwise operator in JS, as Bitwise operators treat their operands as a sequence of 32 bits */ pesPts = (frag[9] & 0x0E) * 536870912 + // 1 << 29 (frag[10] & 0xFF) * 4194304 + // 1 << 22 (frag[11] & 0xFE) * 16384 + // 1 << 14 (frag[12] & 0xFF) * 128 + // 1 << 7 (frag[13] & 0xFE) / 2; // check if greater than 2^32 -1 if (pesPts > 4294967295) { // decrement 2^33 pesPts -= 8589934592; } if (pesFlags & 0x40) { pesDts = (frag[14] & 0x0E) * 536870912 + // 1 << 29 (frag[15] & 0xFF) * 4194304 + // 1 << 22 (frag[16] & 0xFE) * 16384 + // 1 << 14 (frag[17] & 0xFF) * 128 + // 1 << 7 (frag[18] & 0xFE) / 2; // check if greater than 2^32 -1 if (pesDts > 4294967295) { // decrement 2^33 pesDts -= 8589934592; } if (pesPts - pesDts > 60 * 90000) { logger["logger"].warn(Math.round((pesPts - pesDts) / 90000) + "s delta between PTS and DTS, align them"); pesPts = pesDts; } } else { pesDts = pesPts; } } pesHdrLen = frag[8]; // 9 bytes : 6 bytes for PES header + 3 bytes for PES extension payloadStartOffset = pesHdrLen + 9; if (stream.size <= payloadStartOffset) { return null; } stream.size -= payloadStartOffset; // reassemble PES packet pesData = new Uint8Array(stream.size); for (var j = 0, dataLen = data.length; j < dataLen; j++) { frag = data[j]; var len = frag.byteLength; if (payloadStartOffset) { if (payloadStartOffset > len) { // trim full frag if PES header bigger than frag payloadStartOffset -= len; continue; } else { // trim partial frag if PES header smaller than frag frag = frag.subarray(payloadStartOffset); len -= payloadStartOffset; payloadStartOffset = 0; } } pesData.set(frag, i); i += len; } if (pesLen) { // payload size : remove PES header + PES extension pesLen -= pesHdrLen + 3; } return { data: pesData, pts: pesPts, dts: pesDts, len: pesLen }; } else { return null; } }; _proto.pushAccesUnit = function pushAccesUnit(avcSample, avcTrack) { if (avcSample.units.length && avcSample.frame) { var samples = avcTrack.samples; var nbSamples = samples.length; // if sample does not have PTS/DTS, patch with last sample PTS/DTS if (isNaN(avcSample.pts)) { if (nbSamples) { var lastSample = samples[nbSamples - 1]; avcSample.pts = lastSample.pts; avcSample.dts = lastSample.dts; } else { // dropping samples, no timestamp found avcTrack.dropped++; return; } } // only push AVC sample if starting with a keyframe is not mandatory OR // if keyframe already found in this fragment OR // keyframe found in last fragment (track.sps) AND // samples already appended (we already found a keyframe in this fragment) OR fragment is contiguous if (!this.config.forceKeyFrameOnDiscontinuity || avcSample.key === true || avcTrack.sps && (nbSamples || this.contiguous)) { avcSample.id = nbSamples; samples.push(avcSample); } else { // dropped samples, track it avcTrack.dropped++; } } if (avcSample.debug.length) { logger["logger"].log(avcSample.pts + '/' + avcSample.dts + ':' + avcSample.debug); } }; _proto._parseAVCPES = function _parseAVCPES(pes, last) { var _this = this; // logger.log('parse new PES'); var track = this._avcTrack, units = this._parseAVCNALu(pes.data), debug = false, expGolombDecoder, avcSample = this.avcSample, push, spsfound = false, i, pushAccesUnit = this.pushAccesUnit.bind(this), createAVCSample = function createAVCSample(key, pts, dts, debug) { return { key: key, pts: pts, dts: dts, units: [], debug: debug }; }; // free pes.data to save up some memory pes.data = null; // if new NAL units found and last sample still there, let's push ... // this helps parsing streams with missing AUD (only do this if AUD never found) if (avcSample && units.length && !track.audFound) { pushAccesUnit(avcSample, track); avcSample = this.avcSample = createAVCSample(false, pes.pts, pes.dts, ''); } units.forEach(function (unit) { switch (unit.type) { // NDR case 1: push = true; if (!avcSample) { avcSample = _this.avcSample = createAVCSample(true, pes.pts, pes.dts, ''); } if (debug) { avcSample.debug += 'NDR '; } avcSample.frame = true; var data = unit.data; // only check slice type to detect KF in case SPS found in same packet (any keyframe is preceded by SPS ...) if (spsfound && data.length > 4) { // retrieve slice type by parsing beginning of NAL unit (follow H264 spec, slice_header definition) to detect keyframe embedded in NDR var sliceType = new exp_golomb(data).readSliceType(); // 2 : I slice, 4 : SI slice, 7 : I slice, 9: SI slice // SI slice : A slice that is coded using intra prediction only and using quantisation of the prediction samples. // An SI slice can be coded such that its decoded samples can be constructed identically to an SP slice. // I slice: A slice that is not an SI slice that is decoded using intra prediction only. // if (sliceType === 2 || sliceType === 7) { if (sliceType === 2 || sliceType === 4 || sliceType === 7 || sliceType === 9) { avcSample.key = true; } } break; // IDR case 5: push = true; // handle PES not starting with AUD if (!avcSample) { avcSample = _this.avcSample = createAVCSample(true, pes.pts, pes.dts, ''); } if (debug) { avcSample.debug += 'IDR '; } avcSample.key = true; avcSample.frame = true; break; // SEI case 6: push = true; if (debug && avcSample) { avcSample.debug += 'SEI '; } expGolombDecoder = new exp_golomb(_this.discardEPB(unit.data)); // skip frameType expGolombDecoder.readUByte(); var payloadType = 0; var payloadSize = 0; var endOfCaptions = false; var b = 0; while (!endOfCaptions && expGolombDecoder.bytesAvailable > 1) { payloadType = 0; do { b = expGolombDecoder.readUByte(); payloadType += b; } while (b === 0xFF); // Parse payload size. payloadSize = 0; do { b = expGolombDecoder.readUByte(); payloadSize += b; } while (b === 0xFF); // TODO: there can be more than one payload in an SEI packet... // TODO: need to read type and size in a while loop to get them all if (payloadType === 4 && expGolombDecoder.bytesAvailable !== 0) { endOfCaptions = true; var countryCode = expGolombDecoder.readUByte(); if (countryCode === 181) { var providerCode = expGolombDecoder.readUShort(); if (providerCode === 49) { var userStructure = expGolombDecoder.readUInt(); if (userStructure === 0x47413934) { var userDataType = expGolombDecoder.readUByte(); // Raw CEA-608 bytes wrapped in CEA-708 packet if (userDataType === 3) { var firstByte = expGolombDecoder.readUByte(); var secondByte = expGolombDecoder.readUByte(); var totalCCs = 31 & firstByte; var byteArray = [firstByte, secondByte]; for (i = 0; i < totalCCs; i++) { // 3 bytes per CC byteArray.push(expGolombDecoder.readUByte()); byteArray.push(expGolombDecoder.readUByte()); byteArray.push(expGolombDecoder.readUByte()); } _this._insertSampleInOrder(_this._txtTrack.samples, { type: 3, pts: pes.pts, bytes: byteArray }); } } } } } else if (payloadType === 5 && expGolombDecoder.bytesAvailable !== 0) { endOfCaptions = true; if (payloadSize > 16) { var uuidStrArray = []; for (i = 0; i < 16; i++) { uuidStrArray.push(expGolombDecoder.readUByte().toString(16)); if (i === 3 || i === 5 || i === 7 || i === 9) { uuidStrArray.push('-'); } } var length = payloadSize - 16; var userDataPayloadBytes = new Uint8Array(length); for (i = 0; i < length; i++) { userDataPayloadBytes[i] = expGolombDecoder.readUByte(); } _this._insertSampleInOrder(_this._txtTrack.samples, { pts: pes.pts, payloadType: payloadType, uuid: uuidStrArray.join(''), userDataBytes: userDataPayloadBytes, userData: Object(id3["utf8ArrayToStr"])(userDataPayloadBytes.buffer) }); } } else if (payloadSize < expGolombDecoder.bytesAvailable) { for (i = 0; i < payloadSize; i++) { expGolombDecoder.readUByte(); } } } break; // SPS case 7: push = true; spsfound = true; if (debug && avcSample) { avcSample.debug += 'SPS '; } if (!track.sps) { expGolombDecoder = new exp_golomb(unit.data); var config = expGolombDecoder.readSPS(); track.width = config.width; track.height = config.height; track.pixelRatio = config.pixelRatio; track.sps = [unit.data]; track.duration = _this._duration; var codecarray = unit.data.subarray(1, 4); var codecstring = 'avc1.'; for (i = 0; i < 3; i++) { var h = codecarray[i].toString(16); if (h.length < 2) { h = '0' + h; } codecstring += h; } track.codec = codecstring; } break; // PPS case 8: push = true; if (debug && avcSample) { avcSample.debug += 'PPS '; } if (!track.pps) { track.pps = [unit.data]; } break; // AUD case 9: push = false; track.audFound = true; if (avcSample) { pushAccesUnit(avcSample, track); } avcSample = _this.avcSample = createAVCSample(false, pes.pts, pes.dts, debug ? 'AUD ' : ''); break; // Filler Data case 12: push = false; break; default: push = false; if (avcSample) { avcSample.debug += 'unknown NAL ' + unit.type + ' '; } break; } if (avcSample && push) { var _units = avcSample.units; _units.push(unit); } }); // if last PES packet, push samples if (last && avcSample) { pushAccesUnit(avcSample, track); this.avcSample = null; } }; _proto._insertSampleInOrder = function _insertSampleInOrder(arr, data) { var len = arr.length; if (len > 0) { if (data.pts >= arr[len - 1].pts) { arr.push(data); } else { for (var pos = len - 1; pos >= 0; pos--) { if (data.pts < arr[pos].pts) { arr.splice(pos, 0, data); break; } } } } else { arr.push(data); } }; _proto._getLastNalUnit = function _getLastNalUnit() { var avcSample = this.avcSample, lastUnit; // try to fallback to previous sample if current one is empty if (!avcSample || avcSample.units.length === 0) { var track = this._avcTrack, samples = track.samples; avcSample = samples[samples.length - 1]; } if (avcSample) { var units = avcSample.units; lastUnit = units[units.length - 1]; } return lastUnit; }; _proto._parseAVCNALu = function _parseAVCNALu(array) { var i = 0, len = array.byteLength, value, overflow, track = this._avcTrack, state = track.naluState || 0, lastState = state; var units = [], unit, unitType, lastUnitStart = -1, lastUnitType; // logger.log('PES:' + Hex.hexDump(array)); if (state === -1) { // special use case where we found 3 or 4-byte start codes exactly at the end of previous PES packet lastUnitStart = 0; // NALu type is value read from offset 0 lastUnitType = array[0] & 0x1f; state = 0; i = 1; } while (i < len) { value = array[i++]; // optimization. state 0 and 1 are the predominant case. let's handle them outside of the switch/case if (!state) { state = value ? 0 : 1; continue; } if (state === 1) { state = value ? 0 : 2; continue; } // here we have state either equal to 2 or 3 if (!value) { state = 3; } else if (value === 1) { if (lastUnitStart >= 0) { unit = { data: array.subarray(lastUnitStart, i - state - 1), type: lastUnitType }; // logger.log('pushing NALU, type/size:' + unit.type + '/' + unit.data.byteLength); units.push(unit); } else { // lastUnitStart is undefined => this is the first start code found in this PES packet // first check if start code delimiter is overlapping between 2 PES packets, // ie it started in last packet (lastState not zero) // and ended at the beginning of this PES packet (i <= 4 - lastState) var lastUnit = this._getLastNalUnit(); if (lastUnit) { if (lastState && i <= 4 - lastState) { // start delimiter overlapping between PES packets // strip start delimiter bytes from the end of last NAL unit // check if lastUnit had a state different from zero if (lastUnit.state) { // strip last bytes lastUnit.data = lastUnit.data.subarray(0, lastUnit.data.byteLength - lastState); } } // If NAL units are not starting right at the beginning of the PES packet, push preceding data into previous NAL unit. overflow = i - state - 1; if (overflow > 0) { // logger.log('first NALU found with overflow:' + overflow); var tmp = new Uint8Array(lastUnit.data.byteLength + overflow); tmp.set(lastUnit.data, 0); tmp.set(array.subarray(0, overflow), lastUnit.data.byteLength); lastUnit.data = tmp; } } } // check if we can read unit type if (i < len) { unitType = array[i] & 0x1f; // logger.log('find NALU @ offset:' + i + ',type:' + unitType); lastUnitStart = i; lastUnitType = unitType; state = 0; } else { // not enough byte to read unit type. let's read it on next PES parsing state = -1; } } else { state = 0; } } if (lastUnitStart >= 0 && state >= 0) { unit = { data: array.subarray(lastUnitStart, len), type: lastUnitType, state: state }; units.push(unit); // logger.log('pushing NALU, type/size/state:' + unit.type + '/' + unit.data.byteLength + '/' + state); } // no NALu found if (units.length === 0) { // append pes.data to previous NAL unit var _lastUnit = this._getLastNalUnit(); if (_lastUnit) { var _tmp = new Uint8Array(_lastUnit.data.byteLength + array.byteLength); _tmp.set(_lastUnit.data, 0); _tmp.set(array, _lastUnit.data.byteLength); _lastUnit.data = _tmp; } } track.naluState = state; return units; } /** * remove Emulation Prevention bytes from a RBSP */ ; _proto.discardEPB = function discardEPB(data) { var length = data.byteLength, EPBPositions = [], i = 1, newLength, newData; // Find all `Emulation Prevention Bytes` while (i < length - 2) { if (data[i] === 0 && data[i + 1] === 0 && data[i + 2] === 0x03) { EPBPositions.push(i + 2); i += 2; } else { i++; } } // If no Emulation Prevention Bytes were found just return the original // array if (EPBPositions.length === 0) { return data; } // Create a new array to hold the NAL unit data newLength = length - EPBPositions.length; newData = new Uint8Array(newLength); var sourceIndex = 0; for (i = 0; i < newLength; sourceIndex++, i++) { if (sourceIndex === EPBPositions[0]) { // Skip this byte sourceIndex++; // Remove this position index EPBPositions.shift(); } newData[i] = data[sourceIndex]; } return newData; }; _proto._parseAACPES = function _parseAACPES(pes) { var track = this._audioTrack, data = pes.data, pts = pes.pts, startOffset = 0, aacOverFlow = this.aacOverFlow, aacLastPTS = this.aacLastPTS, frameDuration, frameIndex, offset, stamp, len; if (aacOverFlow) { var tmp = new Uint8Array(aacOverFlow.byteLength + data.byteLength); tmp.set(aacOverFlow, 0); tmp.set(data, aacOverFlow.byteLength); // logger.log(`AAC: append overflowing ${aacOverFlow.byteLength} bytes to beginning of new PES`); data = tmp; } // look for ADTS header (0xFFFx) for (offset = startOffset, len = data.length; offset < len - 1; offset++) { if (isHeader(data, offset)) { break; } } // if ADTS header does not start straight from the beginning of the PES payload, raise an error if (offset) { var reason, fatal; if (offset < len - 1) { reason = "AAC PES did not start with ADTS header,offset:" + offset; fatal = false; } else { reason = 'no ADTS header found in AAC PES'; fatal = true; } logger["logger"].warn("parsing error:" + reason); this.observer.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].FRAG_PARSING_ERROR, fatal: fatal, reason: reason }); if (fatal) { return; } } initTrackConfig(track, this.observer, data, offset, this.audioCodec); frameIndex = 0; frameDuration = getFrameDuration(track.samplerate); // if last AAC frame is overflowing, we should ensure timestamps are contiguous: // first sample PTS should be equal to last sample PTS + frameDuration if (aacOverFlow && aacLastPTS) { var newPTS = aacLastPTS + frameDuration; if (Math.abs(newPTS - pts) > 1) { logger["logger"].log("AAC: align PTS for overlapping frames by " + Math.round((newPTS - pts) / 90)); pts = newPTS; } } // scan for aac samples while (offset < len) { if (isHeader(data, offset)) { if (offset + 5 < len) { var frame = appendFrame(track, data, offset, pts, frameIndex); if (frame) { offset += frame.length; stamp = frame.sample.pts; frameIndex++; continue; } } // We are at an ADTS header, but do not have enough data for a frame // Remaining data will be added to aacOverFlow break; } else { // nothing found, keep looking offset++; } } if (offset < len) { aacOverFlow = data.subarray(offset, len); // logger.log(`AAC: overflow detected:${len-offset}`); } else { aacOverFlow = null; } this.aacOverFlow = aacOverFlow; this.aacLastPTS = stamp; }; _proto._parseMPEGPES = function _parseMPEGPES(pes) { var data = pes.data; var length = data.length; var frameIndex = 0; var offset = 0; var pts = pes.pts; while (offset < length) { if (mpegaudio.isHeader(data, offset)) { var frame = mpegaudio.appendFrame(this._audioTrack, data, offset, pts, frameIndex); if (frame) { offset += frame.length; frameIndex++; } else { // logger.log('Unable to parse Mpeg audio frame'); break; } } else { // nothing found, keep looking offset++; } } }; _proto._parseID3PES = function _parseID3PES(pes) { this._id3Track.samples.push(pes); }; return TSDemuxer; }(); /* harmony default export */ var tsdemuxer = (tsdemuxer_TSDemuxer); // CONCATENATED MODULE: ./src/demux/mp3demuxer.js /** * MP3 demuxer */ var mp3demuxer_MP3Demuxer = /*#__PURE__*/function () { function MP3Demuxer(observer, remuxer, config) { this.observer = observer; this.config = config; this.remuxer = remuxer; } var _proto = MP3Demuxer.prototype; _proto.resetInitSegment = function resetInitSegment(initSegment, audioCodec, videoCodec, duration) { this._audioTrack = { container: 'audio/mpeg', type: 'audio', id: -1, sequenceNumber: 0, isAAC: false, samples: [], len: 0, manifestCodec: audioCodec, duration: duration, inputTimeScale: 90000 }; }; _proto.resetTimeStamp = function resetTimeStamp() {}; MP3Demuxer.probe = function probe(data) { // check if data contains ID3 timestamp and MPEG sync word var offset, length; var id3Data = id3["default"].getID3Data(data, 0); if (id3Data && id3["default"].getTimeStamp(id3Data) !== undefined) { // Look for MPEG header | 1111 1111 | 111X XYZX | where X can be either 0 or 1 and Y or Z should be 1 // Layer bits (position 14 and 15) in header should be always different from 0 (Layer I or Layer II or Layer III) // More info http://www.mp3-tech.org/programmer/frame_header.html for (offset = id3Data.length, length = Math.min(data.length - 1, offset + 100); offset < length; offset++) { if (mpegaudio.probe(data, offset)) { logger["logger"].log('MPEG Audio sync word found !'); return true; } } } return false; } // feed incoming data to the front of the parsing pipeline ; _proto.append = function append(data, timeOffset, contiguous, accurateTimeOffset) { var id3Data = id3["default"].getID3Data(data, 0); var timestamp = id3["default"].getTimeStamp(id3Data); var pts = timestamp !== undefined ? 90 * timestamp : timeOffset * 90000; var offset = id3Data.length; var length = data.length; var frameIndex = 0, stamp = 0; var track = this._audioTrack; var id3Samples = [{ pts: pts, dts: pts, data: id3Data }]; while (offset < length) { if (mpegaudio.isHeader(data, offset)) { var frame = mpegaudio.appendFrame(track, data, offset, pts, frameIndex); if (frame) { offset += frame.length; stamp = frame.sample.pts; frameIndex++; } else { // logger.log('Unable to parse Mpeg audio frame'); break; } } else if (id3["default"].isHeader(data, offset)) { id3Data = id3["default"].getID3Data(data, offset); id3Samples.push({ pts: stamp, dts: stamp, data: id3Data }); offset += id3Data.length; } else { // nothing found, keep looking offset++; } } this.remuxer.remux(track, { samples: [] }, { samples: id3Samples, inputTimeScale: 90000 }, { samples: [] }, timeOffset, contiguous, accurateTimeOffset); }; _proto.destroy = function destroy() {}; return MP3Demuxer; }(); /* harmony default export */ var mp3demuxer = (mp3demuxer_MP3Demuxer); // CONCATENATED MODULE: ./src/remux/aac-helper.js /** * AAC helper */ var AAC = /*#__PURE__*/function () { function AAC() {} AAC.getSilentFrame = function getSilentFrame(codec, channelCount) { switch (codec) { case 'mp4a.40.2': if (channelCount === 1) { return new Uint8Array([0x00, 0xc8, 0x00, 0x80, 0x23, 0x80]); } else if (channelCount === 2) { return new Uint8Array([0x21, 0x00, 0x49, 0x90, 0x02, 0x19, 0x00, 0x23, 0x80]); } else if (channelCount === 3) { return new Uint8Array([0x00, 0xc8, 0x00, 0x80, 0x20, 0x84, 0x01, 0x26, 0x40, 0x08, 0x64, 0x00, 0x8e]); } else if (channelCount === 4) { return new Uint8Array([0x00, 0xc8, 0x00, 0x80, 0x20, 0x84, 0x01, 0x26, 0x40, 0x08, 0x64, 0x00, 0x80, 0x2c, 0x80, 0x08, 0x02, 0x38]); } else if (channelCount === 5) { return new Uint8Array([0x00, 0xc8, 0x00, 0x80, 0x20, 0x84, 0x01, 0x26, 0x40, 0x08, 0x64, 0x00, 0x82, 0x30, 0x04, 0x99, 0x00, 0x21, 0x90, 0x02, 0x38]); } else if (channelCount === 6) { return new Uint8Array([0x00, 0xc8, 0x00, 0x80, 0x20, 0x84, 0x01, 0x26, 0x40, 0x08, 0x64, 0x00, 0x82, 0x30, 0x04, 0x99, 0x00, 0x21, 0x90, 0x02, 0x00, 0xb2, 0x00, 0x20, 0x08, 0xe0]); } break; // handle HE-AAC below (mp4a.40.5 / mp4a.40.29) default: if (channelCount === 1) { // ffmpeg -y -f lavfi -i "aevalsrc=0:d=0.05" -c:a libfdk_aac -profile:a aac_he -b:a 4k output.aac && hexdump -v -e '16/1 "0x%x," "\n"' -v output.aac return new Uint8Array([0x1, 0x40, 0x22, 0x80, 0xa3, 0x4e, 0xe6, 0x80, 0xba, 0x8, 0x0, 0x0, 0x0, 0x1c, 0x6, 0xf1, 0xc1, 0xa, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5e]); } else if (channelCount === 2) { // ffmpeg -y -f lavfi -i "aevalsrc=0|0:d=0.05" -c:a libfdk_aac -profile:a aac_he_v2 -b:a 4k output.aac && hexdump -v -e '16/1 "0x%x," "\n"' -v output.aac return new Uint8Array([0x1, 0x40, 0x22, 0x80, 0xa3, 0x5e, 0xe6, 0x80, 0xba, 0x8, 0x0, 0x0, 0x0, 0x0, 0x95, 0x0, 0x6, 0xf1, 0xa1, 0xa, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5e]); } else if (channelCount === 3) { // ffmpeg -y -f lavfi -i "aevalsrc=0|0|0:d=0.05" -c:a libfdk_aac -profile:a aac_he_v2 -b:a 4k output.aac && hexdump -v -e '16/1 "0x%x," "\n"' -v output.aac return new Uint8Array([0x1, 0x40, 0x22, 0x80, 0xa3, 0x5e, 0xe6, 0x80, 0xba, 0x8, 0x0, 0x0, 0x0, 0x0, 0x95, 0x0, 0x6, 0xf1, 0xa1, 0xa, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5a, 0x5e]); } break; } return null; }; return AAC; }(); /* harmony default export */ var aac_helper = (AAC); // CONCATENATED MODULE: ./src/remux/mp4-generator.js /** * Generate MP4 Box */ var UINT32_MAX = Math.pow(2, 32) - 1; var MP4 = /*#__PURE__*/function () { function MP4() {} MP4.init = function init() { MP4.types = { avc1: [], // codingname avcC: [], btrt: [], dinf: [], dref: [], esds: [], ftyp: [], hdlr: [], mdat: [], mdhd: [], mdia: [], mfhd: [], minf: [], moof: [], moov: [], mp4a: [], '.mp3': [], mvex: [], mvhd: [], pasp: [], sdtp: [], stbl: [], stco: [], stsc: [], stsd: [], stsz: [], stts: [], tfdt: [], tfhd: [], traf: [], trak: [], trun: [], trex: [], tkhd: [], vmhd: [], smhd: [] }; var i; for (i in MP4.types) { if (MP4.types.hasOwnProperty(i)) { MP4.types[i] = [i.charCodeAt(0), i.charCodeAt(1), i.charCodeAt(2), i.charCodeAt(3)]; } } var videoHdlr = new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x00, // pre_defined 0x76, 0x69, 0x64, 0x65, // handler_type: 'vide' 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, // reserved 0x56, 0x69, 0x64, 0x65, 0x6f, 0x48, 0x61, 0x6e, 0x64, 0x6c, 0x65, 0x72, 0x00 // name: 'VideoHandler' ]); var audioHdlr = new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x00, // pre_defined 0x73, 0x6f, 0x75, 0x6e, // handler_type: 'soun' 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, // reserved 0x53, 0x6f, 0x75, 0x6e, 0x64, 0x48, 0x61, 0x6e, 0x64, 0x6c, 0x65, 0x72, 0x00 // name: 'SoundHandler' ]); MP4.HDLR_TYPES = { 'video': videoHdlr, 'audio': audioHdlr }; var dref = new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x01, // entry_count 0x00, 0x00, 0x00, 0x0c, // entry_size 0x75, 0x72, 0x6c, 0x20, // 'url' type 0x00, // version 0 0x00, 0x00, 0x01 // entry_flags ]); var stco = new Uint8Array([0x00, // version 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x00 // entry_count ]); MP4.STTS = MP4.STSC = MP4.STCO = stco; MP4.STSZ = new Uint8Array([0x00, // version 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x00, // sample_size 0x00, 0x00, 0x00, 0x00 // sample_count ]); MP4.VMHD = new Uint8Array([0x00, // version 0x00, 0x00, 0x01, // flags 0x00, 0x00, // graphicsmode 0x00, 0x00, 0x00, 0x00, 0x00, 0x00 // opcolor ]); MP4.SMHD = new Uint8Array([0x00, // version 0x00, 0x00, 0x00, // flags 0x00, 0x00, // balance 0x00, 0x00 // reserved ]); MP4.STSD = new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x01]); // entry_count var majorBrand = new Uint8Array([105, 115, 111, 109]); // isom var avc1Brand = new Uint8Array([97, 118, 99, 49]); // avc1 var minorVersion = new Uint8Array([0, 0, 0, 1]); MP4.FTYP = MP4.box(MP4.types.ftyp, majorBrand, minorVersion, majorBrand, avc1Brand); MP4.DINF = MP4.box(MP4.types.dinf, MP4.box(MP4.types.dref, dref)); }; MP4.box = function box(type) { var payload = Array.prototype.slice.call(arguments, 1), size = 8, i = payload.length, len = i, result; // calculate the total size we need to allocate while (i--) { size += payload[i].byteLength; } result = new Uint8Array(size); result[0] = size >> 24 & 0xff; result[1] = size >> 16 & 0xff; result[2] = size >> 8 & 0xff; result[3] = size & 0xff; result.set(type, 4); // copy the payload into the result for (i = 0, size = 8; i < len; i++) { // copy payload[i] array @ offset size result.set(payload[i], size); size += payload[i].byteLength; } return result; }; MP4.hdlr = function hdlr(type) { return MP4.box(MP4.types.hdlr, MP4.HDLR_TYPES[type]); }; MP4.mdat = function mdat(data) { return MP4.box(MP4.types.mdat, data); }; MP4.mdhd = function mdhd(timescale, duration) { duration *= timescale; var upperWordDuration = Math.floor(duration / (UINT32_MAX + 1)); var lowerWordDuration = Math.floor(duration % (UINT32_MAX + 1)); return MP4.box(MP4.types.mdhd, new Uint8Array([0x01, // version 1 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, // creation_time 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x03, // modification_time timescale >> 24 & 0xFF, timescale >> 16 & 0xFF, timescale >> 8 & 0xFF, timescale & 0xFF, // timescale upperWordDuration >> 24, upperWordDuration >> 16 & 0xFF, upperWordDuration >> 8 & 0xFF, upperWordDuration & 0xFF, lowerWordDuration >> 24, lowerWordDuration >> 16 & 0xFF, lowerWordDuration >> 8 & 0xFF, lowerWordDuration & 0xFF, 0x55, 0xc4, // 'und' language (undetermined) 0x00, 0x00])); }; MP4.mdia = function mdia(track) { return MP4.box(MP4.types.mdia, MP4.mdhd(track.timescale, track.duration), MP4.hdlr(track.type), MP4.minf(track)); }; MP4.mfhd = function mfhd(sequenceNumber) { return MP4.box(MP4.types.mfhd, new Uint8Array([0x00, 0x00, 0x00, 0x00, // flags sequenceNumber >> 24, sequenceNumber >> 16 & 0xFF, sequenceNumber >> 8 & 0xFF, sequenceNumber & 0xFF // sequence_number ])); }; MP4.minf = function minf(track) { if (track.type === 'audio') { return MP4.box(MP4.types.minf, MP4.box(MP4.types.smhd, MP4.SMHD), MP4.DINF, MP4.stbl(track)); } else { return MP4.box(MP4.types.minf, MP4.box(MP4.types.vmhd, MP4.VMHD), MP4.DINF, MP4.stbl(track)); } }; MP4.moof = function moof(sn, baseMediaDecodeTime, track) { return MP4.box(MP4.types.moof, MP4.mfhd(sn), MP4.traf(track, baseMediaDecodeTime)); } /** * @param tracks... (optional) {array} the tracks associated with this movie */ ; MP4.moov = function moov(tracks) { var i = tracks.length, boxes = []; while (i--) { boxes[i] = MP4.trak(tracks[i]); } return MP4.box.apply(null, [MP4.types.moov, MP4.mvhd(tracks[0].timescale, tracks[0].duration)].concat(boxes).concat(MP4.mvex(tracks))); }; MP4.mvex = function mvex(tracks) { var i = tracks.length, boxes = []; while (i--) { boxes[i] = MP4.trex(tracks[i]); } return MP4.box.apply(null, [MP4.types.mvex].concat(boxes)); }; MP4.mvhd = function mvhd(timescale, duration) { duration *= timescale; var upperWordDuration = Math.floor(duration / (UINT32_MAX + 1)); var lowerWordDuration = Math.floor(duration % (UINT32_MAX + 1)); var bytes = new Uint8Array([0x01, // version 1 0x00, 0x00, 0x00, // flags 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, // creation_time 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x03, // modification_time timescale >> 24 & 0xFF, timescale >> 16 & 0xFF, timescale >> 8 & 0xFF, timescale & 0xFF, // timescale upperWordDuration >> 24, upperWordDuration >> 16 & 0xFF, upperWordDuration >> 8 & 0xFF, upperWordDuration & 0xFF, lowerWordDuration >> 24, lowerWordDuration >> 16 & 0xFF, lowerWordDuration >> 8 & 0xFF, lowerWordDuration & 0xFF, 0x00, 0x01, 0x00, 0x00, // 1.0 rate 0x01, 0x00, // 1.0 volume 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x40, 0x00, 0x00, 0x00, // transformation: unity matrix 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, // pre_defined 0xff, 0xff, 0xff, 0xff // next_track_ID ]); return MP4.box(MP4.types.mvhd, bytes); }; MP4.sdtp = function sdtp(track) { var samples = track.samples || [], bytes = new Uint8Array(4 + samples.length), flags, i; // leave the full box header (4 bytes) all zero // write the sample table for (i = 0; i < samples.length; i++) { flags = samples[i].flags; bytes[i + 4] = flags.dependsOn << 4 | flags.isDependedOn << 2 | flags.hasRedundancy; } return MP4.box(MP4.types.sdtp, bytes); }; MP4.stbl = function stbl(track) { return MP4.box(MP4.types.stbl, MP4.stsd(track), MP4.box(MP4.types.stts, MP4.STTS), MP4.box(MP4.types.stsc, MP4.STSC), MP4.box(MP4.types.stsz, MP4.STSZ), MP4.box(MP4.types.stco, MP4.STCO)); }; MP4.avc1 = function avc1(track) { var sps = [], pps = [], i, data, len; // assemble the SPSs for (i = 0; i < track.sps.length; i++) { data = track.sps[i]; len = data.byteLength; sps.push(len >>> 8 & 0xFF); sps.push(len & 0xFF); // SPS sps = sps.concat(Array.prototype.slice.call(data)); } // assemble the PPSs for (i = 0; i < track.pps.length; i++) { data = track.pps[i]; len = data.byteLength; pps.push(len >>> 8 & 0xFF); pps.push(len & 0xFF); pps = pps.concat(Array.prototype.slice.call(data)); } var avcc = MP4.box(MP4.types.avcC, new Uint8Array([0x01, // version sps[3], // profile sps[4], // profile compat sps[5], // level 0xfc | 3, // lengthSizeMinusOne, hard-coded to 4 bytes 0xE0 | track.sps.length // 3bit reserved (111) + numOfSequenceParameterSets ].concat(sps).concat([track.pps.length // numOfPictureParameterSets ]).concat(pps))), // "PPS" width = track.width, height = track.height, hSpacing = track.pixelRatio[0], vSpacing = track.pixelRatio[1]; return MP4.box(MP4.types.avc1, new Uint8Array([0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, // reserved 0x00, 0x01, // data_reference_index 0x00, 0x00, // pre_defined 0x00, 0x00, // reserved 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, // pre_defined width >> 8 & 0xFF, width & 0xff, // width height >> 8 & 0xFF, height & 0xff, // height 0x00, 0x48, 0x00, 0x00, // horizresolution 0x00, 0x48, 0x00, 0x00, // vertresolution 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x01, // frame_count 0x12, 0x64, 0x61, 0x69, 0x6C, // dailymotion/hls.js 0x79, 0x6D, 0x6F, 0x74, 0x69, 0x6F, 0x6E, 0x2F, 0x68, 0x6C, 0x73, 0x2E, 0x6A, 0x73, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, // compressorname 0x00, 0x18, // depth = 24 0x11, 0x11]), // pre_defined = -1 avcc, MP4.box(MP4.types.btrt, new Uint8Array([0x00, 0x1c, 0x9c, 0x80, // bufferSizeDB 0x00, 0x2d, 0xc6, 0xc0, // maxBitrate 0x00, 0x2d, 0xc6, 0xc0])), // avgBitrate MP4.box(MP4.types.pasp, new Uint8Array([hSpacing >> 24, // hSpacing hSpacing >> 16 & 0xFF, hSpacing >> 8 & 0xFF, hSpacing & 0xFF, vSpacing >> 24, // vSpacing vSpacing >> 16 & 0xFF, vSpacing >> 8 & 0xFF, vSpacing & 0xFF]))); }; MP4.esds = function esds(track) { var configlen = track.config.length; return new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags 0x03, // descriptor_type 0x17 + configlen, // length 0x00, 0x01, // es_id 0x00, // stream_priority 0x04, // descriptor_type 0x0f + configlen, // length 0x40, // codec : mpeg4_audio 0x15, // stream_type 0x00, 0x00, 0x00, // buffer_size 0x00, 0x00, 0x00, 0x00, // maxBitrate 0x00, 0x00, 0x00, 0x00, // avgBitrate 0x05 // descriptor_type ].concat([configlen]).concat(track.config).concat([0x06, 0x01, 0x02])); // GASpecificConfig)); // length + audio config descriptor }; MP4.mp4a = function mp4a(track) { var samplerate = track.samplerate; return MP4.box(MP4.types.mp4a, new Uint8Array([0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, // reserved 0x00, 0x01, // data_reference_index 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, // reserved 0x00, track.channelCount, // channelcount 0x00, 0x10, // sampleSize:16bits 0x00, 0x00, 0x00, 0x00, // reserved2 samplerate >> 8 & 0xFF, samplerate & 0xff, // 0x00, 0x00]), MP4.box(MP4.types.esds, MP4.esds(track))); }; MP4.mp3 = function mp3(track) { var samplerate = track.samplerate; return MP4.box(MP4.types['.mp3'], new Uint8Array([0x00, 0x00, 0x00, // reserved 0x00, 0x00, 0x00, // reserved 0x00, 0x01, // data_reference_index 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, // reserved 0x00, track.channelCount, // channelcount 0x00, 0x10, // sampleSize:16bits 0x00, 0x00, 0x00, 0x00, // reserved2 samplerate >> 8 & 0xFF, samplerate & 0xff, // 0x00, 0x00])); }; MP4.stsd = function stsd(track) { if (track.type === 'audio') { if (!track.isAAC && track.codec === 'mp3') { return MP4.box(MP4.types.stsd, MP4.STSD, MP4.mp3(track)); } return MP4.box(MP4.types.stsd, MP4.STSD, MP4.mp4a(track)); } else { return MP4.box(MP4.types.stsd, MP4.STSD, MP4.avc1(track)); } }; MP4.tkhd = function tkhd(track) { var id = track.id, duration = track.duration * track.timescale, width = track.width, height = track.height, upperWordDuration = Math.floor(duration / (UINT32_MAX + 1)), lowerWordDuration = Math.floor(duration % (UINT32_MAX + 1)); return MP4.box(MP4.types.tkhd, new Uint8Array([0x01, // version 1 0x00, 0x00, 0x07, // flags 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, // creation_time 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x03, // modification_time id >> 24 & 0xFF, id >> 16 & 0xFF, id >> 8 & 0xFF, id & 0xFF, // track_ID 0x00, 0x00, 0x00, 0x00, // reserved upperWordDuration >> 24, upperWordDuration >> 16 & 0xFF, upperWordDuration >> 8 & 0xFF, upperWordDuration & 0xFF, lowerWordDuration >> 24, lowerWordDuration >> 16 & 0xFF, lowerWordDuration >> 8 & 0xFF, lowerWordDuration & 0xFF, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, // reserved 0x00, 0x00, // layer 0x00, 0x00, // alternate_group 0x00, 0x00, // non-audio track volume 0x00, 0x00, // reserved 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x40, 0x00, 0x00, 0x00, // transformation: unity matrix width >> 8 & 0xFF, width & 0xFF, 0x00, 0x00, // width height >> 8 & 0xFF, height & 0xFF, 0x00, 0x00 // height ])); }; MP4.traf = function traf(track, baseMediaDecodeTime) { var sampleDependencyTable = MP4.sdtp(track), id = track.id, upperWordBaseMediaDecodeTime = Math.floor(baseMediaDecodeTime / (UINT32_MAX + 1)), lowerWordBaseMediaDecodeTime = Math.floor(baseMediaDecodeTime % (UINT32_MAX + 1)); return MP4.box(MP4.types.traf, MP4.box(MP4.types.tfhd, new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags id >> 24, id >> 16 & 0XFF, id >> 8 & 0XFF, id & 0xFF // track_ID ])), MP4.box(MP4.types.tfdt, new Uint8Array([0x01, // version 1 0x00, 0x00, 0x00, // flags upperWordBaseMediaDecodeTime >> 24, upperWordBaseMediaDecodeTime >> 16 & 0XFF, upperWordBaseMediaDecodeTime >> 8 & 0XFF, upperWordBaseMediaDecodeTime & 0xFF, lowerWordBaseMediaDecodeTime >> 24, lowerWordBaseMediaDecodeTime >> 16 & 0XFF, lowerWordBaseMediaDecodeTime >> 8 & 0XFF, lowerWordBaseMediaDecodeTime & 0xFF])), MP4.trun(track, sampleDependencyTable.length + 16 + // tfhd 20 + // tfdt 8 + // traf header 16 + // mfhd 8 + // moof header 8), // mdat header sampleDependencyTable); } /** * Generate a track box. * @param track {object} a track definition * @return {Uint8Array} the track box */ ; MP4.trak = function trak(track) { track.duration = track.duration || 0xffffffff; return MP4.box(MP4.types.trak, MP4.tkhd(track), MP4.mdia(track)); }; MP4.trex = function trex(track) { var id = track.id; return MP4.box(MP4.types.trex, new Uint8Array([0x00, // version 0 0x00, 0x00, 0x00, // flags id >> 24, id >> 16 & 0XFF, id >> 8 & 0XFF, id & 0xFF, // track_ID 0x00, 0x00, 0x00, 0x01, // default_sample_description_index 0x00, 0x00, 0x00, 0x00, // default_sample_duration 0x00, 0x00, 0x00, 0x00, // default_sample_size 0x00, 0x01, 0x00, 0x01 // default_sample_flags ])); }; MP4.trun = function trun(track, offset) { var samples = track.samples || [], len = samples.length, arraylen = 12 + 16 * len, array = new Uint8Array(arraylen), i, sample, duration, size, flags, cts; offset += 8 + arraylen; array.set([0x00, // version 0 0x00, 0x0f, 0x01, // flags len >>> 24 & 0xFF, len >>> 16 & 0xFF, len >>> 8 & 0xFF, len & 0xFF, // sample_count offset >>> 24 & 0xFF, offset >>> 16 & 0xFF, offset >>> 8 & 0xFF, offset & 0xFF // data_offset ], 0); for (i = 0; i < len; i++) { sample = samples[i]; duration = sample.duration; size = sample.size; flags = sample.flags; cts = sample.cts; array.set([duration >>> 24 & 0xFF, duration >>> 16 & 0xFF, duration >>> 8 & 0xFF, duration & 0xFF, // sample_duration size >>> 24 & 0xFF, size >>> 16 & 0xFF, size >>> 8 & 0xFF, size & 0xFF, // sample_size flags.isLeading << 2 | flags.dependsOn, flags.isDependedOn << 6 | flags.hasRedundancy << 4 | flags.paddingValue << 1 | flags.isNonSync, flags.degradPrio & 0xF0 << 8, flags.degradPrio & 0x0F, // sample_flags cts >>> 24 & 0xFF, cts >>> 16 & 0xFF, cts >>> 8 & 0xFF, cts & 0xFF // sample_composition_time_offset ], 12 + 16 * i); } return MP4.box(MP4.types.trun, array); }; MP4.initSegment = function initSegment(tracks) { if (!MP4.types) { MP4.init(); } var movie = MP4.moov(tracks), result; result = new Uint8Array(MP4.FTYP.byteLength + movie.byteLength); result.set(MP4.FTYP); result.set(movie, MP4.FTYP.byteLength); return result; }; return MP4; }(); /* harmony default export */ var mp4_generator = (MP4); // CONCATENATED MODULE: ./src/utils/timescale-conversion.ts var MPEG_TS_CLOCK_FREQ_HZ = 90000; function toTimescaleFromScale(value, destScale, srcScale, round) { if (srcScale === void 0) { srcScale = 1; } if (round === void 0) { round = false; } return toTimescaleFromBase(value, destScale, 1 / srcScale); } function toTimescaleFromBase(value, destScale, srcBase, round) { if (srcBase === void 0) { srcBase = 1; } if (round === void 0) { round = false; } var result = value * destScale * srcBase; // equivalent to `(value * scale) / (1 / base)` return round ? Math.round(result) : result; } function toMsFromMpegTsClock(value, round) { if (round === void 0) { round = false; } return toTimescaleFromBase(value, 1000, 1 / MPEG_TS_CLOCK_FREQ_HZ, round); } function toMpegTsClockFromTimescale(value, srcScale) { if (srcScale === void 0) { srcScale = 1; } return toTimescaleFromBase(value, MPEG_TS_CLOCK_FREQ_HZ, 1 / srcScale); } // CONCATENATED MODULE: ./src/remux/mp4-remuxer.js /** * fMP4 remuxer */ var MAX_SILENT_FRAME_DURATION_90KHZ = toMpegTsClockFromTimescale(10); var PTS_DTS_SHIFT_TOLERANCE_90KHZ = toMpegTsClockFromTimescale(0.2); var mp4_remuxer_MP4Remuxer = /*#__PURE__*/function () { function MP4Remuxer(observer, config, typeSupported, vendor) { this.observer = observer; this.config = config; this.typeSupported = typeSupported; var userAgent = navigator.userAgent; this.isSafari = vendor && vendor.indexOf('Apple') > -1 && userAgent && !userAgent.match('CriOS'); this.ISGenerated = false; } var _proto = MP4Remuxer.prototype; _proto.destroy = function destroy() {}; _proto.resetTimeStamp = function resetTimeStamp(defaultTimeStamp) { this._initPTS = this._initDTS = defaultTimeStamp; }; _proto.resetInitSegment = function resetInitSegment() { this.ISGenerated = false; }; _proto.remux = function remux(audioTrack, videoTrack, id3Track, textTrack, timeOffset, contiguous, accurateTimeOffset) { // generate Init Segment if needed if (!this.ISGenerated) { this.generateIS(audioTrack, videoTrack, timeOffset); } if (this.ISGenerated) { var nbAudioSamples = audioTrack.samples.length; var nbVideoSamples = videoTrack.samples.length; var audioTimeOffset = timeOffset; var videoTimeOffset = timeOffset; if (nbAudioSamples && nbVideoSamples) { // timeOffset is expected to be the offset of the first timestamp of this fragment (first DTS) // if first audio DTS is not aligned with first video DTS then we need to take that into account // when providing timeOffset to remuxAudio / remuxVideo. if we don't do that, there might be a permanent / small // drift between audio and video streams // Use pts at timeOffset 0 so that VOD streams begin at 0 var tsDelta = timeOffset > 0 ? audioTrack.samples[0].dts - videoTrack.samples[0].dts : audioTrack.samples[0].pts - videoTrack.samples[0].pts; var audiovideoTimestampDelta = tsDelta / videoTrack.inputTimeScale; audioTimeOffset += Math.max(0, audiovideoTimestampDelta); videoTimeOffset += Math.max(0, -audiovideoTimestampDelta); } // Purposefully remuxing audio before video, so that remuxVideo can use nextAudioPts, which is // calculated in remuxAudio. // logger.log('nb AAC samples:' + audioTrack.samples.length); if (nbAudioSamples) { // if initSegment was generated without video samples, regenerate it again if (!audioTrack.timescale) { logger["logger"].warn('regenerate InitSegment as audio detected'); this.generateIS(audioTrack, videoTrack, timeOffset); } var audioData = this.remuxAudio(audioTrack, audioTimeOffset, contiguous, accurateTimeOffset); // logger.log('nb AVC samples:' + videoTrack.samples.length); if (nbVideoSamples) { var audioTrackLength; if (audioData) { audioTrackLength = audioData.endPTS - audioData.startPTS; } // if initSegment was generated without video samples, regenerate it again if (!videoTrack.timescale) { logger["logger"].warn('regenerate InitSegment as video detected'); this.generateIS(audioTrack, videoTrack, timeOffset); } this.remuxVideo(videoTrack, videoTimeOffset, contiguous, audioTrackLength, accurateTimeOffset); } } else { // logger.log('nb AVC samples:' + videoTrack.samples.length); if (nbVideoSamples) { var videoData = this.remuxVideo(videoTrack, videoTimeOffset, contiguous, 0, accurateTimeOffset); if (videoData && audioTrack.codec) { this.remuxEmptyAudio(audioTrack, audioTimeOffset, contiguous, videoData); } } } } // logger.log('nb ID3 samples:' + audioTrack.samples.length); if (id3Track.samples.length) { this.remuxID3(id3Track, timeOffset); } // logger.log('nb ID3 samples:' + audioTrack.samples.length); if (textTrack.samples.length) { this.remuxText(textTrack, timeOffset); } // notify end of parsing this.observer.trigger(events["default"].FRAG_PARSED); }; _proto.generateIS = function generateIS(audioTrack, videoTrack, timeOffset) { var observer = this.observer, audioSamples = audioTrack.samples, videoSamples = videoTrack.samples, typeSupported = this.typeSupported, container = 'audio/mp4', tracks = {}, data = { tracks: tracks }, computePTSDTS = this._initPTS === undefined, initPTS, initDTS; if (computePTSDTS) { initPTS = initDTS = Infinity; } if (audioTrack.config && audioSamples.length) { // let's use audio sampling rate as MP4 time scale. // rationale is that there is a integer nb of audio frames per audio sample (1024 for AAC) // using audio sampling rate here helps having an integer MP4 frame duration // this avoids potential rounding issue and AV sync issue audioTrack.timescale = audioTrack.samplerate; logger["logger"].log("audio sampling rate : " + audioTrack.samplerate); if (!audioTrack.isAAC) { if (typeSupported.mpeg) { // Chrome and Safari container = 'audio/mpeg'; audioTrack.codec = ''; } else if (typeSupported.mp3) { // Firefox audioTrack.codec = 'mp3'; } } tracks.audio = { container: container, codec: audioTrack.codec, initSegment: !audioTrack.isAAC && typeSupported.mpeg ? new Uint8Array() : mp4_generator.initSegment([audioTrack]), metadata: { channelCount: audioTrack.channelCount } }; if (computePTSDTS) { // remember first PTS of this demuxing context. for audio, PTS = DTS initPTS = initDTS = audioSamples[0].pts - Math.round(audioTrack.inputTimeScale * timeOffset); } } if (videoTrack.sps && videoTrack.pps && videoSamples.length) { // let's use input time scale as MP4 video timescale // we use input time scale straight away to avoid rounding issues on frame duration / cts computation var inputTimeScale = videoTrack.inputTimeScale; videoTrack.timescale = inputTimeScale; tracks.video = { container: 'video/mp4', codec: videoTrack.codec, initSegment: mp4_generator.initSegment([videoTrack]), metadata: { width: videoTrack.width, height: videoTrack.height } }; if (computePTSDTS) { var startPTS = Math.round(inputTimeScale * timeOffset); initPTS = Math.min(initPTS, videoSamples[0].pts - startPTS); initDTS = Math.min(initDTS, videoSamples[0].dts - startPTS); this.observer.trigger(events["default"].INIT_PTS_FOUND, { initPTS: initPTS }); } } else if (computePTSDTS && tracks.audio) { // initPTS found for audio-only stream with main and alt audio this.observer.trigger(events["default"].INIT_PTS_FOUND, { initPTS: initPTS }); } if (Object.keys(tracks).length) { observer.trigger(events["default"].FRAG_PARSING_INIT_SEGMENT, data); this.ISGenerated = true; if (computePTSDTS) { this._initPTS = initPTS; this._initDTS = initDTS; } } else { observer.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].FRAG_PARSING_ERROR, fatal: false, reason: 'no audio/video samples found' }); } }; _proto.remuxVideo = function remuxVideo(track, timeOffset, contiguous, audioTrackLength, accurateTimeOffset) { var offset = 8; var mp4SampleDuration; var mdat; var moof; var firstDTS; var lastDTS; var minPTS = Number.POSITIVE_INFINITY; var maxPTS = Number.NEGATIVE_INFINITY; var timeScale = track.timescale; var inputSamples = track.samples; var outputSamples = []; var nbSamples = inputSamples.length; var ptsNormalize = this._PTSNormalize; var initPTS = this._initPTS; // if parsed fragment is contiguous with last one, let's use last DTS value as reference var nextAvcDts = this.nextAvcDts; var isSafari = this.isSafari; if (nbSamples === 0) { return; } // Safari does not like overlapping DTS on consecutive fragments. let's use nextAvcDts to overcome this if fragments are consecutive if (isSafari) { // also consider consecutive fragments as being contiguous (even if a level switch occurs), // for sake of clarity: // consecutive fragments are frags with // - less than 100ms gaps between new time offset (if accurate) and next expected PTS OR // - less than 200 ms PTS gaps (timeScale/5) contiguous |= nbSamples && nextAvcDts && (accurateTimeOffset && Math.abs(timeOffset - nextAvcDts / timeScale) < 0.1 || Math.abs(inputSamples[0].pts - nextAvcDts - initPTS) < timeScale / 5); } if (!contiguous) { // if not contiguous, let's use target timeOffset nextAvcDts = timeOffset * timeScale; } // PTS is coded on 33bits, and can loop from -2^32 to 2^32 // ptsNormalize will make PTS/DTS value monotonic, we use last known DTS value as reference value inputSamples.forEach(function (sample) { sample.pts = ptsNormalize(sample.pts - initPTS, nextAvcDts); sample.dts = ptsNormalize(sample.dts - initPTS, nextAvcDts); minPTS = Math.min(sample.pts, minPTS); maxPTS = Math.max(sample.pts, maxPTS); }); // sort video samples by DTS then PTS then demux id order inputSamples.sort(function (a, b) { var deltadts = a.dts - b.dts; var deltapts = a.pts - b.pts; return deltadts || deltapts || a.id - b.id; }); // handle broken streams with PTS < DTS, tolerance up 0.2 seconds var PTSDTSshift = inputSamples.reduce(function (prev, curr) { return Math.max(Math.min(prev, curr.pts - curr.dts), -1 * PTS_DTS_SHIFT_TOLERANCE_90KHZ); }, 0); if (PTSDTSshift < 0) { logger["logger"].warn("PTS < DTS detected in video samples, shifting DTS by " + toMsFromMpegTsClock(PTSDTSshift, true) + " ms to overcome this issue"); for (var i = 0; i < nbSamples; i++) { inputSamples[i].dts = Math.max(0, inputSamples[i].dts + PTSDTSshift); } } // Get first/last DTS firstDTS = inputSamples[0].dts; lastDTS = inputSamples[nbSamples - 1].dts; // on Safari let's signal the same sample duration for all samples // sample duration (as expected by trun MP4 boxes), should be the delta between sample DTS // set this constant duration as being the avg delta between consecutive DTS. var averageSampleDuration = Math.round((lastDTS - firstDTS) / (nbSamples - 1)); // check timestamp continuity across consecutive fragments (this is to remove inter-fragment gap/hole) var delta = firstDTS - nextAvcDts; // if fragment are contiguous, detect hole/overlapping between fragments if (contiguous) { var foundHole = delta > averageSampleDuration; var foundOverlap = delta < -1; if (foundHole || foundOverlap) { if (foundHole) { logger["logger"].warn("AVC: " + toMsFromMpegTsClock(delta, true) + "ms (" + delta + "dts) hole between fragments detected, filling it"); } else { logger["logger"].warn("AVC: " + toMsFromMpegTsClock(-delta, true) + "ms (" + delta + "dts) overlapping between fragments detected"); } firstDTS = nextAvcDts; minPTS -= delta; inputSamples[0].dts = firstDTS; inputSamples[0].pts = minPTS; logger["logger"].log("Video: First PTS/DTS adjusted: " + toMsFromMpegTsClock(minPTS, true) + "/" + toMsFromMpegTsClock(firstDTS, true) + ", delta: " + toMsFromMpegTsClock(delta, true) + " ms"); } } // Clamp first DTS to 0 so that we're still aligning on initPTS, // and not passing negative values to MP4.traf. This will change initial frame compositionTimeOffset! firstDTS = Math.max(firstDTS, 0); var nbNalu = 0, naluLen = 0; for (var _i = 0; _i < nbSamples; _i++) { // compute total/avc sample length and nb of NAL units var sample = inputSamples[_i], units = sample.units, nbUnits = units.length, sampleLen = 0; for (var j = 0; j < nbUnits; j++) { sampleLen += units[j].data.length; } naluLen += sampleLen; nbNalu += nbUnits; sample.length = sampleLen; // normalize PTS/DTS if (isSafari) { // sample DTS is computed using a constant decoding offset (mp4SampleDuration) between samples sample.dts = firstDTS + _i * averageSampleDuration; } else { // ensure sample monotonic DTS sample.dts = Math.max(sample.dts, firstDTS); } // ensure that computed value is greater or equal than sample DTS sample.pts = Math.max(sample.pts, sample.dts); } /* concatenate the video data and construct the mdat in place (need 8 more bytes to fill length and mpdat type) */ var mdatSize = naluLen + 4 * nbNalu + 8; try { mdat = new Uint8Array(mdatSize); } catch (err) { this.observer.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MUX_ERROR, details: errors["ErrorDetails"].REMUX_ALLOC_ERROR, fatal: false, bytes: mdatSize, reason: "fail allocating video mdat " + mdatSize }); return; } var view = new DataView(mdat.buffer); view.setUint32(0, mdatSize); mdat.set(mp4_generator.types.mdat, 4); for (var _i2 = 0; _i2 < nbSamples; _i2++) { var avcSample = inputSamples[_i2], avcSampleUnits = avcSample.units, mp4SampleLength = 0, compositionTimeOffset = void 0; // convert NALU bitstream to MP4 format (prepend NALU with size field) for (var _j = 0, _nbUnits = avcSampleUnits.length; _j < _nbUnits; _j++) { var unit = avcSampleUnits[_j], unitData = unit.data, unitDataLen = unit.data.byteLength; view.setUint32(offset, unitDataLen); offset += 4; mdat.set(unitData, offset); offset += unitDataLen; mp4SampleLength += 4 + unitDataLen; } if (!isSafari) { // expected sample duration is the Decoding Timestamp diff of consecutive samples if (_i2 < nbSamples - 1) { mp4SampleDuration = inputSamples[_i2 + 1].dts - avcSample.dts; } else { var config = this.config, lastFrameDuration = avcSample.dts - inputSamples[_i2 > 0 ? _i2 - 1 : _i2].dts; if (config.stretchShortVideoTrack) { // In some cases, a segment's audio track duration may exceed the video track duration. // Since we've already remuxed audio, and we know how long the audio track is, we look to // see if the delta to the next segment is longer than maxBufferHole. // If so, playback would potentially get stuck, so we artificially inflate // the duration of the last frame to minimize any potential gap between segments. var maxBufferHole = config.maxBufferHole, gapTolerance = Math.floor(maxBufferHole * timeScale), deltaToFrameEnd = (audioTrackLength ? minPTS + audioTrackLength * timeScale : this.nextAudioPts) - avcSample.pts; if (deltaToFrameEnd > gapTolerance) { // We subtract lastFrameDuration from deltaToFrameEnd to try to prevent any video // frame overlap. maxBufferHole should be >> lastFrameDuration anyway. mp4SampleDuration = deltaToFrameEnd - lastFrameDuration; if (mp4SampleDuration < 0) { mp4SampleDuration = lastFrameDuration; } logger["logger"].log("It is approximately " + toMsFromMpegTsClock(deltaToFrameEnd, false) + " ms to the next segment; using duration " + toMsFromMpegTsClock(mp4SampleDuration, false) + " ms for the last video frame."); } else { mp4SampleDuration = lastFrameDuration; } } else { mp4SampleDuration = lastFrameDuration; } } compositionTimeOffset = Math.round(avcSample.pts - avcSample.dts); } else { compositionTimeOffset = Math.max(0, mp4SampleDuration * Math.round((avcSample.pts - avcSample.dts) / mp4SampleDuration)); } // console.log('PTS/DTS/initDTS/normPTS/normDTS/relative PTS : ${avcSample.pts}/${avcSample.dts}/${initDTS}/${ptsnorm}/${dtsnorm}/${(avcSample.pts/4294967296).toFixed(3)}'); outputSamples.push({ size: mp4SampleLength, // constant duration duration: mp4SampleDuration, cts: compositionTimeOffset, flags: { isLeading: 0, isDependedOn: 0, hasRedundancy: 0, degradPrio: 0, dependsOn: avcSample.key ? 2 : 1, isNonSync: avcSample.key ? 0 : 1 } }); } // next AVC sample DTS should be equal to last sample DTS + last sample duration (in PES timescale) this.nextAvcDts = lastDTS + mp4SampleDuration; var dropped = track.dropped; track.nbNalu = 0; track.dropped = 0; if (outputSamples.length && navigator.userAgent.toLowerCase().indexOf('chrome') > -1) { var flags = outputSamples[0].flags; // chrome workaround, mark first sample as being a Random Access Point to avoid sourcebuffer append issue // https://code.google.com/p/chromium/issues/detail?id=229412 flags.dependsOn = 2; flags.isNonSync = 0; } track.samples = outputSamples; moof = mp4_generator.moof(track.sequenceNumber++, firstDTS, track); track.samples = []; var data = { data1: moof, data2: mdat, startPTS: minPTS / timeScale, endPTS: (maxPTS + mp4SampleDuration) / timeScale, startDTS: firstDTS / timeScale, endDTS: this.nextAvcDts / timeScale, type: 'video', hasAudio: false, hasVideo: true, nb: outputSamples.length, dropped: dropped }; this.observer.trigger(events["default"].FRAG_PARSING_DATA, data); return data; }; _proto.remuxAudio = function remuxAudio(track, timeOffset, contiguous, accurateTimeOffset) { var inputTimeScale = track.inputTimeScale; var mp4timeScale = track.timescale; var scaleFactor = inputTimeScale / mp4timeScale; var mp4SampleDuration = track.isAAC ? 1024 : 1152; var inputSampleDuration = mp4SampleDuration * scaleFactor; var ptsNormalize = this._PTSNormalize; var initPTS = this._initPTS; var rawMPEG = !track.isAAC && this.typeSupported.mpeg; var mp4Sample; var fillFrame; var mdat; var moof; var firstPTS; var lastPTS; var offset = rawMPEG ? 0 : 8; var inputSamples = track.samples; var outputSamples = []; var nextAudioPts = this.nextAudioPts; // for audio samples, also consider consecutive fragments as being contiguous (even if a level switch occurs), // for sake of clarity: // consecutive fragments are frags with // - less than 100ms gaps between new time offset (if accurate) and next expected PTS OR // - less than 20 audio frames distance // contiguous fragments are consecutive fragments from same quality level (same level, new SN = old SN + 1) // this helps ensuring audio continuity // and this also avoids audio glitches/cut when switching quality, or reporting wrong duration on first audio frame contiguous |= inputSamples.length && nextAudioPts && (accurateTimeOffset && Math.abs(timeOffset - nextAudioPts / inputTimeScale) < 0.1 || Math.abs(inputSamples[0].pts - nextAudioPts - initPTS) < 20 * inputSampleDuration); // compute normalized PTS inputSamples.forEach(function (sample) { sample.pts = sample.dts = ptsNormalize(sample.pts - initPTS, timeOffset * inputTimeScale); }); // filter out sample with negative PTS that are not playable anyway // if we don't remove these negative samples, they will shift all audio samples forward. // leading to audio overlap between current / next fragment inputSamples = inputSamples.filter(function (sample) { return sample.pts >= 0; }); // in case all samples have negative PTS, and have been filtered out, return now if (inputSamples.length === 0) { return; } if (!contiguous) { if (!accurateTimeOffset) { // if frag are mot contiguous and if we cant trust time offset, let's use first sample PTS as next audio PTS nextAudioPts = inputSamples[0].pts; } else { // if timeOffset is accurate, let's use it as predicted next audio PTS nextAudioPts = timeOffset * inputTimeScale; } } // If the audio track is missing samples, the frames seem to get "left-shifted" within the // resulting mp4 segment, causing sync issues and leaving gaps at the end of the audio segment. // In an effort to prevent this from happening, we inject frames here where there are gaps. // When possible, we inject a silent frame; when that's not possible, we duplicate the last // frame. if (track.isAAC) { var maxAudioFramesDrift = this.config.maxAudioFramesDrift; for (var i = 0, nextPts = nextAudioPts; i < inputSamples.length;) { // First, let's see how far off this frame is from where we expect it to be var sample = inputSamples[i], delta; var pts = sample.pts; delta = pts - nextPts; // If we're overlapping by more than a duration, drop this sample if (delta <= -maxAudioFramesDrift * inputSampleDuration) { if (contiguous) { logger["logger"].warn("Dropping 1 audio frame @ " + toMsFromMpegTsClock(nextPts, true) / 1000 + "s due to " + toMsFromMpegTsClock(delta, true) + " ms overlap."); inputSamples.splice(i, 1); // Don't touch nextPtsNorm or i } else { // When changing qualities we can't trust that audio has been appended up to nextAudioPts // Warn about the overlap but do not drop samples as that can introduce buffer gaps logger["logger"].warn("Audio frame @ " + toMsFromMpegTsClock(pts, true) / 1000 + "s overlaps nextAudioPts by " + toMsFromMpegTsClock(delta, true) + " ms."); nextPts = pts + inputSampleDuration; i++; } } // eslint-disable-line brace-style // Insert missing frames if: // 1: We're more than maxAudioFramesDrift frame away // 2: Not more than MAX_SILENT_FRAME_DURATION away // 3: currentTime (aka nextPtsNorm) is not 0 else if (delta >= maxAudioFramesDrift * inputSampleDuration && delta < MAX_SILENT_FRAME_DURATION_90KHZ && nextPts) { var missing = Math.round(delta / inputSampleDuration); logger["logger"].warn("Injecting " + missing + " audio frames @ " + toMsFromMpegTsClock(nextPts, true) / 1000 + "s due to " + toMsFromMpegTsClock(delta, true) + " ms gap."); for (var j = 0; j < missing; j++) { var newStamp = Math.max(nextPts, 0); fillFrame = aac_helper.getSilentFrame(track.manifestCodec || track.codec, track.channelCount); if (!fillFrame) { logger["logger"].log('Unable to get silent frame for given audio codec; duplicating last frame instead.'); fillFrame = sample.unit.subarray(); } inputSamples.splice(i, 0, { unit: fillFrame, pts: newStamp, dts: newStamp }); nextPts += inputSampleDuration; i++; } // Adjust sample to next expected pts sample.pts = sample.dts = nextPts; nextPts += inputSampleDuration; i++; } else { // Otherwise, just adjust pts if (Math.abs(delta) > 0.1 * inputSampleDuration) {// logger.log(`Invalid frame delta ${Math.round(delta + inputSampleDuration)} at PTS ${Math.round(pts / 90)} (should be ${Math.round(inputSampleDuration)}).`); } sample.pts = sample.dts = nextPts; nextPts += inputSampleDuration; i++; } } } // compute mdat size, as we eventually filtered/added some samples var nbSamples = inputSamples.length; var mdatSize = 0; while (nbSamples--) { mdatSize += inputSamples[nbSamples].unit.byteLength; } for (var _j2 = 0, _nbSamples = inputSamples.length; _j2 < _nbSamples; _j2++) { var audioSample = inputSamples[_j2]; var unit = audioSample.unit; var _pts = audioSample.pts; // logger.log(`Audio/PTS:${toMsFromMpegTsClock(pts, true)}`); // if not first sample if (lastPTS !== undefined && mp4Sample) { mp4Sample.duration = Math.round((_pts - lastPTS) / scaleFactor); } else { var _delta = _pts - nextAudioPts; var numMissingFrames = 0; // if fragment are contiguous, detect hole/overlapping between fragments // contiguous fragments are consecutive fragments from same quality level (same level, new SN = old SN + 1) if (contiguous && track.isAAC) { // log delta if (_delta) { if (_delta > 0 && _delta < MAX_SILENT_FRAME_DURATION_90KHZ) { // Q: why do we have to round here, shouldn't this always result in an integer if timestamps are correct, // and if not, shouldn't we actually Math.ceil() instead? numMissingFrames = Math.round((_pts - nextAudioPts) / inputSampleDuration); logger["logger"].log(toMsFromMpegTsClock(_delta, true) + " ms hole between AAC samples detected,filling it"); if (numMissingFrames > 0) { fillFrame = aac_helper.getSilentFrame(track.manifestCodec || track.codec, track.channelCount); if (!fillFrame) { fillFrame = unit.subarray(); } mdatSize += numMissingFrames * fillFrame.length; } // if we have frame overlap, overlapping for more than half a frame duraion } else if (_delta < -12) { // drop overlapping audio frames... browser will deal with it logger["logger"].log("drop overlapping AAC sample, expected/parsed/delta: " + toMsFromMpegTsClock(nextAudioPts, true) + " ms / " + toMsFromMpegTsClock(_pts, true) + " ms / " + toMsFromMpegTsClock(-_delta, true) + " ms"); mdatSize -= unit.byteLength; continue; } // set PTS/DTS to expected PTS/DTS _pts = nextAudioPts; } } // remember first PTS of our audioSamples firstPTS = _pts; if (mdatSize > 0) { mdatSize += offset; try { mdat = new Uint8Array(mdatSize); } catch (err) { this.observer.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MUX_ERROR, details: errors["ErrorDetails"].REMUX_ALLOC_ERROR, fatal: false, bytes: mdatSize, reason: "fail allocating audio mdat " + mdatSize }); return; } if (!rawMPEG) { var view = new DataView(mdat.buffer); view.setUint32(0, mdatSize); mdat.set(mp4_generator.types.mdat, 4); } } else { // no audio samples return; } for (var _i3 = 0; _i3 < numMissingFrames; _i3++) { fillFrame = aac_helper.getSilentFrame(track.manifestCodec || track.codec, track.channelCount); if (!fillFrame) { logger["logger"].log('Unable to get silent frame for given audio codec; duplicating this frame instead.'); fillFrame = unit.subarray(); } mdat.set(fillFrame, offset); offset += fillFrame.byteLength; mp4Sample = { size: fillFrame.byteLength, cts: 0, duration: 1024, flags: { isLeading: 0, isDependedOn: 0, hasRedundancy: 0, degradPrio: 0, dependsOn: 1 } }; outputSamples.push(mp4Sample); } } mdat.set(unit, offset); var unitLen = unit.byteLength; offset += unitLen; // console.log('PTS/DTS/initDTS/normPTS/normDTS/relative PTS : ${audioSample.pts}/${audioSample.dts}/${initDTS}/${ptsnorm}/${dtsnorm}/${(audioSample.pts/4294967296).toFixed(3)}'); mp4Sample = { size: unitLen, cts: 0, duration: 0, flags: { isLeading: 0, isDependedOn: 0, hasRedundancy: 0, degradPrio: 0, dependsOn: 1 } }; outputSamples.push(mp4Sample); lastPTS = _pts; } var lastSampleDuration = 0; nbSamples = outputSamples.length; // set last sample duration as being identical to previous sample if (nbSamples >= 2) { lastSampleDuration = outputSamples[nbSamples - 2].duration; mp4Sample.duration = lastSampleDuration; } if (nbSamples) { // next audio sample PTS should be equal to last sample PTS + duration this.nextAudioPts = nextAudioPts = lastPTS + scaleFactor * lastSampleDuration; // logger.log('Audio/PTS/PTSend:' + audioSample.pts.toFixed(0) + '/' + this.nextAacDts.toFixed(0)); track.samples = outputSamples; if (rawMPEG) { moof = new Uint8Array(); } else { moof = mp4_generator.moof(track.sequenceNumber++, firstPTS / scaleFactor, track); } track.samples = []; var start = firstPTS / inputTimeScale; var end = nextAudioPts / inputTimeScale; var audioData = { data1: moof, data2: mdat, startPTS: start, endPTS: end, startDTS: start, endDTS: end, type: 'audio', hasAudio: true, hasVideo: false, nb: nbSamples }; this.observer.trigger(events["default"].FRAG_PARSING_DATA, audioData); return audioData; } return null; }; _proto.remuxEmptyAudio = function remuxEmptyAudio(track, timeOffset, contiguous, videoData) { var inputTimeScale = track.inputTimeScale; var mp4timeScale = track.samplerate ? track.samplerate : inputTimeScale; var scaleFactor = inputTimeScale / mp4timeScale; var nextAudioPts = this.nextAudioPts; // sync with video's timestamp var startDTS = (nextAudioPts !== undefined ? nextAudioPts : videoData.startDTS * inputTimeScale) + this._initDTS; var endDTS = videoData.endDTS * inputTimeScale + this._initDTS; // one sample's duration value var sampleDuration = 1024; var frameDuration = scaleFactor * sampleDuration; // samples count of this segment's duration var nbSamples = Math.ceil((endDTS - startDTS) / frameDuration); // silent frame var silentFrame = aac_helper.getSilentFrame(track.manifestCodec || track.codec, track.channelCount); logger["logger"].warn('remux empty Audio'); // Can't remux if we can't generate a silent frame... if (!silentFrame) { logger["logger"].trace('Unable to remuxEmptyAudio since we were unable to get a silent frame for given audio codec!'); return; } var samples = []; for (var i = 0; i < nbSamples; i++) { var stamp = startDTS + i * frameDuration; samples.push({ unit: silentFrame, pts: stamp, dts: stamp }); } track.samples = samples; this.remuxAudio(track, timeOffset, contiguous); }; _proto.remuxID3 = function remuxID3(track) { var length = track.samples.length; if (!length) { return; } var inputTimeScale = track.inputTimeScale; var initPTS = this._initPTS; var initDTS = this._initDTS; // consume samples for (var index = 0; index < length; index++) { var sample = track.samples[index]; // setting id3 pts, dts to relative time // using this._initPTS and this._initDTS to calculate relative time sample.pts = (sample.pts - initPTS) / inputTimeScale; sample.dts = (sample.dts - initDTS) / inputTimeScale; } this.observer.trigger(events["default"].FRAG_PARSING_METADATA, { samples: track.samples }); track.samples = []; }; _proto.remuxText = function remuxText(track) { track.samples.sort(function (a, b) { return a.pts - b.pts; }); var length = track.samples.length, sample; var inputTimeScale = track.inputTimeScale; var initPTS = this._initPTS; // consume samples if (length) { for (var index = 0; index < length; index++) { sample = track.samples[index]; // setting text pts, dts to relative time // using this._initPTS and this._initDTS to calculate relative time sample.pts = (sample.pts - initPTS) / inputTimeScale; } this.observer.trigger(events["default"].FRAG_PARSING_USERDATA, { samples: track.samples }); } track.samples = []; }; _proto._PTSNormalize = function _PTSNormalize(value, reference) { var offset; if (reference === undefined) { return value; } if (reference < value) { // - 2^33 offset = -8589934592; } else { // + 2^33 offset = 8589934592; } /* PTS is 33bit (from 0 to 2^33 -1) if diff between value and reference is bigger than half of the amplitude (2^32) then it means that PTS looping occured. fill the gap */ while (Math.abs(value - reference) > 4294967296) { value += offset; } return value; }; return MP4Remuxer; }(); /* harmony default export */ var mp4_remuxer = (mp4_remuxer_MP4Remuxer); // CONCATENATED MODULE: ./src/remux/passthrough-remuxer.js /** * passthrough remuxer */ var passthrough_remuxer_PassThroughRemuxer = /*#__PURE__*/function () { function PassThroughRemuxer(observer) { this.observer = observer; } var _proto = PassThroughRemuxer.prototype; _proto.destroy = function destroy() {}; _proto.resetTimeStamp = function resetTimeStamp() {}; _proto.resetInitSegment = function resetInitSegment() {}; _proto.remux = function remux(audioTrack, videoTrack, id3Track, textTrack, timeOffset, contiguous, accurateTimeOffset, rawData) { var observer = this.observer; var streamType = ''; if (audioTrack) { streamType += 'audio'; } if (videoTrack) { streamType += 'video'; } observer.trigger(events["default"].FRAG_PARSING_DATA, { data1: rawData, startPTS: timeOffset, startDTS: timeOffset, type: streamType, hasAudio: !!audioTrack, hasVideo: !!videoTrack, nb: 1, dropped: 0 }); // notify end of parsing observer.trigger(events["default"].FRAG_PARSED); }; return PassThroughRemuxer; }(); /* harmony default export */ var passthrough_remuxer = (passthrough_remuxer_PassThroughRemuxer); // CONCATENATED MODULE: ./src/demux/demuxer-inline.js /** * * inline demuxer: probe fragments and instantiate * appropriate demuxer depending on content type (TSDemuxer, AACDemuxer, ...) * */ // see https://stackoverflow.com/a/11237259/589493 var global = Object(get_self_scope["getSelfScope"])(); // safeguard for code that might run both on worker and main thread var now; // performance.now() not available on WebWorker, at least on Safari Desktop try { now = global.performance.now.bind(global.performance); } catch (err) { logger["logger"].debug('Unable to use Performance API on this environment'); now = global.Date.now; } var demuxer_inline_DemuxerInline = /*#__PURE__*/function () { function DemuxerInline(observer, typeSupported, config, vendor) { this.observer = observer; this.typeSupported = typeSupported; this.config = config; this.vendor = vendor; } var _proto = DemuxerInline.prototype; _proto.destroy = function destroy() { var demuxer = this.demuxer; if (demuxer) { demuxer.destroy(); } }; _proto.push = function push(data, decryptdata, initSegment, audioCodec, videoCodec, timeOffset, discontinuity, trackSwitch, contiguous, duration, accurateTimeOffset, defaultInitPTS) { var _this = this; if (data.byteLength > 0 && decryptdata != null && decryptdata.key != null && decryptdata.method === 'AES-128') { var decrypter = this.decrypter; if (decrypter == null) { decrypter = this.decrypter = new crypt_decrypter["default"](this.observer, this.config); } var startTime = now(); decrypter.decrypt(data, decryptdata.key.buffer, decryptdata.iv.buffer, function (decryptedData) { var endTime = now(); _this.observer.trigger(events["default"].FRAG_DECRYPTED, { stats: { tstart: startTime, tdecrypt: endTime } }); _this.pushDecrypted(new Uint8Array(decryptedData), decryptdata, new Uint8Array(initSegment), audioCodec, videoCodec, timeOffset, discontinuity, trackSwitch, contiguous, duration, accurateTimeOffset, defaultInitPTS); }); } else { this.pushDecrypted(new Uint8Array(data), decryptdata, new Uint8Array(initSegment), audioCodec, videoCodec, timeOffset, discontinuity, trackSwitch, contiguous, duration, accurateTimeOffset, defaultInitPTS); } }; _proto.pushDecrypted = function pushDecrypted(data, decryptdata, initSegment, audioCodec, videoCodec, timeOffset, discontinuity, trackSwitch, contiguous, duration, accurateTimeOffset, defaultInitPTS) { var demuxer = this.demuxer; var remuxer = this.remuxer; if (!demuxer || // in case of continuity change, or track switch // we might switch from content type (AAC container to TS container, or TS to fmp4 for example) discontinuity || trackSwitch) { var observer = this.observer; var typeSupported = this.typeSupported; var config = this.config; // probing order is TS/MP4/AAC/MP3 var muxConfig = [{ demux: tsdemuxer, remux: mp4_remuxer }, { demux: mp4demuxer["default"], remux: passthrough_remuxer }, { demux: aacdemuxer, remux: mp4_remuxer }, { demux: mp3demuxer, remux: mp4_remuxer }]; // probe for content type var mux; for (var i = 0, len = muxConfig.length; i < len; i++) { mux = muxConfig[i]; if (mux.demux.probe(data)) { break; } } if (!mux) { observer.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].FRAG_PARSING_ERROR, fatal: true, reason: 'no demux matching with content found' }); return; } // so let's check that current remuxer and demuxer are still valid if (!remuxer || !(remuxer instanceof mux.remux)) { remuxer = new mux.remux(observer, config, typeSupported, this.vendor); } if (!demuxer || !(demuxer instanceof mux.demux)) { demuxer = new mux.demux(observer, remuxer, config, typeSupported); this.probe = mux.demux.probe; } this.demuxer = demuxer; this.remuxer = remuxer; } if (discontinuity || trackSwitch) { demuxer.resetInitSegment(initSegment, audioCodec, videoCodec, duration); remuxer.resetInitSegment(); } if (discontinuity) { demuxer.resetTimeStamp(defaultInitPTS); remuxer.resetTimeStamp(defaultInitPTS); } if (typeof demuxer.setDecryptData === 'function') { demuxer.setDecryptData(decryptdata); } demuxer.append(data, timeOffset, contiguous, accurateTimeOffset); }; return DemuxerInline; }(); /* harmony default export */ var demuxer_inline = __webpack_exports__["default"] = (demuxer_inline_DemuxerInline); /***/ }), /***/ "./src/demux/demuxer-worker.js": /*!*************************************!*\ !*** ./src/demux/demuxer-worker.js ***! \*************************************/ /*! exports provided: default */ /*! ModuleConcatenation bailout: Module is referenced from these modules with unsupported syntax: ./src/demux/demuxer.js (referenced with require.resolve) */ /***/ (function(module, __webpack_exports__, __webpack_require__) { "use strict"; __webpack_require__.r(__webpack_exports__); /* harmony import */ var _demux_demuxer_inline__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../demux/demuxer-inline */ "./src/demux/demuxer-inline.js"); /* harmony import */ var _events__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ../events */ "./src/events.js"); /* harmony import */ var _utils_logger__WEBPACK_IMPORTED_MODULE_2__ = __webpack_require__(/*! ../utils/logger */ "./src/utils/logger.js"); /* harmony import */ var eventemitter3__WEBPACK_IMPORTED_MODULE_3__ = __webpack_require__(/*! eventemitter3 */ "./node_modules/eventemitter3/index.js"); /* harmony import */ var eventemitter3__WEBPACK_IMPORTED_MODULE_3___default = /*#__PURE__*/__webpack_require__.n(eventemitter3__WEBPACK_IMPORTED_MODULE_3__); /* demuxer web worker. * - listen to worker message, and trigger DemuxerInline upon reception of Fragments. * - provides MP4 Boxes back to main thread using [transferable objects](https://developers.google.com/web/updates/2011/12/Transferable-Objects-Lightning-Fast) in order to minimize message passing overhead. */ var DemuxerWorker = function DemuxerWorker(self) { // observer setup var observer = new eventemitter3__WEBPACK_IMPORTED_MODULE_3__["EventEmitter"](); observer.trigger = function trigger(event) { for (var _len = arguments.length, data = new Array(_len > 1 ? _len - 1 : 0), _key = 1; _key < _len; _key++) { data[_key - 1] = arguments[_key]; } observer.emit.apply(observer, [event, event].concat(data)); }; observer.off = function off(event) { for (var _len2 = arguments.length, data = new Array(_len2 > 1 ? _len2 - 1 : 0), _key2 = 1; _key2 < _len2; _key2++) { data[_key2 - 1] = arguments[_key2]; } observer.removeListener.apply(observer, [event].concat(data)); }; var forwardMessage = function forwardMessage(ev, data) { self.postMessage({ event: ev, data: data }); }; self.addEventListener('message', function (ev) { var data = ev.data; // console.log('demuxer cmd:' + data.cmd); switch (data.cmd) { case 'init': var config = JSON.parse(data.config); self.demuxer = new _demux_demuxer_inline__WEBPACK_IMPORTED_MODULE_0__["default"](observer, data.typeSupported, config, data.vendor); Object(_utils_logger__WEBPACK_IMPORTED_MODULE_2__["enableLogs"])(config.debug); // signal end of worker init forwardMessage('init', null); break; case 'demux': self.demuxer.push(data.data, data.decryptdata, data.initSegment, data.audioCodec, data.videoCodec, data.timeOffset, data.discontinuity, data.trackSwitch, data.contiguous, data.duration, data.accurateTimeOffset, data.defaultInitPTS); break; default: break; } }); // forward events to main thread observer.on(_events__WEBPACK_IMPORTED_MODULE_1__["default"].FRAG_DECRYPTED, forwardMessage); observer.on(_events__WEBPACK_IMPORTED_MODULE_1__["default"].FRAG_PARSING_INIT_SEGMENT, forwardMessage); observer.on(_events__WEBPACK_IMPORTED_MODULE_1__["default"].FRAG_PARSED, forwardMessage); observer.on(_events__WEBPACK_IMPORTED_MODULE_1__["default"].ERROR, forwardMessage); observer.on(_events__WEBPACK_IMPORTED_MODULE_1__["default"].FRAG_PARSING_METADATA, forwardMessage); observer.on(_events__WEBPACK_IMPORTED_MODULE_1__["default"].FRAG_PARSING_USERDATA, forwardMessage); observer.on(_events__WEBPACK_IMPORTED_MODULE_1__["default"].INIT_PTS_FOUND, forwardMessage); // special case for FRAG_PARSING_DATA: pass data1/data2 as transferable object (no copy) observer.on(_events__WEBPACK_IMPORTED_MODULE_1__["default"].FRAG_PARSING_DATA, function (ev, data) { var transferable = []; var message = { event: ev, data: data }; if (data.data1) { message.data1 = data.data1.buffer; transferable.push(data.data1.buffer); delete data.data1; } if (data.data2) { message.data2 = data.data2.buffer; transferable.push(data.data2.buffer); delete data.data2; } self.postMessage(message, transferable); }); }; /* harmony default export */ __webpack_exports__["default"] = (DemuxerWorker); /***/ }), /***/ "./src/demux/id3.js": /*!**************************!*\ !*** ./src/demux/id3.js ***! \**************************/ /*! exports provided: default, utf8ArrayToStr */ /***/ (function(module, __webpack_exports__, __webpack_require__) { "use strict"; __webpack_require__.r(__webpack_exports__); /* harmony export (binding) */ __webpack_require__.d(__webpack_exports__, "utf8ArrayToStr", function() { return utf8ArrayToStr; }); /* harmony import */ var _utils_get_self_scope__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../utils/get-self-scope */ "./src/utils/get-self-scope.js"); /** * ID3 parser */ var ID3 = /*#__PURE__*/function () { function ID3() {} /** * Returns true if an ID3 header can be found at offset in data * @param {Uint8Array} data - The data to search in * @param {number} offset - The offset at which to start searching * @return {boolean} - True if an ID3 header is found */ ID3.isHeader = function isHeader(data, offset) { /* * http://id3.org/id3v2.3.0 * [0] = 'I' * [1] = 'D' * [2] = '3' * [3,4] = {Version} * [5] = {Flags} * [6-9] = {ID3 Size} * * An ID3v2 tag can be detected with the following pattern: * $49 44 33 yy yy xx zz zz zz zz * Where yy is less than $FF, xx is the 'flags' byte and zz is less than $80 */ if (offset + 10 <= data.length) { // look for 'ID3' identifier if (data[offset] === 0x49 && data[offset + 1] === 0x44 && data[offset + 2] === 0x33) { // check version is within range if (data[offset + 3] < 0xFF && data[offset + 4] < 0xFF) { // check size is within range if (data[offset + 6] < 0x80 && data[offset + 7] < 0x80 && data[offset + 8] < 0x80 && data[offset + 9] < 0x80) { return true; } } } } return false; } /** * Returns true if an ID3 footer can be found at offset in data * @param {Uint8Array} data - The data to search in * @param {number} offset - The offset at which to start searching * @return {boolean} - True if an ID3 footer is found */ ; ID3.isFooter = function isFooter(data, offset) { /* * The footer is a copy of the header, but with a different identifier */ if (offset + 10 <= data.length) { // look for '3DI' identifier if (data[offset] === 0x33 && data[offset + 1] === 0x44 && data[offset + 2] === 0x49) { // check version is within range if (data[offset + 3] < 0xFF && data[offset + 4] < 0xFF) { // check size is within range if (data[offset + 6] < 0x80 && data[offset + 7] < 0x80 && data[offset + 8] < 0x80 && data[offset + 9] < 0x80) { return true; } } } } return false; } /** * Returns any adjacent ID3 tags found in data starting at offset, as one block of data * @param {Uint8Array} data - The data to search in * @param {number} offset - The offset at which to start searching * @return {Uint8Array} - The block of data containing any ID3 tags found */ ; ID3.getID3Data = function getID3Data(data, offset) { var front = offset; var length = 0; while (ID3.isHeader(data, offset)) { // ID3 header is 10 bytes length += 10; var size = ID3._readSize(data, offset + 6); length += size; if (ID3.isFooter(data, offset + 10)) { // ID3 footer is 10 bytes length += 10; } offset += length; } if (length > 0) { return data.subarray(front, front + length); } return undefined; }; ID3._readSize = function _readSize(data, offset) { var size = 0; size = (data[offset] & 0x7f) << 21; size |= (data[offset + 1] & 0x7f) << 14; size |= (data[offset + 2] & 0x7f) << 7; size |= data[offset + 3] & 0x7f; return size; } /** * Searches for the Elementary Stream timestamp found in the ID3 data chunk * @param {Uint8Array} data - Block of data containing one or more ID3 tags * @return {number} - The timestamp */ ; ID3.getTimeStamp = function getTimeStamp(data) { var frames = ID3.getID3Frames(data); for (var i = 0; i < frames.length; i++) { var frame = frames[i]; if (ID3.isTimeStampFrame(frame)) { return ID3._readTimeStamp(frame); } } return undefined; } /** * Returns true if the ID3 frame is an Elementary Stream timestamp frame * @param {ID3 frame} frame */ ; ID3.isTimeStampFrame = function isTimeStampFrame(frame) { return frame && frame.key === 'PRIV' && frame.info === 'com.apple.streaming.transportStreamTimestamp'; }; ID3._getFrameData = function _getFrameData(data) { /* Frame ID $xx xx xx xx (four characters) Size $xx xx xx xx Flags $xx xx */ var type = String.fromCharCode(data[0], data[1], data[2], data[3]); var size = ID3._readSize(data, 4); // skip frame id, size, and flags var offset = 10; return { type: type, size: size, data: data.subarray(offset, offset + size) }; } /** * Returns an array of ID3 frames found in all the ID3 tags in the id3Data * @param {Uint8Array} id3Data - The ID3 data containing one or more ID3 tags * @return {ID3 frame[]} - Array of ID3 frame objects */ ; ID3.getID3Frames = function getID3Frames(id3Data) { var offset = 0; var frames = []; while (ID3.isHeader(id3Data, offset)) { var size = ID3._readSize(id3Data, offset + 6); // skip past ID3 header offset += 10; var end = offset + size; // loop through frames in the ID3 tag while (offset + 8 < end) { var frameData = ID3._getFrameData(id3Data.subarray(offset)); var frame = ID3._decodeFrame(frameData); if (frame) { frames.push(frame); } // skip frame header and frame data offset += frameData.size + 10; } if (ID3.isFooter(id3Data, offset)) { offset += 10; } } return frames; }; ID3._decodeFrame = function _decodeFrame(frame) { if (frame.type === 'PRIV') { return ID3._decodePrivFrame(frame); } else if (frame.type[0] === 'T') { return ID3._decodeTextFrame(frame); } else if (frame.type[0] === 'W') { return ID3._decodeURLFrame(frame); } return undefined; }; ID3._readTimeStamp = function _readTimeStamp(timeStampFrame) { if (timeStampFrame.data.byteLength === 8) { var data = new Uint8Array(timeStampFrame.data); // timestamp is 33 bit expressed as a big-endian eight-octet number, // with the upper 31 bits set to zero. var pts33Bit = data[3] & 0x1; var timestamp = (data[4] << 23) + (data[5] << 15) + (data[6] << 7) + data[7]; timestamp /= 45; if (pts33Bit) { timestamp += 47721858.84; } // 2^32 / 90 return Math.round(timestamp); } return undefined; }; ID3._decodePrivFrame = function _decodePrivFrame(frame) { /* Format: <text string>\0<binary data> */ if (frame.size < 2) { return undefined; } var owner = ID3._utf8ArrayToStr(frame.data, true); var privateData = new Uint8Array(frame.data.subarray(owner.length + 1)); return { key: frame.type, info: owner, data: privateData.buffer }; }; ID3._decodeTextFrame = function _decodeTextFrame(frame) { if (frame.size < 2) { return undefined; } if (frame.type === 'TXXX') { /* Format: [0] = {Text Encoding} [1-?] = {Description}\0{Value} */ var index = 1; var description = ID3._utf8ArrayToStr(frame.data.subarray(index), true); index += description.length + 1; var value = ID3._utf8ArrayToStr(frame.data.subarray(index)); return { key: frame.type, info: description, data: value }; } else { /* Format: [0] = {Text Encoding} [1-?] = {Value} */ var text = ID3._utf8ArrayToStr(frame.data.subarray(1)); return { key: frame.type, data: text }; } }; ID3._decodeURLFrame = function _decodeURLFrame(frame) { if (frame.type === 'WXXX') { /* Format: [0] = {Text Encoding} [1-?] = {Description}\0{URL} */ if (frame.size < 2) { return undefined; } var index = 1; var description = ID3._utf8ArrayToStr(frame.data.subarray(index)); index += description.length + 1; var value = ID3._utf8ArrayToStr(frame.data.subarray(index)); return { key: frame.type, info: description, data: value }; } else { /* Format: [0-?] = {URL} */ var url = ID3._utf8ArrayToStr(frame.data); return { key: frame.type, data: url }; } } // http://stackoverflow.com/questions/8936984/uint8array-to-string-in-javascript/22373197 // http://www.onicos.com/staff/iz/amuse/javascript/expert/utf.txt /* utf.js - UTF-8 <=> UTF-16 convertion * * Copyright (C) 1999 Masanao Izumo <[email protected]> * Version: 1.0 * LastModified: Dec 25 1999 * This library is free. You can redistribute it and/or modify it. */ ; ID3._utf8ArrayToStr = function _utf8ArrayToStr(array, exitOnNull) { if (exitOnNull === void 0) { exitOnNull = false; } var decoder = getTextDecoder(); if (decoder) { var decoded = decoder.decode(array); if (exitOnNull) { // grab up to the first null var idx = decoded.indexOf('\0'); return idx !== -1 ? decoded.substring(0, idx) : decoded; } // remove any null characters return decoded.replace(/\0/g, ''); } var len = array.length; var c; var char2; var char3; var out = ''; var i = 0; while (i < len) { c = array[i++]; if (c === 0x00 && exitOnNull) { return out; } else if (c === 0x00 || c === 0x03) { // If the character is 3 (END_OF_TEXT) or 0 (NULL) then skip it continue; } switch (c >> 4) { case 0: case 1: case 2: case 3: case 4: case 5: case 6: case 7: // 0xxxxxxx out += String.fromCharCode(c); break; case 12: case 13: // 110x xxxx 10xx xxxx char2 = array[i++]; out += String.fromCharCode((c & 0x1F) << 6 | char2 & 0x3F); break; case 14: // 1110 xxxx 10xx xxxx 10xx xxxx char2 = array[i++]; char3 = array[i++]; out += String.fromCharCode((c & 0x0F) << 12 | (char2 & 0x3F) << 6 | (char3 & 0x3F) << 0); break; default: } } return out; }; return ID3; }(); var decoder; function getTextDecoder() { var global = Object(_utils_get_self_scope__WEBPACK_IMPORTED_MODULE_0__["getSelfScope"])(); // safeguard for code that might run both on worker and main thread if (!decoder && typeof global.TextDecoder !== 'undefined') { decoder = new global.TextDecoder('utf-8'); } return decoder; } var utf8ArrayToStr = ID3._utf8ArrayToStr; /* harmony default export */ __webpack_exports__["default"] = (ID3); /***/ }), /***/ "./src/demux/mp4demuxer.js": /*!*********************************!*\ !*** ./src/demux/mp4demuxer.js ***! \*********************************/ /*! exports provided: default */ /***/ (function(module, __webpack_exports__, __webpack_require__) { "use strict"; __webpack_require__.r(__webpack_exports__); /* harmony import */ var _utils_logger__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ../utils/logger */ "./src/utils/logger.js"); /* harmony import */ var _events__WEBPACK_IMPORTED_MODULE_1__ = __webpack_require__(/*! ../events */ "./src/events.js"); /** * MP4 demuxer */ var UINT32_MAX = Math.pow(2, 32) - 1; var MP4Demuxer = /*#__PURE__*/function () { function MP4Demuxer(observer, remuxer) { this.observer = observer; this.remuxer = remuxer; } var _proto = MP4Demuxer.prototype; _proto.resetTimeStamp = function resetTimeStamp(initPTS) { this.initPTS = initPTS; }; _proto.resetInitSegment = function resetInitSegment(initSegment, audioCodec, videoCodec, duration) { // jshint unused:false if (initSegment && initSegment.byteLength) { var initData = this.initData = MP4Demuxer.parseInitSegment(initSegment); // default audio codec if nothing specified // TODO : extract that from initsegment if (audioCodec == null) { audioCodec = 'mp4a.40.5'; } if (videoCodec == null) { videoCodec = 'avc1.42e01e'; } var tracks = {}; if (initData.audio && initData.video) { tracks.audiovideo = { container: 'video/mp4', codec: audioCodec + ',' + videoCodec, initSegment: duration ? initSegment : null }; } else { if (initData.audio) { tracks.audio = { container: 'audio/mp4', codec: audioCodec, initSegment: duration ? initSegment : null }; } if (initData.video) { tracks.video = { container: 'video/mp4', codec: videoCodec, initSegment: duration ? initSegment : null }; } } this.observer.trigger(_events__WEBPACK_IMPORTED_MODULE_1__["default"].FRAG_PARSING_INIT_SEGMENT, { tracks: tracks }); } else { if (audioCodec) { this.audioCodec = audioCodec; } if (videoCodec) { this.videoCodec = videoCodec; } } }; MP4Demuxer.probe = function probe(data) { // ensure we find a moof box in the first 16 kB return MP4Demuxer.findBox({ data: data, start: 0, end: Math.min(data.length, 16384) }, ['moof']).length > 0; }; MP4Demuxer.bin2str = function bin2str(buffer) { return String.fromCharCode.apply(null, buffer); }; MP4Demuxer.readUint16 = function readUint16(buffer, offset) { if (buffer.data) { offset += buffer.start; buffer = buffer.data; } var val = buffer[offset] << 8 | buffer[offset + 1]; return val < 0 ? 65536 + val : val; }; MP4Demuxer.readUint32 = function readUint32(buffer, offset) { if (buffer.data) { offset += buffer.start; buffer = buffer.data; } var val = buffer[offset] << 24 | buffer[offset + 1] << 16 | buffer[offset + 2] << 8 | buffer[offset + 3]; return val < 0 ? 4294967296 + val : val; }; MP4Demuxer.writeUint32 = function writeUint32(buffer, offset, value) { if (buffer.data) { offset += buffer.start; buffer = buffer.data; } buffer[offset] = value >> 24; buffer[offset + 1] = value >> 16 & 0xff; buffer[offset + 2] = value >> 8 & 0xff; buffer[offset + 3] = value & 0xff; } // Find the data for a box specified by its path ; MP4Demuxer.findBox = function findBox(data, path) { var results = [], i, size, type, end, subresults, start, endbox; if (data.data) { start = data.start; end = data.end; data = data.data; } else { start = 0; end = data.byteLength; } if (!path.length) { // short-circuit the search for empty paths return null; } for (i = start; i < end;) { size = MP4Demuxer.readUint32(data, i); type = MP4Demuxer.bin2str(data.subarray(i + 4, i + 8)); endbox = size > 1 ? i + size : end; if (type === path[0]) { if (path.length === 1) { // this is the end of the path and we've found the box we were // looking for results.push({ data: data, start: i + 8, end: endbox }); } else { // recursively search for the next box along the path subresults = MP4Demuxer.findBox({ data: data, start: i + 8, end: endbox }, path.slice(1)); if (subresults.length) { results = results.concat(subresults); } } } i = endbox; } // we've finished searching all of data return results; }; MP4Demuxer.parseSegmentIndex = function parseSegmentIndex(initSegment) { var moov = MP4Demuxer.findBox(initSegment, ['moov'])[0]; var moovEndOffset = moov ? moov.end : null; // we need this in case we need to chop of garbage of the end of current data var index = 0; var sidx = MP4Demuxer.findBox(initSegment, ['sidx']); var references; if (!sidx || !sidx[0]) { return null; } references = []; sidx = sidx[0]; var version = sidx.data[0]; // set initial offset, we skip the reference ID (not needed) index = version === 0 ? 8 : 16; var timescale = MP4Demuxer.readUint32(sidx, index); index += 4; // TODO: parse earliestPresentationTime and firstOffset // usually zero in our case var earliestPresentationTime = 0; var firstOffset = 0; if (version === 0) { index += 8; } else { index += 16; } // skip reserved index += 2; var startByte = sidx.end + firstOffset; var referencesCount = MP4Demuxer.readUint16(sidx, index); index += 2; for (var i = 0; i < referencesCount; i++) { var referenceIndex = index; var referenceInfo = MP4Demuxer.readUint32(sidx, referenceIndex); referenceIndex += 4; var referenceSize = referenceInfo & 0x7FFFFFFF; var referenceType = (referenceInfo & 0x80000000) >>> 31; if (referenceType === 1) { console.warn('SIDX has hierarchical references (not supported)'); return; } var subsegmentDuration = MP4Demuxer.readUint32(sidx, referenceIndex); referenceIndex += 4; references.push({ referenceSize: referenceSize, subsegmentDuration: subsegmentDuration, // unscaled info: { duration: subsegmentDuration / timescale, start: startByte, end: startByte + referenceSize - 1 } }); startByte += referenceSize; // Skipping 1 bit for |startsWithSap|, 3 bits for |sapType|, and 28 bits // for |sapDelta|. referenceIndex += 4; // skip to next ref index = referenceIndex; } return { earliestPresentationTime: earliestPresentationTime, timescale: timescale, version: version, referencesCount: referencesCount, references: references, moovEndOffset: moovEndOffset }; } /** * Parses an MP4 initialization segment and extracts stream type and * timescale values for any declared tracks. Timescale values indicate the * number of clock ticks per second to assume for time-based values * elsewhere in the MP4. * * To determine the start time of an MP4, you need two pieces of * information: the timescale unit and the earliest base media decode * time. Multiple timescales can be specified within an MP4 but the * base media decode time is always expressed in the timescale from * the media header box for the track: * ``` * moov > trak > mdia > mdhd.timescale * moov > trak > mdia > hdlr * ``` * @param init {Uint8Array} the bytes of the init segment * @return {object} a hash of track type to timescale values or null if * the init segment is malformed. */ ; MP4Demuxer.parseInitSegment = function parseInitSegment(initSegment) { var result = []; var traks = MP4Demuxer.findBox(initSegment, ['moov', 'trak']); traks.forEach(function (trak) { var tkhd = MP4Demuxer.findBox(trak, ['tkhd'])[0]; if (tkhd) { var version = tkhd.data[tkhd.start]; var index = version === 0 ? 12 : 20; var trackId = MP4Demuxer.readUint32(tkhd, index); var mdhd = MP4Demuxer.findBox(trak, ['mdia', 'mdhd'])[0]; if (mdhd) { version = mdhd.data[mdhd.start]; index = version === 0 ? 12 : 20; var timescale = MP4Demuxer.readUint32(mdhd, index); var hdlr = MP4Demuxer.findBox(trak, ['mdia', 'hdlr'])[0]; if (hdlr) { var hdlrType = MP4Demuxer.bin2str(hdlr.data.subarray(hdlr.start + 8, hdlr.start + 12)); var type = { 'soun': 'audio', 'vide': 'video' }[hdlrType]; if (type) { // extract codec info. TODO : parse codec details to be able to build MIME type var codecBox = MP4Demuxer.findBox(trak, ['mdia', 'minf', 'stbl', 'stsd']); if (codecBox.length) { codecBox = codecBox[0]; var codecType = MP4Demuxer.bin2str(codecBox.data.subarray(codecBox.start + 12, codecBox.start + 16)); _utils_logger__WEBPACK_IMPORTED_MODULE_0__["logger"].log("MP4Demuxer:" + type + ":" + codecType + " found"); } result[trackId] = { timescale: timescale, type: type }; result[type] = { timescale: timescale, id: trackId }; } } } } }); return result; } /** * Determine the base media decode start time, in seconds, for an MP4 * fragment. If multiple fragments are specified, the earliest time is * returned. * * The base media decode time can be parsed from track fragment * metadata: * ``` * moof > traf > tfdt.baseMediaDecodeTime * ``` * It requires the timescale value from the mdhd to interpret. * * @param timescale {object} a hash of track ids to timescale values. * @return {number} the earliest base media decode start time for the * fragment, in seconds */ ; MP4Demuxer.getStartDTS = function getStartDTS(initData, fragment) { var trafs, baseTimes, result; // we need info from two childrend of each track fragment box trafs = MP4Demuxer.findBox(fragment, ['moof', 'traf']); // determine the start times for each track baseTimes = [].concat.apply([], trafs.map(function (traf) { return MP4Demuxer.findBox(traf, ['tfhd']).map(function (tfhd) { var id, scale, baseTime; // get the track id from the tfhd id = MP4Demuxer.readUint32(tfhd, 4); // assume a 90kHz clock if no timescale was specified scale = initData[id].timescale || 90e3; // get the base media decode time from the tfdt baseTime = MP4Demuxer.findBox(traf, ['tfdt']).map(function (tfdt) { var version, result; version = tfdt.data[tfdt.start]; result = MP4Demuxer.readUint32(tfdt, 4); if (version === 1) { result *= Math.pow(2, 32); result += MP4Demuxer.readUint32(tfdt, 8); } return result; })[0]; // convert base time to seconds return baseTime / scale; }); })); // return the minimum result = Math.min.apply(null, baseTimes); return isFinite(result) ? result : 0; }; MP4Demuxer.offsetStartDTS = function offsetStartDTS(initData, fragment, timeOffset) { MP4Demuxer.findBox(fragment, ['moof', 'traf']).map(function (traf) { return MP4Demuxer.findBox(traf, ['tfhd']).map(function (tfhd) { // get the track id from the tfhd var id = MP4Demuxer.readUint32(tfhd, 4); // assume a 90kHz clock if no timescale was specified var timescale = initData[id].timescale || 90e3; // get the base media decode time from the tfdt MP4Demuxer.findBox(traf, ['tfdt']).map(function (tfdt) { var version = tfdt.data[tfdt.start]; var baseMediaDecodeTime = MP4Demuxer.readUint32(tfdt, 4); if (version === 0) { MP4Demuxer.writeUint32(tfdt, 4, baseMediaDecodeTime - timeOffset * timescale); } else { baseMediaDecodeTime *= Math.pow(2, 32); baseMediaDecodeTime += MP4Demuxer.readUint32(tfdt, 8); baseMediaDecodeTime -= timeOffset * timescale; baseMediaDecodeTime = Math.max(baseMediaDecodeTime, 0); var upper = Math.floor(baseMediaDecodeTime / (UINT32_MAX + 1)); var lower = Math.floor(baseMediaDecodeTime % (UINT32_MAX + 1)); MP4Demuxer.writeUint32(tfdt, 4, upper); MP4Demuxer.writeUint32(tfdt, 8, lower); } }); }); }); } // feed incoming data to the front of the parsing pipeline ; _proto.append = function append(data, timeOffset, contiguous, accurateTimeOffset) { var initData = this.initData; if (!initData) { this.resetInitSegment(data, this.audioCodec, this.videoCodec, false); initData = this.initData; } var startDTS, initPTS = this.initPTS; if (initPTS === undefined) { var _startDTS = MP4Demuxer.getStartDTS(initData, data); this.initPTS = initPTS = _startDTS - timeOffset; this.observer.trigger(_events__WEBPACK_IMPORTED_MODULE_1__["default"].INIT_PTS_FOUND, { initPTS: initPTS }); } MP4Demuxer.offsetStartDTS(initData, data, initPTS); startDTS = MP4Demuxer.getStartDTS(initData, data); this.remuxer.remux(initData.audio, initData.video, null, null, startDTS, contiguous, accurateTimeOffset, data); }; _proto.destroy = function destroy() {}; return MP4Demuxer; }(); /* harmony default export */ __webpack_exports__["default"] = (MP4Demuxer); /***/ }), /***/ "./src/errors.ts": /*!***********************!*\ !*** ./src/errors.ts ***! \***********************/ /*! exports provided: ErrorTypes, ErrorDetails */ /***/ (function(module, __webpack_exports__, __webpack_require__) { "use strict"; __webpack_require__.r(__webpack_exports__); /* harmony export (binding) */ __webpack_require__.d(__webpack_exports__, "ErrorTypes", function() { return ErrorTypes; }); /* harmony export (binding) */ __webpack_require__.d(__webpack_exports__, "ErrorDetails", function() { return ErrorDetails; }); var ErrorTypes; /** * @enum {ErrorDetails} * @typedef {string} ErrorDetail */ (function (ErrorTypes) { ErrorTypes["NETWORK_ERROR"] = "networkError"; ErrorTypes["MEDIA_ERROR"] = "mediaError"; ErrorTypes["KEY_SYSTEM_ERROR"] = "keySystemError"; ErrorTypes["MUX_ERROR"] = "muxError"; ErrorTypes["OTHER_ERROR"] = "otherError"; })(ErrorTypes || (ErrorTypes = {})); var ErrorDetails; (function (ErrorDetails) { ErrorDetails["KEY_SYSTEM_NO_KEYS"] = "keySystemNoKeys"; ErrorDetails["KEY_SYSTEM_NO_ACCESS"] = "keySystemNoAccess"; ErrorDetails["KEY_SYSTEM_NO_SESSION"] = "keySystemNoSession"; ErrorDetails["KEY_SYSTEM_LICENSE_REQUEST_FAILED"] = "keySystemLicenseRequestFailed"; ErrorDetails["KEY_SYSTEM_NO_INIT_DATA"] = "keySystemNoInitData"; ErrorDetails["MANIFEST_LOAD_ERROR"] = "manifestLoadError"; ErrorDetails["MANIFEST_LOAD_TIMEOUT"] = "manifestLoadTimeOut"; ErrorDetails["MANIFEST_PARSING_ERROR"] = "manifestParsingError"; ErrorDetails["MANIFEST_INCOMPATIBLE_CODECS_ERROR"] = "manifestIncompatibleCodecsError"; ErrorDetails["LEVEL_EMPTY_ERROR"] = "levelEmptyError"; ErrorDetails["LEVEL_LOAD_ERROR"] = "levelLoadError"; ErrorDetails["LEVEL_LOAD_TIMEOUT"] = "levelLoadTimeOut"; ErrorDetails["LEVEL_SWITCH_ERROR"] = "levelSwitchError"; ErrorDetails["AUDIO_TRACK_LOAD_ERROR"] = "audioTrackLoadError"; ErrorDetails["AUDIO_TRACK_LOAD_TIMEOUT"] = "audioTrackLoadTimeOut"; ErrorDetails["FRAG_LOAD_ERROR"] = "fragLoadError"; ErrorDetails["FRAG_LOAD_TIMEOUT"] = "fragLoadTimeOut"; ErrorDetails["FRAG_DECRYPT_ERROR"] = "fragDecryptError"; ErrorDetails["FRAG_PARSING_ERROR"] = "fragParsingError"; ErrorDetails["REMUX_ALLOC_ERROR"] = "remuxAllocError"; ErrorDetails["KEY_LOAD_ERROR"] = "keyLoadError"; ErrorDetails["KEY_LOAD_TIMEOUT"] = "keyLoadTimeOut"; ErrorDetails["BUFFER_ADD_CODEC_ERROR"] = "bufferAddCodecError"; ErrorDetails["BUFFER_APPEND_ERROR"] = "bufferAppendError"; ErrorDetails["BUFFER_APPENDING_ERROR"] = "bufferAppendingError"; ErrorDetails["BUFFER_STALLED_ERROR"] = "bufferStalledError"; ErrorDetails["BUFFER_FULL_ERROR"] = "bufferFullError"; ErrorDetails["BUFFER_SEEK_OVER_HOLE"] = "bufferSeekOverHole"; ErrorDetails["BUFFER_NUDGE_ON_STALL"] = "bufferNudgeOnStall"; ErrorDetails["INTERNAL_EXCEPTION"] = "internalException"; })(ErrorDetails || (ErrorDetails = {})); /***/ }), /***/ "./src/events.js": /*!***********************!*\ !*** ./src/events.js ***! \***********************/ /*! exports provided: default */ /***/ (function(module, __webpack_exports__, __webpack_require__) { "use strict"; __webpack_require__.r(__webpack_exports__); /** * @readonly * @enum {string} */ var HlsEvents = { // fired before MediaSource is attaching to media element - data: { media } MEDIA_ATTACHING: 'hlsMediaAttaching', // fired when MediaSource has been succesfully attached to media element - data: { } MEDIA_ATTACHED: 'hlsMediaAttached', // fired before detaching MediaSource from media element - data: { } MEDIA_DETACHING: 'hlsMediaDetaching', // fired when MediaSource has been detached from media element - data: { } MEDIA_DETACHED: 'hlsMediaDetached', // fired when we buffer is going to be reset - data: { } BUFFER_RESET: 'hlsBufferReset', // fired when we know about the codecs that we need buffers for to push into - data: {tracks : { container, codec, levelCodec, initSegment, metadata }} BUFFER_CODECS: 'hlsBufferCodecs', // fired when sourcebuffers have been created - data: { tracks : tracks } BUFFER_CREATED: 'hlsBufferCreated', // fired when we append a segment to the buffer - data: { segment: segment object } BUFFER_APPENDING: 'hlsBufferAppending', // fired when we are done with appending a media segment to the buffer - data : { parent : segment parent that triggered BUFFER_APPENDING, pending : nb of segments waiting for appending for this segment parent} BUFFER_APPENDED: 'hlsBufferAppended', // fired when the stream is finished and we want to notify the media buffer that there will be no more data - data: { } BUFFER_EOS: 'hlsBufferEos', // fired when the media buffer should be flushed - data { startOffset, endOffset } BUFFER_FLUSHING: 'hlsBufferFlushing', // fired when the media buffer has been flushed - data: { } BUFFER_FLUSHED: 'hlsBufferFlushed', // fired to signal that a manifest loading starts - data: { url : manifestURL} MANIFEST_LOADING: 'hlsManifestLoading', // fired after manifest has been loaded - data: { levels : [available quality levels], audioTracks : [ available audio tracks], url : manifestURL, stats : { trequest, tfirst, tload, mtime}} MANIFEST_LOADED: 'hlsManifestLoaded', // fired after manifest has been parsed - data: { levels : [available quality levels], firstLevel : index of first quality level appearing in Manifest} MANIFEST_PARSED: 'hlsManifestParsed', // fired when a level switch is requested - data: { level : id of new level } LEVEL_SWITCHING: 'hlsLevelSwitching', // fired when a level switch is effective - data: { level : id of new level } LEVEL_SWITCHED: 'hlsLevelSwitched', // fired when a level playlist loading starts - data: { url : level URL, level : id of level being loaded} LEVEL_LOADING: 'hlsLevelLoading', // fired when a level playlist loading finishes - data: { details : levelDetails object, level : id of loaded level, stats : { trequest, tfirst, tload, mtime} } LEVEL_LOADED: 'hlsLevelLoaded', // fired when a level's details have been updated based on previous details, after it has been loaded - data: { details : levelDetails object, level : id of updated level } LEVEL_UPDATED: 'hlsLevelUpdated', // fired when a level's PTS information has been updated after parsing a fragment - data: { details : levelDetails object, level : id of updated level, drift: PTS drift observed when parsing last fragment } LEVEL_PTS_UPDATED: 'hlsLevelPtsUpdated', // fired to notify that levels have changed after removing a level - data: { levels : [available quality levels] } LEVELS_UPDATED: 'hlsLevelsUpdated', // fired to notify that audio track lists has been updated - data: { audioTracks : audioTracks } AUDIO_TRACKS_UPDATED: 'hlsAudioTracksUpdated', // fired when an audio track switching is requested - data: { id : audio track id } AUDIO_TRACK_SWITCHING: 'hlsAudioTrackSwitching', // fired when an audio track switch actually occurs - data: { id : audio track id } AUDIO_TRACK_SWITCHED: 'hlsAudioTrackSwitched', // fired when an audio track loading starts - data: { url : audio track URL, id : audio track id } AUDIO_TRACK_LOADING: 'hlsAudioTrackLoading', // fired when an audio track loading finishes - data: { details : levelDetails object, id : audio track id, stats : { trequest, tfirst, tload, mtime } } AUDIO_TRACK_LOADED: 'hlsAudioTrackLoaded', // fired to notify that subtitle track lists has been updated - data: { subtitleTracks : subtitleTracks } SUBTITLE_TRACKS_UPDATED: 'hlsSubtitleTracksUpdated', // fired when an subtitle track switch occurs - data: { id : subtitle track id } SUBTITLE_TRACK_SWITCH: 'hlsSubtitleTrackSwitch', // fired when a subtitle track loading starts - data: { url : subtitle track URL, id : subtitle track id } SUBTITLE_TRACK_LOADING: 'hlsSubtitleTrackLoading', // fired when a subtitle track loading finishes - data: { details : levelDetails object, id : subtitle track id, stats : { trequest, tfirst, tload, mtime } } SUBTITLE_TRACK_LOADED: 'hlsSubtitleTrackLoaded', // fired when a subtitle fragment has been processed - data: { success : boolean, frag : the processed frag } SUBTITLE_FRAG_PROCESSED: 'hlsSubtitleFragProcessed', // fired when a set of VTTCues to be managed externally has been parsed - data: { type: string, track: string, cues: [ VTTCue ] } CUES_PARSED: 'hlsCuesParsed', // fired when a text track to be managed externally is found - data: { tracks: [ { label: string, kind: string, default: boolean } ] } NON_NATIVE_TEXT_TRACKS_FOUND: 'hlsNonNativeTextTracksFound', // fired when the first timestamp is found - data: { id : demuxer id, initPTS: initPTS, frag : fragment object } INIT_PTS_FOUND: 'hlsInitPtsFound', // fired when a fragment loading starts - data: { frag : fragment object } FRAG_LOADING: 'hlsFragLoading', // fired when a fragment loading is progressing - data: { frag : fragment object, { trequest, tfirst, loaded } } FRAG_LOAD_PROGRESS: 'hlsFragLoadProgress', // Identifier for fragment load aborting for emergency switch down - data: { frag : fragment object } FRAG_LOAD_EMERGENCY_ABORTED: 'hlsFragLoadEmergencyAborted', // fired when a fragment loading is completed - data: { frag : fragment object, payload : fragment payload, stats : { trequest, tfirst, tload, length } } FRAG_LOADED: 'hlsFragLoaded', // fired when a fragment has finished decrypting - data: { id : demuxer id, frag: fragment object, payload : fragment payload, stats : { tstart, tdecrypt } } FRAG_DECRYPTED: 'hlsFragDecrypted', // fired when Init Segment has been extracted from fragment - data: { id : demuxer id, frag: fragment object, moov : moov MP4 box, codecs : codecs found while parsing fragment } FRAG_PARSING_INIT_SEGMENT: 'hlsFragParsingInitSegment', // fired when parsing sei text is completed - data: { id : demuxer id, frag: fragment object, samples : [ sei samples pes ] } FRAG_PARSING_USERDATA: 'hlsFragParsingUserdata', // fired when parsing id3 is completed - data: { id : demuxer id, frag: fragment object, samples : [ id3 samples pes ] } FRAG_PARSING_METADATA: 'hlsFragParsingMetadata', // fired when data have been extracted from fragment - data: { id : demuxer id, frag: fragment object, data1 : moof MP4 box or TS fragments, data2 : mdat MP4 box or null} FRAG_PARSING_DATA: 'hlsFragParsingData', // fired when fragment parsing is completed - data: { id : demuxer id, frag: fragment object } FRAG_PARSED: 'hlsFragParsed', // fired when fragment remuxed MP4 boxes have all been appended into SourceBuffer - data: { id : demuxer id, frag : fragment object, stats : { trequest, tfirst, tload, tparsed, tbuffered, length, bwEstimate } } FRAG_BUFFERED: 'hlsFragBuffered', // fired when fragment matching with current media position is changing - data : { id : demuxer id, frag : fragment object } FRAG_CHANGED: 'hlsFragChanged', // Identifier for a FPS drop event - data: { curentDropped, currentDecoded, totalDroppedFrames } FPS_DROP: 'hlsFpsDrop', // triggered when FPS drop triggers auto level capping - data: { level, droppedlevel } FPS_DROP_LEVEL_CAPPING: 'hlsFpsDropLevelCapping', // Identifier for an error event - data: { type : error type, details : error details, fatal : if true, hls.js cannot/will not try to recover, if false, hls.js will try to recover,other error specific data } ERROR: 'hlsError', // fired when hls.js instance starts destroying. Different from MEDIA_DETACHED as one could want to detach and reattach a media to the instance of hls.js to handle mid-rolls for example - data: { } DESTROYING: 'hlsDestroying', // fired when a decrypt key loading starts - data: { frag : fragment object } KEY_LOADING: 'hlsKeyLoading', // fired when a decrypt key loading is completed - data: { frag : fragment object, payload : key payload, stats : { trequest, tfirst, tload, length } } KEY_LOADED: 'hlsKeyLoaded', // fired upon stream controller state transitions - data: { previousState, nextState } STREAM_STATE_TRANSITION: 'hlsStreamStateTransition', // fired when the live back buffer is reached defined by the liveBackBufferLength config option - data : { bufferEnd: number } LIVE_BACK_BUFFER_REACHED: 'hlsLiveBackBufferReached' }; /* harmony default export */ __webpack_exports__["default"] = (HlsEvents); /***/ }), /***/ "./src/hls.ts": /*!*********************************!*\ !*** ./src/hls.ts + 50 modules ***! \*********************************/ /*! exports provided: default */ /*! ModuleConcatenation bailout: Cannot concat with ./src/crypt/decrypter.js because of ./src/demux/demuxer-worker.js */ /*! ModuleConcatenation bailout: Cannot concat with ./src/demux/demuxer-inline.js because of ./src/demux/demuxer-worker.js */ /*! ModuleConcatenation bailout: Cannot concat with ./src/demux/id3.js because of ./src/demux/demuxer-worker.js */ /*! ModuleConcatenation bailout: Cannot concat with ./src/demux/mp4demuxer.js because of ./src/demux/demuxer-worker.js */ /*! ModuleConcatenation bailout: Cannot concat with ./src/errors.ts because of ./src/demux/demuxer-worker.js */ /*! ModuleConcatenation bailout: Cannot concat with ./src/events.js because of ./src/demux/demuxer-worker.js */ /*! ModuleConcatenation bailout: Cannot concat with ./src/polyfills/number.js because of ./src/demux/demuxer-worker.js */ /*! ModuleConcatenation bailout: Cannot concat with ./src/utils/get-self-scope.js because of ./src/demux/demuxer-worker.js */ /*! ModuleConcatenation bailout: Cannot concat with ./src/utils/logger.js because of ./src/demux/demuxer-worker.js */ /*! ModuleConcatenation bailout: Cannot concat with ./node_modules/eventemitter3/index.js (<- Module is not an ECMAScript module) */ /*! ModuleConcatenation bailout: Cannot concat with ./node_modules/url-toolkit/src/url-toolkit.js (<- Module is not an ECMAScript module) */ /***/ (function(module, __webpack_exports__, __webpack_require__) { "use strict"; // ESM COMPAT FLAG __webpack_require__.r(__webpack_exports__); // EXPORTS __webpack_require__.d(__webpack_exports__, "default", function() { return /* binding */ hls_Hls; }); // NAMESPACE OBJECT: ./src/utils/cues.ts var cues_namespaceObject = {}; __webpack_require__.r(cues_namespaceObject); __webpack_require__.d(cues_namespaceObject, "newCue", function() { return newCue; }); // EXTERNAL MODULE: ./node_modules/url-toolkit/src/url-toolkit.js var url_toolkit = __webpack_require__("./node_modules/url-toolkit/src/url-toolkit.js"); // EXTERNAL MODULE: ./src/errors.ts var errors = __webpack_require__("./src/errors.ts"); // EXTERNAL MODULE: ./src/polyfills/number.js var number = __webpack_require__("./src/polyfills/number.js"); // EXTERNAL MODULE: ./src/events.js var events = __webpack_require__("./src/events.js"); // EXTERNAL MODULE: ./src/utils/logger.js var logger = __webpack_require__("./src/utils/logger.js"); // CONCATENATED MODULE: ./src/event-handler.ts /* * * All objects in the event handling chain should inherit from this class * */ var FORBIDDEN_EVENT_NAMES = { 'hlsEventGeneric': true, 'hlsHandlerDestroying': true, 'hlsHandlerDestroyed': true }; var event_handler_EventHandler = /*#__PURE__*/function () { function EventHandler(hls) { this.hls = void 0; this.handledEvents = void 0; this.useGenericHandler = void 0; this.hls = hls; this.onEvent = this.onEvent.bind(this); for (var _len = arguments.length, events = new Array(_len > 1 ? _len - 1 : 0), _key = 1; _key < _len; _key++) { events[_key - 1] = arguments[_key]; } this.handledEvents = events; this.useGenericHandler = true; this.registerListeners(); } var _proto = EventHandler.prototype; _proto.destroy = function destroy() { this.onHandlerDestroying(); this.unregisterListeners(); this.onHandlerDestroyed(); }; _proto.onHandlerDestroying = function onHandlerDestroying() {}; _proto.onHandlerDestroyed = function onHandlerDestroyed() {}; _proto.isEventHandler = function isEventHandler() { return typeof this.handledEvents === 'object' && this.handledEvents.length && typeof this.onEvent === 'function'; }; _proto.registerListeners = function registerListeners() { if (this.isEventHandler()) { this.handledEvents.forEach(function (event) { if (FORBIDDEN_EVENT_NAMES[event]) { throw new Error('Forbidden event-name: ' + event); } this.hls.on(event, this.onEvent); }, this); } }; _proto.unregisterListeners = function unregisterListeners() { if (this.isEventHandler()) { this.handledEvents.forEach(function (event) { this.hls.off(event, this.onEvent); }, this); } } /** * arguments: event (string), data (any) */ ; _proto.onEvent = function onEvent(event, data) { this.onEventGeneric(event, data); }; _proto.onEventGeneric = function onEventGeneric(event, data) { var eventToFunction = function eventToFunction(event, data) { var funcName = 'on' + event.replace('hls', ''); if (typeof this[funcName] !== 'function') { throw new Error("Event " + event + " has no generic handler in this " + this.constructor.name + " class (tried " + funcName + ")"); } return this[funcName].bind(this, data); }; try { eventToFunction.call(this, event, data).call(); } catch (err) { logger["logger"].error("An internal error happened while handling event " + event + ". Error message: \"" + err.message + "\". Here is a stacktrace:", err); this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].OTHER_ERROR, details: errors["ErrorDetails"].INTERNAL_EXCEPTION, fatal: false, event: event, err: err }); } }; return EventHandler; }(); /* harmony default export */ var event_handler = (event_handler_EventHandler); // CONCATENATED MODULE: ./src/types/loader.ts /** * `type` property values for this loaders' context object * @enum * */ var PlaylistContextType; /** * @enum {string} */ (function (PlaylistContextType) { PlaylistContextType["MANIFEST"] = "manifest"; PlaylistContextType["LEVEL"] = "level"; PlaylistContextType["AUDIO_TRACK"] = "audioTrack"; PlaylistContextType["SUBTITLE_TRACK"] = "subtitleTrack"; })(PlaylistContextType || (PlaylistContextType = {})); var PlaylistLevelType; (function (PlaylistLevelType) { PlaylistLevelType["MAIN"] = "main"; PlaylistLevelType["AUDIO"] = "audio"; PlaylistLevelType["SUBTITLE"] = "subtitle"; })(PlaylistLevelType || (PlaylistLevelType = {})); // EXTERNAL MODULE: ./src/demux/mp4demuxer.js var mp4demuxer = __webpack_require__("./src/demux/mp4demuxer.js"); // CONCATENATED MODULE: ./src/loader/level-key.ts function _defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } function _createClass(Constructor, protoProps, staticProps) { if (protoProps) _defineProperties(Constructor.prototype, protoProps); if (staticProps) _defineProperties(Constructor, staticProps); return Constructor; } var level_key_LevelKey = /*#__PURE__*/function () { function LevelKey(baseURI, relativeURI) { this._uri = null; this.baseuri = void 0; this.reluri = void 0; this.method = null; this.key = null; this.iv = null; this.baseuri = baseURI; this.reluri = relativeURI; } _createClass(LevelKey, [{ key: "uri", get: function get() { if (!this._uri && this.reluri) { this._uri = Object(url_toolkit["buildAbsoluteURL"])(this.baseuri, this.reluri, { alwaysNormalize: true }); } return this._uri; } }]); return LevelKey; }(); // CONCATENATED MODULE: ./src/loader/fragment.ts function fragment_defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } function fragment_createClass(Constructor, protoProps, staticProps) { if (protoProps) fragment_defineProperties(Constructor.prototype, protoProps); if (staticProps) fragment_defineProperties(Constructor, staticProps); return Constructor; } var ElementaryStreamTypes; (function (ElementaryStreamTypes) { ElementaryStreamTypes["AUDIO"] = "audio"; ElementaryStreamTypes["VIDEO"] = "video"; })(ElementaryStreamTypes || (ElementaryStreamTypes = {})); var fragment_Fragment = /*#__PURE__*/function () { function Fragment() { var _this$_elementaryStre; this._url = null; this._byteRange = null; this._decryptdata = null; this._elementaryStreams = (_this$_elementaryStre = {}, _this$_elementaryStre[ElementaryStreamTypes.AUDIO] = false, _this$_elementaryStre[ElementaryStreamTypes.VIDEO] = false, _this$_elementaryStre); this.deltaPTS = 0; this.rawProgramDateTime = null; this.programDateTime = null; this.title = null; this.tagList = []; this.cc = void 0; this.type = void 0; this.relurl = void 0; this.baseurl = void 0; this.duration = void 0; this.start = void 0; this.sn = 0; this.urlId = 0; this.level = 0; this.levelkey = void 0; this.loader = void 0; } var _proto = Fragment.prototype; // setByteRange converts a EXT-X-BYTERANGE attribute into a two element array _proto.setByteRange = function setByteRange(value, previousFrag) { var params = value.split('@', 2); var byteRange = []; if (params.length === 1) { byteRange[0] = previousFrag ? previousFrag.byteRangeEndOffset : 0; } else { byteRange[0] = parseInt(params[1]); } byteRange[1] = parseInt(params[0]) + byteRange[0]; this._byteRange = byteRange; }; /** * @param {ElementaryStreamTypes} type */ _proto.addElementaryStream = function addElementaryStream(type) { this._elementaryStreams[type] = true; } /** * @param {ElementaryStreamTypes} type */ ; _proto.hasElementaryStream = function hasElementaryStream(type) { return this._elementaryStreams[type] === true; } /** * Utility method for parseLevelPlaylist to create an initialization vector for a given segment * @param {number} segmentNumber - segment number to generate IV with * @returns {Uint8Array} */ ; _proto.createInitializationVector = function createInitializationVector(segmentNumber) { var uint8View = new Uint8Array(16); for (var i = 12; i < 16; i++) { uint8View[i] = segmentNumber >> 8 * (15 - i) & 0xff; } return uint8View; } /** * Utility method for parseLevelPlaylist to get a fragment's decryption data from the currently parsed encryption key data * @param levelkey - a playlist's encryption info * @param segmentNumber - the fragment's segment number * @returns {LevelKey} - an object to be applied as a fragment's decryptdata */ ; _proto.setDecryptDataFromLevelKey = function setDecryptDataFromLevelKey(levelkey, segmentNumber) { var decryptdata = levelkey; if ((levelkey === null || levelkey === void 0 ? void 0 : levelkey.method) && levelkey.uri && !levelkey.iv) { decryptdata = new level_key_LevelKey(levelkey.baseuri, levelkey.reluri); decryptdata.method = levelkey.method; decryptdata.iv = this.createInitializationVector(segmentNumber); } return decryptdata; }; fragment_createClass(Fragment, [{ key: "url", get: function get() { if (!this._url && this.relurl) { this._url = Object(url_toolkit["buildAbsoluteURL"])(this.baseurl, this.relurl, { alwaysNormalize: true }); } return this._url; }, set: function set(value) { this._url = value; } }, { key: "byteRange", get: function get() { if (!this._byteRange) { return []; } return this._byteRange; } /** * @type {number} */ }, { key: "byteRangeStartOffset", get: function get() { return this.byteRange[0]; } }, { key: "byteRangeEndOffset", get: function get() { return this.byteRange[1]; } }, { key: "decryptdata", get: function get() { if (!this.levelkey && !this._decryptdata) { return null; } if (!this._decryptdata && this.levelkey) { var sn = this.sn; if (typeof sn !== 'number') { // We are fetching decryption data for a initialization segment // If the segment was encrypted with AES-128 // It must have an IV defined. We cannot substitute the Segment Number in. if (this.levelkey && this.levelkey.method === 'AES-128' && !this.levelkey.iv) { logger["logger"].warn("missing IV for initialization segment with method=\"" + this.levelkey.method + "\" - compliance issue"); } /* Be converted to a Number. 'initSegment' will become NaN. NaN, which when converted through ToInt32() -> +0. --- Explicitly set sn to resulting value from implicit conversions 'initSegment' values for IV generation. */ sn = 0; } this._decryptdata = this.setDecryptDataFromLevelKey(this.levelkey, sn); } return this._decryptdata; } }, { key: "endProgramDateTime", get: function get() { if (this.programDateTime === null) { return null; } if (!Object(number["isFiniteNumber"])(this.programDateTime)) { return null; } var duration = !Object(number["isFiniteNumber"])(this.duration) ? 0 : this.duration; return this.programDateTime + duration * 1000; } }, { key: "encrypted", get: function get() { return !!(this.decryptdata && this.decryptdata.uri !== null && this.decryptdata.key === null); } }]); return Fragment; }(); // CONCATENATED MODULE: ./src/loader/level.js function level_defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } function level_createClass(Constructor, protoProps, staticProps) { if (protoProps) level_defineProperties(Constructor.prototype, protoProps); if (staticProps) level_defineProperties(Constructor, staticProps); return Constructor; } var level_Level = /*#__PURE__*/function () { function Level(baseUrl) { // Please keep properties in alphabetical order this.endCC = 0; this.endSN = 0; this.fragments = []; this.initSegment = null; this.live = true; this.needSidxRanges = false; this.startCC = 0; this.startSN = 0; this.startTimeOffset = null; this.targetduration = 0; this.totalduration = 0; this.type = null; this.url = baseUrl; this.version = null; } level_createClass(Level, [{ key: "hasProgramDateTime", get: function get() { return !!(this.fragments[0] && Object(number["isFiniteNumber"])(this.fragments[0].programDateTime)); } }]); return Level; }(); // CONCATENATED MODULE: ./src/utils/attr-list.js var DECIMAL_RESOLUTION_REGEX = /^(\d+)x(\d+)$/; // eslint-disable-line no-useless-escape var ATTR_LIST_REGEX = /\s*(.+?)\s*=((?:\".*?\")|.*?)(?:,|$)/g; // eslint-disable-line no-useless-escape // adapted from https://github.com/kanongil/node-m3u8parse/blob/master/attrlist.js var AttrList = /*#__PURE__*/function () { function AttrList(attrs) { if (typeof attrs === 'string') { attrs = AttrList.parseAttrList(attrs); } for (var attr in attrs) { if (attrs.hasOwnProperty(attr)) { this[attr] = attrs[attr]; } } } var _proto = AttrList.prototype; _proto.decimalInteger = function decimalInteger(attrName) { var intValue = parseInt(this[attrName], 10); if (intValue > Number.MAX_SAFE_INTEGER) { return Infinity; } return intValue; }; _proto.hexadecimalInteger = function hexadecimalInteger(attrName) { if (this[attrName]) { var stringValue = (this[attrName] || '0x').slice(2); stringValue = (stringValue.length & 1 ? '0' : '') + stringValue; var value = new Uint8Array(stringValue.length / 2); for (var i = 0; i < stringValue.length / 2; i++) { value[i] = parseInt(stringValue.slice(i * 2, i * 2 + 2), 16); } return value; } else { return null; } }; _proto.hexadecimalIntegerAsNumber = function hexadecimalIntegerAsNumber(attrName) { var intValue = parseInt(this[attrName], 16); if (intValue > Number.MAX_SAFE_INTEGER) { return Infinity; } return intValue; }; _proto.decimalFloatingPoint = function decimalFloatingPoint(attrName) { return parseFloat(this[attrName]); }; _proto.enumeratedString = function enumeratedString(attrName) { return this[attrName]; }; _proto.decimalResolution = function decimalResolution(attrName) { var res = DECIMAL_RESOLUTION_REGEX.exec(this[attrName]); if (res === null) { return undefined; } return { width: parseInt(res[1], 10), height: parseInt(res[2], 10) }; }; AttrList.parseAttrList = function parseAttrList(input) { var match, attrs = {}; ATTR_LIST_REGEX.lastIndex = 0; while ((match = ATTR_LIST_REGEX.exec(input)) !== null) { var value = match[2], quote = '"'; if (value.indexOf(quote) === 0 && value.lastIndexOf(quote) === value.length - 1) { value = value.slice(1, -1); } attrs[match[1]] = value; } return attrs; }; return AttrList; }(); /* harmony default export */ var attr_list = (AttrList); // CONCATENATED MODULE: ./src/utils/codecs.ts // from http://mp4ra.org/codecs.html var sampleEntryCodesISO = { audio: { 'a3ds': true, 'ac-3': true, 'ac-4': true, 'alac': true, 'alaw': true, 'dra1': true, 'dts+': true, 'dts-': true, 'dtsc': true, 'dtse': true, 'dtsh': true, 'ec-3': true, 'enca': true, 'g719': true, 'g726': true, 'm4ae': true, 'mha1': true, 'mha2': true, 'mhm1': true, 'mhm2': true, 'mlpa': true, 'mp4a': true, 'raw ': true, 'Opus': true, 'samr': true, 'sawb': true, 'sawp': true, 'sevc': true, 'sqcp': true, 'ssmv': true, 'twos': true, 'ulaw': true }, video: { 'avc1': true, 'avc2': true, 'avc3': true, 'avc4': true, 'avcp': true, 'drac': true, 'dvav': true, 'dvhe': true, 'encv': true, 'hev1': true, 'hvc1': true, 'mjp2': true, 'mp4v': true, 'mvc1': true, 'mvc2': true, 'mvc3': true, 'mvc4': true, 'resv': true, 'rv60': true, 's263': true, 'svc1': true, 'svc2': true, 'vc-1': true, 'vp08': true, 'vp09': true } }; function isCodecType(codec, type) { var typeCodes = sampleEntryCodesISO[type]; return !!typeCodes && typeCodes[codec.slice(0, 4)] === true; } function isCodecSupportedInMp4(codec, type) { return MediaSource.isTypeSupported((type || 'video') + "/mp4;codecs=\"" + codec + "\""); } // CONCATENATED MODULE: ./src/loader/m3u8-parser.ts /** * M3U8 parser * @module */ // https://regex101.com is your friend var MASTER_PLAYLIST_REGEX = /(?:#EXT-X-STREAM-INF:([^\n\r]*)[\r\n]+([^\r\n]+)|#EXT-X-SESSION-DATA:([^\n\r]*)[\r\n]+)/g; var MASTER_PLAYLIST_MEDIA_REGEX = /#EXT-X-MEDIA:(.*)/g; var LEVEL_PLAYLIST_REGEX_FAST = new RegExp([/#EXTINF:\s*(\d*(?:\.\d+)?)(?:,(.*)\s+)?/.source, // duration (#EXTINF:<duration>,<title>), group 1 => duration, group 2 => title /|(?!#)([\S+ ?]+)/.source, // segment URI, group 3 => the URI (note newline is not eaten) /|#EXT-X-BYTERANGE:*(.+)/.source, // next segment's byterange, group 4 => range spec (x@y) /|#EXT-X-PROGRAM-DATE-TIME:(.+)/.source, // next segment's program date/time group 5 => the datetime spec /|#.*/.source // All other non-segment oriented tags will match with all groups empty ].join(''), 'g'); var LEVEL_PLAYLIST_REGEX_SLOW = /(?:(?:#(EXTM3U))|(?:#EXT-X-(PLAYLIST-TYPE):(.+))|(?:#EXT-X-(MEDIA-SEQUENCE): *(\d+))|(?:#EXT-X-(TARGETDURATION): *(\d+))|(?:#EXT-X-(KEY):(.+))|(?:#EXT-X-(START):(.+))|(?:#EXT-X-(ENDLIST))|(?:#EXT-X-(DISCONTINUITY-SEQ)UENCE:(\d+))|(?:#EXT-X-(DIS)CONTINUITY))|(?:#EXT-X-(VERSION):(\d+))|(?:#EXT-X-(MAP):(.+))|(?:(#)([^:]*):(.*))|(?:(#)(.*))(?:.*)\r?\n?/; var MP4_REGEX_SUFFIX = /\.(mp4|m4s|m4v|m4a)$/i; var m3u8_parser_M3U8Parser = /*#__PURE__*/function () { function M3U8Parser() {} M3U8Parser.findGroup = function findGroup(groups, mediaGroupId) { for (var i = 0; i < groups.length; i++) { var group = groups[i]; if (group.id === mediaGroupId) { return group; } } }; M3U8Parser.convertAVC1ToAVCOTI = function convertAVC1ToAVCOTI(codec) { var avcdata = codec.split('.'); var result; if (avcdata.length > 2) { result = avcdata.shift() + '.'; result += parseInt(avcdata.shift()).toString(16); result += ('000' + parseInt(avcdata.shift()).toString(16)).substr(-4); } else { result = codec; } return result; }; M3U8Parser.resolve = function resolve(url, baseUrl) { return url_toolkit["buildAbsoluteURL"](baseUrl, url, { alwaysNormalize: true }); }; M3U8Parser.parseMasterPlaylist = function parseMasterPlaylist(string, baseurl) { // TODO(typescript-level) var levels = []; var sessionData = {}; var hasSessionData = false; MASTER_PLAYLIST_REGEX.lastIndex = 0; // TODO(typescript-level) function setCodecs(codecs, level) { ['video', 'audio'].forEach(function (type) { var filtered = codecs.filter(function (codec) { return isCodecType(codec, type); }); if (filtered.length) { var preferred = filtered.filter(function (codec) { return codec.lastIndexOf('avc1', 0) === 0 || codec.lastIndexOf('mp4a', 0) === 0; }); level[type + "Codec"] = preferred.length > 0 ? preferred[0] : filtered[0]; // remove from list codecs = codecs.filter(function (codec) { return filtered.indexOf(codec) === -1; }); } }); level.unknownCodecs = codecs; } var result; while ((result = MASTER_PLAYLIST_REGEX.exec(string)) != null) { if (result[1]) { // '#EXT-X-STREAM-INF' is found, parse level tag in group 1 // TODO(typescript-level) var level = {}; var attrs = level.attrs = new attr_list(result[1]); level.url = M3U8Parser.resolve(result[2], baseurl); var resolution = attrs.decimalResolution('RESOLUTION'); if (resolution) { level.width = resolution.width; level.height = resolution.height; } level.bitrate = attrs.decimalInteger('AVERAGE-BANDWIDTH') || attrs.decimalInteger('BANDWIDTH'); level.name = attrs.NAME; setCodecs([].concat((attrs.CODECS || '').split(/[ ,]+/)), level); if (level.videoCodec && level.videoCodec.indexOf('avc1') !== -1) { level.videoCodec = M3U8Parser.convertAVC1ToAVCOTI(level.videoCodec); } levels.push(level); } else if (result[3]) { // '#EXT-X-SESSION-DATA' is found, parse session data in group 3 var sessionAttrs = new attr_list(result[3]); if (sessionAttrs['DATA-ID']) { hasSessionData = true; sessionData[sessionAttrs['DATA-ID']] = sessionAttrs; } } } return { levels: levels, sessionData: hasSessionData ? sessionData : null }; }; M3U8Parser.parseMasterPlaylistMedia = function parseMasterPlaylistMedia(string, baseurl, type, audioGroups) { if (audioGroups === void 0) { audioGroups = []; } var result; var medias = []; var id = 0; MASTER_PLAYLIST_MEDIA_REGEX.lastIndex = 0; while ((result = MASTER_PLAYLIST_MEDIA_REGEX.exec(string)) !== null) { var attrs = new attr_list(result[1]); if (attrs.TYPE === type) { var media = { attrs: attrs, id: id++, groupId: attrs['GROUP-ID'], instreamId: attrs['INSTREAM-ID'], name: attrs.NAME || attrs.LANGUAGE, type: type, default: attrs.DEFAULT === 'YES', autoselect: attrs.AUTOSELECT === 'YES', forced: attrs.FORCED === 'YES', lang: attrs.LANGUAGE }; if (attrs.URI) { media.url = M3U8Parser.resolve(attrs.URI, baseurl); } if (audioGroups.length) { // If there are audio groups signalled in the manifest, let's look for a matching codec string for this track var groupCodec = M3U8Parser.findGroup(audioGroups, media.groupId); // If we don't find the track signalled, lets use the first audio groups codec we have // Acting as a best guess media.audioCodec = groupCodec ? groupCodec.codec : audioGroups[0].codec; } medias.push(media); } } return medias; }; M3U8Parser.parseLevelPlaylist = function parseLevelPlaylist(string, baseurl, id, type, levelUrlId) { var currentSN = 0; var totalduration = 0; var level = new level_Level(baseurl); var discontinuityCounter = 0; var prevFrag = null; var frag = new fragment_Fragment(); var result; var i; var levelkey; var firstPdtIndex = null; LEVEL_PLAYLIST_REGEX_FAST.lastIndex = 0; while ((result = LEVEL_PLAYLIST_REGEX_FAST.exec(string)) !== null) { var duration = result[1]; if (duration) { // INF frag.duration = parseFloat(duration); // avoid sliced strings https://github.com/video-dev/hls.js/issues/939 var title = (' ' + result[2]).slice(1); frag.title = title || null; frag.tagList.push(title ? ['INF', duration, title] : ['INF', duration]); } else if (result[3]) { // url if (Object(number["isFiniteNumber"])(frag.duration)) { var sn = currentSN++; frag.type = type; frag.start = totalduration; if (levelkey) { frag.levelkey = levelkey; } frag.sn = sn; frag.level = id; frag.cc = discontinuityCounter; frag.urlId = levelUrlId; frag.baseurl = baseurl; // avoid sliced strings https://github.com/video-dev/hls.js/issues/939 frag.relurl = (' ' + result[3]).slice(1); assignProgramDateTime(frag, prevFrag); level.fragments.push(frag); prevFrag = frag; totalduration += frag.duration; frag = new fragment_Fragment(); } } else if (result[4]) { // X-BYTERANGE var data = (' ' + result[4]).slice(1); if (prevFrag) { frag.setByteRange(data, prevFrag); } else { frag.setByteRange(data); } } else if (result[5]) { // PROGRAM-DATE-TIME // avoid sliced strings https://github.com/video-dev/hls.js/issues/939 frag.rawProgramDateTime = (' ' + result[5]).slice(1); frag.tagList.push(['PROGRAM-DATE-TIME', frag.rawProgramDateTime]); if (firstPdtIndex === null) { firstPdtIndex = level.fragments.length; } } else { result = result[0].match(LEVEL_PLAYLIST_REGEX_SLOW); if (!result) { logger["logger"].warn('No matches on slow regex match for level playlist!'); continue; } for (i = 1; i < result.length; i++) { if (typeof result[i] !== 'undefined') { break; } } // avoid sliced strings https://github.com/video-dev/hls.js/issues/939 var value1 = (' ' + result[i + 1]).slice(1); var value2 = (' ' + result[i + 2]).slice(1); switch (result[i]) { case '#': frag.tagList.push(value2 ? [value1, value2] : [value1]); break; case 'PLAYLIST-TYPE': level.type = value1.toUpperCase(); break; case 'MEDIA-SEQUENCE': currentSN = level.startSN = parseInt(value1); break; case 'TARGETDURATION': level.targetduration = parseFloat(value1); break; case 'VERSION': level.version = parseInt(value1); break; case 'EXTM3U': break; case 'ENDLIST': level.live = false; break; case 'DIS': discontinuityCounter++; frag.tagList.push(['DIS']); break; case 'DISCONTINUITY-SEQ': discontinuityCounter = parseInt(value1); break; case 'KEY': { // https://tools.ietf.org/html/rfc8216#section-4.3.2.4 var decryptparams = value1; var keyAttrs = new attr_list(decryptparams); var decryptmethod = keyAttrs.enumeratedString('METHOD'); var decrypturi = keyAttrs.URI; var decryptiv = keyAttrs.hexadecimalInteger('IV'); // From RFC: This attribute is OPTIONAL; its absence indicates an implicit value of "identity". var decryptkeyformat = keyAttrs.KEYFORMAT || 'identity'; if (decryptkeyformat === 'com.apple.streamingkeydelivery') { logger["logger"].warn('Keyformat com.apple.streamingkeydelivery is not supported'); continue; } if (decryptmethod) { levelkey = new level_key_LevelKey(baseurl, decrypturi); if (decrypturi && ['AES-128', 'SAMPLE-AES', 'SAMPLE-AES-CENC'].indexOf(decryptmethod) >= 0) { levelkey.method = decryptmethod; levelkey.key = null; // Initialization Vector (IV) levelkey.iv = decryptiv; } } break; } case 'START': { var startAttrs = new attr_list(value1); var startTimeOffset = startAttrs.decimalFloatingPoint('TIME-OFFSET'); // TIME-OFFSET can be 0 if (Object(number["isFiniteNumber"])(startTimeOffset)) { level.startTimeOffset = startTimeOffset; } break; } case 'MAP': { var mapAttrs = new attr_list(value1); frag.relurl = mapAttrs.URI; if (mapAttrs.BYTERANGE) { frag.setByteRange(mapAttrs.BYTERANGE); } frag.baseurl = baseurl; frag.level = id; frag.type = type; frag.sn = 'initSegment'; level.initSegment = frag; frag = new fragment_Fragment(); frag.rawProgramDateTime = level.initSegment.rawProgramDateTime; break; } default: logger["logger"].warn("line parsed but not handled: " + result); break; } } } frag = prevFrag; // logger.log('found ' + level.fragments.length + ' fragments'); if (frag && !frag.relurl) { level.fragments.pop(); totalduration -= frag.duration; } level.totalduration = totalduration; level.averagetargetduration = totalduration / level.fragments.length; level.endSN = currentSN - 1; level.startCC = level.fragments[0] ? level.fragments[0].cc : 0; level.endCC = discontinuityCounter; if (!level.initSegment && level.fragments.length) { // this is a bit lurky but HLS really has no other way to tell us // if the fragments are TS or MP4, except if we download them :/ // but this is to be able to handle SIDX. if (level.fragments.every(function (frag) { return MP4_REGEX_SUFFIX.test(frag.relurl); })) { logger["logger"].warn('MP4 fragments found but no init segment (probably no MAP, incomplete M3U8), trying to fetch SIDX'); frag = new fragment_Fragment(); frag.relurl = level.fragments[0].relurl; frag.baseurl = baseurl; frag.level = id; frag.type = type; frag.sn = 'initSegment'; level.initSegment = frag; level.needSidxRanges = true; } } /** * Backfill any missing PDT values "If the first EXT-X-PROGRAM-DATE-TIME tag in a Playlist appears after one or more Media Segment URIs, the client SHOULD extrapolate backward from that tag (using EXTINF durations and/or media timestamps) to associate dates with those segments." * We have already extrapolated forward, but all fragments up to the first instance of PDT do not have their PDTs * computed. */ if (firstPdtIndex) { backfillProgramDateTimes(level.fragments, firstPdtIndex); } return level; }; return M3U8Parser; }(); function backfillProgramDateTimes(fragments, startIndex) { var fragPrev = fragments[startIndex]; for (var i = startIndex - 1; i >= 0; i--) { var frag = fragments[i]; frag.programDateTime = fragPrev.programDateTime - frag.duration * 1000; fragPrev = frag; } } function assignProgramDateTime(frag, prevFrag) { if (frag.rawProgramDateTime) { frag.programDateTime = Date.parse(frag.rawProgramDateTime); } else if (prevFrag === null || prevFrag === void 0 ? void 0 : prevFrag.programDateTime) { frag.programDateTime = prevFrag.endProgramDateTime; } if (!Object(number["isFiniteNumber"])(frag.programDateTime)) { frag.programDateTime = null; frag.rawProgramDateTime = null; } } // CONCATENATED MODULE: ./src/loader/playlist-loader.ts function _inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /** * PlaylistLoader - delegate for media manifest/playlist loading tasks. Takes care of parsing media to internal data-models. * * Once loaded, dispatches events with parsed data-models of manifest/levels/audio/subtitle tracks. * * Uses loader(s) set in config to do actual internal loading of resource tasks. * * @module * */ var _window = window, performance = _window.performance; /** * @constructor */ var playlist_loader_PlaylistLoader = /*#__PURE__*/function (_EventHandler) { _inheritsLoose(PlaylistLoader, _EventHandler); /** * @constructs * @param {Hls} hls */ function PlaylistLoader(hls) { var _this; _this = _EventHandler.call(this, hls, events["default"].MANIFEST_LOADING, events["default"].LEVEL_LOADING, events["default"].AUDIO_TRACK_LOADING, events["default"].SUBTITLE_TRACK_LOADING) || this; _this.loaders = {}; return _this; } /** * @param {PlaylistContextType} type * @returns {boolean} */ PlaylistLoader.canHaveQualityLevels = function canHaveQualityLevels(type) { return type !== PlaylistContextType.AUDIO_TRACK && type !== PlaylistContextType.SUBTITLE_TRACK; } /** * Map context.type to LevelType * @param {PlaylistLoaderContext} context * @returns {LevelType} */ ; PlaylistLoader.mapContextToLevelType = function mapContextToLevelType(context) { var type = context.type; switch (type) { case PlaylistContextType.AUDIO_TRACK: return PlaylistLevelType.AUDIO; case PlaylistContextType.SUBTITLE_TRACK: return PlaylistLevelType.SUBTITLE; default: return PlaylistLevelType.MAIN; } }; PlaylistLoader.getResponseUrl = function getResponseUrl(response, context) { var url = response.url; // responseURL not supported on some browsers (it is used to detect URL redirection) // data-uri mode also not supported (but no need to detect redirection) if (url === undefined || url.indexOf('data:') === 0) { // fallback to initial URL url = context.url; } return url; } /** * Returns defaults or configured loader-type overloads (pLoader and loader config params) * Default loader is XHRLoader (see utils) * @param {PlaylistLoaderContext} context * @returns {Loader} or other compatible configured overload */ ; var _proto = PlaylistLoader.prototype; _proto.createInternalLoader = function createInternalLoader(context) { var config = this.hls.config; var PLoader = config.pLoader; var Loader = config.loader; // TODO(typescript-config): Verify once config is typed that InternalLoader always returns a Loader var InternalLoader = PLoader || Loader; var loader = new InternalLoader(config); // TODO - Do we really need to assign the instance or if the dep has been lost context.loader = loader; this.loaders[context.type] = loader; return loader; }; _proto.getInternalLoader = function getInternalLoader(context) { return this.loaders[context.type]; }; _proto.resetInternalLoader = function resetInternalLoader(contextType) { if (this.loaders[contextType]) { delete this.loaders[contextType]; } } /** * Call `destroy` on all internal loader instances mapped (one per context type) */ ; _proto.destroyInternalLoaders = function destroyInternalLoaders() { for (var contextType in this.loaders) { var loader = this.loaders[contextType]; if (loader) { loader.destroy(); } this.resetInternalLoader(contextType); } }; _proto.destroy = function destroy() { this.destroyInternalLoaders(); _EventHandler.prototype.destroy.call(this); }; _proto.onManifestLoading = function onManifestLoading(data) { this.load({ url: data.url, type: PlaylistContextType.MANIFEST, level: 0, id: null, responseType: 'text' }); }; _proto.onLevelLoading = function onLevelLoading(data) { this.load({ url: data.url, type: PlaylistContextType.LEVEL, level: data.level, id: data.id, responseType: 'text' }); }; _proto.onAudioTrackLoading = function onAudioTrackLoading(data) { this.load({ url: data.url, type: PlaylistContextType.AUDIO_TRACK, level: null, id: data.id, responseType: 'text' }); }; _proto.onSubtitleTrackLoading = function onSubtitleTrackLoading(data) { this.load({ url: data.url, type: PlaylistContextType.SUBTITLE_TRACK, level: null, id: data.id, responseType: 'text' }); }; _proto.load = function load(context) { var config = this.hls.config; logger["logger"].debug("Loading playlist of type " + context.type + ", level: " + context.level + ", id: " + context.id); // Check if a loader for this context already exists var loader = this.getInternalLoader(context); if (loader) { var loaderContext = loader.context; if (loaderContext && loaderContext.url === context.url) { // same URL can't overlap logger["logger"].trace('playlist request ongoing'); return false; } else { logger["logger"].warn("aborting previous loader for type: " + context.type); loader.abort(); } } var maxRetry; var timeout; var retryDelay; var maxRetryDelay; // apply different configs for retries depending on // context (manifest, level, audio/subs playlist) switch (context.type) { case PlaylistContextType.MANIFEST: maxRetry = config.manifestLoadingMaxRetry; timeout = config.manifestLoadingTimeOut; retryDelay = config.manifestLoadingRetryDelay; maxRetryDelay = config.manifestLoadingMaxRetryTimeout; break; case PlaylistContextType.LEVEL: // Disable internal loader retry logic, since we are managing retries in Level Controller maxRetry = 0; maxRetryDelay = 0; retryDelay = 0; timeout = config.levelLoadingTimeOut; // TODO Introduce retry settings for audio-track and subtitle-track, it should not use level retry config break; default: maxRetry = config.levelLoadingMaxRetry; timeout = config.levelLoadingTimeOut; retryDelay = config.levelLoadingRetryDelay; maxRetryDelay = config.levelLoadingMaxRetryTimeout; break; } loader = this.createInternalLoader(context); var loaderConfig = { timeout: timeout, maxRetry: maxRetry, retryDelay: retryDelay, maxRetryDelay: maxRetryDelay }; var loaderCallbacks = { onSuccess: this.loadsuccess.bind(this), onError: this.loaderror.bind(this), onTimeout: this.loadtimeout.bind(this) }; logger["logger"].debug("Calling internal loader delegate for URL: " + context.url); loader.load(context, loaderConfig, loaderCallbacks); return true; }; _proto.loadsuccess = function loadsuccess(response, stats, context, networkDetails) { if (networkDetails === void 0) { networkDetails = null; } if (context.isSidxRequest) { this._handleSidxRequest(response, context); this._handlePlaylistLoaded(response, stats, context, networkDetails); return; } this.resetInternalLoader(context.type); if (typeof response.data !== 'string') { throw new Error('expected responseType of "text" for PlaylistLoader'); } var string = response.data; stats.tload = performance.now(); // stats.mtime = new Date(target.getResponseHeader('Last-Modified')); // Validate if it is an M3U8 at all if (string.indexOf('#EXTM3U') !== 0) { this._handleManifestParsingError(response, context, 'no EXTM3U delimiter', networkDetails); return; } // Check if chunk-list or master. handle empty chunk list case (first EXTINF not signaled, but TARGETDURATION present) if (string.indexOf('#EXTINF:') > 0 || string.indexOf('#EXT-X-TARGETDURATION:') > 0) { this._handleTrackOrLevelPlaylist(response, stats, context, networkDetails); } else { this._handleMasterPlaylist(response, stats, context, networkDetails); } }; _proto.loaderror = function loaderror(response, context, networkDetails) { if (networkDetails === void 0) { networkDetails = null; } this._handleNetworkError(context, networkDetails, false, response); }; _proto.loadtimeout = function loadtimeout(stats, context, networkDetails) { if (networkDetails === void 0) { networkDetails = null; } this._handleNetworkError(context, networkDetails, true); } // TODO(typescript-config): networkDetails can currently be a XHR or Fetch impl, // but with custom loaders it could be generic investigate this further when config is typed ; _proto._handleMasterPlaylist = function _handleMasterPlaylist(response, stats, context, networkDetails) { var hls = this.hls; var string = response.data; var url = PlaylistLoader.getResponseUrl(response, context); var _M3U8Parser$parseMast = m3u8_parser_M3U8Parser.parseMasterPlaylist(string, url), levels = _M3U8Parser$parseMast.levels, sessionData = _M3U8Parser$parseMast.sessionData; if (!levels.length) { this._handleManifestParsingError(response, context, 'no level found in manifest', networkDetails); return; } // multi level playlist, parse level info var audioGroups = levels.map(function (level) { return { id: level.attrs.AUDIO, codec: level.audioCodec }; }); var audioTracks = m3u8_parser_M3U8Parser.parseMasterPlaylistMedia(string, url, 'AUDIO', audioGroups); var subtitles = m3u8_parser_M3U8Parser.parseMasterPlaylistMedia(string, url, 'SUBTITLES'); var captions = m3u8_parser_M3U8Parser.parseMasterPlaylistMedia(string, url, 'CLOSED-CAPTIONS'); if (audioTracks.length) { // check if we have found an audio track embedded in main playlist (audio track without URI attribute) var embeddedAudioFound = false; audioTracks.forEach(function (audioTrack) { if (!audioTrack.url) { embeddedAudioFound = true; } }); // if no embedded audio track defined, but audio codec signaled in quality level, // we need to signal this main audio track this could happen with playlists with // alt audio rendition in which quality levels (main) // contains both audio+video. but with mixed audio track not signaled if (embeddedAudioFound === false && levels[0].audioCodec && !levels[0].attrs.AUDIO) { logger["logger"].log('audio codec signaled in quality level, but no embedded audio track signaled, create one'); audioTracks.unshift({ type: 'main', name: 'main', default: false, autoselect: false, forced: false, id: -1, attrs: {}, url: '' }); } } hls.trigger(events["default"].MANIFEST_LOADED, { levels: levels, audioTracks: audioTracks, subtitles: subtitles, captions: captions, url: url, stats: stats, networkDetails: networkDetails, sessionData: sessionData }); }; _proto._handleTrackOrLevelPlaylist = function _handleTrackOrLevelPlaylist(response, stats, context, networkDetails) { var hls = this.hls; var id = context.id, level = context.level, type = context.type; var url = PlaylistLoader.getResponseUrl(response, context); // if the values are null, they will result in the else conditional var levelUrlId = Object(number["isFiniteNumber"])(id) ? id : 0; var levelId = Object(number["isFiniteNumber"])(level) ? level : levelUrlId; var levelType = PlaylistLoader.mapContextToLevelType(context); var levelDetails = m3u8_parser_M3U8Parser.parseLevelPlaylist(response.data, url, levelId, levelType, levelUrlId); // set stats on level structure // TODO(jstackhouse): why? mixing concerns, is it just treated as value bag? levelDetails.tload = stats.tload; if (!levelDetails.fragments.length) { hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].NETWORK_ERROR, details: errors["ErrorDetails"].LEVEL_EMPTY_ERROR, fatal: false, url: url, reason: 'no fragments found in level', level: typeof context.level === 'number' ? context.level : undefined }); return; } // We have done our first request (Manifest-type) and receive // not a master playlist but a chunk-list (track/level) // We fire the manifest-loaded event anyway with the parsed level-details // by creating a single-level structure for it. if (type === PlaylistContextType.MANIFEST) { var singleLevel = { url: url, details: levelDetails }; hls.trigger(events["default"].MANIFEST_LOADED, { levels: [singleLevel], audioTracks: [], url: url, stats: stats, networkDetails: networkDetails, sessionData: null }); } // save parsing time stats.tparsed = performance.now(); // in case we need SIDX ranges // return early after calling load for // the SIDX box. if (levelDetails.needSidxRanges) { var sidxUrl = levelDetails.initSegment.url; this.load({ url: sidxUrl, isSidxRequest: true, type: type, level: level, levelDetails: levelDetails, id: id, rangeStart: 0, rangeEnd: 2048, responseType: 'arraybuffer' }); return; } // extend the context with the new levelDetails property context.levelDetails = levelDetails; this._handlePlaylistLoaded(response, stats, context, networkDetails); }; _proto._handleSidxRequest = function _handleSidxRequest(response, context) { if (typeof response.data === 'string') { throw new Error('sidx request must be made with responseType of array buffer'); } var sidxInfo = mp4demuxer["default"].parseSegmentIndex(new Uint8Array(response.data)); // if provided fragment does not contain sidx, early return if (!sidxInfo) { return; } var sidxReferences = sidxInfo.references; var levelDetails = context.levelDetails; sidxReferences.forEach(function (segmentRef, index) { var segRefInfo = segmentRef.info; if (!levelDetails) { return; } var frag = levelDetails.fragments[index]; if (frag.byteRange.length === 0) { frag.setByteRange(String(1 + segRefInfo.end - segRefInfo.start) + '@' + String(segRefInfo.start)); } }); if (levelDetails) { levelDetails.initSegment.setByteRange(String(sidxInfo.moovEndOffset) + '@0'); } }; _proto._handleManifestParsingError = function _handleManifestParsingError(response, context, reason, networkDetails) { this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].NETWORK_ERROR, details: errors["ErrorDetails"].MANIFEST_PARSING_ERROR, fatal: true, url: response.url, reason: reason, networkDetails: networkDetails }); }; _proto._handleNetworkError = function _handleNetworkError(context, networkDetails, timeout, response) { if (timeout === void 0) { timeout = false; } if (response === void 0) { response = null; } logger["logger"].info("A network error occured while loading a " + context.type + "-type playlist"); var details; var fatal; var loader = this.getInternalLoader(context); switch (context.type) { case PlaylistContextType.MANIFEST: details = timeout ? errors["ErrorDetails"].MANIFEST_LOAD_TIMEOUT : errors["ErrorDetails"].MANIFEST_LOAD_ERROR; fatal = true; break; case PlaylistContextType.LEVEL: details = timeout ? errors["ErrorDetails"].LEVEL_LOAD_TIMEOUT : errors["ErrorDetails"].LEVEL_LOAD_ERROR; fatal = false; break; case PlaylistContextType.AUDIO_TRACK: details = timeout ? errors["ErrorDetails"].AUDIO_TRACK_LOAD_TIMEOUT : errors["ErrorDetails"].AUDIO_TRACK_LOAD_ERROR; fatal = false; break; default: // details = ...? fatal = false; } if (loader) { loader.abort(); this.resetInternalLoader(context.type); } // TODO(typescript-events): when error events are handled, type this var errorData = { type: errors["ErrorTypes"].NETWORK_ERROR, details: details, fatal: fatal, url: context.url, loader: loader, context: context, networkDetails: networkDetails }; if (response) { errorData.response = response; } this.hls.trigger(events["default"].ERROR, errorData); }; _proto._handlePlaylistLoaded = function _handlePlaylistLoaded(response, stats, context, networkDetails) { var type = context.type, level = context.level, id = context.id, levelDetails = context.levelDetails; if (!levelDetails || !levelDetails.targetduration) { this._handleManifestParsingError(response, context, 'invalid target duration', networkDetails); return; } var canHaveLevels = PlaylistLoader.canHaveQualityLevels(context.type); if (canHaveLevels) { this.hls.trigger(events["default"].LEVEL_LOADED, { details: levelDetails, level: level || 0, id: id || 0, stats: stats, networkDetails: networkDetails }); } else { switch (type) { case PlaylistContextType.AUDIO_TRACK: this.hls.trigger(events["default"].AUDIO_TRACK_LOADED, { details: levelDetails, id: id, stats: stats, networkDetails: networkDetails }); break; case PlaylistContextType.SUBTITLE_TRACK: this.hls.trigger(events["default"].SUBTITLE_TRACK_LOADED, { details: levelDetails, id: id, stats: stats, networkDetails: networkDetails }); break; } } }; return PlaylistLoader; }(event_handler); /* harmony default export */ var playlist_loader = (playlist_loader_PlaylistLoader); // CONCATENATED MODULE: ./src/loader/fragment-loader.js function fragment_loader_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /* * Fragment Loader */ var fragment_loader_FragmentLoader = /*#__PURE__*/function (_EventHandler) { fragment_loader_inheritsLoose(FragmentLoader, _EventHandler); function FragmentLoader(hls) { var _this; _this = _EventHandler.call(this, hls, events["default"].FRAG_LOADING) || this; _this.loaders = {}; return _this; } var _proto = FragmentLoader.prototype; _proto.destroy = function destroy() { var loaders = this.loaders; for (var loaderName in loaders) { var loader = loaders[loaderName]; if (loader) { loader.destroy(); } } this.loaders = {}; _EventHandler.prototype.destroy.call(this); }; _proto.onFragLoading = function onFragLoading(data) { var frag = data.frag, type = frag.type, loaders = this.loaders, config = this.hls.config, FragmentILoader = config.fLoader, DefaultILoader = config.loader; // reset fragment state frag.loaded = 0; var loader = loaders[type]; if (loader) { logger["logger"].warn("abort previous fragment loader for type: " + type); loader.abort(); } loader = loaders[type] = frag.loader = config.fLoader ? new FragmentILoader(config) : new DefaultILoader(config); var loaderContext, loaderConfig, loaderCallbacks; loaderContext = { url: frag.url, frag: frag, responseType: 'arraybuffer', progressData: false }; var start = frag.byteRangeStartOffset, end = frag.byteRangeEndOffset; if (Object(number["isFiniteNumber"])(start) && Object(number["isFiniteNumber"])(end)) { loaderContext.rangeStart = start; loaderContext.rangeEnd = end; } loaderConfig = { timeout: config.fragLoadingTimeOut, maxRetry: 0, retryDelay: 0, maxRetryDelay: config.fragLoadingMaxRetryTimeout }; loaderCallbacks = { onSuccess: this.loadsuccess.bind(this), onError: this.loaderror.bind(this), onTimeout: this.loadtimeout.bind(this), onProgress: this.loadprogress.bind(this) }; loader.load(loaderContext, loaderConfig, loaderCallbacks); }; _proto.loadsuccess = function loadsuccess(response, stats, context, networkDetails) { if (networkDetails === void 0) { networkDetails = null; } var payload = response.data, frag = context.frag; // detach fragment loader on load success frag.loader = undefined; this.loaders[frag.type] = undefined; this.hls.trigger(events["default"].FRAG_LOADED, { payload: payload, frag: frag, stats: stats, networkDetails: networkDetails }); }; _proto.loaderror = function loaderror(response, context, networkDetails) { if (networkDetails === void 0) { networkDetails = null; } var frag = context.frag; var loader = frag.loader; if (loader) { loader.abort(); } this.loaders[frag.type] = undefined; this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].NETWORK_ERROR, details: errors["ErrorDetails"].FRAG_LOAD_ERROR, fatal: false, frag: context.frag, response: response, networkDetails: networkDetails }); }; _proto.loadtimeout = function loadtimeout(stats, context, networkDetails) { if (networkDetails === void 0) { networkDetails = null; } var frag = context.frag; var loader = frag.loader; if (loader) { loader.abort(); } this.loaders[frag.type] = undefined; this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].NETWORK_ERROR, details: errors["ErrorDetails"].FRAG_LOAD_TIMEOUT, fatal: false, frag: context.frag, networkDetails: networkDetails }); } // data will be used for progressive parsing ; _proto.loadprogress = function loadprogress(stats, context, data, networkDetails) { if (networkDetails === void 0) { networkDetails = null; } // jshint ignore:line var frag = context.frag; frag.loaded = stats.loaded; this.hls.trigger(events["default"].FRAG_LOAD_PROGRESS, { frag: frag, stats: stats, networkDetails: networkDetails }); }; return FragmentLoader; }(event_handler); /* harmony default export */ var fragment_loader = (fragment_loader_FragmentLoader); // CONCATENATED MODULE: ./src/loader/key-loader.ts function key_loader_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /* * Decrypt key Loader */ var key_loader_KeyLoader = /*#__PURE__*/function (_EventHandler) { key_loader_inheritsLoose(KeyLoader, _EventHandler); function KeyLoader(hls) { var _this; _this = _EventHandler.call(this, hls, events["default"].KEY_LOADING) || this; _this.loaders = {}; _this.decryptkey = null; _this.decrypturl = null; return _this; } var _proto = KeyLoader.prototype; _proto.destroy = function destroy() { for (var loaderName in this.loaders) { var loader = this.loaders[loaderName]; if (loader) { loader.destroy(); } } this.loaders = {}; _EventHandler.prototype.destroy.call(this); }; _proto.onKeyLoading = function onKeyLoading(data) { var frag = data.frag; var type = frag.type; var loader = this.loaders[type]; if (!frag.decryptdata) { logger["logger"].warn('Missing decryption data on fragment in onKeyLoading'); return; } // Load the key if the uri is different from previous one, or if the decrypt key has not yet been retrieved var uri = frag.decryptdata.uri; if (uri !== this.decrypturl || this.decryptkey === null) { var config = this.hls.config; if (loader) { logger["logger"].warn("abort previous key loader for type:" + type); loader.abort(); } if (!uri) { logger["logger"].warn('key uri is falsy'); return; } frag.loader = this.loaders[type] = new config.loader(config); this.decrypturl = uri; this.decryptkey = null; var loaderContext = { url: uri, frag: frag, responseType: 'arraybuffer' }; // maxRetry is 0 so that instead of retrying the same key on the same variant multiple times, // key-loader will trigger an error and rely on stream-controller to handle retry logic. // this will also align retry logic with fragment-loader var loaderConfig = { timeout: config.fragLoadingTimeOut, maxRetry: 0, retryDelay: config.fragLoadingRetryDelay, maxRetryDelay: config.fragLoadingMaxRetryTimeout }; var loaderCallbacks = { onSuccess: this.loadsuccess.bind(this), onError: this.loaderror.bind(this), onTimeout: this.loadtimeout.bind(this) }; frag.loader.load(loaderContext, loaderConfig, loaderCallbacks); } else if (this.decryptkey) { // Return the key if it's already been loaded frag.decryptdata.key = this.decryptkey; this.hls.trigger(events["default"].KEY_LOADED, { frag: frag }); } }; _proto.loadsuccess = function loadsuccess(response, stats, context) { var frag = context.frag; if (!frag.decryptdata) { logger["logger"].error('after key load, decryptdata unset'); return; } this.decryptkey = frag.decryptdata.key = new Uint8Array(response.data); // detach fragment loader on load success frag.loader = undefined; delete this.loaders[frag.type]; this.hls.trigger(events["default"].KEY_LOADED, { frag: frag }); }; _proto.loaderror = function loaderror(response, context) { var frag = context.frag; var loader = frag.loader; if (loader) { loader.abort(); } delete this.loaders[frag.type]; this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].NETWORK_ERROR, details: errors["ErrorDetails"].KEY_LOAD_ERROR, fatal: false, frag: frag, response: response }); }; _proto.loadtimeout = function loadtimeout(stats, context) { var frag = context.frag; var loader = frag.loader; if (loader) { loader.abort(); } delete this.loaders[frag.type]; this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].NETWORK_ERROR, details: errors["ErrorDetails"].KEY_LOAD_TIMEOUT, fatal: false, frag: frag }); }; return KeyLoader; }(event_handler); /* harmony default export */ var key_loader = (key_loader_KeyLoader); // CONCATENATED MODULE: ./src/controller/fragment-tracker.js function fragment_tracker_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } var FragmentState = { NOT_LOADED: 'NOT_LOADED', APPENDING: 'APPENDING', PARTIAL: 'PARTIAL', OK: 'OK' }; var fragment_tracker_FragmentTracker = /*#__PURE__*/function (_EventHandler) { fragment_tracker_inheritsLoose(FragmentTracker, _EventHandler); function FragmentTracker(hls) { var _this; _this = _EventHandler.call(this, hls, events["default"].BUFFER_APPENDED, events["default"].FRAG_BUFFERED, events["default"].FRAG_LOADED) || this; _this.bufferPadding = 0.2; _this.fragments = Object.create(null); _this.timeRanges = Object.create(null); _this.config = hls.config; return _this; } var _proto = FragmentTracker.prototype; _proto.destroy = function destroy() { this.fragments = Object.create(null); this.timeRanges = Object.create(null); this.config = null; event_handler.prototype.destroy.call(this); _EventHandler.prototype.destroy.call(this); } /** * Return a Fragment that match the position and levelType. * If not found any Fragment, return null * @param {number} position * @param {LevelType} levelType * @returns {Fragment|null} */ ; _proto.getBufferedFrag = function getBufferedFrag(position, levelType) { var fragments = this.fragments; var bufferedFrags = Object.keys(fragments).filter(function (key) { var fragmentEntity = fragments[key]; if (fragmentEntity.body.type !== levelType) { return false; } if (!fragmentEntity.buffered) { return false; } var frag = fragmentEntity.body; return frag.startPTS <= position && position <= frag.endPTS; }); if (bufferedFrags.length === 0) { return null; } else { // https://github.com/video-dev/hls.js/pull/1545#discussion_r166229566 var bufferedFragKey = bufferedFrags.pop(); return fragments[bufferedFragKey].body; } } /** * Partial fragments effected by coded frame eviction will be removed * The browser will unload parts of the buffer to free up memory for new buffer data * Fragments will need to be reloaded when the buffer is freed up, removing partial fragments will allow them to reload(since there might be parts that are still playable) * @param {String} elementaryStream The elementaryStream of media this is (eg. video/audio) * @param {TimeRanges} timeRange TimeRange object from a sourceBuffer */ ; _proto.detectEvictedFragments = function detectEvictedFragments(elementaryStream, timeRange) { var _this2 = this; // Check if any flagged fragments have been unloaded Object.keys(this.fragments).forEach(function (key) { var fragmentEntity = _this2.fragments[key]; if (!fragmentEntity || !fragmentEntity.buffered) { return; } var esData = fragmentEntity.range[elementaryStream]; if (!esData) { return; } var fragmentTimes = esData.time; for (var i = 0; i < fragmentTimes.length; i++) { var time = fragmentTimes[i]; if (!_this2.isTimeBuffered(time.startPTS, time.endPTS, timeRange)) { // Unregister partial fragment as it needs to load again to be reused _this2.removeFragment(fragmentEntity.body); break; } } }); } /** * Checks if the fragment passed in is loaded in the buffer properly * Partially loaded fragments will be registered as a partial fragment * @param {Object} fragment Check the fragment against all sourceBuffers loaded */ ; _proto.detectPartialFragments = function detectPartialFragments(fragment) { var _this3 = this; var fragKey = this.getFragmentKey(fragment); var fragmentEntity = this.fragments[fragKey]; if (fragmentEntity) { fragmentEntity.buffered = true; Object.keys(this.timeRanges).forEach(function (elementaryStream) { if (fragment.hasElementaryStream(elementaryStream)) { var timeRange = _this3.timeRanges[elementaryStream]; // Check for malformed fragments // Gaps need to be calculated for each elementaryStream fragmentEntity.range[elementaryStream] = _this3.getBufferedTimes(fragment.startPTS, fragment.endPTS, timeRange); } }); } }; _proto.getBufferedTimes = function getBufferedTimes(startPTS, endPTS, timeRange) { var fragmentTimes = []; var startTime, endTime; var fragmentPartial = false; for (var i = 0; i < timeRange.length; i++) { startTime = timeRange.start(i) - this.bufferPadding; endTime = timeRange.end(i) + this.bufferPadding; if (startPTS >= startTime && endPTS <= endTime) { // Fragment is entirely contained in buffer // No need to check the other timeRange times since it's completely playable fragmentTimes.push({ startPTS: Math.max(startPTS, timeRange.start(i)), endPTS: Math.min(endPTS, timeRange.end(i)) }); break; } else if (startPTS < endTime && endPTS > startTime) { // Check for intersection with buffer // Get playable sections of the fragment fragmentTimes.push({ startPTS: Math.max(startPTS, timeRange.start(i)), endPTS: Math.min(endPTS, timeRange.end(i)) }); fragmentPartial = true; } else if (endPTS <= startTime) { // No need to check the rest of the timeRange as it is in order break; } } return { time: fragmentTimes, partial: fragmentPartial }; }; _proto.getFragmentKey = function getFragmentKey(fragment) { return fragment.type + "_" + fragment.level + "_" + fragment.urlId + "_" + fragment.sn; } /** * Gets the partial fragment for a certain time * @param {Number} time * @returns {Object} fragment Returns a partial fragment at a time or null if there is no partial fragment */ ; _proto.getPartialFragment = function getPartialFragment(time) { var _this4 = this; var timePadding, startTime, endTime; var bestFragment = null; var bestOverlap = 0; Object.keys(this.fragments).forEach(function (key) { var fragmentEntity = _this4.fragments[key]; if (_this4.isPartial(fragmentEntity)) { startTime = fragmentEntity.body.startPTS - _this4.bufferPadding; endTime = fragmentEntity.body.endPTS + _this4.bufferPadding; if (time >= startTime && time <= endTime) { // Use the fragment that has the most padding from start and end time timePadding = Math.min(time - startTime, endTime - time); if (bestOverlap <= timePadding) { bestFragment = fragmentEntity.body; bestOverlap = timePadding; } } } }); return bestFragment; } /** * @param {Object} fragment The fragment to check * @returns {String} Returns the fragment state when a fragment never loaded or if it partially loaded */ ; _proto.getState = function getState(fragment) { var fragKey = this.getFragmentKey(fragment); var fragmentEntity = this.fragments[fragKey]; var state = FragmentState.NOT_LOADED; if (fragmentEntity !== undefined) { if (!fragmentEntity.buffered) { state = FragmentState.APPENDING; } else if (this.isPartial(fragmentEntity) === true) { state = FragmentState.PARTIAL; } else { state = FragmentState.OK; } } return state; }; _proto.isPartial = function isPartial(fragmentEntity) { return fragmentEntity.buffered === true && (fragmentEntity.range.video !== undefined && fragmentEntity.range.video.partial === true || fragmentEntity.range.audio !== undefined && fragmentEntity.range.audio.partial === true); }; _proto.isTimeBuffered = function isTimeBuffered(startPTS, endPTS, timeRange) { var startTime, endTime; for (var i = 0; i < timeRange.length; i++) { startTime = timeRange.start(i) - this.bufferPadding; endTime = timeRange.end(i) + this.bufferPadding; if (startPTS >= startTime && endPTS <= endTime) { return true; } if (endPTS <= startTime) { // No need to check the rest of the timeRange as it is in order return false; } } return false; } /** * Fires when a fragment loading is completed */ ; _proto.onFragLoaded = function onFragLoaded(e) { var fragment = e.frag; // don't track initsegment (for which sn is not a number) // don't track frags used for bitrateTest, they're irrelevant. if (!Object(number["isFiniteNumber"])(fragment.sn) || fragment.bitrateTest) { return; } this.fragments[this.getFragmentKey(fragment)] = { body: fragment, range: Object.create(null), buffered: false }; } /** * Fires when the buffer is updated */ ; _proto.onBufferAppended = function onBufferAppended(e) { var _this5 = this; // Store the latest timeRanges loaded in the buffer this.timeRanges = e.timeRanges; Object.keys(this.timeRanges).forEach(function (elementaryStream) { var timeRange = _this5.timeRanges[elementaryStream]; _this5.detectEvictedFragments(elementaryStream, timeRange); }); } /** * Fires after a fragment has been loaded into the source buffer */ ; _proto.onFragBuffered = function onFragBuffered(e) { this.detectPartialFragments(e.frag); } /** * Return true if fragment tracker has the fragment. * @param {Object} fragment * @returns {boolean} */ ; _proto.hasFragment = function hasFragment(fragment) { var fragKey = this.getFragmentKey(fragment); return this.fragments[fragKey] !== undefined; } /** * Remove a fragment from fragment tracker until it is loaded again * @param {Object} fragment The fragment to remove */ ; _proto.removeFragment = function removeFragment(fragment) { var fragKey = this.getFragmentKey(fragment); delete this.fragments[fragKey]; } /** * Remove all fragments from fragment tracker. */ ; _proto.removeAllFragments = function removeAllFragments() { this.fragments = Object.create(null); }; return FragmentTracker; }(event_handler); // CONCATENATED MODULE: ./src/utils/binary-search.ts var BinarySearch = { /** * Searches for an item in an array which matches a certain condition. * This requires the condition to only match one item in the array, * and for the array to be ordered. * * @param {Array<T>} list The array to search. * @param {BinarySearchComparison<T>} comparisonFn * Called and provided a candidate item as the first argument. * Should return: * > -1 if the item should be located at a lower index than the provided item. * > 1 if the item should be located at a higher index than the provided item. * > 0 if the item is the item you're looking for. * * @return {T | null} The object if it is found or null otherwise. */ search: function search(list, comparisonFn) { var minIndex = 0; var maxIndex = list.length - 1; var currentIndex = null; var currentElement = null; while (minIndex <= maxIndex) { currentIndex = (minIndex + maxIndex) / 2 | 0; currentElement = list[currentIndex]; var comparisonResult = comparisonFn(currentElement); if (comparisonResult > 0) { minIndex = currentIndex + 1; } else if (comparisonResult < 0) { maxIndex = currentIndex - 1; } else { return currentElement; } } return null; } }; /* harmony default export */ var binary_search = (BinarySearch); // CONCATENATED MODULE: ./src/utils/buffer-helper.ts /** * @module BufferHelper * * Providing methods dealing with buffer length retrieval for example. * * In general, a helper around HTML5 MediaElement TimeRanges gathered from `buffered` property. * * Also @see https://developer.mozilla.org/en-US/docs/Web/API/HTMLMediaElement/buffered */ var BufferHelper = /*#__PURE__*/function () { function BufferHelper() {} /** * Return true if `media`'s buffered include `position` * @param {Bufferable} media * @param {number} position * @returns {boolean} */ BufferHelper.isBuffered = function isBuffered(media, position) { try { if (media) { var buffered = media.buffered; for (var i = 0; i < buffered.length; i++) { if (position >= buffered.start(i) && position <= buffered.end(i)) { return true; } } } } catch (error) {// this is to catch // InvalidStateError: Failed to read the 'buffered' property from 'SourceBuffer': // This SourceBuffer has been removed from the parent media source } return false; }; BufferHelper.bufferInfo = function bufferInfo(media, pos, maxHoleDuration) { try { if (media) { var vbuffered = media.buffered; var buffered = []; var i; for (i = 0; i < vbuffered.length; i++) { buffered.push({ start: vbuffered.start(i), end: vbuffered.end(i) }); } return this.bufferedInfo(buffered, pos, maxHoleDuration); } } catch (error) {// this is to catch // InvalidStateError: Failed to read the 'buffered' property from 'SourceBuffer': // This SourceBuffer has been removed from the parent media source } return { len: 0, start: pos, end: pos, nextStart: undefined }; }; BufferHelper.bufferedInfo = function bufferedInfo(buffered, pos, maxHoleDuration) { // sort on buffer.start/smaller end (IE does not always return sorted buffered range) buffered.sort(function (a, b) { var diff = a.start - b.start; if (diff) { return diff; } else { return b.end - a.end; } }); var buffered2 = []; if (maxHoleDuration) { // there might be some small holes between buffer time range // consider that holes smaller than maxHoleDuration are irrelevant and build another // buffer time range representations that discards those holes for (var i = 0; i < buffered.length; i++) { var buf2len = buffered2.length; if (buf2len) { var buf2end = buffered2[buf2len - 1].end; // if small hole (value between 0 or maxHoleDuration ) or overlapping (negative) if (buffered[i].start - buf2end < maxHoleDuration) { // merge overlapping time ranges // update lastRange.end only if smaller than item.end // e.g. [ 1, 15] with [ 2,8] => [ 1,15] (no need to modify lastRange.end) // whereas [ 1, 8] with [ 2,15] => [ 1,15] ( lastRange should switch from [1,8] to [1,15]) if (buffered[i].end > buf2end) { buffered2[buf2len - 1].end = buffered[i].end; } } else { // big hole buffered2.push(buffered[i]); } } else { // first value buffered2.push(buffered[i]); } } } else { buffered2 = buffered; } var bufferLen = 0; // bufferStartNext can possibly be undefined based on the conditional logic below var bufferStartNext; // bufferStart and bufferEnd are buffer boundaries around current video position var bufferStart = pos; var bufferEnd = pos; for (var _i = 0; _i < buffered2.length; _i++) { var start = buffered2[_i].start, end = buffered2[_i].end; // logger.log('buf start/end:' + buffered.start(i) + '/' + buffered.end(i)); if (pos + maxHoleDuration >= start && pos < end) { // play position is inside this buffer TimeRange, retrieve end of buffer position and buffer length bufferStart = start; bufferEnd = end; bufferLen = bufferEnd - pos; } else if (pos + maxHoleDuration < start) { bufferStartNext = start; break; } } return { len: bufferLen, start: bufferStart, end: bufferEnd, nextStart: bufferStartNext }; }; return BufferHelper; }(); // EXTERNAL MODULE: ./node_modules/eventemitter3/index.js var eventemitter3 = __webpack_require__("./node_modules/eventemitter3/index.js"); // EXTERNAL MODULE: ./node_modules/webworkify-webpack/index.js var webworkify_webpack = __webpack_require__("./node_modules/webworkify-webpack/index.js"); // EXTERNAL MODULE: ./src/demux/demuxer-inline.js + 12 modules var demuxer_inline = __webpack_require__("./src/demux/demuxer-inline.js"); // CONCATENATED MODULE: ./src/utils/mediasource-helper.ts /** * MediaSource helper */ function getMediaSource() { return window.MediaSource || window.WebKitMediaSource; } // EXTERNAL MODULE: ./src/utils/get-self-scope.js var get_self_scope = __webpack_require__("./src/utils/get-self-scope.js"); // CONCATENATED MODULE: ./src/observer.ts function observer_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /** * Simple adapter sub-class of Nodejs-like EventEmitter. */ var Observer = /*#__PURE__*/function (_EventEmitter) { observer_inheritsLoose(Observer, _EventEmitter); function Observer() { return _EventEmitter.apply(this, arguments) || this; } var _proto = Observer.prototype; /** * We simply want to pass along the event-name itself * in every call to a handler, which is the purpose of our `trigger` method * extending the standard API. */ _proto.trigger = function trigger(event) { for (var _len = arguments.length, data = new Array(_len > 1 ? _len - 1 : 0), _key = 1; _key < _len; _key++) { data[_key - 1] = arguments[_key]; } this.emit.apply(this, [event, event].concat(data)); }; return Observer; }(eventemitter3["EventEmitter"]); // CONCATENATED MODULE: ./src/demux/demuxer.js // see https://stackoverflow.com/a/11237259/589493 var global = Object(get_self_scope["getSelfScope"])(); // safeguard for code that might run both on worker and main thread var demuxer_MediaSource = getMediaSource() || { isTypeSupported: function isTypeSupported() { return false; } }; var demuxer_Demuxer = /*#__PURE__*/function () { function Demuxer(hls, id) { var _this = this; this.hls = hls; this.id = id; var observer = this.observer = new Observer(); var config = hls.config; var forwardMessage = function forwardMessage(ev, data) { data = data || {}; data.frag = _this.frag; data.id = _this.id; hls.trigger(ev, data); }; // forward events to main thread observer.on(events["default"].FRAG_DECRYPTED, forwardMessage); observer.on(events["default"].FRAG_PARSING_INIT_SEGMENT, forwardMessage); observer.on(events["default"].FRAG_PARSING_DATA, forwardMessage); observer.on(events["default"].FRAG_PARSED, forwardMessage); observer.on(events["default"].ERROR, forwardMessage); observer.on(events["default"].FRAG_PARSING_METADATA, forwardMessage); observer.on(events["default"].FRAG_PARSING_USERDATA, forwardMessage); observer.on(events["default"].INIT_PTS_FOUND, forwardMessage); var typeSupported = { mp4: demuxer_MediaSource.isTypeSupported('video/mp4'), mpeg: demuxer_MediaSource.isTypeSupported('audio/mpeg'), mp3: demuxer_MediaSource.isTypeSupported('audio/mp4; codecs="mp3"') }; // navigator.vendor is not always available in Web Worker // refer to https://developer.mozilla.org/en-US/docs/Web/API/WorkerGlobalScope/navigator var vendor = navigator.vendor; if (config.enableWorker && typeof Worker !== 'undefined') { logger["logger"].log('demuxing in webworker'); var w; try { w = this.w = webworkify_webpack(/*require.resolve*/(/*! ../demux/demuxer-worker.js */ "./src/demux/demuxer-worker.js")); this.onwmsg = this.onWorkerMessage.bind(this); w.addEventListener('message', this.onwmsg); w.onerror = function (event) { hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].OTHER_ERROR, details: errors["ErrorDetails"].INTERNAL_EXCEPTION, fatal: true, event: 'demuxerWorker', err: { message: event.message + ' (' + event.filename + ':' + event.lineno + ')' } }); }; w.postMessage({ cmd: 'init', typeSupported: typeSupported, vendor: vendor, id: id, config: JSON.stringify(config) }); } catch (err) { logger["logger"].warn('Error in worker:', err); logger["logger"].error('Error while initializing DemuxerWorker, fallback on DemuxerInline'); if (w) { // revoke the Object URL that was used to create demuxer worker, so as not to leak it global.URL.revokeObjectURL(w.objectURL); } this.demuxer = new demuxer_inline["default"](observer, typeSupported, config, vendor); this.w = undefined; } } else { this.demuxer = new demuxer_inline["default"](observer, typeSupported, config, vendor); } } var _proto = Demuxer.prototype; _proto.destroy = function destroy() { var w = this.w; if (w) { w.removeEventListener('message', this.onwmsg); w.terminate(); this.w = null; } else { var demuxer = this.demuxer; if (demuxer) { demuxer.destroy(); this.demuxer = null; } } var observer = this.observer; if (observer) { observer.removeAllListeners(); this.observer = null; } }; _proto.push = function push(data, initSegment, audioCodec, videoCodec, frag, duration, accurateTimeOffset, defaultInitPTS) { var w = this.w; var timeOffset = Object(number["isFiniteNumber"])(frag.startPTS) ? frag.startPTS : frag.start; var decryptdata = frag.decryptdata; var lastFrag = this.frag; var discontinuity = !(lastFrag && frag.cc === lastFrag.cc); var trackSwitch = !(lastFrag && frag.level === lastFrag.level); var nextSN = lastFrag && frag.sn === lastFrag.sn + 1; var contiguous = !trackSwitch && nextSN; if (discontinuity) { logger["logger"].log(this.id + ":discontinuity detected"); } if (trackSwitch) { logger["logger"].log(this.id + ":switch detected"); } this.frag = frag; if (w) { // post fragment payload as transferable objects for ArrayBuffer (no copy) w.postMessage({ cmd: 'demux', data: data, decryptdata: decryptdata, initSegment: initSegment, audioCodec: audioCodec, videoCodec: videoCodec, timeOffset: timeOffset, discontinuity: discontinuity, trackSwitch: trackSwitch, contiguous: contiguous, duration: duration, accurateTimeOffset: accurateTimeOffset, defaultInitPTS: defaultInitPTS }, data instanceof ArrayBuffer ? [data] : []); } else { var demuxer = this.demuxer; if (demuxer) { demuxer.push(data, decryptdata, initSegment, audioCodec, videoCodec, timeOffset, discontinuity, trackSwitch, contiguous, duration, accurateTimeOffset, defaultInitPTS); } } }; _proto.onWorkerMessage = function onWorkerMessage(ev) { var data = ev.data, hls = this.hls; switch (data.event) { case 'init': // revoke the Object URL that was used to create demuxer worker, so as not to leak it global.URL.revokeObjectURL(this.w.objectURL); break; // special case for FRAG_PARSING_DATA: data1 and data2 are transferable objects case events["default"].FRAG_PARSING_DATA: data.data.data1 = new Uint8Array(data.data1); if (data.data2) { data.data.data2 = new Uint8Array(data.data2); } /* falls through */ default: data.data = data.data || {}; data.data.frag = this.frag; data.data.id = this.id; hls.trigger(data.event, data.data); break; } }; return Demuxer; }(); /* harmony default export */ var demux_demuxer = (demuxer_Demuxer); // CONCATENATED MODULE: ./src/controller/level-helper.js /** * @module LevelHelper * * Providing methods dealing with playlist sliding and drift * * TODO: Create an actual `Level` class/model that deals with all this logic in an object-oriented-manner. * * */ function addGroupId(level, type, id) { switch (type) { case 'audio': if (!level.audioGroupIds) { level.audioGroupIds = []; } level.audioGroupIds.push(id); break; case 'text': if (!level.textGroupIds) { level.textGroupIds = []; } level.textGroupIds.push(id); break; } } function updatePTS(fragments, fromIdx, toIdx) { var fragFrom = fragments[fromIdx], fragTo = fragments[toIdx], fragToPTS = fragTo.startPTS; // if we know startPTS[toIdx] if (Object(number["isFiniteNumber"])(fragToPTS)) { // update fragment duration. // it helps to fix drifts between playlist reported duration and fragment real duration if (toIdx > fromIdx) { fragFrom.duration = fragToPTS - fragFrom.start; if (fragFrom.duration < 0) { logger["logger"].warn("negative duration computed for frag " + fragFrom.sn + ",level " + fragFrom.level + ", there should be some duration drift between playlist and fragment!"); } } else { fragTo.duration = fragFrom.start - fragToPTS; if (fragTo.duration < 0) { logger["logger"].warn("negative duration computed for frag " + fragTo.sn + ",level " + fragTo.level + ", there should be some duration drift between playlist and fragment!"); } } } else { // we dont know startPTS[toIdx] if (toIdx > fromIdx) { fragTo.start = fragFrom.start + fragFrom.duration; } else { fragTo.start = Math.max(fragFrom.start - fragTo.duration, 0); } } } function updateFragPTSDTS(details, frag, startPTS, endPTS, startDTS, endDTS) { // update frag PTS/DTS var maxStartPTS = startPTS; if (Object(number["isFiniteNumber"])(frag.startPTS)) { // delta PTS between audio and video var deltaPTS = Math.abs(frag.startPTS - startPTS); if (!Object(number["isFiniteNumber"])(frag.deltaPTS)) { frag.deltaPTS = deltaPTS; } else { frag.deltaPTS = Math.max(deltaPTS, frag.deltaPTS); } maxStartPTS = Math.max(startPTS, frag.startPTS); startPTS = Math.min(startPTS, frag.startPTS); endPTS = Math.max(endPTS, frag.endPTS); startDTS = Math.min(startDTS, frag.startDTS); endDTS = Math.max(endDTS, frag.endDTS); } var drift = startPTS - frag.start; frag.start = frag.startPTS = startPTS; frag.maxStartPTS = maxStartPTS; frag.endPTS = endPTS; frag.startDTS = startDTS; frag.endDTS = endDTS; frag.duration = endPTS - startPTS; var sn = frag.sn; // exit if sn out of range if (!details || sn < details.startSN || sn > details.endSN) { return 0; } var fragIdx, fragments, i; fragIdx = sn - details.startSN; fragments = details.fragments; // update frag reference in fragments array // rationale is that fragments array might not contain this frag object. // this will happen if playlist has been refreshed between frag loading and call to updateFragPTSDTS() // if we don't update frag, we won't be able to propagate PTS info on the playlist // resulting in invalid sliding computation fragments[fragIdx] = frag; // adjust fragment PTS/duration from seqnum-1 to frag 0 for (i = fragIdx; i > 0; i--) { updatePTS(fragments, i, i - 1); } // adjust fragment PTS/duration from seqnum to last frag for (i = fragIdx; i < fragments.length - 1; i++) { updatePTS(fragments, i, i + 1); } details.PTSKnown = true; return drift; } function mergeDetails(oldDetails, newDetails) { // potentially retrieve cached initsegment if (newDetails.initSegment && oldDetails.initSegment) { newDetails.initSegment = oldDetails.initSegment; } // check if old/new playlists have fragments in common // loop through overlapping SN and update startPTS , cc, and duration if any found var ccOffset = 0; var PTSFrag; mapFragmentIntersection(oldDetails, newDetails, function (oldFrag, newFrag) { ccOffset = oldFrag.cc - newFrag.cc; if (Object(number["isFiniteNumber"])(oldFrag.startPTS)) { newFrag.start = newFrag.startPTS = oldFrag.startPTS; newFrag.endPTS = oldFrag.endPTS; newFrag.duration = oldFrag.duration; newFrag.backtracked = oldFrag.backtracked; newFrag.dropped = oldFrag.dropped; PTSFrag = newFrag; } // PTS is known when there are overlapping segments newDetails.PTSKnown = true; }); if (!newDetails.PTSKnown) { return; } if (ccOffset) { logger["logger"].log('discontinuity sliding from playlist, take drift into account'); var newFragments = newDetails.fragments; for (var i = 0; i < newFragments.length; i++) { newFragments[i].cc += ccOffset; } } // if at least one fragment contains PTS info, recompute PTS information for all fragments if (PTSFrag) { updateFragPTSDTS(newDetails, PTSFrag, PTSFrag.startPTS, PTSFrag.endPTS, PTSFrag.startDTS, PTSFrag.endDTS); } else { // ensure that delta is within oldFragments range // also adjust sliding in case delta is 0 (we could have old=[50-60] and new=old=[50-61]) // in that case we also need to adjust start offset of all fragments adjustSliding(oldDetails, newDetails); } // if we are here, it means we have fragments overlapping between // old and new level. reliable PTS info is thus relying on old level newDetails.PTSKnown = oldDetails.PTSKnown; } function mergeSubtitlePlaylists(oldPlaylist, newPlaylist, referenceStart) { if (referenceStart === void 0) { referenceStart = 0; } var lastIndex = -1; mapFragmentIntersection(oldPlaylist, newPlaylist, function (oldFrag, newFrag, index) { newFrag.start = oldFrag.start; lastIndex = index; }); var frags = newPlaylist.fragments; if (lastIndex < 0) { frags.forEach(function (frag) { frag.start += referenceStart; }); return; } for (var i = lastIndex + 1; i < frags.length; i++) { frags[i].start = frags[i - 1].start + frags[i - 1].duration; } } function mapFragmentIntersection(oldPlaylist, newPlaylist, intersectionFn) { if (!oldPlaylist || !newPlaylist) { return; } var start = Math.max(oldPlaylist.startSN, newPlaylist.startSN) - newPlaylist.startSN; var end = Math.min(oldPlaylist.endSN, newPlaylist.endSN) - newPlaylist.startSN; var delta = newPlaylist.startSN - oldPlaylist.startSN; for (var i = start; i <= end; i++) { var oldFrag = oldPlaylist.fragments[delta + i]; var newFrag = newPlaylist.fragments[i]; if (!oldFrag || !newFrag) { break; } intersectionFn(oldFrag, newFrag, i); } } function adjustSliding(oldPlaylist, newPlaylist) { var delta = newPlaylist.startSN - oldPlaylist.startSN; var oldFragments = oldPlaylist.fragments; var newFragments = newPlaylist.fragments; if (delta < 0 || delta > oldFragments.length) { return; } for (var i = 0; i < newFragments.length; i++) { newFragments[i].start += oldFragments[delta].start; } } function computeReloadInterval(currentPlaylist, newPlaylist, lastRequestTime) { var reloadInterval = 1000 * (newPlaylist.averagetargetduration ? newPlaylist.averagetargetduration : newPlaylist.targetduration); var minReloadInterval = reloadInterval / 2; if (currentPlaylist && newPlaylist.endSN === currentPlaylist.endSN) { // follow HLS Spec, If the client reloads a Playlist file and finds that it has not // changed then it MUST wait for a period of one-half the target // duration before retrying. reloadInterval = minReloadInterval; } if (lastRequestTime) { reloadInterval = Math.max(minReloadInterval, reloadInterval - (window.performance.now() - lastRequestTime)); } // in any case, don't reload more than half of target duration return Math.round(reloadInterval); } // CONCATENATED MODULE: ./src/utils/time-ranges.ts /** * TimeRanges to string helper */ var TimeRanges = { toString: function toString(r) { var log = ''; var len = r.length; for (var i = 0; i < len; i++) { log += '[' + r.start(i).toFixed(3) + ',' + r.end(i).toFixed(3) + ']'; } return log; } }; /* harmony default export */ var time_ranges = (TimeRanges); // CONCATENATED MODULE: ./src/utils/discontinuities.js function findFirstFragWithCC(fragments, cc) { var firstFrag = null; for (var i = 0; i < fragments.length; i += 1) { var currentFrag = fragments[i]; if (currentFrag && currentFrag.cc === cc) { firstFrag = currentFrag; break; } } return firstFrag; } function findFragWithCC(fragments, CC) { return binary_search.search(fragments, function (candidate) { if (candidate.cc < CC) { return 1; } else if (candidate.cc > CC) { return -1; } else { return 0; } }); } function shouldAlignOnDiscontinuities(lastFrag, lastLevel, details) { var shouldAlign = false; if (lastLevel && lastLevel.details && details) { if (details.endCC > details.startCC || lastFrag && lastFrag.cc < details.startCC) { shouldAlign = true; } } return shouldAlign; } // Find the first frag in the previous level which matches the CC of the first frag of the new level function findDiscontinuousReferenceFrag(prevDetails, curDetails) { var prevFrags = prevDetails.fragments; var curFrags = curDetails.fragments; if (!curFrags.length || !prevFrags.length) { logger["logger"].log('No fragments to align'); return; } var prevStartFrag = findFirstFragWithCC(prevFrags, curFrags[0].cc); if (!prevStartFrag || prevStartFrag && !prevStartFrag.startPTS) { logger["logger"].log('No frag in previous level to align on'); return; } return prevStartFrag; } function adjustPts(sliding, details) { details.fragments.forEach(function (frag) { if (frag) { var start = frag.start + sliding; frag.start = frag.startPTS = start; frag.endPTS = start + frag.duration; } }); details.PTSKnown = true; } /** * Using the parameters of the last level, this function computes PTS' of the new fragments so that they form a * contiguous stream with the last fragments. * The PTS of a fragment lets Hls.js know where it fits into a stream - by knowing every PTS, we know which fragment to * download at any given time. PTS is normally computed when the fragment is demuxed, so taking this step saves us time * and an extra download. * @param lastFrag * @param lastLevel * @param details */ function alignStream(lastFrag, lastLevel, details) { alignDiscontinuities(lastFrag, details, lastLevel); if (!details.PTSKnown && lastLevel) { // If the PTS wasn't figured out via discontinuity sequence that means there was no CC increase within the level. // Aligning via Program Date Time should therefore be reliable, since PDT should be the same within the same // discontinuity sequence. alignPDT(details, lastLevel.details); } } /** * Computes the PTS if a new level's fragments using the PTS of a fragment in the last level which shares the same * discontinuity sequence. * @param lastLevel - The details of the last loaded level * @param details - The details of the new level */ function alignDiscontinuities(lastFrag, details, lastLevel) { if (shouldAlignOnDiscontinuities(lastFrag, lastLevel, details)) { var referenceFrag = findDiscontinuousReferenceFrag(lastLevel.details, details); if (referenceFrag) { logger["logger"].log('Adjusting PTS using last level due to CC increase within current level'); adjustPts(referenceFrag.start, details); } } } /** * Computes the PTS of a new level's fragments using the difference in Program Date Time from the last level. * @param details - The details of the new level * @param lastDetails - The details of the last loaded level */ function alignPDT(details, lastDetails) { if (lastDetails && lastDetails.fragments.length) { if (!details.hasProgramDateTime || !lastDetails.hasProgramDateTime) { return; } // if last level sliding is 1000 and its first frag PROGRAM-DATE-TIME is 2017-08-20 1:10:00 AM // and if new details first frag PROGRAM DATE-TIME is 2017-08-20 1:10:08 AM // then we can deduce that playlist B sliding is 1000+8 = 1008s var lastPDT = lastDetails.fragments[0].programDateTime; var newPDT = details.fragments[0].programDateTime; // date diff is in ms. frag.start is in seconds var sliding = (newPDT - lastPDT) / 1000 + lastDetails.fragments[0].start; if (Object(number["isFiniteNumber"])(sliding)) { logger["logger"].log("adjusting PTS using programDateTime delta, sliding:" + sliding.toFixed(3)); adjustPts(sliding, details); } } } // CONCATENATED MODULE: ./src/controller/fragment-finders.ts /** * Returns first fragment whose endPdt value exceeds the given PDT. * @param {Array<Fragment>} fragments - The array of candidate fragments * @param {number|null} [PDTValue = null] - The PDT value which must be exceeded * @param {number} [maxFragLookUpTolerance = 0] - The amount of time that a fragment's start/end can be within in order to be considered contiguous * @returns {*|null} fragment - The best matching fragment */ function findFragmentByPDT(fragments, PDTValue, maxFragLookUpTolerance) { if (PDTValue === null || !Array.isArray(fragments) || !fragments.length || !Object(number["isFiniteNumber"])(PDTValue)) { return null; } // if less than start var startPDT = fragments[0].programDateTime; if (PDTValue < (startPDT || 0)) { return null; } var endPDT = fragments[fragments.length - 1].endProgramDateTime; if (PDTValue >= (endPDT || 0)) { return null; } maxFragLookUpTolerance = maxFragLookUpTolerance || 0; for (var seg = 0; seg < fragments.length; ++seg) { var frag = fragments[seg]; if (pdtWithinToleranceTest(PDTValue, maxFragLookUpTolerance, frag)) { return frag; } } return null; } /** * Finds a fragment based on the SN of the previous fragment; or based on the needs of the current buffer. * This method compensates for small buffer gaps by applying a tolerance to the start of any candidate fragment, thus * breaking any traps which would cause the same fragment to be continuously selected within a small range. * @param {*} fragPrevious - The last frag successfully appended * @param {Array<Fragment>} fragments - The array of candidate fragments * @param {number} [bufferEnd = 0] - The end of the contiguous buffered range the playhead is currently within * @param {number} maxFragLookUpTolerance - The amount of time that a fragment's start/end can be within in order to be considered contiguous * @returns {*} foundFrag - The best matching fragment */ function findFragmentByPTS(fragPrevious, fragments, bufferEnd, maxFragLookUpTolerance) { if (bufferEnd === void 0) { bufferEnd = 0; } if (maxFragLookUpTolerance === void 0) { maxFragLookUpTolerance = 0; } var fragNext = null; if (fragPrevious) { fragNext = fragments[fragPrevious.sn - fragments[0].sn + 1]; } else if (bufferEnd === 0 && fragments[0].start === 0) { fragNext = fragments[0]; } // Prefer the next fragment if it's within tolerance if (fragNext && fragment_finders_fragmentWithinToleranceTest(bufferEnd, maxFragLookUpTolerance, fragNext) === 0) { return fragNext; } // We might be seeking past the tolerance so find the best match var foundFragment = binary_search.search(fragments, fragment_finders_fragmentWithinToleranceTest.bind(null, bufferEnd, maxFragLookUpTolerance)); if (foundFragment) { return foundFragment; } // If no match was found return the next fragment after fragPrevious, or null return fragNext; } /** * The test function used by the findFragmentBySn's BinarySearch to look for the best match to the current buffer conditions. * @param {*} candidate - The fragment to test * @param {number} [bufferEnd = 0] - The end of the current buffered range the playhead is currently within * @param {number} [maxFragLookUpTolerance = 0] - The amount of time that a fragment's start can be within in order to be considered contiguous * @returns {number} - 0 if it matches, 1 if too low, -1 if too high */ function fragment_finders_fragmentWithinToleranceTest(bufferEnd, maxFragLookUpTolerance, candidate) { if (bufferEnd === void 0) { bufferEnd = 0; } if (maxFragLookUpTolerance === void 0) { maxFragLookUpTolerance = 0; } // offset should be within fragment boundary - config.maxFragLookUpTolerance // this is to cope with situations like // bufferEnd = 9.991 // frag[Ø] : [0,10] // frag[1] : [10,20] // bufferEnd is within frag[0] range ... although what we are expecting is to return frag[1] here // frag start frag start+duration // |-----------------------------| // <---> <---> // ...--------><-----------------------------><---------.... // previous frag matching fragment next frag // return -1 return 0 return 1 // logger.log(`level/sn/start/end/bufEnd:${level}/${candidate.sn}/${candidate.start}/${(candidate.start+candidate.duration)}/${bufferEnd}`); // Set the lookup tolerance to be small enough to detect the current segment - ensures we don't skip over very small segments var candidateLookupTolerance = Math.min(maxFragLookUpTolerance, candidate.duration + (candidate.deltaPTS ? candidate.deltaPTS : 0)); if (candidate.start + candidate.duration - candidateLookupTolerance <= bufferEnd) { return 1; } else if (candidate.start - candidateLookupTolerance > bufferEnd && candidate.start) { // if maxFragLookUpTolerance will have negative value then don't return -1 for first element return -1; } return 0; } /** * The test function used by the findFragmentByPdt's BinarySearch to look for the best match to the current buffer conditions. * This function tests the candidate's program date time values, as represented in Unix time * @param {*} candidate - The fragment to test * @param {number} [pdtBufferEnd = 0] - The Unix time representing the end of the current buffered range * @param {number} [maxFragLookUpTolerance = 0] - The amount of time that a fragment's start can be within in order to be considered contiguous * @returns {boolean} True if contiguous, false otherwise */ function pdtWithinToleranceTest(pdtBufferEnd, maxFragLookUpTolerance, candidate) { var candidateLookupTolerance = Math.min(maxFragLookUpTolerance, candidate.duration + (candidate.deltaPTS ? candidate.deltaPTS : 0)) * 1000; // endProgramDateTime can be null, default to zero var endProgramDateTime = candidate.endProgramDateTime || 0; return endProgramDateTime - candidateLookupTolerance > pdtBufferEnd; } // CONCATENATED MODULE: ./src/controller/gap-controller.js var STALL_MINIMUM_DURATION_MS = 250; var MAX_START_GAP_JUMP = 2.0; var SKIP_BUFFER_HOLE_STEP_SECONDS = 0.1; var SKIP_BUFFER_RANGE_START = 0.05; var gap_controller_GapController = /*#__PURE__*/function () { function GapController(config, media, fragmentTracker, hls) { this.config = config; this.media = media; this.fragmentTracker = fragmentTracker; this.hls = hls; this.nudgeRetry = 0; this.stallReported = false; this.stalled = null; this.moved = false; this.seeking = false; } /** * Checks if the playhead is stuck within a gap, and if so, attempts to free it. * A gap is an unbuffered range between two buffered ranges (or the start and the first buffered range). * * @param {number} lastCurrentTime Previously read playhead position */ var _proto = GapController.prototype; _proto.poll = function poll(lastCurrentTime) { var config = this.config, media = this.media, stalled = this.stalled; var currentTime = media.currentTime, seeking = media.seeking; var seeked = this.seeking && !seeking; var beginSeek = !this.seeking && seeking; this.seeking = seeking; // The playhead is moving, no-op if (currentTime !== lastCurrentTime) { this.moved = true; if (stalled !== null) { // The playhead is now moving, but was previously stalled if (this.stallReported) { var _stalledDuration = self.performance.now() - stalled; logger["logger"].warn("playback not stuck anymore @" + currentTime + ", after " + Math.round(_stalledDuration) + "ms"); this.stallReported = false; } this.stalled = null; this.nudgeRetry = 0; } return; } // Clear stalled state when beginning or finishing seeking so that we don't report stalls coming out of a seek if (beginSeek || seeked) { this.stalled = null; } // The playhead should not be moving if (media.paused || media.ended || media.playbackRate === 0 || !media.buffered.length) { return; } var bufferInfo = BufferHelper.bufferInfo(media, currentTime, 0); var isBuffered = bufferInfo.len > 0; var nextStart = bufferInfo.nextStart || 0; // There is no playable buffer (waiting for buffer append) if (!isBuffered && !nextStart) { return; } if (seeking) { // Waiting for seeking in a buffered range to complete var hasEnoughBuffer = bufferInfo.len > MAX_START_GAP_JUMP; // Next buffered range is too far ahead to jump to while still seeking var noBufferGap = !nextStart || nextStart - currentTime > MAX_START_GAP_JUMP && !this.fragmentTracker.getPartialFragment(currentTime); if (hasEnoughBuffer || noBufferGap) { return; } // Reset moved state when seeking to a point in or before a gap this.moved = false; } // Skip start gaps if we haven't played, but the last poll detected the start of a stall // The addition poll gives the browser a chance to jump the gap for us if (!this.moved && this.stalled) { // Jump start gaps within jump threshold var startJump = Math.max(nextStart, bufferInfo.start || 0) - currentTime; if (startJump > 0 && startJump <= MAX_START_GAP_JUMP) { this._trySkipBufferHole(null); return; } } // Start tracking stall time var tnow = self.performance.now(); if (stalled === null) { this.stalled = tnow; return; } var stalledDuration = tnow - stalled; if (!seeking && stalledDuration >= STALL_MINIMUM_DURATION_MS) { // Report stalling after trying to fix this._reportStall(bufferInfo.len); } var bufferedWithHoles = BufferHelper.bufferInfo(media, currentTime, config.maxBufferHole); this._tryFixBufferStall(bufferedWithHoles, stalledDuration); } /** * Detects and attempts to fix known buffer stalling issues. * @param bufferInfo - The properties of the current buffer. * @param stalledDurationMs - The amount of time Hls.js has been stalling for. * @private */ ; _proto._tryFixBufferStall = function _tryFixBufferStall(bufferInfo, stalledDurationMs) { var config = this.config, fragmentTracker = this.fragmentTracker, media = this.media; var currentTime = media.currentTime; var partial = fragmentTracker.getPartialFragment(currentTime); if (partial) { // Try to skip over the buffer hole caused by a partial fragment // This method isn't limited by the size of the gap between buffered ranges var targetTime = this._trySkipBufferHole(partial); // we return here in this case, meaning // the branch below only executes when we don't handle a partial fragment if (targetTime) { return; } } // if we haven't had to skip over a buffer hole of a partial fragment // we may just have to "nudge" the playlist as the browser decoding/rendering engine // needs to cross some sort of threshold covering all source-buffers content // to start playing properly. if (bufferInfo.len > config.maxBufferHole && stalledDurationMs > config.highBufferWatchdogPeriod * 1000) { logger["logger"].warn('Trying to nudge playhead over buffer-hole'); // Try to nudge currentTime over a buffer hole if we've been stalling for the configured amount of seconds // We only try to jump the hole if it's under the configured size // Reset stalled so to rearm watchdog timer this.stalled = null; this._tryNudgeBuffer(); } } /** * Triggers a BUFFER_STALLED_ERROR event, but only once per stall period. * @param bufferLen - The playhead distance from the end of the current buffer segment. * @private */ ; _proto._reportStall = function _reportStall(bufferLen) { var hls = this.hls, media = this.media, stallReported = this.stallReported; if (!stallReported) { // Report stalled error once this.stallReported = true; logger["logger"].warn("Playback stalling at @" + media.currentTime + " due to low buffer (buffer=" + bufferLen + ")"); hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].BUFFER_STALLED_ERROR, fatal: false, buffer: bufferLen }); } } /** * Attempts to fix buffer stalls by jumping over known gaps caused by partial fragments * @param partial - The partial fragment found at the current time (where playback is stalling). * @private */ ; _proto._trySkipBufferHole = function _trySkipBufferHole(partial) { var config = this.config, hls = this.hls, media = this.media; var currentTime = media.currentTime; var lastEndTime = 0; // Check if currentTime is between unbuffered regions of partial fragments for (var i = 0; i < media.buffered.length; i++) { var startTime = media.buffered.start(i); if (currentTime + config.maxBufferHole >= lastEndTime && currentTime < startTime) { var targetTime = Math.max(startTime + SKIP_BUFFER_RANGE_START, media.currentTime + SKIP_BUFFER_HOLE_STEP_SECONDS); logger["logger"].warn("skipping hole, adjusting currentTime from " + currentTime + " to " + targetTime); this.moved = true; this.stalled = null; media.currentTime = targetTime; if (partial) { hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].BUFFER_SEEK_OVER_HOLE, fatal: false, reason: "fragment loaded with buffer holes, seeking from " + currentTime + " to " + targetTime, frag: partial }); } return targetTime; } lastEndTime = media.buffered.end(i); } return 0; } /** * Attempts to fix buffer stalls by advancing the mediaElement's current time by a small amount. * @private */ ; _proto._tryNudgeBuffer = function _tryNudgeBuffer() { var config = this.config, hls = this.hls, media = this.media; var currentTime = media.currentTime; var nudgeRetry = (this.nudgeRetry || 0) + 1; this.nudgeRetry = nudgeRetry; if (nudgeRetry < config.nudgeMaxRetry) { var targetTime = currentTime + nudgeRetry * config.nudgeOffset; // playback stalled in buffered area ... let's nudge currentTime to try to overcome this logger["logger"].warn("Nudging 'currentTime' from " + currentTime + " to " + targetTime); media.currentTime = targetTime; hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].BUFFER_NUDGE_ON_STALL, fatal: false }); } else { logger["logger"].error("Playhead still not moving while enough data buffered @" + currentTime + " after " + config.nudgeMaxRetry + " nudges"); hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].BUFFER_STALLED_ERROR, fatal: true }); } }; return GapController; }(); // CONCATENATED MODULE: ./src/task-loop.ts function _assertThisInitialized(self) { if (self === void 0) { throw new ReferenceError("this hasn't been initialised - super() hasn't been called"); } return self; } function task_loop_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /** * Sub-class specialization of EventHandler base class. * * TaskLoop allows to schedule a task function being called (optionnaly repeatedly) on the main loop, * scheduled asynchroneously, avoiding recursive calls in the same tick. * * The task itself is implemented in `doTick`. It can be requested and called for single execution * using the `tick` method. * * It will be assured that the task execution method (`tick`) only gets called once per main loop "tick", * no matter how often it gets requested for execution. Execution in further ticks will be scheduled accordingly. * * If further execution requests have already been scheduled on the next tick, it can be checked with `hasNextTick`, * and cancelled with `clearNextTick`. * * The task can be scheduled as an interval repeatedly with a period as parameter (see `setInterval`, `clearInterval`). * * Sub-classes need to implement the `doTick` method which will effectively have the task execution routine. * * Further explanations: * * The baseclass has a `tick` method that will schedule the doTick call. It may be called synchroneously * only for a stack-depth of one. On re-entrant calls, sub-sequent calls are scheduled for next main loop ticks. * * When the task execution (`tick` method) is called in re-entrant way this is detected and * we are limiting the task execution per call stack to exactly one, but scheduling/post-poning further * task processing on the next main loop iteration (also known as "next tick" in the Node/JS runtime lingo). */ var TaskLoop = /*#__PURE__*/function (_EventHandler) { task_loop_inheritsLoose(TaskLoop, _EventHandler); function TaskLoop(hls) { var _this; for (var _len = arguments.length, events = new Array(_len > 1 ? _len - 1 : 0), _key = 1; _key < _len; _key++) { events[_key - 1] = arguments[_key]; } _this = _EventHandler.call.apply(_EventHandler, [this, hls].concat(events)) || this; _this._boundTick = void 0; _this._tickTimer = null; _this._tickInterval = null; _this._tickCallCount = 0; _this._boundTick = _this.tick.bind(_assertThisInitialized(_this)); return _this; } /** * @override */ var _proto = TaskLoop.prototype; _proto.onHandlerDestroying = function onHandlerDestroying() { // clear all timers before unregistering from event bus this.clearNextTick(); this.clearInterval(); } /** * @returns {boolean} */ ; _proto.hasInterval = function hasInterval() { return !!this._tickInterval; } /** * @returns {boolean} */ ; _proto.hasNextTick = function hasNextTick() { return !!this._tickTimer; } /** * @param {number} millis Interval time (ms) * @returns {boolean} True when interval has been scheduled, false when already scheduled (no effect) */ ; _proto.setInterval = function setInterval(millis) { if (!this._tickInterval) { this._tickInterval = self.setInterval(this._boundTick, millis); return true; } return false; } /** * @returns {boolean} True when interval was cleared, false when none was set (no effect) */ ; _proto.clearInterval = function clearInterval() { if (this._tickInterval) { self.clearInterval(this._tickInterval); this._tickInterval = null; return true; } return false; } /** * @returns {boolean} True when timeout was cleared, false when none was set (no effect) */ ; _proto.clearNextTick = function clearNextTick() { if (this._tickTimer) { self.clearTimeout(this._tickTimer); this._tickTimer = null; return true; } return false; } /** * Will call the subclass doTick implementation in this main loop tick * or in the next one (via setTimeout(,0)) in case it has already been called * in this tick (in case this is a re-entrant call). */ ; _proto.tick = function tick() { this._tickCallCount++; if (this._tickCallCount === 1) { this.doTick(); // re-entrant call to tick from previous doTick call stack // -> schedule a call on the next main loop iteration to process this task processing request if (this._tickCallCount > 1) { // make sure only one timer exists at any time at max this.clearNextTick(); this._tickTimer = self.setTimeout(this._boundTick, 0); } this._tickCallCount = 0; } } /** * For subclass to implement task logic * @abstract */ ; _proto.doTick = function doTick() {}; return TaskLoop; }(event_handler); // CONCATENATED MODULE: ./src/controller/base-stream-controller.js function base_stream_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } var State = { STOPPED: 'STOPPED', STARTING: 'STARTING', IDLE: 'IDLE', PAUSED: 'PAUSED', KEY_LOADING: 'KEY_LOADING', FRAG_LOADING: 'FRAG_LOADING', FRAG_LOADING_WAITING_RETRY: 'FRAG_LOADING_WAITING_RETRY', WAITING_TRACK: 'WAITING_TRACK', PARSING: 'PARSING', PARSED: 'PARSED', BUFFER_FLUSHING: 'BUFFER_FLUSHING', ENDED: 'ENDED', ERROR: 'ERROR', WAITING_INIT_PTS: 'WAITING_INIT_PTS', WAITING_LEVEL: 'WAITING_LEVEL' }; var base_stream_controller_BaseStreamController = /*#__PURE__*/function (_TaskLoop) { base_stream_controller_inheritsLoose(BaseStreamController, _TaskLoop); function BaseStreamController() { return _TaskLoop.apply(this, arguments) || this; } var _proto = BaseStreamController.prototype; _proto.doTick = function doTick() {}; _proto.startLoad = function startLoad() {}; _proto.stopLoad = function stopLoad() { var frag = this.fragCurrent; if (frag) { if (frag.loader) { frag.loader.abort(); } this.fragmentTracker.removeFragment(frag); } if (this.demuxer) { this.demuxer.destroy(); this.demuxer = null; } this.fragCurrent = null; this.fragPrevious = null; this.clearInterval(); this.clearNextTick(); this.state = State.STOPPED; }; _proto._streamEnded = function _streamEnded(bufferInfo, levelDetails) { var fragCurrent = this.fragCurrent, fragmentTracker = this.fragmentTracker; // we just got done loading the final fragment and there is no other buffered range after ... // rationale is that in case there are any buffered ranges after, it means that there are unbuffered portion in between // so we should not switch to ENDED in that case, to be able to buffer them // dont switch to ENDED if we need to backtrack last fragment if (!levelDetails.live && fragCurrent && !fragCurrent.backtracked && fragCurrent.sn === levelDetails.endSN && !bufferInfo.nextStart) { var fragState = fragmentTracker.getState(fragCurrent); return fragState === FragmentState.PARTIAL || fragState === FragmentState.OK; } return false; }; _proto.onMediaSeeking = function onMediaSeeking() { var config = this.config, media = this.media, mediaBuffer = this.mediaBuffer, state = this.state; var currentTime = media ? media.currentTime : null; var bufferInfo = BufferHelper.bufferInfo(mediaBuffer || media, currentTime, this.config.maxBufferHole); if (Object(number["isFiniteNumber"])(currentTime)) { logger["logger"].log("media seeking to " + currentTime.toFixed(3)); } if (state === State.FRAG_LOADING) { var fragCurrent = this.fragCurrent; // check if we are seeking to a unbuffered area AND if frag loading is in progress if (bufferInfo.len === 0 && fragCurrent) { var tolerance = config.maxFragLookUpTolerance; var fragStartOffset = fragCurrent.start - tolerance; var fragEndOffset = fragCurrent.start + fragCurrent.duration + tolerance; // check if we seek position will be out of currently loaded frag range : if out cancel frag load, if in, don't do anything if (currentTime < fragStartOffset || currentTime > fragEndOffset) { if (fragCurrent.loader) { logger["logger"].log('seeking outside of buffer while fragment load in progress, cancel fragment load'); fragCurrent.loader.abort(); } this.fragCurrent = null; this.fragPrevious = null; // switch to IDLE state to load new fragment this.state = State.IDLE; } else { logger["logger"].log('seeking outside of buffer but within currently loaded fragment range'); } } } else if (state === State.ENDED) { // if seeking to unbuffered area, clean up fragPrevious if (bufferInfo.len === 0) { this.fragPrevious = null; this.fragCurrent = null; } // switch to IDLE state to check for potential new fragment this.state = State.IDLE; } if (media) { this.lastCurrentTime = currentTime; } // in case seeking occurs although no media buffered, adjust startPosition and nextLoadPosition to seek target if (!this.loadedmetadata) { this.nextLoadPosition = this.startPosition = currentTime; } // tick to speed up processing this.tick(); }; _proto.onMediaEnded = function onMediaEnded() { // reset startPosition and lastCurrentTime to restart playback @ stream beginning this.startPosition = this.lastCurrentTime = 0; }; _proto.onHandlerDestroying = function onHandlerDestroying() { this.stopLoad(); _TaskLoop.prototype.onHandlerDestroying.call(this); }; _proto.onHandlerDestroyed = function onHandlerDestroyed() { this.state = State.STOPPED; this.fragmentTracker = null; }; _proto.computeLivePosition = function computeLivePosition(sliding, levelDetails) { var targetLatency = this.config.liveSyncDuration !== undefined ? this.config.liveSyncDuration : this.config.liveSyncDurationCount * levelDetails.targetduration; return sliding + Math.max(0, levelDetails.totalduration - targetLatency); }; return BaseStreamController; }(TaskLoop); // CONCATENATED MODULE: ./src/controller/stream-controller.js function stream_controller_defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } function stream_controller_createClass(Constructor, protoProps, staticProps) { if (protoProps) stream_controller_defineProperties(Constructor.prototype, protoProps); if (staticProps) stream_controller_defineProperties(Constructor, staticProps); return Constructor; } function stream_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /* * Stream Controller */ var TICK_INTERVAL = 100; // how often to tick in ms var stream_controller_StreamController = /*#__PURE__*/function (_BaseStreamController) { stream_controller_inheritsLoose(StreamController, _BaseStreamController); function StreamController(hls, fragmentTracker) { var _this; _this = _BaseStreamController.call(this, hls, events["default"].MEDIA_ATTACHED, events["default"].MEDIA_DETACHING, events["default"].MANIFEST_LOADING, events["default"].MANIFEST_PARSED, events["default"].LEVEL_LOADED, events["default"].LEVELS_UPDATED, events["default"].KEY_LOADED, events["default"].FRAG_LOADED, events["default"].FRAG_LOAD_EMERGENCY_ABORTED, events["default"].FRAG_PARSING_INIT_SEGMENT, events["default"].FRAG_PARSING_DATA, events["default"].FRAG_PARSED, events["default"].ERROR, events["default"].AUDIO_TRACK_SWITCHING, events["default"].AUDIO_TRACK_SWITCHED, events["default"].BUFFER_CREATED, events["default"].BUFFER_APPENDED, events["default"].BUFFER_FLUSHED) || this; _this.fragmentTracker = fragmentTracker; _this.config = hls.config; _this.audioCodecSwap = false; _this._state = State.STOPPED; _this.stallReported = false; _this.gapController = null; _this.altAudio = false; _this.audioOnly = false; _this.bitrateTest = false; return _this; } var _proto = StreamController.prototype; _proto.startLoad = function startLoad(startPosition) { if (this.levels) { var lastCurrentTime = this.lastCurrentTime, hls = this.hls; this.stopLoad(); this.setInterval(TICK_INTERVAL); this.level = -1; this.fragLoadError = 0; if (!this.startFragRequested) { // determine load level var startLevel = hls.startLevel; if (startLevel === -1) { if (hls.config.testBandwidth) { // -1 : guess start Level by doing a bitrate test by loading first fragment of lowest quality level startLevel = 0; this.bitrateTest = true; } else { startLevel = hls.nextAutoLevel; } } // set new level to playlist loader : this will trigger start level load // hls.nextLoadLevel remains until it is set to a new value or until a new frag is successfully loaded this.level = hls.nextLoadLevel = startLevel; this.loadedmetadata = false; } // if startPosition undefined but lastCurrentTime set, set startPosition to last currentTime if (lastCurrentTime > 0 && startPosition === -1) { logger["logger"].log("override startPosition with lastCurrentTime @" + lastCurrentTime.toFixed(3)); startPosition = lastCurrentTime; } this.state = State.IDLE; this.nextLoadPosition = this.startPosition = this.lastCurrentTime = startPosition; this.tick(); } else { this.forceStartLoad = true; this.state = State.STOPPED; } }; _proto.stopLoad = function stopLoad() { this.forceStartLoad = false; _BaseStreamController.prototype.stopLoad.call(this); }; _proto.doTick = function doTick() { switch (this.state) { case State.BUFFER_FLUSHING: // in buffer flushing state, reset fragLoadError counter this.fragLoadError = 0; break; case State.IDLE: this._doTickIdle(); break; case State.WAITING_LEVEL: var level = this.levels[this.level]; // check if playlist is already loaded if (level && level.details) { this.state = State.IDLE; } break; case State.FRAG_LOADING_WAITING_RETRY: var now = window.performance.now(); var retryDate = this.retryDate; // if current time is gt than retryDate, or if media seeking let's switch to IDLE state to retry loading if (!retryDate || now >= retryDate || this.media && this.media.seeking) { logger["logger"].log('mediaController: retryDate reached, switch back to IDLE state'); this.state = State.IDLE; } break; case State.ERROR: case State.STOPPED: case State.FRAG_LOADING: case State.PARSING: case State.PARSED: case State.ENDED: break; default: break; } // check buffer this._checkBuffer(); // check/update current fragment this._checkFragmentChanged(); } // Ironically the "idle" state is the on we do the most logic in it seems .... // NOTE: Maybe we could rather schedule a check for buffer length after half of the currently // played segment, or on pause/play/seek instead of naively checking every 100ms? ; _proto._doTickIdle = function _doTickIdle() { var hls = this.hls, config = hls.config, media = this.media; // if start level not parsed yet OR // if video not attached AND start fragment already requested OR start frag prefetch disable // exit loop, as we either need more info (level not parsed) or we need media to be attached to load new fragment if (this.levelLastLoaded === undefined || !media && (this.startFragRequested || !config.startFragPrefetch)) { return; } // If the "main" level is audio-only but we are loading an alternate track in the same group, do not load anything if (this.altAudio && this.audioOnly) { // Clear audio demuxer state so when switching back to main audio we're not still appending where we left off this.demuxer.frag = null; return; } // if we have not yet loaded any fragment, start loading from start position var pos; if (this.loadedmetadata) { pos = media.currentTime; } else { pos = this.nextLoadPosition; } // determine next load level var level = hls.nextLoadLevel, levelInfo = this.levels[level]; if (!levelInfo) { return; } var levelBitrate = levelInfo.bitrate, maxBufLen; // compute max Buffer Length that we could get from this load level, based on level bitrate. if (levelBitrate) { maxBufLen = Math.max(8 * config.maxBufferSize / levelBitrate, config.maxBufferLength); } else { maxBufLen = config.maxBufferLength; } maxBufLen = Math.min(maxBufLen, config.maxMaxBufferLength); // determine next candidate fragment to be loaded, based on current position and end of buffer position // ensure up to `config.maxMaxBufferLength` of buffer upfront var maxBufferHole = pos < config.maxBufferHole ? Math.max(MAX_START_GAP_JUMP, config.maxBufferHole) : config.maxBufferHole; var bufferInfo = BufferHelper.bufferInfo(this.mediaBuffer ? this.mediaBuffer : media, pos, maxBufferHole); var bufferLen = bufferInfo.len; // Stay idle if we are still with buffer margins if (bufferLen >= maxBufLen) { return; } // if buffer length is less than maxBufLen try to load a new fragment ... logger["logger"].trace("buffer length of " + bufferLen.toFixed(3) + " is below max of " + maxBufLen.toFixed(3) + ". checking for more payload ..."); // set next load level : this will trigger a playlist load if needed this.level = hls.nextLoadLevel = level; var levelDetails = levelInfo.details; // if level info not retrieved yet, switch state and wait for level retrieval // if live playlist, ensure that new playlist has been refreshed to avoid loading/try to load // a useless and outdated fragment (that might even introduce load error if it is already out of the live playlist) if (!levelDetails || levelDetails.live && this.levelLastLoaded !== level) { this.state = State.WAITING_LEVEL; return; } if (this._streamEnded(bufferInfo, levelDetails)) { var data = {}; if (this.altAudio) { data.type = 'video'; } this.hls.trigger(events["default"].BUFFER_EOS, data); this.state = State.ENDED; return; } // if we have the levelDetails for the selected variant, lets continue enrichen our stream (load keys/fragments or trigger EOS, etc..) this._fetchPayloadOrEos(pos, bufferInfo, levelDetails); }; _proto._fetchPayloadOrEos = function _fetchPayloadOrEos(pos, bufferInfo, levelDetails) { var fragPrevious = this.fragPrevious, level = this.level, fragments = levelDetails.fragments, fragLen = fragments.length; // empty playlist if (fragLen === 0) { return; } // find fragment index, contiguous with end of buffer position var start = fragments[0].start, end = fragments[fragLen - 1].start + fragments[fragLen - 1].duration, bufferEnd = bufferInfo.end, frag; if (levelDetails.initSegment && !levelDetails.initSegment.data) { frag = levelDetails.initSegment; } else { // in case of live playlist we need to ensure that requested position is not located before playlist start if (levelDetails.live) { var initialLiveManifestSize = this.config.initialLiveManifestSize; if (fragLen < initialLiveManifestSize) { logger["logger"].warn("Can not start playback of a level, reason: not enough fragments " + fragLen + " < " + initialLiveManifestSize); return; } frag = this._ensureFragmentAtLivePoint(levelDetails, bufferEnd, start, end, fragPrevious, fragments); // if it explicitely returns null don't load any fragment and exit function now if (frag === null) { return; } } else { // VoD playlist: if bufferEnd before start of playlist, load first fragment if (bufferEnd < start) { frag = fragments[0]; } } } if (!frag) { frag = this._findFragment(start, fragPrevious, fragLen, fragments, bufferEnd, end, levelDetails); } if (frag) { if (frag.encrypted) { this._loadKey(frag, levelDetails); } else { this._loadFragment(frag, levelDetails, pos, bufferEnd); } } }; _proto._ensureFragmentAtLivePoint = function _ensureFragmentAtLivePoint(levelDetails, bufferEnd, start, end, fragPrevious, fragments) { var config = this.hls.config, media = this.media; var frag; // check if requested position is within seekable boundaries : // logger.log(`start/pos/bufEnd/seeking:${start.toFixed(3)}/${pos.toFixed(3)}/${bufferEnd.toFixed(3)}/${this.media.seeking}`); var maxLatency = Infinity; if (config.liveMaxLatencyDuration !== undefined) { maxLatency = config.liveMaxLatencyDuration; } else if (Object(number["isFiniteNumber"])(config.liveMaxLatencyDurationCount)) { maxLatency = config.liveMaxLatencyDurationCount * levelDetails.targetduration; } if (bufferEnd < Math.max(start - config.maxFragLookUpTolerance, end - maxLatency)) { var liveSyncPosition = this.liveSyncPosition = this.computeLivePosition(start, levelDetails); bufferEnd = liveSyncPosition; if (media && !media.paused && media.readyState && media.duration > liveSyncPosition && liveSyncPosition > media.currentTime) { logger["logger"].log("buffer end: " + bufferEnd.toFixed(3) + " is located too far from the end of live sliding playlist, reset currentTime to : " + liveSyncPosition.toFixed(3)); media.currentTime = liveSyncPosition; } this.nextLoadPosition = liveSyncPosition; } // if end of buffer greater than live edge, don't load any fragment // this could happen if live playlist intermittently slides in the past. // level 1 loaded [182580161,182580167] // level 1 loaded [182580162,182580169] // Loading 182580168 of [182580162 ,182580169],level 1 .. // Loading 182580169 of [182580162 ,182580169],level 1 .. // level 1 loaded [182580162,182580168] <============= here we should have bufferEnd > end. in that case break to avoid reloading 182580168 // level 1 loaded [182580164,182580171] // // don't return null in case media not loaded yet (readystate === 0) if (levelDetails.PTSKnown && bufferEnd > end && media && media.readyState) { return null; } if (this.startFragRequested && !levelDetails.PTSKnown) { /* we are switching level on live playlist, but we don't have any PTS info for that quality level ... try to load frag matching with next SN. even if SN are not synchronized between playlists, loading this frag will help us compute playlist sliding and find the right one after in case it was not the right consecutive one */ if (fragPrevious) { if (levelDetails.hasProgramDateTime) { // Relies on PDT in order to switch bitrates (Support EXT-X-DISCONTINUITY without EXT-X-DISCONTINUITY-SEQUENCE) logger["logger"].log("live playlist, switching playlist, load frag with same PDT: " + fragPrevious.programDateTime); frag = findFragmentByPDT(fragments, fragPrevious.endProgramDateTime, config.maxFragLookUpTolerance); } else { // Uses buffer and sequence number to calculate switch segment (required if using EXT-X-DISCONTINUITY-SEQUENCE) var targetSN = fragPrevious.sn + 1; if (targetSN >= levelDetails.startSN && targetSN <= levelDetails.endSN) { var fragNext = fragments[targetSN - levelDetails.startSN]; if (fragPrevious.cc === fragNext.cc) { frag = fragNext; logger["logger"].log("live playlist, switching playlist, load frag with next SN: " + frag.sn); } } // next frag SN not available (or not with same continuity counter) // look for a frag sharing the same CC if (!frag) { frag = binary_search.search(fragments, function (frag) { return fragPrevious.cc - frag.cc; }); if (frag) { logger["logger"].log("live playlist, switching playlist, load frag with same CC: " + frag.sn); } } } } } return frag; }; _proto._findFragment = function _findFragment(start, fragPreviousLoad, fragmentIndexRange, fragments, bufferEnd, end, levelDetails) { var config = this.hls.config; var fragNextLoad; if (bufferEnd < end) { var lookupTolerance = bufferEnd > end - config.maxFragLookUpTolerance ? 0 : config.maxFragLookUpTolerance; // Remove the tolerance if it would put the bufferEnd past the actual end of stream // Uses buffer and sequence number to calculate switch segment (required if using EXT-X-DISCONTINUITY-SEQUENCE) fragNextLoad = findFragmentByPTS(fragPreviousLoad, fragments, bufferEnd, lookupTolerance); } else { // reach end of playlist fragNextLoad = fragments[fragmentIndexRange - 1]; } if (fragNextLoad) { var curSNIdx = fragNextLoad.sn - levelDetails.startSN; var sameLevel = fragPreviousLoad && fragNextLoad.level === fragPreviousLoad.level; var prevSnFrag = fragments[curSNIdx - 1]; var nextSnFrag = fragments[curSNIdx + 1]; // logger.log('find SN matching with pos:' + bufferEnd + ':' + frag.sn); if (fragPreviousLoad && fragNextLoad.sn === fragPreviousLoad.sn) { if (sameLevel && !fragNextLoad.backtracked) { if (fragNextLoad.sn < levelDetails.endSN) { var deltaPTS = fragPreviousLoad.deltaPTS; // if there is a significant delta between audio and video, larger than max allowed hole, // and if previous remuxed fragment did not start with a keyframe. (fragPrevious.dropped) // let's try to load previous fragment again to get last keyframe // then we will reload again current fragment (that way we should be able to fill the buffer hole ...) if (deltaPTS && deltaPTS > config.maxBufferHole && fragPreviousLoad.dropped && curSNIdx) { fragNextLoad = prevSnFrag; logger["logger"].warn('Previous fragment was dropped with large PTS gap between audio and video. Maybe fragment is not starting with a keyframe? Loading previous one to try to overcome this'); } else { fragNextLoad = nextSnFrag; logger["logger"].log("Re-loading fragment with SN: " + fragNextLoad.sn); } } else { fragNextLoad = null; } } else if (fragNextLoad.backtracked) { // Only backtrack a max of 1 consecutive fragment to prevent sliding back too far when little or no frags start with keyframes if (nextSnFrag && nextSnFrag.backtracked) { logger["logger"].warn("Already backtracked from fragment " + nextSnFrag.sn + ", will not backtrack to fragment " + fragNextLoad.sn + ". Loading fragment " + nextSnFrag.sn); fragNextLoad = nextSnFrag; } else { // If a fragment has dropped frames and it's in a same level/sequence, load the previous fragment to try and find the keyframe // Reset the dropped count now since it won't be reset until we parse the fragment again, which prevents infinite backtracking on the same segment logger["logger"].warn('Loaded fragment with dropped frames, backtracking 1 segment to find a keyframe'); fragNextLoad.dropped = 0; if (prevSnFrag) { fragNextLoad = prevSnFrag; fragNextLoad.backtracked = true; } else if (curSNIdx) { // can't backtrack on very first fragment fragNextLoad = null; } } } } } return fragNextLoad; }; _proto._loadKey = function _loadKey(frag, levelDetails) { logger["logger"].log("Loading key for " + frag.sn + " of [" + levelDetails.startSN + " ," + levelDetails.endSN + "],level " + this.level); this.state = State.KEY_LOADING; this.hls.trigger(events["default"].KEY_LOADING, { frag: frag }); }; _proto._loadFragment = function _loadFragment(frag, levelDetails, pos, bufferEnd) { // Check if fragment is not loaded var fragState = this.fragmentTracker.getState(frag); this.fragCurrent = frag; if (frag.sn !== 'initSegment') { this.startFragRequested = true; } // Don't update nextLoadPosition for fragments which are not buffered if (Object(number["isFiniteNumber"])(frag.sn) && !frag.bitrateTest) { this.nextLoadPosition = frag.start + frag.duration; } // Allow backtracked fragments to load if (frag.backtracked || fragState === FragmentState.NOT_LOADED || fragState === FragmentState.PARTIAL) { frag.autoLevel = this.hls.autoLevelEnabled; frag.bitrateTest = this.bitrateTest; logger["logger"].log("Loading " + frag.sn + " of [" + levelDetails.startSN + " ," + levelDetails.endSN + "],level " + this.level + ", currentTime:" + pos.toFixed(3) + ",bufferEnd:" + bufferEnd.toFixed(3)); this.hls.trigger(events["default"].FRAG_LOADING, { frag: frag }); // lazy demuxer init, as this could take some time ... do it during frag loading if (!this.demuxer) { this.demuxer = new demux_demuxer(this.hls, 'main'); } this.state = State.FRAG_LOADING; } else if (fragState === FragmentState.APPENDING) { // Lower the buffer size and try again if (this._reduceMaxBufferLength(frag.duration)) { this.fragmentTracker.removeFragment(frag); } } }; _proto.getBufferedFrag = function getBufferedFrag(position) { return this.fragmentTracker.getBufferedFrag(position, PlaylistLevelType.MAIN); }; _proto.followingBufferedFrag = function followingBufferedFrag(frag) { if (frag) { // try to get range of next fragment (500ms after this range) return this.getBufferedFrag(frag.endPTS + 0.5); } return null; }; _proto._checkFragmentChanged = function _checkFragmentChanged() { var fragPlayingCurrent, currentTime, video = this.media; if (video && video.readyState && video.seeking === false) { currentTime = video.currentTime; /* if video element is in seeked state, currentTime can only increase. (assuming that playback rate is positive ...) As sometimes currentTime jumps back to zero after a media decode error, check this, to avoid seeking back to wrong position after a media decode error */ if (currentTime > this.lastCurrentTime) { this.lastCurrentTime = currentTime; } if (BufferHelper.isBuffered(video, currentTime)) { fragPlayingCurrent = this.getBufferedFrag(currentTime); } else if (BufferHelper.isBuffered(video, currentTime + 0.1)) { /* ensure that FRAG_CHANGED event is triggered at startup, when first video frame is displayed and playback is paused. add a tolerance of 100ms, in case current position is not buffered, check if current pos+100ms is buffered and use that buffer range for FRAG_CHANGED event reporting */ fragPlayingCurrent = this.getBufferedFrag(currentTime + 0.1); } if (fragPlayingCurrent) { var fragPlaying = fragPlayingCurrent; if (fragPlaying !== this.fragPlaying) { this.hls.trigger(events["default"].FRAG_CHANGED, { frag: fragPlaying }); var fragPlayingLevel = fragPlaying.level; if (!this.fragPlaying || this.fragPlaying.level !== fragPlayingLevel) { this.hls.trigger(events["default"].LEVEL_SWITCHED, { level: fragPlayingLevel }); } this.fragPlaying = fragPlaying; } } } } /* on immediate level switch : - pause playback if playing - cancel any pending load request - and trigger a buffer flush */ ; _proto.immediateLevelSwitch = function immediateLevelSwitch() { logger["logger"].log('immediateLevelSwitch'); if (!this.immediateSwitch) { this.immediateSwitch = true; var media = this.media, previouslyPaused; if (media) { previouslyPaused = media.paused; media.pause(); } else { // don't restart playback after instant level switch in case media not attached previouslyPaused = true; } this.previouslyPaused = previouslyPaused; } var fragCurrent = this.fragCurrent; if (fragCurrent && fragCurrent.loader) { fragCurrent.loader.abort(); } this.fragCurrent = null; // flush everything this.flushMainBuffer(0, Number.POSITIVE_INFINITY); } /** * on immediate level switch end, after new fragment has been buffered: * - nudge video decoder by slightly adjusting video currentTime (if currentTime buffered) * - resume the playback if needed */ ; _proto.immediateLevelSwitchEnd = function immediateLevelSwitchEnd() { var media = this.media; if (media && media.buffered.length) { this.immediateSwitch = false; if (BufferHelper.isBuffered(media, media.currentTime)) { // only nudge if currentTime is buffered media.currentTime -= 0.0001; } if (!this.previouslyPaused) { media.play(); } } } /** * try to switch ASAP without breaking video playback: * in order to ensure smooth but quick level switching, * we need to find the next flushable buffer range * we should take into account new segment fetch time */ ; _proto.nextLevelSwitch = function nextLevelSwitch() { var media = this.media; // ensure that media is defined and that metadata are available (to retrieve currentTime) if (media && media.readyState) { var fetchdelay, fragPlayingCurrent, nextBufferedFrag; fragPlayingCurrent = this.getBufferedFrag(media.currentTime); if (fragPlayingCurrent && fragPlayingCurrent.startPTS > 1) { // flush buffer preceding current fragment (flush until current fragment start offset) // minus 1s to avoid video freezing, that could happen if we flush keyframe of current video ... this.flushMainBuffer(0, fragPlayingCurrent.startPTS - 1); } if (!media.paused) { // add a safety delay of 1s var nextLevelId = this.hls.nextLoadLevel, nextLevel = this.levels[nextLevelId], fragLastKbps = this.fragLastKbps; if (fragLastKbps && this.fragCurrent) { fetchdelay = this.fragCurrent.duration * nextLevel.bitrate / (1000 * fragLastKbps) + 1; } else { fetchdelay = 0; } } else { fetchdelay = 0; } // logger.log('fetchdelay:'+fetchdelay); // find buffer range that will be reached once new fragment will be fetched nextBufferedFrag = this.getBufferedFrag(media.currentTime + fetchdelay); if (nextBufferedFrag) { // we can flush buffer range following this one without stalling playback nextBufferedFrag = this.followingBufferedFrag(nextBufferedFrag); if (nextBufferedFrag) { // if we are here, we can also cancel any loading/demuxing in progress, as they are useless var fragCurrent = this.fragCurrent; if (fragCurrent && fragCurrent.loader) { fragCurrent.loader.abort(); } this.fragCurrent = null; // start flush position is the start PTS of next buffered frag. // we use frag.naxStartPTS which is max(audio startPTS, video startPTS). // in case there is a small PTS Delta between audio and video, using maxStartPTS avoids flushing last samples from current fragment this.flushMainBuffer(nextBufferedFrag.maxStartPTS, Number.POSITIVE_INFINITY); } } } }; _proto.flushMainBuffer = function flushMainBuffer(startOffset, endOffset) { this.state = State.BUFFER_FLUSHING; var flushScope = { startOffset: startOffset, endOffset: endOffset }; // if alternate audio tracks are used, only flush video, otherwise flush everything if (this.altAudio) { flushScope.type = 'video'; } this.hls.trigger(events["default"].BUFFER_FLUSHING, flushScope); }; _proto.onMediaAttached = function onMediaAttached(data) { var media = this.media = this.mediaBuffer = data.media; this.onvseeking = this.onMediaSeeking.bind(this); this.onvseeked = this.onMediaSeeked.bind(this); this.onvended = this.onMediaEnded.bind(this); media.addEventListener('seeking', this.onvseeking); media.addEventListener('seeked', this.onvseeked); media.addEventListener('ended', this.onvended); var config = this.config; if (this.levels && config.autoStartLoad) { this.hls.startLoad(config.startPosition); } this.gapController = new gap_controller_GapController(config, media, this.fragmentTracker, this.hls); }; _proto.onMediaDetaching = function onMediaDetaching() { var media = this.media; if (media && media.ended) { logger["logger"].log('MSE detaching and video ended, reset startPosition'); this.startPosition = this.lastCurrentTime = 0; } // reset fragment backtracked flag var levels = this.levels; if (levels) { levels.forEach(function (level) { if (level.details) { level.details.fragments.forEach(function (fragment) { fragment.backtracked = undefined; }); } }); } // remove video listeners if (media) { media.removeEventListener('seeking', this.onvseeking); media.removeEventListener('seeked', this.onvseeked); media.removeEventListener('ended', this.onvended); this.onvseeking = this.onvseeked = this.onvended = null; } this.fragmentTracker.removeAllFragments(); this.media = this.mediaBuffer = null; this.loadedmetadata = false; this.stopLoad(); }; _proto.onMediaSeeked = function onMediaSeeked() { var media = this.media; var currentTime = media ? media.currentTime : undefined; if (Object(number["isFiniteNumber"])(currentTime)) { logger["logger"].log("media seeked to " + currentTime.toFixed(3)); } // tick to speed up FRAGMENT_PLAYING triggering this.tick(); }; _proto.onManifestLoading = function onManifestLoading() { // reset buffer on manifest loading logger["logger"].log('trigger BUFFER_RESET'); this.hls.trigger(events["default"].BUFFER_RESET); this.fragmentTracker.removeAllFragments(); this.stalled = false; this.startPosition = this.lastCurrentTime = 0; }; _proto.onManifestParsed = function onManifestParsed(data) { var aac = false, heaac = false, codec; data.levels.forEach(function (level) { // detect if we have different kind of audio codecs used amongst playlists codec = level.audioCodec; if (codec) { if (codec.indexOf('mp4a.40.2') !== -1) { aac = true; } if (codec.indexOf('mp4a.40.5') !== -1) { heaac = true; } } }); this.audioCodecSwitch = aac && heaac; if (this.audioCodecSwitch) { logger["logger"].log('both AAC/HE-AAC audio found in levels; declaring level codec as HE-AAC'); } this.altAudio = data.altAudio; this.levels = data.levels; this.startFragRequested = false; var config = this.config; if (config.autoStartLoad || this.forceStartLoad) { this.hls.startLoad(config.startPosition); } }; _proto.onLevelLoaded = function onLevelLoaded(data) { var newDetails = data.details; var newLevelId = data.level; var lastLevel = this.levels[this.levelLastLoaded]; var curLevel = this.levels[newLevelId]; var duration = newDetails.totalduration; var sliding = 0; logger["logger"].log("level " + newLevelId + " loaded [" + newDetails.startSN + "," + newDetails.endSN + "],duration:" + duration); if (newDetails.live) { var curDetails = curLevel.details; if (curDetails && newDetails.fragments.length > 0) { // we already have details for that level, merge them mergeDetails(curDetails, newDetails); sliding = newDetails.fragments[0].start; this.liveSyncPosition = this.computeLivePosition(sliding, curDetails); if (newDetails.PTSKnown && Object(number["isFiniteNumber"])(sliding)) { logger["logger"].log("live playlist sliding:" + sliding.toFixed(3)); } else { logger["logger"].log('live playlist - outdated PTS, unknown sliding'); alignStream(this.fragPrevious, lastLevel, newDetails); } } else { logger["logger"].log('live playlist - first load, unknown sliding'); newDetails.PTSKnown = false; alignStream(this.fragPrevious, lastLevel, newDetails); } } else { newDetails.PTSKnown = false; } // override level info curLevel.details = newDetails; this.levelLastLoaded = newLevelId; this.hls.trigger(events["default"].LEVEL_UPDATED, { details: newDetails, level: newLevelId }); if (this.startFragRequested === false) { // compute start position if set to -1. use it straight away if value is defined if (this.startPosition === -1 || this.lastCurrentTime === -1) { // first, check if start time offset has been set in playlist, if yes, use this value var startTimeOffset = newDetails.startTimeOffset; if (Object(number["isFiniteNumber"])(startTimeOffset)) { if (startTimeOffset < 0) { logger["logger"].log("negative start time offset " + startTimeOffset + ", count from end of last fragment"); startTimeOffset = sliding + duration + startTimeOffset; } logger["logger"].log("start time offset found in playlist, adjust startPosition to " + startTimeOffset); this.startPosition = startTimeOffset; } else { // if live playlist, set start position to be fragment N-this.config.liveSyncDurationCount (usually 3) if (newDetails.live) { this.startPosition = this.computeLivePosition(sliding, newDetails); logger["logger"].log("configure startPosition to " + this.startPosition); } else { this.startPosition = 0; } } this.lastCurrentTime = this.startPosition; } this.nextLoadPosition = this.startPosition; } // only switch batck to IDLE state if we were waiting for level to start downloading a new fragment if (this.state === State.WAITING_LEVEL) { this.state = State.IDLE; } // trigger handler right now this.tick(); }; _proto.onKeyLoaded = function onKeyLoaded() { if (this.state === State.KEY_LOADING) { this.state = State.IDLE; this.tick(); } }; _proto.onFragLoaded = function onFragLoaded(data) { var fragCurrent = this.fragCurrent, hls = this.hls, levels = this.levels, media = this.media; var fragLoaded = data.frag; if (this.state === State.FRAG_LOADING && fragCurrent && fragLoaded.type === 'main' && fragLoaded.level === fragCurrent.level && fragLoaded.sn === fragCurrent.sn) { var stats = data.stats; var currentLevel = levels[fragCurrent.level]; var details = currentLevel.details; // reset frag bitrate test in any case after frag loaded event // if this frag was loaded to perform a bitrate test AND if hls.nextLoadLevel is greater than 0 // then this means that we should be able to load a fragment at a higher quality level this.bitrateTest = false; this.stats = stats; logger["logger"].log("Loaded " + fragCurrent.sn + " of [" + details.startSN + " ," + details.endSN + "],level " + fragCurrent.level); if (fragLoaded.bitrateTest && hls.nextLoadLevel) { // switch back to IDLE state ... we just loaded a fragment to determine adequate start bitrate and initialize autoswitch algo this.state = State.IDLE; this.startFragRequested = false; stats.tparsed = stats.tbuffered = window.performance.now(); hls.trigger(events["default"].FRAG_BUFFERED, { stats: stats, frag: fragCurrent, id: 'main' }); this.tick(); } else if (fragLoaded.sn === 'initSegment') { this.state = State.IDLE; stats.tparsed = stats.tbuffered = window.performance.now(); details.initSegment.data = data.payload; hls.trigger(events["default"].FRAG_BUFFERED, { stats: stats, frag: fragCurrent, id: 'main' }); this.tick(); } else { logger["logger"].log("Parsing " + fragCurrent.sn + " of [" + details.startSN + " ," + details.endSN + "],level " + fragCurrent.level + ", cc " + fragCurrent.cc); this.state = State.PARSING; this.pendingBuffering = true; this.appended = false; // Bitrate test frags are not usually buffered so the fragment tracker ignores them. If Hls.js decides to buffer // it (and therefore ends up at this line), then the fragment tracker needs to be manually informed. if (fragLoaded.bitrateTest) { fragLoaded.bitrateTest = false; this.fragmentTracker.onFragLoaded({ frag: fragLoaded }); } // time Offset is accurate if level PTS is known, or if playlist is not sliding (not live) and if media is not seeking (this is to overcome potential timestamp drifts between playlists and fragments) var accurateTimeOffset = !(media && media.seeking) && (details.PTSKnown || !details.live); var initSegmentData = details.initSegment ? details.initSegment.data : []; var audioCodec = this._getAudioCodec(currentLevel); // transmux the MPEG-TS data to ISO-BMFF segments var demuxer = this.demuxer = this.demuxer || new demux_demuxer(this.hls, 'main'); demuxer.push(data.payload, initSegmentData, audioCodec, currentLevel.videoCodec, fragCurrent, details.totalduration, accurateTimeOffset); } } this.fragLoadError = 0; }; _proto.onFragParsingInitSegment = function onFragParsingInitSegment(data) { var fragCurrent = this.fragCurrent; var fragNew = data.frag; if (fragCurrent && data.id === 'main' && fragNew.sn === fragCurrent.sn && fragNew.level === fragCurrent.level && this.state === State.PARSING) { var tracks = data.tracks, trackName, track; this.audioOnly = tracks.audio && !tracks.video; // if audio track is expected to come from audio stream controller, discard any coming from main if (this.altAudio && !this.audioOnly) { delete tracks.audio; } // include levelCodec in audio and video tracks track = tracks.audio; if (track) { var audioCodec = this.levels[this.level].audioCodec, ua = navigator.userAgent.toLowerCase(); if (audioCodec && this.audioCodecSwap) { logger["logger"].log('swapping playlist audio codec'); if (audioCodec.indexOf('mp4a.40.5') !== -1) { audioCodec = 'mp4a.40.2'; } else { audioCodec = 'mp4a.40.5'; } } // in case AAC and HE-AAC audio codecs are signalled in manifest // force HE-AAC , as it seems that most browsers prefers that way, // except for mono streams OR on FF // these conditions might need to be reviewed ... if (this.audioCodecSwitch) { // don't force HE-AAC if mono stream if (track.metadata.channelCount !== 1 && // don't force HE-AAC if firefox ua.indexOf('firefox') === -1) { audioCodec = 'mp4a.40.5'; } } // HE-AAC is broken on Android, always signal audio codec as AAC even if variant manifest states otherwise if (ua.indexOf('android') !== -1 && track.container !== 'audio/mpeg') { // Exclude mpeg audio audioCodec = 'mp4a.40.2'; logger["logger"].log("Android: force audio codec to " + audioCodec); } track.levelCodec = audioCodec; track.id = data.id; } track = tracks.video; if (track) { track.levelCodec = this.levels[this.level].videoCodec; track.id = data.id; } this.hls.trigger(events["default"].BUFFER_CODECS, tracks); // loop through tracks that are going to be provided to bufferController for (trackName in tracks) { track = tracks[trackName]; logger["logger"].log("main track:" + trackName + ",container:" + track.container + ",codecs[level/parsed]=[" + track.levelCodec + "/" + track.codec + "]"); var initSegment = track.initSegment; if (initSegment) { this.appended = true; // arm pending Buffering flag before appending a segment this.pendingBuffering = true; this.hls.trigger(events["default"].BUFFER_APPENDING, { type: trackName, data: initSegment, parent: 'main', content: 'initSegment' }); } } // trigger handler right now this.tick(); } }; _proto.onFragParsingData = function onFragParsingData(data) { var _this2 = this; var fragCurrent = this.fragCurrent; var fragNew = data.frag; if (fragCurrent && data.id === 'main' && fragNew.sn === fragCurrent.sn && fragNew.level === fragCurrent.level && !(data.type === 'audio' && this.altAudio) && // filter out main audio if audio track is loaded through audio stream controller this.state === State.PARSING) { var level = this.levels[this.level], frag = fragCurrent; if (!Object(number["isFiniteNumber"])(data.endPTS)) { data.endPTS = data.startPTS + fragCurrent.duration; data.endDTS = data.startDTS + fragCurrent.duration; } if (data.hasAudio === true) { frag.addElementaryStream(ElementaryStreamTypes.AUDIO); } if (data.hasVideo === true) { frag.addElementaryStream(ElementaryStreamTypes.VIDEO); } logger["logger"].log("Parsed " + data.type + ",PTS:[" + data.startPTS.toFixed(3) + "," + data.endPTS.toFixed(3) + "],DTS:[" + data.startDTS.toFixed(3) + "/" + data.endDTS.toFixed(3) + "],nb:" + data.nb + ",dropped:" + (data.dropped || 0)); // Detect gaps in a fragment and try to fix it by finding a keyframe in the previous fragment (see _findFragments) if (data.type === 'video') { frag.dropped = data.dropped; if (frag.dropped) { if (!frag.backtracked) { var levelDetails = level.details; if (levelDetails && frag.sn === levelDetails.startSN) { logger["logger"].warn('missing video frame(s) on first frag, appending with gap', frag.sn); } else { logger["logger"].warn('missing video frame(s), backtracking fragment', frag.sn); // Return back to the IDLE state without appending to buffer // Causes findFragments to backtrack a segment and find the keyframe // Audio fragments arriving before video sets the nextLoadPosition, causing _findFragments to skip the backtracked fragment this.fragmentTracker.removeFragment(frag); frag.backtracked = true; this.nextLoadPosition = data.startPTS; this.state = State.IDLE; this.fragPrevious = frag; if (this.demuxer) { this.demuxer.destroy(); this.demuxer = null; } this.tick(); return; } } else { logger["logger"].warn('Already backtracked on this fragment, appending with the gap', frag.sn); } } else { // Only reset the backtracked flag if we've loaded the frag without any dropped frames frag.backtracked = false; } } var drift = updateFragPTSDTS(level.details, frag, data.startPTS, data.endPTS, data.startDTS, data.endDTS), hls = this.hls; hls.trigger(events["default"].LEVEL_PTS_UPDATED, { details: level.details, level: this.level, drift: drift, type: data.type, start: data.startPTS, end: data.endPTS }); // has remuxer dropped video frames located before first keyframe ? [data.data1, data.data2].forEach(function (buffer) { // only append in PARSING state (rationale is that an appending error could happen synchronously on first segment appending) // in that case it is useless to append following segments if (buffer && buffer.length && _this2.state === State.PARSING) { _this2.appended = true; // arm pending Buffering flag before appending a segment _this2.pendingBuffering = true; hls.trigger(events["default"].BUFFER_APPENDING, { type: data.type, data: buffer, parent: 'main', content: 'data' }); } }); // trigger handler right now this.tick(); } }; _proto.onFragParsed = function onFragParsed(data) { var fragCurrent = this.fragCurrent; var fragNew = data.frag; if (fragCurrent && data.id === 'main' && fragNew.sn === fragCurrent.sn && fragNew.level === fragCurrent.level && this.state === State.PARSING) { this.stats.tparsed = window.performance.now(); this.state = State.PARSED; this._checkAppendedParsed(); } }; _proto.onAudioTrackSwitching = function onAudioTrackSwitching(data) { // if any URL found on new audio track, it is an alternate audio track var altAudio = !!data.url, trackId = data.id; // if we switch on main audio, ensure that main fragment scheduling is synced with media.buffered // don't do anything if we switch to alt audio: audio stream controller is handling it. // we will just have to change buffer scheduling on audioTrackSwitched if (!altAudio) { if (this.mediaBuffer !== this.media) { logger["logger"].log('switching on main audio, use media.buffered to schedule main fragment loading'); this.mediaBuffer = this.media; var fragCurrent = this.fragCurrent; // we need to refill audio buffer from main: cancel any frag loading to speed up audio switch if (fragCurrent.loader) { logger["logger"].log('switching to main audio track, cancel main fragment load'); fragCurrent.loader.abort(); } this.fragCurrent = null; this.fragPrevious = null; // destroy demuxer to force init segment generation (following audio switch) if (this.demuxer) { this.demuxer.destroy(); this.demuxer = null; } // switch to IDLE state to load new fragment this.state = State.IDLE; } var hls = this.hls; // switching to main audio, flush all audio and trigger track switched hls.trigger(events["default"].BUFFER_FLUSHING, { startOffset: 0, endOffset: Number.POSITIVE_INFINITY, type: 'audio' }); hls.trigger(events["default"].AUDIO_TRACK_SWITCHED, { id: trackId }); this.altAudio = false; } }; _proto.onAudioTrackSwitched = function onAudioTrackSwitched(data) { var trackId = data.id, altAudio = !!this.hls.audioTracks[trackId].url; if (altAudio) { var videoBuffer = this.videoBuffer; // if we switched on alternate audio, ensure that main fragment scheduling is synced with video sourcebuffer buffered if (videoBuffer && this.mediaBuffer !== videoBuffer) { logger["logger"].log('switching on alternate audio, use video.buffered to schedule main fragment loading'); this.mediaBuffer = videoBuffer; } } this.altAudio = altAudio; this.tick(); }; _proto.onBufferCreated = function onBufferCreated(data) { var tracks = data.tracks, mediaTrack, name, alternate = false; for (var type in tracks) { var track = tracks[type]; if (track.id === 'main') { name = type; mediaTrack = track; // keep video source buffer reference if (type === 'video') { this.videoBuffer = tracks[type].buffer; } } else { alternate = true; } } if (alternate && mediaTrack) { logger["logger"].log("alternate track found, use " + name + ".buffered to schedule main fragment loading"); this.mediaBuffer = mediaTrack.buffer; } else { this.mediaBuffer = this.media; } }; _proto.onBufferAppended = function onBufferAppended(data) { if (data.parent === 'main') { var state = this.state; if (state === State.PARSING || state === State.PARSED) { // check if all buffers have been appended this.pendingBuffering = data.pending > 0; this._checkAppendedParsed(); } } }; _proto._checkAppendedParsed = function _checkAppendedParsed() { // trigger handler right now if (this.state === State.PARSED && (!this.appended || !this.pendingBuffering)) { var frag = this.fragCurrent; if (frag) { var media = this.mediaBuffer ? this.mediaBuffer : this.media; logger["logger"].log("main buffered : " + time_ranges.toString(media.buffered)); this.fragPrevious = frag; var stats = this.stats; stats.tbuffered = window.performance.now(); // we should get rid of this.fragLastKbps this.fragLastKbps = Math.round(8 * stats.total / (stats.tbuffered - stats.tfirst)); this.hls.trigger(events["default"].FRAG_BUFFERED, { stats: stats, frag: frag, id: 'main' }); this.state = State.IDLE; } // Do not tick when _seekToStartPos needs to be called as seeking to the start can fail on live streams at this point if (this.loadedmetadata || this.startPosition <= 0) { this.tick(); } } }; _proto.onError = function onError(data) { var frag = data.frag || this.fragCurrent; // don't handle frag error not related to main fragment if (frag && frag.type !== 'main') { return; } // 0.5 : tolerance needed as some browsers stalls playback before reaching buffered end var mediaBuffered = !!this.media && BufferHelper.isBuffered(this.media, this.media.currentTime) && BufferHelper.isBuffered(this.media, this.media.currentTime + 0.5); switch (data.details) { case errors["ErrorDetails"].FRAG_LOAD_ERROR: case errors["ErrorDetails"].FRAG_LOAD_TIMEOUT: case errors["ErrorDetails"].KEY_LOAD_ERROR: case errors["ErrorDetails"].KEY_LOAD_TIMEOUT: if (!data.fatal) { // keep retrying until the limit will be reached if (this.fragLoadError + 1 <= this.config.fragLoadingMaxRetry) { // exponential backoff capped to config.fragLoadingMaxRetryTimeout var delay = Math.min(Math.pow(2, this.fragLoadError) * this.config.fragLoadingRetryDelay, this.config.fragLoadingMaxRetryTimeout); logger["logger"].warn("mediaController: frag loading failed, retry in " + delay + " ms"); this.retryDate = window.performance.now() + delay; // retry loading state // if loadedmetadata is not set, it means that we are emergency switch down on first frag // in that case, reset startFragRequested flag if (!this.loadedmetadata) { this.startFragRequested = false; this.nextLoadPosition = this.startPosition; } this.fragLoadError++; this.state = State.FRAG_LOADING_WAITING_RETRY; } else { logger["logger"].error("mediaController: " + data.details + " reaches max retry, redispatch as fatal ..."); // switch error to fatal data.fatal = true; this.state = State.ERROR; } } break; case errors["ErrorDetails"].LEVEL_LOAD_ERROR: case errors["ErrorDetails"].LEVEL_LOAD_TIMEOUT: if (this.state !== State.ERROR) { if (data.fatal) { // if fatal error, stop processing this.state = State.ERROR; logger["logger"].warn("streamController: " + data.details + ",switch to " + this.state + " state ..."); } else { // in case of non fatal error while loading level, if level controller is not retrying to load level , switch back to IDLE if (!data.levelRetry && this.state === State.WAITING_LEVEL) { this.state = State.IDLE; } } } break; case errors["ErrorDetails"].BUFFER_FULL_ERROR: // if in appending state if (data.parent === 'main' && (this.state === State.PARSING || this.state === State.PARSED)) { // reduce max buf len if current position is buffered if (mediaBuffered) { this._reduceMaxBufferLength(this.config.maxBufferLength); this.state = State.IDLE; } else { // current position is not buffered, but browser is still complaining about buffer full error // this happens on IE/Edge, refer to https://github.com/video-dev/hls.js/pull/708 // in that case flush the whole buffer to recover logger["logger"].warn('buffer full error also media.currentTime is not buffered, flush everything'); this.fragCurrent = null; // flush everything this.flushMainBuffer(0, Number.POSITIVE_INFINITY); } } break; default: break; } }; _proto._reduceMaxBufferLength = function _reduceMaxBufferLength(minLength) { var config = this.config; if (config.maxMaxBufferLength >= minLength) { // reduce max buffer length as it might be too high. we do this to avoid loop flushing ... config.maxMaxBufferLength /= 2; logger["logger"].warn("main:reduce max buffer length to " + config.maxMaxBufferLength + "s"); return true; } return false; } /** * Checks the health of the buffer and attempts to resolve playback stalls. * @private */ ; _proto._checkBuffer = function _checkBuffer() { var media = this.media; if (!media || media.readyState === 0) { // Exit early if we don't have media or if the media hasn't bufferd anything yet (readyState 0) return; } var mediaBuffer = this.mediaBuffer ? this.mediaBuffer : media; var buffered = mediaBuffer.buffered; if (!this.loadedmetadata && buffered.length) { this.loadedmetadata = true; this._seekToStartPos(); } else if (this.immediateSwitch) { this.immediateLevelSwitchEnd(); } else { this.gapController.poll(this.lastCurrentTime, buffered); } }; _proto.onFragLoadEmergencyAborted = function onFragLoadEmergencyAborted() { this.state = State.IDLE; // if loadedmetadata is not set, it means that we are emergency switch down on first frag // in that case, reset startFragRequested flag if (!this.loadedmetadata) { this.startFragRequested = false; this.nextLoadPosition = this.startPosition; } this.tick(); }; _proto.onBufferFlushed = function onBufferFlushed() { /* after successful buffer flushing, filter flushed fragments from bufferedFrags use mediaBuffered instead of media (so that we will check against video.buffered ranges in case of alt audio track) */ var media = this.mediaBuffer ? this.mediaBuffer : this.media; if (media) { // filter fragments potentially evicted from buffer. this is to avoid memleak on live streams var elementaryStreamType = this.audioOnly ? ElementaryStreamTypes.AUDIO : ElementaryStreamTypes.VIDEO; this.fragmentTracker.detectEvictedFragments(elementaryStreamType, media.buffered); } // move to IDLE once flush complete. this should trigger new fragment loading this.state = State.IDLE; // reset reference to frag this.fragPrevious = null; }; _proto.onLevelsUpdated = function onLevelsUpdated(data) { this.levels = data.levels; }; _proto.swapAudioCodec = function swapAudioCodec() { this.audioCodecSwap = !this.audioCodecSwap; } /** * Seeks to the set startPosition if not equal to the mediaElement's current time. * @private */ ; _proto._seekToStartPos = function _seekToStartPos() { var media = this.media, startPosition = this.startPosition; var currentTime = media.currentTime; // only adjust currentTime if different from startPosition or if startPosition not buffered // at that stage, there should be only one buffered range, as we reach that code after first fragment has been buffered if (currentTime !== startPosition && startPosition >= 0) { if (media.seeking) { logger["logger"].log("could not seek to " + startPosition + ", already seeking at " + currentTime); return; } logger["logger"].log("seek to target start position " + startPosition + " from current time " + currentTime + ". ready state " + media.readyState); media.currentTime = startPosition; } }; _proto._getAudioCodec = function _getAudioCodec(currentLevel) { var audioCodec = this.config.defaultAudioCodec || currentLevel.audioCodec; if (this.audioCodecSwap) { logger["logger"].log('swapping playlist audio codec'); if (audioCodec) { if (audioCodec.indexOf('mp4a.40.5') !== -1) { audioCodec = 'mp4a.40.2'; } else { audioCodec = 'mp4a.40.5'; } } } return audioCodec; }; stream_controller_createClass(StreamController, [{ key: "state", set: function set(nextState) { if (this.state !== nextState) { var previousState = this.state; this._state = nextState; logger["logger"].log("main stream-controller: " + previousState + "->" + nextState); this.hls.trigger(events["default"].STREAM_STATE_TRANSITION, { previousState: previousState, nextState: nextState }); } }, get: function get() { return this._state; } }, { key: "currentLevel", get: function get() { var media = this.media; if (media) { var frag = this.getBufferedFrag(media.currentTime); if (frag) { return frag.level; } } return -1; } }, { key: "nextBufferedFrag", get: function get() { var media = this.media; if (media) { // first get end range of current fragment return this.followingBufferedFrag(this.getBufferedFrag(media.currentTime)); } else { return null; } } }, { key: "nextLevel", get: function get() { var frag = this.nextBufferedFrag; if (frag) { return frag.level; } else { return -1; } } }, { key: "liveSyncPosition", get: function get() { return this._liveSyncPosition; }, set: function set(value) { this._liveSyncPosition = value; } }]); return StreamController; }(base_stream_controller_BaseStreamController); /* harmony default export */ var stream_controller = (stream_controller_StreamController); // CONCATENATED MODULE: ./src/controller/level-controller.js function level_controller_defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } function level_controller_createClass(Constructor, protoProps, staticProps) { if (protoProps) level_controller_defineProperties(Constructor.prototype, protoProps); if (staticProps) level_controller_defineProperties(Constructor, staticProps); return Constructor; } function level_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /* * Level Controller */ var chromeOrFirefox; var level_controller_LevelController = /*#__PURE__*/function (_EventHandler) { level_controller_inheritsLoose(LevelController, _EventHandler); function LevelController(hls) { var _this; _this = _EventHandler.call(this, hls, events["default"].MANIFEST_LOADED, events["default"].LEVEL_LOADED, events["default"].AUDIO_TRACK_SWITCHED, events["default"].FRAG_LOADED, events["default"].ERROR) || this; _this.canload = false; _this.currentLevelIndex = null; _this.manualLevelIndex = -1; _this.timer = null; chromeOrFirefox = /chrome|firefox/.test(navigator.userAgent.toLowerCase()); return _this; } var _proto = LevelController.prototype; _proto.onHandlerDestroying = function onHandlerDestroying() { this.clearTimer(); this.manualLevelIndex = -1; }; _proto.clearTimer = function clearTimer() { if (this.timer !== null) { clearTimeout(this.timer); this.timer = null; } }; _proto.startLoad = function startLoad() { var levels = this._levels; this.canload = true; this.levelRetryCount = 0; // clean up live level details to force reload them, and reset load errors if (levels) { levels.forEach(function (level) { level.loadError = 0; var levelDetails = level.details; if (levelDetails && levelDetails.live) { level.details = undefined; } }); } // speed up live playlist refresh if timer exists if (this.timer !== null) { this.loadLevel(); } }; _proto.stopLoad = function stopLoad() { this.canload = false; }; _proto.onManifestLoaded = function onManifestLoaded(data) { var levels = []; var audioTracks = []; var bitrateStart; var levelSet = {}; var levelFromSet = null; var videoCodecFound = false; var audioCodecFound = false; // regroup redundant levels together data.levels.forEach(function (level) { var attributes = level.attrs; level.loadError = 0; level.fragmentError = false; videoCodecFound = videoCodecFound || !!level.videoCodec; audioCodecFound = audioCodecFound || !!level.audioCodec; // erase audio codec info if browser does not support mp4a.40.34. // demuxer will autodetect codec and fallback to mpeg/audio if (chromeOrFirefox && level.audioCodec && level.audioCodec.indexOf('mp4a.40.34') !== -1) { level.audioCodec = undefined; } levelFromSet = levelSet[level.bitrate]; // FIXME: we would also have to match the resolution here if (!levelFromSet) { level.url = [level.url]; level.urlId = 0; levelSet[level.bitrate] = level; levels.push(level); } else { levelFromSet.url.push(level.url); } if (attributes) { if (attributes.AUDIO) { addGroupId(levelFromSet || level, 'audio', attributes.AUDIO); } if (attributes.SUBTITLES) { addGroupId(levelFromSet || level, 'text', attributes.SUBTITLES); } } }); // remove audio-only level if we also have levels with audio+video codecs signalled if (videoCodecFound && audioCodecFound) { levels = levels.filter(function (_ref) { var videoCodec = _ref.videoCodec; return !!videoCodec; }); } // only keep levels with supported audio/video codecs levels = levels.filter(function (_ref2) { var audioCodec = _ref2.audioCodec, videoCodec = _ref2.videoCodec; return (!audioCodec || isCodecSupportedInMp4(audioCodec, 'audio')) && (!videoCodec || isCodecSupportedInMp4(videoCodec, 'video')); }); if (data.audioTracks) { audioTracks = data.audioTracks.filter(function (track) { return !track.audioCodec || isCodecSupportedInMp4(track.audioCodec, 'audio'); }); // Reassign id's after filtering since they're used as array indices audioTracks.forEach(function (track, index) { track.id = index; }); } if (levels.length > 0) { // start bitrate is the first bitrate of the manifest bitrateStart = levels[0].bitrate; // sort level on bitrate levels.sort(function (a, b) { return a.bitrate - b.bitrate; }); this._levels = levels; // find index of first level in sorted levels for (var i = 0; i < levels.length; i++) { if (levels[i].bitrate === bitrateStart) { this._firstLevel = i; logger["logger"].log("manifest loaded," + levels.length + " level(s) found, first bitrate:" + bitrateStart); break; } } // Audio is only alternate if manifest include a URI along with the audio group tag, // and this is not an audio-only stream where levels contain audio-only var audioOnly = audioCodecFound && !videoCodecFound; this.hls.trigger(events["default"].MANIFEST_PARSED, { levels: levels, audioTracks: audioTracks, firstLevel: this._firstLevel, stats: data.stats, audio: audioCodecFound, video: videoCodecFound, altAudio: !audioOnly && audioTracks.some(function (t) { return !!t.url; }) }); } else { this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].MANIFEST_INCOMPATIBLE_CODECS_ERROR, fatal: true, url: this.hls.url, reason: 'no level with compatible codecs found in manifest' }); } }; _proto.setLevelInternal = function setLevelInternal(newLevel) { var levels = this._levels; var hls = this.hls; // check if level idx is valid if (newLevel >= 0 && newLevel < levels.length) { // stopping live reloading timer if any this.clearTimer(); if (this.currentLevelIndex !== newLevel) { logger["logger"].log("switching to level " + newLevel); this.currentLevelIndex = newLevel; var levelProperties = levels[newLevel]; levelProperties.level = newLevel; hls.trigger(events["default"].LEVEL_SWITCHING, levelProperties); } var level = levels[newLevel]; var levelDetails = level.details; // check if we need to load playlist for this level if (!levelDetails || levelDetails.live) { // level not retrieved yet, or live playlist we need to (re)load it var urlId = level.urlId; hls.trigger(events["default"].LEVEL_LOADING, { url: level.url[urlId], level: newLevel, id: urlId }); } } else { // invalid level id given, trigger error hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].OTHER_ERROR, details: errors["ErrorDetails"].LEVEL_SWITCH_ERROR, level: newLevel, fatal: false, reason: 'invalid level idx' }); } }; _proto.onError = function onError(data) { if (data.fatal) { if (data.type === errors["ErrorTypes"].NETWORK_ERROR) { this.clearTimer(); } return; } var levelError = false, fragmentError = false; var levelIndex; // try to recover not fatal errors switch (data.details) { case errors["ErrorDetails"].FRAG_LOAD_ERROR: case errors["ErrorDetails"].FRAG_LOAD_TIMEOUT: case errors["ErrorDetails"].KEY_LOAD_ERROR: case errors["ErrorDetails"].KEY_LOAD_TIMEOUT: levelIndex = data.frag.level; fragmentError = true; break; case errors["ErrorDetails"].LEVEL_LOAD_ERROR: case errors["ErrorDetails"].LEVEL_LOAD_TIMEOUT: levelIndex = data.context.level; levelError = true; break; case errors["ErrorDetails"].REMUX_ALLOC_ERROR: levelIndex = data.level; levelError = true; break; } if (levelIndex !== undefined) { this.recoverLevel(data, levelIndex, levelError, fragmentError); } } /** * Switch to a redundant stream if any available. * If redundant stream is not available, emergency switch down if ABR mode is enabled. * * @param {Object} errorEvent * @param {Number} levelIndex current level index * @param {Boolean} levelError * @param {Boolean} fragmentError */ // FIXME Find a better abstraction where fragment/level retry management is well decoupled ; _proto.recoverLevel = function recoverLevel(errorEvent, levelIndex, levelError, fragmentError) { var _this2 = this; var config = this.hls.config; var errorDetails = errorEvent.details; var level = this._levels[levelIndex]; var redundantLevels, delay, nextLevel; level.loadError++; level.fragmentError = fragmentError; if (levelError) { if (this.levelRetryCount + 1 <= config.levelLoadingMaxRetry) { // exponential backoff capped to max retry timeout delay = Math.min(Math.pow(2, this.levelRetryCount) * config.levelLoadingRetryDelay, config.levelLoadingMaxRetryTimeout); // Schedule level reload this.timer = setTimeout(function () { return _this2.loadLevel(); }, delay); // boolean used to inform stream controller not to switch back to IDLE on non fatal error errorEvent.levelRetry = true; this.levelRetryCount++; logger["logger"].warn("level controller, " + errorDetails + ", retry in " + delay + " ms, current retry count is " + this.levelRetryCount); } else { logger["logger"].error("level controller, cannot recover from " + errorDetails + " error"); this.currentLevelIndex = null; // stopping live reloading timer if any this.clearTimer(); // switch error to fatal errorEvent.fatal = true; return; } } // Try any redundant streams if available for both errors: level and fragment // If level.loadError reaches redundantLevels it means that we tried them all, no hope => let's switch down if (levelError || fragmentError) { redundantLevels = level.url.length; if (redundantLevels > 1 && level.loadError < redundantLevels) { level.urlId = (level.urlId + 1) % redundantLevels; level.details = undefined; logger["logger"].warn("level controller, " + errorDetails + " for level " + levelIndex + ": switching to redundant URL-id " + level.urlId); // console.log('Current audio track group ID:', this.hls.audioTracks[this.hls.audioTrack].groupId); // console.log('New video quality level audio group id:', level.attrs.AUDIO); } else { // Search for available level if (this.manualLevelIndex === -1) { // When lowest level has been reached, let's start hunt from the top nextLevel = levelIndex === 0 ? this._levels.length - 1 : levelIndex - 1; logger["logger"].warn("level controller, " + errorDetails + ": switch to " + nextLevel); this.hls.nextAutoLevel = this.currentLevelIndex = nextLevel; } else if (fragmentError) { // Allow fragment retry as long as configuration allows. // reset this._level so that another call to set level() will trigger again a frag load logger["logger"].warn("level controller, " + errorDetails + ": reload a fragment"); this.currentLevelIndex = null; } } } } // reset errors on the successful load of a fragment ; _proto.onFragLoaded = function onFragLoaded(_ref3) { var frag = _ref3.frag; if (frag !== undefined && frag.type === 'main') { var level = this._levels[frag.level]; if (level !== undefined) { level.fragmentError = false; level.loadError = 0; this.levelRetryCount = 0; } } }; _proto.onLevelLoaded = function onLevelLoaded(data) { var _this3 = this; var level = data.level, details = data.details; // only process level loaded events matching with expected level if (level !== this.currentLevelIndex) { return; } var curLevel = this._levels[level]; // reset level load error counter on successful level loaded only if there is no issues with fragments if (!curLevel.fragmentError) { curLevel.loadError = 0; this.levelRetryCount = 0; } // if current playlist is a live playlist, arm a timer to reload it if (details.live) { var reloadInterval = computeReloadInterval(curLevel.details, details, data.stats.trequest); logger["logger"].log("live playlist, reload in " + Math.round(reloadInterval) + " ms"); this.timer = setTimeout(function () { return _this3.loadLevel(); }, reloadInterval); } else { this.clearTimer(); } }; _proto.onAudioTrackSwitched = function onAudioTrackSwitched(data) { var audioGroupId = this.hls.audioTracks[data.id].groupId; var currentLevel = this.hls.levels[this.currentLevelIndex]; if (!currentLevel) { return; } if (currentLevel.audioGroupIds) { var urlId = -1; for (var i = 0; i < currentLevel.audioGroupIds.length; i++) { if (currentLevel.audioGroupIds[i] === audioGroupId) { urlId = i; break; } } if (urlId !== currentLevel.urlId) { currentLevel.urlId = urlId; this.startLoad(); } } }; _proto.loadLevel = function loadLevel() { logger["logger"].debug('call to loadLevel'); if (this.currentLevelIndex !== null && this.canload) { var levelObject = this._levels[this.currentLevelIndex]; if (typeof levelObject === 'object' && levelObject.url.length > 0) { var level = this.currentLevelIndex; var id = levelObject.urlId; var url = levelObject.url[id]; logger["logger"].log("Attempt loading level index " + level + " with URL-id " + id); // console.log('Current audio track group ID:', this.hls.audioTracks[this.hls.audioTrack].groupId); // console.log('New video quality level audio group id:', levelObject.attrs.AUDIO, level); this.hls.trigger(events["default"].LEVEL_LOADING, { url: url, level: level, id: id }); } } }; _proto.removeLevel = function removeLevel(levelIndex, urlId) { var levels = this.levels.filter(function (level, index) { if (index !== levelIndex) { return true; } if (level.url.length > 1 && urlId !== undefined) { level.url = level.url.filter(function (url, id) { return id !== urlId; }); level.urlId = 0; return true; } return false; }).map(function (level, index) { var details = level.details; if (details && details.fragments) { details.fragments.forEach(function (fragment) { fragment.level = index; }); } return level; }); this._levels = levels; this.hls.trigger(events["default"].LEVELS_UPDATED, { levels: levels }); }; level_controller_createClass(LevelController, [{ key: "levels", get: function get() { return this._levels; } }, { key: "level", get: function get() { return this.currentLevelIndex; }, set: function set(newLevel) { var levels = this._levels; if (levels) { newLevel = Math.min(newLevel, levels.length - 1); if (this.currentLevelIndex !== newLevel || !levels[newLevel].details) { this.setLevelInternal(newLevel); } } } }, { key: "manualLevel", get: function get() { return this.manualLevelIndex; }, set: function set(newLevel) { this.manualLevelIndex = newLevel; if (this._startLevel === undefined) { this._startLevel = newLevel; } if (newLevel !== -1) { this.level = newLevel; } } }, { key: "firstLevel", get: function get() { return this._firstLevel; }, set: function set(newLevel) { this._firstLevel = newLevel; } }, { key: "startLevel", get: function get() { // hls.startLevel takes precedence over config.startLevel // if none of these values are defined, fallback on this._firstLevel (first quality level appearing in variant manifest) if (this._startLevel === undefined) { var configStartLevel = this.hls.config.startLevel; if (configStartLevel !== undefined) { return configStartLevel; } else { return this._firstLevel; } } else { return this._startLevel; } }, set: function set(newLevel) { this._startLevel = newLevel; } }, { key: "nextLoadLevel", get: function get() { if (this.manualLevelIndex !== -1) { return this.manualLevelIndex; } else { return this.hls.nextAutoLevel; } }, set: function set(nextLevel) { this.level = nextLevel; if (this.manualLevelIndex === -1) { this.hls.nextAutoLevel = nextLevel; } } }]); return LevelController; }(event_handler); // EXTERNAL MODULE: ./src/demux/id3.js var id3 = __webpack_require__("./src/demux/id3.js"); // CONCATENATED MODULE: ./src/utils/texttrack-utils.ts function sendAddTrackEvent(track, videoEl) { var event; try { event = new Event('addtrack'); } catch (err) { // for IE11 event = document.createEvent('Event'); event.initEvent('addtrack', false, false); } event.track = track; videoEl.dispatchEvent(event); } function clearCurrentCues(track) { if (track === null || track === void 0 ? void 0 : track.cues) { while (track.cues.length > 0) { track.removeCue(track.cues[0]); } } } /** * Given a list of Cues, finds the closest cue matching the given time. * Modified verison of binary search O(log(n)). * * @export * @param {(TextTrackCueList | TextTrackCue[])} cues - List of cues. * @param {number} time - Target time, to find closest cue to. * @returns {TextTrackCue} */ function getClosestCue(cues, time) { // If the offset is less than the first element, the first element is the closest. if (time < cues[0].endTime) { return cues[0]; } // If the offset is greater than the last cue, the last is the closest. if (time > cues[cues.length - 1].endTime) { return cues[cues.length - 1]; } var left = 0; var right = cues.length - 1; while (left <= right) { var mid = Math.floor((right + left) / 2); if (time < cues[mid].endTime) { right = mid - 1; } else if (time > cues[mid].endTime) { left = mid + 1; } else { // If it's not lower or higher, it must be equal. return cues[mid]; } } // At this point, left and right have swapped. // No direct match was found, left or right element must be the closest. Check which one has the smallest diff. return cues[left].endTime - time < time - cues[right].endTime ? cues[left] : cues[right]; } // CONCATENATED MODULE: ./src/controller/id3-track-controller.js function id3_track_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /* * id3 metadata track controller */ var id3_track_controller_ID3TrackController = /*#__PURE__*/function (_EventHandler) { id3_track_controller_inheritsLoose(ID3TrackController, _EventHandler); function ID3TrackController(hls) { var _this; _this = _EventHandler.call(this, hls, events["default"].MEDIA_ATTACHED, events["default"].MEDIA_DETACHING, events["default"].FRAG_PARSING_METADATA, events["default"].LIVE_BACK_BUFFER_REACHED) || this; _this.id3Track = undefined; _this.media = undefined; return _this; } var _proto = ID3TrackController.prototype; _proto.destroy = function destroy() { event_handler.prototype.destroy.call(this); } // Add ID3 metatadata text track. ; _proto.onMediaAttached = function onMediaAttached(data) { this.media = data.media; if (!this.media) {} }; _proto.onMediaDetaching = function onMediaDetaching() {<|fim▁hole|> }; _proto.getID3Track = function getID3Track(textTracks) { for (var i = 0; i < textTracks.length; i++) { var textTrack = textTracks[i]; if (textTrack.kind === 'metadata' && textTrack.label === 'id3') { // send 'addtrack' when reusing the textTrack for metadata, // same as what we do for captions sendAddTrackEvent(textTrack, this.media); return textTrack; } } return this.media.addTextTrack('metadata', 'id3'); }; _proto.onFragParsingMetadata = function onFragParsingMetadata(data) { var fragment = data.frag; var samples = data.samples; // create track dynamically if (!this.id3Track) { this.id3Track = this.getID3Track(this.media.textTracks); this.id3Track.mode = 'hidden'; } // Attempt to recreate Safari functionality by creating // WebKitDataCue objects when available and store the decoded // ID3 data in the value property of the cue var Cue = window.WebKitDataCue || window.VTTCue || window.TextTrackCue; for (var i = 0; i < samples.length; i++) { var frames = id3["default"].getID3Frames(samples[i].data); if (frames) { // Ensure the pts is positive - sometimes it's reported as a small negative number var startTime = Math.max(samples[i].pts, 0); var endTime = i < samples.length - 1 ? samples[i + 1].pts : fragment.endPTS; if (!endTime) { endTime = fragment.start + fragment.duration; } if (startTime === endTime) { // Give a slight bump to the endTime if it's equal to startTime to avoid a SyntaxError in IE endTime += 0.0001; } else if (startTime > endTime) { logger["logger"].warn('detected an id3 sample with endTime < startTime, adjusting endTime to (startTime + 0.25)'); endTime = startTime + 0.25; } for (var j = 0; j < frames.length; j++) { var frame = frames[j]; // Safari doesn't put the timestamp frame in the TextTrack if (!id3["default"].isTimeStampFrame(frame)) { var cue = new Cue(startTime, endTime, ''); cue.value = frame; this.id3Track.addCue(cue); } } } } }; _proto.onLiveBackBufferReached = function onLiveBackBufferReached(_ref) { var bufferEnd = _ref.bufferEnd; var id3Track = this.id3Track; if (!id3Track || !id3Track.cues || !id3Track.cues.length) { return; } var foundCue = getClosestCue(id3Track.cues, bufferEnd); if (!foundCue) { return; } while (id3Track.cues[0] !== foundCue) { id3Track.removeCue(id3Track.cues[0]); } }; return ID3TrackController; }(event_handler); /* harmony default export */ var id3_track_controller = (id3_track_controller_ID3TrackController); // CONCATENATED MODULE: ./src/is-supported.ts function is_supported_isSupported() { var mediaSource = getMediaSource(); if (!mediaSource) { return false; } var sourceBuffer = self.SourceBuffer || self.WebKitSourceBuffer; var isTypeSupported = mediaSource && typeof mediaSource.isTypeSupported === 'function' && mediaSource.isTypeSupported('video/mp4; codecs="avc1.42E01E,mp4a.40.2"'); // if SourceBuffer is exposed ensure its API is valid // safari and old version of Chrome doe not expose SourceBuffer globally so checking SourceBuffer.prototype is impossible var sourceBufferValidAPI = !sourceBuffer || sourceBuffer.prototype && typeof sourceBuffer.prototype.appendBuffer === 'function' && typeof sourceBuffer.prototype.remove === 'function'; return !!isTypeSupported && !!sourceBufferValidAPI; } // CONCATENATED MODULE: ./src/utils/ewma.ts /* * compute an Exponential Weighted moving average * - https://en.wikipedia.org/wiki/Moving_average#Exponential_moving_average * - heavily inspired from shaka-player */ var EWMA = /*#__PURE__*/function () { // About half of the estimated value will be from the last |halfLife| samples by weight. function EWMA(halfLife) { this.alpha_ = void 0; this.estimate_ = void 0; this.totalWeight_ = void 0; // Larger values of alpha expire historical data more slowly. this.alpha_ = halfLife ? Math.exp(Math.log(0.5) / halfLife) : 0; this.estimate_ = 0; this.totalWeight_ = 0; } var _proto = EWMA.prototype; _proto.sample = function sample(weight, value) { var adjAlpha = Math.pow(this.alpha_, weight); this.estimate_ = value * (1 - adjAlpha) + adjAlpha * this.estimate_; this.totalWeight_ += weight; }; _proto.getTotalWeight = function getTotalWeight() { return this.totalWeight_; }; _proto.getEstimate = function getEstimate() { if (this.alpha_) { var zeroFactor = 1 - Math.pow(this.alpha_, this.totalWeight_); return this.estimate_ / zeroFactor; } else { return this.estimate_; } }; return EWMA; }(); /* harmony default export */ var ewma = (EWMA); // CONCATENATED MODULE: ./src/utils/ewma-bandwidth-estimator.ts /* * EWMA Bandwidth Estimator * - heavily inspired from shaka-player * Tracks bandwidth samples and estimates available bandwidth. * Based on the minimum of two exponentially-weighted moving averages with * different half-lives. */ var ewma_bandwidth_estimator_EwmaBandWidthEstimator = /*#__PURE__*/function () { // TODO(typescript-hls) function EwmaBandWidthEstimator(hls, slow, fast, defaultEstimate) { this.hls = void 0; this.defaultEstimate_ = void 0; this.minWeight_ = void 0; this.minDelayMs_ = void 0; this.slow_ = void 0; this.fast_ = void 0; this.hls = hls; this.defaultEstimate_ = defaultEstimate; this.minWeight_ = 0.001; this.minDelayMs_ = 50; this.slow_ = new ewma(slow); this.fast_ = new ewma(fast); } var _proto = EwmaBandWidthEstimator.prototype; _proto.sample = function sample(durationMs, numBytes) { durationMs = Math.max(durationMs, this.minDelayMs_); var numBits = 8 * numBytes, // weight is duration in seconds durationS = durationMs / 1000, // value is bandwidth in bits/s bandwidthInBps = numBits / durationS; this.fast_.sample(durationS, bandwidthInBps); this.slow_.sample(durationS, bandwidthInBps); }; _proto.canEstimate = function canEstimate() { var fast = this.fast_; return fast && fast.getTotalWeight() >= this.minWeight_; }; _proto.getEstimate = function getEstimate() { if (this.canEstimate()) { // console.log('slow estimate:'+ Math.round(this.slow_.getEstimate())); // console.log('fast estimate:'+ Math.round(this.fast_.getEstimate())); // Take the minimum of these two estimates. This should have the effect of // adapting down quickly, but up more slowly. return Math.min(this.fast_.getEstimate(), this.slow_.getEstimate()); } else { return this.defaultEstimate_; } }; _proto.destroy = function destroy() {}; return EwmaBandWidthEstimator; }(); /* harmony default export */ var ewma_bandwidth_estimator = (ewma_bandwidth_estimator_EwmaBandWidthEstimator); // CONCATENATED MODULE: ./src/controller/abr-controller.js function abr_controller_defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } function abr_controller_createClass(Constructor, protoProps, staticProps) { if (protoProps) abr_controller_defineProperties(Constructor.prototype, protoProps); if (staticProps) abr_controller_defineProperties(Constructor, staticProps); return Constructor; } function abr_controller_assertThisInitialized(self) { if (self === void 0) { throw new ReferenceError("this hasn't been initialised - super() hasn't been called"); } return self; } function abr_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /* * simple ABR Controller * - compute next level based on last fragment bw heuristics * - implement an abandon rules triggered if we have less than 2 frag buffered and if computed bw shows that we risk buffer stalling */ var abr_controller_window = window, abr_controller_performance = abr_controller_window.performance; var abr_controller_AbrController = /*#__PURE__*/function (_EventHandler) { abr_controller_inheritsLoose(AbrController, _EventHandler); function AbrController(hls) { var _this; _this = _EventHandler.call(this, hls, events["default"].FRAG_LOADING, events["default"].FRAG_LOADED, events["default"].FRAG_BUFFERED, events["default"].ERROR) || this; _this.lastLoadedFragLevel = 0; _this._nextAutoLevel = -1; _this.hls = hls; _this.timer = null; _this._bwEstimator = null; _this.onCheck = _this._abandonRulesCheck.bind(abr_controller_assertThisInitialized(_this)); return _this; } var _proto = AbrController.prototype; _proto.destroy = function destroy() { this.clearTimer(); event_handler.prototype.destroy.call(this); }; _proto.onFragLoading = function onFragLoading(data) { var frag = data.frag; if (frag.type === 'main') { if (!this.timer) { this.fragCurrent = frag; this.timer = setInterval(this.onCheck, 100); } // lazy init of BwEstimator, rationale is that we use different params for Live/VoD // so we need to wait for stream manifest / playlist type to instantiate it. if (!this._bwEstimator) { var hls = this.hls; var config = hls.config; var level = frag.level; var isLive = hls.levels[level].details.live; var ewmaFast; var ewmaSlow; if (isLive) { ewmaFast = config.abrEwmaFastLive; ewmaSlow = config.abrEwmaSlowLive; } else { ewmaFast = config.abrEwmaFastVoD; ewmaSlow = config.abrEwmaSlowVoD; } this._bwEstimator = new ewma_bandwidth_estimator(hls, ewmaSlow, ewmaFast, config.abrEwmaDefaultEstimate); } } }; _proto._abandonRulesCheck = function _abandonRulesCheck() { /* monitor fragment retrieval time... we compute expected time of arrival of the complete fragment. we compare it to expected time of buffer starvation */ var hls = this.hls; var video = hls.media; var frag = this.fragCurrent; if (!frag) { return; } var loader = frag.loader; // if loader has been destroyed or loading has been aborted, stop timer and return if (!loader || loader.stats && loader.stats.aborted) { logger["logger"].warn('frag loader destroy or aborted, disarm abandonRules'); this.clearTimer(); // reset forced auto level value so that next level will be selected this._nextAutoLevel = -1; return; } var stats = loader.stats; /* only monitor frag retrieval time if (video not paused OR first fragment being loaded(ready state === HAVE_NOTHING = 0)) AND autoswitching enabled AND not lowest level (=> means that we have several levels) */ if (video && stats && (!video.paused && video.playbackRate !== 0 || !video.readyState) && frag.autoLevel && frag.level) { var requestDelay = abr_controller_performance.now() - stats.trequest; var playbackRate = Math.abs(video.playbackRate); // monitor fragment load progress after half of expected fragment duration,to stabilize bitrate if (requestDelay > 500 * frag.duration / playbackRate) { var levels = hls.levels; var loadRate = Math.max(1, stats.bw ? stats.bw / 8 : stats.loaded * 1000 / requestDelay); // byte/s; at least 1 byte/s to avoid division by zero // compute expected fragment length using frag duration and level bitrate. also ensure that expected len is gte than already loaded size var level = levels[frag.level]; if (!level) { return; } var levelBitrate = level.realBitrate ? Math.max(level.realBitrate, level.bitrate) : level.bitrate; var expectedLen = stats.total ? stats.total : Math.max(stats.loaded, Math.round(frag.duration * levelBitrate / 8)); var pos = video.currentTime; var fragLoadedDelay = (expectedLen - stats.loaded) / loadRate; var bufferStarvationDelay = (BufferHelper.bufferInfo(video, pos, hls.config.maxBufferHole).end - pos) / playbackRate; // consider emergency switch down only if we have less than 2 frag buffered AND // time to finish loading current fragment is bigger than buffer starvation delay // ie if we risk buffer starvation if bw does not increase quickly if (bufferStarvationDelay < 2 * frag.duration / playbackRate && fragLoadedDelay > bufferStarvationDelay) { var minAutoLevel = hls.minAutoLevel; var fragLevelNextLoadedDelay; var nextLoadLevel; // lets iterate through lower level and try to find the biggest one that could avoid rebuffering // we start from current level - 1 and we step down , until we find a matching level for (nextLoadLevel = frag.level - 1; nextLoadLevel > minAutoLevel; nextLoadLevel--) { // compute time to load next fragment at lower level // 0.8 : consider only 80% of current bw to be conservative // 8 = bits per byte (bps/Bps) var levelNextBitrate = levels[nextLoadLevel].realBitrate ? Math.max(levels[nextLoadLevel].realBitrate, levels[nextLoadLevel].bitrate) : levels[nextLoadLevel].bitrate; var _fragLevelNextLoadedDelay = frag.duration * levelNextBitrate / (8 * 0.8 * loadRate); if (_fragLevelNextLoadedDelay < bufferStarvationDelay) { // we found a lower level that be rebuffering free with current estimated bw ! break; } } // only emergency switch down if it takes less time to load new fragment at lowest level instead // of finishing loading current one ... if (fragLevelNextLoadedDelay < fragLoadedDelay) { logger["logger"].warn("loading too slow, abort fragment loading and switch to level " + nextLoadLevel + ":fragLoadedDelay[" + nextLoadLevel + "]<fragLoadedDelay[" + (frag.level - 1) + "];bufferStarvationDelay:" + fragLevelNextLoadedDelay.toFixed(1) + "<" + fragLoadedDelay.toFixed(1) + ":" + bufferStarvationDelay.toFixed(1)); // force next load level in auto mode hls.nextLoadLevel = nextLoadLevel; // update bw estimate for this fragment before cancelling load (this will help reducing the bw) this._bwEstimator.sample(requestDelay, stats.loaded); // abort fragment loading loader.abort(); // stop abandon rules timer this.clearTimer(); hls.trigger(events["default"].FRAG_LOAD_EMERGENCY_ABORTED, { frag: frag, stats: stats }); } } } } }; _proto.onFragLoaded = function onFragLoaded(data) { var frag = data.frag; if (frag.type === 'main' && Object(number["isFiniteNumber"])(frag.sn)) { // stop monitoring bw once frag loaded this.clearTimer(); // store level id after successful fragment load this.lastLoadedFragLevel = frag.level; // reset forced auto level value so that next level will be selected this._nextAutoLevel = -1; // compute level average bitrate if (this.hls.config.abrMaxWithRealBitrate) { var level = this.hls.levels[frag.level]; var loadedBytes = (level.loaded ? level.loaded.bytes : 0) + data.stats.loaded; var loadedDuration = (level.loaded ? level.loaded.duration : 0) + data.frag.duration; level.loaded = { bytes: loadedBytes, duration: loadedDuration }; level.realBitrate = Math.round(8 * loadedBytes / loadedDuration); } // if fragment has been loaded to perform a bitrate test, if (data.frag.bitrateTest) { var stats = data.stats; stats.tparsed = stats.tbuffered = stats.tload; this.onFragBuffered(data); } } }; _proto.onFragBuffered = function onFragBuffered(data) { var stats = data.stats; var frag = data.frag; // only update stats on first frag buffering // if same frag is loaded multiple times, it might be in browser cache, and loaded quickly // and leading to wrong bw estimation // on bitrate test, also only update stats once (if tload = tbuffered == on FRAG_LOADED) if (stats.aborted !== true && frag.type === 'main' && Object(number["isFiniteNumber"])(frag.sn) && (!frag.bitrateTest || stats.tload === stats.tbuffered)) { // use tparsed-trequest instead of tbuffered-trequest to compute fragLoadingProcessing; rationale is that buffer appending only happens once media is attached // in case we use config.startFragPrefetch while media is not attached yet, fragment might be parsed while media not attached yet, but it will only be buffered on media attached // as a consequence it could happen really late in the process. meaning that appending duration might appears huge ... leading to underestimated throughput estimation var fragLoadingProcessingMs = stats.tparsed - stats.trequest; logger["logger"].log("latency/loading/parsing/append/kbps:" + Math.round(stats.tfirst - stats.trequest) + "/" + Math.round(stats.tload - stats.tfirst) + "/" + Math.round(stats.tparsed - stats.tload) + "/" + Math.round(stats.tbuffered - stats.tparsed) + "/" + Math.round(8 * stats.loaded / (stats.tbuffered - stats.trequest))); this._bwEstimator.sample(fragLoadingProcessingMs, stats.loaded); stats.bwEstimate = this._bwEstimator.getEstimate(); // if fragment has been loaded to perform a bitrate test, (hls.startLevel = -1), store bitrate test delay duration if (frag.bitrateTest) { this.bitrateTestDelay = fragLoadingProcessingMs / 1000; } else { this.bitrateTestDelay = 0; } } }; _proto.onError = function onError(data) { // stop timer in case of frag loading error switch (data.details) { case errors["ErrorDetails"].FRAG_LOAD_ERROR: case errors["ErrorDetails"].FRAG_LOAD_TIMEOUT: this.clearTimer(); break; default: break; } }; _proto.clearTimer = function clearTimer() { clearInterval(this.timer); this.timer = null; } // return next auto level ; _proto._findBestLevel = function _findBestLevel(currentLevel, currentFragDuration, currentBw, minAutoLevel, maxAutoLevel, maxFetchDuration, bwFactor, bwUpFactor, levels) { for (var i = maxAutoLevel; i >= minAutoLevel; i--) { var levelInfo = levels[i]; if (!levelInfo) { continue; } var levelDetails = levelInfo.details; var avgDuration = levelDetails ? levelDetails.totalduration / levelDetails.fragments.length : currentFragDuration; var live = levelDetails ? levelDetails.live : false; var adjustedbw = void 0; // follow algorithm captured from stagefright : // https://android.googlesource.com/platform/frameworks/av/+/master/media/libstagefright/httplive/LiveSession.cpp // Pick the highest bandwidth stream below or equal to estimated bandwidth. // consider only 80% of the available bandwidth, but if we are switching up, // be even more conservative (70%) to avoid overestimating and immediately // switching back. if (i <= currentLevel) { adjustedbw = bwFactor * currentBw; } else { adjustedbw = bwUpFactor * currentBw; } var bitrate = levels[i].realBitrate ? Math.max(levels[i].realBitrate, levels[i].bitrate) : levels[i].bitrate; var fetchDuration = bitrate * avgDuration / adjustedbw; logger["logger"].trace("level/adjustedbw/bitrate/avgDuration/maxFetchDuration/fetchDuration: " + i + "/" + Math.round(adjustedbw) + "/" + bitrate + "/" + avgDuration + "/" + maxFetchDuration + "/" + fetchDuration); // if adjusted bw is greater than level bitrate AND if (adjustedbw > bitrate && ( // fragment fetchDuration unknown OR live stream OR fragment fetchDuration less than max allowed fetch duration, then this level matches // we don't account for max Fetch Duration for live streams, this is to avoid switching down when near the edge of live sliding window ... // special case to support startLevel = -1 (bitrateTest) on live streams : in that case we should not exit loop so that _findBestLevel will return -1 !fetchDuration || live && !this.bitrateTestDelay || fetchDuration < maxFetchDuration)) { // as we are looping from highest to lowest, this will return the best achievable quality level return i; } } // not enough time budget even with quality level 0 ... rebuffering might happen return -1; }; abr_controller_createClass(AbrController, [{ key: "nextAutoLevel", get: function get() { var forcedAutoLevel = this._nextAutoLevel; var bwEstimator = this._bwEstimator; // in case next auto level has been forced, and bw not available or not reliable, return forced value if (forcedAutoLevel !== -1 && (!bwEstimator || !bwEstimator.canEstimate())) { return forcedAutoLevel; } // compute next level using ABR logic var nextABRAutoLevel = this._nextABRAutoLevel; // if forced auto level has been defined, use it to cap ABR computed quality level if (forcedAutoLevel !== -1) { nextABRAutoLevel = Math.min(forcedAutoLevel, nextABRAutoLevel); } return nextABRAutoLevel; }, set: function set(nextLevel) { this._nextAutoLevel = nextLevel; } }, { key: "_nextABRAutoLevel", get: function get() { var hls = this.hls; var maxAutoLevel = hls.maxAutoLevel, levels = hls.levels, config = hls.config, minAutoLevel = hls.minAutoLevel; var video = hls.media; var currentLevel = this.lastLoadedFragLevel; var currentFragDuration = this.fragCurrent ? this.fragCurrent.duration : 0; var pos = video ? video.currentTime : 0; // playbackRate is the absolute value of the playback rate; if video.playbackRate is 0, we use 1 to load as // if we're playing back at the normal rate. var playbackRate = video && video.playbackRate !== 0 ? Math.abs(video.playbackRate) : 1.0; var avgbw = this._bwEstimator ? this._bwEstimator.getEstimate() : config.abrEwmaDefaultEstimate; // bufferStarvationDelay is the wall-clock time left until the playback buffer is exhausted. var bufferStarvationDelay = (BufferHelper.bufferInfo(video, pos, config.maxBufferHole).end - pos) / playbackRate; // First, look to see if we can find a level matching with our avg bandwidth AND that could also guarantee no rebuffering at all var bestLevel = this._findBestLevel(currentLevel, currentFragDuration, avgbw, minAutoLevel, maxAutoLevel, bufferStarvationDelay, config.abrBandWidthFactor, config.abrBandWidthUpFactor, levels); if (bestLevel >= 0) { return bestLevel; } else { logger["logger"].trace('rebuffering expected to happen, lets try to find a quality level minimizing the rebuffering'); // not possible to get rid of rebuffering ... let's try to find level that will guarantee less than maxStarvationDelay of rebuffering // if no matching level found, logic will return 0 var maxStarvationDelay = currentFragDuration ? Math.min(currentFragDuration, config.maxStarvationDelay) : config.maxStarvationDelay; var bwFactor = config.abrBandWidthFactor; var bwUpFactor = config.abrBandWidthUpFactor; if (bufferStarvationDelay === 0) { // in case buffer is empty, let's check if previous fragment was loaded to perform a bitrate test var bitrateTestDelay = this.bitrateTestDelay; if (bitrateTestDelay) { // if it is the case, then we need to adjust our max starvation delay using maxLoadingDelay config value // max video loading delay used in automatic start level selection : // in that mode ABR controller will ensure that video loading time (ie the time to fetch the first fragment at lowest quality level + // the time to fetch the fragment at the appropriate quality level is less than ```maxLoadingDelay``` ) // cap maxLoadingDelay and ensure it is not bigger 'than bitrate test' frag duration var maxLoadingDelay = currentFragDuration ? Math.min(currentFragDuration, config.maxLoadingDelay) : config.maxLoadingDelay; maxStarvationDelay = maxLoadingDelay - bitrateTestDelay; logger["logger"].trace("bitrate test took " + Math.round(1000 * bitrateTestDelay) + "ms, set first fragment max fetchDuration to " + Math.round(1000 * maxStarvationDelay) + " ms"); // don't use conservative factor on bitrate test bwFactor = bwUpFactor = 1; } } bestLevel = this._findBestLevel(currentLevel, currentFragDuration, avgbw, minAutoLevel, maxAutoLevel, bufferStarvationDelay + maxStarvationDelay, bwFactor, bwUpFactor, levels); return Math.max(bestLevel, 0); } } }]); return AbrController; }(event_handler); /* harmony default export */ var abr_controller = (abr_controller_AbrController); // CONCATENATED MODULE: ./src/controller/buffer-controller.ts function buffer_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /* * Buffer Controller */ var buffer_controller_MediaSource = getMediaSource(); var buffer_controller_BufferController = /*#__PURE__*/function (_EventHandler) { buffer_controller_inheritsLoose(BufferController, _EventHandler); // the value that we have set mediasource.duration to // (the actual duration may be tweaked slighly by the browser) // the value that we want to set mediaSource.duration to // the target duration of the current media playlist // current stream state: true - for live broadcast, false - for VoD content // cache the self generated object url to detect hijack of video tag // signals that the sourceBuffers need to be flushed // signals that mediaSource should have endOfStream called // this is optional because this property is removed from the class sometimes // The number of BUFFER_CODEC events received before any sourceBuffers are created // The total number of BUFFER_CODEC events received // A reference to the attached media element // A reference to the active media source // List of pending segments to be appended to source buffer // A guard to see if we are currently appending to the source buffer // counters function BufferController(hls) { var _this; _this = _EventHandler.call(this, hls, events["default"].MEDIA_ATTACHING, events["default"].MEDIA_DETACHING, events["default"].MANIFEST_PARSED, events["default"].BUFFER_RESET, events["default"].BUFFER_APPENDING, events["default"].BUFFER_CODECS, events["default"].BUFFER_EOS, events["default"].BUFFER_FLUSHING, events["default"].LEVEL_PTS_UPDATED, events["default"].LEVEL_UPDATED) || this; _this._msDuration = null; _this._levelDuration = null; _this._levelTargetDuration = 10; _this._live = null; _this._objectUrl = null; _this._needsFlush = false; _this._needsEos = false; _this.config = void 0; _this.audioTimestampOffset = void 0; _this.bufferCodecEventsExpected = 0; _this._bufferCodecEventsTotal = 0; _this.media = null; _this.mediaSource = null; _this.segments = []; _this.parent = void 0; _this.appending = false; _this.appended = 0; _this.appendError = 0; _this.flushBufferCounter = 0; _this.tracks = {}; _this.pendingTracks = {}; _this.sourceBuffer = {}; _this.flushRange = []; _this._onMediaSourceOpen = function () { logger["logger"].log('media source opened'); _this.hls.trigger(events["default"].MEDIA_ATTACHED, { media: _this.media }); var mediaSource = _this.mediaSource; if (mediaSource) { // once received, don't listen anymore to sourceopen event mediaSource.removeEventListener('sourceopen', _this._onMediaSourceOpen); } _this.checkPendingTracks(); }; _this._onMediaSourceClose = function () { logger["logger"].log('media source closed'); }; _this._onMediaSourceEnded = function () { logger["logger"].log('media source ended'); }; _this._onSBUpdateEnd = function () { // update timestampOffset if (_this.audioTimestampOffset && _this.sourceBuffer.audio) { var audioBuffer = _this.sourceBuffer.audio; logger["logger"].warn("change mpeg audio timestamp offset from " + audioBuffer.timestampOffset + " to " + _this.audioTimestampOffset); audioBuffer.timestampOffset = _this.audioTimestampOffset; delete _this.audioTimestampOffset; } if (_this._needsFlush) { _this.doFlush(); } if (_this._needsEos) { _this.checkEos(); } _this.appending = false; var parent = _this.parent; // count nb of pending segments waiting for appending on this sourcebuffer var pending = _this.segments.reduce(function (counter, segment) { return segment.parent === parent ? counter + 1 : counter; }, 0); // this.sourceBuffer is better to use than media.buffered as it is closer to the PTS data from the fragments var timeRanges = {}; var sbSet = _this.sourceBuffer; for (var streamType in sbSet) { var sb = sbSet[streamType]; if (!sb) { throw Error("handling source buffer update end error: source buffer for " + streamType + " uninitilized and unable to update buffered TimeRanges."); } timeRanges[streamType] = sb.buffered; } _this.hls.trigger(events["default"].BUFFER_APPENDED, { parent: parent, pending: pending, timeRanges: timeRanges }); // don't append in flushing mode if (!_this._needsFlush) { _this.doAppending(); } _this.updateMediaElementDuration(); // appending goes first if (pending === 0) { _this.flushLiveBackBuffer(); } }; _this._onSBUpdateError = function (event) { logger["logger"].error('sourceBuffer error:', event); // according to http://www.w3.org/TR/media-source/#sourcebuffer-append-error // this error might not always be fatal (it is fatal if decode error is set, in that case // it will be followed by a mediaElement error ...) _this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].BUFFER_APPENDING_ERROR, fatal: false }); // we don't need to do more than that, as accordin to the spec, updateend will be fired just after }; _this.config = hls.config; return _this; } var _proto = BufferController.prototype; _proto.destroy = function destroy() { event_handler.prototype.destroy.call(this); }; _proto.onLevelPtsUpdated = function onLevelPtsUpdated(data) { var type = data.type; var audioTrack = this.tracks.audio; // Adjusting `SourceBuffer.timestampOffset` (desired point in the timeline where the next frames should be appended) // in Chrome browser when we detect MPEG audio container and time delta between level PTS and `SourceBuffer.timestampOffset` // is greater than 100ms (this is enough to handle seek for VOD or level change for LIVE videos). At the time of change we issue // `SourceBuffer.abort()` and adjusting `SourceBuffer.timestampOffset` if `SourceBuffer.updating` is false or awaiting `updateend` // event if SB is in updating state. // More info here: https://github.com/video-dev/hls.js/issues/332#issuecomment-257986486 if (type === 'audio' && audioTrack && audioTrack.container === 'audio/mpeg') { // Chrome audio mp3 track var audioBuffer = this.sourceBuffer.audio; if (!audioBuffer) { throw Error('Level PTS Updated and source buffer for audio uninitalized'); } var delta = Math.abs(audioBuffer.timestampOffset - data.start); // adjust timestamp offset if time delta is greater than 100ms if (delta > 0.1) { var updating = audioBuffer.updating; try { audioBuffer.abort(); } catch (err) { logger["logger"].warn('can not abort audio buffer: ' + err); } if (!updating) { logger["logger"].warn('change mpeg audio timestamp offset from ' + audioBuffer.timestampOffset + ' to ' + data.start); audioBuffer.timestampOffset = data.start; } else { this.audioTimestampOffset = data.start; } } } }; _proto.onManifestParsed = function onManifestParsed(data) { // in case of alt audio (where all tracks have urls) 2 BUFFER_CODECS events will be triggered, one per stream controller // sourcebuffers will be created all at once when the expected nb of tracks will be reached // in case alt audio is not used, only one BUFFER_CODEC event will be fired from main stream controller // it will contain the expected nb of source buffers, no need to compute it var codecEvents = 2; if (data.audio && !data.video || !data.altAudio) { codecEvents = 1; } this.bufferCodecEventsExpected = this._bufferCodecEventsTotal = codecEvents; logger["logger"].log(this.bufferCodecEventsExpected + " bufferCodec event(s) expected"); }; _proto.onMediaAttaching = function onMediaAttaching(data) { var media = this.media = data.media; if (media && buffer_controller_MediaSource) { // setup the media source var ms = this.mediaSource = new buffer_controller_MediaSource(); // Media Source listeners ms.addEventListener('sourceopen', this._onMediaSourceOpen); ms.addEventListener('sourceended', this._onMediaSourceEnded); ms.addEventListener('sourceclose', this._onMediaSourceClose); // link video and media Source media.src = window.URL.createObjectURL(ms); // cache the locally generated object url this._objectUrl = media.src; } }; _proto.onMediaDetaching = function onMediaDetaching() { logger["logger"].log('media source detaching'); var ms = this.mediaSource; if (ms) { if (ms.readyState === 'open') { try { // endOfStream could trigger exception if any sourcebuffer is in updating state // we don't really care about checking sourcebuffer state here, // as we are anyway detaching the MediaSource // let's just avoid this exception to propagate ms.endOfStream(); } catch (err) { logger["logger"].warn("onMediaDetaching:" + err.message + " while calling endOfStream"); } } ms.removeEventListener('sourceopen', this._onMediaSourceOpen); ms.removeEventListener('sourceended', this._onMediaSourceEnded); ms.removeEventListener('sourceclose', this._onMediaSourceClose); // Detach properly the MediaSource from the HTMLMediaElement as // suggested in https://github.com/w3c/media-source/issues/53. if (this.media) { if (this._objectUrl) { window.URL.revokeObjectURL(this._objectUrl); } // clean up video tag src only if it's our own url. some external libraries might // hijack the video tag and change its 'src' without destroying the Hls instance first if (this.media.src === this._objectUrl) { this.media.removeAttribute('src'); this.media.load(); } else { logger["logger"].warn('media.src was changed by a third party - skip cleanup'); } } this.mediaSource = null; this.media = null; this._objectUrl = null; this.bufferCodecEventsExpected = this._bufferCodecEventsTotal; this.pendingTracks = {}; this.tracks = {}; this.sourceBuffer = {}; this.flushRange = []; this.segments = []; this.appended = 0; } this.hls.trigger(events["default"].MEDIA_DETACHED); }; _proto.checkPendingTracks = function checkPendingTracks() { var bufferCodecEventsExpected = this.bufferCodecEventsExpected, pendingTracks = this.pendingTracks; // Check if we've received all of the expected bufferCodec events. When none remain, create all the sourceBuffers at once. // This is important because the MSE spec allows implementations to throw QuotaExceededErrors if creating new sourceBuffers after // data has been appended to existing ones. // 2 tracks is the max (one for audio, one for video). If we've reach this max go ahead and create the buffers. var pendingTracksCount = Object.keys(pendingTracks).length; if (pendingTracksCount && !bufferCodecEventsExpected || pendingTracksCount === 2) { // ok, let's create them now ! this.createSourceBuffers(pendingTracks); this.pendingTracks = {}; // append any pending segments now ! this.doAppending(); } }; _proto.onBufferReset = function onBufferReset() { var sourceBuffer = this.sourceBuffer; for (var type in sourceBuffer) { var sb = sourceBuffer[type]; try { if (sb) { if (this.mediaSource) { this.mediaSource.removeSourceBuffer(sb); } sb.removeEventListener('updateend', this._onSBUpdateEnd); sb.removeEventListener('error', this._onSBUpdateError); } } catch (err) {} } this.sourceBuffer = {}; this.flushRange = []; this.segments = []; this.appended = 0; }; _proto.onBufferCodecs = function onBufferCodecs(tracks) { var _this2 = this; // if source buffer(s) not created yet, appended buffer tracks in this.pendingTracks // if sourcebuffers already created, do nothing ... if (Object.keys(this.sourceBuffer).length) { return; } Object.keys(tracks).forEach(function (trackName) { _this2.pendingTracks[trackName] = tracks[trackName]; }); this.bufferCodecEventsExpected = Math.max(this.bufferCodecEventsExpected - 1, 0); if (this.mediaSource && this.mediaSource.readyState === 'open') { this.checkPendingTracks(); } }; _proto.createSourceBuffers = function createSourceBuffers(tracks) { var sourceBuffer = this.sourceBuffer, mediaSource = this.mediaSource; if (!mediaSource) { throw Error('createSourceBuffers called when mediaSource was null'); } for (var trackName in tracks) { if (!sourceBuffer[trackName]) { var track = tracks[trackName]; if (!track) { throw Error("source buffer exists for track " + trackName + ", however track does not"); } // use levelCodec as first priority var codec = track.levelCodec || track.codec; var mimeType = track.container + ";codecs=" + codec; logger["logger"].log("creating sourceBuffer(" + mimeType + ")"); try { var sb = sourceBuffer[trackName] = mediaSource.addSourceBuffer(mimeType); sb.addEventListener('updateend', this._onSBUpdateEnd); sb.addEventListener('error', this._onSBUpdateError); this.tracks[trackName] = { buffer: sb, codec: codec, id: track.id, container: track.container, levelCodec: track.levelCodec }; } catch (err) { logger["logger"].error("error while trying to add sourceBuffer:" + err.message); this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].BUFFER_ADD_CODEC_ERROR, fatal: false, err: err, mimeType: mimeType }); } } } this.hls.trigger(events["default"].BUFFER_CREATED, { tracks: this.tracks }); }; _proto.onBufferAppending = function onBufferAppending(data) { if (!this._needsFlush) { if (!this.segments) { this.segments = [data]; } else { this.segments.push(data); } this.doAppending(); } } // on BUFFER_EOS mark matching sourcebuffer(s) as ended and trigger checkEos() // an undefined data.type will mark all buffers as EOS. ; _proto.onBufferEos = function onBufferEos(data) { for (var type in this.sourceBuffer) { if (!data.type || data.type === type) { var sb = this.sourceBuffer[type]; if (sb && !sb.ended) { sb.ended = true; logger["logger"].log(type + " sourceBuffer now EOS"); } } } this.checkEos(); } // if all source buffers are marked as ended, signal endOfStream() to MediaSource. ; _proto.checkEos = function checkEos() { var sourceBuffer = this.sourceBuffer, mediaSource = this.mediaSource; if (!mediaSource || mediaSource.readyState !== 'open') { this._needsEos = false; return; } for (var type in sourceBuffer) { var sb = sourceBuffer[type]; if (!sb) continue; if (!sb.ended) { return; } if (sb.updating) { this._needsEos = true; return; } } logger["logger"].log('all media data are available, signal endOfStream() to MediaSource and stop loading fragment'); // Notify the media element that it now has all of the media data try { mediaSource.endOfStream(); } catch (e) { logger["logger"].warn('exception while calling mediaSource.endOfStream()'); } this._needsEos = false; }; _proto.onBufferFlushing = function onBufferFlushing(data) { if (data.type) { this.flushRange.push({ start: data.startOffset, end: data.endOffset, type: data.type }); } else { this.flushRange.push({ start: data.startOffset, end: data.endOffset, type: 'video' }); this.flushRange.push({ start: data.startOffset, end: data.endOffset, type: 'audio' }); } // attempt flush immediately this.flushBufferCounter = 0; this.doFlush(); }; _proto.flushLiveBackBuffer = function flushLiveBackBuffer() { // clear back buffer for live only if (!this._live) { return; } var liveBackBufferLength = this.config.liveBackBufferLength; if (!isFinite(liveBackBufferLength) || liveBackBufferLength < 0) { return; } if (!this.media) { logger["logger"].error('flushLiveBackBuffer called without attaching media'); return; } var currentTime = this.media.currentTime; var sourceBuffer = this.sourceBuffer; var bufferTypes = Object.keys(sourceBuffer); var targetBackBufferPosition = currentTime - Math.max(liveBackBufferLength, this._levelTargetDuration); for (var index = bufferTypes.length - 1; index >= 0; index--) { var bufferType = bufferTypes[index]; var sb = sourceBuffer[bufferType]; if (sb) { var buffered = sb.buffered; // when target buffer start exceeds actual buffer start if (buffered.length > 0 && targetBackBufferPosition > buffered.start(0)) { // remove buffer up until current time minus minimum back buffer length (removing buffer too close to current // time will lead to playback freezing) // credits for level target duration - https://github.com/videojs/http-streaming/blob/3132933b6aa99ddefab29c10447624efd6fd6e52/src/segment-loader.js#L91 if (this.removeBufferRange(bufferType, sb, 0, targetBackBufferPosition)) { this.hls.trigger(events["default"].LIVE_BACK_BUFFER_REACHED, { bufferEnd: targetBackBufferPosition }); } } } } }; _proto.onLevelUpdated = function onLevelUpdated(_ref) { var details = _ref.details; if (details.fragments.length > 0) { this._levelDuration = details.totalduration + details.fragments[0].start; this._levelTargetDuration = details.averagetargetduration || details.targetduration || 10; this._live = details.live; this.updateMediaElementDuration(); } } /** * Update Media Source duration to current level duration or override to Infinity if configuration parameter * 'liveDurationInfinity` is set to `true` * More details: https://github.com/video-dev/hls.js/issues/355 */ ; _proto.updateMediaElementDuration = function updateMediaElementDuration() { var config = this.config; var duration; if (this._levelDuration === null || !this.media || !this.mediaSource || !this.sourceBuffer || this.media.readyState === 0 || this.mediaSource.readyState !== 'open') { return; } for (var type in this.sourceBuffer) { var sb = this.sourceBuffer[type]; if (sb && sb.updating === true) { // can't set duration whilst a buffer is updating return; } } duration = this.media.duration; // initialise to the value that the media source is reporting if (this._msDuration === null) { this._msDuration = this.mediaSource.duration; } if (this._live === true && config.liveDurationInfinity === true) { // Override duration to Infinity logger["logger"].log('Media Source duration is set to Infinity'); this._msDuration = this.mediaSource.duration = Infinity; } else if (this._levelDuration > this._msDuration && this._levelDuration > duration || !Object(number["isFiniteNumber"])(duration)) { // levelDuration was the last value we set. // not using mediaSource.duration as the browser may tweak this value // only update Media Source duration if its value increase, this is to avoid // flushing already buffered portion when switching between quality level logger["logger"].log("Updating Media Source duration to " + this._levelDuration.toFixed(3)); this._msDuration = this.mediaSource.duration = this._levelDuration; } }; _proto.doFlush = function doFlush() { // loop through all buffer ranges to flush while (this.flushRange.length) { var range = this.flushRange[0]; // flushBuffer will abort any buffer append in progress and flush Audio/Video Buffer if (this.flushBuffer(range.start, range.end, range.type)) { // range flushed, remove from flush array this.flushRange.shift(); this.flushBufferCounter = 0; } else { this._needsFlush = true; // avoid looping, wait for SB update end to retrigger a flush return; } } if (this.flushRange.length === 0) { // everything flushed this._needsFlush = false; // let's recompute this.appended, which is used to avoid flush looping var appended = 0; var sourceBuffer = this.sourceBuffer; try { for (var type in sourceBuffer) { var sb = sourceBuffer[type]; if (sb) { appended += sb.buffered.length; } } } catch (error) { // error could be thrown while accessing buffered, in case sourcebuffer has already been removed from MediaSource // this is harmess at this stage, catch this to avoid reporting an internal exception logger["logger"].error('error while accessing sourceBuffer.buffered'); } this.appended = appended; this.hls.trigger(events["default"].BUFFER_FLUSHED); } }; _proto.doAppending = function doAppending() { var config = this.config, hls = this.hls, segments = this.segments, sourceBuffer = this.sourceBuffer; if (!Object.keys(sourceBuffer).length) { // early exit if no source buffers have been initialized yet return; } if (!this.media || this.media.error) { this.segments = []; logger["logger"].error('trying to append although a media error occured, flush segment and abort'); return; } if (this.appending) { // logger.log(`sb appending in progress`); return; } var segment = segments.shift(); if (!segment) { // handle undefined shift return; } try { var sb = sourceBuffer[segment.type]; if (!sb) { // in case we don't have any source buffer matching with this segment type, // it means that Mediasource fails to create sourcebuffer // discard this segment, and trigger update end this._onSBUpdateEnd(); return; } if (sb.updating) { // if we are still updating the source buffer from the last segment, place this back at the front of the queue segments.unshift(segment); return; } // reset sourceBuffer ended flag before appending segment sb.ended = false; // logger.log(`appending ${segment.content} ${type} SB, size:${segment.data.length}, ${segment.parent}`); this.parent = segment.parent; sb.appendBuffer(segment.data); this.appendError = 0; this.appended++; this.appending = true; } catch (err) { // in case any error occured while appending, put back segment in segments table logger["logger"].error("error while trying to append buffer:" + err.message); segments.unshift(segment); var event = { type: errors["ErrorTypes"].MEDIA_ERROR, parent: segment.parent, details: '', fatal: false }; if (err.code === 22) { // QuotaExceededError: http://www.w3.org/TR/html5/infrastructure.html#quotaexceedederror // let's stop appending any segments, and report BUFFER_FULL_ERROR error this.segments = []; event.details = errors["ErrorDetails"].BUFFER_FULL_ERROR; } else { this.appendError++; event.details = errors["ErrorDetails"].BUFFER_APPEND_ERROR; /* with UHD content, we could get loop of quota exceeded error until browser is able to evict some data from sourcebuffer. retrying help recovering this */ if (this.appendError > config.appendErrorMaxRetry) { logger["logger"].log("fail " + config.appendErrorMaxRetry + " times to append segment in sourceBuffer"); this.segments = []; event.fatal = true; } } hls.trigger(events["default"].ERROR, event); } } /* flush specified buffered range, return true once range has been flushed. as sourceBuffer.remove() is asynchronous, flushBuffer will be retriggered on sourceBuffer update end */ ; _proto.flushBuffer = function flushBuffer(startOffset, endOffset, sbType) { var sourceBuffer = this.sourceBuffer; // exit if no sourceBuffers are initialized if (!Object.keys(sourceBuffer).length) { return true; } var currentTime = 'null'; if (this.media) { currentTime = this.media.currentTime.toFixed(3); } logger["logger"].log("flushBuffer,pos/start/end: " + currentTime + "/" + startOffset + "/" + endOffset); // safeguard to avoid infinite looping : don't try to flush more than the nb of appended segments if (this.flushBufferCounter >= this.appended) { logger["logger"].warn('abort flushing too many retries'); return true; } var sb = sourceBuffer[sbType]; // we are going to flush buffer, mark source buffer as 'not ended' if (sb) { sb.ended = false; if (!sb.updating) { if (this.removeBufferRange(sbType, sb, startOffset, endOffset)) { this.flushBufferCounter++; return false; } } else { logger["logger"].warn('cannot flush, sb updating in progress'); return false; } } logger["logger"].log('buffer flushed'); // everything flushed ! return true; } /** * Removes first buffered range from provided source buffer that lies within given start and end offsets. * * @param {string} type Type of the source buffer, logging purposes only. * @param {SourceBuffer} sb Target SourceBuffer instance. * @param {number} startOffset * @param {number} endOffset * * @returns {boolean} True when source buffer remove requested. */ ; _proto.removeBufferRange = function removeBufferRange(type, sb, startOffset, endOffset) { try { for (var i = 0; i < sb.buffered.length; i++) { var bufStart = sb.buffered.start(i); var bufEnd = sb.buffered.end(i); var removeStart = Math.max(bufStart, startOffset); var removeEnd = Math.min(bufEnd, endOffset); /* sometimes sourcebuffer.remove() does not flush the exact expected time range. to avoid rounding issues/infinite loop, only flush buffer range of length greater than 500ms. */ if (Math.min(removeEnd, bufEnd) - removeStart > 0.5) { var currentTime = 'null'; if (this.media) { currentTime = this.media.currentTime.toString(); } logger["logger"].log("sb remove " + type + " [" + removeStart + "," + removeEnd + "], of [" + bufStart + "," + bufEnd + "], pos:" + currentTime); sb.remove(removeStart, removeEnd); return true; } } } catch (error) { logger["logger"].warn('removeBufferRange failed', error); } return false; }; return BufferController; }(event_handler); /* harmony default export */ var buffer_controller = (buffer_controller_BufferController); // CONCATENATED MODULE: ./src/controller/cap-level-controller.js function cap_level_controller_defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } function cap_level_controller_createClass(Constructor, protoProps, staticProps) { if (protoProps) cap_level_controller_defineProperties(Constructor.prototype, protoProps); if (staticProps) cap_level_controller_defineProperties(Constructor, staticProps); return Constructor; } function cap_level_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /* * cap stream level to media size dimension controller */ var cap_level_controller_CapLevelController = /*#__PURE__*/function (_EventHandler) { cap_level_controller_inheritsLoose(CapLevelController, _EventHandler); function CapLevelController(hls) { var _this; _this = _EventHandler.call(this, hls, events["default"].FPS_DROP_LEVEL_CAPPING, events["default"].MEDIA_ATTACHING, events["default"].MANIFEST_PARSED, events["default"].LEVELS_UPDATED, events["default"].BUFFER_CODECS, events["default"].MEDIA_DETACHING) || this; _this.autoLevelCapping = Number.POSITIVE_INFINITY; _this.firstLevel = null; _this.levels = []; _this.media = null; _this.restrictedLevels = []; _this.timer = null; _this.clientRect = null; return _this; } var _proto = CapLevelController.prototype; _proto.destroy = function destroy() { if (this.hls.config.capLevelToPlayerSize) { this.media = null; this.clientRect = null; this.stopCapping(); } }; _proto.onFpsDropLevelCapping = function onFpsDropLevelCapping(data) { // Don't add a restricted level more than once if (CapLevelController.isLevelAllowed(data.droppedLevel, this.restrictedLevels)) { this.restrictedLevels.push(data.droppedLevel); } }; _proto.onMediaAttaching = function onMediaAttaching(data) { this.media = data.media instanceof window.HTMLVideoElement ? data.media : null; }; _proto.onManifestParsed = function onManifestParsed(data) { var hls = this.hls; this.restrictedLevels = []; this.levels = data.levels; this.firstLevel = data.firstLevel; if (hls.config.capLevelToPlayerSize && data.video) { // Start capping immediately if the manifest has signaled video codecs this.startCapping(); } } // Only activate capping when playing a video stream; otherwise, multi-bitrate audio-only streams will be restricted // to the first level ; _proto.onBufferCodecs = function onBufferCodecs(data) { var hls = this.hls; if (hls.config.capLevelToPlayerSize && data.video) { // If the manifest did not signal a video codec capping has been deferred until we're certain video is present this.startCapping(); } }; _proto.onLevelsUpdated = function onLevelsUpdated(data) { this.levels = data.levels; }; _proto.onMediaDetaching = function onMediaDetaching() { this.stopCapping(); }; _proto.detectPlayerSize = function detectPlayerSize() { if (this.media) { var levelsLength = this.levels ? this.levels.length : 0; if (levelsLength) { var hls = this.hls; hls.autoLevelCapping = this.getMaxLevel(levelsLength - 1); if (hls.autoLevelCapping > this.autoLevelCapping) { // if auto level capping has a higher value for the previous one, flush the buffer using nextLevelSwitch // usually happen when the user go to the fullscreen mode. hls.streamController.nextLevelSwitch(); } this.autoLevelCapping = hls.autoLevelCapping; } } } /* * returns level should be the one with the dimensions equal or greater than the media (player) dimensions (so the video will be downscaled) */ ; _proto.getMaxLevel = function getMaxLevel(capLevelIndex) { var _this2 = this; if (!this.levels) { return -1; } var validLevels = this.levels.filter(function (level, index) { return CapLevelController.isLevelAllowed(index, _this2.restrictedLevels) && index <= capLevelIndex; }); this.clientRect = null; return CapLevelController.getMaxLevelByMediaSize(validLevels, this.mediaWidth, this.mediaHeight); }; _proto.startCapping = function startCapping() { if (this.timer) { // Don't reset capping if started twice; this can happen if the manifest signals a video codec return; } this.autoLevelCapping = Number.POSITIVE_INFINITY; this.hls.firstLevel = this.getMaxLevel(this.firstLevel); clearInterval(this.timer); this.timer = setInterval(this.detectPlayerSize.bind(this), 1000); this.detectPlayerSize(); }; _proto.stopCapping = function stopCapping() { this.restrictedLevels = []; this.firstLevel = null; this.autoLevelCapping = Number.POSITIVE_INFINITY; if (this.timer) { this.timer = clearInterval(this.timer); this.timer = null; } }; _proto.getDimensions = function getDimensions() { if (this.clientRect) { return this.clientRect; } var media = this.media; var boundsRect = { width: 0, height: 0 }; if (media) { var clientRect = media.getBoundingClientRect(); boundsRect.width = clientRect.width; boundsRect.height = clientRect.height; if (!boundsRect.width && !boundsRect.height) { // When the media element has no width or height (equivalent to not being in the DOM), // then use its width and height attributes (media.width, media.height) boundsRect.width = clientRect.right - clientRect.left || media.width || 0; boundsRect.height = clientRect.bottom - clientRect.top || media.height || 0; } } this.clientRect = boundsRect; return boundsRect; }; CapLevelController.isLevelAllowed = function isLevelAllowed(level, restrictedLevels) { if (restrictedLevels === void 0) { restrictedLevels = []; } return restrictedLevels.indexOf(level) === -1; }; CapLevelController.getMaxLevelByMediaSize = function getMaxLevelByMediaSize(levels, width, height) { if (!levels || levels && !levels.length) { return -1; } // Levels can have the same dimensions but differing bandwidths - since levels are ordered, we can look to the next // to determine whether we've chosen the greatest bandwidth for the media's dimensions var atGreatestBandiwdth = function atGreatestBandiwdth(curLevel, nextLevel) { if (!nextLevel) { return true; } return curLevel.width !== nextLevel.width || curLevel.height !== nextLevel.height; }; // If we run through the loop without breaking, the media's dimensions are greater than every level, so default to // the max level var maxLevelIndex = levels.length - 1; for (var i = 0; i < levels.length; i += 1) { var level = levels[i]; if ((level.width >= width || level.height >= height) && atGreatestBandiwdth(level, levels[i + 1])) { maxLevelIndex = i; break; } } return maxLevelIndex; }; cap_level_controller_createClass(CapLevelController, [{ key: "mediaWidth", get: function get() { return this.getDimensions().width * CapLevelController.contentScaleFactor; } }, { key: "mediaHeight", get: function get() { return this.getDimensions().height * CapLevelController.contentScaleFactor; } }], [{ key: "contentScaleFactor", get: function get() { var pixelRatio = 1; try { pixelRatio = window.devicePixelRatio; } catch (e) {} return pixelRatio; } }]); return CapLevelController; }(event_handler); /* harmony default export */ var cap_level_controller = (cap_level_controller_CapLevelController); // CONCATENATED MODULE: ./src/controller/fps-controller.js function fps_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /* * FPS Controller */ var fps_controller_window = window, fps_controller_performance = fps_controller_window.performance; var fps_controller_FPSController = /*#__PURE__*/function (_EventHandler) { fps_controller_inheritsLoose(FPSController, _EventHandler); function FPSController(hls) { return _EventHandler.call(this, hls, events["default"].MEDIA_ATTACHING) || this; } var _proto = FPSController.prototype; _proto.destroy = function destroy() { if (this.timer) { clearInterval(this.timer); } this.isVideoPlaybackQualityAvailable = false; }; _proto.onMediaAttaching = function onMediaAttaching(data) { var config = this.hls.config; if (config.capLevelOnFPSDrop) { var video = this.video = data.media instanceof window.HTMLVideoElement ? data.media : null; if (typeof video.getVideoPlaybackQuality === 'function') { this.isVideoPlaybackQualityAvailable = true; } clearInterval(this.timer); this.timer = setInterval(this.checkFPSInterval.bind(this), config.fpsDroppedMonitoringPeriod); } }; _proto.checkFPS = function checkFPS(video, decodedFrames, droppedFrames) { var currentTime = fps_controller_performance.now(); if (decodedFrames) { if (this.lastTime) { var currentPeriod = currentTime - this.lastTime, currentDropped = droppedFrames - this.lastDroppedFrames, currentDecoded = decodedFrames - this.lastDecodedFrames, droppedFPS = 1000 * currentDropped / currentPeriod, hls = this.hls; hls.trigger(events["default"].FPS_DROP, { currentDropped: currentDropped, currentDecoded: currentDecoded, totalDroppedFrames: droppedFrames }); if (droppedFPS > 0) { // logger.log('checkFPS : droppedFPS/decodedFPS:' + droppedFPS/(1000 * currentDecoded / currentPeriod)); if (currentDropped > hls.config.fpsDroppedMonitoringThreshold * currentDecoded) { var currentLevel = hls.currentLevel; logger["logger"].warn('drop FPS ratio greater than max allowed value for currentLevel: ' + currentLevel); if (currentLevel > 0 && (hls.autoLevelCapping === -1 || hls.autoLevelCapping >= currentLevel)) { currentLevel = currentLevel - 1; hls.trigger(events["default"].FPS_DROP_LEVEL_CAPPING, { level: currentLevel, droppedLevel: hls.currentLevel }); hls.autoLevelCapping = currentLevel; hls.streamController.nextLevelSwitch(); } } } } this.lastTime = currentTime; this.lastDroppedFrames = droppedFrames; this.lastDecodedFrames = decodedFrames; } }; _proto.checkFPSInterval = function checkFPSInterval() { var video = this.video; if (video) { if (this.isVideoPlaybackQualityAvailable) { var videoPlaybackQuality = video.getVideoPlaybackQuality(); this.checkFPS(video, videoPlaybackQuality.totalVideoFrames, videoPlaybackQuality.droppedVideoFrames); } else { this.checkFPS(video, video.webkitDecodedFrameCount, video.webkitDroppedFrameCount); } } }; return FPSController; }(event_handler); /* harmony default export */ var fps_controller = (fps_controller_FPSController); // CONCATENATED MODULE: ./src/utils/xhr-loader.js /** * XHR based logger */ var xhr_loader_XhrLoader = /*#__PURE__*/function () { function XhrLoader(config) { if (config && config.xhrSetup) { this.xhrSetup = config.xhrSetup; } } var _proto = XhrLoader.prototype; _proto.destroy = function destroy() { this.abort(); this.loader = null; }; _proto.abort = function abort() { var loader = this.loader; if (loader && loader.readyState !== 4) { this.stats.aborted = true; loader.abort(); } window.clearTimeout(this.requestTimeout); this.requestTimeout = null; window.clearTimeout(this.retryTimeout); this.retryTimeout = null; }; _proto.load = function load(context, config, callbacks) { this.context = context; this.config = config; this.callbacks = callbacks; this.stats = { trequest: window.performance.now(), retry: 0 }; this.retryDelay = config.retryDelay; this.loadInternal(); }; _proto.loadInternal = function loadInternal() { var xhr, context = this.context; xhr = this.loader = new window.XMLHttpRequest(); var stats = this.stats; stats.tfirst = 0; stats.loaded = 0; var xhrSetup = this.xhrSetup; try { if (xhrSetup) { try { xhrSetup(xhr, context.url); } catch (e) { // fix xhrSetup: (xhr, url) => {xhr.setRequestHeader("Content-Language", "test");} // not working, as xhr.setRequestHeader expects xhr.readyState === OPEN xhr.open('GET', context.url, true); xhrSetup(xhr, context.url); } } if (!xhr.readyState) { xhr.open('GET', context.url, true); } } catch (e) { // IE11 throws an exception on xhr.open if attempting to access an HTTP resource over HTTPS this.callbacks.onError({ code: xhr.status, text: e.message }, context, xhr); return; } if (context.rangeEnd) { xhr.setRequestHeader('Range', 'bytes=' + context.rangeStart + '-' + (context.rangeEnd - 1)); } xhr.onreadystatechange = this.readystatechange.bind(this); xhr.onprogress = this.loadprogress.bind(this); xhr.responseType = context.responseType; // setup timeout before we perform request this.requestTimeout = window.setTimeout(this.loadtimeout.bind(this), this.config.timeout); xhr.send(); }; _proto.readystatechange = function readystatechange(event) { var xhr = event.currentTarget, readyState = xhr.readyState, stats = this.stats, context = this.context, config = this.config; // don't proceed if xhr has been aborted if (stats.aborted) { return; } // >= HEADERS_RECEIVED if (readyState >= 2) { // clear xhr timeout and rearm it if readyState less than 4 window.clearTimeout(this.requestTimeout); if (stats.tfirst === 0) { stats.tfirst = Math.max(window.performance.now(), stats.trequest); } if (readyState === 4) { var status = xhr.status; // http status between 200 to 299 are all successful if (status >= 200 && status < 300) { stats.tload = Math.max(stats.tfirst, window.performance.now()); var data, len; if (context.responseType === 'arraybuffer') { data = xhr.response; len = data.byteLength; } else { data = xhr.responseText; len = data.length; } stats.loaded = stats.total = len; var response = { url: xhr.responseURL, data: data }; this.callbacks.onSuccess(response, stats, context, xhr); } else { // if max nb of retries reached or if http status between 400 and 499 (such error cannot be recovered, retrying is useless), return error if (stats.retry >= config.maxRetry || status >= 400 && status < 499) { logger["logger"].error(status + " while loading " + context.url); this.callbacks.onError({ code: status, text: xhr.statusText }, context, xhr); } else { // retry logger["logger"].warn(status + " while loading " + context.url + ", retrying in " + this.retryDelay + "..."); // aborts and resets internal state this.destroy(); // schedule retry this.retryTimeout = window.setTimeout(this.loadInternal.bind(this), this.retryDelay); // set exponential backoff this.retryDelay = Math.min(2 * this.retryDelay, config.maxRetryDelay); stats.retry++; } } } else { // readyState >= 2 AND readyState !==4 (readyState = HEADERS_RECEIVED || LOADING) rearm timeout as xhr not finished yet this.requestTimeout = window.setTimeout(this.loadtimeout.bind(this), config.timeout); } } }; _proto.loadtimeout = function loadtimeout() { logger["logger"].warn("timeout while loading " + this.context.url); this.callbacks.onTimeout(this.stats, this.context, null); }; _proto.loadprogress = function loadprogress(event) { var xhr = event.currentTarget, stats = this.stats; stats.loaded = event.loaded; if (event.lengthComputable) { stats.total = event.total; } var onProgress = this.callbacks.onProgress; if (onProgress) { // third arg is to provide on progress data onProgress(stats, this.context, null, xhr); } }; return XhrLoader; }(); /* harmony default export */ var xhr_loader = (xhr_loader_XhrLoader); // CONCATENATED MODULE: ./src/controller/audio-track-controller.js function audio_track_controller_defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } function audio_track_controller_createClass(Constructor, protoProps, staticProps) { if (protoProps) audio_track_controller_defineProperties(Constructor.prototype, protoProps); if (staticProps) audio_track_controller_defineProperties(Constructor, staticProps); return Constructor; } function audio_track_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /** * @class AudioTrackController * @implements {EventHandler} * * Handles main manifest and audio-track metadata loaded, * owns and exposes the selectable audio-tracks data-models. * * Exposes internal interface to select available audio-tracks. * * Handles errors on loading audio-track playlists. Manages fallback mechanism * with redundants tracks (group-IDs). * * Handles level-loading and group-ID switches for video (fallback on video levels), * and eventually adapts the audio-track group-ID to match. * * @fires AUDIO_TRACK_LOADING * @fires AUDIO_TRACK_SWITCHING * @fires AUDIO_TRACKS_UPDATED * @fires ERROR * */ var audio_track_controller_AudioTrackController = /*#__PURE__*/function (_TaskLoop) { audio_track_controller_inheritsLoose(AudioTrackController, _TaskLoop); function AudioTrackController(hls) { var _this; _this = _TaskLoop.call(this, hls, events["default"].MANIFEST_LOADING, events["default"].MANIFEST_PARSED, events["default"].AUDIO_TRACK_LOADED, events["default"].AUDIO_TRACK_SWITCHED, events["default"].LEVEL_LOADED, events["default"].ERROR) || this; /** * @private * Currently selected index in `tracks` * @member {number} trackId */ _this._trackId = -1; /** * @private * If should select tracks according to default track attribute * @member {boolean} _selectDefaultTrack */ _this._selectDefaultTrack = true; /** * @public * All tracks available * @member {AudioTrack[]} */ _this.tracks = []; /** * @public * List of blacklisted audio track IDs (that have caused failure) * @member {number[]} */ _this.trackIdBlacklist = Object.create(null); /** * @public * The currently running group ID for audio * (we grab this on manifest-parsed and new level-loaded) * @member {string} */ _this.audioGroupId = null; return _this; } /** * Reset audio tracks on new manifest loading. */ var _proto = AudioTrackController.prototype; _proto.onManifestLoading = function onManifestLoading() { this.tracks = []; this._trackId = -1; this._selectDefaultTrack = true; } /** * Store tracks data from manifest parsed data. * * Trigger AUDIO_TRACKS_UPDATED event. * * @param {*} data */ ; _proto.onManifestParsed = function onManifestParsed(data) { var tracks = this.tracks = data.audioTracks || []; this.hls.trigger(events["default"].AUDIO_TRACKS_UPDATED, { audioTracks: tracks }); this._selectAudioGroup(this.hls.nextLoadLevel); } /** * Store track details of loaded track in our data-model. * * Set-up metadata update interval task for live-mode streams. * * @param {*} data */ ; _proto.onAudioTrackLoaded = function onAudioTrackLoaded(data) { if (data.id >= this.tracks.length) { logger["logger"].warn('Invalid audio track id:', data.id); return; } logger["logger"].log("audioTrack " + data.id + " loaded"); this.tracks[data.id].details = data.details; // check if current playlist is a live playlist // and if we have already our reload interval setup if (data.details.live && !this.hasInterval()) { // if live playlist we will have to reload it periodically // set reload period to playlist target duration var updatePeriodMs = data.details.targetduration * 1000; this.setInterval(updatePeriodMs); } if (!data.details.live && this.hasInterval()) { // playlist is not live and timer is scheduled: cancel it this.clearInterval(); } } /** * Update the internal group ID to any audio-track we may have set manually * or because of a failure-handling fallback. * * Quality-levels should update to that group ID in this case. * * @param {*} data */ ; _proto.onAudioTrackSwitched = function onAudioTrackSwitched(data) { var audioGroupId = this.tracks[data.id].groupId; if (audioGroupId && this.audioGroupId !== audioGroupId) { this.audioGroupId = audioGroupId; } } /** * When a level gets loaded, if it has redundant audioGroupIds (in the same ordinality as it's redundant URLs) * we are setting our audio-group ID internally to the one set, if it is different from the group ID currently set. * * If group-ID got update, we re-select the appropriate audio-track with this group-ID matching the currently * selected one (based on NAME property). * * @param {*} data */ ; _proto.onLevelLoaded = function onLevelLoaded(data) { this._selectAudioGroup(data.level); } /** * Handle network errors loading audio track manifests * and also pausing on any netwok errors. * * @param {ErrorEventData} data */ ; _proto.onError = function onError(data) { // Only handle network errors if (data.type !== errors["ErrorTypes"].NETWORK_ERROR) { return; } // If fatal network error, cancel update task if (data.fatal) { this.clearInterval(); } // If not an audio-track loading error don't handle further if (data.details !== errors["ErrorDetails"].AUDIO_TRACK_LOAD_ERROR) { return; } logger["logger"].warn('Network failure on audio-track id:', data.context.id); this._handleLoadError(); } /** * @type {AudioTrack[]} Audio-track list we own */ ; /** * @private * @param {number} newId */ _proto._setAudioTrack = function _setAudioTrack(newId) { // noop on same audio track id as already set if (this._trackId === newId && this.tracks[this._trackId].details) { logger["logger"].debug('Same id as current audio-track passed, and track details available -> no-op'); return; } // check if level idx is valid if (newId < 0 || newId >= this.tracks.length) { logger["logger"].warn('Invalid id passed to audio-track controller'); return; } var audioTrack = this.tracks[newId]; logger["logger"].log("Now switching to audio-track index " + newId); // stopping live reloading timer if any this.clearInterval(); this._trackId = newId; var url = audioTrack.url, type = audioTrack.type, id = audioTrack.id; this.hls.trigger(events["default"].AUDIO_TRACK_SWITCHING, { id: id, type: type, url: url }); this._loadTrackDetailsIfNeeded(audioTrack); } /** * @override */ ; _proto.doTick = function doTick() { this._updateTrack(this._trackId); } /** * @param levelId * @private */ ; _proto._selectAudioGroup = function _selectAudioGroup(levelId) { var levelInfo = this.hls.levels[levelId]; if (!levelInfo || !levelInfo.audioGroupIds) { return; } var audioGroupId = levelInfo.audioGroupIds[levelInfo.urlId]; if (this.audioGroupId !== audioGroupId) { this.audioGroupId = audioGroupId; this._selectInitialAudioTrack(); } } /** * Select initial track * @private */ ; _proto._selectInitialAudioTrack = function _selectInitialAudioTrack() { var _this2 = this; var tracks = this.tracks; if (!tracks.length) { return; } var currentAudioTrack = this.tracks[this._trackId]; var name = null; if (currentAudioTrack) { name = currentAudioTrack.name; } // Pre-select default tracks if there are any if (this._selectDefaultTrack) { var defaultTracks = tracks.filter(function (track) { return track.default; }); if (defaultTracks.length) { tracks = defaultTracks; } else { logger["logger"].warn('No default audio tracks defined'); } } var trackFound = false; var traverseTracks = function traverseTracks() { // Select track with right group ID tracks.forEach(function (track) { if (trackFound) { return; } // We need to match the (pre-)selected group ID // and the NAME of the current track. if ((!_this2.audioGroupId || track.groupId === _this2.audioGroupId) && (!name || name === track.name)) { // If there was a previous track try to stay with the same `NAME`. // It should be unique across tracks of same group, and consistent through redundant track groups. _this2._setAudioTrack(track.id); trackFound = true; } }); }; traverseTracks(); if (!trackFound) { name = null; traverseTracks(); } if (!trackFound) { logger["logger"].error("No track found for running audio group-ID: " + this.audioGroupId); this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: errors["ErrorDetails"].AUDIO_TRACK_LOAD_ERROR, fatal: true }); } } /** * @private * @param {AudioTrack} audioTrack * @returns {boolean} */ ; _proto._needsTrackLoading = function _needsTrackLoading(audioTrack) { var details = audioTrack.details, url = audioTrack.url; if (!details || details.live) { // check if we face an audio track embedded in main playlist (audio track without URI attribute) return !!url; } return false; } /** * @private * @param {AudioTrack} audioTrack */ ; _proto._loadTrackDetailsIfNeeded = function _loadTrackDetailsIfNeeded(audioTrack) { if (this._needsTrackLoading(audioTrack)) { var url = audioTrack.url, id = audioTrack.id; // track not retrieved yet, or live playlist we need to (re)load it logger["logger"].log("loading audio-track playlist for id: " + id); this.hls.trigger(events["default"].AUDIO_TRACK_LOADING, { url: url, id: id }); } } /** * @private * @param {number} newId */ ; _proto._updateTrack = function _updateTrack(newId) { // check if level idx is valid if (newId < 0 || newId >= this.tracks.length) { return; } // stopping live reloading timer if any this.clearInterval(); this._trackId = newId; logger["logger"].log("trying to update audio-track " + newId); var audioTrack = this.tracks[newId]; this._loadTrackDetailsIfNeeded(audioTrack); } /** * @private */ ; _proto._handleLoadError = function _handleLoadError() { // First, let's black list current track id this.trackIdBlacklist[this._trackId] = true; // Let's try to fall back on a functional audio-track with the same group ID var previousId = this._trackId; var _this$tracks$previous = this.tracks[previousId], name = _this$tracks$previous.name, language = _this$tracks$previous.language, groupId = _this$tracks$previous.groupId; logger["logger"].warn("Loading failed on audio track id: " + previousId + ", group-id: " + groupId + ", name/language: \"" + name + "\" / \"" + language + "\""); // Find a non-blacklisted track ID with the same NAME // At least a track that is not blacklisted, thus on another group-ID. var newId = previousId; for (var i = 0; i < this.tracks.length; i++) { if (this.trackIdBlacklist[i]) { continue; } var newTrack = this.tracks[i]; if (newTrack.name === name) { newId = i; break; } } if (newId === previousId) { logger["logger"].warn("No fallback audio-track found for name/language: \"" + name + "\" / \"" + language + "\""); return; } logger["logger"].log('Attempting audio-track fallback id:', newId, 'group-id:', this.tracks[newId].groupId); this._setAudioTrack(newId); }; audio_track_controller_createClass(AudioTrackController, [{ key: "audioTracks", get: function get() { return this.tracks; } /** * @type {number} Index into audio-tracks list of currently selected track. */ }, { key: "audioTrack", get: function get() { return this._trackId; } /** * Select current track by index */ , set: function set(newId) { this._setAudioTrack(newId); // If audio track is selected from API then don't choose from the manifest default track this._selectDefaultTrack = false; } }]); return AudioTrackController; }(TaskLoop); /* harmony default export */ var audio_track_controller = (audio_track_controller_AudioTrackController); // CONCATENATED MODULE: ./src/controller/audio-stream-controller.js function audio_stream_controller_defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } function audio_stream_controller_createClass(Constructor, protoProps, staticProps) { if (protoProps) audio_stream_controller_defineProperties(Constructor.prototype, protoProps); if (staticProps) audio_stream_controller_defineProperties(Constructor, staticProps); return Constructor; } function audio_stream_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /* * Audio Stream Controller */ var audio_stream_controller_window = window, audio_stream_controller_performance = audio_stream_controller_window.performance; var audio_stream_controller_TICK_INTERVAL = 100; // how often to tick in ms var audio_stream_controller_AudioStreamController = /*#__PURE__*/function (_BaseStreamController) { audio_stream_controller_inheritsLoose(AudioStreamController, _BaseStreamController); function AudioStreamController(hls, fragmentTracker) { var _this; _this = _BaseStreamController.call(this, hls, events["default"].MEDIA_ATTACHED, events["default"].MEDIA_DETACHING, events["default"].AUDIO_TRACKS_UPDATED, events["default"].AUDIO_TRACK_SWITCHING, events["default"].AUDIO_TRACK_LOADED, events["default"].KEY_LOADED, events["default"].FRAG_LOADED, events["default"].FRAG_PARSING_INIT_SEGMENT, events["default"].FRAG_PARSING_DATA, events["default"].FRAG_PARSED, events["default"].ERROR, events["default"].BUFFER_RESET, events["default"].BUFFER_CREATED, events["default"].BUFFER_APPENDED, events["default"].BUFFER_FLUSHED, events["default"].INIT_PTS_FOUND) || this; _this.fragmentTracker = fragmentTracker; _this.config = hls.config; _this.audioCodecSwap = false; _this._state = State.STOPPED; _this.initPTS = []; _this.waitingFragment = null; _this.videoTrackCC = null; return _this; } // Signal that video PTS was found var _proto = AudioStreamController.prototype; _proto.onInitPtsFound = function onInitPtsFound(data) { var demuxerId = data.id, cc = data.frag.cc, initPTS = data.initPTS; if (demuxerId === 'main') { // Always update the new INIT PTS // Can change due level switch this.initPTS[cc] = initPTS; this.videoTrackCC = cc; logger["logger"].log("InitPTS for cc: " + cc + " found from video track: " + initPTS); // If we are waiting we need to demux/remux the waiting frag // With the new initPTS if (this.state === State.WAITING_INIT_PTS) { this.tick(); } } }; _proto.startLoad = function startLoad(startPosition) { if (this.tracks) { var lastCurrentTime = this.lastCurrentTime; this.stopLoad(); this.setInterval(audio_stream_controller_TICK_INTERVAL); this.fragLoadError = 0; if (lastCurrentTime > 0 && startPosition === -1) { logger["logger"].log("audio:override startPosition with lastCurrentTime @" + lastCurrentTime.toFixed(3)); this.state = State.IDLE; } else { this.lastCurrentTime = this.startPosition ? this.startPosition : startPosition; this.state = State.STARTING; } this.nextLoadPosition = this.startPosition = this.lastCurrentTime; this.tick(); } else { this.startPosition = startPosition; this.state = State.STOPPED; } }; _proto.doTick = function doTick() { var pos, track, trackDetails, hls = this.hls, config = hls.config; // logger.log('audioStream:' + this.state); switch (this.state) { case State.ERROR: // don't do anything in error state to avoid breaking further ... case State.PAUSED: // don't do anything in paused state either ... case State.BUFFER_FLUSHING: break; case State.STARTING: this.state = State.WAITING_TRACK; this.loadedmetadata = false; break; case State.IDLE: var tracks = this.tracks; // audio tracks not received => exit loop if (!tracks) { break; } // if video not attached AND // start fragment already requested OR start frag prefetch disable // exit loop // => if media not attached but start frag prefetch is enabled and start frag not requested yet, we will not exit loop if (!this.media && (this.startFragRequested || !config.startFragPrefetch)) { break; } // determine next candidate fragment to be loaded, based on current position and // end of buffer position // if we have not yet loaded any fragment, start loading from start position if (this.loadedmetadata) { pos = this.media.currentTime; } else { pos = this.nextLoadPosition; if (pos === undefined) { break; } } var media = this.mediaBuffer ? this.mediaBuffer : this.media; var videoBuffer = this.videoBuffer ? this.videoBuffer : this.media; var maxBufferHole = pos < config.maxBufferHole ? Math.max(MAX_START_GAP_JUMP, config.maxBufferHole) : config.maxBufferHole; var bufferInfo = BufferHelper.bufferInfo(media, pos, maxBufferHole); var mainBufferInfo = BufferHelper.bufferInfo(videoBuffer, pos, maxBufferHole); var bufferLen = bufferInfo.len; var bufferEnd = bufferInfo.end; var fragPrevious = this.fragPrevious; // ensure we buffer at least config.maxBufferLength (default 30s) or config.maxMaxBufferLength (default: 600s) // whichever is smaller. // once we reach that threshold, don't buffer more than video (mainBufferInfo.len) var maxConfigBuffer = Math.min(config.maxBufferLength, config.maxMaxBufferLength); var maxBufLen = Math.max(maxConfigBuffer, mainBufferInfo.len); var audioSwitch = this.audioSwitch; var trackId = this.trackId; // if buffer length is less than maxBufLen try to load a new fragment if ((bufferLen < maxBufLen || audioSwitch) && trackId < tracks.length) { trackDetails = tracks[trackId].details; // if track info not retrieved yet, switch state and wait for track retrieval if (typeof trackDetails === 'undefined') { this.state = State.WAITING_TRACK; break; } if (!audioSwitch && this._streamEnded(bufferInfo, trackDetails)) { this.hls.trigger(events["default"].BUFFER_EOS, { type: 'audio' }); this.state = State.ENDED; return; } // find fragment index, contiguous with end of buffer position var fragments = trackDetails.fragments, fragLen = fragments.length, start = fragments[0].start, end = fragments[fragLen - 1].start + fragments[fragLen - 1].duration, frag; // When switching audio track, reload audio as close as possible to currentTime if (audioSwitch) { if (trackDetails.live && !trackDetails.PTSKnown) { logger["logger"].log('switching audiotrack, live stream, unknown PTS,load first fragment'); bufferEnd = 0; } else { bufferEnd = pos; // if currentTime (pos) is less than alt audio playlist start time, it means that alt audio is ahead of currentTime if (trackDetails.PTSKnown && pos < start) { // if everything is buffered from pos to start or if audio buffer upfront, let's seek to start if (bufferInfo.end > start || bufferInfo.nextStart) { logger["logger"].log('alt audio track ahead of main track, seek to start of alt audio track'); this.media.currentTime = start + 0.05; } else { return; } } } } if (trackDetails.initSegment && !trackDetails.initSegment.data) { frag = trackDetails.initSegment; } // eslint-disable-line brace-style // if bufferEnd before start of playlist, load first fragment else if (bufferEnd <= start) { frag = fragments[0]; if (this.videoTrackCC !== null && frag.cc !== this.videoTrackCC) { // Ensure we find a fragment which matches the continuity of the video track frag = findFragWithCC(fragments, this.videoTrackCC); } if (trackDetails.live && frag.loadIdx && frag.loadIdx === this.fragLoadIdx) { // we just loaded this first fragment, and we are still lagging behind the start of the live playlist // let's force seek to start var nextBuffered = bufferInfo.nextStart ? bufferInfo.nextStart : start; logger["logger"].log("no alt audio available @currentTime:" + this.media.currentTime + ", seeking @" + (nextBuffered + 0.05)); this.media.currentTime = nextBuffered + 0.05; return; } } else { var foundFrag; var maxFragLookUpTolerance = config.maxFragLookUpTolerance; var fragNext = fragPrevious ? fragments[fragPrevious.sn - fragments[0].sn + 1] : undefined; var fragmentWithinToleranceTest = function fragmentWithinToleranceTest(candidate) { // offset should be within fragment boundary - config.maxFragLookUpTolerance // this is to cope with situations like // bufferEnd = 9.991 // frag[Ø] : [0,10] // frag[1] : [10,20] // bufferEnd is within frag[0] range ... although what we are expecting is to return frag[1] here // frag start frag start+duration // |-----------------------------| // <---> <---> // ...--------><-----------------------------><---------.... // previous frag matching fragment next frag // return -1 return 0 return 1 // logger.log(`level/sn/start/end/bufEnd:${level}/${candidate.sn}/${candidate.start}/${(candidate.start+candidate.duration)}/${bufferEnd}`); // Set the lookup tolerance to be small enough to detect the current segment - ensures we don't skip over very small segments var candidateLookupTolerance = Math.min(maxFragLookUpTolerance, candidate.duration); if (candidate.start + candidate.duration - candidateLookupTolerance <= bufferEnd) { return 1; } else if (candidate.start - candidateLookupTolerance > bufferEnd && candidate.start) { // if maxFragLookUpTolerance will have negative value then don't return -1 for first element return -1; } return 0; }; if (bufferEnd < end) { if (bufferEnd > end - maxFragLookUpTolerance) { maxFragLookUpTolerance = 0; } // Prefer the next fragment if it's within tolerance if (fragNext && !fragmentWithinToleranceTest(fragNext)) { foundFrag = fragNext; } else { foundFrag = binary_search.search(fragments, fragmentWithinToleranceTest); } } else { // reach end of playlist foundFrag = fragments[fragLen - 1]; } if (foundFrag) { frag = foundFrag; start = foundFrag.start; // logger.log('find SN matching with pos:' + bufferEnd + ':' + frag.sn); if (fragPrevious && frag.level === fragPrevious.level && frag.sn === fragPrevious.sn) { if (frag.sn < trackDetails.endSN) { frag = fragments[frag.sn + 1 - trackDetails.startSN]; logger["logger"].log("SN just loaded, load next one: " + frag.sn); } else { frag = null; } } } } if (frag) { // logger.log(' loading frag ' + i +',pos/bufEnd:' + pos.toFixed(3) + '/' + bufferEnd.toFixed(3)); if (frag.encrypted) { logger["logger"].log("Loading key for " + frag.sn + " of [" + trackDetails.startSN + " ," + trackDetails.endSN + "],track " + trackId); this.state = State.KEY_LOADING; hls.trigger(events["default"].KEY_LOADING, { frag: frag }); } else { // only load if fragment is not loaded or if in audio switch // we force a frag loading in audio switch as fragment tracker might not have evicted previous frags in case of quick audio switch this.fragCurrent = frag; if (audioSwitch || this.fragmentTracker.getState(frag) === FragmentState.NOT_LOADED) { logger["logger"].log("Loading " + frag.sn + ", cc: " + frag.cc + " of [" + trackDetails.startSN + " ," + trackDetails.endSN + "],track " + trackId + ", currentTime:" + pos + ",bufferEnd:" + bufferEnd.toFixed(3)); if (frag.sn !== 'initSegment') { this.startFragRequested = true; } if (Object(number["isFiniteNumber"])(frag.sn)) { this.nextLoadPosition = frag.start + frag.duration; } hls.trigger(events["default"].FRAG_LOADING, { frag: frag }); this.state = State.FRAG_LOADING; } } } } break; case State.WAITING_TRACK: track = this.tracks[this.trackId]; // check if playlist is already loaded if (track && track.details) { this.state = State.IDLE; } break; case State.FRAG_LOADING_WAITING_RETRY: var now = audio_stream_controller_performance.now(); var retryDate = this.retryDate; media = this.media; var isSeeking = media && media.seeking; // if current time is gt than retryDate, or if media seeking let's switch to IDLE state to retry loading if (!retryDate || now >= retryDate || isSeeking) { logger["logger"].log('audioStreamController: retryDate reached, switch back to IDLE state'); this.state = State.IDLE; } break; case State.WAITING_INIT_PTS: var videoTrackCC = this.videoTrackCC; if (this.initPTS[videoTrackCC] === undefined) { break; } // Ensure we don't get stuck in the WAITING_INIT_PTS state if the waiting frag CC doesn't match any initPTS var waitingFrag = this.waitingFragment; if (waitingFrag) { var waitingFragCC = waitingFrag.frag.cc; if (videoTrackCC !== waitingFragCC) { track = this.tracks[this.trackId]; if (track.details && track.details.live) { logger["logger"].warn("Waiting fragment CC (" + waitingFragCC + ") does not match video track CC (" + videoTrackCC + ")"); this.waitingFragment = null; this.state = State.IDLE; } } else { this.state = State.FRAG_LOADING; this.onFragLoaded(this.waitingFragment); this.waitingFragment = null; } } else { this.state = State.IDLE; } break; case State.STOPPED: case State.FRAG_LOADING: case State.PARSING: case State.PARSED: case State.ENDED: break; default: break; } }; _proto.onMediaAttached = function onMediaAttached(data) { var media = this.media = this.mediaBuffer = data.media; this.onvseeking = this.onMediaSeeking.bind(this); this.onvended = this.onMediaEnded.bind(this); media.addEventListener('seeking', this.onvseeking); media.addEventListener('ended', this.onvended); var config = this.config; if (this.tracks && config.autoStartLoad) { this.startLoad(config.startPosition); } }; _proto.onMediaDetaching = function onMediaDetaching() { var media = this.media; if (media && media.ended) { logger["logger"].log('MSE detaching and video ended, reset startPosition'); this.startPosition = this.lastCurrentTime = 0; } // remove video listeners if (media) { media.removeEventListener('seeking', this.onvseeking); media.removeEventListener('ended', this.onvended); this.onvseeking = this.onvseeked = this.onvended = null; } this.media = this.mediaBuffer = this.videoBuffer = null; this.loadedmetadata = false; this.fragmentTracker.removeAllFragments(); this.stopLoad(); }; _proto.onAudioTracksUpdated = function onAudioTracksUpdated(data) { logger["logger"].log('audio tracks updated'); this.tracks = data.audioTracks; }; _proto.onAudioTrackSwitching = function onAudioTrackSwitching(data) { // if any URL found on new audio track, it is an alternate audio track var altAudio = !!data.url; this.trackId = data.id; this.fragCurrent = null; this.state = State.PAUSED; this.waitingFragment = null; // destroy useless demuxer when switching audio to main if (!altAudio) { if (this.demuxer) { this.demuxer.destroy(); this.demuxer = null; } } else { // switching to audio track, start timer if not already started this.setInterval(audio_stream_controller_TICK_INTERVAL); } // should we switch tracks ? if (altAudio) { this.audioSwitch = true; // main audio track are handled by stream-controller, just do something if switching to alt audio track this.state = State.IDLE; } this.tick(); }; _proto.onAudioTrackLoaded = function onAudioTrackLoaded(data) { var newDetails = data.details, trackId = data.id, track = this.tracks[trackId], duration = newDetails.totalduration, sliding = 0; logger["logger"].log("track " + trackId + " loaded [" + newDetails.startSN + "," + newDetails.endSN + "],duration:" + duration); if (newDetails.live) { var curDetails = track.details; if (curDetails && newDetails.fragments.length > 0) { // we already have details for that level, merge them mergeDetails(curDetails, newDetails); sliding = newDetails.fragments[0].start; // TODO // this.liveSyncPosition = this.computeLivePosition(sliding, curDetails); if (newDetails.PTSKnown) { logger["logger"].log("live audio playlist sliding:" + sliding.toFixed(3)); } else { logger["logger"].log('live audio playlist - outdated PTS, unknown sliding'); } } else { newDetails.PTSKnown = false; logger["logger"].log('live audio playlist - first load, unknown sliding'); } } else { newDetails.PTSKnown = false; } track.details = newDetails; // compute start position if (!this.startFragRequested) { // compute start position if set to -1. use it straight away if value is defined if (this.startPosition === -1) { // first, check if start time offset has been set in playlist, if yes, use this value var startTimeOffset = newDetails.startTimeOffset; if (Object(number["isFiniteNumber"])(startTimeOffset)) { logger["logger"].log("start time offset found in playlist, adjust startPosition to " + startTimeOffset); this.startPosition = startTimeOffset; } else { if (newDetails.live) { this.startPosition = this.computeLivePosition(sliding, newDetails); logger["logger"].log("compute startPosition for audio-track to " + this.startPosition); } else { this.startPosition = 0; } } } this.nextLoadPosition = this.startPosition; } // only switch batck to IDLE state if we were waiting for track to start downloading a new fragment if (this.state === State.WAITING_TRACK) { this.state = State.IDLE; } // trigger handler right now this.tick(); }; _proto.onKeyLoaded = function onKeyLoaded() { if (this.state === State.KEY_LOADING) { this.state = State.IDLE; this.tick(); } }; _proto.onFragLoaded = function onFragLoaded(data) { var fragCurrent = this.fragCurrent, fragLoaded = data.frag; if (this.state === State.FRAG_LOADING && fragCurrent && fragLoaded.type === 'audio' && fragLoaded.level === fragCurrent.level && fragLoaded.sn === fragCurrent.sn) { var track = this.tracks[this.trackId], details = track.details, duration = details.totalduration, trackId = fragCurrent.level, sn = fragCurrent.sn, cc = fragCurrent.cc, audioCodec = this.config.defaultAudioCodec || track.audioCodec || 'mp4a.40.2', stats = this.stats = data.stats; if (sn === 'initSegment') { this.state = State.IDLE; stats.tparsed = stats.tbuffered = audio_stream_controller_performance.now(); details.initSegment.data = data.payload; this.hls.trigger(events["default"].FRAG_BUFFERED, { stats: stats, frag: fragCurrent, id: 'audio' }); this.tick(); } else { this.state = State.PARSING; // transmux the MPEG-TS data to ISO-BMFF segments this.appended = false; if (!this.demuxer) { this.demuxer = new demux_demuxer(this.hls, 'audio'); } // Check if we have video initPTS // If not we need to wait for it var initPTS = this.initPTS[cc]; var initSegmentData = details.initSegment ? details.initSegment.data : []; if (details.initSegment || initPTS !== undefined) { this.pendingBuffering = true; logger["logger"].log("Demuxing " + sn + " of [" + details.startSN + " ," + details.endSN + "],track " + trackId); // time Offset is accurate if level PTS is known, or if playlist is not sliding (not live) var accurateTimeOffset = false; // details.PTSKnown || !details.live; this.demuxer.push(data.payload, initSegmentData, audioCodec, null, fragCurrent, duration, accurateTimeOffset, initPTS); } else { logger["logger"].log("unknown video PTS for continuity counter " + cc + ", waiting for video PTS before demuxing audio frag " + sn + " of [" + details.startSN + " ," + details.endSN + "],track " + trackId); this.waitingFragment = data; this.state = State.WAITING_INIT_PTS; } } } this.fragLoadError = 0; }; _proto.onFragParsingInitSegment = function onFragParsingInitSegment(data) { var fragCurrent = this.fragCurrent; var fragNew = data.frag; if (fragCurrent && data.id === 'audio' && fragNew.sn === fragCurrent.sn && fragNew.level === fragCurrent.level && this.state === State.PARSING) { var tracks = data.tracks, track; // delete any video track found on audio demuxer if (tracks.video) { delete tracks.video; } // include levelCodec in audio and video tracks track = tracks.audio; if (track) { track.levelCodec = track.codec; track.id = data.id; this.hls.trigger(events["default"].BUFFER_CODECS, tracks); logger["logger"].log("audio track:audio,container:" + track.container + ",codecs[level/parsed]=[" + track.levelCodec + "/" + track.codec + "]"); var initSegment = track.initSegment; if (initSegment) { var appendObj = { type: 'audio', data: initSegment, parent: 'audio', content: 'initSegment' }; if (this.audioSwitch) { this.pendingData = [appendObj]; } else { this.appended = true; // arm pending Buffering flag before appending a segment this.pendingBuffering = true; this.hls.trigger(events["default"].BUFFER_APPENDING, appendObj); } } // trigger handler right now this.tick(); } } }; _proto.onFragParsingData = function onFragParsingData(data) { var _this2 = this; var fragCurrent = this.fragCurrent; var fragNew = data.frag; if (fragCurrent && data.id === 'audio' && data.type === 'audio' && fragNew.sn === fragCurrent.sn && fragNew.level === fragCurrent.level && this.state === State.PARSING) { var trackId = this.trackId, track = this.tracks[trackId], hls = this.hls; if (!Object(number["isFiniteNumber"])(data.endPTS)) { data.endPTS = data.startPTS + fragCurrent.duration; data.endDTS = data.startDTS + fragCurrent.duration; } fragCurrent.addElementaryStream(ElementaryStreamTypes.AUDIO); logger["logger"].log("parsed " + data.type + ",PTS:[" + data.startPTS.toFixed(3) + "," + data.endPTS.toFixed(3) + "],DTS:[" + data.startDTS.toFixed(3) + "/" + data.endDTS.toFixed(3) + "],nb:" + data.nb); updateFragPTSDTS(track.details, fragCurrent, data.startPTS, data.endPTS); var audioSwitch = this.audioSwitch, media = this.media, appendOnBufferFlush = false; // Only flush audio from old audio tracks when PTS is known on new audio track if (audioSwitch) { if (media && media.readyState) { var currentTime = media.currentTime; logger["logger"].log('switching audio track : currentTime:' + currentTime); if (currentTime >= data.startPTS) { logger["logger"].log('switching audio track : flushing all audio'); this.state = State.BUFFER_FLUSHING; hls.trigger(events["default"].BUFFER_FLUSHING, { startOffset: 0, endOffset: Number.POSITIVE_INFINITY, type: 'audio' }); appendOnBufferFlush = true; // Lets announce that the initial audio track switch flush occur this.audioSwitch = false; hls.trigger(events["default"].AUDIO_TRACK_SWITCHED, { id: trackId }); } } else { // Lets announce that the initial audio track switch flush occur this.audioSwitch = false; hls.trigger(events["default"].AUDIO_TRACK_SWITCHED, { id: trackId }); } } var pendingData = this.pendingData; if (!pendingData) { logger["logger"].warn('Apparently attempt to enqueue media payload without codec initialization data upfront'); hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].MEDIA_ERROR, details: null, fatal: true }); return; } if (!this.audioSwitch) { [data.data1, data.data2].forEach(function (buffer) { if (buffer && buffer.length) { pendingData.push({ type: data.type, data: buffer, parent: 'audio', content: 'data' }); } }); if (!appendOnBufferFlush && pendingData.length) { pendingData.forEach(function (appendObj) { // only append in PARSING state (rationale is that an appending error could happen synchronously on first segment appending) // in that case it is useless to append following segments if (_this2.state === State.PARSING) { // arm pending Buffering flag before appending a segment _this2.pendingBuffering = true; _this2.hls.trigger(events["default"].BUFFER_APPENDING, appendObj); } }); this.pendingData = []; this.appended = true; } } // trigger handler right now this.tick(); } }; _proto.onFragParsed = function onFragParsed(data) { var fragCurrent = this.fragCurrent; var fragNew = data.frag; if (fragCurrent && data.id === 'audio' && fragNew.sn === fragCurrent.sn && fragNew.level === fragCurrent.level && this.state === State.PARSING) { this.stats.tparsed = audio_stream_controller_performance.now(); this.state = State.PARSED; this._checkAppendedParsed(); } }; _proto.onBufferReset = function onBufferReset() { // reset reference to sourcebuffers this.mediaBuffer = this.videoBuffer = null; this.loadedmetadata = false; }; _proto.onBufferCreated = function onBufferCreated(data) { var audioTrack = data.tracks.audio; if (audioTrack) { this.mediaBuffer = audioTrack.buffer; this.loadedmetadata = true; } if (data.tracks.video) { this.videoBuffer = data.tracks.video.buffer; } }; _proto.onBufferAppended = function onBufferAppended(data) { if (data.parent === 'audio') { var state = this.state; if (state === State.PARSING || state === State.PARSED) { // check if all buffers have been appended this.pendingBuffering = data.pending > 0; this._checkAppendedParsed(); } } }; _proto._checkAppendedParsed = function _checkAppendedParsed() { // trigger handler right now if (this.state === State.PARSED && (!this.appended || !this.pendingBuffering)) { var frag = this.fragCurrent, stats = this.stats, hls = this.hls; if (frag) { this.fragPrevious = frag; stats.tbuffered = audio_stream_controller_performance.now(); hls.trigger(events["default"].FRAG_BUFFERED, { stats: stats, frag: frag, id: 'audio' }); var media = this.mediaBuffer ? this.mediaBuffer : this.media; if (media) { logger["logger"].log("audio buffered : " + time_ranges.toString(media.buffered)); } if (this.audioSwitch && this.appended) { this.audioSwitch = false; hls.trigger(events["default"].AUDIO_TRACK_SWITCHED, { id: this.trackId }); } this.state = State.IDLE; } this.tick(); } }; _proto.onError = function onError(data) { var frag = data.frag; // don't handle frag error not related to audio fragment if (frag && frag.type !== 'audio') { return; } switch (data.details) { case errors["ErrorDetails"].FRAG_LOAD_ERROR: case errors["ErrorDetails"].FRAG_LOAD_TIMEOUT: var _frag = data.frag; // don't handle frag error not related to audio fragment if (_frag && _frag.type !== 'audio') { break; } if (!data.fatal) { var loadError = this.fragLoadError; if (loadError) { loadError++; } else { loadError = 1; } var config = this.config; if (loadError <= config.fragLoadingMaxRetry) { this.fragLoadError = loadError; // exponential backoff capped to config.fragLoadingMaxRetryTimeout var delay = Math.min(Math.pow(2, loadError - 1) * config.fragLoadingRetryDelay, config.fragLoadingMaxRetryTimeout); logger["logger"].warn("AudioStreamController: frag loading failed, retry in " + delay + " ms"); this.retryDate = audio_stream_controller_performance.now() + delay; // retry loading state this.state = State.FRAG_LOADING_WAITING_RETRY; } else { logger["logger"].error("AudioStreamController: " + data.details + " reaches max retry, redispatch as fatal ..."); // switch error to fatal data.fatal = true; this.state = State.ERROR; } } break; case errors["ErrorDetails"].AUDIO_TRACK_LOAD_ERROR: case errors["ErrorDetails"].AUDIO_TRACK_LOAD_TIMEOUT: case errors["ErrorDetails"].KEY_LOAD_ERROR: case errors["ErrorDetails"].KEY_LOAD_TIMEOUT: // when in ERROR state, don't switch back to IDLE state in case a non-fatal error is received if (this.state !== State.ERROR) { // if fatal error, stop processing, otherwise move to IDLE to retry loading this.state = data.fatal ? State.ERROR : State.IDLE; logger["logger"].warn("AudioStreamController: " + data.details + " while loading frag, now switching to " + this.state + " state ..."); } break; case errors["ErrorDetails"].BUFFER_FULL_ERROR: // if in appending state if (data.parent === 'audio' && (this.state === State.PARSING || this.state === State.PARSED)) { var media = this.mediaBuffer, currentTime = this.media.currentTime, mediaBuffered = media && BufferHelper.isBuffered(media, currentTime) && BufferHelper.isBuffered(media, currentTime + 0.5); // reduce max buf len if current position is buffered if (mediaBuffered) { var _config = this.config; if (_config.maxMaxBufferLength >= _config.maxBufferLength) { // reduce max buffer length as it might be too high. we do this to avoid loop flushing ... _config.maxMaxBufferLength /= 2; logger["logger"].warn("AudioStreamController: reduce max buffer length to " + _config.maxMaxBufferLength + "s"); } this.state = State.IDLE; } else { // current position is not buffered, but browser is still complaining about buffer full error // this happens on IE/Edge, refer to https://github.com/video-dev/hls.js/pull/708 // in that case flush the whole audio buffer to recover logger["logger"].warn('AudioStreamController: buffer full error also media.currentTime is not buffered, flush audio buffer'); this.fragCurrent = null; // flush everything this.state = State.BUFFER_FLUSHING; this.hls.trigger(events["default"].BUFFER_FLUSHING, { startOffset: 0, endOffset: Number.POSITIVE_INFINITY, type: 'audio' }); } } break; default: break; } }; _proto.onBufferFlushed = function onBufferFlushed() { var _this3 = this; var pendingData = this.pendingData; if (pendingData && pendingData.length) { logger["logger"].log('AudioStreamController: appending pending audio data after buffer flushed'); pendingData.forEach(function (appendObj) { _this3.hls.trigger(events["default"].BUFFER_APPENDING, appendObj); }); this.appended = true; this.pendingData = []; this.state = State.PARSED; } else { // move to IDLE once flush complete. this should trigger new fragment loading this.state = State.IDLE; // reset reference to frag this.fragPrevious = null; this.tick(); } }; audio_stream_controller_createClass(AudioStreamController, [{ key: "state", set: function set(nextState) { if (this.state !== nextState) { var previousState = this.state; this._state = nextState; logger["logger"].log("audio stream:" + previousState + "->" + nextState); } }, get: function get() { return this._state; } }]); return AudioStreamController; }(base_stream_controller_BaseStreamController); /* harmony default export */ var audio_stream_controller = (audio_stream_controller_AudioStreamController); // CONCATENATED MODULE: ./src/utils/vttcue.js /** * Copyright 2013 vtt.js Contributors * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /* harmony default export */ var vttcue = ((function () { if (typeof window !== 'undefined' && window.VTTCue) { return window.VTTCue; } var autoKeyword = 'auto'; var directionSetting = { '': true, lr: true, rl: true }; var alignSetting = { start: true, middle: true, end: true, left: true, right: true }; function findDirectionSetting(value) { if (typeof value !== 'string') { return false; } var dir = directionSetting[value.toLowerCase()]; return dir ? value.toLowerCase() : false; } function findAlignSetting(value) { if (typeof value !== 'string') { return false; } var align = alignSetting[value.toLowerCase()]; return align ? value.toLowerCase() : false; } function extend(obj) { var i = 1; for (; i < arguments.length; i++) { var cobj = arguments[i]; for (var p in cobj) { obj[p] = cobj[p]; } } return obj; } function VTTCue(startTime, endTime, text) { var cue = this; var baseObj = {}; baseObj.enumerable = true; /** * Shim implementation specific properties. These properties are not in * the spec. */ // Lets us know when the VTTCue's data has changed in such a way that we need // to recompute its display state. This lets us compute its display state // lazily. cue.hasBeenReset = false; /** * VTTCue and TextTrackCue properties * http://dev.w3.org/html5/webvtt/#vttcue-interface */ var _id = ''; var _pauseOnExit = false; var _startTime = startTime; var _endTime = endTime; var _text = text; var _region = null; var _vertical = ''; var _snapToLines = true; var _line = 'auto'; var _lineAlign = 'start'; var _position = 50; var _positionAlign = 'middle'; var _size = 50; var _align = 'middle'; Object.defineProperty(cue, 'id', extend({}, baseObj, { get: function get() { return _id; }, set: function set(value) { _id = '' + value; } })); Object.defineProperty(cue, 'pauseOnExit', extend({}, baseObj, { get: function get() { return _pauseOnExit; }, set: function set(value) { _pauseOnExit = !!value; } })); Object.defineProperty(cue, 'startTime', extend({}, baseObj, { get: function get() { return _startTime; }, set: function set(value) { if (typeof value !== 'number') { throw new TypeError('Start time must be set to a number.'); } _startTime = value; this.hasBeenReset = true; } })); Object.defineProperty(cue, 'endTime', extend({}, baseObj, { get: function get() { return _endTime; }, set: function set(value) { if (typeof value !== 'number') { throw new TypeError('End time must be set to a number.'); } _endTime = value; this.hasBeenReset = true; } })); Object.defineProperty(cue, 'text', extend({}, baseObj, { get: function get() { return _text; }, set: function set(value) { _text = '' + value; this.hasBeenReset = true; } })); Object.defineProperty(cue, 'region', extend({}, baseObj, { get: function get() { return _region; }, set: function set(value) { _region = value; this.hasBeenReset = true; } })); Object.defineProperty(cue, 'vertical', extend({}, baseObj, { get: function get() { return _vertical; }, set: function set(value) { var setting = findDirectionSetting(value); // Have to check for false because the setting an be an empty string. if (setting === false) { throw new SyntaxError('An invalid or illegal string was specified.'); } _vertical = setting; this.hasBeenReset = true; } })); Object.defineProperty(cue, 'snapToLines', extend({}, baseObj, { get: function get() { return _snapToLines; }, set: function set(value) { _snapToLines = !!value; this.hasBeenReset = true; } })); Object.defineProperty(cue, 'line', extend({}, baseObj, { get: function get() { return _line; }, set: function set(value) { if (typeof value !== 'number' && value !== autoKeyword) { throw new SyntaxError('An invalid number or illegal string was specified.'); } _line = value; this.hasBeenReset = true; } })); Object.defineProperty(cue, 'lineAlign', extend({}, baseObj, { get: function get() { return _lineAlign; }, set: function set(value) { var setting = findAlignSetting(value); if (!setting) { throw new SyntaxError('An invalid or illegal string was specified.'); } _lineAlign = setting; this.hasBeenReset = true; } })); Object.defineProperty(cue, 'position', extend({}, baseObj, { get: function get() { return _position; }, set: function set(value) { if (value < 0 || value > 100) { throw new Error('Position must be between 0 and 100.'); } _position = value; this.hasBeenReset = true; } })); Object.defineProperty(cue, 'positionAlign', extend({}, baseObj, { get: function get() { return _positionAlign; }, set: function set(value) { var setting = findAlignSetting(value); if (!setting) { throw new SyntaxError('An invalid or illegal string was specified.'); } _positionAlign = setting; this.hasBeenReset = true; } })); Object.defineProperty(cue, 'size', extend({}, baseObj, { get: function get() { return _size; }, set: function set(value) { if (value < 0 || value > 100) { throw new Error('Size must be between 0 and 100.'); } _size = value; this.hasBeenReset = true; } })); Object.defineProperty(cue, 'align', extend({}, baseObj, { get: function get() { return _align; }, set: function set(value) { var setting = findAlignSetting(value); if (!setting) { throw new SyntaxError('An invalid or illegal string was specified.'); } _align = setting; this.hasBeenReset = true; } })); /** * Other <track> spec defined properties */ // http://www.whatwg.org/specs/web-apps/current-work/multipage/the-video-element.html#text-track-cue-display-state cue.displayState = void 0; } /** * VTTCue methods */ VTTCue.prototype.getCueAsHTML = function () { // Assume WebVTT.convertCueToDOMTree is on the global. var WebVTT = window.WebVTT; return WebVTT.convertCueToDOMTree(window, this.text); }; return VTTCue; })()); // CONCATENATED MODULE: ./src/utils/vttparser.js /* * Source: https://github.com/mozilla/vtt.js/blob/master/dist/vtt.js#L1716 */ var StringDecoder = function StringDecoder() { return { decode: function decode(data) { if (!data) { return ''; } if (typeof data !== 'string') { throw new Error('Error - expected string data.'); } return decodeURIComponent(encodeURIComponent(data)); } }; }; function VTTParser() { this.window = window; this.state = 'INITIAL'; this.buffer = ''; this.decoder = new StringDecoder(); this.regionList = []; } // Try to parse input as a time stamp. function parseTimeStamp(input) { function computeSeconds(h, m, s, f) { return (h | 0) * 3600 + (m | 0) * 60 + (s | 0) + (f | 0) / 1000; } var m = input.match(/^(\d+):(\d{2})(:\d{2})?\.(\d{3})/); if (!m) { return null; } if (m[3]) { // Timestamp takes the form of [hours]:[minutes]:[seconds].[milliseconds] return computeSeconds(m[1], m[2], m[3].replace(':', ''), m[4]); } else if (m[1] > 59) { // Timestamp takes the form of [hours]:[minutes].[milliseconds] // First position is hours as it's over 59. return computeSeconds(m[1], m[2], 0, m[4]); } else { // Timestamp takes the form of [minutes]:[seconds].[milliseconds] return computeSeconds(0, m[1], m[2], m[4]); } } // A settings object holds key/value pairs and will ignore anything but the first // assignment to a specific key. function Settings() { this.values = Object.create(null); } Settings.prototype = { // Only accept the first assignment to any key. set: function set(k, v) { if (!this.get(k) && v !== '') { this.values[k] = v; } }, // Return the value for a key, or a default value. // If 'defaultKey' is passed then 'dflt' is assumed to be an object with // a number of possible default values as properties where 'defaultKey' is // the key of the property that will be chosen; otherwise it's assumed to be // a single value. get: function get(k, dflt, defaultKey) { if (defaultKey) { return this.has(k) ? this.values[k] : dflt[defaultKey]; } return this.has(k) ? this.values[k] : dflt; }, // Check whether we have a value for a key. has: function has(k) { return k in this.values; }, // Accept a setting if its one of the given alternatives. alt: function alt(k, v, a) { for (var n = 0; n < a.length; ++n) { if (v === a[n]) { this.set(k, v); break; } } }, // Accept a setting if its a valid (signed) integer. integer: function integer(k, v) { if (/^-?\d+$/.test(v)) { // integer this.set(k, parseInt(v, 10)); } }, // Accept a setting if its a valid percentage. percent: function percent(k, v) { var m; if (m = v.match(/^([\d]{1,3})(\.[\d]*)?%$/)) { v = parseFloat(v); if (v >= 0 && v <= 100) { this.set(k, v); return true; } } return false; } }; // Helper function to parse input into groups separated by 'groupDelim', and // interprete each group as a key/value pair separated by 'keyValueDelim'. function parseOptions(input, callback, keyValueDelim, groupDelim) { var groups = groupDelim ? input.split(groupDelim) : [input]; for (var i in groups) { if (typeof groups[i] !== 'string') { continue; } var kv = groups[i].split(keyValueDelim); if (kv.length !== 2) { continue; } var k = kv[0]; var v = kv[1]; callback(k, v); } } var defaults = new vttcue(0, 0, 0); // 'middle' was changed to 'center' in the spec: https://github.com/w3c/webvtt/pull/244 // Safari doesn't yet support this change, but FF and Chrome do. var center = defaults.align === 'middle' ? 'middle' : 'center'; function parseCue(input, cue, regionList) { // Remember the original input if we need to throw an error. var oInput = input; // 4.1 WebVTT timestamp function consumeTimeStamp() { var ts = parseTimeStamp(input); if (ts === null) { throw new Error('Malformed timestamp: ' + oInput); } // Remove time stamp from input. input = input.replace(/^[^\sa-zA-Z-]+/, ''); return ts; } // 4.4.2 WebVTT cue settings function consumeCueSettings(input, cue) { var settings = new Settings(); parseOptions(input, function (k, v) { switch (k) { case 'region': // Find the last region we parsed with the same region id. for (var i = regionList.length - 1; i >= 0; i--) { if (regionList[i].id === v) { settings.set(k, regionList[i].region); break; } } break; case 'vertical': settings.alt(k, v, ['rl', 'lr']); break; case 'line': var vals = v.split(','), vals0 = vals[0]; settings.integer(k, vals0); if (settings.percent(k, vals0)) { settings.set('snapToLines', false); } settings.alt(k, vals0, ['auto']); if (vals.length === 2) { settings.alt('lineAlign', vals[1], ['start', center, 'end']); } break; case 'position': vals = v.split(','); settings.percent(k, vals[0]); if (vals.length === 2) { settings.alt('positionAlign', vals[1], ['start', center, 'end', 'line-left', 'line-right', 'auto']); } break; case 'size': settings.percent(k, v); break; case 'align': settings.alt(k, v, ['start', center, 'end', 'left', 'right']); break; } }, /:/, /\s/); // Apply default values for any missing fields. cue.region = settings.get('region', null); cue.vertical = settings.get('vertical', ''); var line = settings.get('line', 'auto'); if (line === 'auto' && defaults.line === -1) { // set numeric line number for Safari line = -1; } cue.line = line; cue.lineAlign = settings.get('lineAlign', 'start'); cue.snapToLines = settings.get('snapToLines', true); cue.size = settings.get('size', 100); cue.align = settings.get('align', center); var position = settings.get('position', 'auto'); if (position === 'auto' && defaults.position === 50) { // set numeric position for Safari position = cue.align === 'start' || cue.align === 'left' ? 0 : cue.align === 'end' || cue.align === 'right' ? 100 : 50; } cue.position = position; } function skipWhitespace() { input = input.replace(/^\s+/, ''); } // 4.1 WebVTT cue timings. skipWhitespace(); cue.startTime = consumeTimeStamp(); // (1) collect cue start time skipWhitespace(); if (input.substr(0, 3) !== '-->') { // (3) next characters must match '-->' throw new Error('Malformed time stamp (time stamps must be separated by \'-->\'): ' + oInput); } input = input.substr(3); skipWhitespace(); cue.endTime = consumeTimeStamp(); // (5) collect cue end time // 4.1 WebVTT cue settings list. skipWhitespace(); consumeCueSettings(input, cue); } function fixLineBreaks(input) { return input.replace(/<br(?: \/)?>/gi, '\n'); } VTTParser.prototype = { parse: function parse(data) { var self = this; // If there is no data then we won't decode it, but will just try to parse // whatever is in buffer already. This may occur in circumstances, for // example when flush() is called. if (data) { // Try to decode the data that we received. self.buffer += self.decoder.decode(data, { stream: true }); } function collectNextLine() { var buffer = self.buffer; var pos = 0; buffer = fixLineBreaks(buffer); while (pos < buffer.length && buffer[pos] !== '\r' && buffer[pos] !== '\n') { ++pos; } var line = buffer.substr(0, pos); // Advance the buffer early in case we fail below. if (buffer[pos] === '\r') { ++pos; } if (buffer[pos] === '\n') { ++pos; } self.buffer = buffer.substr(pos); return line; } // 3.2 WebVTT metadata header syntax function parseHeader(input) { parseOptions(input, function (k, v) { switch (k) { case 'Region': // 3.3 WebVTT region metadata header syntax // console.log('parse region', v); // parseRegion(v); break; } }, /:/); } // 5.1 WebVTT file parsing. try { var line; if (self.state === 'INITIAL') { // We can't start parsing until we have the first line. if (!/\r\n|\n/.test(self.buffer)) { return this; } line = collectNextLine(); // strip of UTF-8 BOM if any // https://en.wikipedia.org/wiki/Byte_order_mark#UTF-8 var m = line.match(/^()?WEBVTT([ \t].*)?$/); if (!m || !m[0]) { throw new Error('Malformed WebVTT signature.'); } self.state = 'HEADER'; } var alreadyCollectedLine = false; while (self.buffer) { // We can't parse a line until we have the full line. if (!/\r\n|\n/.test(self.buffer)) { return this; } if (!alreadyCollectedLine) { line = collectNextLine(); } else { alreadyCollectedLine = false; } switch (self.state) { case 'HEADER': // 13-18 - Allow a header (metadata) under the WEBVTT line. if (/:/.test(line)) { parseHeader(line); } else if (!line) { // An empty line terminates the header and starts the body (cues). self.state = 'ID'; } continue; case 'NOTE': // Ignore NOTE blocks. if (!line) { self.state = 'ID'; } continue; case 'ID': // Check for the start of NOTE blocks. if (/^NOTE($|[ \t])/.test(line)) { self.state = 'NOTE'; break; } // 19-29 - Allow any number of line terminators, then initialize new cue values. if (!line) { continue; } self.cue = new vttcue(0, 0, ''); self.state = 'CUE'; // 30-39 - Check if self line contains an optional identifier or timing data. if (line.indexOf('-->') === -1) { self.cue.id = line; continue; } // Process line as start of a cue. /* falls through */ case 'CUE': // 40 - Collect cue timings and settings. try { parseCue(line, self.cue, self.regionList); } catch (e) { // In case of an error ignore rest of the cue. self.cue = null; self.state = 'BADCUE'; continue; } self.state = 'CUETEXT'; continue; case 'CUETEXT': var hasSubstring = line.indexOf('-->') !== -1; // 34 - If we have an empty line then report the cue. // 35 - If we have the special substring '-->' then report the cue, // but do not collect the line as we need to process the current // one as a new cue. if (!line || hasSubstring && (alreadyCollectedLine = true)) { // We are done parsing self cue. if (self.oncue) { self.oncue(self.cue); } self.cue = null; self.state = 'ID'; continue; } if (self.cue.text) { self.cue.text += '\n'; } self.cue.text += line; continue; case 'BADCUE': // BADCUE // 54-62 - Collect and discard the remaining cue. if (!line) { self.state = 'ID'; } continue; } } } catch (e) { // If we are currently parsing a cue, report what we have. if (self.state === 'CUETEXT' && self.cue && self.oncue) { self.oncue(self.cue); } self.cue = null; // Enter BADWEBVTT state if header was not parsed correctly otherwise // another exception occurred so enter BADCUE state. self.state = self.state === 'INITIAL' ? 'BADWEBVTT' : 'BADCUE'; } return this; }, flush: function flush() { var self = this; try { // Finish decoding the stream. self.buffer += self.decoder.decode(); // Synthesize the end of the current cue or region. if (self.cue || self.state === 'HEADER') { self.buffer += '\n\n'; self.parse(); } // If we've flushed, parsed, and we're still on the INITIAL state then // that means we don't have enough of the stream to parse the first // line. if (self.state === 'INITIAL') { throw new Error('Malformed WebVTT signature.'); } } catch (e) { throw e; } if (self.onflush) { self.onflush(); } return this; } }; /* harmony default export */ var vttparser = (VTTParser); // CONCATENATED MODULE: ./src/utils/cues.ts function newCue(track, startTime, endTime, captionScreen) { var result = []; var row; // the type data states this is VTTCue, but it can potentially be a TextTrackCue on old browsers var cue; var indenting; var indent; var text; var VTTCue = window.VTTCue || TextTrackCue; for (var r = 0; r < captionScreen.rows.length; r++) { row = captionScreen.rows[r]; indenting = true; indent = 0; text = ''; if (!row.isEmpty()) { for (var c = 0; c < row.chars.length; c++) { if (row.chars[c].uchar.match(/\s/) && indenting) { indent++; } else { text += row.chars[c].uchar; indenting = false; } } // To be used for cleaning-up orphaned roll-up captions row.cueStartTime = startTime; // Give a slight bump to the endTime if it's equal to startTime to avoid a SyntaxError in IE if (startTime === endTime) { endTime += 0.0001; } cue = new VTTCue(startTime, endTime, fixLineBreaks(text.trim())); if (indent >= 16) { indent--; } else { indent++; } // VTTCue.line get's flakey when using controls, so let's now include line 13&14 // also, drop line 1 since it's to close to the top if (navigator.userAgent.match(/Firefox\//)) { cue.line = r + 1; } else { cue.line = r > 7 ? r - 2 : r + 1; } cue.align = 'left'; // Clamp the position between 0 and 100 - if out of these bounds, Firefox throws an exception and captions break cue.position = Math.max(0, Math.min(100, 100 * (indent / 32))); result.push(cue); if (track) { track.addCue(cue); } } } return result; } // CONCATENATED MODULE: ./src/utils/cea-608-parser.ts /** * * This code was ported from the dash.js project at: * https://github.com/Dash-Industry-Forum/dash.js/blob/development/externals/cea608-parser.js * https://github.com/Dash-Industry-Forum/dash.js/commit/8269b26a761e0853bb21d78780ed945144ecdd4d#diff-71bc295a2d6b6b7093a1d3290d53a4b2 * * The original copyright appears below: * * The copyright in this software is being made available under the BSD License, * included below. This software may be subject to other third party and contributor * rights, including patent rights, and no such rights are granted under this license. * * Copyright (c) 2015-2016, DASH Industry Forum. * All rights reserved. * * Redistribution and use in source and binary forms, with or without modification, * are permitted provided that the following conditions are met: * 1. Redistributions of source code must retain the above copyright notice, this * list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation and/or * other materials provided with the distribution. * 2. Neither the name of Dash Industry Forum nor the names of its * contributors may be used to endorse or promote products derived from this software * without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS AS IS AND ANY * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. */ /** * Exceptions from regular ASCII. CodePoints are mapped to UTF-16 codes */ var specialCea608CharsCodes = { 0x2a: 0xe1, // lowercase a, acute accent 0x5c: 0xe9, // lowercase e, acute accent 0x5e: 0xed, // lowercase i, acute accent 0x5f: 0xf3, // lowercase o, acute accent 0x60: 0xfa, // lowercase u, acute accent 0x7b: 0xe7, // lowercase c with cedilla 0x7c: 0xf7, // division symbol 0x7d: 0xd1, // uppercase N tilde 0x7e: 0xf1, // lowercase n tilde 0x7f: 0x2588, // Full block // THIS BLOCK INCLUDES THE 16 EXTENDED (TWO-BYTE) LINE 21 CHARACTERS // THAT COME FROM HI BYTE=0x11 AND LOW BETWEEN 0x30 AND 0x3F // THIS MEANS THAT \x50 MUST BE ADDED TO THE VALUES 0x80: 0xae, // Registered symbol (R) 0x81: 0xb0, // degree sign 0x82: 0xbd, // 1/2 symbol 0x83: 0xbf, // Inverted (open) question mark 0x84: 0x2122, // Trademark symbol (TM) 0x85: 0xa2, // Cents symbol 0x86: 0xa3, // Pounds sterling 0x87: 0x266a, // Music 8'th note 0x88: 0xe0, // lowercase a, grave accent 0x89: 0x20, // transparent space (regular) 0x8a: 0xe8, // lowercase e, grave accent 0x8b: 0xe2, // lowercase a, circumflex accent 0x8c: 0xea, // lowercase e, circumflex accent 0x8d: 0xee, // lowercase i, circumflex accent 0x8e: 0xf4, // lowercase o, circumflex accent 0x8f: 0xfb, // lowercase u, circumflex accent // THIS BLOCK INCLUDES THE 32 EXTENDED (TWO-BYTE) LINE 21 CHARACTERS // THAT COME FROM HI BYTE=0x12 AND LOW BETWEEN 0x20 AND 0x3F 0x90: 0xc1, // capital letter A with acute 0x91: 0xc9, // capital letter E with acute 0x92: 0xd3, // capital letter O with acute 0x93: 0xda, // capital letter U with acute 0x94: 0xdc, // capital letter U with diaresis 0x95: 0xfc, // lowercase letter U with diaeresis 0x96: 0x2018, // opening single quote 0x97: 0xa1, // inverted exclamation mark 0x98: 0x2a, // asterisk 0x99: 0x2019, // closing single quote 0x9a: 0x2501, // box drawings heavy horizontal 0x9b: 0xa9, // copyright sign 0x9c: 0x2120, // Service mark 0x9d: 0x2022, // (round) bullet 0x9e: 0x201c, // Left double quotation mark 0x9f: 0x201d, // Right double quotation mark 0xa0: 0xc0, // uppercase A, grave accent 0xa1: 0xc2, // uppercase A, circumflex 0xa2: 0xc7, // uppercase C with cedilla 0xa3: 0xc8, // uppercase E, grave accent 0xa4: 0xca, // uppercase E, circumflex 0xa5: 0xcb, // capital letter E with diaresis 0xa6: 0xeb, // lowercase letter e with diaresis 0xa7: 0xce, // uppercase I, circumflex 0xa8: 0xcf, // uppercase I, with diaresis 0xa9: 0xef, // lowercase i, with diaresis 0xaa: 0xd4, // uppercase O, circumflex 0xab: 0xd9, // uppercase U, grave accent 0xac: 0xf9, // lowercase u, grave accent 0xad: 0xdb, // uppercase U, circumflex 0xae: 0xab, // left-pointing double angle quotation mark 0xaf: 0xbb, // right-pointing double angle quotation mark // THIS BLOCK INCLUDES THE 32 EXTENDED (TWO-BYTE) LINE 21 CHARACTERS // THAT COME FROM HI BYTE=0x13 AND LOW BETWEEN 0x20 AND 0x3F 0xb0: 0xc3, // Uppercase A, tilde 0xb1: 0xe3, // Lowercase a, tilde 0xb2: 0xcd, // Uppercase I, acute accent 0xb3: 0xcc, // Uppercase I, grave accent 0xb4: 0xec, // Lowercase i, grave accent 0xb5: 0xd2, // Uppercase O, grave accent 0xb6: 0xf2, // Lowercase o, grave accent 0xb7: 0xd5, // Uppercase O, tilde 0xb8: 0xf5, // Lowercase o, tilde 0xb9: 0x7b, // Open curly brace 0xba: 0x7d, // Closing curly brace 0xbb: 0x5c, // Backslash 0xbc: 0x5e, // Caret 0xbd: 0x5f, // Underscore 0xbe: 0x7c, // Pipe (vertical line) 0xbf: 0x223c, // Tilde operator 0xc0: 0xc4, // Uppercase A, umlaut 0xc1: 0xe4, // Lowercase A, umlaut 0xc2: 0xd6, // Uppercase O, umlaut 0xc3: 0xf6, // Lowercase o, umlaut 0xc4: 0xdf, // Esszett (sharp S) 0xc5: 0xa5, // Yen symbol 0xc6: 0xa4, // Generic currency sign 0xc7: 0x2503, // Box drawings heavy vertical 0xc8: 0xc5, // Uppercase A, ring 0xc9: 0xe5, // Lowercase A, ring 0xca: 0xd8, // Uppercase O, stroke 0xcb: 0xf8, // Lowercase o, strok 0xcc: 0x250f, // Box drawings heavy down and right 0xcd: 0x2513, // Box drawings heavy down and left 0xce: 0x2517, // Box drawings heavy up and right 0xcf: 0x251b // Box drawings heavy up and left }; /** * Utils */ var getCharForByte = function getCharForByte(_byte) { var charCode = _byte; if (specialCea608CharsCodes.hasOwnProperty(_byte)) { charCode = specialCea608CharsCodes[_byte]; } return String.fromCharCode(charCode); }; var NR_ROWS = 15; var NR_COLS = 100; // Tables to look up row from PAC data var rowsLowCh1 = { 0x11: 1, 0x12: 3, 0x15: 5, 0x16: 7, 0x17: 9, 0x10: 11, 0x13: 12, 0x14: 14 }; var rowsHighCh1 = { 0x11: 2, 0x12: 4, 0x15: 6, 0x16: 8, 0x17: 10, 0x13: 13, 0x14: 15 }; var rowsLowCh2 = { 0x19: 1, 0x1A: 3, 0x1D: 5, 0x1E: 7, 0x1F: 9, 0x18: 11, 0x1B: 12, 0x1C: 14 }; var rowsHighCh2 = { 0x19: 2, 0x1A: 4, 0x1D: 6, 0x1E: 8, 0x1F: 10, 0x1B: 13, 0x1C: 15 }; var backgroundColors = ['white', 'green', 'blue', 'cyan', 'red', 'yellow', 'magenta', 'black', 'transparent']; var VerboseLevel; (function (VerboseLevel) { VerboseLevel[VerboseLevel["ERROR"] = 0] = "ERROR"; VerboseLevel[VerboseLevel["TEXT"] = 1] = "TEXT"; VerboseLevel[VerboseLevel["WARNING"] = 2] = "WARNING"; VerboseLevel[VerboseLevel["INFO"] = 2] = "INFO"; VerboseLevel[VerboseLevel["DEBUG"] = 3] = "DEBUG"; VerboseLevel[VerboseLevel["DATA"] = 3] = "DATA"; })(VerboseLevel || (VerboseLevel = {})); var cea_608_parser_CaptionsLogger = /*#__PURE__*/function () { function CaptionsLogger() { this.time = null; this.verboseLevel = VerboseLevel.ERROR; } var _proto = CaptionsLogger.prototype; _proto.log = function log(severity, msg) { if (this.verboseLevel >= severity) { logger["logger"].log(this.time + " [" + severity + "] " + msg); } }; return CaptionsLogger; }(); var numArrayToHexArray = function numArrayToHexArray(numArray) { var hexArray = []; for (var j = 0; j < numArray.length; j++) { hexArray.push(numArray[j].toString(16)); } return hexArray; }; var PenState = /*#__PURE__*/function () { function PenState(foreground, underline, italics, background, flash) { this.foreground = void 0; this.underline = void 0; this.italics = void 0; this.background = void 0; this.flash = void 0; this.foreground = foreground || 'white'; this.underline = underline || false; this.italics = italics || false; this.background = background || 'black'; this.flash = flash || false; } var _proto2 = PenState.prototype; _proto2.reset = function reset() { this.foreground = 'white'; this.underline = false; this.italics = false; this.background = 'black'; this.flash = false; }; _proto2.setStyles = function setStyles(styles) { var attribs = ['foreground', 'underline', 'italics', 'background', 'flash']; for (var i = 0; i < attribs.length; i++) { var style = attribs[i]; if (styles.hasOwnProperty(style)) { this[style] = styles[style]; } } }; _proto2.isDefault = function isDefault() { return this.foreground === 'white' && !this.underline && !this.italics && this.background === 'black' && !this.flash; }; _proto2.equals = function equals(other) { return this.foreground === other.foreground && this.underline === other.underline && this.italics === other.italics && this.background === other.background && this.flash === other.flash; }; _proto2.copy = function copy(newPenState) { this.foreground = newPenState.foreground; this.underline = newPenState.underline; this.italics = newPenState.italics; this.background = newPenState.background; this.flash = newPenState.flash; }; _proto2.toString = function toString() { return 'color=' + this.foreground + ', underline=' + this.underline + ', italics=' + this.italics + ', background=' + this.background + ', flash=' + this.flash; }; return PenState; }(); /** * Unicode character with styling and background. * @constructor */ var StyledUnicodeChar = /*#__PURE__*/function () { function StyledUnicodeChar(uchar, foreground, underline, italics, background, flash) { this.uchar = void 0; this.penState = void 0; this.uchar = uchar || ' '; // unicode character this.penState = new PenState(foreground, underline, italics, background, flash); } var _proto3 = StyledUnicodeChar.prototype; _proto3.reset = function reset() { this.uchar = ' '; this.penState.reset(); }; _proto3.setChar = function setChar(uchar, newPenState) { this.uchar = uchar; this.penState.copy(newPenState); }; _proto3.setPenState = function setPenState(newPenState) { this.penState.copy(newPenState); }; _proto3.equals = function equals(other) { return this.uchar === other.uchar && this.penState.equals(other.penState); }; _proto3.copy = function copy(newChar) { this.uchar = newChar.uchar; this.penState.copy(newChar.penState); }; _proto3.isEmpty = function isEmpty() { return this.uchar === ' ' && this.penState.isDefault(); }; return StyledUnicodeChar; }(); /** * CEA-608 row consisting of NR_COLS instances of StyledUnicodeChar. * @constructor */ var Row = /*#__PURE__*/function () { function Row(logger) { this.chars = void 0; this.pos = void 0; this.currPenState = void 0; this.cueStartTime = void 0; this.logger = void 0; this.chars = []; for (var i = 0; i < NR_COLS; i++) { this.chars.push(new StyledUnicodeChar()); } this.logger = logger; this.pos = 0; this.currPenState = new PenState(); } var _proto4 = Row.prototype; _proto4.equals = function equals(other) { var equal = true; for (var i = 0; i < NR_COLS; i++) { if (!this.chars[i].equals(other.chars[i])) { equal = false; break; } } return equal; }; _proto4.copy = function copy(other) { for (var i = 0; i < NR_COLS; i++) { this.chars[i].copy(other.chars[i]); } }; _proto4.isEmpty = function isEmpty() { var empty = true; for (var i = 0; i < NR_COLS; i++) { if (!this.chars[i].isEmpty()) { empty = false; break; } } return empty; } /** * Set the cursor to a valid column. */ ; _proto4.setCursor = function setCursor(absPos) { if (this.pos !== absPos) { this.pos = absPos; } if (this.pos < 0) { this.logger.log(VerboseLevel.DEBUG, 'Negative cursor position ' + this.pos); this.pos = 0; } else if (this.pos > NR_COLS) { this.logger.log(VerboseLevel.DEBUG, 'Too large cursor position ' + this.pos); this.pos = NR_COLS; } } /** * Move the cursor relative to current position. */ ; _proto4.moveCursor = function moveCursor(relPos) { var newPos = this.pos + relPos; if (relPos > 1) { for (var i = this.pos + 1; i < newPos + 1; i++) { this.chars[i].setPenState(this.currPenState); } } this.setCursor(newPos); } /** * Backspace, move one step back and clear character. */ ; _proto4.backSpace = function backSpace() { this.moveCursor(-1); this.chars[this.pos].setChar(' ', this.currPenState); }; _proto4.insertChar = function insertChar(_byte2) { if (_byte2 >= 0x90) { // Extended char this.backSpace(); } var _char = getCharForByte(_byte2); if (this.pos >= NR_COLS) { this.logger.log(VerboseLevel.ERROR, 'Cannot insert ' + _byte2.toString(16) + ' (' + _char + ') at position ' + this.pos + '. Skipping it!'); return; } this.chars[this.pos].setChar(_char, this.currPenState); this.moveCursor(1); }; _proto4.clearFromPos = function clearFromPos(startPos) { var i; for (i = startPos; i < NR_COLS; i++) { this.chars[i].reset(); } }; _proto4.clear = function clear() { this.clearFromPos(0); this.pos = 0; this.currPenState.reset(); }; _proto4.clearToEndOfRow = function clearToEndOfRow() { this.clearFromPos(this.pos); }; _proto4.getTextString = function getTextString() { var chars = []; var empty = true; for (var i = 0; i < NR_COLS; i++) { var _char2 = this.chars[i].uchar; if (_char2 !== ' ') { empty = false; } chars.push(_char2); } if (empty) { return ''; } else { return chars.join(''); } }; _proto4.setPenStyles = function setPenStyles(styles) { this.currPenState.setStyles(styles); var currChar = this.chars[this.pos]; currChar.setPenState(this.currPenState); }; return Row; }(); /** * Keep a CEA-608 screen of 32x15 styled characters * @constructor */ var CaptionScreen = /*#__PURE__*/function () { function CaptionScreen(logger) { this.rows = void 0; this.currRow = void 0; this.nrRollUpRows = void 0; this.lastOutputScreen = void 0; this.logger = void 0; this.rows = []; for (var i = 0; i < NR_ROWS; i++) { this.rows.push(new Row(logger)); } // Note that we use zero-based numbering (0-14) this.logger = logger; this.currRow = NR_ROWS - 1; this.nrRollUpRows = null; this.lastOutputScreen = null; this.reset(); } var _proto5 = CaptionScreen.prototype; _proto5.reset = function reset() { for (var i = 0; i < NR_ROWS; i++) { this.rows[i].clear(); } this.currRow = NR_ROWS - 1; }; _proto5.equals = function equals(other) { var equal = true; for (var i = 0; i < NR_ROWS; i++) { if (!this.rows[i].equals(other.rows[i])) { equal = false; break; } } return equal; }; _proto5.copy = function copy(other) { for (var i = 0; i < NR_ROWS; i++) { this.rows[i].copy(other.rows[i]); } }; _proto5.isEmpty = function isEmpty() { var empty = true; for (var i = 0; i < NR_ROWS; i++) { if (!this.rows[i].isEmpty()) { empty = false; break; } } return empty; }; _proto5.backSpace = function backSpace() { var row = this.rows[this.currRow]; row.backSpace(); }; _proto5.clearToEndOfRow = function clearToEndOfRow() { var row = this.rows[this.currRow]; row.clearToEndOfRow(); } /** * Insert a character (without styling) in the current row. */ ; _proto5.insertChar = function insertChar(_char3) { var row = this.rows[this.currRow]; row.insertChar(_char3); }; _proto5.setPen = function setPen(styles) { var row = this.rows[this.currRow]; row.setPenStyles(styles); }; _proto5.moveCursor = function moveCursor(relPos) { var row = this.rows[this.currRow]; row.moveCursor(relPos); }; _proto5.setCursor = function setCursor(absPos) { this.logger.log(VerboseLevel.INFO, 'setCursor: ' + absPos); var row = this.rows[this.currRow]; row.setCursor(absPos); }; _proto5.setPAC = function setPAC(pacData) { this.logger.log(VerboseLevel.INFO, 'pacData = ' + JSON.stringify(pacData)); var newRow = pacData.row - 1; if (this.nrRollUpRows && newRow < this.nrRollUpRows - 1) { newRow = this.nrRollUpRows - 1; } // Make sure this only affects Roll-up Captions by checking this.nrRollUpRows if (this.nrRollUpRows && this.currRow !== newRow) { // clear all rows first for (var i = 0; i < NR_ROWS; i++) { this.rows[i].clear(); } // Copy this.nrRollUpRows rows from lastOutputScreen and place it in the newRow location // topRowIndex - the start of rows to copy (inclusive index) var topRowIndex = this.currRow + 1 - this.nrRollUpRows; // We only copy if the last position was already shown. // We use the cueStartTime value to check this. var lastOutputScreen = this.lastOutputScreen; if (lastOutputScreen) { var prevLineTime = lastOutputScreen.rows[topRowIndex].cueStartTime; var time = this.logger.time; if (prevLineTime && time !== null && prevLineTime < time) { for (var _i = 0; _i < this.nrRollUpRows; _i++) { this.rows[newRow - this.nrRollUpRows + _i + 1].copy(lastOutputScreen.rows[topRowIndex + _i]); } } } } this.currRow = newRow; var row = this.rows[this.currRow]; if (pacData.indent !== null) { var indent = pacData.indent; var prevPos = Math.max(indent - 1, 0); row.setCursor(pacData.indent); pacData.color = row.chars[prevPos].penState.foreground; } var styles = { foreground: pacData.color, underline: pacData.underline, italics: pacData.italics, background: 'black', flash: false }; this.setPen(styles); } /** * Set background/extra foreground, but first do back_space, and then insert space (backwards compatibility). */ ; _proto5.setBkgData = function setBkgData(bkgData) { this.logger.log(VerboseLevel.INFO, 'bkgData = ' + JSON.stringify(bkgData)); this.backSpace(); this.setPen(bkgData); this.insertChar(0x20); // Space }; _proto5.setRollUpRows = function setRollUpRows(nrRows) { this.nrRollUpRows = nrRows; }; _proto5.rollUp = function rollUp() { if (this.nrRollUpRows === null) { this.logger.log(VerboseLevel.DEBUG, 'roll_up but nrRollUpRows not set yet'); return; // Not properly setup } this.logger.log(VerboseLevel.TEXT, this.getDisplayText()); var topRowIndex = this.currRow + 1 - this.nrRollUpRows; var topRow = this.rows.splice(topRowIndex, 1)[0]; topRow.clear(); this.rows.splice(this.currRow, 0, topRow); this.logger.log(VerboseLevel.INFO, 'Rolling up'); // this.logger.log(VerboseLevel.TEXT, this.get_display_text()) } /** * Get all non-empty rows with as unicode text. */ ; _proto5.getDisplayText = function getDisplayText(asOneRow) { asOneRow = asOneRow || false; var displayText = []; var text = ''; var rowNr = -1; for (var i = 0; i < NR_ROWS; i++) { var rowText = this.rows[i].getTextString(); if (rowText) { rowNr = i + 1; if (asOneRow) { displayText.push('Row ' + rowNr + ': \'' + rowText + '\''); } else { displayText.push(rowText.trim()); } } } if (displayText.length > 0) { if (asOneRow) { text = '[' + displayText.join(' | ') + ']'; } else { text = displayText.join('\n'); } } return text; }; _proto5.getTextAndFormat = function getTextAndFormat() { return this.rows; }; return CaptionScreen; }(); // var modes = ['MODE_ROLL-UP', 'MODE_POP-ON', 'MODE_PAINT-ON', 'MODE_TEXT']; var Cea608Channel = /*#__PURE__*/function () { function Cea608Channel(channelNumber, outputFilter, logger) { this.chNr = void 0; this.outputFilter = void 0; this.mode = void 0; this.verbose = void 0; this.displayedMemory = void 0; this.nonDisplayedMemory = void 0; this.lastOutputScreen = void 0; this.currRollUpRow = void 0; this.writeScreen = void 0; this.cueStartTime = void 0; this.logger = void 0; this.chNr = channelNumber; this.outputFilter = outputFilter; this.mode = null; this.verbose = 0; this.displayedMemory = new CaptionScreen(logger); this.nonDisplayedMemory = new CaptionScreen(logger); this.lastOutputScreen = new CaptionScreen(logger); this.currRollUpRow = this.displayedMemory.rows[NR_ROWS - 1]; this.writeScreen = this.displayedMemory; this.mode = null; this.cueStartTime = null; // Keeps track of where a cue started. this.logger = logger; } var _proto6 = Cea608Channel.prototype; _proto6.reset = function reset() { this.mode = null; this.displayedMemory.reset(); this.nonDisplayedMemory.reset(); this.lastOutputScreen.reset(); this.outputFilter.reset(); this.currRollUpRow = this.displayedMemory.rows[NR_ROWS - 1]; this.writeScreen = this.displayedMemory; this.mode = null; this.cueStartTime = null; }; _proto6.getHandler = function getHandler() { return this.outputFilter; }; _proto6.setHandler = function setHandler(newHandler) { this.outputFilter = newHandler; }; _proto6.setPAC = function setPAC(pacData) { this.writeScreen.setPAC(pacData); }; _proto6.setBkgData = function setBkgData(bkgData) { this.writeScreen.setBkgData(bkgData); }; _proto6.setMode = function setMode(newMode) { if (newMode === this.mode) { return; } this.mode = newMode; this.logger.log(VerboseLevel.INFO, 'MODE=' + newMode); if (this.mode === 'MODE_POP-ON') { this.writeScreen = this.nonDisplayedMemory; } else { this.writeScreen = this.displayedMemory; this.writeScreen.reset(); } if (this.mode !== 'MODE_ROLL-UP') { this.displayedMemory.nrRollUpRows = null; this.nonDisplayedMemory.nrRollUpRows = null; } this.mode = newMode; }; _proto6.insertChars = function insertChars(chars) { for (var i = 0; i < chars.length; i++) { this.writeScreen.insertChar(chars[i]); } var screen = this.writeScreen === this.displayedMemory ? 'DISP' : 'NON_DISP'; this.logger.log(VerboseLevel.INFO, screen + ': ' + this.writeScreen.getDisplayText(true)); if (this.mode === 'MODE_PAINT-ON' || this.mode === 'MODE_ROLL-UP') { this.logger.log(VerboseLevel.TEXT, 'DISPLAYED: ' + this.displayedMemory.getDisplayText(true)); this.outputDataUpdate(); } }; _proto6.ccRCL = function ccRCL() { // Resume Caption Loading (switch mode to Pop On) this.logger.log(VerboseLevel.INFO, 'RCL - Resume Caption Loading'); this.setMode('MODE_POP-ON'); }; _proto6.ccBS = function ccBS() { // BackSpace this.logger.log(VerboseLevel.INFO, 'BS - BackSpace'); if (this.mode === 'MODE_TEXT') { return; } this.writeScreen.backSpace(); if (this.writeScreen === this.displayedMemory) { this.outputDataUpdate(); } }; _proto6.ccAOF = function ccAOF() {// Reserved (formerly Alarm Off) }; _proto6.ccAON = function ccAON() {// Reserved (formerly Alarm On) }; _proto6.ccDER = function ccDER() { // Delete to End of Row this.logger.log(VerboseLevel.INFO, 'DER- Delete to End of Row'); this.writeScreen.clearToEndOfRow(); this.outputDataUpdate(); }; _proto6.ccRU = function ccRU(nrRows) { // Roll-Up Captions-2,3,or 4 Rows this.logger.log(VerboseLevel.INFO, 'RU(' + nrRows + ') - Roll Up'); this.writeScreen = this.displayedMemory; this.setMode('MODE_ROLL-UP'); this.writeScreen.setRollUpRows(nrRows); }; _proto6.ccFON = function ccFON() { // Flash On this.logger.log(VerboseLevel.INFO, 'FON - Flash On'); this.writeScreen.setPen({ flash: true }); }; _proto6.ccRDC = function ccRDC() { // Resume Direct Captioning (switch mode to PaintOn) this.logger.log(VerboseLevel.INFO, 'RDC - Resume Direct Captioning'); this.setMode('MODE_PAINT-ON'); }; _proto6.ccTR = function ccTR() { // Text Restart in text mode (not supported, however) this.logger.log(VerboseLevel.INFO, 'TR'); this.setMode('MODE_TEXT'); }; _proto6.ccRTD = function ccRTD() { // Resume Text Display in Text mode (not supported, however) this.logger.log(VerboseLevel.INFO, 'RTD'); this.setMode('MODE_TEXT'); }; _proto6.ccEDM = function ccEDM() { // Erase Displayed Memory this.logger.log(VerboseLevel.INFO, 'EDM - Erase Displayed Memory'); this.displayedMemory.reset(); this.outputDataUpdate(true); }; _proto6.ccCR = function ccCR() { // Carriage Return this.logger.log(VerboseLevel.INFO, 'CR - Carriage Return'); this.writeScreen.rollUp(); this.outputDataUpdate(true); }; _proto6.ccENM = function ccENM() { // Erase Non-Displayed Memory this.logger.log(VerboseLevel.INFO, 'ENM - Erase Non-displayed Memory'); this.nonDisplayedMemory.reset(); }; _proto6.ccEOC = function ccEOC() { // End of Caption (Flip Memories) this.logger.log(VerboseLevel.INFO, 'EOC - End Of Caption'); if (this.mode === 'MODE_POP-ON') { var tmp = this.displayedMemory; this.displayedMemory = this.nonDisplayedMemory; this.nonDisplayedMemory = tmp; this.writeScreen = this.nonDisplayedMemory; this.logger.log(VerboseLevel.TEXT, 'DISP: ' + this.displayedMemory.getDisplayText()); } this.outputDataUpdate(true); }; _proto6.ccTO = function ccTO(nrCols) { // Tab Offset 1,2, or 3 columns this.logger.log(VerboseLevel.INFO, 'TO(' + nrCols + ') - Tab Offset'); this.writeScreen.moveCursor(nrCols); }; _proto6.ccMIDROW = function ccMIDROW(secondByte) { // Parse MIDROW command var styles = { flash: false }; styles.underline = secondByte % 2 === 1; styles.italics = secondByte >= 0x2e; if (!styles.italics) { var colorIndex = Math.floor(secondByte / 2) - 0x10; var colors = ['white', 'green', 'blue', 'cyan', 'red', 'yellow', 'magenta']; styles.foreground = colors[colorIndex]; } else { styles.foreground = 'white'; } this.logger.log(VerboseLevel.INFO, 'MIDROW: ' + JSON.stringify(styles)); this.writeScreen.setPen(styles); }; _proto6.outputDataUpdate = function outputDataUpdate(dispatch) { if (dispatch === void 0) { dispatch = false; } var time = this.logger.time; if (time === null) { return; } if (this.outputFilter) { if (this.cueStartTime === null && !this.displayedMemory.isEmpty()) { // Start of a new cue this.cueStartTime = time; } else { if (!this.displayedMemory.equals(this.lastOutputScreen)) { this.outputFilter.newCue(this.cueStartTime, time, this.lastOutputScreen); if (dispatch && this.outputFilter.dispatchCue) { this.outputFilter.dispatchCue(); } this.cueStartTime = this.displayedMemory.isEmpty() ? null : time; } } this.lastOutputScreen.copy(this.displayedMemory); } }; _proto6.cueSplitAtTime = function cueSplitAtTime(t) { if (this.outputFilter) { if (!this.displayedMemory.isEmpty()) { if (this.outputFilter.newCue) { this.outputFilter.newCue(this.cueStartTime, t, this.displayedMemory); } this.cueStartTime = t; } } }; return Cea608Channel; }(); var Cea608Parser = /*#__PURE__*/function () { function Cea608Parser(field, out1, out2) { this.channels = void 0; this.currentChannel = 0; this.cmdHistory = void 0; this.logger = void 0; var logger = new cea_608_parser_CaptionsLogger(); this.channels = [null, new Cea608Channel(field, out1, logger), new Cea608Channel(field + 1, out2, logger)]; this.cmdHistory = createCmdHistory(); this.logger = logger; } var _proto7 = Cea608Parser.prototype; _proto7.getHandler = function getHandler(channel) { return this.channels[channel].getHandler(); }; _proto7.setHandler = function setHandler(channel, newHandler) { this.channels[channel].setHandler(newHandler); } /** * Add data for time t in forms of list of bytes (unsigned ints). The bytes are treated as pairs. */ ; _proto7.addData = function addData(time, byteList) { var cmdFound; var a; var b; var charsFound = false; this.logger.time = time; for (var i = 0; i < byteList.length; i += 2) { a = byteList[i] & 0x7f; b = byteList[i + 1] & 0x7f; if (a === 0 && b === 0) { continue; } else { this.logger.log(VerboseLevel.DATA, '[' + numArrayToHexArray([byteList[i], byteList[i + 1]]) + '] -> (' + numArrayToHexArray([a, b]) + ')'); } cmdFound = this.parseCmd(a, b); if (!cmdFound) { cmdFound = this.parseMidrow(a, b); } if (!cmdFound) { cmdFound = this.parsePAC(a, b); } if (!cmdFound) { cmdFound = this.parseBackgroundAttributes(a, b); } if (!cmdFound) { charsFound = this.parseChars(a, b); if (charsFound) { var currChNr = this.currentChannel; if (currChNr && currChNr > 0) { var channel = this.channels[currChNr]; channel.insertChars(charsFound); } else { this.logger.log(VerboseLevel.WARNING, 'No channel found yet. TEXT-MODE?'); } } } if (!cmdFound && !charsFound) { this.logger.log(VerboseLevel.WARNING, 'Couldn\'t parse cleaned data ' + numArrayToHexArray([a, b]) + ' orig: ' + numArrayToHexArray([byteList[i], byteList[i + 1]])); } } } /** * Parse Command. * @returns {Boolean} Tells if a command was found */ ; _proto7.parseCmd = function parseCmd(a, b) { var cmdHistory = this.cmdHistory; var cond1 = (a === 0x14 || a === 0x1C || a === 0x15 || a === 0x1D) && b >= 0x20 && b <= 0x2F; var cond2 = (a === 0x17 || a === 0x1F) && b >= 0x21 && b <= 0x23; if (!(cond1 || cond2)) { return false; } if (hasCmdRepeated(a, b, cmdHistory)) { setLastCmd(null, null, cmdHistory); this.logger.log(VerboseLevel.DEBUG, 'Repeated command (' + numArrayToHexArray([a, b]) + ') is dropped'); return true; } var chNr = a === 0x14 || a === 0x15 || a === 0x17 ? 1 : 2; var channel = this.channels[chNr]; if (a === 0x14 || a === 0x15 || a === 0x1C || a === 0x1D) { if (b === 0x20) { channel.ccRCL(); } else if (b === 0x21) { channel.ccBS(); } else if (b === 0x22) { channel.ccAOF(); } else if (b === 0x23) { channel.ccAON(); } else if (b === 0x24) { channel.ccDER(); } else if (b === 0x25) { channel.ccRU(2); } else if (b === 0x26) { channel.ccRU(3); } else if (b === 0x27) { channel.ccRU(4); } else if (b === 0x28) { channel.ccFON(); } else if (b === 0x29) { channel.ccRDC(); } else if (b === 0x2A) { channel.ccTR(); } else if (b === 0x2B) { channel.ccRTD(); } else if (b === 0x2C) { channel.ccEDM(); } else if (b === 0x2D) { channel.ccCR(); } else if (b === 0x2E) { channel.ccENM(); } else if (b === 0x2F) { channel.ccEOC(); } } else { // a == 0x17 || a == 0x1F channel.ccTO(b - 0x20); } setLastCmd(a, b, cmdHistory); this.currentChannel = chNr; return true; } /** * Parse midrow styling command * @returns {Boolean} */ ; _proto7.parseMidrow = function parseMidrow(a, b) { var chNr = 0; if ((a === 0x11 || a === 0x19) && b >= 0x20 && b <= 0x2f) { if (a === 0x11) { chNr = 1; } else { chNr = 2; } if (chNr !== this.currentChannel) { this.logger.log(VerboseLevel.ERROR, 'Mismatch channel in midrow parsing'); return false; } var channel = this.channels[chNr]; if (!channel) { return false; } channel.ccMIDROW(b); this.logger.log(VerboseLevel.DEBUG, 'MIDROW (' + numArrayToHexArray([a, b]) + ')'); return true; } return false; } /** * Parse Preable Access Codes (Table 53). * @returns {Boolean} Tells if PAC found */ ; _proto7.parsePAC = function parsePAC(a, b) { var row; var cmdHistory = this.cmdHistory; var case1 = (a >= 0x11 && a <= 0x17 || a >= 0x19 && a <= 0x1F) && b >= 0x40 && b <= 0x7F; var case2 = (a === 0x10 || a === 0x18) && b >= 0x40 && b <= 0x5F; if (!(case1 || case2)) { return false; } if (hasCmdRepeated(a, b, cmdHistory)) { setLastCmd(null, null, cmdHistory); return true; // Repeated commands are dropped (once) } var chNr = a <= 0x17 ? 1 : 2; if (b >= 0x40 && b <= 0x5F) { row = chNr === 1 ? rowsLowCh1[a] : rowsLowCh2[a]; } else { // 0x60 <= b <= 0x7F row = chNr === 1 ? rowsHighCh1[a] : rowsHighCh2[a]; } var channel = this.channels[chNr]; if (!channel) { return false; } channel.setPAC(this.interpretPAC(row, b)); setLastCmd(a, b, cmdHistory); this.currentChannel = chNr; return true; } /** * Interpret the second byte of the pac, and return the information. * @returns {Object} pacData with style parameters. */ ; _proto7.interpretPAC = function interpretPAC(row, _byte3) { var pacIndex = _byte3; var pacData = { color: null, italics: false, indent: null, underline: false, row: row }; if (_byte3 > 0x5F) { pacIndex = _byte3 - 0x60; } else { pacIndex = _byte3 - 0x40; } pacData.underline = (pacIndex & 1) === 1; if (pacIndex <= 0xd) { pacData.color = ['white', 'green', 'blue', 'cyan', 'red', 'yellow', 'magenta', 'white'][Math.floor(pacIndex / 2)]; } else if (pacIndex <= 0xf) { pacData.italics = true; pacData.color = 'white'; } else { pacData.indent = Math.floor((pacIndex - 0x10) / 2) * 4; } return pacData; // Note that row has zero offset. The spec uses 1. } /** * Parse characters. * @returns An array with 1 to 2 codes corresponding to chars, if found. null otherwise. */ ; _proto7.parseChars = function parseChars(a, b) { var channelNr; var charCodes = null; var charCode1 = null; if (a >= 0x19) { channelNr = 2; charCode1 = a - 8; } else { channelNr = 1; charCode1 = a; } if (charCode1 >= 0x11 && charCode1 <= 0x13) { // Special character var oneCode = b; if (charCode1 === 0x11) { oneCode = b + 0x50; } else if (charCode1 === 0x12) { oneCode = b + 0x70; } else { oneCode = b + 0x90; } this.logger.log(VerboseLevel.INFO, 'Special char \'' + getCharForByte(oneCode) + '\' in channel ' + channelNr); charCodes = [oneCode]; } else if (a >= 0x20 && a <= 0x7f) { charCodes = b === 0 ? [a] : [a, b]; } if (charCodes) { var hexCodes = numArrayToHexArray(charCodes); this.logger.log(VerboseLevel.DEBUG, 'Char codes = ' + hexCodes.join(',')); setLastCmd(a, b, this.cmdHistory); } return charCodes; } /** * Parse extended background attributes as well as new foreground color black. * @returns {Boolean} Tells if background attributes are found */ ; _proto7.parseBackgroundAttributes = function parseBackgroundAttributes(a, b) { var case1 = (a === 0x10 || a === 0x18) && b >= 0x20 && b <= 0x2f; var case2 = (a === 0x17 || a === 0x1f) && b >= 0x2d && b <= 0x2f; if (!(case1 || case2)) { return false; } var index; var bkgData = {}; if (a === 0x10 || a === 0x18) { index = Math.floor((b - 0x20) / 2); bkgData.background = backgroundColors[index]; if (b % 2 === 1) { bkgData.background = bkgData.background + '_semi'; } } else if (b === 0x2d) { bkgData.background = 'transparent'; } else { bkgData.foreground = 'black'; if (b === 0x2f) { bkgData.underline = true; } } var chNr = a <= 0x17 ? 1 : 2; var channel = this.channels[chNr]; channel.setBkgData(bkgData); setLastCmd(a, b, this.cmdHistory); return true; } /** * Reset state of parser and its channels. */ ; _proto7.reset = function reset() { for (var i = 0; i < Object.keys(this.channels).length; i++) { var channel = this.channels[i]; if (channel) { channel.reset(); } } this.cmdHistory = createCmdHistory(); } /** * Trigger the generation of a cue, and the start of a new one if displayScreens are not empty. */ ; _proto7.cueSplitAtTime = function cueSplitAtTime(t) { for (var i = 0; i < this.channels.length; i++) { var channel = this.channels[i]; if (channel) { channel.cueSplitAtTime(t); } } }; return Cea608Parser; }(); function setLastCmd(a, b, cmdHistory) { cmdHistory.a = a; cmdHistory.b = b; } function hasCmdRepeated(a, b, cmdHistory) { return cmdHistory.a === a && cmdHistory.b === b; } function createCmdHistory() { return { a: null, b: null }; } /* harmony default export */ var cea_608_parser = (Cea608Parser); // CONCATENATED MODULE: ./src/utils/output-filter.ts var OutputFilter = /*#__PURE__*/function () { function OutputFilter(timelineController, trackName) { this.timelineController = void 0; this.cueRanges = []; this.trackName = void 0; this.startTime = null; this.endTime = null; this.screen = null; this.timelineController = timelineController; this.trackName = trackName; } var _proto = OutputFilter.prototype; _proto.dispatchCue = function dispatchCue() { if (this.startTime === null) { return; } this.timelineController.addCues(this.trackName, this.startTime, this.endTime, this.screen, this.cueRanges); this.startTime = null; }; _proto.newCue = function newCue(startTime, endTime, screen) { if (this.startTime === null || this.startTime > startTime) { this.startTime = startTime; } this.endTime = endTime; this.screen = screen; this.timelineController.createCaptionsTrack(this.trackName); }; _proto.reset = function reset() { this.cueRanges = []; }; return OutputFilter; }(); // CONCATENATED MODULE: ./src/utils/webvtt-parser.js // String.prototype.startsWith is not supported in IE11 var startsWith = function startsWith(inputString, searchString, position) { return inputString.substr(position || 0, searchString.length) === searchString; }; var webvtt_parser_cueString2millis = function cueString2millis(timeString) { var ts = parseInt(timeString.substr(-3)); var secs = parseInt(timeString.substr(-6, 2)); var mins = parseInt(timeString.substr(-9, 2)); var hours = timeString.length > 9 ? parseInt(timeString.substr(0, timeString.indexOf(':'))) : 0; if (!Object(number["isFiniteNumber"])(ts) || !Object(number["isFiniteNumber"])(secs) || !Object(number["isFiniteNumber"])(mins) || !Object(number["isFiniteNumber"])(hours)) { throw Error("Malformed X-TIMESTAMP-MAP: Local:" + timeString); } ts += 1000 * secs; ts += 60 * 1000 * mins; ts += 60 * 60 * 1000 * hours; return ts; }; // From https://github.com/darkskyapp/string-hash var hash = function hash(text) { var hash = 5381; var i = text.length; while (i) { hash = hash * 33 ^ text.charCodeAt(--i); } return (hash >>> 0).toString(); }; var calculateOffset = function calculateOffset(vttCCs, cc, presentationTime) { var currCC = vttCCs[cc]; var prevCC = vttCCs[currCC.prevCC]; // This is the first discontinuity or cues have been processed since the last discontinuity // Offset = current discontinuity time if (!prevCC || !prevCC.new && currCC.new) { vttCCs.ccOffset = vttCCs.presentationOffset = currCC.start; currCC.new = false; return; } // There have been discontinuities since cues were last parsed. // Offset = time elapsed while (prevCC && prevCC.new) { vttCCs.ccOffset += currCC.start - prevCC.start; currCC.new = false; currCC = prevCC; prevCC = vttCCs[currCC.prevCC]; } vttCCs.presentationOffset = presentationTime; }; var WebVTTParser = { parse: function parse(vttByteArray, syncPTS, vttCCs, cc, callBack, errorCallBack) { // Convert byteArray into string, replacing any somewhat exotic linefeeds with "\n", then split on that character. var re = /\r\n|\n\r|\n|\r/g; // Uint8Array.prototype.reduce is not implemented in IE11 var vttLines = Object(id3["utf8ArrayToStr"])(new Uint8Array(vttByteArray)).trim().replace(re, '\n').split('\n'); var cueTime = '00:00.000'; var mpegTs = 0; var localTime = 0; var presentationTime = 0; var cues = []; var parsingError; var inHeader = true; var timestampMap = false; // let VTTCue = VTTCue || window.TextTrackCue; // Create parser object using VTTCue with TextTrackCue fallback on certain browsers. var parser = new vttparser(); parser.oncue = function (cue) { // Adjust cue timing; clamp cues to start no earlier than - and drop cues that don't end after - 0 on timeline. var currCC = vttCCs[cc]; var cueOffset = vttCCs.ccOffset; // Update offsets for new discontinuities if (currCC && currCC.new) { if (localTime !== undefined) { // When local time is provided, offset = discontinuity start time - local time cueOffset = vttCCs.ccOffset = currCC.start; } else { calculateOffset(vttCCs, cc, presentationTime); } } if (presentationTime) { // If we have MPEGTS, offset = presentation time + discontinuity offset cueOffset = presentationTime - vttCCs.presentationOffset; } if (timestampMap) { cue.startTime += cueOffset - localTime; cue.endTime += cueOffset - localTime; } // Create a unique hash id for a cue based on start/end times and text. // This helps timeline-controller to avoid showing repeated captions. cue.id = hash(cue.startTime.toString()) + hash(cue.endTime.toString()) + hash(cue.text); // Fix encoding of special characters. TODO: Test with all sorts of weird characters. cue.text = decodeURIComponent(encodeURIComponent(cue.text)); if (cue.endTime > 0) { cues.push(cue); } }; parser.onparsingerror = function (e) { parsingError = e; }; parser.onflush = function () { if (parsingError && errorCallBack) { errorCallBack(parsingError); return; } callBack(cues); }; // Go through contents line by line. vttLines.forEach(function (line) { if (inHeader) { // Look for X-TIMESTAMP-MAP in header. if (startsWith(line, 'X-TIMESTAMP-MAP=')) { // Once found, no more are allowed anyway, so stop searching. inHeader = false; timestampMap = true; // Extract LOCAL and MPEGTS. line.substr(16).split(',').forEach(function (timestamp) { if (startsWith(timestamp, 'LOCAL:')) { cueTime = timestamp.substr(6); } else if (startsWith(timestamp, 'MPEGTS:')) { mpegTs = parseInt(timestamp.substr(7)); } }); try { // Calculate subtitle offset in milliseconds. if (syncPTS + (vttCCs[cc].start * 90000 || 0) < 0) { syncPTS += 8589934592; } // Adjust MPEGTS by sync PTS. mpegTs -= syncPTS; // Convert cue time to seconds localTime = webvtt_parser_cueString2millis(cueTime) / 1000; // Convert MPEGTS to seconds from 90kHz. presentationTime = mpegTs / 90000; } catch (e) { timestampMap = false; parsingError = e; } // Return without parsing X-TIMESTAMP-MAP line. return; } else if (line === '') { inHeader = false; } } // Parse line by default. parser.parse(line + '\n'); }); parser.flush(); } }; /* harmony default export */ var webvtt_parser = (WebVTTParser); // CONCATENATED MODULE: ./src/controller/timeline-controller.ts function timeline_controller_assertThisInitialized(self) { if (self === void 0) { throw new ReferenceError("this hasn't been initialised - super() hasn't been called"); } return self; } function timeline_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } var timeline_controller_TimelineController = /*#__PURE__*/function (_EventHandler) { timeline_controller_inheritsLoose(TimelineController, _EventHandler); function TimelineController(hls) { var _this; _this = _EventHandler.call(this, hls, events["default"].MEDIA_ATTACHING, events["default"].MEDIA_DETACHING, events["default"].FRAG_PARSING_USERDATA, events["default"].FRAG_DECRYPTED, events["default"].MANIFEST_LOADING, events["default"].MANIFEST_LOADED, events["default"].FRAG_LOADED, events["default"].INIT_PTS_FOUND) || this; _this.media = null; _this.config = void 0; _this.enabled = true; _this.Cues = void 0; _this.textTracks = []; _this.tracks = []; _this.initPTS = []; _this.unparsedVttFrags = []; _this.captionsTracks = {}; _this.nonNativeCaptionsTracks = {}; _this.captionsProperties = void 0; _this.cea608Parser1 = void 0; _this.cea608Parser2 = void 0; _this.lastSn = -1; _this.prevCC = -1; _this.vttCCs = newVTTCCs(); _this.hls = hls; _this.config = hls.config; _this.Cues = hls.config.cueHandler; _this.captionsProperties = { textTrack1: { label: _this.config.captionsTextTrack1Label, languageCode: _this.config.captionsTextTrack1LanguageCode }, textTrack2: { label: _this.config.captionsTextTrack2Label, languageCode: _this.config.captionsTextTrack2LanguageCode }, textTrack3: { label: _this.config.captionsTextTrack3Label, languageCode: _this.config.captionsTextTrack3LanguageCode }, textTrack4: { label: _this.config.captionsTextTrack4Label, languageCode: _this.config.captionsTextTrack4LanguageCode } }; if (_this.config.enableCEA708Captions) { var channel1 = new OutputFilter(timeline_controller_assertThisInitialized(_this), 'textTrack1'); var channel2 = new OutputFilter(timeline_controller_assertThisInitialized(_this), 'textTrack2'); var channel3 = new OutputFilter(timeline_controller_assertThisInitialized(_this), 'textTrack3'); var channel4 = new OutputFilter(timeline_controller_assertThisInitialized(_this), 'textTrack4'); _this.cea608Parser1 = new cea_608_parser(1, channel1, channel2); _this.cea608Parser2 = new cea_608_parser(3, channel3, channel4); } return _this; } var _proto = TimelineController.prototype; _proto.addCues = function addCues(trackName, startTime, endTime, screen, cueRanges) { // skip cues which overlap more than 50% with previously parsed time ranges var merged = false; for (var i = cueRanges.length; i--;) { var cueRange = cueRanges[i]; var overlap = intersection(cueRange[0], cueRange[1], startTime, endTime); if (overlap >= 0) { cueRange[0] = Math.min(cueRange[0], startTime); cueRange[1] = Math.max(cueRange[1], endTime); merged = true; if (overlap / (endTime - startTime) > 0.5) { return; } } } if (!merged) { cueRanges.push([startTime, endTime]); } if (this.config.renderTextTracksNatively) { this.Cues.newCue(this.captionsTracks[trackName], startTime, endTime, screen); } else { var cues = this.Cues.newCue(null, startTime, endTime, screen); this.hls.trigger(events["default"].CUES_PARSED, { type: 'captions', cues: cues, track: trackName }); } } // Triggered when an initial PTS is found; used for synchronisation of WebVTT. ; _proto.onInitPtsFound = function onInitPtsFound(data) { var _this2 = this; var frag = data.frag, id = data.id, initPTS = data.initPTS; var unparsedVttFrags = this.unparsedVttFrags; if (id === 'main') { this.initPTS[frag.cc] = initPTS; } // Due to asynchronous processing, initial PTS may arrive later than the first VTT fragments are loaded. // Parse any unparsed fragments upon receiving the initial PTS. if (unparsedVttFrags.length) { this.unparsedVttFrags = []; unparsedVttFrags.forEach(function (frag) { _this2.onFragLoaded(frag); }); } }; _proto.getExistingTrack = function getExistingTrack(trackName) { var media = this.media; if (media) { for (var i = 0; i < media.textTracks.length; i++) { var textTrack = media.textTracks[i]; if (textTrack[trackName]) { return textTrack; } } } return null; }; _proto.createCaptionsTrack = function createCaptionsTrack(trackName) { if (this.config.renderTextTracksNatively) { this.createNativeTrack(trackName); } else { this.createNonNativeTrack(trackName); } }; _proto.createNativeTrack = function createNativeTrack(trackName) { if (this.captionsTracks[trackName]) { return; } var captionsProperties = this.captionsProperties, captionsTracks = this.captionsTracks, media = this.media; var _captionsProperties$t = captionsProperties[trackName], label = _captionsProperties$t.label, languageCode = _captionsProperties$t.languageCode; // Enable reuse of existing text track. var existingTrack = this.getExistingTrack(trackName); if (!existingTrack) { var textTrack = this.createTextTrack('captions', label, languageCode); if (textTrack) { // Set a special property on the track so we know it's managed by Hls.js textTrack[trackName] = true; captionsTracks[trackName] = textTrack; } } else { captionsTracks[trackName] = existingTrack; clearCurrentCues(captionsTracks[trackName]); sendAddTrackEvent(captionsTracks[trackName], media); } }; _proto.createNonNativeTrack = function createNonNativeTrack(trackName) { if (this.nonNativeCaptionsTracks[trackName]) { return; } // Create a list of a single track for the provider to consume var trackProperties = this.captionsProperties[trackName]; if (!trackProperties) { return; } var label = trackProperties.label; var track = { _id: trackName, label: label, kind: 'captions', default: trackProperties.media ? !!trackProperties.media.default : false, closedCaptions: trackProperties.media }; this.nonNativeCaptionsTracks[trackName] = track; this.hls.trigger(events["default"].NON_NATIVE_TEXT_TRACKS_FOUND, { tracks: [track] }); }; _proto.createTextTrack = function createTextTrack(kind, label, lang) { var media = this.media; if (!media) { return; } return media.addTextTrack(kind, label, lang); }; _proto.destroy = function destroy() { _EventHandler.prototype.destroy.call(this); }; _proto.onMediaAttaching = function onMediaAttaching(data) { this.media = data.media; this._cleanTracks(); }; _proto.onMediaDetaching = function onMediaDetaching() { var captionsTracks = this.captionsTracks; Object.keys(captionsTracks).forEach(function (trackName) { clearCurrentCues(captionsTracks[trackName]); delete captionsTracks[trackName]; }); this.nonNativeCaptionsTracks = {}; }; _proto.onManifestLoading = function onManifestLoading() { this.lastSn = -1; // Detect discontiguity in fragment parsing this.prevCC = -1; this.vttCCs = newVTTCCs(); // Detect discontinuity in subtitle manifests this._cleanTracks(); this.tracks = []; this.captionsTracks = {}; this.nonNativeCaptionsTracks = {}; }; _proto._cleanTracks = function _cleanTracks() { // clear outdated subtitles var media = this.media; if (!media) { return; } var textTracks = media.textTracks; if (textTracks) { for (var i = 0; i < textTracks.length; i++) { clearCurrentCues(textTracks[i]); } } }; _proto.onManifestLoaded = function onManifestLoaded(data) { var _this3 = this; this.textTracks = []; this.unparsedVttFrags = this.unparsedVttFrags || []; this.initPTS = []; if (this.cea608Parser1 && this.cea608Parser2) { this.cea608Parser1.reset(); this.cea608Parser2.reset(); } if (this.config.enableWebVTT) { var tracks = data.subtitles || []; var sameTracks = this.tracks && tracks && this.tracks.length === tracks.length; this.tracks = data.subtitles || []; if (this.config.renderTextTracksNatively) { var inUseTracks = this.media ? this.media.textTracks : []; this.tracks.forEach(function (track, index) { var textTrack; if (index < inUseTracks.length) { var inUseTrack = null; for (var i = 0; i < inUseTracks.length; i++) { if (canReuseVttTextTrack(inUseTracks[i], track)) { inUseTrack = inUseTracks[i]; break; } } // Reuse tracks with the same label, but do not reuse 608/708 tracks if (inUseTrack) { textTrack = inUseTrack; } } if (!textTrack) { textTrack = _this3.createTextTrack('subtitles', track.name, track.lang); } if (track.default) { textTrack.mode = _this3.hls.subtitleDisplay ? 'showing' : 'hidden'; } else { textTrack.mode = 'disabled'; } _this3.textTracks.push(textTrack); }); } else if (!sameTracks && this.tracks && this.tracks.length) { // Create a list of tracks for the provider to consume var tracksList = this.tracks.map(function (track) { return { label: track.name, kind: track.type.toLowerCase(), default: track.default, subtitleTrack: track }; }); this.hls.trigger(events["default"].NON_NATIVE_TEXT_TRACKS_FOUND, { tracks: tracksList }); } } if (this.config.enableCEA708Captions && data.captions) { data.captions.forEach(function (captionsTrack) { var instreamIdMatch = /(?:CC|SERVICE)([1-4])/.exec(captionsTrack.instreamId); if (!instreamIdMatch) { return; } var trackName = "textTrack" + instreamIdMatch[1]; var trackProperties = _this3.captionsProperties[trackName]; if (!trackProperties) { return; } trackProperties.label = captionsTrack.name; if (captionsTrack.lang) { // optional attribute trackProperties.languageCode = captionsTrack.lang; } trackProperties.media = captionsTrack; }); } }; _proto.onFragLoaded = function onFragLoaded(data) { var frag = data.frag, payload = data.payload; var cea608Parser1 = this.cea608Parser1, cea608Parser2 = this.cea608Parser2, initPTS = this.initPTS, lastSn = this.lastSn, unparsedVttFrags = this.unparsedVttFrags; if (frag.type === 'main') { var sn = frag.sn; // if this frag isn't contiguous, clear the parser so cues with bad start/end times aren't added to the textTrack if (frag.sn !== lastSn + 1) { if (cea608Parser1 && cea608Parser2) { cea608Parser1.reset(); cea608Parser2.reset(); } } this.lastSn = sn; } // eslint-disable-line brace-style // If fragment is subtitle type, parse as WebVTT. else if (frag.type === 'subtitle') { if (payload.byteLength) { // We need an initial synchronisation PTS. Store fragments as long as none has arrived. if (!Object(number["isFiniteNumber"])(initPTS[frag.cc])) { unparsedVttFrags.push(data); if (initPTS.length) { // finish unsuccessfully, otherwise the subtitle-stream-controller could be blocked from loading new frags. this.hls.trigger(events["default"].SUBTITLE_FRAG_PROCESSED, { success: false, frag: frag }); } return; } var decryptData = frag.decryptdata; // If the subtitles are not encrypted, parse VTTs now. Otherwise, we need to wait. if (decryptData == null || decryptData.key == null || decryptData.method !== 'AES-128') { this._parseVTTs(frag, payload); } } else { // In case there is no payload, finish unsuccessfully. this.hls.trigger(events["default"].SUBTITLE_FRAG_PROCESSED, { success: false, frag: frag }); } } }; _proto._parseVTTs = function _parseVTTs(frag, payload) { var _this4 = this; var hls = this.hls, prevCC = this.prevCC, textTracks = this.textTracks, vttCCs = this.vttCCs; if (!vttCCs[frag.cc]) { vttCCs[frag.cc] = { start: frag.start, prevCC: prevCC, new: true }; this.prevCC = frag.cc; } // Parse the WebVTT file contents. webvtt_parser.parse(payload, this.initPTS[frag.cc], vttCCs, frag.cc, function (cues) { if (_this4.config.renderTextTracksNatively) { var currentTrack = textTracks[frag.level]; // WebVTTParser.parse is an async method and if the currently selected text track mode is set to "disabled" // before parsing is done then don't try to access currentTrack.cues.getCueById as cues will be null // and trying to access getCueById method of cues will throw an exception // Because we check if the mode is diabled, we can force check `cues` below. They can't be null. if (currentTrack.mode === 'disabled') { hls.trigger(events["default"].SUBTITLE_FRAG_PROCESSED, { success: false, frag: frag }); return; } // Add cues and trigger event with success true. cues.forEach(function (cue) { // Sometimes there are cue overlaps on segmented vtts so the same // cue can appear more than once in different vtt files. // This avoid showing duplicated cues with same timecode and text. if (!currentTrack.cues.getCueById(cue.id)) { try { currentTrack.addCue(cue); if (!currentTrack.cues.getCueById(cue.id)) { throw new Error("addCue is failed for: " + cue); } } catch (err) { logger["logger"].debug("Failed occurred on adding cues: " + err); var textTrackCue = new window.TextTrackCue(cue.startTime, cue.endTime, cue.text); textTrackCue.id = cue.id; currentTrack.addCue(textTrackCue); } } }); } else { var trackId = _this4.tracks[frag.level].default ? 'default' : 'subtitles' + frag.level; hls.trigger(events["default"].CUES_PARSED, { type: 'subtitles', cues: cues, track: trackId }); } hls.trigger(events["default"].SUBTITLE_FRAG_PROCESSED, { success: true, frag: frag }); }, function (e) { // Something went wrong while parsing. Trigger event with success false. logger["logger"].log("Failed to parse VTT cue: " + e); hls.trigger(events["default"].SUBTITLE_FRAG_PROCESSED, { success: false, frag: frag }); }); }; _proto.onFragDecrypted = function onFragDecrypted(data) { var frag = data.frag, payload = data.payload; if (frag.type === 'subtitle') { if (!Object(number["isFiniteNumber"])(this.initPTS[frag.cc])) { this.unparsedVttFrags.push(data); return; } this._parseVTTs(frag, payload); } }; _proto.onFragParsingUserdata = function onFragParsingUserdata(data) { var cea608Parser1 = this.cea608Parser1, cea608Parser2 = this.cea608Parser2; if (!this.enabled || !(cea608Parser1 && cea608Parser2)) { return; } // If the event contains captions (found in the bytes property), push all bytes into the parser immediately // It will create the proper timestamps based on the PTS value for (var i = 0; i < data.samples.length; i++) { var ccBytes = data.samples[i].bytes; if (ccBytes) { var ccdatas = this.extractCea608Data(ccBytes); cea608Parser1.addData(data.samples[i].pts, ccdatas[0]); cea608Parser2.addData(data.samples[i].pts, ccdatas[1]); } } }; _proto.extractCea608Data = function extractCea608Data(byteArray) { var count = byteArray[0] & 31; var position = 2; var actualCCBytes = [[], []]; for (var j = 0; j < count; j++) { var tmpByte = byteArray[position++]; var ccbyte1 = 0x7F & byteArray[position++]; var ccbyte2 = 0x7F & byteArray[position++]; var ccValid = (4 & tmpByte) !== 0; var ccType = 3 & tmpByte; if (ccbyte1 === 0 && ccbyte2 === 0) { continue; } if (ccValid) { if (ccType === 0 || ccType === 1) { actualCCBytes[ccType].push(ccbyte1); actualCCBytes[ccType].push(ccbyte2); } } } return actualCCBytes; }; return TimelineController; }(event_handler); function canReuseVttTextTrack(inUseTrack, manifestTrack) { return inUseTrack && inUseTrack.label === manifestTrack.name && !(inUseTrack.textTrack1 || inUseTrack.textTrack2); } function intersection(x1, x2, y1, y2) { return Math.min(x2, y2) - Math.max(x1, y1); } function newVTTCCs() { return { ccOffset: 0, presentationOffset: 0, 0: { start: 0, prevCC: -1, new: false } }; } /* harmony default export */ var timeline_controller = (timeline_controller_TimelineController); // CONCATENATED MODULE: ./src/controller/subtitle-track-controller.js function subtitle_track_controller_defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } function subtitle_track_controller_createClass(Constructor, protoProps, staticProps) { if (protoProps) subtitle_track_controller_defineProperties(Constructor.prototype, protoProps); if (staticProps) subtitle_track_controller_defineProperties(Constructor, staticProps); return Constructor; } function subtitle_track_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } var subtitle_track_controller_SubtitleTrackController = /*#__PURE__*/function (_EventHandler) { subtitle_track_controller_inheritsLoose(SubtitleTrackController, _EventHandler); function SubtitleTrackController(hls) { var _this; _this = _EventHandler.call(this, hls, events["default"].MEDIA_ATTACHED, events["default"].MEDIA_DETACHING, events["default"].MANIFEST_LOADED, events["default"].SUBTITLE_TRACK_LOADED) || this; _this.tracks = []; _this.trackId = -1; _this.media = null; _this.stopped = true; /** * @member {boolean} subtitleDisplay Enable/disable subtitle display rendering */ _this.subtitleDisplay = true; /** * Keeps reference to a default track id when media has not been attached yet * @member {number} */ _this.queuedDefaultTrack = null; return _this; } var _proto = SubtitleTrackController.prototype; _proto.destroy = function destroy() { event_handler.prototype.destroy.call(this); } // Listen for subtitle track change, then extract the current track ID. ; _proto.onMediaAttached = function onMediaAttached(data) { var _this2 = this; this.media = data.media; if (!this.media) { return; } if (Object(number["isFiniteNumber"])(this.queuedDefaultTrack)) { this.subtitleTrack = this.queuedDefaultTrack; this.queuedDefaultTrack = null; } this.trackChangeListener = this._onTextTracksChanged.bind(this); this.useTextTrackPolling = !(this.media.textTracks && 'onchange' in this.media.textTracks); if (this.useTextTrackPolling) { this.subtitlePollingInterval = setInterval(function () { _this2.trackChangeListener(); }, 500); } else { this.media.textTracks.addEventListener('change', this.trackChangeListener); } }; _proto.onMediaDetaching = function onMediaDetaching() { if (!this.media) { return; } if (this.useTextTrackPolling) { clearInterval(this.subtitlePollingInterval); } else { this.media.textTracks.removeEventListener('change', this.trackChangeListener); } if (Object(number["isFiniteNumber"])(this.subtitleTrack)) { this.queuedDefaultTrack = this.subtitleTrack; } var textTracks = filterSubtitleTracks(this.media.textTracks); // Clear loaded cues on media detachment from tracks textTracks.forEach(function (track) { clearCurrentCues(track); }); // Disable all subtitle tracks before detachment so when reattached only tracks in that content are enabled. this.subtitleTrack = -1; this.media = null; } // Fired whenever a new manifest is loaded. ; _proto.onManifestLoaded = function onManifestLoaded(data) { var _this3 = this; var tracks = data.subtitles || []; this.tracks = tracks; this.hls.trigger(events["default"].SUBTITLE_TRACKS_UPDATED, { subtitleTracks: tracks }); // loop through available subtitle tracks and autoselect default if needed // TODO: improve selection logic to handle forced, etc tracks.forEach(function (track) { if (track.default) { // setting this.subtitleTrack will trigger internal logic // if media has not been attached yet, it will fail // we keep a reference to the default track id // and we'll set subtitleTrack when onMediaAttached is triggered if (_this3.media) { _this3.subtitleTrack = track.id; } else { _this3.queuedDefaultTrack = track.id; } } }); }; _proto.onSubtitleTrackLoaded = function onSubtitleTrackLoaded(data) { var _this4 = this; var id = data.id, details = data.details; var trackId = this.trackId, tracks = this.tracks; var currentTrack = tracks[trackId]; if (id >= tracks.length || id !== trackId || !currentTrack || this.stopped) { this._clearReloadTimer(); return; } logger["logger"].log("subtitle track " + id + " loaded"); if (details.live) { var reloadInterval = computeReloadInterval(currentTrack.details, details, data.stats.trequest); logger["logger"].log("Reloading live subtitle playlist in " + reloadInterval + "ms"); this.timer = setTimeout(function () { _this4._loadCurrentTrack(); }, reloadInterval); } else { this._clearReloadTimer(); } }; _proto.startLoad = function startLoad() { this.stopped = false; this._loadCurrentTrack(); }; _proto.stopLoad = function stopLoad() { this.stopped = true; this._clearReloadTimer(); } /** get alternate subtitle tracks list from playlist **/ ; _proto._clearReloadTimer = function _clearReloadTimer() { if (this.timer) { clearTimeout(this.timer); this.timer = null; } }; _proto._loadCurrentTrack = function _loadCurrentTrack() { var trackId = this.trackId, tracks = this.tracks, hls = this.hls; var currentTrack = tracks[trackId]; if (trackId < 0 || !currentTrack || currentTrack.details && !currentTrack.details.live) { return; } logger["logger"].log("Loading subtitle track " + trackId); hls.trigger(events["default"].SUBTITLE_TRACK_LOADING, { url: currentTrack.url, id: trackId }); } /** * Disables the old subtitleTrack and sets current mode on the next subtitleTrack. * This operates on the DOM textTracks. * A value of -1 will disable all subtitle tracks. * @param newId - The id of the next track to enable * @private */ ; _proto._toggleTrackModes = function _toggleTrackModes(newId) { var media = this.media, subtitleDisplay = this.subtitleDisplay, trackId = this.trackId; if (!media) { return; } var textTracks = filterSubtitleTracks(media.textTracks); if (newId === -1) { [].slice.call(textTracks).forEach(function (track) { track.mode = 'disabled'; }); } else { var oldTrack = textTracks[trackId]; if (oldTrack) { oldTrack.mode = 'disabled'; } } var nextTrack = textTracks[newId]; if (nextTrack) { nextTrack.mode = subtitleDisplay ? 'showing' : 'hidden'; } } /** * This method is responsible for validating the subtitle index and periodically reloading if live. * Dispatches the SUBTITLE_TRACK_SWITCH event, which instructs the subtitle-stream-controller to load the selected track. * @param newId - The id of the subtitle track to activate. */ ; _proto._setSubtitleTrackInternal = function _setSubtitleTrackInternal(newId) { var hls = this.hls, tracks = this.tracks; if (!Object(number["isFiniteNumber"])(newId) || newId < -1 || newId >= tracks.length) { return; } this.trackId = newId; logger["logger"].log("Switching to subtitle track " + newId); hls.trigger(events["default"].SUBTITLE_TRACK_SWITCH, { id: newId }); this._loadCurrentTrack(); }; _proto._onTextTracksChanged = function _onTextTracksChanged() { // Media is undefined when switching streams via loadSource() if (!this.media || !this.hls.config.renderTextTracksNatively) { return; } var trackId = -1; var tracks = filterSubtitleTracks(this.media.textTracks); for (var id = 0; id < tracks.length; id++) { if (tracks[id].mode === 'hidden') { // Do not break in case there is a following track with showing. trackId = id; } else if (tracks[id].mode === 'showing') { trackId = id; break; } } // Setting current subtitleTrack will invoke code. this.subtitleTrack = trackId; }; subtitle_track_controller_createClass(SubtitleTrackController, [{ key: "subtitleTracks", get: function get() { return this.tracks; } /** get index of the selected subtitle track (index in subtitle track lists) **/ }, { key: "subtitleTrack", get: function get() { return this.trackId; } /** select a subtitle track, based on its index in subtitle track lists**/ , set: function set(subtitleTrackId) { if (this.trackId !== subtitleTrackId) { this._toggleTrackModes(subtitleTrackId); this._setSubtitleTrackInternal(subtitleTrackId); } } }]); return SubtitleTrackController; }(event_handler); function filterSubtitleTracks(textTrackList) { var tracks = []; for (var i = 0; i < textTrackList.length; i++) { var track = textTrackList[i]; // Edge adds a track without a label; we don't want to use it if (track.kind === 'subtitles' && track.label) { tracks.push(textTrackList[i]); } } return tracks; } /* harmony default export */ var subtitle_track_controller = (subtitle_track_controller_SubtitleTrackController); // EXTERNAL MODULE: ./src/crypt/decrypter.js + 3 modules var decrypter = __webpack_require__("./src/crypt/decrypter.js"); // CONCATENATED MODULE: ./src/controller/subtitle-stream-controller.js function subtitle_stream_controller_assertThisInitialized(self) { if (self === void 0) { throw new ReferenceError("this hasn't been initialised - super() hasn't been called"); } return self; } function subtitle_stream_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /** * @class SubtitleStreamController */ var subtitle_stream_controller_window = window, subtitle_stream_controller_performance = subtitle_stream_controller_window.performance; var subtitle_stream_controller_TICK_INTERVAL = 500; // how often to tick in ms var subtitle_stream_controller_SubtitleStreamController = /*#__PURE__*/function (_BaseStreamController) { subtitle_stream_controller_inheritsLoose(SubtitleStreamController, _BaseStreamController); function SubtitleStreamController(hls, fragmentTracker) { var _this; _this = _BaseStreamController.call(this, hls, events["default"].MEDIA_ATTACHED, events["default"].MEDIA_DETACHING, events["default"].ERROR, events["default"].KEY_LOADED, events["default"].FRAG_LOADED, events["default"].SUBTITLE_TRACKS_UPDATED, events["default"].SUBTITLE_TRACK_SWITCH, events["default"].SUBTITLE_TRACK_LOADED, events["default"].SUBTITLE_FRAG_PROCESSED, events["default"].LEVEL_UPDATED) || this; _this.fragmentTracker = fragmentTracker; _this.config = hls.config; _this.state = State.STOPPED; _this.tracks = []; _this.tracksBuffered = []; _this.currentTrackId = -1; _this.decrypter = new decrypter["default"](hls, hls.config); // lastAVStart stores the time in seconds for the start time of a level load _this.lastAVStart = 0; _this._onMediaSeeking = _this.onMediaSeeking.bind(subtitle_stream_controller_assertThisInitialized(_this)); return _this; } var _proto = SubtitleStreamController.prototype; _proto.onSubtitleFragProcessed = function onSubtitleFragProcessed(data) { var frag = data.frag, success = data.success; this.fragPrevious = frag; this.state = State.IDLE; if (!success) { return; } var buffered = this.tracksBuffered[this.currentTrackId]; if (!buffered) { return; } // Create/update a buffered array matching the interface used by BufferHelper.bufferedInfo // so we can re-use the logic used to detect how much have been buffered var timeRange; var fragStart = frag.start; for (var i = 0; i < buffered.length; i++) { if (fragStart >= buffered[i].start && fragStart <= buffered[i].end) { timeRange = buffered[i]; break; } } var fragEnd = frag.start + frag.duration; if (timeRange) { timeRange.end = fragEnd; } else { timeRange = { start: fragStart, end: fragEnd }; buffered.push(timeRange); } }; _proto.onMediaAttached = function onMediaAttached(_ref) { var media = _ref.media; this.media = media; media.addEventListener('seeking', this._onMediaSeeking); this.state = State.IDLE; }; _proto.onMediaDetaching = function onMediaDetaching() { var _this2 = this; if (!this.media) { return; } this.media.removeEventListener('seeking', this._onMediaSeeking); this.fragmentTracker.removeAllFragments(); this.currentTrackId = -1; this.tracks.forEach(function (track) { _this2.tracksBuffered[track.id] = []; }); this.media = null; this.state = State.STOPPED; } // If something goes wrong, proceed to next frag, if we were processing one. ; _proto.onError = function onError(data) { var frag = data.frag; // don't handle error not related to subtitle fragment if (!frag || frag.type !== 'subtitle') { return; } this.state = State.IDLE; } // Got all new subtitle tracks. ; _proto.onSubtitleTracksUpdated = function onSubtitleTracksUpdated(data) { var _this3 = this; logger["logger"].log('subtitle tracks updated'); this.tracksBuffered = []; this.tracks = data.subtitleTracks; this.tracks.forEach(function (track) { _this3.tracksBuffered[track.id] = []; }); }; _proto.onSubtitleTrackSwitch = function onSubtitleTrackSwitch(data) { this.currentTrackId = data.id; if (!this.tracks || !this.tracks.length || this.currentTrackId === -1) { this.clearInterval(); return; } // Check if track has the necessary details to load fragments var currentTrack = this.tracks[this.currentTrackId]; if (currentTrack && currentTrack.details) { this.setInterval(subtitle_stream_controller_TICK_INTERVAL); } } // Got a new set of subtitle fragments. ; _proto.onSubtitleTrackLoaded = function onSubtitleTrackLoaded(data) { var id = data.id, details = data.details; var currentTrackId = this.currentTrackId, tracks = this.tracks; var currentTrack = tracks[currentTrackId]; if (id >= tracks.length || id !== currentTrackId || !currentTrack) { return; } if (details.live) { mergeSubtitlePlaylists(currentTrack.details, details, this.lastAVStart); } currentTrack.details = details; this.setInterval(subtitle_stream_controller_TICK_INTERVAL); }; _proto.onKeyLoaded = function onKeyLoaded() { if (this.state === State.KEY_LOADING) { this.state = State.IDLE; } }; _proto.onFragLoaded = function onFragLoaded(data) { var fragCurrent = this.fragCurrent; var decryptData = data.frag.decryptdata; var fragLoaded = data.frag; var hls = this.hls; if (this.state === State.FRAG_LOADING && fragCurrent && data.frag.type === 'subtitle' && fragCurrent.sn === data.frag.sn) { // check to see if the payload needs to be decrypted if (data.payload.byteLength > 0 && decryptData && decryptData.key && decryptData.method === 'AES-128') { var startTime = subtitle_stream_controller_performance.now(); // decrypt the subtitles this.decrypter.decrypt(data.payload, decryptData.key.buffer, decryptData.iv.buffer, function (decryptedData) { var endTime = subtitle_stream_controller_performance.now(); hls.trigger(events["default"].FRAG_DECRYPTED, { frag: fragLoaded, payload: decryptedData, stats: { tstart: startTime, tdecrypt: endTime } }); }); } } }; _proto.onLevelUpdated = function onLevelUpdated(_ref2) { var details = _ref2.details; var frags = details.fragments; this.lastAVStart = frags.length ? frags[0].start : 0; }; _proto.doTick = function doTick() { if (!this.media) { this.state = State.IDLE; return; } switch (this.state) { case State.IDLE: { var config = this.config, currentTrackId = this.currentTrackId, fragmentTracker = this.fragmentTracker, media = this.media, tracks = this.tracks; if (!tracks || !tracks[currentTrackId] || !tracks[currentTrackId].details) { break; } var maxBufferHole = config.maxBufferHole, maxFragLookUpTolerance = config.maxFragLookUpTolerance; var maxConfigBuffer = Math.min(config.maxBufferLength, config.maxMaxBufferLength); var bufferedInfo = BufferHelper.bufferedInfo(this._getBuffered(), media.currentTime, maxBufferHole); var bufferEnd = bufferedInfo.end, bufferLen = bufferedInfo.len; var trackDetails = tracks[currentTrackId].details; var fragments = trackDetails.fragments; var fragLen = fragments.length; var end = fragments[fragLen - 1].start + fragments[fragLen - 1].duration; if (bufferLen > maxConfigBuffer) { return; } var foundFrag; var fragPrevious = this.fragPrevious; if (bufferEnd < end) { if (fragPrevious && trackDetails.hasProgramDateTime) { foundFrag = findFragmentByPDT(fragments, fragPrevious.endProgramDateTime, maxFragLookUpTolerance); } if (!foundFrag) { foundFrag = findFragmentByPTS(fragPrevious, fragments, bufferEnd, maxFragLookUpTolerance); } } else { foundFrag = fragments[fragLen - 1]; } if (foundFrag && foundFrag.encrypted) { logger["logger"].log("Loading key for " + foundFrag.sn); this.state = State.KEY_LOADING; this.hls.trigger(events["default"].KEY_LOADING, { frag: foundFrag }); } else if (foundFrag && fragmentTracker.getState(foundFrag) === FragmentState.NOT_LOADED) { // only load if fragment is not loaded this.fragCurrent = foundFrag; this.state = State.FRAG_LOADING; this.hls.trigger(events["default"].FRAG_LOADING, { frag: foundFrag }); } } } }; _proto.stopLoad = function stopLoad() { this.lastAVStart = 0; _BaseStreamController.prototype.stopLoad.call(this); }; _proto._getBuffered = function _getBuffered() { return this.tracksBuffered[this.currentTrackId] || []; }; _proto.onMediaSeeking = function onMediaSeeking() { this.fragPrevious = null; }; return SubtitleStreamController; }(base_stream_controller_BaseStreamController); // CONCATENATED MODULE: ./src/utils/mediakeys-helper.ts /** * @see https://developer.mozilla.org/en-US/docs/Web/API/Navigator/requestMediaKeySystemAccess */ var KeySystems; (function (KeySystems) { KeySystems["WIDEVINE"] = "com.widevine.alpha"; KeySystems["PLAYREADY"] = "com.microsoft.playready"; })(KeySystems || (KeySystems = {})); var requestMediaKeySystemAccess = function () { if (typeof window !== 'undefined' && window.navigator && window.navigator.requestMediaKeySystemAccess) { return window.navigator.requestMediaKeySystemAccess.bind(window.navigator); } else { return null; } }(); // CONCATENATED MODULE: ./src/controller/eme-controller.ts function eme_controller_defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } function eme_controller_createClass(Constructor, protoProps, staticProps) { if (protoProps) eme_controller_defineProperties(Constructor.prototype, protoProps); if (staticProps) eme_controller_defineProperties(Constructor, staticProps); return Constructor; } function eme_controller_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /** * @author Stephan Hesse <[email protected]> | <[email protected]> * * DRM support for Hls.js */ var MAX_LICENSE_REQUEST_FAILURES = 3; /** * @see https://developer.mozilla.org/en-US/docs/Web/API/MediaKeySystemConfiguration * @param {Array<string>} audioCodecs List of required audio codecs to support * @param {Array<string>} videoCodecs List of required video codecs to support * @param {object} drmSystemOptions Optional parameters/requirements for the key-system * @returns {Array<MediaSystemConfiguration>} An array of supported configurations */ var createWidevineMediaKeySystemConfigurations = function createWidevineMediaKeySystemConfigurations(audioCodecs, videoCodecs, drmSystemOptions) { /* jshint ignore:line */ var baseConfig = { // initDataTypes: ['keyids', 'mp4'], // label: "", // persistentState: "not-allowed", // or "required" ? // distinctiveIdentifier: "not-allowed", // or "required" ? // sessionTypes: ['temporary'], audioCapabilities: [], // { contentType: 'audio/mp4; codecs="mp4a.40.2"' } videoCapabilities: [] // { contentType: 'video/mp4; codecs="avc1.42E01E"' } }; audioCodecs.forEach(function (codec) { baseConfig.audioCapabilities.push({ contentType: "audio/mp4; codecs=\"" + codec + "\"", robustness: drmSystemOptions.audioRobustness || '' }); }); videoCodecs.forEach(function (codec) { baseConfig.videoCapabilities.push({ contentType: "video/mp4; codecs=\"" + codec + "\"", robustness: drmSystemOptions.videoRobustness || '' }); }); return [baseConfig]; }; /** * The idea here is to handle key-system (and their respective platforms) specific configuration differences * in order to work with the local requestMediaKeySystemAccess method. * * We can also rule-out platform-related key-system support at this point by throwing an error. * * @param {string} keySystem Identifier for the key-system, see `KeySystems` enum * @param {Array<string>} audioCodecs List of required audio codecs to support * @param {Array<string>} videoCodecs List of required video codecs to support * @throws will throw an error if a unknown key system is passed * @returns {Array<MediaSystemConfiguration>} A non-empty Array of MediaKeySystemConfiguration objects */ var eme_controller_getSupportedMediaKeySystemConfigurations = function getSupportedMediaKeySystemConfigurations(keySystem, audioCodecs, videoCodecs, drmSystemOptions) { switch (keySystem) { case KeySystems.WIDEVINE: return createWidevineMediaKeySystemConfigurations(audioCodecs, videoCodecs, drmSystemOptions); default: throw new Error("Unknown key-system: " + keySystem); } }; /** * Controller to deal with encrypted media extensions (EME) * @see https://developer.mozilla.org/en-US/docs/Web/API/Encrypted_Media_Extensions_API * * @class * @constructor */ var eme_controller_EMEController = /*#__PURE__*/function (_EventHandler) { eme_controller_inheritsLoose(EMEController, _EventHandler); /** * @constructs * @param {Hls} hls Our Hls.js instance */ function EMEController(hls) { var _this; _this = _EventHandler.call(this, hls, events["default"].MEDIA_ATTACHED, events["default"].MEDIA_DETACHED, events["default"].MANIFEST_PARSED) || this; _this._widevineLicenseUrl = void 0; _this._licenseXhrSetup = void 0; _this._emeEnabled = void 0; _this._requestMediaKeySystemAccess = void 0; _this._drmSystemOptions = void 0; _this._config = void 0; _this._mediaKeysList = []; _this._media = null; _this._hasSetMediaKeys = false; _this._requestLicenseFailureCount = 0; _this.mediaKeysPromise = null; _this._onMediaEncrypted = function (e) { logger["logger"].log("Media is encrypted using \"" + e.initDataType + "\" init data type"); if (!_this.mediaKeysPromise) { logger["logger"].error('Fatal: Media is encrypted but no CDM access or no keys have been requested'); _this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].KEY_SYSTEM_ERROR, details: errors["ErrorDetails"].KEY_SYSTEM_NO_KEYS, fatal: true }); return; } var finallySetKeyAndStartSession = function finallySetKeyAndStartSession(mediaKeys) { if (!_this._media) { return; } _this._attemptSetMediaKeys(mediaKeys); _this._generateRequestWithPreferredKeySession(e.initDataType, e.initData); }; // Could use `Promise.finally` but some Promise polyfills are missing it _this.mediaKeysPromise.then(finallySetKeyAndStartSession).catch(finallySetKeyAndStartSession); }; _this._config = hls.config; _this._widevineLicenseUrl = _this._config.widevineLicenseUrl; _this._licenseXhrSetup = _this._config.licenseXhrSetup; _this._emeEnabled = _this._config.emeEnabled; _this._requestMediaKeySystemAccess = _this._config.requestMediaKeySystemAccessFunc; _this._drmSystemOptions = hls.config.drmSystemOptions; return _this; } /** * @param {string} keySystem Identifier for the key-system, see `KeySystems` enum * @returns {string} License server URL for key-system (if any configured, otherwise causes error) * @throws if a unsupported keysystem is passed */ var _proto = EMEController.prototype; _proto.getLicenseServerUrl = function getLicenseServerUrl(keySystem) { switch (keySystem) { case KeySystems.WIDEVINE: if (!this._widevineLicenseUrl) { break; } return this._widevineLicenseUrl; } throw new Error("no license server URL configured for key-system \"" + keySystem + "\""); } /** * Requests access object and adds it to our list upon success * @private * @param {string} keySystem System ID (see `KeySystems`) * @param {Array<string>} audioCodecs List of required audio codecs to support * @param {Array<string>} videoCodecs List of required video codecs to support * @throws When a unsupported KeySystem is passed */ ; _proto._attemptKeySystemAccess = function _attemptKeySystemAccess(keySystem, audioCodecs, videoCodecs) { var _this2 = this; // This can throw, but is caught in event handler callpath var mediaKeySystemConfigs = eme_controller_getSupportedMediaKeySystemConfigurations(keySystem, audioCodecs, videoCodecs, this._drmSystemOptions); logger["logger"].log('Requesting encrypted media key-system access'); // expecting interface like window.navigator.requestMediaKeySystemAccess var keySystemAccessPromise = this.requestMediaKeySystemAccess(keySystem, mediaKeySystemConfigs); this.mediaKeysPromise = keySystemAccessPromise.then(function (mediaKeySystemAccess) { return _this2._onMediaKeySystemAccessObtained(keySystem, mediaKeySystemAccess); }); keySystemAccessPromise.catch(function (err) { logger["logger"].error("Failed to obtain key-system \"" + keySystem + "\" access:", err); }); }; /** * Handles obtaining access to a key-system * @private * @param {string} keySystem * @param {MediaKeySystemAccess} mediaKeySystemAccess https://developer.mozilla.org/en-US/docs/Web/API/MediaKeySystemAccess */ _proto._onMediaKeySystemAccessObtained = function _onMediaKeySystemAccessObtained(keySystem, mediaKeySystemAccess) { var _this3 = this; logger["logger"].log("Access for key-system \"" + keySystem + "\" obtained"); var mediaKeysListItem = { mediaKeysSessionInitialized: false, mediaKeySystemAccess: mediaKeySystemAccess, mediaKeySystemDomain: keySystem }; this._mediaKeysList.push(mediaKeysListItem); var mediaKeysPromise = Promise.resolve().then(function () { return mediaKeySystemAccess.createMediaKeys(); }).then(function (mediaKeys) { mediaKeysListItem.mediaKeys = mediaKeys; logger["logger"].log("Media-keys created for key-system \"" + keySystem + "\""); _this3._onMediaKeysCreated(); return mediaKeys; }); mediaKeysPromise.catch(function (err) { logger["logger"].error('Failed to create media-keys:', err); }); return mediaKeysPromise; } /** * Handles key-creation (represents access to CDM). We are going to create key-sessions upon this * for all existing keys where no session exists yet. * * @private */ ; _proto._onMediaKeysCreated = function _onMediaKeysCreated() { var _this4 = this; // check for all key-list items if a session exists, otherwise, create one this._mediaKeysList.forEach(function (mediaKeysListItem) { if (!mediaKeysListItem.mediaKeysSession) { // mediaKeys is definitely initialized here mediaKeysListItem.mediaKeysSession = mediaKeysListItem.mediaKeys.createSession(); _this4._onNewMediaKeySession(mediaKeysListItem.mediaKeysSession); } }); } /** * @private * @param {*} keySession */ ; _proto._onNewMediaKeySession = function _onNewMediaKeySession(keySession) { var _this5 = this; logger["logger"].log("New key-system session " + keySession.sessionId); keySession.addEventListener('message', function (event) { _this5._onKeySessionMessage(keySession, event.message); }, false); } /** * @private * @param {MediaKeySession} keySession * @param {ArrayBuffer} message */ ; _proto._onKeySessionMessage = function _onKeySessionMessage(keySession, message) { logger["logger"].log('Got EME message event, creating license request'); this._requestLicense(message, function (data) { logger["logger"].log("Received license data (length: " + (data ? data.byteLength : data) + "), updating key-session"); keySession.update(data); }); } /** * @private * @param e {MediaEncryptedEvent} */ ; /** * @private */ _proto._attemptSetMediaKeys = function _attemptSetMediaKeys(mediaKeys) { if (!this._media) { throw new Error('Attempted to set mediaKeys without first attaching a media element'); } if (!this._hasSetMediaKeys) { // FIXME: see if we can/want/need-to really to deal with several potential key-sessions? var keysListItem = this._mediaKeysList[0]; if (!keysListItem || !keysListItem.mediaKeys) { logger["logger"].error('Fatal: Media is encrypted but no CDM access or no keys have been obtained yet'); this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].KEY_SYSTEM_ERROR, details: errors["ErrorDetails"].KEY_SYSTEM_NO_KEYS, fatal: true }); return; } logger["logger"].log('Setting keys for encrypted media'); this._media.setMediaKeys(keysListItem.mediaKeys); this._hasSetMediaKeys = true; } } /** * @private */ ; _proto._generateRequestWithPreferredKeySession = function _generateRequestWithPreferredKeySession(initDataType, initData) { var _this6 = this; // FIXME: see if we can/want/need-to really to deal with several potential key-sessions? var keysListItem = this._mediaKeysList[0]; if (!keysListItem) { logger["logger"].error('Fatal: Media is encrypted but not any key-system access has been obtained yet'); this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].KEY_SYSTEM_ERROR, details: errors["ErrorDetails"].KEY_SYSTEM_NO_ACCESS, fatal: true }); return; } if (keysListItem.mediaKeysSessionInitialized) { logger["logger"].warn('Key-Session already initialized but requested again'); return; } var keySession = keysListItem.mediaKeysSession; if (!keySession) { logger["logger"].error('Fatal: Media is encrypted but no key-session existing'); this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].KEY_SYSTEM_ERROR, details: errors["ErrorDetails"].KEY_SYSTEM_NO_SESSION, fatal: true }); return; } // initData is null if the media is not CORS-same-origin if (!initData) { logger["logger"].warn('Fatal: initData required for generating a key session is null'); this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].KEY_SYSTEM_ERROR, details: errors["ErrorDetails"].KEY_SYSTEM_NO_INIT_DATA, fatal: true }); return; } logger["logger"].log("Generating key-session request for \"" + initDataType + "\" init data type"); keysListItem.mediaKeysSessionInitialized = true; keySession.generateRequest(initDataType, initData).then(function () { logger["logger"].debug('Key-session generation succeeded'); }).catch(function (err) { logger["logger"].error('Error generating key-session request:', err); _this6.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].KEY_SYSTEM_ERROR, details: errors["ErrorDetails"].KEY_SYSTEM_NO_SESSION, fatal: false }); }); } /** * @private * @param {string} url License server URL * @param {ArrayBuffer} keyMessage Message data issued by key-system * @param {function} callback Called when XHR has succeeded * @returns {XMLHttpRequest} Unsent (but opened state) XHR object * @throws if XMLHttpRequest construction failed */ ; _proto._createLicenseXhr = function _createLicenseXhr(url, keyMessage, callback) { var xhr = new XMLHttpRequest(); var licenseXhrSetup = this._licenseXhrSetup; try { if (licenseXhrSetup) { try { licenseXhrSetup(xhr, url); } catch (e) { // let's try to open before running setup xhr.open('POST', url, true); licenseXhrSetup(xhr, url); } } // if licenseXhrSetup did not yet call open, let's do it now if (!xhr.readyState) { xhr.open('POST', url, true); } } catch (e) { // IE11 throws an exception on xhr.open if attempting to access an HTTP resource over HTTPS throw new Error("issue setting up KeySystem license XHR " + e); } // Because we set responseType to ArrayBuffer here, callback is typed as handling only array buffers xhr.responseType = 'arraybuffer'; xhr.onreadystatechange = this._onLicenseRequestReadyStageChange.bind(this, xhr, url, keyMessage, callback); return xhr; } /** * @private * @param {XMLHttpRequest} xhr * @param {string} url License server URL * @param {ArrayBuffer} keyMessage Message data issued by key-system * @param {function} callback Called when XHR has succeeded */ ; _proto._onLicenseRequestReadyStageChange = function _onLicenseRequestReadyStageChange(xhr, url, keyMessage, callback) { switch (xhr.readyState) { case 4: if (xhr.status === 200) { this._requestLicenseFailureCount = 0; logger["logger"].log('License request succeeded'); if (xhr.responseType !== 'arraybuffer') { logger["logger"].warn('xhr response type was not set to the expected arraybuffer for license request'); } callback(xhr.response); } else { logger["logger"].error("License Request XHR failed (" + url + "). Status: " + xhr.status + " (" + xhr.statusText + ")"); this._requestLicenseFailureCount++; if (this._requestLicenseFailureCount > MAX_LICENSE_REQUEST_FAILURES) { this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].KEY_SYSTEM_ERROR, details: errors["ErrorDetails"].KEY_SYSTEM_LICENSE_REQUEST_FAILED, fatal: true }); return; } var attemptsLeft = MAX_LICENSE_REQUEST_FAILURES - this._requestLicenseFailureCount + 1; logger["logger"].warn("Retrying license request, " + attemptsLeft + " attempts left"); this._requestLicense(keyMessage, callback); } break; } } /** * @private * @param {MediaKeysListItem} keysListItem * @param {ArrayBuffer} keyMessage * @returns {ArrayBuffer} Challenge data posted to license server * @throws if KeySystem is unsupported */ ; _proto._generateLicenseRequestChallenge = function _generateLicenseRequestChallenge(keysListItem, keyMessage) { switch (keysListItem.mediaKeySystemDomain) { // case KeySystems.PLAYREADY: // from https://github.com/MicrosoftEdge/Demos/blob/master/eme/scripts/demo.js /* if (this.licenseType !== this.LICENSE_TYPE_WIDEVINE) { // For PlayReady CDMs, we need to dig the Challenge out of the XML. var keyMessageXml = new DOMParser().parseFromString(String.fromCharCode.apply(null, new Uint16Array(keyMessage)), 'application/xml'); if (keyMessageXml.getElementsByTagName('Challenge')[0]) { challenge = atob(keyMessageXml.getElementsByTagName('Challenge')[0].childNodes[0].nodeValue); } else { throw 'Cannot find <Challenge> in key message'; } var headerNames = keyMessageXml.getElementsByTagName('name'); var headerValues = keyMessageXml.getElementsByTagName('value'); if (headerNames.length !== headerValues.length) { throw 'Mismatched header <name>/<value> pair in key message'; } for (var i = 0; i < headerNames.length; i++) { xhr.setRequestHeader(headerNames[i].childNodes[0].nodeValue, headerValues[i].childNodes[0].nodeValue); } } break; */ case KeySystems.WIDEVINE: // For Widevine CDMs, the challenge is the keyMessage. return keyMessage; } throw new Error("unsupported key-system: " + keysListItem.mediaKeySystemDomain); } /** * @private * @param keyMessage * @param callback */ ; _proto._requestLicense = function _requestLicense(keyMessage, callback) { logger["logger"].log('Requesting content license for key-system'); var keysListItem = this._mediaKeysList[0]; if (!keysListItem) { logger["logger"].error('Fatal error: Media is encrypted but no key-system access has been obtained yet'); this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].KEY_SYSTEM_ERROR, details: errors["ErrorDetails"].KEY_SYSTEM_NO_ACCESS, fatal: true }); return; } try { var _url = this.getLicenseServerUrl(keysListItem.mediaKeySystemDomain); var _xhr = this._createLicenseXhr(_url, keyMessage, callback); logger["logger"].log("Sending license request to URL: " + _url); var challenge = this._generateLicenseRequestChallenge(keysListItem, keyMessage); _xhr.send(challenge); } catch (e) { logger["logger"].error("Failure requesting DRM license: " + e); this.hls.trigger(events["default"].ERROR, { type: errors["ErrorTypes"].KEY_SYSTEM_ERROR, details: errors["ErrorDetails"].KEY_SYSTEM_LICENSE_REQUEST_FAILED, fatal: true }); } }; _proto.onMediaAttached = function onMediaAttached(data) { if (!this._emeEnabled) { return; } var media = data.media; // keep reference of media this._media = media; media.addEventListener('encrypted', this._onMediaEncrypted); }; _proto.onMediaDetached = function onMediaDetached() { var media = this._media; var mediaKeysList = this._mediaKeysList; if (!media) { return; } media.removeEventListener('encrypted', this._onMediaEncrypted); this._media = null; this._mediaKeysList = []; // Close all sessions and remove media keys from the video element. Promise.all(mediaKeysList.map(function (mediaKeysListItem) { if (mediaKeysListItem.mediaKeysSession) { return mediaKeysListItem.mediaKeysSession.close().catch(function () {// Ignore errors when closing the sessions. Closing a session that // generated no key requests will throw an error. }); } })).then(function () { return media.setMediaKeys(null); }).catch(function () {// Ignore any failures while removing media keys from the video element. }); } // TODO: Use manifest types here when they are defined ; _proto.onManifestParsed = function onManifestParsed(data) { if (!this._emeEnabled) { return; } var audioCodecs = data.levels.map(function (level) { return level.audioCodec; }); var videoCodecs = data.levels.map(function (level) { return level.videoCodec; }); this._attemptKeySystemAccess(KeySystems.WIDEVINE, audioCodecs, videoCodecs); }; eme_controller_createClass(EMEController, [{ key: "requestMediaKeySystemAccess", get: function get() { if (!this._requestMediaKeySystemAccess) { throw new Error('No requestMediaKeySystemAccess function configured'); } return this._requestMediaKeySystemAccess; } }]); return EMEController; }(event_handler); /* harmony default export */ var eme_controller = (eme_controller_EMEController); // CONCATENATED MODULE: ./src/config.ts function ownKeys(object, enumerableOnly) { var keys = Object.keys(object); if (Object.getOwnPropertySymbols) { var symbols = Object.getOwnPropertySymbols(object); if (enumerableOnly) symbols = symbols.filter(function (sym) { return Object.getOwnPropertyDescriptor(object, sym).enumerable; }); keys.push.apply(keys, symbols); } return keys; } function _objectSpread(target) { for (var i = 1; i < arguments.length; i++) { var source = arguments[i] != null ? arguments[i] : {}; if (i % 2) { ownKeys(Object(source), true).forEach(function (key) { _defineProperty(target, key, source[key]); }); } else if (Object.getOwnPropertyDescriptors) { Object.defineProperties(target, Object.getOwnPropertyDescriptors(source)); } else { ownKeys(Object(source)).forEach(function (key) { Object.defineProperty(target, key, Object.getOwnPropertyDescriptor(source, key)); }); } } return target; } function _defineProperty(obj, key, value) { if (key in obj) { Object.defineProperty(obj, key, { value: value, enumerable: true, configurable: true, writable: true }); } else { obj[key] = value; } return obj; } /** * HLS config */ // import FetchLoader from './utils/fetch-loader'; // If possible, keep hlsDefaultConfig shallow // It is cloned whenever a new Hls instance is created, by keeping the config // shallow the properties are cloned, and we don't end up manipulating the default var hlsDefaultConfig = _objectSpread(_objectSpread({ autoStartLoad: true, // used by stream-controller startPosition: -1, // used by stream-controller defaultAudioCodec: void 0, // used by stream-controller debug: false, // used by logger capLevelOnFPSDrop: false, // used by fps-controller capLevelToPlayerSize: false, // used by cap-level-controller initialLiveManifestSize: 1, // used by stream-controller maxBufferLength: 30, // used by stream-controller maxBufferSize: 60 * 1000 * 1000, // used by stream-controller maxBufferHole: 0.5, // used by stream-controller lowBufferWatchdogPeriod: 0.5, // used by stream-controller highBufferWatchdogPeriod: 3, // used by stream-controller nudgeOffset: 0.1, // used by stream-controller nudgeMaxRetry: 3, // used by stream-controller maxFragLookUpTolerance: 0.25, // used by stream-controller liveSyncDurationCount: 3, // used by stream-controller liveMaxLatencyDurationCount: Infinity, // used by stream-controller liveSyncDuration: void 0, // used by stream-controller liveMaxLatencyDuration: void 0, // used by stream-controller liveDurationInfinity: false, // used by buffer-controller liveBackBufferLength: Infinity, // used by buffer-controller maxMaxBufferLength: 600, // used by stream-controller enableWorker: true, // used by demuxer enableSoftwareAES: true, // used by decrypter manifestLoadingTimeOut: 10000, // used by playlist-loader manifestLoadingMaxRetry: 1, // used by playlist-loader manifestLoadingRetryDelay: 1000, // used by playlist-loader manifestLoadingMaxRetryTimeout: 64000, // used by playlist-loader startLevel: void 0, // used by level-controller levelLoadingTimeOut: 10000, // used by playlist-loader levelLoadingMaxRetry: 4, // used by playlist-loader levelLoadingRetryDelay: 1000, // used by playlist-loader levelLoadingMaxRetryTimeout: 64000, // used by playlist-loader fragLoadingTimeOut: 20000, // used by fragment-loader fragLoadingMaxRetry: 6, // used by fragment-loader fragLoadingRetryDelay: 1000, // used by fragment-loader fragLoadingMaxRetryTimeout: 64000, // used by fragment-loader startFragPrefetch: false, // used by stream-controller fpsDroppedMonitoringPeriod: 5000, // used by fps-controller fpsDroppedMonitoringThreshold: 0.2, // used by fps-controller appendErrorMaxRetry: 3, // used by buffer-controller loader: xhr_loader, // loader: FetchLoader, fLoader: void 0, // used by fragment-loader pLoader: void 0, // used by playlist-loader xhrSetup: void 0, // used by xhr-loader licenseXhrSetup: void 0, // used by eme-controller // fetchSetup: void 0, abrController: abr_controller, bufferController: buffer_controller, capLevelController: cap_level_controller, fpsController: fps_controller, stretchShortVideoTrack: false, // used by mp4-remuxer maxAudioFramesDrift: 1, // used by mp4-remuxer forceKeyFrameOnDiscontinuity: true, // used by ts-demuxer abrEwmaFastLive: 3, // used by abr-controller abrEwmaSlowLive: 9, // used by abr-controller abrEwmaFastVoD: 3, // used by abr-controller abrEwmaSlowVoD: 9, // used by abr-controller abrEwmaDefaultEstimate: 5e5, // 500 kbps // used by abr-controller abrBandWidthFactor: 0.95, // used by abr-controller abrBandWidthUpFactor: 0.7, // used by abr-controller abrMaxWithRealBitrate: false, // used by abr-controller maxStarvationDelay: 4, // used by abr-controller maxLoadingDelay: 4, // used by abr-controller minAutoBitrate: 0, // used by hls emeEnabled: false, // used by eme-controller widevineLicenseUrl: void 0, // used by eme-controller drmSystemOptions: {}, // used by eme-controller requestMediaKeySystemAccessFunc: requestMediaKeySystemAccess, // used by eme-controller testBandwidth: true }, timelineConfig()), {}, { subtitleStreamController: true ? subtitle_stream_controller_SubtitleStreamController : undefined, subtitleTrackController: true ? subtitle_track_controller : undefined, timelineController: true ? timeline_controller : undefined, audioStreamController: true ? audio_stream_controller : undefined, audioTrackController: true ? audio_track_controller : undefined, emeController: true ? eme_controller : undefined }); function timelineConfig() { return { cueHandler: cues_namespaceObject, // used by timeline-controller enableCEA708Captions: true, // used by timeline-controller enableWebVTT: true, // used by timeline-controller captionsTextTrack1Label: 'English', // used by timeline-controller captionsTextTrack1LanguageCode: 'en', // used by timeline-controller captionsTextTrack2Label: 'Spanish', // used by timeline-controller captionsTextTrack2LanguageCode: 'es', // used by timeline-controller captionsTextTrack3Label: 'Unknown CC', // used by timeline-controller captionsTextTrack3LanguageCode: '', // used by timeline-controller captionsTextTrack4Label: 'Unknown CC', // used by timeline-controller captionsTextTrack4LanguageCode: '', // used by timeline-controller renderTextTracksNatively: true }; } // CONCATENATED MODULE: ./src/hls.ts function hls_ownKeys(object, enumerableOnly) { var keys = Object.keys(object); if (Object.getOwnPropertySymbols) { var symbols = Object.getOwnPropertySymbols(object); if (enumerableOnly) symbols = symbols.filter(function (sym) { return Object.getOwnPropertyDescriptor(object, sym).enumerable; }); keys.push.apply(keys, symbols); } return keys; } function hls_objectSpread(target) { for (var i = 1; i < arguments.length; i++) { var source = arguments[i] != null ? arguments[i] : {}; if (i % 2) { hls_ownKeys(Object(source), true).forEach(function (key) { hls_defineProperty(target, key, source[key]); }); } else if (Object.getOwnPropertyDescriptors) { Object.defineProperties(target, Object.getOwnPropertyDescriptors(source)); } else { hls_ownKeys(Object(source)).forEach(function (key) { Object.defineProperty(target, key, Object.getOwnPropertyDescriptor(source, key)); }); } } return target; } function hls_defineProperty(obj, key, value) { if (key in obj) { Object.defineProperty(obj, key, { value: value, enumerable: true, configurable: true, writable: true }); } else { obj[key] = value; } return obj; } function hls_assertThisInitialized(self) { if (self === void 0) { throw new ReferenceError("this hasn't been initialised - super() hasn't been called"); } return self; } function hls_defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } } function hls_createClass(Constructor, protoProps, staticProps) { if (protoProps) hls_defineProperties(Constructor.prototype, protoProps); if (staticProps) hls_defineProperties(Constructor, staticProps); return Constructor; } function hls_inheritsLoose(subClass, superClass) { subClass.prototype = Object.create(superClass.prototype); subClass.prototype.constructor = subClass; subClass.__proto__ = superClass; } /** * @module Hls * @class * @constructor */ var hls_Hls = /*#__PURE__*/function (_Observer) { hls_inheritsLoose(Hls, _Observer); /** * @type {boolean} */ Hls.isSupported = function isSupported() { return is_supported_isSupported(); } /** * @type {HlsEvents} */ ; hls_createClass(Hls, null, [{ key: "version", /** * @type {string} */ get: function get() { return "0.14.4-0.alpha.5747"; } }, { key: "Events", get: function get() { return events["default"]; } /** * @type {HlsErrorTypes} */ }, { key: "ErrorTypes", get: function get() { return errors["ErrorTypes"]; } /** * @type {HlsErrorDetails} */ }, { key: "ErrorDetails", get: function get() { return errors["ErrorDetails"]; } /** * @type {HlsConfig} */ }, { key: "DefaultConfig", get: function get() { if (!Hls.defaultConfig) { return hlsDefaultConfig; } return Hls.defaultConfig; } /** * @type {HlsConfig} */ , set: function set(defaultConfig) { Hls.defaultConfig = defaultConfig; } /** * Creates an instance of an HLS client that can attach to exactly one `HTMLMediaElement`. * * @constructs Hls * @param {HlsConfig} config */ }]); function Hls(userConfig) { var _this; if (userConfig === void 0) { userConfig = {}; } _this = _Observer.call(this) || this; _this.config = void 0; _this._autoLevelCapping = void 0; _this.abrController = void 0; _this.capLevelController = void 0; _this.levelController = void 0; _this.streamController = void 0; _this.networkControllers = void 0; _this.audioTrackController = void 0; _this.subtitleTrackController = void 0; _this.emeController = void 0; _this.coreComponents = void 0; _this.media = null; _this.url = null; var defaultConfig = Hls.DefaultConfig; if ((userConfig.liveSyncDurationCount || userConfig.liveMaxLatencyDurationCount) && (userConfig.liveSyncDuration || userConfig.liveMaxLatencyDuration)) { throw new Error('Illegal hls.js config: don\'t mix up liveSyncDurationCount/liveMaxLatencyDurationCount and liveSyncDuration/liveMaxLatencyDuration'); } // Shallow clone _this.config = hls_objectSpread(hls_objectSpread({}, defaultConfig), userConfig); var _assertThisInitialize = hls_assertThisInitialized(_this), config = _assertThisInitialize.config; if (config.liveMaxLatencyDurationCount !== void 0 && config.liveMaxLatencyDurationCount <= config.liveSyncDurationCount) { throw new Error('Illegal hls.js config: "liveMaxLatencyDurationCount" must be gt "liveSyncDurationCount"'); } if (config.liveMaxLatencyDuration !== void 0 && (config.liveSyncDuration === void 0 || config.liveMaxLatencyDuration <= config.liveSyncDuration)) { throw new Error('Illegal hls.js config: "liveMaxLatencyDuration" must be gt "liveSyncDuration"'); } Object(logger["enableLogs"])(config.debug); _this._autoLevelCapping = -1; // core controllers and network loaders /** * @member {AbrController} abrController */ var abrController = _this.abrController = new config.abrController(hls_assertThisInitialized(_this)); // eslint-disable-line new-cap var bufferController = new config.bufferController(hls_assertThisInitialized(_this)); // eslint-disable-line new-cap var capLevelController = _this.capLevelController = new config.capLevelController(hls_assertThisInitialized(_this)); // eslint-disable-line new-cap var fpsController = new config.fpsController(hls_assertThisInitialized(_this)); // eslint-disable-line new-cap var playListLoader = new playlist_loader(hls_assertThisInitialized(_this)); var fragmentLoader = new fragment_loader(hls_assertThisInitialized(_this)); var keyLoader = new key_loader(hls_assertThisInitialized(_this)); var id3TrackController = new id3_track_controller(hls_assertThisInitialized(_this)); // network controllers /** * @member {LevelController} levelController */ var levelController = _this.levelController = new level_controller_LevelController(hls_assertThisInitialized(_this)); // FIXME: FragmentTracker must be defined before StreamController because the order of event handling is important var fragmentTracker = new fragment_tracker_FragmentTracker(hls_assertThisInitialized(_this)); /** * @member {StreamController} streamController */ var streamController = _this.streamController = new stream_controller(hls_assertThisInitialized(_this), fragmentTracker); var networkControllers = [levelController, streamController]; // optional audio stream controller /** * @var {ICoreComponent | Controller} */ var Controller = config.audioStreamController; if (Controller) { networkControllers.push(new Controller(hls_assertThisInitialized(_this), fragmentTracker)); } /** * @member {INetworkController[]} networkControllers */ _this.networkControllers = networkControllers; /** * @var {ICoreComponent[]} */ var coreComponents = [playListLoader, fragmentLoader, keyLoader, abrController, bufferController, capLevelController, fpsController, id3TrackController, fragmentTracker]; // optional audio track and subtitle controller Controller = config.audioTrackController; if (Controller) { var audioTrackController = new Controller(hls_assertThisInitialized(_this)); /** * @member {AudioTrackController} audioTrackController */ _this.audioTrackController = audioTrackController; coreComponents.push(audioTrackController); } Controller = config.subtitleTrackController; if (Controller) { var subtitleTrackController = new Controller(hls_assertThisInitialized(_this)); /** * @member {SubtitleTrackController} subtitleTrackController */ _this.subtitleTrackController = subtitleTrackController; networkControllers.push(subtitleTrackController); } Controller = config.emeController; if (Controller) { var emeController = new Controller(hls_assertThisInitialized(_this)); /** * @member {EMEController} emeController */ _this.emeController = emeController; coreComponents.push(emeController); } // optional subtitle controllers Controller = config.subtitleStreamController; if (Controller) { networkControllers.push(new Controller(hls_assertThisInitialized(_this), fragmentTracker)); } Controller = config.timelineController; if (Controller) { coreComponents.push(new Controller(hls_assertThisInitialized(_this))); } /** * @member {ICoreComponent[]} */ _this.coreComponents = coreComponents; return _this; } /** * Dispose of the instance */ var _proto = Hls.prototype; _proto.destroy = function destroy() { logger["logger"].log('destroy'); this.trigger(events["default"].DESTROYING); this.detachMedia(); this.coreComponents.concat(this.networkControllers).forEach(function (component) { component.destroy(); }); this.url = null; this.removeAllListeners(); this._autoLevelCapping = -1; } /** * Attach a media element * @param {HTMLMediaElement} media */ ; _proto.attachMedia = function attachMedia(media) { logger["logger"].log('attachMedia'); this.media = media; this.trigger(events["default"].MEDIA_ATTACHING, { media: media }); } /** * Detach from the media */ ; _proto.detachMedia = function detachMedia() { logger["logger"].log('detachMedia'); this.trigger(events["default"].MEDIA_DETACHING); this.media = null; } /** * Set the source URL. Can be relative or absolute. * @param {string} url */ ; _proto.loadSource = function loadSource(url) { url = url_toolkit["buildAbsoluteURL"](window.location.href, url, { alwaysNormalize: true }); logger["logger"].log("loadSource:" + url); this.url = url; // when attaching to a source URL, trigger a playlist load this.trigger(events["default"].MANIFEST_LOADING, { url: url }); } /** * Start loading data from the stream source. * Depending on default config, client starts loading automatically when a source is set. * * @param {number} startPosition Set the start position to stream from * @default -1 None (from earliest point) */ ; _proto.startLoad = function startLoad(startPosition) { if (startPosition === void 0) { startPosition = -1; } logger["logger"].log("startLoad(" + startPosition + ")"); this.networkControllers.forEach(function (controller) { controller.startLoad(startPosition); }); } /** * Stop loading of any stream data. */ ; _proto.stopLoad = function stopLoad() { logger["logger"].log('stopLoad'); this.networkControllers.forEach(function (controller) { controller.stopLoad(); }); } /** * Swap through possible audio codecs in the stream (for example to switch from stereo to 5.1) */ ; _proto.swapAudioCodec = function swapAudioCodec() { logger["logger"].log('swapAudioCodec'); this.streamController.swapAudioCodec(); } /** * When the media-element fails, this allows to detach and then re-attach it * as one call (convenience method). * * Automatic recovery of media-errors by this process is configurable. */ ; _proto.recoverMediaError = function recoverMediaError() { logger["logger"].log('recoverMediaError'); var media = this.media; this.detachMedia(); if (media) { this.attachMedia(media); } } /** * Remove a loaded level from the list of levels, or a level url in from a list of redundant level urls. * This can be used to remove a rendition or playlist url that errors frequently from the list of levels that a user * or hls.js can choose from. * * @param levelIndex {number} The quality level index to of the level to remove * @param urlId {number} The quality level url index in the case that fallback levels are available. Defaults to 0. */ ; _proto.removeLevel = function removeLevel(levelIndex, urlId) { if (urlId === void 0) { urlId = 0; } this.levelController.removeLevel(levelIndex, urlId); } /** * @type {QualityLevel[]} */ // todo(typescript-levelController) ; hls_createClass(Hls, [{ key: "levels", get: function get() { return this.levelController.levels; } /** * Index of quality level currently played * @type {number} */ }, { key: "currentLevel", get: function get() { return this.streamController.currentLevel; } /** * Set quality level index immediately . * This will flush the current buffer to replace the quality asap. * That means playback will interrupt at least shortly to re-buffer and re-sync eventually. * @param newLevel {number} -1 for automatic level selection */ , set: function set(newLevel) { logger["logger"].log("set currentLevel:" + newLevel); this.loadLevel = newLevel; this.streamController.immediateLevelSwitch(); } /** * Index of next quality level loaded as scheduled by stream controller. * @type {number} */ }, { key: "nextLevel", get: function get() { return this.streamController.nextLevel; } /** * Set quality level index for next loaded data. * This will switch the video quality asap, without interrupting playback. * May abort current loading of data, and flush parts of buffer (outside currently played fragment region). * @type {number} -1 for automatic level selection */ , set: function set(newLevel) { logger["logger"].log("set nextLevel:" + newLevel); this.levelController.manualLevel = newLevel; this.streamController.nextLevelSwitch(); } /** * Return the quality level of the currently or last (of none is loaded currently) segment * @type {number} */ }, { key: "loadLevel", get: function get() { return this.levelController.level; } /** * Set quality level index for next loaded data in a conservative way. * This will switch the quality without flushing, but interrupt current loading. * Thus the moment when the quality switch will appear in effect will only be after the already existing buffer. * @type {number} newLevel -1 for automatic level selection */ , set: function set(newLevel) { logger["logger"].log("set loadLevel:" + newLevel); this.levelController.manualLevel = newLevel; } /** * get next quality level loaded * @type {number} */ }, { key: "nextLoadLevel", get: function get() { return this.levelController.nextLoadLevel; } /** * Set quality level of next loaded segment in a fully "non-destructive" way. * Same as `loadLevel` but will wait for next switch (until current loading is done). * @type {number} level */ , set: function set(level) { this.levelController.nextLoadLevel = level; } /** * Return "first level": like a default level, if not set, * falls back to index of first level referenced in manifest * @type {number} */ }, { key: "firstLevel", get: function get() { return Math.max(this.levelController.firstLevel, this.minAutoLevel); } /** * Sets "first-level", see getter. * @type {number} */ , set: function set(newLevel) { logger["logger"].log("set firstLevel:" + newLevel); this.levelController.firstLevel = newLevel; } /** * Return start level (level of first fragment that will be played back) * if not overrided by user, first level appearing in manifest will be used as start level * if -1 : automatic start level selection, playback will start from level matching download bandwidth * (determined from download of first segment) * @type {number} */ }, { key: "startLevel", get: function get() { return this.levelController.startLevel; } /** * set start level (level of first fragment that will be played back) * if not overrided by user, first level appearing in manifest will be used as start level * if -1 : automatic start level selection, playback will start from level matching download bandwidth * (determined from download of first segment) * @type {number} newLevel */ , set: function set(newLevel) { logger["logger"].log("set startLevel:" + newLevel); // if not in automatic start level detection, ensure startLevel is greater than minAutoLevel if (newLevel !== -1) { newLevel = Math.max(newLevel, this.minAutoLevel); } this.levelController.startLevel = newLevel; } /** * set dynamically set capLevelToPlayerSize against (`CapLevelController`) * * @type {boolean} */ }, { key: "capLevelToPlayerSize", set: function set(shouldStartCapping) { var newCapLevelToPlayerSize = !!shouldStartCapping; if (newCapLevelToPlayerSize !== this.config.capLevelToPlayerSize) { if (newCapLevelToPlayerSize) { this.capLevelController.startCapping(); // If capping occurs, nextLevelSwitch will happen based on size. } else { this.capLevelController.stopCapping(); this.autoLevelCapping = -1; this.streamController.nextLevelSwitch(); // Now we're uncapped, get the next level asap. } this.config.capLevelToPlayerSize = newCapLevelToPlayerSize; } } /** * Capping/max level value that should be used by automatic level selection algorithm (`ABRController`) * @type {number} */ }, { key: "autoLevelCapping", get: function get() { return this._autoLevelCapping; } /** * get bandwidth estimate * @type {number} */ , /** * Capping/max level value that should be used by automatic level selection algorithm (`ABRController`) * @type {number} */ set: function set(newLevel) { logger["logger"].log("set autoLevelCapping:" + newLevel); this._autoLevelCapping = newLevel; } /** * True when automatic level selection enabled * @type {boolean} */ }, { key: "bandwidthEstimate", get: function get() { var bwEstimator = this.abrController._bwEstimator; return bwEstimator ? bwEstimator.getEstimate() : NaN; } }, { key: "autoLevelEnabled", get: function get() { return this.levelController.manualLevel === -1; } /** * Level set manually (if any) * @type {number} */ }, { key: "manualLevel", get: function get() { return this.levelController.manualLevel; } /** * min level selectable in auto mode according to config.minAutoBitrate * @type {number} */ }, { key: "minAutoLevel", get: function get() { var levels = this.levels, minAutoBitrate = this.config.minAutoBitrate; var len = levels ? levels.length : 0; for (var i = 0; i < len; i++) { var levelNextBitrate = levels[i].realBitrate ? Math.max(levels[i].realBitrate, levels[i].bitrate) : levels[i].bitrate; if (levelNextBitrate > minAutoBitrate) { return i; } } return 0; } /** * max level selectable in auto mode according to autoLevelCapping * @type {number} */ }, { key: "maxAutoLevel", get: function get() { var levels = this.levels, autoLevelCapping = this.autoLevelCapping; var maxAutoLevel; if (autoLevelCapping === -1 && levels && levels.length) { maxAutoLevel = levels.length - 1; } else { maxAutoLevel = autoLevelCapping; } return maxAutoLevel; } /** * next automatically selected quality level * @type {number} */ }, { key: "nextAutoLevel", get: function get() { // ensure next auto level is between min and max auto level return Math.min(Math.max(this.abrController.nextAutoLevel, this.minAutoLevel), this.maxAutoLevel); } /** * this setter is used to force next auto level. * this is useful to force a switch down in auto mode: * in case of load error on level N, hls.js can set nextAutoLevel to N-1 for example) * forced value is valid for one fragment. upon succesful frag loading at forced level, * this value will be resetted to -1 by ABR controller. * @type {number} */ , set: function set(nextLevel) { this.abrController.nextAutoLevel = Math.max(this.minAutoLevel, nextLevel); } /** * @type {AudioTrack[]} */ // todo(typescript-audioTrackController) }, { key: "audioTracks", get: function get() { var audioTrackController = this.audioTrackController; return audioTrackController ? audioTrackController.audioTracks : []; } /** * index of the selected audio track (index in audio track lists) * @type {number} */ }, { key: "audioTrack", get: function get() { var audioTrackController = this.audioTrackController; return audioTrackController ? audioTrackController.audioTrack : -1; } /** * selects an audio track, based on its index in audio track lists * @type {number} */ , set: function set(audioTrackId) { var audioTrackController = this.audioTrackController; if (audioTrackController) { audioTrackController.audioTrack = audioTrackId; } } /** * @type {Seconds} */ }, { key: "liveSyncPosition", get: function get() { return this.streamController.liveSyncPosition; } /** * get alternate subtitle tracks list from playlist * @type {SubtitleTrack[]} */ // todo(typescript-subtitleTrackController) }, { key: "subtitleTracks", get: function get() { var subtitleTrackController = this.subtitleTrackController; return subtitleTrackController ? subtitleTrackController.subtitleTracks : []; } /** * index of the selected subtitle track (index in subtitle track lists) * @type {number} */ }, { key: "subtitleTrack", get: function get() { var subtitleTrackController = this.subtitleTrackController; return subtitleTrackController ? subtitleTrackController.subtitleTrack : -1; } /** * select an subtitle track, based on its index in subtitle track lists * @type {number} */ , set: function set(subtitleTrackId) { var subtitleTrackController = this.subtitleTrackController; if (subtitleTrackController) { subtitleTrackController.subtitleTrack = subtitleTrackId; } } /** * @type {boolean} */ }, { key: "subtitleDisplay", get: function get() { var subtitleTrackController = this.subtitleTrackController; return subtitleTrackController ? subtitleTrackController.subtitleDisplay : false; } /** * Enable/disable subtitle display rendering * @type {boolean} */ , set: function set(value) { var subtitleTrackController = this.subtitleTrackController; if (subtitleTrackController) { subtitleTrackController.subtitleDisplay = value; } } }]); return Hls; }(Observer); hls_Hls.defaultConfig = void 0; /***/ }), /***/ "./src/polyfills/number.js": /*!*********************************!*\ !*** ./src/polyfills/number.js ***! \*********************************/ /*! exports provided: isFiniteNumber, MAX_SAFE_INTEGER */ /***/ (function(module, __webpack_exports__, __webpack_require__) { "use strict"; __webpack_require__.r(__webpack_exports__); /* harmony export (binding) */ __webpack_require__.d(__webpack_exports__, "isFiniteNumber", function() { return isFiniteNumber; }); /* harmony export (binding) */ __webpack_require__.d(__webpack_exports__, "MAX_SAFE_INTEGER", function() { return MAX_SAFE_INTEGER; }); var isFiniteNumber = Number.isFinite || function (value) { return typeof value === 'number' && isFinite(value); }; var MAX_SAFE_INTEGER = Number.MAX_SAFE_INTEGER || 9007199254740991; /***/ }), /***/ "./src/utils/get-self-scope.js": /*!*************************************!*\ !*** ./src/utils/get-self-scope.js ***! \*************************************/ /*! exports provided: getSelfScope */ /***/ (function(module, __webpack_exports__, __webpack_require__) { "use strict"; __webpack_require__.r(__webpack_exports__); /* harmony export (binding) */ __webpack_require__.d(__webpack_exports__, "getSelfScope", function() { return getSelfScope; }); function getSelfScope() { // see https://stackoverflow.com/a/11237259/589493 if (typeof window === 'undefined') { /* eslint-disable-next-line no-undef */ return self; } else { return window; } } /***/ }), /***/ "./src/utils/logger.js": /*!*****************************!*\ !*** ./src/utils/logger.js ***! \*****************************/ /*! exports provided: enableLogs, logger */ /***/ (function(module, __webpack_exports__, __webpack_require__) { "use strict"; __webpack_require__.r(__webpack_exports__); /* harmony export (binding) */ __webpack_require__.d(__webpack_exports__, "enableLogs", function() { return enableLogs; }); /* harmony export (binding) */ __webpack_require__.d(__webpack_exports__, "logger", function() { return logger; }); /* harmony import */ var _get_self_scope__WEBPACK_IMPORTED_MODULE_0__ = __webpack_require__(/*! ./get-self-scope */ "./src/utils/get-self-scope.js"); function noop() {} var fakeLogger = { trace: noop, debug: noop, log: noop, warn: noop, info: noop, error: noop }; var exportedLogger = fakeLogger; // let lastCallTime; // function formatMsgWithTimeInfo(type, msg) { // const now = Date.now(); // const diff = lastCallTime ? '+' + (now - lastCallTime) : '0'; // lastCallTime = now; // msg = (new Date(now)).toISOString() + ' | [' + type + '] > ' + msg + ' ( ' + diff + ' ms )'; // return msg; // } function formatMsg(type, msg) { msg = '[' + type + '] > ' + msg; return msg; } var global = Object(_get_self_scope__WEBPACK_IMPORTED_MODULE_0__["getSelfScope"])(); function consolePrintFn(type) { var func = global.console[type]; if (func) { return function () { for (var _len = arguments.length, args = new Array(_len), _key = 0; _key < _len; _key++) { args[_key] = arguments[_key]; } if (args[0]) { args[0] = formatMsg(type, args[0]); } func.apply(global.console, args); }; } return noop; } function exportLoggerFunctions(debugConfig) { for (var _len2 = arguments.length, functions = new Array(_len2 > 1 ? _len2 - 1 : 0), _key2 = 1; _key2 < _len2; _key2++) { functions[_key2 - 1] = arguments[_key2]; } functions.forEach(function (type) { exportedLogger[type] = debugConfig[type] ? debugConfig[type].bind(debugConfig) : consolePrintFn(type); }); } var enableLogs = function enableLogs(debugConfig) { // check that console is available if (global.console && debugConfig === true || typeof debugConfig === 'object') { exportLoggerFunctions(debugConfig, // Remove out from list here to hard-disable a log-level // 'trace', 'debug', 'log', 'info', 'warn', 'error'); // Some browsers don't allow to use bind on console object anyway // fallback to default if needed try { exportedLogger.log(); } catch (e) { exportedLogger = fakeLogger; } } else { exportedLogger = fakeLogger; } }; var logger = exportedLogger; /***/ }) /******/ })["default"]; }); //# sourceMappingURL=hls.js.map<|fim▁end|>
clearCurrentCues(this.id3Track); this.id3Track = undefined; this.media = undefined;
<|file_name|>docscrape.py<|end_file_name|><|fim▁begin|>"""Extract reference documentation from the NumPy source tree. """ import inspect import textwrap import re import pydoc from io import StringIO from warnings import warn import collections class Reader(object): """A line-based string reader. """ def __init__(self, data): """ Parameters ---------- data : str String with lines separated by '\n'. """ if isinstance(data, list): self._str = data else: self._str = data.split('\n') # store string as list of lines self.reset() def __getitem__(self, n): return self._str[n] def reset(self): self._l = 0 # current line nr def read(self): if not self.eof(): out = self[self._l] self._l += 1 return out else: return '' def seek_next_non_empty_line(self): for l in self[self._l:]: if l.strip(): break else: self._l += 1 def eof(self): return self._l >= len(self._str) def read_to_condition(self, condition_func): start = self._l for line in self[start:]: if condition_func(line): return self[start:self._l] self._l += 1 if self.eof(): return self[start:self._l + 1] return [] def read_to_next_empty_line(self): self.seek_next_non_empty_line() def is_empty(line): return not line.strip() return self.read_to_condition(is_empty) def read_to_next_unindented_line(self): def is_unindented(line): return (line.strip() and (len(line.lstrip()) == len(line))) return self.read_to_condition(is_unindented) def peek(self, n=0): if self._l + n < len(self._str): return self[self._l + n] else: return '' def is_empty(self): return not ''.join(self._str).strip() class NumpyDocString(object): def __init__(self, docstring, config={}): docstring = textwrap.dedent(docstring).split('\n') self._doc = Reader(docstring) self._parsed_data = { 'Signature': '', 'Summary': [''], 'Extended Summary': [], 'Parameters': [], 'Returns': [], 'Raises': [], 'Warns': [], 'Other Parameters': [], 'Attributes': [], 'Methods': [], 'See Also': [], 'Notes': [], 'Warnings': [], 'References': '', 'Examples': '', 'index': {} } self._parse() def __getitem__(self, key): return self._parsed_data[key] def __setitem__(self, key, val): if key not in self._parsed_data: warn("Unknown section %s" % key) else: self._parsed_data[key] = val def _is_at_section(self): self._doc.seek_next_non_empty_line() if self._doc.eof(): return False l1 = self._doc.peek().strip() # e.g. Parameters if l1.startswith('.. index::'): return True l2 = self._doc.peek(1).strip() # ---------- or ========== return l2.startswith('-' * len(l1)) or l2.startswith('=' * len(l1)) def _strip(self, doc): i = 0 j = 0 for i, line in enumerate(doc): if line.strip(): break for j, line in enumerate(doc[::-1]): if line.strip(): break return doc[i:len(doc) - j] def _read_to_next_section(self): section = self._doc.read_to_next_empty_line() while not self._is_at_section() and not self._doc.eof(): if not self._doc.peek(-1).strip(): # previous line was empty section += [''] section += self._doc.read_to_next_empty_line() return section def _read_sections(self): while not self._doc.eof(): data = self._read_to_next_section() name = data[0].strip() if name.startswith('..'): # index section yield name, data[1:] elif len(data) < 2: yield StopIteration else: yield name, self._strip(data[2:]) def _parse_param_list(self, content): r = Reader(content) params = [] while not r.eof(): header = r.read().strip() if ' : ' in header: arg_name, arg_type = header.split(' : ')[:2] else: arg_name, arg_type = header, '' desc = r.read_to_next_unindented_line() desc = dedent_lines(desc) params.append((arg_name, arg_type, desc)) return params _name_rgx = re.compile(r"^\s*(:(?P<role>\w+):`(?P<name>[a-zA-Z0-9_.-]+)`|" r" (?P<name2>[a-zA-Z0-9_.-]+))\s*", re.X) def _parse_see_also(self, content): """ func_name : Descriptive text continued text another_func_name : Descriptive text func_name1, func_name2, :meth:`func_name`, func_name3 """ items = [] def parse_item_name(text): """Match ':role:`name`' or 'name'""" m = self._name_rgx.match(text) if m: g = m.groups() if g[1] is None: return g[3], None else: return g[2], g[1] raise ValueError("%s is not a item name" % text) def push_item(name, rest): if not name: return name, role = parse_item_name(name) items.append((name, list(rest), role)) del rest[:] current_func = None rest = [] for line in content: if not line.strip(): continue m = self._name_rgx.match(line) if m and line[m.end():].strip().startswith(':'): push_item(current_func, rest) current_func, line = line[:m.end()], line[m.end():] rest = [line.split(':', 1)[1].strip()] if not rest[0]: rest = [] elif not line.startswith(' '): push_item(current_func, rest) current_func = None if ',' in line: for func in line.split(','): if func.strip(): push_item(func, []) elif line.strip(): current_func = line elif current_func is not None: rest.append(line.strip()) push_item(current_func, rest) return items def _parse_index(self, section, content): """ .. index: default :refguide: something, else, and more """ def strip_each_in(lst): return [s.strip() for s in lst] out = {} section = section.split('::') if len(section) > 1: out['default'] = strip_each_in(section[1].split(','))[0] for line in content: line = line.split(':') if len(line) > 2: out[line[1]] = strip_each_in(line[2].split(',')) return out def _parse_summary(self): """Grab signature (if given) and summary""" if self._is_at_section(): return summary = self._doc.read_to_next_empty_line() summary_str = " ".join([s.strip() for s in summary]).strip() if re.compile('^([\w., ]+=)?\s*[\w\.]+\(.*\)$').match(summary_str): self['Signature'] = summary_str if not self._is_at_section(): self['Summary'] = self._doc.read_to_next_empty_line() else: self['Summary'] = summary if not self._is_at_section(): self['Extended Summary'] = self._read_to_next_section() def _parse(self): self._doc.reset() self._parse_summary() for (section, content) in self._read_sections(): if not section.startswith('..'): section = ' '.join([s.capitalize() for s in section.split(' ')]) if section in ('Parameters', 'Returns', 'Raises', 'Warns', 'Other Parameters', 'Attributes', 'Methods'): self[section] = self._parse_param_list(content) elif section.startswith('.. index::'): self['index'] = self._parse_index(section, content) elif section == 'See Also': self['See Also'] = self._parse_see_also(content) else: self[section] = content # string conversion routines def _str_header(self, name, symbol='-'): return [name, len(name) * symbol] def _str_indent(self, doc, indent=4): out = [] for line in doc: out += [' ' * indent + line] return out def _str_signature(self): if self['Signature']: return [self['Signature'].replace('*', '\*')] + [''] else: return [''] def _str_summary(self): if self['Summary']: return self['Summary'] + [''] else: return [] def _str_extended_summary(self): if self['Extended Summary']: return self['Extended Summary'] + [''] else: return [] def _str_param_list(self, name): out = [] if self[name]: out += self._str_header(name) for param, param_type, desc in self[name]: out += ['%s : %s' % (param, param_type)] out += self._str_indent(desc) out += [''] return out def _str_section(self, name): out = [] if self[name]: out += self._str_header(name) out += self[name] out += [''] return out def _str_see_also(self, func_role): if not self['See Also']: return [] out = [] out += self._str_header("See Also") last_had_desc = True for func, desc, role in self['See Also']: if role: link = ':%s:`%s`' % (role, func) elif func_role: link = ':%s:`%s`' % (func_role, func) else: link = "`%s`_" % func if desc or last_had_desc: out += [''] out += [link] else: out[-1] += ", %s" % link if desc: out += self._str_indent([' '.join(desc)]) last_had_desc = True else: last_had_desc = False out += [''] return out def _str_index(self): idx = self['index'] out = [] out += ['.. index:: %s' % idx.get('default', '')] for section, references in idx.items(): if section == 'default': continue out += [' :%s: %s' % (section, ', '.join(references))] return out def __str__(self, func_role=''): out = [] out += self._str_signature() out += self._str_summary() out += self._str_extended_summary() for param_list in ('Parameters', 'Returns', 'Other Parameters', 'Raises', 'Warns'): out += self._str_param_list(param_list) out += self._str_section('Warnings') out += self._str_see_also(func_role) for s in ('Notes', 'References', 'Examples'): out += self._str_section(s) for param_list in ('Attributes', 'Methods'): out += self._str_param_list(param_list) out += self._str_index() return '\n'.join(out) def indent(str, indent=4): indent_str = ' ' * indent if str is None: return indent_str lines = str.split('\n') return '\n'.join(indent_str + l for l in lines) def dedent_lines(lines): """Deindent a list of lines maximally""" return textwrap.dedent("\n".join(lines)).split("\n") def header(text, style='-'):<|fim▁hole|> def __init__(self, func, role='func', doc=None, config={}): self._f = func self._role = role # e.g. "func" or "meth" if doc is None: if func is None: raise ValueError("No function or docstring given") doc = inspect.getdoc(func) or '' NumpyDocString.__init__(self, doc) if not self['Signature'] and func is not None: func, func_name = self.get_func() try: # try to read signature argspec = inspect.signature(func) argspec = inspect.formatargspec(*argspec) argspec = argspec.replace('*', '\*') signature = '%s%s' % (func_name, argspec) except TypeError as e: signature = '%s()' % func_name self['Signature'] = signature def get_func(self): func_name = getattr(self._f, '__name__', self.__class__.__name__) if inspect.isclass(self._f): func = getattr(self._f, '__call__', self._f.__init__) else: func = self._f return func, func_name def __str__(self): out = '' func, func_name = self.get_func() signature = self['Signature'].replace('*', '\*') roles = {'func': 'function', 'meth': 'method'} if self._role: if self._role not in roles: print("Warning: invalid role %s" % self._role) out += '.. %s:: %s\n \n\n' % (roles.get(self._role, ''), func_name) out += super(FunctionDoc, self).__str__(func_role=self._role) return out class ClassDoc(NumpyDocString): extra_public_methods = ['__call__'] def __init__(self, cls, doc=None, modulename='', func_doc=FunctionDoc, config={}): if not inspect.isclass(cls) and cls is not None: raise ValueError("Expected a class or None, but got %r" % cls) self._cls = cls if modulename and not modulename.endswith('.'): modulename += '.' self._mod = modulename if doc is None: if cls is None: raise ValueError("No class or documentation string given") doc = pydoc.getdoc(cls) NumpyDocString.__init__(self, doc) if config.get('show_class_members', True): if not self['Methods']: self['Methods'] = [(name, '', '') for name in sorted(self.methods)] if not self['Attributes']: self['Attributes'] = [(name, '', '') for name in sorted(self.properties)] @property def methods(self): if self._cls is None: return [] return [name for name, func in inspect.getmembers(self._cls) if ((not name.startswith('_') or name in self.extra_public_methods) and isinstance(func, collections.Callable))] @property def properties(self): if self._cls is None: return [] return [name for name, func in inspect.getmembers(self._cls) if not name.startswith('_') and func is None]<|fim▁end|>
return text + '\n' + style * len(text) + '\n' class FunctionDoc(NumpyDocString):
<|file_name|>en.js<|end_file_name|><|fim▁begin|>/* * /MathJax-v2/localization/en/en.js * * Copyright (c) 2009-2018 The MathJax Consortium * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ <|fim▁hole|><|fim▁end|>
MathJax.Localization.addTranslation("en",null,{menuTitle:"English",version:"2.7.8",isLoaded:true,domains:{_:{version:"2.7.8",isLoaded:true,strings:{CookieConfig:"MathJax has found a user-configuration cookie that includes code to be run. Do you want to run it?\n\n(You should press Cancel unless you set up the cookie yourself.)",MathProcessingError:"Math processing error",MathError:"Math error",LoadFile:"Loading %1",Loading:"Loading",LoadFailed:"File failed to load: %1",ProcessMath:"Processing math: %1%%",Processing:"Processing",TypesetMath:"Typesetting math: %1%%",Typesetting:"Typesetting",MathJaxNotSupported:"Your browser does not support MathJax",ErrorTips:"Debugging tips: use %%1, inspect %%2 in the browser console"}},FontWarnings:{},"HTML-CSS":{},HelpDialog:{},MathML:{},MathMenu:{},TeX:{}},plural:function(a){if(a===1){return 1}return 2},number:function(a){return a}});MathJax.Ajax.loadComplete("[MathJax]/localization/en/en.js");
<|file_name|>package.js<|end_file_name|><|fim▁begin|>/* global Package, Npm */ Package.describe({ name: 'procempa:keycloak-auth', version: '1.0.0', summary: 'Meteor Keycloak Handshake flow', git: 'https://github.com/Procempa/meteor-keycloak-auth.git', documentation: 'README.md' }); Package.onUse(function(api) { api.use('[email protected]'); api.use('[email protected]'); api.export('KeycloakServer', 'server'); api.export('KeycloakClient', 'client'); api.mainModule('client-main.js', 'client'); api.mainModule('server-main.js', 'server'); }); Npm.depends({ 'lodash': '4.16.1', 'fallbackjs': '1.1.8', 'localforage': '1.4.2', 'keycloak-auth-utils': '2.2.1', 'babel-plugin-transform-decorators-legacy': '1.3.4', 'babel-plugin-transform-class-properties': '6.11.5', 'babel-plugin-transform-strict-mode': '6.11.3',<|fim▁hole|> 'q': '1.4.1' });<|fim▁end|>
<|file_name|>iron_hippo.go<|end_file_name|><|fim▁begin|>package iron_hippo_exe import ( "fmt" "io" "net/http" "golang.org/x/oauth2" "golang.org/x/oauth2/google" "google.golang.org/appengine" "google.golang.org/appengine/log" "google.golang.org/appengine/urlfetch" ) const ProjectID = "cpb101demo1" type DataflowTemplatePostBody struct { JobName string `json:"jobName"` GcsPath string `json:"gcsPath"` Parameters struct { InputTable string `json:"inputTable"` OutputProjectID string `json:"outputProjectId"` OutputKind string `json:"outputKind"` } `json:"parameters"` Environment struct { TempLocation string `json:"tempLocation"` Zone string `json:"zone"`<|fim▁hole|> http.HandleFunc("/cron/start", handler) } func handler(w http.ResponseWriter, r *http.Request) { ctx := appengine.NewContext(r) client := &http.Client{ Transport: &oauth2.Transport{ Source: google.AppEngineTokenSource(ctx, "https://www.googleapis.com/auth/cloud-platform"), Base: &urlfetch.Transport{Context: ctx}, }, } res, err := client.Post(fmt.Sprintf("https://dataflow.googleapis.com/v1b3/projects/%s/templates", ProjectID), "application/json", r.Body) if err != nil { log.Errorf(ctx, "ERROR dataflow: %s", err) w.WriteHeader(http.StatusInternalServerError) return } _, err = io.Copy(w, res.Body) if err != nil { log.Errorf(ctx, "ERROR Copy API response: %s", err) w.WriteHeader(http.StatusInternalServerError) return } w.WriteHeader(res.StatusCode) }<|fim▁end|>
} `json:"environment"` } func init() {
<|file_name|>aso_bin_log.go<|end_file_name|><|fim▁begin|>package service import ( "context" "encoding/json" "strconv" "time"<|fim▁hole|> "go-common/app/job/main/passport-game-local/model" "go-common/library/log" "go-common/library/queue/databus" ) func (s *Service) asobinlogconsumeproc() { mergeNum := int64(s.c.Group.AsoBinLog.Num) for { msg, ok := <-s.dsAsoBinLogSub.Messages() if !ok { log.Error("asobinlogconsumeproc closed") return } // marked head to first commit m := &message{data: msg} s.mu.Lock() if s.head == nil { s.head = m s.last = m } else { s.last.next = m s.last = m } s.mu.Unlock() bmsg := new(model.BMsg) if err := json.Unmarshal(msg.Value, bmsg); err != nil { log.Error("json.Unmarshal(%s) error(%v)", string(msg.Value), err) continue } mid := int64(0) if bmsg.Table == _asoAccountTable { t := new(model.AsoAccount) if err := json.Unmarshal(bmsg.New, t); err != nil { log.Error("json.Unmarshal(%s) error(%v)", string(bmsg.New), err) } mid = t.Mid m.object = bmsg log.Info("asobinlogconsumeproc table:%s key:%s partition:%d offset:%d", bmsg.Table, msg.Key, msg.Partition, msg.Offset) } else { continue } s.merges[mid%mergeNum] <- m } } func (s *Service) asobinlogcommitproc() { for { done := <-s.done commits := make(map[int32]*databus.Message) for _, d := range done { d.done = true } s.mu.Lock() for ; s.head != nil && s.head.done; s.head = s.head.next { commits[s.head.data.Partition] = s.head.data } s.mu.Unlock() for _, m := range commits { m.Commit() } } } func (s *Service) asobinlogmergeproc(c chan *message) { var ( max = s.c.Group.AsoBinLog.Size merges = make([]*model.BMsg, 0, max) marked = make([]*message, 0, max) ticker = time.NewTicker(time.Duration(s.c.Group.AsoBinLog.Ticker)) ) for { select { case msg, ok := <-c: if !ok { log.Error("asobinlogmergeproc closed") return } p, assertOk := msg.object.(*model.BMsg) if assertOk && p.Action != "" && (p.Table == _asoAccountTable) { merges = append(merges, p) } marked = append(marked, msg) if len(marked) < max && len(merges) < max { continue } case <-ticker.C: } if len(merges) > 0 { s.processAsoAccLogInfo(merges) merges = make([]*model.BMsg, 0, max) } if len(marked) > 0 { s.done <- marked marked = make([]*message, 0, max) } } } func (s *Service) processAsoAccLogInfo(bmsgs []*model.BMsg) { for _, msg := range bmsgs { s.processAsoAccLog(msg) } } func (s *Service) processAsoAccLog(msg *model.BMsg) { aso := new(model.OriginAsoAccount) if err := json.Unmarshal(msg.New, aso); err != nil { log.Error("failed to parse binlog new, json.Unmarshal(%s) error(%v)", string(msg.New), err) return } pmsg := new(model.PMsg) if "update" == msg.Action { old := new(model.AsoAccount) if err := json.Unmarshal(msg.Old, old); err != nil { log.Error("failed to parse binlog new, json.Unmarshal(%s) error(%v)", string(msg.New), err) return } if old.Pwd != aso.Pwd { pmsg.Flag = 1 } } pmsg.Action = msg.Action pmsg.Table = msg.Table pmsg.Data = model.Default(aso) key := strconv.FormatInt(aso.Mid, 10) for { if err := s.dsAsoEncryptTransPub.Send(context.TODO(), key, pmsg); err == nil { return } time.Sleep(time.Second) } }<|fim▁end|>
<|file_name|>GrayCompositeConverter.java<|end_file_name|><|fim▁begin|>/** * Logback: the reliable, generic, fast and flexible logging framework. * Copyright (C) 1999-2015, QOS.ch. All rights reserved. * <|fim▁hole|> * or (per the licensee's choosing) * * under the terms of the GNU Lesser General Public License version 2.1 * as published by the Free Software Foundation. */ package ch.qos.logback.core.pattern.color; import static ch.qos.logback.core.pattern.color.ANSIConstants.*; /** * Encloses a given set of converter output in gray using the appropriate ANSI * escape codes. * * @param <E> * @author Ceki G&uuml;lc&uuml; * @since 1.0.5 */ public class GrayCompositeConverter<E> extends ForegroundCompositeConverterBase<E> { @Override protected String getForegroundColorCode(E event) { return BOLD + BLACK_FG; } }<|fim▁end|>
* This program and the accompanying materials are dual-licensed under * either the terms of the Eclipse Public License v1.0 as published by * the Eclipse Foundation *
<|file_name|>qa_vector_map.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python # # Copyright 2012 Free Software Foundation, Inc. # # This file is part of GNU Radio # # GNU Radio is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 3, or (at your option) # any later version. # # GNU Radio is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of<|fim▁hole|># along with GNU Radio; see the file COPYING. If not, write to # the Free Software Foundation, Inc., 51 Franklin Street, # Boston, MA 02110-1301, USA. # from gnuradio import gr, gr_unittest import math class test_vector_map(gr_unittest.TestCase): def setUp (self): self.tb = gr.top_block() def tearDown (self): self.tb = None def test_reversing(self): # Chunk data in blocks of N and reverse the block contents. N = 5 src_data = range(0, 20) expected_result = [] for i in range(N-1, len(src_data), N): for j in range(0, N): expected_result.append(1.0*(i-j)) mapping = [list(reversed([(0, i) for i in range(0, N)]))] src = gr.vector_source_f(src_data, False, N) vmap = gr.vector_map(gr.sizeof_float, (N, ), mapping) dst = gr.vector_sink_f(N) self.tb.connect(src, vmap, dst) self.tb.run() result_data = list(dst.data()) self.assertEqual(expected_result, result_data) def test_vector_to_streams(self): # Split an input vector into N streams. N = 5 M = 20 src_data = range(0, M) expected_results = [] for n in range(0, N): expected_results.append(range(n, M, N)) mapping = [[(0, n)] for n in range(0, N)] src = gr.vector_source_f(src_data, False, N) vmap = gr.vector_map(gr.sizeof_float, (N, ), mapping) dsts = [gr.vector_sink_f(1) for n in range(0, N)] self.tb.connect(src, vmap) for n in range(0, N): self.tb.connect((vmap, n), dsts[n]) self.tb.run() for n in range(0, N): result_data = list(dsts[n].data()) self.assertEqual(expected_results[n], result_data) def test_interleaving(self): # Takes 3 streams (a, b and c) # Outputs 2 streams. # First (d) is interleaving of a and b. # Second (e) is interleaving of a and b and c. c is taken in # chunks of 2 which are reversed. A = (1, 2, 3, 4, 5) B = (11, 12, 13, 14, 15) C = (99, 98, 97, 96, 95, 94, 93, 92, 91, 90) expected_D = (1, 11, 2, 12, 3, 13, 4, 14, 5, 15) expected_E = (1, 11, 98, 99, 2, 12, 96, 97, 3, 13, 94, 95, 4, 14, 92, 93, 5, 15, 90, 91) mapping = [[(0, 0), (1, 0)], # mapping to produce D [(0, 0), (1, 0), (2, 1), (2, 0)], # mapping to produce E ] srcA = gr.vector_source_f(A, False, 1) srcB = gr.vector_source_f(B, False, 1) srcC = gr.vector_source_f(C, False, 2) vmap = gr.vector_map(gr.sizeof_int, (1, 1, 2), mapping) dstD = gr.vector_sink_f(2) dstE = gr.vector_sink_f(4) self.tb.connect(srcA, (vmap, 0)) self.tb.connect(srcB, (vmap, 1)) self.tb.connect(srcC, (vmap, 2)) self.tb.connect((vmap, 0), dstD) self.tb.connect((vmap, 1), dstE) self.tb.run() self.assertEqual(expected_D, dstD.data()) self.assertEqual(expected_E, dstE.data()) if __name__ == '__main__': gr_unittest.run(test_vector_map, "test_vector_map.xml")<|fim▁end|>
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License
<|file_name|>progress_bar.cpp<|end_file_name|><|fim▁begin|>#include <progress_bar.hh> PROGRESS_BAR :: PROGRESS_BAR(int x, int y, int w, int h, void(*opf)(void), void(*orf)(void), int ea, int ia, int l) :<|fim▁hole|> this->int_angle = ia; this->level = l; }<|fim▁end|>
WIDGET(x,y,w,h,opf,orf) { this->ext_angle = ea;
<|file_name|>album_endpoints.py<|end_file_name|><|fim▁begin|>import flask; from flask import request import os import urllib.parse from voussoirkit import flasktools from voussoirkit import gentools from voussoirkit import stringtools import etiquette from .. import common site = common.site session_manager = common.session_manager # Individual albums ################################################################################ @site.route('/album/<album_id>') def get_album_html(album_id): album = common.P_album(album_id, response_type='html') response = common.render_template( request, 'album.html', album=album, view=request.args.get('view', 'grid'), ) return response @site.route('/album/<album_id>.json') def get_album_json(album_id): album = common.P_album(album_id, response_type='json') album = album.jsonify() return flasktools.json_response(album) @site.route('/album/<album_id>.zip') def get_album_zip(album_id): album = common.P_album(album_id, response_type='html') recursive = request.args.get('recursive', True) recursive = stringtools.truthystring(recursive) streamed_zip = etiquette.helpers.zip_album(album, recursive=recursive) if album.title: download_as = f'album {album.id} - {album.title}.zip' else: download_as = f'album {album.id}.zip' download_as = etiquette.helpers.remove_path_badchars(download_as) download_as = urllib.parse.quote(download_as) outgoing_headers = { 'Content-Type': 'application/octet-stream', 'Content-Disposition': f'attachment; filename*=UTF-8\'\'{download_as}', } return flask.Response(streamed_zip, headers=outgoing_headers) @site.route('/album/<album_id>/add_child', methods=['POST']) @flasktools.required_fields(['child_id'], forbid_whitespace=True) def post_album_add_child(album_id): album = common.P_album(album_id, response_type='json') child_ids = stringtools.comma_space_split(request.form['child_id']) children = list(common.P_albums(child_ids, response_type='json')) print(children) album.add_children(children, commit=True) response = album.jsonify() return flasktools.json_response(response) @site.route('/album/<album_id>/remove_child', methods=['POST']) @flasktools.required_fields(['child_id'], forbid_whitespace=True) def post_album_remove_child(album_id): album = common.P_album(album_id, response_type='json') child_ids = stringtools.comma_space_split(request.form['child_id']) children = list(common.P_albums(child_ids, response_type='json')) album.remove_children(children, commit=True) response = album.jsonify() return flasktools.json_response(response) @site.route('/album/<album_id>/remove_thumbnail_photo', methods=['POST']) def post_album_remove_thumbnail_photo(album_id): album = common.P_album(album_id, response_type='json') album.set_thumbnail_photo(None) common.P.commit(message='album remove thumbnail photo endpoint') return flasktools.json_response(album.jsonify()) @site.route('/album/<album_id>/refresh_directories', methods=['POST']) def post_album_refresh_directories(album_id): album = common.P_album(album_id, response_type='json') for directory in album.get_associated_directories(): if not directory.is_dir: continue digest = common.P.digest_directory(directory, new_photo_ratelimit=0.1) gentools.run(digest) common.P.commit(message='refresh album directories endpoint') return flasktools.json_response({}) @site.route('/album/<album_id>/set_thumbnail_photo', methods=['POST']) @flasktools.required_fields(['photo_id'], forbid_whitespace=True) def post_album_set_thumbnail_photo(album_id): album = common.P_album(album_id, response_type='json') photo = common.P_photo(request.form['photo_id'], response_type='json') album.set_thumbnail_photo(photo) common.P.commit(message='album set thumbnail photo endpoint') return flasktools.json_response(album.jsonify()) # Album photo operations ########################################################################### @site.route('/album/<album_id>/add_photo', methods=['POST']) @flasktools.required_fields(['photo_id'], forbid_whitespace=True) def post_album_add_photo(album_id): ''' Add a photo or photos to this album. ''' album = common.P_album(album_id, response_type='json') photo_ids = stringtools.comma_space_split(request.form['photo_id']) photos = list(common.P_photos(photo_ids, response_type='json')) album.add_photos(photos, commit=True) response = album.jsonify() return flasktools.json_response(response) @site.route('/album/<album_id>/remove_photo', methods=['POST']) @flasktools.required_fields(['photo_id'], forbid_whitespace=True) def post_album_remove_photo(album_id): ''' Remove a photo or photos from this album. ''' album = common.P_album(album_id, response_type='json') photo_ids = stringtools.comma_space_split(request.form['photo_id']) photos = list(common.P_photos(photo_ids, response_type='json')) album.remove_photos(photos, commit=True) response = album.jsonify() return flasktools.json_response(response) # Album tag operations ############################################################################# @site.route('/album/<album_id>/add_tag', methods=['POST']) def post_album_add_tag(album_id): ''' Apply a tag to every photo in the album. ''' response = {} album = common.P_album(album_id, response_type='json') tag = request.form['tagname'].strip() try: tag = common.P_tag(tag, response_type='json') except etiquette.exceptions.NoSuchTag as exc: response = exc.jsonify() return flasktools.json_response(response, status=404) recursive = request.form.get('recursive', False) recursive = stringtools.truthystring(recursive) album.add_tag_to_all(tag, nested_children=recursive, commit=True) response['action'] = 'add_tag' response['tagname'] = tag.name return flasktools.json_response(response) # Album metadata operations ######################################################################## @site.route('/album/<album_id>/edit', methods=['POST']) def post_album_edit(album_id): ''' Edit the title / description. ''' album = common.P_album(album_id, response_type='json') title = request.form.get('title', None) description = request.form.get('description', None) album.edit(title=title, description=description, commit=True) response = album.jsonify(minimal=True) return flasktools.json_response(response) @site.route('/album/<album_id>/show_in_folder', methods=['POST']) def post_album_show_in_folder(album_id): if not request.is_localhost: flask.abort(403) album = common.P_album(album_id, response_type='json') directories = album.get_associated_directories() if len(directories) != 1: flask.abort(400) directory = directories.pop() if os.name == 'nt': command = f'start explorer.exe "{directory.absolute_path}"' os.system(command) return flasktools.json_response({}) flask.abort(501) # Album listings ################################################################################### @site.route('/all_albums.json') @flasktools.cached_endpoint(max_age=15) def get_all_album_names(): all_albums = {album.id: album.display_name for album in common.P.get_albums()} response = {'albums': all_albums} return flasktools.json_response(response) def get_albums_core(): albums = list(common.P.get_root_albums()) albums.sort(key=lambda x: x.display_name.lower()) return albums @site.route('/albums') def get_albums_html(): albums = get_albums_core() response = common.render_template( request, 'album.html', albums=albums, view=request.args.get('view', 'grid'), ) return response @site.route('/albums.json') def get_albums_json(): albums = get_albums_core() albums = [album.jsonify(minimal=True) for album in albums] return flasktools.json_response(albums) # Album create and delete ########################################################################## @site.route('/albums/create_album', methods=['POST']) def post_albums_create(): title = request.form.get('title', None) description = request.form.get('description', None) parent_id = request.form.get('parent_id', None) if parent_id is not None: parent = common.P_album(parent_id, response_type='json') user = session_manager.get(request).user<|fim▁hole|> album = common.P.new_album(title=title, description=description, author=user) if parent_id is not None: parent.add_child(album) common.P.commit('create album endpoint') response = album.jsonify(minimal=False) return flasktools.json_response(response) @site.route('/album/<album_id>/delete', methods=['POST']) def post_album_delete(album_id): album = common.P_album(album_id, response_type='json') album.delete(commit=True) return flasktools.json_response({})<|fim▁end|>
<|file_name|>exp_XGB_CF_tc_mb_mf_ntree.py<|end_file_name|><|fim▁begin|>""" Experiment for XGBoost + CF Aim: To find the best tc(max_depth), mb(min_child_weight), mf(colsample_bytree * 93), ntree tc: [13, 15, 17] mb: [5, 7, 9] mf: [40, 45, 50, 55, 60] ntree: [160, 180, 200, 220, 240, 260, 280, 300, 320, 340, 360] Averaging 20 models Summary Best loss ntree mf 40 45 50 55 60 40 45 50 55 60 tc mb 13 5 0.4471 0.4471 0.4473 0.4471 0.4476 300 300 280 280 260 7 0.4477 0.4475 0.4469 0.4472 0.4481 340 320 300 300 300 9 0.4485 0.4484 0.4487 0.4488 0.4487 360 360 340 340 340 15 5 0.4471 *0.4465* 0.4471 0.4476 0.4478 260 *260* 240 240 240 7 0.4473 0.4468 0.4473 0.4474 0.4478 300 280 260 260 260 9 0.4483 0.4480 0.4483 0.4484 0.4492 340 320 300 300 280 17 5 0.4471 0.4472 0.4474 0.4476 0.4478 240 240 220 220 200 7 0.4474 0.4470 0.4468 0.4475 0.4473 280 260 260 240 240 9 0.4481 0.4480 0.4476 0.4480 0.4486 320 300 280 260 260 Time: 1 day, 7:37:21 on i7-4790k 32G MEM GTX660 """ import numpy as np import scipy as sp import pandas as pd from sklearn.cross_validation import StratifiedKFold from sklearn.metrics import log_loss from datetime import datetime import os from sklearn.grid_search import ParameterGrid import xgboost as xgb from utility import * path = os.getcwd() + '/' path_log = path + 'logs/' file_train = path + 'train.csv' training = pd.read_csv(file_train, index_col = 0) num_train = training.shape[0] y = training['target'].values yMat = pd.get_dummies(training['target']).values X = training.iloc[:,:93].values <|fim▁hole|>for train_idx, valid_idx in kf: break y_train_1 = yMat[train_idx].argmax(1) y_train = yMat[train_idx] y_valid = yMat[valid_idx] X2, ignore = count_feature(X) dtrain , dvalid= xgb.DMatrix(X2[train_idx], label = y_train_1), xgb.DMatrix(X2[valid_idx]) # nIter = 20 nt = 360 nt_lst = range(160, 370, 20) nt_len = len(nt_lst) bf = .8 # subsample sh = .1 # eta # tc:max_depth, mb:min_child_weight, mf(max features):colsample_bytree * 93 param_grid = {'tc':[13, 15, 17], 'mb':[5, 7, 9], 'mf':[40, 45, 50, 55, 60]} scores = [] t0 = datetime.now() for params in ParameterGrid(param_grid): tc = params['tc'] mb = params['mb'] mf = params['mf'] cs = float(mf) / X.shape[1] print tc, mb, mf predAll = [np.zeros(y_valid.shape) for k in range(nt_len)] for i in range(nIter): seed = 112233 + i param = {'bst:max_depth':tc, 'bst:eta':sh,'objective':'multi:softprob','num_class':9, 'min_child_weight':mb, 'subsample':bf, 'colsample_bytree':cs, 'silent':1, 'nthread':8, 'seed':seed} plst = param.items() bst = xgb.train(plst, dtrain, nt) for s in range(nt_len): ntree = nt_lst[s] pred = bst.predict(dvalid, ntree_limit = ntree).reshape(y_valid.shape) predAll[s] += pred scores.append({'tc':tc, 'mb':mb, 'mf':mf, 'ntree':ntree, 'nModels':i+1, 'seed':seed, 'valid':log_loss(y_valid, pred), 'valid_avg':log_loss(y_valid, predAll[s] / (i+1))}) print scores[-4], datetime.now() - t0 df = pd.DataFrame(scores) if os.path.exists(path_log) is False: print 'mkdir', path_log os.mkdir(path_log) df.to_csv(path_log + 'exp_XGB_CF_tc_mb_mf_ntree.csv') keys = ['tc', 'mb', 'mf', 'ntree'] grouped = df.groupby(keys) pd.set_option('display.precision', 5) print pd.DataFrame({'loss':grouped['valid_avg'].last().unstack().min(1), 'ntree':grouped['valid_avg'].last().unstack().idxmin(1)}).unstack() # loss ntree # mf 40 45 50 55 60 40 45 50 55 60 # tc mb # 13 5 0.4471 0.4471 0.4473 0.4471 0.4476 300 300 280 280 260 # 7 0.4477 0.4475 0.4469 0.4472 0.4481 340 320 300 300 300 # 9 0.4485 0.4484 0.4487 0.4488 0.4487 360 360 340 340 340 # 15 5 0.4471 0.4465 0.4471 0.4476 0.4478 260 260 240 240 240 # 7 0.4473 0.4468 0.4473 0.4474 0.4478 300 280 260 260 260 # 9 0.4483 0.4480 0.4483 0.4484 0.4492 340 320 300 300 280 # 17 5 0.4471 0.4472 0.4474 0.4476 0.4478 240 240 220 220 200 # 7 0.4474 0.4470 0.4468 0.4475 0.4473 280 260 260 240 240 # 9 0.4481 0.4480 0.4476 0.4480 0.4486 320 300 280 260 260 print pd.DataFrame({'loss':grouped['valid'].mean().unstack().min(1), 'ntree':grouped['valid'].mean().unstack().idxmin(1)}).unstack() # loss ntree # mf 40 45 50 55 60 40 45 50 55 60 # tc mb # 13 5 0.4563 0.4564 0.4564 0.4561 0.4566 280 260 260 260 240 # 7 0.4565 0.4563 0.4557 0.4561 0.4569 320 300 300 300 280 # 9 0.4571 0.4569 0.4571 0.4573 0.4570 340 340 320 300 300 # 15 5 0.4567 0.4559 0.4565 0.4571 0.4571 260 240 240 220 220 # 7 0.4565 0.4558 0.4562 0.4564 0.4568 280 260 260 260 240 # 9 0.4570 0.4567 0.4570 0.4570 0.4577 300 300 280 280 260 # 17 5 0.4568 0.4569 0.4570 0.4572 0.4574 220 220 200 200 200 # 7 0.4567 0.4563 0.4559 0.4567 0.4564 260 240 240 220 220 # 9 0.4571 0.4569 0.4565 0.4567 0.4573 280 280 260 260 240 # criterion = df.apply(lambda x: x['tc']==15 and x['mb']==5 and x['mf']==45, axis = 1) grouped = df[criterion].groupby('ntree') g = grouped[['valid']].mean() g['valid_avg'] = grouped['valid_avg'].last() print g # valid valid_avg # ntree # 160 0.461023 0.452912 # 180 0.458513 0.450111 # 200 0.456939 0.448232 # 220 0.456147 0.447141 # 240 0.455870 0.446598 # 260 0.456097 0.446525 # 280 0.456657 0.446827 # 300 0.457434 0.447327 # 320 0.458462 0.448101 # 340 0.459635 0.449036 # 360 0.460977 0.450160 ax = g.plot() ax.set_title('XGB+CF max_depth=15\n min_child_weight=5, colsample_bytree=45/93.') ax.set_ylabel('Logloss') fig = ax.get_figure() fig.savefig(path_log + 'exp_XGB_CF_tc_mb_mf_ntree.png')<|fim▁end|>
kf = StratifiedKFold(y, n_folds=5, shuffle = True, random_state = 345)
<|file_name|>vcpw15-5.py<|end_file_name|><|fim▁begin|># vcpw15-5.py # Memory Puzzle # Original By Al Sweigart [email protected] # http://inventwithpython.com/pygame # Released under a "Simplified BSD" license import random, pygame, sys from pygame.locals import * FPS = 30 # frames per second, the general speed of the program WINDOWWIDTH = 640 # size of window's width in pixels WINDOWHEIGHT = 480 # size of windows' height in pixels REVEALSPEED = 8 # speed boxes' sliding reveals and covers BOXSIZE = 40 # size of box height & width in pixels GAPSIZE = 10 # size of gap between boxes in pixels BOARDWIDTH = 5 # number of columns of icons BOARDHEIGHT = 2 # number of rows of icons <|fim▁hole|> # R G B GRAY = (100, 100, 100) NAVYBLUE = ( 60, 60, 100) WHITE = (255, 255, 255) RED = (255, 0, 0) GREEN = ( 0, 255, 0) BLUE = ( 0, 0, 255) YELLOW = (255, 255, 0) ORANGE = (255, 128, 0) PURPLE = (255, 0, 255) CYAN = ( 0, 255, 255) BGCOLOR = NAVYBLUE LIGHTBGCOLOR = GRAY BOXCOLOR = WHITE HIGHLIGHTCOLOR = BLUE DONUT = 'donut' SQUARE = 'square' DIAMOND = 'diamond' LINES = 'lines' OVAL = 'oval' ALLCOLORS = (RED, GREEN, BLUE, YELLOW, ORANGE, PURPLE, CYAN) ALLSHAPES = (DONUT, SQUARE, DIAMOND, LINES, OVAL) assert len(ALLCOLORS) * len(ALLSHAPES) * 2 >= BOARDWIDTH * BOARDHEIGHT, "Board is too big for the number of shapes/colors defined." def main(): global FPSCLOCK, DISPLAYSURF pygame.init() FPSCLOCK = pygame.time.Clock() DISPLAYSURF = pygame.display.set_mode((WINDOWWIDTH, WINDOWHEIGHT)) mousex = 0 # used to store x coordinate of mouse event mousey = 0 # used to store y coordinate of mouse event pygame.display.set_caption('Memory Game') mainBoard = getRandomizedBoard() revealedBoxes = generateRevealedBoxesData(False) firstSelection = None # stores the (x, y) of the first box clicked. DISPLAYSURF.fill(BGCOLOR) startGameAnimation(mainBoard) while True: # main game loop mouseClicked = False DISPLAYSURF.fill(BGCOLOR) # drawing the window drawBoard(mainBoard, revealedBoxes) for event in pygame.event.get(): # event handling loop if event.type == QUIT or (event.type == KEYUP and event.key == K_ESCAPE): pygame.quit() sys.exit() elif event.type == MOUSEMOTION: mousex, mousey = event.pos elif event.type == MOUSEBUTTONUP: mousex, mousey = event.pos mouseClicked = True boxx, boxy = getBoxAtPixel(mousex, mousey) if boxx != None and boxy != None: # The mouse is currently over a box. if not revealedBoxes[boxx][boxy]: drawHighlightBox(boxx, boxy) if not revealedBoxes[boxx][boxy] and mouseClicked: revealBoxesAnimation(mainBoard, [(boxx, boxy)]) revealedBoxes[boxx][boxy] = True # set the box as "revealed" if firstSelection == None: # the current box was the first box clicked firstSelection = (boxx, boxy) else: # the current box was the second box clicked # Check if there is a match between the two icons. icon1shape, icon1color = getShapeAndColor(mainBoard, firstSelection[0], firstSelection[1]) icon2shape, icon2color = getShapeAndColor(mainBoard, boxx, boxy) if icon1shape != icon2shape or icon1color != icon2color: # Icons don't match. Re-cover up both selections. pygame.time.wait(1000) # 1000 milliseconds = 1 sec coverBoxesAnimation(mainBoard, [(firstSelection[0], firstSelection[1]), (boxx, boxy)]) revealedBoxes[firstSelection[0]][firstSelection[1]] = False revealedBoxes[boxx][boxy] = False elif hasWon(revealedBoxes): # check if all pairs found gameWonAnimation(mainBoard) pygame.time.wait(2000) # Reset the board mainBoard = getRandomizedBoard() revealedBoxes = generateRevealedBoxesData(False) # Show the fully unrevealed board for a second. drawBoard(mainBoard, revealedBoxes) pygame.display.update() pygame.time.wait(1000) # Replay the start game animation. startGameAnimation(mainBoard) firstSelection = None # reset firstSelection variable # Redraw the screen and wait a clock tick. pygame.display.update() FPSCLOCK.tick(FPS) def generateRevealedBoxesData(val): revealedBoxes = [] for i in range(BOARDWIDTH): revealedBoxes.append([val] * BOARDHEIGHT) return revealedBoxes def getRandomizedBoard(): # Get a list of every possible shape in every possible color. icons = [] for color in ALLCOLORS: for shape in ALLSHAPES: icons.append( (shape, color) ) random.shuffle(icons) # randomize the order of the icons list numIconsUsed = int(BOARDWIDTH * BOARDHEIGHT / 2) # calculate how many icons are needed icons = icons[:numIconsUsed] * 2 # make two of each random.shuffle(icons) # Create the board data structure, with randomly placed icons. board = [] for x in range(BOARDWIDTH): column = [] for y in range(BOARDHEIGHT): column.append(icons[0]) del icons[0] # remove the icons as we assign them board.append(column) return board def splitIntoGroupsOf(groupSize, theList): # splits a list into a list of lists, where the inner lists have at # most groupSize number of items. result = [] for i in range(0, len(theList), groupSize): result.append(theList[i:i + groupSize]) return result def leftTopCoordsOfBox(boxx, boxy): # Convert board coordinates to pixel coordinates left = boxx * (BOXSIZE + GAPSIZE) + XMARGIN top = boxy * (BOXSIZE + GAPSIZE) + YMARGIN return (left, top) def getBoxAtPixel(x, y): for boxx in range(BOARDWIDTH): for boxy in range(BOARDHEIGHT): left, top = leftTopCoordsOfBox(boxx, boxy) boxRect = pygame.Rect(left, top, BOXSIZE, BOXSIZE) if boxRect.collidepoint(x, y): return (boxx, boxy) return (None, None) def drawIcon(shape, color, boxx, boxy): quarter = int(BOXSIZE * 0.25) # syntactic sugar half = int(BOXSIZE * 0.5) # syntactic sugar left, top = leftTopCoordsOfBox(boxx, boxy) # get pixel coords from board coords # Draw the shapes if shape == DONUT: pygame.draw.circle(DISPLAYSURF, color, (left + half, top + half), half - 5) pygame.draw.circle(DISPLAYSURF, BGCOLOR, (left + half, top + half), quarter - 5) elif shape == SQUARE: pygame.draw.rect(DISPLAYSURF, color, (left + quarter, top + quarter, BOXSIZE - half, BOXSIZE - half)) elif shape == DIAMOND: pygame.draw.polygon(DISPLAYSURF, color, ((left + half, top), (left + BOXSIZE - 1, top + half), (left + half, top + BOXSIZE - 1), (left, top + half))) elif shape == LINES: for i in range(0, BOXSIZE, 4): pygame.draw.line(DISPLAYSURF, color, (left, top + i), (left + i, top)) pygame.draw.line(DISPLAYSURF, color, (left + i, top + BOXSIZE - 1), (left + BOXSIZE - 1, top + i)) elif shape == OVAL: pygame.draw.ellipse(DISPLAYSURF, color, (left, top + quarter, BOXSIZE, half)) def getShapeAndColor(board, boxx, boxy): # shape value for x, y spot is stored in board[x][y][0] # color value for x, y spot is stored in board[x][y][1] return board[boxx][boxy][0], board[boxx][boxy][1] def drawBoxCovers(board, boxes, coverage): # Draws boxes being covered/revealed. "boxes" is a list # of two-item lists, which have the x & y spot of the box. for box in boxes: left, top = leftTopCoordsOfBox(box[0], box[1]) pygame.draw.rect(DISPLAYSURF, BGCOLOR, (left, top, BOXSIZE, BOXSIZE)) shape, color = getShapeAndColor(board, box[0], box[1]) drawIcon(shape, color, box[0], box[1]) if coverage > 0: # only draw the cover if there is an coverage pygame.draw.rect(DISPLAYSURF, BOXCOLOR, (left, top, coverage, BOXSIZE)) pygame.display.update() FPSCLOCK.tick(FPS) def revealBoxesAnimation(board, boxesToReveal): # Do the "box reveal" animation. for coverage in range(BOXSIZE, (-REVEALSPEED) - 1, -REVEALSPEED): drawBoxCovers(board, boxesToReveal, coverage) def coverBoxesAnimation(board, boxesToCover): # Do the "box cover" animation. for coverage in range(0, BOXSIZE + REVEALSPEED, REVEALSPEED): drawBoxCovers(board, boxesToCover, coverage) def drawBoard(board, revealed): # Draws all of the boxes in their covered or revealed state. for boxx in range(BOARDWIDTH): for boxy in range(BOARDHEIGHT): left, top = leftTopCoordsOfBox(boxx, boxy) if not revealed[boxx][boxy]: # Draw a covered box. pygame.draw.rect(DISPLAYSURF, BOXCOLOR, (left, top, BOXSIZE, BOXSIZE)) else: # Draw the (revealed) icon. shape, color = getShapeAndColor(board, boxx, boxy) drawIcon(shape, color, boxx, boxy) def drawHighlightBox(boxx, boxy): left, top = leftTopCoordsOfBox(boxx, boxy) pygame.draw.rect(DISPLAYSURF, HIGHLIGHTCOLOR, (left - 5, top - 5, BOXSIZE + 10, BOXSIZE + 10), 4) def startGameAnimation(board): # Randomly reveal the boxes 8 at a time. coveredBoxes = generateRevealedBoxesData(False) boxes = [] for x in range(BOARDWIDTH): for y in range(BOARDHEIGHT): boxes.append( (x, y) ) random.shuffle(boxes) boxGroups = splitIntoGroupsOf(8, boxes) drawBoard(board, coveredBoxes) for boxGroup in boxGroups: revealBoxesAnimation(board, boxGroup) coverBoxesAnimation(board, boxGroup) def gameWonAnimation(board): # flash the background color when the player has won coveredBoxes = generateRevealedBoxesData(True) color1 = LIGHTBGCOLOR color2 = BGCOLOR for i in range(13): color1, color2 = color2, color1 # swap colors DISPLAYSURF.fill(color1) drawBoard(board, coveredBoxes) pygame.display.update() pygame.time.wait(300) def hasWon(revealedBoxes): # Returns True if all the boxes have been revealed, otherwise False for i in revealedBoxes: if False in i: return False # return False if any boxes are covered. return True if __name__ == '__main__': main()<|fim▁end|>
assert (BOARDWIDTH * BOARDHEIGHT) % 2 == 0, 'Board needs to have an even number of boxes for pairs of matches.' XMARGIN = int((WINDOWWIDTH - (BOARDWIDTH * (BOXSIZE + GAPSIZE))) / 2) YMARGIN = int((WINDOWHEIGHT - (BOARDHEIGHT * (BOXSIZE + GAPSIZE))) / 2)
<|file_name|>htmldialogelement.rs<|end_file_name|><|fim▁begin|>/* This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this * file, You can obtain one at https://mozilla.org/MPL/2.0/. */ use crate::dom::bindings::cell::DomRefCell; use crate::dom::bindings::codegen::Bindings::HTMLDialogElementBinding; use crate::dom::bindings::codegen::Bindings::HTMLDialogElementBinding::HTMLDialogElementMethods; use crate::dom::bindings::inheritance::Castable; use crate::dom::bindings::root::DomRoot; use crate::dom::bindings::str::DOMString; use crate::dom::document::Document; use crate::dom::element::Element; use crate::dom::eventtarget::EventTarget; use crate::dom::htmlelement::HTMLElement; use crate::dom::node::{window_from_node, Node}; use dom_struct::dom_struct; use html5ever::{LocalName, Prefix}; #[dom_struct] pub struct HTMLDialogElement { htmlelement: HTMLElement, return_value: DomRefCell<DOMString>, } impl HTMLDialogElement { fn new_inherited( local_name: LocalName,<|fim▁hole|> prefix: Option<Prefix>, document: &Document, ) -> HTMLDialogElement { HTMLDialogElement { htmlelement: HTMLElement::new_inherited(local_name, prefix, document), return_value: DomRefCell::new(DOMString::new()), } } #[allow(unrooted_must_root)] pub fn new( local_name: LocalName, prefix: Option<Prefix>, document: &Document, ) -> DomRoot<HTMLDialogElement> { Node::reflect_node( Box::new(HTMLDialogElement::new_inherited( local_name, prefix, document, )), document, HTMLDialogElementBinding::Wrap, ) } } impl HTMLDialogElementMethods for HTMLDialogElement { // https://html.spec.whatwg.org/multipage/#dom-dialog-open make_bool_getter!(Open, "open"); // https://html.spec.whatwg.org/multipage/#dom-dialog-open make_bool_setter!(SetOpen, "open"); // https://html.spec.whatwg.org/multipage/#dom-dialog-returnvalue fn ReturnValue(&self) -> DOMString { let return_value = self.return_value.borrow(); return_value.clone() } // https://html.spec.whatwg.org/multipage/#dom-dialog-returnvalue fn SetReturnValue(&self, return_value: DOMString) { *self.return_value.borrow_mut() = return_value; } // https://html.spec.whatwg.org/multipage/#dom-dialog-close fn Close(&self, return_value: Option<DOMString>) { let element = self.upcast::<Element>(); let target = self.upcast::<EventTarget>(); let win = window_from_node(self); // Step 1 & 2 if element .remove_attribute(&ns!(), &local_name!("open")) .is_none() { return; } // Step 3 if let Some(new_value) = return_value { *self.return_value.borrow_mut() = new_value; } // TODO: Step 4 implement pending dialog stack removal // Step 5 win.task_manager() .dom_manipulation_task_source() .queue_simple_event(target, atom!("close"), &win); } }<|fim▁end|>
<|file_name|>default.py<|end_file_name|><|fim▁begin|>import os,sys,urllib2 import xbmcplugin,xbmcgui import xml.etree.ElementTree as ET __addon__ = "SomaFM" __addonid__ = "plugin.audio.somafm" __version__ = "0.0.2" def log(msg): print "[PLUGIN] '%s (%s)' " % (__addon__, __version__) + str(msg) log("Initialized!") log(sys.argv) rootURL = "http://somafm.com/" <|fim▁hole|>#pluginPath = sys.argv[0] handle = int(sys.argv[1]) query = sys.argv[2] def getHeaders(withReferrer=None): headers = {} headers['User-Agent'] = 'Mozilla/5.0 (Windows; U; Windows NT 6.1; en-US; rv:1.9.2.3) Gecko/20100401 Firefox/3.6.3' if withReferrer: headers['Referrer'] = withReferrer return headers def getHTMLFor(url, withData=None, withReferrer=None): url = rootURL + url log("Get HTML for URL: " + url) req = urllib2.Request(url, withData, getHeaders(withReferrer)) response = urllib2.urlopen(req) data = response.read() response.close() return data def addEntries(): somaXML = getHTMLFor(url="channels.xml") channelsContainer = ET.fromstring(somaXML) for stations in channelsContainer.findall(".//channel"): title = stations.find('title').text description = stations.find('description').text if stations.find('largeimage') is not None: img = rootURL + stations.find('largeimage').text.replace(rootURL,"") else: img = rootURL + stations.find('image').text.replace(rootURL,"") url = rootURL + stations.find('fastpls').text.replace(rootURL,"") log(title) log(description) log(img) log(url) li = xbmcgui.ListItem(title, description, thumbnailImage=img) li.setProperty("IsPlayable","true") xbmcplugin.addDirectoryItem( handle=handle, url=url, listitem=li) addEntries() xbmcplugin.endOfDirectory(handle)<|fim▁end|>
<|file_name|>__init__.py<|end_file_name|><|fim▁begin|>from django.conf import settings def MAX_USERNAME_LENGTH(): return getattr(settings, "MAX_USERNAME_LENGTH", 255) def MAX_EMAIL_LENGTH(): return getattr(settings, "MAX_EMAIL_LENGTH", 255) <|fim▁hole|> def REQUIRE_UNIQUE_EMAIL(): return getattr(settings, "REQUIRE_UNIQUE_EMAIL", True)<|fim▁end|>
<|file_name|>SearchLayoutButton.js<|end_file_name|><|fim▁begin|>/** * Copyright IBM Corp. 2016, 2018 * * This source code is licensed under the Apache-2.0 license found in the * LICENSE file in the root directory of this source tree. */ import React, { Component } from 'react'; import PropTypes from 'prop-types'; import { settings } from 'carbon-components'; import ListBulleted16 from '@carbon/icons-react/lib/list--bulleted/16'; import Grid16 from '@carbon/icons-react/lib/grid/16'; const { prefix } = settings; /** * The layout button for `<Search>`. */ class SearchLayoutButton extends Component { state = { format: 'list' }; static propTypes = { /** * The layout. */ format: PropTypes.oneOf(['list', 'grid']), /** * The a11y label text. */ labelText: PropTypes.string, /** * The description for the "list" icon. */ iconDescriptionList: PropTypes.string, /** * The description for the "grid" icon. */ iconDescriptionGrid: PropTypes.string, /** * The callback called when layout switches. */ onChangeFormat: PropTypes.func, }; static defaultProps = { labelText: 'Filter', iconDescriptionList: 'list', iconDescriptionGrid: 'grid', }; static getDerivedStateFromProps({ format }, state) { const { prevFormat } = state; return prevFormat === format ? null : { format: format || 'list', prevFormat: format, }; } /** * Toggles the button state upon user-initiated event. */ toggleLayout = () => { const format = this.state.format === 'list' ? 'grid' : 'list'; this.setState({ format }, () => { const { onChangeFormat } = this.props; if (typeof onChangeFormat === 'function') { onChangeFormat({ format }); } });<|fim▁hole|> const SearchLayoutButtonIcon = () => { if (this.state.format === 'list') { return ( <ListBulleted16 className={`${prefix}--search-view`} aria-label={iconDescriptionList} /> ); } return ( <Grid16 className={`${prefix}--search-view`} aria-label={iconDescriptionGrid} /> ); }; return ( <button className={`${prefix}--search-button`} type="button" onClick={this.toggleLayout} aria-label={labelText} title={labelText}> <div className={`${prefix}--search__toggle-layout__container`}> <SearchLayoutButtonIcon /> </div> </button> ); } } export default SearchLayoutButton;<|fim▁end|>
}; render() { const { labelText, iconDescriptionList, iconDescriptionGrid } = this.props;
<|file_name|>urls.py<|end_file_name|><|fim▁begin|>"""cmput404project URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/1.10/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """<|fim▁hole|>from rest_framework import routers from service import views urlpatterns = [ url(r'^', include('service.urls')), url(r'^docs/', include('rest_framework_docs.urls')), url(r'^admin/', admin.site.urls), url(r'^api-auth/', include('rest_framework.urls', namespace='rest_framework')), ]<|fim▁end|>
from django.conf.urls import include, url from django.contrib import admin from django.contrib.auth import views as auth_views
<|file_name|>data.go<|end_file_name|><|fim▁begin|>package main<|fim▁hole|> "github.com/gokyle/adn" "os" "path/filepath" ) const ( AppName = "The Social Gopher" AppUnixName = "socialgopher" AppVersion = "0.1" ) var ( homeDir = os.ExpandEnv("${HOME}/." + AppUnixName) authDatabaseFile = filepath.Join(homeDir, "profiles.db") ) // The desktop experience should mimic in functionality the web experience. // Accordingly, we need to be able to do everything. var Scopes = []string{ adn.ScopeBasic, adn.ScopeStream, adn.ScopeEmail, adn.ScopeWritePost, adn.ScopeFollow, adn.ScopeMessages, adn.ScopeExport, } // App represents the Social Gopher, and stores our client information. var App = &adn.Application{ "STUdQHPA8EC9zeKpqd3hWGUC2VzJASxG", // client ID "placeholder_secret", // client secret "", // redirect URI Scopes, // scopes "hgq6MjvtePat6ZZwDsJxAMX4Vvq5PvCE", // password secret } // Profiles is a list of all the users the app knows about. var Profiles []*Profile var ( BodyTypeJSON = "application/json" )<|fim▁end|>
// store global values here import (
<|file_name|>__init__.py<|end_file_name|><|fim▁begin|>""" Base backends structures. This module defines base classes needed to define custom OpenID or OAuth auth services from third parties. This customs must subclass an Auth and and Backend class, check current implementation for examples. Also the modules *must* define a BACKENDS dictionary with the backend name (which is used for URLs matching) and Auth class, otherwise it won't be enabled. """ from urllib2 import Request, urlopen, HTTPError from urllib import urlencode from urlparse import urlsplit from openid.consumer.consumer import Consumer, SUCCESS, CANCEL, FAILURE from openid.consumer.discover import DiscoveryFailure from openid.extensions import sreg, ax from oauth2 import Consumer as OAuthConsumer, Token, Request as OAuthRequest from django.db import models from django.contrib.auth import authenticate from django.contrib.auth.backends import ModelBackend from django.utils import simplejson from django.utils.importlib import import_module from django.utils.crypto import constant_time_compare, get_random_string from django.middleware.csrf import CSRF_KEY_LENGTH from social_auth.utils import setting, log, model_to_ctype, ctype_to_model, \ clean_partial_pipeline from social_auth.store import DjangoOpenIDStore from social_auth.backends.exceptions import StopPipeline, AuthException, \ AuthFailed, AuthCanceled, \ AuthUnknownError, AuthTokenError, \ AuthMissingParameter, \ AuthForbidden from social_auth.backends.utils import build_consumer_oauth_request if setting('SOCIAL_AUTH_USER_MODEL'): User = models.get_model(*setting('SOCIAL_AUTH_USER_MODEL').rsplit('.', 1)) else: from django.contrib.auth.models import User # OpenID configuration OLD_AX_ATTRS = [ ('http://schema.openid.net/contact/email', 'old_email'), ('http://schema.openid.net/namePerson', 'old_fullname'), ('http://schema.openid.net/namePerson/friendly', 'old_nickname') ] AX_SCHEMA_ATTRS = [ # Request both the full name and first/last components since some # providers offer one but not the other. ('http://axschema.org/contact/email', 'email'), ('http://axschema.org/namePerson', 'fullname'), ('http://axschema.org/namePerson/first', 'first_name'), ('http://axschema.org/namePerson/last', 'last_name'), ('http://axschema.org/namePerson/friendly', 'nickname'), ] SREG_ATTR = [ ('email', 'email'), ('fullname', 'fullname'), ('nickname', 'nickname') ] OPENID_ID_FIELD = 'openid_identifier' SESSION_NAME = 'openid' # key for username in user details dict used around, see get_user_details # method USERNAME = 'username' PIPELINE = setting('SOCIAL_AUTH_PIPELINE', ( 'social_auth.backends.pipeline.social.social_auth_user', 'social_auth.backends.pipeline.associate.associate_by_email', 'social_auth.backends.pipeline.user.get_username', 'social_auth.backends.pipeline.user.create_user', 'social_auth.backends.pipeline.social.associate_user', 'social_auth.backends.pipeline.social.load_extra_data', 'social_auth.backends.pipeline.user.update_user_details', )) class SocialAuthBackend(ModelBackend): """A django.contrib.auth backend that authenticates the user based on a authentication provider response""" name = '' # provider name, it's stored in database def authenticate(self, *args, **kwargs): """Authenticate user using social credentials Authentication is made if this is the correct backend, backend verification is made by kwargs inspection for current backend name presence. """ # Validate backend and arguments. Require that the Social Auth # response be passed in as a keyword argument, to make sure we # don't match the username/password calling conventions of # authenticate. if not (self.name and kwargs.get(self.name) and 'response' in kwargs): return None response = kwargs.get('response') pipeline = PIPELINE kwargs = kwargs.copy() kwargs['backend'] = self if 'pipeline_index' in kwargs: pipeline = pipeline[kwargs['pipeline_index']:] else: kwargs['details'] = self.get_user_details(response) kwargs['uid'] = self.get_user_id(kwargs['details'], response) kwargs['is_new'] = False out = self.pipeline(pipeline, *args, **kwargs) if not isinstance(out, dict): return out social_user = out.get('social_user') if social_user: # define user.social_user attribute to track current social # account user = social_user.user user.social_user = social_user user.is_new = out.get('is_new') return user def pipeline(self, pipeline, *args, **kwargs): """Pipeline""" out = kwargs.copy() if 'pipeline_index' in kwargs: base_index = int(kwargs['pipeline_index']) else: base_index = 0 for idx, name in enumerate(pipeline): out['pipeline_index'] = base_index + idx mod_name, func_name = name.rsplit('.', 1) try: mod = import_module(mod_name) except ImportError: log('exception', 'Error importing pipeline %s', name) else: func = getattr(mod, func_name, None) if callable(func): try: result = func(*args, **out) or {} except StopPipeline: # Clean partial pipeline on stop if 'request' in kwargs: clean_partial_pipeline(kwargs['request']) break if isinstance(result, dict): out.update(result) else: return result return out def extra_data(self, user, uid, response, details): """Return default blank user extra data""" return {} def get_user_id(self, details, response): """Must return a unique ID from values returned on details""" raise NotImplementedError('Implement in subclass') def get_user_details(self, response): """Must return user details in a know internal struct: {USERNAME: <username if any>, 'email': <user email if any>, 'fullname': <user full name if any>, 'first_name': <user first name if any>, 'last_name': <user last name if any>} """ raise NotImplementedError('Implement in subclass') @classmethod def tokens(cls, instance): """Return the tokens needed to authenticate the access to any API the service might provide. The return value will be a dictionary with the token type name as key and the token value. instance must be a UserSocialAuth instance. """ if instance.extra_data and 'access_token' in instance.extra_data: return { 'access_token': instance.extra_data['access_token'] } else: return {} def get_user(self, user_id): """ Return user with given ID from the User model used by this backend """ try: return User.objects.get(pk=user_id) except User.DoesNotExist: return None <|fim▁hole|> extra_data field. It must be a list of tuples with name and alias. Also settings will be inspected to get more values names that should be stored on extra_data field. Setting name is created from current backend name (all uppercase) plus _EXTRA_DATA. access_token is always stored. """ EXTRA_DATA = None def get_user_id(self, details, response): """OAuth providers return an unique user id in response""" return response['id'] def extra_data(self, user, uid, response, details): """Return access_token and extra defined names to store in extra_data field""" data = {'access_token': response.get('access_token', '')} name = self.name.replace('-', '_').upper() names = (self.EXTRA_DATA or []) + setting(name + '_EXTRA_DATA', []) for entry in names: if len(entry) == 2: (name, alias), discard = entry, False elif len(entry) == 3: name, alias, discard = entry elif len(entry) == 1: name = alias = entry else: # ??? continue value = response.get(name) if discard and not value: continue data[alias] = value return data class OpenIDBackend(SocialAuthBackend): """Generic OpenID authentication backend""" name = 'openid' def get_user_id(self, details, response): """Return user unique id provided by service""" return response.identity_url def values_from_response(self, response, sreg_names=None, ax_names=None): """Return values from SimpleRegistration response or AttributeExchange response if present. @sreg_names and @ax_names must be a list of name and aliases for such name. The alias will be used as mapping key. """ values = {} # Use Simple Registration attributes if provided if sreg_names: resp = sreg.SRegResponse.fromSuccessResponse(response) if resp: values.update((alias, resp.get(name) or '') for name, alias in sreg_names) # Use Attribute Exchange attributes if provided if ax_names: resp = ax.FetchResponse.fromSuccessResponse(response) if resp: for src, alias in ax_names: name = alias.replace('old_', '') values[name] = resp.getSingle(src, '') or values.get(name) return values def get_user_details(self, response): """Return user details from an OpenID request""" values = {USERNAME: '', 'email': '', 'fullname': '', 'first_name': '', 'last_name': ''} # update values using SimpleRegistration or AttributeExchange # values values.update(self.values_from_response(response, SREG_ATTR, OLD_AX_ATTRS + \ AX_SCHEMA_ATTRS)) fullname = values.get('fullname') or '' first_name = values.get('first_name') or '' last_name = values.get('last_name') or '' if not fullname and first_name and last_name: fullname = first_name + ' ' + last_name elif fullname: try: # Try to split name for django user storage first_name, last_name = fullname.rsplit(' ', 1) except ValueError: last_name = fullname values.update({'fullname': fullname, 'first_name': first_name, 'last_name': last_name, USERNAME: values.get(USERNAME) or \ (first_name.title() + last_name.title())}) return values def extra_data(self, user, uid, response, details): """Return defined extra data names to store in extra_data field. Settings will be inspected to get more values names that should be stored on extra_data field. Setting name is created from current backend name (all uppercase) plus _SREG_EXTRA_DATA and _AX_EXTRA_DATA because values can be returned by SimpleRegistration or AttributeExchange schemas. Both list must be a value name and an alias mapping similar to SREG_ATTR, OLD_AX_ATTRS or AX_SCHEMA_ATTRS """ name = self.name.replace('-', '_').upper() sreg_names = setting(name + '_SREG_EXTRA_DATA') ax_names = setting(name + '_AX_EXTRA_DATA') data = self.values_from_response(response, sreg_names, ax_names) return data class BaseAuth(object): """Base authentication class, new authenticators should subclass and implement needed methods. AUTH_BACKEND Authorization backend related with this service """ AUTH_BACKEND = None def __init__(self, request, redirect): self.request = request # Use request because some auth providers use POST urls with needed # GET parameters on it self.data = request.REQUEST self.redirect = redirect def auth_url(self): """Must return redirect URL to auth provider""" raise NotImplementedError('Implement in subclass') def auth_html(self): """Must return login HTML content returned by provider""" raise NotImplementedError('Implement in subclass') def auth_complete(self, *args, **kwargs): """Completes loging process, must return user instance""" raise NotImplementedError('Implement in subclass') def to_session_dict(self, next_idx, *args, **kwargs): """Returns dict to store on session for partial pipeline.""" return { 'next': next_idx, 'backend': self.AUTH_BACKEND.name, 'args': tuple(map(model_to_ctype, args)), 'kwargs': dict((key, model_to_ctype(val)) for key, val in kwargs.iteritems()) } def from_session_dict(self, entry, *args, **kwargs): """Takes session saved entry to continue pipeline and merges with any new extra argument needed. Returns tuple with next pipeline index entry, arguments and keyword arguments to continue the process.""" args = args[:] + tuple(map(ctype_to_model, entry['args'])) kwargs = kwargs.copy() kwargs.update((key, ctype_to_model(val)) for key, val in entry['kwargs'].iteritems()) return (entry['next'], args, kwargs) def continue_pipeline(self, *args, **kwargs): """Continue previous halted pipeline""" kwargs.update({ 'auth': self, self.AUTH_BACKEND.name: True }) return authenticate(*args, **kwargs) def request_token_extra_arguments(self): """Return extra arguments needed on request-token process, setting is per backend and defined by: <backend name in uppercase>_REQUEST_TOKEN_EXTRA_ARGUMENTS. """ backend_name = self.AUTH_BACKEND.name.upper().replace('-', '_') return setting(backend_name + '_REQUEST_TOKEN_EXTRA_ARGUMENTS', {}) def auth_extra_arguments(self): """Return extra arguments needed on auth process, setting is per backend and defined by: <backend name in uppercase>_AUTH_EXTRA_ARGUMENTS. """ backend_name = self.AUTH_BACKEND.name.upper().replace('-', '_') return setting(backend_name + '_AUTH_EXTRA_ARGUMENTS', {}) @property def uses_redirect(self): """Return True if this provider uses redirect url method, otherwise return false.""" return True @classmethod def enabled(cls): """Return backend enabled status, all enabled by default""" return True def disconnect(self, user, association_id=None): """Deletes current backend from user if associated. Override if extra operations are needed. """ if association_id: user.social_auth.get(id=association_id).delete() else: user.social_auth.filter(provider=self.AUTH_BACKEND.name).delete() def build_absolute_uri(self, path=None): """Build absolute URI for given path. Replace http:// schema with https:// if SOCIAL_AUTH_REDIRECT_IS_HTTPS is defined. """ uri = self.request.build_absolute_uri(path) if setting('SOCIAL_AUTH_REDIRECT_IS_HTTPS'): uri = uri.replace('http://', 'https://') return uri class OpenIdAuth(BaseAuth): """OpenId process handling""" AUTH_BACKEND = OpenIDBackend def auth_url(self): """Return auth URL returned by service""" openid_request = self.setup_request(self.auth_extra_arguments()) # Construct completion URL, including page we should redirect to return_to = self.build_absolute_uri(self.redirect) return openid_request.redirectURL(self.trust_root(), return_to) def auth_html(self): """Return auth HTML returned by service""" openid_request = self.setup_request(self.auth_extra_arguments()) return_to = self.build_absolute_uri(self.redirect) form_tag = {'id': 'openid_message'} return openid_request.htmlMarkup(self.trust_root(), return_to, form_tag_attrs=form_tag) def trust_root(self): """Return trust-root option""" return setting('OPENID_TRUST_ROOT') or self.build_absolute_uri('/') def continue_pipeline(self, *args, **kwargs): """Continue previous halted pipeline""" response = self.consumer().complete(dict(self.data.items()), self.build_absolute_uri()) kwargs.update({ 'auth': self, 'response': response, self.AUTH_BACKEND.name: True }) return authenticate(*args, **kwargs) def auth_complete(self, *args, **kwargs): """Complete auth process""" response = self.consumer().complete(dict(self.data.items()), self.build_absolute_uri()) if not response: raise AuthException(self, 'OpenID relying party endpoint') elif response.status == SUCCESS: kwargs.update({ 'auth': self, 'response': response, self.AUTH_BACKEND.name: True }) return authenticate(*args, **kwargs) elif response.status == FAILURE: raise AuthFailed(self, response.message) elif response.status == CANCEL: raise AuthCanceled(self) else: raise AuthUnknownError(self, response.status) def setup_request(self, extra_params=None): """Setup request""" openid_request = self.openid_request(extra_params) # Request some user details. Use attribute exchange if provider # advertises support. if openid_request.endpoint.supportsType(ax.AXMessage.ns_uri): fetch_request = ax.FetchRequest() # Mark all attributes as required, Google ignores optional ones for attr, alias in (AX_SCHEMA_ATTRS + OLD_AX_ATTRS): fetch_request.add(ax.AttrInfo(attr, alias=alias, required=True)) else: fetch_request = sreg.SRegRequest(optional=dict(SREG_ATTR).keys()) openid_request.addExtension(fetch_request) return openid_request def consumer(self): """Create an OpenID Consumer object for the given Django request.""" return Consumer(self.request.session.setdefault(SESSION_NAME, {}), DjangoOpenIDStore()) @property def uses_redirect(self): """Return true if openid request will be handled with redirect or HTML content will be returned. """ return self.openid_request().shouldSendRedirect() def openid_request(self, extra_params=None): """Return openid request""" openid_url = self.openid_url() if extra_params: query = urlsplit(openid_url).query openid_url += (query and '&' or '?') + urlencode(extra_params) try: return self.consumer().begin(openid_url) except DiscoveryFailure, err: raise AuthException(self, 'OpenID discovery error: %s' % err) def openid_url(self): """Return service provider URL. This base class is generic accepting a POST parameter that specifies provider URL.""" if OPENID_ID_FIELD not in self.data: raise AuthMissingParameter(self, OPENID_ID_FIELD) return self.data[OPENID_ID_FIELD] class BaseOAuth(BaseAuth): """OAuth base class""" SETTINGS_KEY_NAME = '' SETTINGS_SECRET_NAME = '' def __init__(self, request, redirect): """Init method""" super(BaseOAuth, self).__init__(request, redirect) self.redirect_uri = self.build_absolute_uri(self.redirect) @classmethod def get_key_and_secret(cls): """Return tuple with Consumer Key and Consumer Secret for current service provider. Must return (key, secret), order *must* be respected. """ return setting(cls.SETTINGS_KEY_NAME), \ setting(cls.SETTINGS_SECRET_NAME) @classmethod def enabled(cls): """Return backend enabled status by checking basic settings""" return setting(cls.SETTINGS_KEY_NAME) and \ setting(cls.SETTINGS_SECRET_NAME) class ConsumerBasedOAuth(BaseOAuth): """Consumer based mechanism OAuth authentication, fill the needed parameters to communicate properly with authentication service. AUTHORIZATION_URL Authorization service url REQUEST_TOKEN_URL Request token URL ACCESS_TOKEN_URL Access token URL SERVER_URL Authorization server URL """ AUTHORIZATION_URL = '' REQUEST_TOKEN_URL = '' ACCESS_TOKEN_URL = '' SERVER_URL = '' def auth_url(self): """Return redirect url""" token = self.unauthorized_token() name = self.AUTH_BACKEND.name + 'unauthorized_token_name' self.request.session[name] = token.to_string() return self.oauth_authorization_request(token).to_url() def auth_complete(self, *args, **kwargs): """Return user, might be logged in""" name = self.AUTH_BACKEND.name + 'unauthorized_token_name' unauthed_token = self.request.session.get(name) if not unauthed_token: raise AuthTokenError('Missing unauthorized token') token = Token.from_string(unauthed_token) if token.key != self.data.get('oauth_token', 'no-token'): raise AuthTokenError('Incorrect tokens') try: access_token = self.access_token(token) except HTTPError, e: if e.code == 400: raise AuthCanceled(self) else: raise data = self.user_data(access_token) if data is not None: data['access_token'] = access_token.to_string() kwargs.update({ 'auth': self, 'response': data, self.AUTH_BACKEND.name: True }) return authenticate(*args, **kwargs) def unauthorized_token(self): """Return request for unauthorized token (first stage)""" request = self.oauth_request(token=None, url=self.REQUEST_TOKEN_URL, extra_params=self.request_token_extra_arguments()) response = self.fetch_response(request) return Token.from_string(response) def oauth_authorization_request(self, token): """Generate OAuth request to authorize token.""" return OAuthRequest.from_token_and_callback(token=token, callback=self.redirect_uri, http_url=self.AUTHORIZATION_URL, parameters=self.auth_extra_arguments()) def oauth_request(self, token, url, extra_params=None): """Generate OAuth request, setups callback url""" return build_consumer_oauth_request(self, token, url, self.redirect_uri, self.data.get('oauth_verifier'), extra_params) def fetch_response(self, request): """Executes request and fetchs service response""" response = urlopen(request.to_url()) return '\n'.join(response.readlines()) def access_token(self, token): """Return request for access token value""" request = self.oauth_request(token, self.ACCESS_TOKEN_URL) return Token.from_string(self.fetch_response(request)) def user_data(self, access_token, *args, **kwargs): """Loads user data from service""" raise NotImplementedError('Implement in subclass') @property def consumer(self): """Setups consumer""" return OAuthConsumer(*self.get_key_and_secret()) class BaseOAuth2(BaseOAuth): """Base class for OAuth2 providers. OAuth2 draft details at: http://tools.ietf.org/html/draft-ietf-oauth-v2-10 Attributes: AUTHORIZATION_URL Authorization service url ACCESS_TOKEN_URL Token URL FORCE_STATE_CHECK Ensure state argument check (check issue #386 for further details) """ AUTHORIZATION_URL = None ACCESS_TOKEN_URL = None SCOPE_SEPARATOR = ' ' RESPONSE_TYPE = 'code' SCOPE_VAR_NAME = None DEFAULT_SCOPE = None FORCE_STATE_CHECK = True def csrf_token(self): """Generate csrf token to include as state parameter.""" return get_random_string(CSRF_KEY_LENGTH) def auth_url(self): """Return redirect url""" client_id, client_secret = self.get_key_and_secret() args = {'client_id': client_id, 'redirect_uri': self.redirect_uri} if self.FORCE_STATE_CHECK: state = self.csrf_token() args['state'] = state self.request.session[self.AUTH_BACKEND.name + '_state'] = state scope = self.get_scope() if scope: args['scope'] = self.SCOPE_SEPARATOR.join(self.get_scope()) if self.RESPONSE_TYPE: args['response_type'] = self.RESPONSE_TYPE args.update(self.auth_extra_arguments()) return self.AUTHORIZATION_URL + '?' + urlencode(args) def auth_complete(self, *args, **kwargs): """Completes loging process, must return user instance""" if self.data.get('error'): error = self.data.get('error_description') or self.data['error'] raise AuthFailed(self, error) if self.FORCE_STATE_CHECK: if 'state' not in self.data: raise AuthMissingParameter(self, 'state') state = self.request.session[self.AUTH_BACKEND.name + '_state'] if not constant_time_compare(self.data['state'], state): raise AuthForbidden(self) client_id, client_secret = self.get_key_and_secret() params = {'grant_type': 'authorization_code', # request auth code 'code': self.data.get('code', ''), # server response code 'client_id': client_id, 'client_secret': client_secret, 'redirect_uri': self.redirect_uri} headers = {'Content-Type': 'application/x-www-form-urlencoded', 'Accept': 'application/json'} request = Request(self.ACCESS_TOKEN_URL, data=urlencode(params), headers=headers) try: response = simplejson.loads(urlopen(request).read()) except HTTPError, e: if e.code == 400: raise AuthCanceled(self) else: raise except (ValueError, KeyError): raise AuthUnknownError(self) if response.get('error'): error = response.get('error_description') or response.get('error') raise AuthFailed(self, error) else: data = self.user_data(response['access_token'], response) response.update(data or {}) kwargs.update({ 'auth': self, 'response': response, self.AUTH_BACKEND.name: True }) return authenticate(*args, **kwargs) def get_scope(self): """Return list with needed access scope""" scope = self.DEFAULT_SCOPE or [] if self.SCOPE_VAR_NAME: scope = scope + setting(self.SCOPE_VAR_NAME, []) return scope # Backend loading was previously performed via the # SOCIAL_AUTH_IMPORT_BACKENDS setting - as it's no longer used, # provide a deprecation warning. if setting('SOCIAL_AUTH_IMPORT_BACKENDS'): from warnings import warn warn("SOCIAL_AUTH_IMPORT_SOURCES is deprecated") # Cache for discovered backends. BACKENDSCACHE = {} def get_backends(force_load=False): """ Entry point to the BACKENDS cache. If BACKENDSCACHE hasn't been populated, each of the modules referenced in AUTHENTICATION_BACKENDS is imported and checked for a BACKENDS definition and if enabled, added to the cache. Previously all backends were attempted to be loaded at import time of this module, which meant that backends that subclass bases found in this module would not have the chance to be loaded by the time they were added to this module's BACKENDS dict. See: https://github.com/omab/django-social-auth/issues/204 This new approach ensures that backends are allowed to subclass from bases in this module and still be picked up. A force_load boolean arg is also provided so that get_backend below can retry a requested backend that may not yet be discovered. """ if not BACKENDSCACHE or force_load: for auth_backend in setting('AUTHENTICATION_BACKENDS'): mod, cls_name = auth_backend.rsplit('.', 1) module = import_module(mod) backend = getattr(module, cls_name) if issubclass(backend, SocialAuthBackend): name = backend.name backends = getattr(module, 'BACKENDS', {}) if name in backends and backends[name].enabled(): BACKENDSCACHE[name] = backends[name] return BACKENDSCACHE def get_backend(name, *args, **kwargs): """Returns a backend by name. Backends are stored in the BACKENDSCACHE cache dict. If not found, each of the modules referenced in AUTHENTICATION_BACKENDS is imported and checked for a BACKENDS definition. If the named backend is found in the module's BACKENDS definition, it's then stored in the cache for future access. """ try: # Cached backend which has previously been discovered. return BACKENDSCACHE[name](*args, **kwargs) except KeyError: # Force a reload of BACKENDS to ensure a missing # backend hasn't been missed. get_backends(force_load=True) try: return BACKENDSCACHE[name](*args, **kwargs) except KeyError: return None BACKENDS = { 'openid': OpenIdAuth }<|fim▁end|>
class OAuthBackend(SocialAuthBackend): """OAuth authentication backend base class. EXTRA_DATA defines a set of name that will be stored in
<|file_name|>cases.py<|end_file_name|><|fim▁begin|>import pingo ''' In order to use this set of cases, it is necessary to set the following attributes on your TestCase setUp: self.analog_input_pin_number = 0 self.expected_analog_input = 1004 self.expected_analog_ratio = 0.98 ''' <|fim▁hole|> Wire a 10K Ohm resistence from the AnalogPin to the GND. Then wire a 200 Ohm from the AnalogPin to the VND. This schema will provide a read of ~98% ''' def test_200ohmRead(self): pin = self.board.pins[self.analog_input_pin_number] pin.mode = pingo.ANALOG _input = pin.value # print "Value Read: ", _input assert self.expected_analog_input - 3 <= _input <= self.expected_analog_input + 3 def test_pin_ratio(self): pin = self.board.pins[self.analog_input_pin_number] pin.mode = pingo.ANALOG bits_resolution = (2 ** pin.bits) - 1 _input = pin.ratio(0, bits_resolution, 0.0, 1.0) # print "Value Read: ", _input # Two decimal places check assert abs(_input - self.expected_analog_ratio) < 10e-1 class AnalogExceptions(object): def test_wrong_output_mode(self): pin = self.board.pins[self.analog_input_pin_number] with self.assertRaises(pingo.ModeNotSuported): pin.mode = pingo.OUT<|fim▁end|>
class AnalogReadBasics(object): '''
<|file_name|>sourcefilter.pipe.ts<|end_file_name|><|fim▁begin|>import { Pipe, PipeTransform} from '@angular/core'; @Pipe({ name: 'sourceFilter' }) export class sourceFilter implements PipeTransform { <|fim▁hole|> public regexp: RegExp; public inputsource : string; transform(value: any, args: any[]): any { if(!args[0]){ alert(args[0]); return value; }else{ this.inputsource = args.toString(); this.regexp= new RegExp(this.inputsource); return value.filter(item=>{if(this.regexp.test(item.toLowerCase())){return true;}else{return false;}}); } } }<|fim▁end|>
<|file_name|>TestSyncGConfFolder.py<|end_file_name|><|fim▁begin|>#common sets up the conduit environment from common import * #setup test test = SimpleSyncTest() #Setup the key to sync gconf = test.get_dataprovider("GConfTwoWay") gconf.module.whitelist = ['/apps/metacity/general/num_workspaces'] folder = test.get_dataprovider("TestFolderTwoWay") test.prepare(gconf, folder) test.set_two_way_policy({"conflict":"ask","deleted":"ask"}) test.set_two_way_sync(True) a = test.get_source_count() b = test.get_sink_count() ok("Got items to sync (%s,%s)" % (a,b), a == 1 and b == 0) for i in (1,2,3,4): if i > 1: #Now modify the file f = folder.module.get( folder.module.get_all()[0] ) f._set_file_mtime(datetime.datetime(2008,1,i)) a,b = test.sync() aborted,errored,conflicted = test.get_sync_result() ok("Sync #%s: Completed without conflicts" % i, aborted == False and errored == False and conflicted == False) ok("Sync #%s: All items (%s,%s)" % (i,a,b), a == b and a == 1)<|fim▁hole|><|fim▁end|>
finished()
<|file_name|>cmd_listobj.cc<|end_file_name|><|fim▁begin|>/* * Copyright (c) 2012 Stanford University * * Permission to use, copy, modify, and distribute this software for any * purpose with or without fee is hereby granted, provided that the above * copyright notice and this permission notice appear in all copies. * * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR(S) DISCLAIM ALL WARRANTIES * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL AUTHORS BE LIABLE FOR * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. */ #include <stdint.h> #include <set> #include <string> #include <iostream> #include <oriutil/debug.h> #include <ori/localrepo.h> using namespace std; extern LocalRepo repository; int<|fim▁hole|> for (it = objects.begin(); it != objects.end(); it++) { const char *type; switch ((*it).type) { case ObjectInfo::Commit: type = "Commit"; break; case ObjectInfo::Tree: type = "Tree"; break; case ObjectInfo::Blob: type = "Blob"; break; case ObjectInfo::LargeBlob: type = "LargeBlob"; break; case ObjectInfo::Purged: type = "Purged"; break; default: cout << "Unknown object type (id " << (*it).hash.hex() << ")!" << endl; PANIC(); } cout << (*it).hash.hex() << " # " << type << endl; } return 0; }<|fim▁end|>
cmd_listobj(int argc, char * const argv[]) { set<ObjectInfo> objects = repository.listObjects(); set<ObjectInfo>::iterator it;
<|file_name|>libstdc++.a-gdb.py<|end_file_name|><|fim▁begin|># -*- python -*- # Copyright (C) 2009 Free Software Foundation, Inc. # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. import sys import gdb import os import os.path pythondir = '/opt/codesourcery/arm-none-eabi/share/gcc-4.5.1/python' libdir = '/opt/codesourcery/arm-none-eabi/lib/armv6-m' # This file might be loaded when there is no current objfile. This # can happen if the user loads it manually. In this case we don't # update sys.path; instead we just hope the user managed to do that # beforehand. if gdb.current_objfile () is not None: # Update module path. We want to find the relative path from libdir # to pythondir, and then we want to apply that relative path to the # directory holding the objfile with which this file is associated. # This preserves relocatability of the gcc tree. # Do a simple normalization that removes duplicate separators. pythondir = os.path.normpath (pythondir) libdir = os.path.normpath (libdir) prefix = os.path.commonprefix ([libdir, pythondir]) # In some bizarre configuration we might have found a match in the # middle of a directory name. if prefix[-1] != '/': prefix = os.path.dirname (prefix) + '/' # Strip off the prefix.<|fim▁hole|> # Compute the ".."s needed to get from libdir to the prefix. dotdots = ('..' + os.sep) * len (libdir.split (os.sep)) objfile = gdb.current_objfile ().filename dir = os.path.join (os.path.dirname (objfile), dotdots, pythondir) if not dir in sys.path: sys.path.insert(0, dir) # Load the pretty-printers. from libstdcxx.v6.printers import register_libstdcxx_printers register_libstdcxx_printers (gdb.current_objfile ())<|fim▁end|>
pythondir = pythondir[len (prefix):] libdir = libdir[len (prefix):]
<|file_name|>lib.rs<|end_file_name|><|fim▁begin|>mod map;<|fim▁hole|> pub use map::{pos, Map, Pos};<|fim▁end|>
<|file_name|>bitcoin_id_ID.ts<|end_file_name|><|fim▁begin|><?xml version="1.0" ?><!DOCTYPE TS><TS language="id_ID" version="2.1"> <context> <name>AboutDialog</name> <message> <location filename="../forms/aboutdialog.ui" line="+14"/> <source>About phreak</source> <translation type="unfinished"/> </message> <message> <location line="+39"/> <source>&lt;b&gt;phreak&lt;/b&gt; version</source> <translation type="unfinished"/> </message> <message> <location line="+41"/> <source>Copyright © 2009-2014 The Bitcoin developers Copyright © 2012-2014 The NovaCoin developers Copyright © 2014 The phreak developers</source> <translation type="unfinished"/> </message> <message> <location line="+15"/> <source> This is experimental software. Distributed under the MIT/X11 software license, see the accompanying file COPYING or http://www.opensource.org/licenses/mit-license.php. This product includes software developed by the OpenSSL Project for use in the OpenSSL Toolkit (http://www.openssl.org/) and cryptographic software written by Eric Young ([email protected]) and UPnP software written by Thomas Bernard.</source> <translation type="unfinished"/> </message> </context> <context> <name>AddressBookPage</name> <message> <location filename="../forms/addressbookpage.ui" line="+14"/> <source>Address Book</source> <translation type="unfinished"/> </message> <message> <location line="+22"/> <source>Double-click to edit address or label</source> <translation>Klik-ganda untuk mengubah alamat atau label</translation> </message> <message> <location line="+27"/> <source>Create a new address</source> <translation>Buat alamat baru</translation> </message> <message> <location line="+14"/> <source>Copy the currently selected address to the system clipboard</source> <translation>Salin alamat yang dipilih ke clipboard</translation> </message> <message> <location line="-11"/> <source>&amp;New Address</source> <translation type="unfinished"/> </message> <message> <location line="-46"/> <source>These are your phreak addresses for receiving payments. You may want to give a different one to each sender so you can keep track of who is paying you.</source> <translation type="unfinished"/> </message> <message> <location line="+60"/> <source>&amp;Copy Address</source> <translation>&amp;Salin Alamat</translation> </message> <message> <location line="+11"/> <source>Show &amp;QR Code</source> <translation type="unfinished"/> </message> <message> <location line="+11"/> <source>Sign a message to prove you own a phreak address</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Sign &amp;Message</source> <translation type="unfinished"/> </message> <message> <location line="+25"/> <source>Delete the currently selected address from the list</source> <translation type="unfinished"/> </message> <message> <location line="-14"/> <source>Verify a message to ensure it was signed with a specified phreak address</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>&amp;Verify Message</source> <translation type="unfinished"/> </message> <message> <location line="+14"/> <source>&amp;Delete</source> <translation>&amp;Hapus</translation> </message> <message> <location filename="../addressbookpage.cpp" line="+65"/> <source>Copy &amp;Label</source> <translation>Salin &amp;Label</translation> </message> <message> <location line="+2"/> <source>&amp;Edit</source> <translation>&amp;Ubah</translation> </message> <message> <location line="+250"/> <source>Export Address Book Data</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Comma separated file (*.csv)</source> <translation>File CSV (*.csv)</translation> </message> <message> <location line="+13"/> <source>Error exporting</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>Could not write to file %1.</source> <translation type="unfinished"/> </message> </context> <context> <name>AddressTableModel</name> <message> <location filename="../addresstablemodel.cpp" line="+144"/> <source>Label</source> <translation>Label</translation> </message> <message> <location line="+0"/> <source>Address</source> <translation>Alamat</translation> </message> <message> <location line="+36"/> <source>(no label)</source> <translation>(tidak ada label)</translation> </message> </context> <context> <name>AskPassphraseDialog</name> <message> <location filename="../forms/askpassphrasedialog.ui" line="+26"/> <source>Passphrase Dialog</source> <translation>Dialog Kata kunci</translation> </message> <message> <location line="+21"/> <source>Enter passphrase</source> <translation>Masukkan kata kunci</translation> </message> <message> <location line="+14"/> <source>New passphrase</source> <translation>Kata kunci baru</translation> </message> <message> <location line="+14"/> <source>Repeat new passphrase</source> <translation>Ulangi kata kunci baru</translation> </message> <message> <location line="+33"/> <source>Serves to disable the trivial sendmoney when OS account compromised. Provides no real security.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>For staking only</source> <translation type="unfinished"/> </message> <message> <location filename="../askpassphrasedialog.cpp" line="+35"/> <source>Enter the new passphrase to the wallet.&lt;br/&gt;Please use a passphrase of &lt;b&gt;10 or more random characters&lt;/b&gt;, or &lt;b&gt;eight or more words&lt;/b&gt;.</source> <translation>Masukkan kata kunci baru ke dompet.&lt;br/&gt;Mohon gunakan kata kunci dengan &lt;b&gt;10 karakter atau lebih dengan acak&lt;/b&gt;, atau &lt;b&gt;delapan kata atau lebih&lt;/b&gt;.</translation> </message> <message> <location line="+1"/> <source>Encrypt wallet</source> <translation>Enkripsi dompet</translation> </message> <message> <location line="+7"/> <source>This operation needs your wallet passphrase to unlock the wallet.</source> <translation>Operasi ini memerlukan kata kunci dompet Anda untuk membuka dompet ini.</translation> </message> <message> <location line="+5"/> <source>Unlock wallet</source> <translation>Buka dompet</translation> </message> <message> <location line="+3"/> <source>This operation needs your wallet passphrase to decrypt the wallet.</source> <translation>Operasi ini memerlukan kata kunci dompet Anda untuk mendekripsi dompet ini.</translation> </message> <message> <location line="+5"/> <source>Decrypt wallet</source> <translation>Dekripsi dompet</translation> </message> <message> <location line="+3"/> <source>Change passphrase</source> <translation>Ubah kata kunci</translation> </message> <message> <location line="+1"/> <source>Enter the old and new passphrase to the wallet.</source> <translation>Masukkan kata kunci lama dan baru ke dompet ini.</translation> </message> <message> <location line="+46"/> <source>Confirm wallet encryption</source> <translation>Konfirmasi enkripsi dompet</translation> </message> <message> <location line="+1"/> <source>Warning: If you encrypt your wallet and lose your passphrase, you will &lt;b&gt;LOSE ALL OF YOUR COINS&lt;/b&gt;!</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>Are you sure you wish to encrypt your wallet?</source> <translation type="unfinished"/> </message> <message> <location line="+15"/> <source>IMPORTANT: Any previous backups you have made of your wallet file should be replaced with the newly generated, encrypted wallet file. For security reasons, previous backups of the unencrypted wallet file will become useless as soon as you start using the new, encrypted wallet.</source> <translation type="unfinished"/> </message> <message> <location line="+103"/> <location line="+24"/> <source>Warning: The Caps Lock key is on!</source> <translation type="unfinished"/> </message> <message> <location line="-133"/> <location line="+60"/> <source>Wallet encrypted</source> <translation>Dompet terenkripsi</translation> </message> <message> <location line="-58"/> <source>phreak will close now to finish the encryption process. Remember that encrypting your wallet cannot fully protect your coins from being stolen by malware infecting your computer.</source> <translation type="unfinished"/> </message> <message> <location line="+13"/> <location line="+7"/> <location line="+44"/> <location line="+6"/> <source>Wallet encryption failed</source> <translation>Enkripsi dompet gagal</translation> </message> <message> <location line="-56"/> <source>Wallet encryption failed due to an internal error. Your wallet was not encrypted.</source> <translation>Enkripsi dompet gagal karena kesalahan internal. Dompet Anda tidak dienkripsi.</translation> </message> <message> <location line="+7"/> <location line="+50"/> <source>The supplied passphrases do not match.</source> <translation>Kata kunci yang dimasukkan tidak cocok.</translation> </message> <message> <location line="-38"/> <source>Wallet unlock failed</source> <translation>Gagal buka dompet</translation> </message> <message> <location line="+1"/> <location line="+12"/> <location line="+19"/> <source>The passphrase entered for the wallet decryption was incorrect.</source> <translation>Kata kunci yang dimasukkan untuk dekripsi dompet tidak cocok.</translation> </message> <message> <location line="-20"/> <source>Wallet decryption failed</source> <translation>Dekripsi dompet gagal</translation> </message> <message> <location line="+14"/> <source>Wallet passphrase was successfully changed.</source> <translation type="unfinished"/> </message> </context> <context> <name>BitcoinGUI</name> <message> <location filename="../bitcoingui.cpp" line="+282"/> <source>Sign &amp;message...</source> <translation>Pesan &amp;penanda...</translation> </message> <message> <location line="+251"/> <source>Synchronizing with network...</source> <translation>Sinkronisasi dengan jaringan...</translation> </message> <message> <location line="-319"/> <source>&amp;Overview</source> <translation>&amp;Kilasan</translation> </message> <message> <location line="+1"/> <source>Show general overview of wallet</source> <translation>Tampilkan kilasan umum dari dompet</translation> </message> <message> <location line="+17"/> <source>&amp;Transactions</source> <translation>&amp;Transaksi</translation> </message> <message> <location line="+1"/> <source>Browse transaction history</source> <translation>Jelajah sejarah transaksi</translation> </message> <message> <location line="+5"/> <source>&amp;Address Book</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Edit the list of stored addresses and labels</source> <translation type="unfinished"/> </message> <message> <location line="-13"/> <source>&amp;Receive coins</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Show the list of addresses for receiving payments</source> <translation type="unfinished"/> </message> <message> <location line="-7"/> <source>&amp;Send coins</source> <translation type="unfinished"/> </message> <message> <location line="+35"/> <source>E&amp;xit</source> <translation>K&amp;eluar</translation> </message> <message> <location line="+1"/> <source>Quit application</source> <translation>Keluar dari aplikasi</translation> </message> <message> <location line="+6"/> <source>Show information about phreak</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>About &amp;Qt</source> <translation>Mengenai &amp;Qt</translation> </message> <message> <location line="+1"/> <source>Show information about Qt</source> <translation>Tampilkan informasi mengenai Qt</translation> </message> <message> <location line="+2"/> <source>&amp;Options...</source> <translation>&amp;Pilihan...</translation> </message> <message> <location line="+4"/> <source>&amp;Encrypt Wallet...</source> <translation>%Enkripsi Dompet...</translation> </message> <message> <location line="+3"/> <source>&amp;Backup Wallet...</source> <translation>&amp;Cadangkan Dompet...</translation> </message> <message> <location line="+2"/> <source>&amp;Change Passphrase...</source> <translation>&amp;Ubah Kata Kunci...</translation> </message> <message numerus="yes"> <location line="+259"/> <source>~%n block(s) remaining</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message> <location line="+6"/> <source>Downloaded %1 of %2 blocks of transaction history (%3% done).</source> <translation type="unfinished"/> </message> <message> <location line="-256"/> <source>&amp;Export...</source> <translation type="unfinished"/> </message> <message> <location line="-64"/> <source>Send coins to a phreak address</source> <translation type="unfinished"/> </message> <message> <location line="+47"/> <source>Modify configuration options for phreak</source> <translation type="unfinished"/> </message> <message> <location line="+18"/> <source>Export the data in the current tab to a file</source> <translation type="unfinished"/> </message> <message> <location line="-14"/> <source>Encrypt or decrypt wallet</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Backup wallet to another location</source> <translation>Cadangkan dompet ke lokasi lain</translation> </message> <message> <location line="+2"/> <source>Change the passphrase used for wallet encryption</source> <translation>Ubah kata kunci yang digunakan untuk enkripsi dompet</translation> </message> <message> <location line="+10"/> <source>&amp;Debug window</source> <translation>&amp;Jendela Debug</translation> </message> <message> <location line="+1"/> <source>Open debugging and diagnostic console</source> <translation>Buka konsol debug dan diagnosa</translation> </message> <message> <location line="-5"/> <source>&amp;Verify message...</source> <translation>&amp;Verifikasi pesan...</translation> </message> <message> <location line="-202"/> <source>phreak</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>Wallet</source> <translation>Dompet</translation> </message> <message> <location line="+180"/> <source>&amp;About phreak</source> <translation type="unfinished"/> </message> <message> <location line="+9"/> <source>&amp;Show / Hide</source> <translation type="unfinished"/> </message> <message> <location line="+9"/> <source>Unlock wallet</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>&amp;Lock Wallet</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Lock wallet</source> <translation type="unfinished"/> </message> <message> <location line="+35"/> <source>&amp;File</source> <translation>&amp;Berkas</translation> </message> <message> <location line="+8"/> <source>&amp;Settings</source> <translation>&amp;Pengaturan</translation> </message> <message> <location line="+8"/> <source>&amp;Help</source> <translation>&amp;Bantuan</translation> </message> <message> <location line="+12"/> <source>Tabs toolbar</source> <translation>Baris tab</translation> </message> <message> <location line="+8"/> <source>Actions toolbar</source> <translation type="unfinished"/> </message> <message> <location line="+13"/> <location line="+9"/> <source>[testnet]</source> <translation>[testnet]</translation> </message> <message> <location line="+0"/> <location line="+60"/> <source>phreak client</source> <translation type="unfinished"/> </message> <message numerus="yes"> <location line="+75"/> <source>%n active connection(s) to phreak network</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message> <location line="+40"/> <source>Downloaded %1 blocks of transaction history.</source> <translation type="unfinished"/> </message> <message> <location line="+413"/> <source>Staking.&lt;br&gt;Your weight is %1&lt;br&gt;Network weight is %2&lt;br&gt;Expected time to earn reward is %3</source> <translation type="unfinished"/> </message> <message> <location line="+6"/> <source>Not staking because wallet is locked</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Not staking because wallet is offline</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Not staking because wallet is syncing</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Not staking because you don&apos;t have mature coins</source> <translation type="unfinished"/> </message> <message numerus="yes"> <location line="-403"/> <source>%n second(s) ago</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message> <location line="-312"/> <source>About phreak card</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Show information about phreak card</source> <translation type="unfinished"/> </message> <message> <location line="+18"/> <source>&amp;Unlock Wallet...</source> <translation type="unfinished"/> </message> <message numerus="yes"> <location line="+297"/> <source>%n minute(s) ago</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message numerus="yes"> <location line="+4"/> <source>%n hour(s) ago</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message numerus="yes"> <location line="+4"/> <source>%n day(s) ago</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message> <location line="+6"/> <source>Up to date</source> <translation>Terbaru</translation> </message> <message> <location line="+7"/> <source>Catching up...</source> <translation>Menyusul...</translation> </message> <message> <location line="+10"/> <source>Last received block was generated %1.</source> <translation type="unfinished"/> </message> <message> <location line="+59"/> <source>This transaction is over the size limit. You can still send it for a fee of %1, which goes to the nodes that process your transaction and helps to support the network. Do you want to pay the fee?</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Confirm transaction fee</source> <translation type="unfinished"/> </message> <message> <location line="+27"/> <source>Sent transaction</source> <translation>Transaksi terkirim</translation> </message> <message> <location line="+1"/> <source>Incoming transaction</source> <translation>Transaksi diterima</translation> </message> <message> <location line="+1"/> <source>Date: %1 Amount: %2 Type: %3 Address: %4 </source> <translation>Tanggal: %1 Jumlah: %2 Jenis: %3 Alamat: %4 </translation> </message> <message> <location line="+100"/> <location line="+15"/> <source>URI handling</source> <translation type="unfinished"/> </message> <message> <location line="-15"/> <location line="+15"/> <source>URI can not be parsed! This can be caused by an invalid phreak address or malformed URI parameters.</source> <translation type="unfinished"/> </message> <message> <location line="+18"/> <source>Wallet is &lt;b&gt;encrypted&lt;/b&gt; and currently &lt;b&gt;unlocked&lt;/b&gt;</source> <translation>Dompet saat ini &lt;b&gt;terenkripsi&lt;/b&gt; dan &lt;b&gt;terbuka&lt;/b&gt;</translation> </message> <message> <location line="+10"/> <source>Wallet is &lt;b&gt;encrypted&lt;/b&gt; and currently &lt;b&gt;locked&lt;/b&gt;</source> <translation>Dompet saat ini &lt;b&gt;terenkripsi&lt;/b&gt; dan &lt;b&gt;terkunci&lt;/b&gt;</translation> </message> <message> <location line="+25"/> <source>Backup Wallet</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>Wallet Data (*.dat)</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Backup Failed</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>There was an error trying to save the wallet data to the new location.</source> <translation type="unfinished"/> </message> <message numerus="yes"> <location line="+76"/> <source>%n second(s)</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message numerus="yes"> <location line="+4"/> <source>%n minute(s)</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message numerus="yes"> <location line="+4"/> <source>%n hour(s)</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message numerus="yes"> <location line="+4"/> <source>%n day(s)</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message> <location line="+18"/> <source>Not staking</source> <translation type="unfinished"/> </message> <message> <location filename="../bitcoin.cpp" line="+109"/> <source>A fatal error occurred. phreak can no longer continue safely and will quit.</source> <translation type="unfinished"/> </message> </context> <context> <name>ClientModel</name> <message> <location filename="../clientmodel.cpp" line="+90"/> <source>Network Alert</source> <translation>Notifikasi Jaringan</translation> </message> </context> <context> <name>CoinControlDialog</name> <message> <location filename="../forms/coincontroldialog.ui" line="+14"/> <source>Coin Control</source> <translation type="unfinished"/> </message> <message> <location line="+31"/> <source>Quantity:</source> <translation type="unfinished"/> </message> <message> <location line="+32"/> <source>Bytes:</source> <translation type="unfinished"/> </message> <message> <location line="+48"/> <source>Amount:</source> <translation>Jumlah:</translation> </message> <message> <location line="+32"/> <source>Priority:</source> <translation type="unfinished"/> </message> <message> <location line="+48"/> <source>Fee:</source> <translation type="unfinished"/> </message> <message> <location line="+35"/> <source>Low Output:</source> <translation type="unfinished"/> </message> <message> <location filename="../coincontroldialog.cpp" line="+551"/> <source>no</source> <translation type="unfinished"/> </message> <message> <location filename="../forms/coincontroldialog.ui" line="+51"/> <source>After Fee:</source> <translation type="unfinished"/> </message> <message> <location line="+35"/> <source>Change:</source> <translation type="unfinished"/> </message> <message> <location line="+69"/> <source>(un)select all</source> <translation type="unfinished"/> </message> <message> <location line="+13"/> <source>Tree mode</source> <translation type="unfinished"/> </message> <message> <location line="+16"/> <source>List mode</source> <translation type="unfinished"/> </message> <message> <location line="+45"/> <source>Amount</source> <translation>Jumlah</translation> </message> <message> <location line="+5"/> <source>Label</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Address</source> <translation>Alamat</translation> </message> <message> <location line="+5"/> <source>Date</source> <translation>Tanggal</translation> </message> <message> <location line="+5"/> <source>Confirmations</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Confirmed</source> <translation>Terkonfirmasi</translation> </message> <message> <location line="+5"/> <source>Priority</source> <translation type="unfinished"/> </message> <message> <location filename="../coincontroldialog.cpp" line="-515"/> <source>Copy address</source> <translation>Salin alamat</translation> </message> <message> <location line="+1"/> <source>Copy label</source> <translation>Salin label</translation> </message> <message> <location line="+1"/> <location line="+26"/> <source>Copy amount</source> <translation>Salin jumlah</translation> </message> <message> <location line="-25"/> <source>Copy transaction ID</source> <translation type="unfinished"/> </message> <message> <location line="+24"/> <source>Copy quantity</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Copy fee</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Copy after fee</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Copy bytes</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Copy priority</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Copy low output</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Copy change</source> <translation type="unfinished"/> </message> <message> <location line="+317"/> <source>highest</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>high</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>medium-high</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>medium</source> <translation type="unfinished"/> </message> <message> <location line="+4"/> <source>low-medium</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>low</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>lowest</source> <translation type="unfinished"/> </message> <message> <location line="+155"/> <source>DUST</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>yes</source> <translation type="unfinished"/> </message> <message> <location line="+10"/> <source>This label turns red, if the transaction size is bigger than 10000 bytes. This means a fee of at least %1 per kb is required. Can vary +/- 1 Byte per input.</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Transactions with higher priority get more likely into a block. This label turns red, if the priority is smaller than &quot;medium&quot;. This means a fee of at least %1 per kb is required.</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>This label turns red, if any recipient receives an amount smaller than %1. This means a fee of at least %2 is required. Amounts below 0.546 times the minimum relay fee are shown as DUST.</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>This label turns red, if the change is smaller than %1. This means a fee of at least %2 is required.</source> <translation type="unfinished"/> </message> <message> <location line="+37"/> <location line="+66"/> <source>(no label)</source> <translation>(tidak ada label)</translation> </message> <message> <location line="-9"/> <source>change from %1 (%2)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>(change)</source> <translation type="unfinished"/> </message> </context> <context> <name>EditAddressDialog</name> <message> <location filename="../forms/editaddressdialog.ui" line="+14"/> <source>Edit Address</source> <translation>Ubah Alamat</translation> </message> <message> <location line="+11"/> <source>&amp;Label</source> <translation>&amp;Label</translation> </message> <message> <location line="+10"/> <source>The label associated with this address book entry</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>&amp;Address</source> <translation>&amp;Alamat</translation> </message> <message> <location line="+10"/> <source>The address associated with this address book entry. This can only be modified for sending addresses.</source> <translation type="unfinished"/> </message> <message> <location filename="../editaddressdialog.cpp" line="+20"/> <source>New receiving address</source> <translation>Alamat menerima baru</translation> </message> <message> <location line="+4"/> <source>New sending address</source> <translation>Alamat mengirim baru</translation> </message> <message> <location line="+3"/> <source>Edit receiving address</source> <translation>Ubah alamat menerima</translation> </message> <message> <location line="+4"/> <source>Edit sending address</source> <translation>Ubah alamat mengirim</translation> </message> <message> <location line="+76"/> <source>The entered address &quot;%1&quot; is already in the address book.</source> <translation>Alamat yang dimasukkan &quot;%1&quot; sudah ada di dalam buku alamat.</translation> </message> <message> <location line="-5"/> <source>The entered address &quot;%1&quot; is not a valid phreak address.</source> <translation type="unfinished"/> </message> <message> <location line="+10"/> <source>Could not unlock wallet.</source> <translation>Tidak dapat membuka dompet.</translation> </message> <message> <location line="+5"/> <source>New key generation failed.</source> <translation>Pembuatan kunci baru gagal.</translation> </message> </context> <context> <name>GUIUtil::HelpMessageBox</name> <message> <location filename="../guiutil.cpp" line="+420"/> <location line="+12"/> <source>phreak-Qt</source> <translation type="unfinished"/> </message> <message> <location line="-12"/> <source>version</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Usage:</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>command-line options</source> <translation type="unfinished"/> </message> <message> <location line="+4"/> <source>UI options</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Set language, for example &quot;de_DE&quot; (default: system locale)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Start minimized</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Show splash screen on startup (default: 1)</source> <translation type="unfinished"/> </message> </context> <context> <name>OptionsDialog</name> <message> <location filename="../forms/optionsdialog.ui" line="+14"/> <source>Options</source> <translation>Pilihan</translation> </message> <message> <location line="+16"/> <source>&amp;Main</source> <translation>&amp;Utama</translation> </message> <message> <location line="+6"/> <source>Optional transaction fee per kB that helps make sure your transactions are processed quickly. Most transactions are 1 kB. Fee 0.01 recommended.</source> <translation type="unfinished"/> </message> <message> <location line="+15"/> <source>Pay transaction &amp;fee</source> <translation>Bayar &amp;biaya transaksi</translation> </message> <message> <location line="+31"/> <source>Reserved amount does not participate in staking and is therefore spendable at any time.</source> <translation type="unfinished"/> </message> <message> <location line="+15"/> <source>Reserve</source> <translation type="unfinished"/> </message> <message> <location line="+31"/> <source>Automatically start phreak after logging in to the system.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>&amp;Start phreak on system login</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>Detach block and address databases at shutdown. This means they can be moved to another data directory, but it slows down shutdown. The wallet is always detached.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>&amp;Detach databases at shutdown</source> <translation type="unfinished"/> </message> <message> <location line="+21"/> <source>&amp;Network</source> <translation>&amp;Jaringan</translation> </message> <message> <location line="+6"/> <source>Automatically open the phreak client port on the router. This only works when your router supports UPnP and it is enabled.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Map port using &amp;UPnP</source> <translation>Petakan port dengan &amp;UPnP</translation> </message> <message> <location line="+7"/> <source>Connect to the phreak network through a SOCKS proxy (e.g. when connecting through Tor).</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>&amp;Connect through SOCKS proxy:</source> <translation type="unfinished"/> </message> <message> <location line="+9"/> <source>Proxy &amp;IP:</source> <translation>IP Proxy:</translation> </message> <message> <location line="+19"/> <source>IP address of the proxy (e.g. 127.0.0.1)</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>&amp;Port:</source> <translation>&amp;Port:</translation> </message> <message> <location line="+19"/> <source>Port of the proxy (e.g. 9050)</source> <translation>Port proxy (cth. 9050)</translation> </message> <message> <location line="+7"/> <source>SOCKS &amp;Version:</source> <translation>Versi &amp;SOCKS:</translation> </message> <message> <location line="+13"/> <source>SOCKS version of the proxy (e.g. 5)</source> <translation>Versi SOCKS proxy (cth. 5)</translation> </message> <message> <location line="+36"/> <source>&amp;Window</source> <translation>&amp;Jendela</translation> </message> <message> <location line="+6"/> <source>Show only a tray icon after minimizing the window.</source> <translation>Hanya tampilkan ikon tray setelah meminilisasi jendela</translation> </message> <message> <location line="+3"/> <source>&amp;Minimize to the tray instead of the taskbar</source> <translation>&amp;Meminilisasi ke tray daripada taskbar</translation> </message> <message> <location line="+7"/> <source>Minimize instead of exit the application when the window is closed. When this option is enabled, the application will be closed only after selecting Quit in the menu.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>M&amp;inimize on close</source> <translation>M&amp;eminilisasi saat tutup</translation> </message> <message> <location line="+21"/> <source>&amp;Display</source> <translation>&amp;Tampilan</translation> </message> <message> <location line="+8"/> <source>User Interface &amp;language:</source> <translation>&amp;Bahasa Antarmuka Pengguna:</translation> </message> <message> <location line="+13"/> <source>The user interface language can be set here. This setting will take effect after restarting phreak.</source> <translation type="unfinished"/> </message> <message> <location line="+11"/> <source>&amp;Unit to show amounts in:</source> <translation>&amp;Unit untuk menunjukkan jumlah:</translation> </message> <message> <location line="+13"/> <source>Choose the default subdivision unit to show in the interface and when sending coins.</source> <translation type="unfinished"/> </message> <message> <location line="+9"/> <source>Whether to show phreak addresses in the transaction list or not.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>&amp;Display addresses in transaction list</source> <translation>&amp;Tampilkan alamat dalam daftar transaksi</translation> </message> <message> <location line="+7"/> <source>Whether to show coin control features or not.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Display coin &amp;control features (experts only!)</source> <translation type="unfinished"/> </message> <message> <location line="+71"/> <source>&amp;OK</source> <translation>&amp;YA</translation> </message> <message> <location line="+7"/> <source>&amp;Cancel</source> <translation>&amp;Batal</translation> </message> <message> <location line="+10"/> <source>&amp;Apply</source> <translation type="unfinished"/> </message> <message> <location filename="../optionsdialog.cpp" line="+55"/> <source>default</source> <translation>standar</translation> </message> <message> <location line="+149"/> <location line="+9"/> <source>Warning</source> <translation type="unfinished"/> </message> <message> <location line="-9"/> <location line="+9"/> <source>This setting will take effect after restarting phreak.</source> <translation type="unfinished"/> </message> <message> <location line="+29"/> <source>The supplied proxy address is invalid.</source> <translation>Alamat proxy yang diisi tidak valid.</translation> </message> </context> <context> <name>OverviewPage</name> <message> <location filename="../forms/overviewpage.ui" line="+14"/> <source>Form</source> <translation>Formulir</translation> </message> <message> <location line="+33"/> <location line="+231"/> <source>The displayed information may be out of date. Your wallet automatically synchronizes with the phreak network after a connection is established, but this process has not completed yet.</source> <translation type="unfinished"/> </message> <message> <location line="-160"/> <source>Stake:</source> <translation type="unfinished"/> </message> <message> <location line="+29"/> <source>Unconfirmed:</source> <translation type="unfinished"/> </message> <message> <location line="-107"/> <source>Wallet</source> <translation>Dompet</translation> </message> <message> <location line="+49"/> <source>Spendable:</source> <translation type="unfinished"/> </message> <message> <location line="+16"/> <source>Your current spendable balance</source> <translation type="unfinished"/> </message> <message> <location line="+71"/> <source>Immature:</source> <translation type="unfinished"/> </message> <message> <location line="+13"/> <source>Mined balance that has not yet matured</source> <translation type="unfinished"/> </message> <message> <location line="+20"/> <source>Total:</source> <translation type="unfinished"/> </message> <message> <location line="+16"/> <source>Your current total balance</source> <translation type="unfinished"/> </message> <message> <location line="+46"/> <source>&lt;b&gt;Recent transactions&lt;/b&gt;</source> <translation>&lt;b&gt;Transaksi sebelumnya&lt;/b&gt;</translation> </message> <message> <location line="-108"/> <source>Total of transactions that have yet to be confirmed, and do not yet count toward the current balance</source> <translation type="unfinished"/> </message> <message> <location line="-29"/> <source>Total of coins that was staked, and do not yet count toward the current balance</source> <translation type="unfinished"/> </message> <message> <location filename="../overviewpage.cpp" line="+113"/> <location line="+1"/> <source>out of sync</source> <translation>tidak tersinkron</translation> </message> </context> <context> <name>QRCodeDialog</name> <message> <location filename="../forms/qrcodedialog.ui" line="+14"/> <source>QR Code Dialog</source> <translation type="unfinished"/> </message> <message> <location line="+59"/> <source>Request Payment</source> <translation type="unfinished"/> </message> <message> <location line="+56"/> <source>Amount:</source> <translation type="unfinished"/> </message> <message> <location line="-44"/> <source>Label:</source> <translation type="unfinished"/> </message> <message> <location line="+19"/> <source>Message:</source> <translation type="unfinished"/> </message> <message> <location line="+71"/> <source>&amp;Save As...</source> <translation type="unfinished"/> </message> <message> <location filename="../qrcodedialog.cpp" line="+62"/> <source>Error encoding URI into QR Code.</source> <translation type="unfinished"/> </message> <message> <location line="+40"/> <source>The entered amount is invalid, please check.</source> <translation type="unfinished"/> </message> <message> <location line="+23"/> <source>Resulting URI too long, try to reduce the text for label / message.</source> <translation type="unfinished"/> </message> <message> <location line="+25"/> <source>Save QR Code</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>PNG Images (*.png)</source> <translation type="unfinished"/> </message> </context> <context> <name>RPCConsole</name> <message> <location filename="../forms/rpcconsole.ui" line="+46"/> <source>Client name</source> <translation>Nama Klien</translation> </message> <message> <location line="+10"/> <location line="+23"/> <location line="+26"/> <location line="+23"/> <location line="+23"/> <location line="+36"/> <location line="+53"/> <location line="+23"/> <location line="+23"/> <location filename="../rpcconsole.cpp" line="+348"/> <source>N/A</source> <translation>T/S</translation> </message> <message> <location line="-217"/> <source>Client version</source> <translation>Versi Klien</translation> </message> <message> <location line="-45"/> <source>&amp;Information</source> <translation>&amp;Informasi</translation> </message> <message> <location line="+68"/> <source>Using OpenSSL version</source> <translation type="unfinished"/> </message> <message> <location line="+49"/> <source>Startup time</source> <translation>Waktu nyala</translation> </message> <message> <location line="+29"/> <source>Network</source> <translation>Jaringan</translation> </message> <message> <location line="+7"/> <source>Number of connections</source> <translation>Jumlah hubungan</translation> </message> <message> <location line="+23"/> <source>On testnet</source> <translation type="unfinished"/> </message> <message> <location line="+23"/> <source>Block chain</source> <translation>Rantai blok</translation> </message> <message> <location line="+7"/> <source>Current number of blocks</source> <translation>Jumlah blok terkini</translation> </message> <message> <location line="+23"/> <source>Estimated total blocks</source> <translation>Perkiraan blok total</translation> </message> <message> <location line="+23"/> <source>Last block time</source> <translation>Waktu blok terakhir</translation> </message> <message> <location line="+52"/> <source>&amp;Open</source> <translation>&amp;Buka</translation> </message> <message> <location line="+16"/> <source>Command-line options</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>Show the phreak-Qt help message to get a list with possible phreak command-line options.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>&amp;Show</source> <translation type="unfinished"/> </message> <message> <location line="+24"/> <source>&amp;Console</source> <translation>&amp;Konsol</translation> </message> <message> <location line="-260"/> <source>Build date</source> <translation>Tanggal pembuatan</translation> </message> <message> <location line="-104"/> <source>phreak - Debug window</source> <translation type="unfinished"/> </message> <message> <location line="+25"/> <source>phreak Core</source> <translation type="unfinished"/> </message> <message> <location line="+279"/> <source>Debug log file</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>Open the phreak debug log file from the current data directory. This can take a few seconds for large log files.</source> <translation type="unfinished"/> </message> <message> <location line="+102"/> <source>Clear console</source> <translation>Bersihkan konsol</translation> </message> <message> <location filename="../rpcconsole.cpp" line="-33"/> <source>Welcome to the phreak RPC console.</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Use up and down arrows to navigate history, and &lt;b&gt;Ctrl-L&lt;/b&gt; to clear screen.</source> <translation>Gunakan panah keatas dan kebawah untuk menampilkan sejarah, dan &lt;b&gt;Ctrl-L&lt;/b&gt; untuk bersihkan layar.</translation> </message> <message> <location line="+1"/> <source>Type &lt;b&gt;help&lt;/b&gt; for an overview of available commands.</source> <translation>Ketik &lt;b&gt;help&lt;/b&gt; untuk menampilkan perintah tersedia.</translation> </message> </context> <context> <name>SendCoinsDialog</name> <message> <location filename="../forms/sendcoinsdialog.ui" line="+14"/> <location filename="../sendcoinsdialog.cpp" line="+182"/> <location line="+5"/> <location line="+5"/> <location line="+5"/> <location line="+6"/> <location line="+5"/> <location line="+5"/> <source>Send Coins</source> <translation>Kirim Koin</translation> </message> <message> <location line="+76"/> <source>Coin Control Features</source> <translation type="unfinished"/> </message> <message> <location line="+20"/> <source>Inputs...</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>automatically selected</source> <translation type="unfinished"/> </message> <message> <location line="+19"/> <source>Insufficient funds!</source> <translation type="unfinished"/> </message> <message> <location line="+77"/> <source>Quantity:</source> <translation type="unfinished"/> </message> <message> <location line="+22"/> <location line="+35"/> <source>0</source> <translation type="unfinished"/> </message> <message> <location line="-19"/> <source>Bytes:</source> <translation type="unfinished"/> </message> <message> <location line="+51"/> <source>Amount:</source> <translation>Jumlah:</translation> </message> <message> <location line="+22"/> <location line="+86"/> <location line="+86"/> <location line="+32"/> <source>0.00 PHR</source> <translation type="unfinished"/> </message> <message> <location line="-191"/> <source>Priority:</source> <translation type="unfinished"/> </message> <message> <location line="+19"/> <source>medium</source> <translation type="unfinished"/> </message> <message> <location line="+32"/> <source>Fee:</source> <translation type="unfinished"/> </message> <message> <location line="+35"/> <source>Low Output:</source> <translation type="unfinished"/> </message> <message> <location line="+19"/> <source>no</source> <translation type="unfinished"/> </message> <message> <location line="+32"/> <source>After Fee:</source> <translation type="unfinished"/> </message> <message> <location line="+35"/> <source>Change</source> <translation type="unfinished"/> </message> <message> <location line="+50"/> <source>custom change address</source> <translation type="unfinished"/> </message> <message> <location line="+106"/> <source>Send to multiple recipients at once</source> <translation>Kirim ke beberapa penerima sekaligus</translation> </message> <message> <location line="+3"/> <source>Add &amp;Recipient</source> <translation type="unfinished"/> </message> <message> <location line="+20"/> <source>Remove all transaction fields</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Clear &amp;All</source> <translation>Hapus %Semua</translation> </message> <message> <location line="+28"/> <source>Balance:</source> <translation>Saldo:</translation> </message> <message> <location line="+16"/> <source>123.456 PHR</source> <translation type="unfinished"/> </message> <message> <location line="+31"/> <source>Confirm the send action</source> <translation>Konfirmasi aksi pengiriman</translation> </message> <message> <location line="+3"/> <source>S&amp;end</source> <translation type="unfinished"/> </message> <message> <location filename="../sendcoinsdialog.cpp" line="-173"/> <source>Enter a phreak address (e.g. PHR1xGeKnTkaAotEVgs2rnUfVsFv8LVSM)</source> <translation type="unfinished"/> </message> <message> <location line="+15"/> <source>Copy quantity</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Copy amount</source> <translation>Salin jumlah</translation> </message> <message> <location line="+1"/> <source>Copy fee</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Copy after fee</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Copy bytes</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Copy priority</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Copy low output</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Copy change</source> <translation type="unfinished"/> </message> <message> <location line="+86"/> <source>&lt;b&gt;%1&lt;/b&gt; to %2 (%3)</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Confirm send coins</source> <translation>Konfirmasi pengiriman koin</translation> </message> <message> <location line="+1"/> <source>Are you sure you want to send %1?</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source> and </source> <translation type="unfinished"/> </message> <message> <location line="+29"/> <source>The recipient address is not valid, please recheck.</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>The amount to pay must be larger than 0.</source> <translation>Jumlah yang dibayar harus lebih besar dari 0.</translation> </message> <message> <location line="+5"/> <source>The amount exceeds your balance.</source> <translation>Jumlah melebihi saldo Anda.</translation> </message> <message> <location line="+5"/> <source>The total exceeds your balance when the %1 transaction fee is included.</source> <translation>Kelebihan total saldo Anda ketika biaya transaksi %1 ditambahkan.</translation> </message> <message> <location line="+6"/> <source>Duplicate address found, can only send to each address once per send operation.</source> <translation>Ditemukan alamat ganda, hanya dapat mengirim ke tiap alamat sekali per operasi pengiriman.</translation> </message> <message> <location line="+5"/> <source>Error: Transaction creation failed.</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Error: The transaction was rejected. This might happen if some of the coins in your wallet were already spent, such as if you used a copy of wallet.dat and coins were spent in the copy but not marked as spent here.</source> <translation type="unfinished"/> </message> <message> <location line="+251"/> <source>WARNING: Invalid phreak address</source> <translation type="unfinished"/> </message> <message> <location line="+13"/> <source>(no label)</source> <translation>(tidak ada label)</translation> </message> <message> <location line="+4"/> <source>WARNING: unknown change address</source> <translation type="unfinished"/> </message> </context> <context> <name>SendCoinsEntry</name> <message> <location filename="../forms/sendcoinsentry.ui" line="+14"/> <source>Form</source> <translation type="unfinished"/> </message> <message> <location line="+15"/> <source>A&amp;mount:</source> <translation>J&amp;umlah:</translation> </message> <message> <location line="+13"/> <source>Pay &amp;To:</source> <translation>Kirim &amp;Ke:</translation> </message> <message> <location line="+24"/> <location filename="../sendcoinsentry.cpp" line="+25"/> <source>Enter a label for this address to add it to your address book</source> <translation>Masukkan label bagi alamat ini untuk menambahkannya ke buku alamat Anda</translation> </message> <message> <location line="+9"/> <source>&amp;Label:</source> <translation>&amp;Label:</translation> </message> <message> <location line="+18"/> <source>The address to send the payment to (e.g. PHR1xGeKnTkaAotEVgs2rnUfVsFv8LVSM)</source> <translation type="unfinished"/> </message> <message> <location line="+10"/> <source>Choose address from address book</source> <translation type="unfinished"/> </message> <message> <location line="+10"/> <source>Alt+A</source> <translation>Alt+J</translation> </message> <message> <location line="+7"/> <source>Paste address from clipboard</source> <translation>Tempel alamat dari salinan</translation> </message> <message> <location line="+10"/> <source>Alt+P</source> <translation>Alt+B</translation> </message> <message> <location line="+7"/> <source>Remove this recipient</source> <translation type="unfinished"/> </message> <message> <location filename="../sendcoinsentry.cpp" line="+1"/> <source>Enter a phreak address (e.g. PHR1xGeKnTkaAotEVgs2rnUfVsFv8LVSM)</source> <translation type="unfinished"/> </message> </context> <context> <name>SignVerifyMessageDialog</name> <message> <location filename="../forms/signverifymessagedialog.ui" line="+14"/> <source>Signatures - Sign / Verify a Message</source> <translation type="unfinished"/> </message> <message> <location line="+13"/> <location line="+124"/> <source>&amp;Sign Message</source> <translation type="unfinished"/> </message> <message> <location line="-118"/> <source>You can sign messages with your addresses to prove you own them. Be careful not to sign anything vague, as phishing attacks may try to trick you into signing your identity over to them. Only sign fully-detailed statements you agree to.</source> <translation type="unfinished"/> </message> <message> <location line="+18"/> <source>The address to sign the message with (e.g. PHR1xGeKnTkaAotEVgs2rnUfVsFv8LVSM)</source> <translation type="unfinished"/> </message> <message> <location line="+10"/> <location line="+203"/> <source>Choose an address from the address book</source> <translation type="unfinished"/> </message> <message> <location line="-193"/> <location line="+203"/> <source>Alt+A</source> <translation>Alt+J</translation> </message> <message> <location line="-193"/> <source>Paste address from clipboard</source> <translation>Tempel alamat dari salinan</translation> </message> <message> <location line="+10"/> <source>Alt+P</source> <translation>Alt+B</translation> </message> <message> <location line="+12"/> <source>Enter the message you want to sign here</source> <translation type="unfinished"/> </message> <message> <location line="+24"/> <source>Copy the current signature to the system clipboard</source> <translation type="unfinished"/> </message> <message> <location line="+21"/> <source>Sign the message to prove you own this phreak address</source> <translation type="unfinished"/> </message> <message> <location line="+17"/> <source>Reset all sign message fields</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <location line="+146"/> <source>Clear &amp;All</source> <translation>Hapus %Semua</translation> </message> <message> <location line="-87"/> <location line="+70"/> <source>&amp;Verify Message</source> <translation type="unfinished"/> </message> <message> <location line="-64"/> <source>Enter the signing address, message (ensure you copy line breaks, spaces, tabs, etc. exactly) and signature below to verify the message. Be careful not to read more into the signature than what is in the signed message itself, to avoid being tricked by a man-in-the-middle attack.</source> <translation type="unfinished"/> </message> <message> <location line="+21"/> <source>The address the message was signed with (e.g. PHR1xGeKnTkaAotEVgs2rnUfVsFv8LVSM)</source> <translation type="unfinished"/> </message> <message> <location line="+40"/> <source>Verify the message to ensure it was signed with the specified phreak address</source> <translation type="unfinished"/> </message> <message> <location line="+17"/> <source>Reset all verify message fields</source> <translation type="unfinished"/> </message> <message> <location filename="../signverifymessagedialog.cpp" line="+27"/> <location line="+3"/> <source>Enter a phreak address (e.g. PHR1xGeKnTkaAotEVgs2rnUfVsFv8LVSM)</source> <translation type="unfinished"/> </message> <message> <location line="-2"/> <source>Click &quot;Sign Message&quot; to generate signature</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Enter phreak signature</source> <translation type="unfinished"/> </message> <message> <location line="+82"/> <location line="+81"/> <source>The entered address is invalid.</source> <translation>Alamat yang dimasukkan tidak sesuai.</translation> </message> <message> <location line="-81"/> <location line="+8"/> <location line="+73"/> <location line="+8"/> <source>Please check the address and try again.</source> <translation>Silahkan periksa alamat dan coba lagi.</translation> </message> <message> <location line="-81"/> <location line="+81"/> <source>The entered address does not refer to a key.</source> <translation type="unfinished"/> </message> <message> <location line="-73"/> <source>Wallet unlock was cancelled.</source> <translation type="unfinished"/> </message> <message> <location line="+8"/> <source>Private key for the entered address is not available.</source> <translation type="unfinished"/> </message> <message> <location line="+12"/> <source>Message signing failed.</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Message signed.</source> <translation type="unfinished"/> </message> <message> <location line="+59"/> <source>The signature could not be decoded.</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <location line="+13"/> <source>Please check the signature and try again.</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>The signature did not match the message digest.</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>Message verification failed.</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Message verified.</source> <translation type="unfinished"/> </message> </context> <context> <name>TransactionDesc</name> <message> <location filename="../transactiondesc.cpp" line="+19"/> <source>Open until %1</source> <translation>Buka hingga %1</translation> </message> <message numerus="yes"> <location line="-2"/> <source>Open for %n block(s)</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message> <location line="+8"/> <source>conflicted</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>%1/offline</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>%1/unconfirmed</source> <translation>%1/tidak terkonfirmasi</translation> </message> <message> <location line="+2"/> <source>%1 confirmations</source> <translation>%1 konfirmasi</translation> </message> <message> <location line="+18"/> <source>Status</source> <translation>Status</translation> </message> <message numerus="yes"> <location line="+7"/> <source>, broadcast through %n node(s)</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message> <location line="+4"/> <source>Date</source> <translation>Tanggal</translation> </message> <message> <location line="+7"/> <source>Source</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>Generated</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <location line="+17"/> <source>From</source> <translation>Dari</translation> </message> <message> <location line="+1"/> <location line="+22"/> <location line="+58"/> <source>To</source> <translation>Untuk</translation> </message> <message> <location line="-77"/> <location line="+2"/> <source>own address</source> <translation type="unfinished"/> </message> <message> <location line="-2"/> <source>label</source> <translation type="unfinished"/> </message> <message> <location line="+37"/> <location line="+12"/> <location line="+45"/> <location line="+17"/> <location line="+30"/> <source>Credit</source> <translation type="unfinished"/> </message> <message numerus="yes"> <location line="-102"/> <source>matures in %n more block(s)</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message> <location line="+2"/> <source>not accepted</source> <translation type="unfinished"/> </message> <message> <location line="+44"/> <location line="+8"/> <location line="+15"/> <location line="+30"/> <source>Debit</source> <translation type="unfinished"/> </message> <message> <location line="-39"/> <source>Transaction fee</source> <translation type="unfinished"/> </message> <message> <location line="+16"/> <source>Net amount</source> <translation type="unfinished"/> </message> <message> <location line="+6"/> <source>Message</source> <translation>Pesan:</translation> </message> <message> <location line="+2"/> <source>Comment</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Transaction ID</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Generated coins must mature 110 blocks before they can be spent. When you generated this block, it was broadcast to the network to be added to the block chain. If it fails to get into the chain, its state will change to &quot;not accepted&quot; and it won&apos;t be spendable. This may occasionally happen if another node generates a block within a few seconds of yours.</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>Debug information</source> <translation type="unfinished"/> </message> <message> <location line="+8"/> <source>Transaction</source> <translation>Transaksi</translation> </message> <message> <location line="+5"/> <source>Inputs</source> <translation type="unfinished"/> </message> <message> <location line="+23"/> <source>Amount</source> <translation>Jumlah</translation> </message> <message> <location line="+1"/> <source>true</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>false</source> <translation type="unfinished"/> </message> <message> <location line="-211"/> <source>, has not been successfully broadcast yet</source> <translation>, belum berhasil disiarkan</translation> </message> <message> <location line="+35"/> <source>unknown</source> <translation>tidak diketahui</translation> </message> </context> <context> <name>TransactionDescDialog</name> <message> <location filename="../forms/transactiondescdialog.ui" line="+14"/> <source>Transaction details</source> <translation>Rincian transaksi</translation> </message> <message> <location line="+6"/> <source>This pane shows a detailed description of the transaction</source> <translation>Jendela ini menampilkan deskripsi rinci dari transaksi tersebut</translation> </message> </context> <context> <name>TransactionTableModel</name> <message> <location filename="../transactiontablemodel.cpp" line="+226"/> <source>Date</source> <translation>Tanggal</translation> </message> <message> <location line="+0"/> <source>Type</source> <translation>Jenis</translation> </message> <message> <location line="+0"/> <source>Address</source> <translation>Alamat</translation> </message> <message> <location line="+0"/> <source>Amount</source> <translation>Jumlah</translation> </message> <message> <location line="+60"/> <source>Open until %1</source> <translation>Buka hingga %1</translation> </message> <message> <location line="+12"/> <source>Confirmed (%1 confirmations)</source> <translation>Terkonfirmasi (%1 konfirmasi)</translation> </message> <message numerus="yes"> <location line="-15"/> <source>Open for %n more block(s)</source> <translation type="unfinished"><numerusform></numerusform></translation> </message> <message> <location line="+6"/> <source>Offline</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Unconfirmed</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Confirming (%1 of %2 recommended confirmations)</source> <translation type="unfinished"/> </message> <message> <location line="+6"/> <source>Conflicted</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Immature (%1 confirmations, will be available after %2)</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>This block was not received by any other nodes and will probably not be accepted!</source> <translation>Blok ini tidak diterima oleh node lainnya dan kemungkinan tidak akan diterima!</translation> </message> <message> <location line="+3"/> <source>Generated but not accepted</source> <translation>Terbuat tetapi tidak diterima</translation> </message> <message> <location line="+42"/> <source>Received with</source> <translation>Diterima dengan</translation> </message> <message> <location line="+2"/> <source>Received from</source> <translation>Diterima dari</translation> </message> <message> <location line="+3"/> <source>Sent to</source> <translation>Terkirim ke</translation> </message> <message> <location line="+2"/> <source>Payment to yourself</source> <translation>Pembayaran ke Anda sendiri</translation> </message> <message> <location line="+2"/> <source>Mined</source> <translation>Tertambang</translation> </message> <message> <location line="+38"/> <source>(n/a)</source> <translation>(t/s)</translation> </message> <message> <location line="+190"/> <source>Transaction status. Hover over this field to show number of confirmations.</source> <translation>Status transaksi. Arahkan ke bagian ini untuk menampilkan jumlah konfrimasi.</translation> </message> <message> <location line="+2"/> <source>Date and time that the transaction was received.</source> <translation>Tanggal dan waktu transaksi tersebut diterima.</translation> </message> <message> <location line="+2"/> <source>Type of transaction.</source> <translation>Jenis transaksi.</translation> </message> <message> <location line="+2"/> <source>Destination address of transaction.</source> <translation>Alamat tujuan dari transaksi.</translation> </message> <message> <location line="+2"/> <source>Amount removed from or added to balance.</source> <translation>Jumlah terbuang dari atau ditambahkan ke saldo.</translation> </message> </context> <context> <name>TransactionView</name> <message> <location filename="../transactionview.cpp" line="+55"/> <location line="+16"/> <source>All</source> <translation>Semua</translation> </message> <message> <location line="-15"/> <source>Today</source> <translation>Hari ini</translation> </message> <message> <location line="+1"/> <source>This week</source> <translation>Minggu ini</translation> </message> <message> <location line="+1"/> <source>This month</source> <translation>Bulan ini</translation> </message> <message> <location line="+1"/> <source>Last month</source> <translation>Bulan kemarin</translation> </message> <message> <location line="+1"/> <source>This year</source> <translation>Tahun ini</translation> </message> <message> <location line="+1"/> <source>Range...</source> <translation>Jarak...</translation> </message> <message> <location line="+11"/> <source>Received with</source> <translation>DIterima dengan</translation> </message> <message> <location line="+2"/> <source>Sent to</source> <translation>Terkirim ke</translation> </message> <message> <location line="+2"/> <source>To yourself</source> <translation>Ke Anda sendiri</translation> </message> <message> <location line="+1"/> <source>Mined</source> <translation>Ditambang</translation> </message> <message> <location line="+1"/> <source>Other</source> <translation>Lainnya</translation> </message> <message> <location line="+7"/> <source>Enter address or label to search</source> <translation>Masukkan alamat atau label untuk mencari</translation> </message> <message> <location line="+7"/> <source>Min amount</source> <translation>Jumlah min</translation> </message> <message> <location line="+34"/> <source>Copy address</source> <translation>Salin alamat</translation> </message> <message> <location line="+1"/> <source>Copy label</source> <translation>Salin label</translation> </message> <message> <location line="+1"/> <source>Copy amount</source> <translation>Salin jumlah</translation> </message> <message> <location line="+1"/> <source>Copy transaction ID</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Edit label</source> <translation>Ubah label</translation> </message> <message> <location line="+1"/> <source>Show transaction details</source> <translation>Tampilkan rincian transaksi</translation> </message> <message> <location line="+144"/> <source>Export Transaction Data</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Comma separated file (*.csv)</source> <translation>Berkas CSV (*.csv)</translation> </message> <message> <location line="+8"/> <source>Confirmed</source> <translation>Terkonfirmasi</translation> </message> <message> <location line="+1"/> <source>Date</source> <translation>Tanggal</translation> </message> <message> <location line="+1"/> <source>Type</source> <translation>Jenis</translation> </message> <message> <location line="+1"/> <source>Label</source> <translation>Label</translation> </message> <message> <location line="+1"/> <source>Address</source> <translation>Alamat</translation> </message> <message> <location line="+1"/> <source>Amount</source> <translation>Jumlah</translation> </message> <message> <location line="+1"/> <source>ID</source> <translation>ID</translation> </message> <message> <location line="+4"/> <source>Error exporting</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>Could not write to file %1.</source> <translation type="unfinished"/> </message> <message> <location line="+100"/> <source>Range:</source> <translation>Jarak:</translation> </message> <message> <location line="+8"/> <source>to</source> <translation>ke</translation> </message> </context> <context> <name>WalletModel</name> <message> <location filename="../walletmodel.cpp" line="+206"/> <source>Sending...</source> <translation type="unfinished"/> </message> </context> <context> <name>bitcoin-core</name> <message> <location filename="../bitcoinstrings.cpp" line="+33"/> <source>phreak version</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Usage:</source> <translation>Penggunaan:</translation> </message> <message> <location line="+1"/> <source>Send command to -server or phreakd</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>List commands</source> <translation>Daftar perintah</translation> </message> <message> <location line="+1"/> <source>Get help for a command</source> <translation>Dapatkan bantuan untuk perintah</translation> </message> <message> <location line="+2"/> <source>Options:</source> <translation>Pilihan:</translation> </message> <message> <location line="+2"/> <source>Specify configuration file (default: phreak.conf)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Specify pid file (default: phreakd.pid)</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Specify wallet file (within data directory)</source> <translation type="unfinished"/> </message> <message> <location line="-1"/> <source>Specify data directory</source> <translation>Tentukan direktori data</translation> </message> <message> <location line="+2"/> <source>Set database cache size in megabytes (default: 25)</source> <translation>Atur ukuran tembolok dalam megabyte (standar: 25)</translation> </message> <message> <location line="+1"/> <source>Set database disk log size in megabytes (default: 100)</source> <translation type="unfinished"/> </message> <message> <location line="+6"/> <source>Listen for connections on &lt;port&gt; (default: 15714 or testnet: 25714)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Maintain at most &lt;n&gt; connections to peers (default: 125)</source> <translation>Mengatur hubungan paling banyak &lt;n&gt; ke peer (standar: 125)</translation> </message> <message> <location line="+3"/> <source>Connect to a node to retrieve peer addresses, and disconnect</source> <translation>Hubungkan ke node untuk menerima alamat peer, dan putuskan</translation> </message> <message> <location line="+1"/> <source>Specify your own public address</source> <translation>Tentukan alamat publik Anda sendiri</translation> </message> <message> <location line="+5"/> <source>Bind to given address. Use [host]:port notation for IPv6</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Stake your coins to support network and gain reward (default: 1)</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Threshold for disconnecting misbehaving peers (default: 100)</source> <translation>Batas untuk memutuskan peer buruk (standar: 100)</translation> </message> <message> <location line="+1"/> <source>Number of seconds to keep misbehaving peers from reconnecting (default: 86400)</source> <translation>Jumlah kedua untuk menjaga peer buruk dari hubung-ulang (standar: 86400)</translation> </message> <message> <location line="-44"/> <source>An error occurred while setting up the RPC port %u for listening on IPv4: %s</source> <translation type="unfinished"/> </message> <message> <location line="+51"/> <source>Detach block and address databases. Increases shutdown time (default: 0)</source> <translation type="unfinished"/> </message> <message> <location line="+109"/> <source>Error: The transaction was rejected. This might happen if some of the coins in your wallet were already spent, such as if you used a copy of wallet.dat and coins were spent in the copy but not marked as spent here.</source> <translation type="unfinished"/> </message> <message> <location line="-5"/> <source>Error: This transaction requires a transaction fee of at least %s because of its amount, complexity, or use of recently received funds </source> <translation type="unfinished"/> </message> <message> <location line="-87"/> <source>Listen for JSON-RPC connections on &lt;port&gt; (default: 15715 or testnet: 25715)</source> <translation type="unfinished"/> </message> <message> <location line="-11"/> <source>Accept command line and JSON-RPC commands</source> <translation>Menerima perintah baris perintah dan JSON-RPC</translation> </message> <message> <location line="+101"/> <source>Error: Transaction creation failed </source> <translation type="unfinished"/> </message> <message> <location line="-5"/> <source>Error: Wallet locked, unable to create transaction </source> <translation type="unfinished"/> </message> <message> <location line="-8"/> <source>Importing blockchain data file.</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Importing bootstrap blockchain data file.</source> <translation type="unfinished"/> </message> <message> <location line="-88"/> <source>Run in the background as a daemon and accept commands</source> <translation>Berjalan dibelakang sebagai daemin dan menerima perintah</translation> </message> <message> <location line="+1"/> <source>Use the test network</source> <translation>Gunakan jaringan uji</translation> </message> <message> <location line="-24"/> <source>Accept connections from outside (default: 1 if no -proxy or -connect)</source> <translation type="unfinished"/> </message> <message> <location line="-38"/> <source>An error occurred while setting up the RPC port %u for listening on IPv6, falling back to IPv4: %s</source> <translation type="unfinished"/> </message> <message> <location line="+117"/> <source>Error initializing database environment %s! To recover, BACKUP THAT DIRECTORY, then remove everything from it except for wallet.dat.</source> <translation type="unfinished"/> </message> <message> <location line="-20"/> <source>Set maximum size of high-priority/low-fee transactions in bytes (default: 27000)</source> <translation type="unfinished"/> </message> <message> <location line="+11"/> <source>Warning: -paytxfee is set very high! This is the transaction fee you will pay if you send a transaction.</source> <translation type="unfinished"/> </message> <message> <location line="+61"/> <source>Warning: Please check that your computer&apos;s date and time are correct! If your clock is wrong phreak will not work properly.</source> <translation type="unfinished"/> </message> <message> <location line="-31"/> <source>Warning: error reading wallet.dat! All keys read correctly, but transaction data or address book entries might be missing or incorrect.</source> <translation type="unfinished"/> </message> <message> <location line="-18"/> <source>Warning: wallet.dat corrupt, data salvaged! Original wallet.dat saved as wallet.{timestamp}.bak in %s; if your balance or transactions are incorrect you should restore from a backup.</source> <translation type="unfinished"/> </message> <message> <location line="-30"/> <source>Attempt to recover private keys from a corrupt wallet.dat</source> <translation type="unfinished"/> </message> <message> <location line="+4"/> <source>Block creation options:</source> <translation type="unfinished"/> </message> <message> <location line="-62"/> <source>Connect only to the specified node(s)</source> <translation type="unfinished"/> </message> <message> <location line="+4"/> <source>Discover own IP address (default: 1 when listening and no -externalip)</source> <translation type="unfinished"/> </message> <message> <location line="+94"/> <source>Failed to listen on any port. Use -listen=0 if you want this.</source> <translation type="unfinished"/> </message> <message> <location line="-90"/> <source>Find peers using DNS lookup (default: 1)</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Sync checkpoints policy (default: strict)</source> <translation type="unfinished"/> </message> <message> <location line="+83"/> <source>Invalid -tor address: &apos;%s&apos;</source> <translation type="unfinished"/> </message> <message> <location line="+4"/> <source>Invalid amount for -reservebalance=&lt;amount&gt;</source> <translation type="unfinished"/> </message> <message> <location line="-82"/> <source>Maximum per-connection receive buffer, &lt;n&gt;*1000 bytes (default: 5000)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Maximum per-connection send buffer, &lt;n&gt;*1000 bytes (default: 1000)</source> <translation type="unfinished"/> </message> <message> <location line="-16"/> <source>Only connect to nodes in network &lt;net&gt; (IPv4, IPv6 or Tor)</source> <translation type="unfinished"/> </message> <message> <location line="+28"/> <source>Output extra debugging information. Implies all other -debug* options</source> <translation type="unfinished"/> </message> <message><|fim▁hole|> <message> <location line="+1"/> <source>Prepend debug output with timestamp</source> <translation type="unfinished"/> </message> <message> <location line="+35"/> <source>SSL options: (see the Bitcoin Wiki for SSL setup instructions)</source> <translation type="unfinished"/> </message> <message> <location line="-74"/> <source>Select the version of socks proxy to use (4-5, default: 5)</source> <translation type="unfinished"/> </message> <message> <location line="+41"/> <source>Send trace/debug info to console instead of debug.log file</source> <translation>Kirim info lacak/debug ke konsol sebaliknya dari berkas debug.log</translation> </message> <message> <location line="+1"/> <source>Send trace/debug info to debugger</source> <translation type="unfinished"/> </message> <message> <location line="+28"/> <source>Set maximum block size in bytes (default: 250000)</source> <translation type="unfinished"/> </message> <message> <location line="-1"/> <source>Set minimum block size in bytes (default: 0)</source> <translation type="unfinished"/> </message> <message> <location line="-29"/> <source>Shrink debug.log file on client startup (default: 1 when no -debug)</source> <translation type="unfinished"/> </message> <message> <location line="-42"/> <source>Specify connection timeout in milliseconds (default: 5000)</source> <translation type="unfinished"/> </message> <message> <location line="+109"/> <source>Unable to sign checkpoint, wrong checkpointkey? </source> <translation type="unfinished"/> </message> <message> <location line="-80"/> <source>Use UPnP to map the listening port (default: 0)</source> <translation type="unfinished"/> </message> <message> <location line="-1"/> <source>Use UPnP to map the listening port (default: 1 when listening)</source> <translation type="unfinished"/> </message> <message> <location line="-25"/> <source>Use proxy to reach tor hidden services (default: same as -proxy)</source> <translation type="unfinished"/> </message> <message> <location line="+42"/> <source>Username for JSON-RPC connections</source> <translation>Nama pengguna untuk hubungan JSON-RPC</translation> </message> <message> <location line="+47"/> <source>Verifying database integrity...</source> <translation type="unfinished"/> </message> <message> <location line="+57"/> <source>WARNING: syncronized checkpoint violation detected, but skipped!</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Warning: Disk space is low!</source> <translation type="unfinished"/> </message> <message> <location line="-2"/> <source>Warning: This version is obsolete, upgrade required!</source> <translation type="unfinished"/> </message> <message> <location line="-48"/> <source>wallet.dat corrupt, salvage failed</source> <translation type="unfinished"/> </message> <message> <location line="-54"/> <source>Password for JSON-RPC connections</source> <translation>Kata sandi untuk hubungan JSON-RPC</translation> </message> <message> <location line="-84"/> <source>%s, you must set a rpcpassword in the configuration file: %s It is recommended you use the following random password: rpcuser=phreakrpc rpcpassword=%s (you do not need to remember this password) The username and password MUST NOT be the same. If the file does not exist, create it with owner-readable-only file permissions. It is also recommended to set alertnotify so you are notified of problems; for example: alertnotify=echo %%s | mail -s &quot;phreak Alert&quot; [email protected] </source> <translation type="unfinished"/> </message> <message> <location line="+51"/> <source>Find peers using internet relay chat (default: 0)</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Sync time with other nodes. Disable if time on your system is precise e.g. syncing with NTP (default: 1)</source> <translation type="unfinished"/> </message> <message> <location line="+15"/> <source>When creating transactions, ignore inputs with value less than this (default: 0.01)</source> <translation type="unfinished"/> </message> <message> <location line="+16"/> <source>Allow JSON-RPC connections from specified IP address</source> <translation>Izinkan hubungan JSON-RPC dari alamat IP yang ditentukan</translation> </message> <message> <location line="+1"/> <source>Send commands to node running on &lt;ip&gt; (default: 127.0.0.1)</source> <translation>Kirim perintah ke node berjalan pada &lt;ip&gt; (standar: 127.0.0.1)</translation> </message> <message> <location line="+1"/> <source>Execute command when the best block changes (%s in cmd is replaced by block hash)</source> <translation>Menjalankan perintah ketika perubahan blok terbaik (%s dalam cmd digantikan oleh hash blok)</translation> </message> <message> <location line="+3"/> <source>Execute command when a wallet transaction changes (%s in cmd is replaced by TxID)</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Require a confirmations for change (default: 0)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Enforce transaction scripts to use canonical PUSH operators (default: 1)</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Execute command when a relevant alert is received (%s in cmd is replaced by message)</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Upgrade wallet to latest format</source> <translation>Perbarui dompet ke format terbaru</translation> </message> <message> <location line="+1"/> <source>Set key pool size to &lt;n&gt; (default: 100)</source> <translation>Kirim ukuran kolam kunci ke &lt;n&gt; (standar: 100)</translation> </message> <message> <location line="+1"/> <source>Rescan the block chain for missing wallet transactions</source> <translation>Pindai ulang rantai-blok untuk transaksi dompet yang hilang</translation> </message> <message> <location line="+2"/> <source>How many blocks to check at startup (default: 2500, 0 = all)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>How thorough the block verification is (0-6, default: 1)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Imports blocks from external blk000?.dat file</source> <translation type="unfinished"/> </message> <message> <location line="+8"/> <source>Use OpenSSL (https) for JSON-RPC connections</source> <translation>Gunakan OpenSSL (https) untuk hubungan JSON-RPC</translation> </message> <message> <location line="+1"/> <source>Server certificate file (default: server.cert)</source> <translation>Berkas sertifikat server (standar: server.cert)</translation> </message> <message> <location line="+1"/> <source>Server private key (default: server.pem)</source> <translation>Kunci pribadi server (standar: server.pem)</translation> </message> <message> <location line="+1"/> <source>Acceptable ciphers (default: TLSv1+HIGH:!SSLv2:!aNULL:!eNULL:!AH:!3DES:@STRENGTH)</source> <translation type="unfinished"/> </message> <message> <location line="+53"/> <source>Error: Wallet unlocked for staking only, unable to create transaction.</source> <translation type="unfinished"/> </message> <message> <location line="+18"/> <source>WARNING: Invalid checkpoint found! Displayed transactions may not be correct! You may need to upgrade, or notify developers.</source> <translation type="unfinished"/> </message> <message> <location line="-158"/> <source>This help message</source> <translation>Pesan bantuan ini</translation> </message> <message> <location line="+95"/> <source>Wallet %s resides outside data directory %s.</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Cannot obtain a lock on data directory %s. phreak is probably already running.</source> <translation type="unfinished"/> </message> <message> <location line="-98"/> <source>phreak</source> <translation type="unfinished"/> </message> <message> <location line="+140"/> <source>Unable to bind to %s on this computer (bind returned error %d, %s)</source> <translation>Tidak dapat mengikat ke %s dengan komputer ini (ikatan gagal %d, %s)</translation> </message> <message> <location line="-130"/> <source>Connect through socks proxy</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Allow DNS lookups for -addnode, -seednode and -connect</source> <translation>Izinkan peninjauan DNS untuk -addnote, -seednode dan -connect</translation> </message> <message> <location line="+122"/> <source>Loading addresses...</source> <translation>Memuat alamat...</translation> </message> <message> <location line="-15"/> <source>Error loading blkindex.dat</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Error loading wallet.dat: Wallet corrupted</source> <translation>Gagal memuat wallet.dat: Dompet rusak</translation> </message> <message> <location line="+4"/> <source>Error loading wallet.dat: Wallet requires newer version of phreak</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Wallet needed to be rewritten: restart phreak to complete</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Error loading wallet.dat</source> <translation>Gagal memuat wallet.dat</translation> </message> <message> <location line="-16"/> <source>Invalid -proxy address: &apos;%s&apos;</source> <translation>Alamat -proxy salah: &apos;%s&apos;</translation> </message> <message> <location line="-1"/> <source>Unknown network specified in -onlynet: &apos;%s&apos;</source> <translation>Jaringan tidak diketahui yang ditentukan dalam -onlynet: &apos;%s&apos;</translation> </message> <message> <location line="-1"/> <source>Unknown -socks proxy version requested: %i</source> <translation>Diminta versi proxy -socks tidak diketahui: %i</translation> </message> <message> <location line="+4"/> <source>Cannot resolve -bind address: &apos;%s&apos;</source> <translation>Tidak dapat menyelesaikan alamat -bind: &apos;%s&apos;</translation> </message> <message> <location line="+2"/> <source>Cannot resolve -externalip address: &apos;%s&apos;</source> <translation>Tidak dapat menyelesaikan alamat -externalip: &apos;%s&apos;</translation> </message> <message> <location line="-24"/> <source>Invalid amount for -paytxfee=&lt;amount&gt;: &apos;%s&apos;</source> <translation>Jumlah salah untuk -paytxfee=&lt;amount&gt;: &apos;%s&apos;</translation> </message> <message> <location line="+44"/> <source>Error: could not start node</source> <translation type="unfinished"/> </message> <message> <location line="+11"/> <source>Sending...</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Invalid amount</source> <translation>Jumlah salah</translation> </message> <message> <location line="+1"/> <source>Insufficient funds</source> <translation>Saldo tidak mencukupi</translation> </message> <message> <location line="-34"/> <source>Loading block index...</source> <translation>Memuat indeks blok...</translation> </message> <message> <location line="-103"/> <source>Add a node to connect to and attempt to keep the connection open</source> <translation>Tambahkan node untuk dihubungkan dan upaya untuk menjaga hubungan tetap terbuka</translation> </message> <message> <location line="+122"/> <source>Unable to bind to %s on this computer. phreak is probably already running.</source> <translation type="unfinished"/> </message> <message> <location line="-97"/> <source>Fee per KB to add to transactions you send</source> <translation type="unfinished"/> </message> <message> <location line="+55"/> <source>Invalid amount for -mininput=&lt;amount&gt;: &apos;%s&apos;</source> <translation type="unfinished"/> </message> <message> <location line="+25"/> <source>Loading wallet...</source> <translation>Memuat dompet...</translation> </message> <message> <location line="+8"/> <source>Cannot downgrade wallet</source> <translation>Tidak dapat menurunkan versi dompet</translation> </message> <message> <location line="+1"/> <source>Cannot initialize keypool</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Cannot write default address</source> <translation>Tidak dapat menyimpan alamat standar</translation> </message> <message> <location line="+1"/> <source>Rescanning...</source> <translation>Memindai ulang...</translation> </message> <message> <location line="+5"/> <source>Done loading</source> <translation>Memuat selesai</translation> </message> <message> <location line="-167"/> <source>To use the %s option</source> <translation>Gunakan pilihan %s</translation> </message> <message> <location line="+14"/> <source>Error</source> <translation>Gagal</translation> </message> <message> <location line="+6"/> <source>You must set rpcpassword=&lt;password&gt; in the configuration file: %s If the file does not exist, create it with owner-readable-only file permissions.</source> <translation>Anda harus mengatur rpcpassword=&lt;kata sandi&gt; dalam berkas konfigurasi: %s Jika berkas tidak ada, buatlah dengan permisi berkas hanya-dapat-dibaca-oleh-pemilik.</translation> </message> </context> </TS><|fim▁end|>
<location line="+1"/> <source>Output extra network debugging information</source> <translation type="unfinished"/> </message>
<|file_name|>saveable_object.py<|end_file_name|><|fim▁begin|># Copyright 2015 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Types for specifying saving and loading behavior.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function class SaveSpec(object): """Class used to describe tensor slices that need to be saved.""" def __init__(self, tensor, slice_spec, name, dtype=None, device=None): """Creates a `SaveSpec` object. Args: tensor: the tensor to save or callable that produces a tensor to save. If the value is `None`, the `SaveSpec` is ignored. slice_spec: the slice to be saved. See `Variable.SaveSliceInfo`. name: the name to save the tensor under. dtype: The data type of the Tensor. Required if `tensor` is callable. Used for error checking in the restore op. device: The device generating and consuming this tensor. Required if `tensor` is callable. Used to group objects to save by device. """ self._tensor = tensor self.slice_spec = slice_spec self.name = name if callable(self._tensor): if dtype is None or device is None: raise AssertionError( "When passing a callable `tensor` to a SaveSpec, an explicit " "dtype and device must be provided.") self.dtype = dtype self.device = device else: self.dtype = tensor.dtype if device is not None: self.device = device else: self.device = tensor.device @property def tensor(self): return self._tensor() if callable(self._tensor) else self._tensor <|fim▁hole|> def __init__(self, op, specs, name): """Creates a `SaveableObject` object. Args: op: the "producer" object that this class wraps; it produces a list of tensors to save. E.g., a "Variable" object saving its backing tensor. specs: a list of SaveSpec, each element of which describes one tensor to save under this object. All Tensors must be on the same device. name: the name to save the object under. """ self.op = op self.specs = specs self.name = name @property def optional_restore(self): """A hint to restore assertions that this object is optional.""" return False # Default to required @property def device(self): """The device for SaveSpec Tensors.""" return self.specs[0].device def restore(self, restored_tensors, restored_shapes): """Restores this object from 'restored_tensors'. Args: restored_tensors: the tensors that were loaded from a checkpoint restored_shapes: the shapes this object should conform to after restore, or None. Returns: An operation that restores the state of the object. Raises: ValueError: If the object cannot be restored using the provided parameters. """ # pylint: disable=unused-argument raise ValueError("Calling an abstract method.")<|fim▁end|>
class SaveableObject(object): """Base class for saving and restoring saveable objects."""
<|file_name|>lightbox_plus.js<|end_file_name|><|fim▁begin|>// lightbox_plus.js // == written by Takuya Otani <[email protected]> === // == Copyright (C) 2006 SimpleBoxes/SerendipityNZ Ltd. == /* Copyright (C) 2006 Takuya Otani/SimpleBoxes - http://serennz.cool.ne.jp/sb/ Copyright (C) 2006 SerendipityNZ - http://serennz.cool.ne.jp/snz/ This script is licensed under the Creative Commons Attribution 2.5 License http://creativecommons.org/licenses/by/2.5/ basically, do anything you want, just leave my name and link. */ /* Original script : Lightbox JS : Fullsize Image Overlays Copyright (C) 2005 Lokesh Dhakar - http://www.huddletogether.com For more information on this script, visit: http://huddletogether.com/projects/lightbox/ */ // ver. 20061027 - fixed a bug ( not work at xhml documents on Netscape7 ) // ver. 20061026 - fixed bugs // ver. 20061010 - implemented image set feature // ver. 20060921 - fixed a bug / added overall view // ver. 20060920 - added flag to prevent mouse wheel event // ver. 20060919 - fixed a bug // ver. 20060918 - implemented functionality of wheel zoom & drag'n drop // ver. 20060131 - fixed a bug to work correctly on Internet Explorer for Windows // ver. 20060128 - implemented functionality of echoic word // ver. 20060120 - implemented functionality of caption and close button function WindowSize() { // window size object this.w = 0; this.h = 0; return this.update(); } WindowSize.prototype.update = function() { var d = document; this.w = (window.innerWidth) ? window.innerWidth : (d.documentElement && d.documentElement.clientWidth) ? d.documentElement.clientWidth : d.body.clientWidth; this.h = (window.innerHeight) ? window.innerHeight : (d.documentElement && d.documentElement.clientHeight) ? d.documentElement.clientHeight : d.body.clientHeight; return this; }; function PageSize() { // page size object this.win = new WindowSize(); this.w = 0; this.h = 0; return this.update(); } PageSize.prototype.update = function() { var d = document; this.w = (window.innerWidth && window.scrollMaxX) ? window.innerWidth + window.scrollMaxX : (d.body.scrollWidth > d.body.offsetWidth) ? d.body.scrollWidth : d.body.offsetWidt; this.h = (window.innerHeight && window.scrollMaxY) ? window.innerHeight + window.scrollMaxY : (d.body.scrollHeight > d.body.offsetHeight) ? d.body.scrollHeight : d.body.offsetHeight; this.win.update(); if (this.w < this.win.w) this.w = this.win.w; if (this.h < this.win.h) this.h = this.win.h; return this; }; function PagePos() { // page position object this.x = 0; this.y = 0; return this.update(); } PagePos.prototype.update = function() { var d = document; this.x = (window.pageXOffset) ? window.pageXOffset : (d.documentElement && d.documentElement.scrollLeft) ? d.documentElement.scrollLeft : (d.body) ? d.body.scrollLeft : 0; this.y = (window.pageYOffset) ? window.pageYOffset : (d.documentElement && d.documentElement.scrollTop) ? d.documentElement.scrollTop : (d.body) ? d.body.scrollTop : 0; return this; }; function LightBox(option) { var self = this; self._imgs = []; self._sets = []; self._wrap = null; self._box = null; self._img = null; self._open = -1; self._page = new PageSize(); self._pos = new PagePos(); self._zoomimg = null; self._expandable = false; self._expanded = false; self._funcs = {'move':null,'up':null,'drag':null,'wheel':null,'dbl':null}; self._level = 1; self._curpos = {x:0,y:0}; self._imgpos = {x:0,y:0}; self._minpos = {x:0,y:0}; self._expand = option.expandimg; self._shrink = option.shrinkimg; self._resizable = option.resizable; self._timer = null; self._indicator = null; self._overall = null; self._openedset = null; self._prev = null; self._next = null; self._hiding = []; self._first = false; return self._init(option); } LightBox.prototype = { _init : function(option) { var self = this; var d = document; if (!d.getElementsByTagName) return; if (Browser.isMacIE) return self; var links = d.getElementsByTagName("a"); for (var i = 0; i < links.length; i++) { var anchor = links[i]; var num = self._imgs.length; var rel = String(anchor.getAttribute("rel")).toLowerCase(); if (!anchor.getAttribute("href") || !rel.match('lightbox')) continue; // initialize item self._imgs[num] = { src:anchor.getAttribute("href"), w:-1, h:-1, title:'', cls:anchor.className, set:rel }; if (anchor.getAttribute("title")) self._imgs[num].title = anchor.getAttribute("title"); else if (anchor.firstChild && anchor.firstChild.getAttribute && anchor.firstChild.getAttribute("title")) self._imgs[num].title = anchor.firstChild.getAttribute("title"); anchor.onclick = self._genOpener(num); // set closure to onclick event if (rel != 'lightbox') { if (!self._sets[rel]) self._sets[rel] = []; self._sets[rel].push(num); } } var body = d.getElementsByTagName("body")[0]; self._wrap = self._createWrapOn(body, option.loadingimg); self._box = self._createBoxOn(body, option); self._img = self._box.firstChild; self._zoomimg = d.getElementById('actionImage'); return self; }, _genOpener : function(num) { var self = this; return function() { self._show(num); return false; } }, _createWrapOn : function(obj, imagePath) { var self = this; if (!obj) return null; // create wrapper object, translucent background var wrap = document.createElement('div'); obj.appendChild(wrap); wrap.id = 'overlay'; wrap.style.display = 'none'; wrap.style.position = 'fixed'; wrap.style.top = '0px'; wrap.style.left = '0px'; wrap.style.zIndex = '50'; wrap.style.width = '100%'; wrap.style.height = '100%'; if (Browser.isWinIE) wrap.style.position = 'absolute'; Event.register(wrap, "click", function(evt) { self._close(evt); }); // create loading image, animated image var imag = new Image; imag.onload = function() { var spin = document.createElement('img'); wrap.appendChild(spin); spin.id = 'loadingImage'; spin.src = imag.src; spin.style.position = 'relative'; self._set_cursor(spin); Event.register(spin, 'click', function(evt) { self._close(evt); }); imag.onload = function() { }; }; if (imagePath != '') imag.src = imagePath; return wrap; }, _createBoxOn : function(obj, option) { var self = this; if (!obj) return null; // create lightbox object, frame rectangle var box = document.createElement('div'); obj.appendChild(box); box.id = 'lightbox'; box.style.display = 'none'; box.style.position = 'absolute'; box.style.zIndex = '60'; // create image object to display a target image var img = document.createElement('img'); box.appendChild(img); img.id = 'lightboxImage'; self._set_cursor(img); Event.register(img, 'mouseover', function() { self._show_action(); }); Event.register(img, 'mouseout', function() { self._hide_action(); }); Event.register(img, 'click', function(evt) { self._close(evt); }); // create hover navi - prev if (option.previmg) { var prevLink = document.createElement('img'); box.appendChild(prevLink); prevLink.id = 'prevLink'; prevLink.style.display = 'none'; prevLink.style.position = 'absolute'; prevLink.style.left = '9px'; prevLink.style.zIndex = '70'; prevLink.src = option.previmg; self._prev = prevLink; Event.register(prevLink, 'mouseover', function() { self._show_action(); }); Event.register(prevLink, 'click', function() { self._show_next(-1); }); } // create hover navi - next if (option.nextimg) { var nextLink = document.createElement('img'); box.appendChild(nextLink); nextLink.id = 'nextLink'; nextLink.style.display = 'none'; nextLink.style.position = 'absolute'; nextLink.style.right = '9px'; nextLink.style.zIndex = '70'; nextLink.src = option.nextimg; self._next = nextLink; Event.register(nextLink, 'mouseover', function() { self._show_action(); }); Event.register(nextLink, 'click', function() { self._show_next(+1); }); } // create zoom indicator var zoom = document.createElement('img'); box.appendChild(zoom); zoom.id = 'actionImage'; zoom.style.display = 'none'; zoom.style.position = 'absolute'; zoom.style.top = '15px'; zoom.style.left = '15px'; zoom.style.zIndex = '70'; self._set_cursor(zoom); zoom.src = self._expand; Event.register(zoom, 'mouseover', function() { self._show_action(); }); Event.register(zoom, 'click', function() { self._zoom(); }); Event.register(window, 'resize', function() { self._set_size(true); }); // create close button if (option.closeimg) { var btn = document.createElement('img'); box.appendChild(btn); btn.id = 'closeButton'; btn.style.display = 'inline'; btn.style.position = 'absolute'; btn.style.right = '9px'; btn.style.top = '10px'; btn.style.zIndex = '80'; btn.src = option.closeimg; self._set_cursor(btn); Event.register(btn, 'click', function(evt) { self._close(evt); }); } // caption text var caption = document.createElement('span'); box.appendChild(caption); caption.id = 'lightboxCaption'; caption.style.display = 'none'; caption.style.position = 'absolute'; caption.style.zIndex = '80'; // create effect image /* if (!option.effectpos) option.effectpos = {x:0,y:0}; else { if (option.effectpos.x == '') option.effectpos.x = 0; if (option.effectpos.y == '') option.effectpos.y = 0; } var effect = new Image; effect.onload = function() { var effectImg = document.createElement('img'); box.appendChild(effectImg); effectImg.id = 'effectImage'; effectImg.src = effect.src; if (option.effectclass) effectImg.className = option.effectclass; effectImg.style.position = 'absolute'; effectImg.style.display = 'none'; effectImg.style.left = [option.effectpos.x,'px'].join('');; effectImg.style.top = [option.effectpos.y,'px'].join(''); effectImg.style.zIndex = '90'; self._set_cursor(effectImg); Event.register(effectImg,'click',function() { effectImg.style.display = 'none'; }); }; if (option.effectimg != '') effect.src = option.effectimg;*/ if (self._resizable) { var overall = document.createElement('div'); obj.appendChild(overall); overall.id = 'lightboxOverallView'; overall.style.display = 'none'; overall.style.position = 'absolute'; overall.style.zIndex = '70'; self._overall = overall; var indicator = document.createElement('div'); obj.appendChild(indicator); indicator.id = 'lightboxIndicator'; indicator.style.display = 'none'; indicator.style.position = 'absolute'; indicator.style.zIndex = '80'; self._indicator = indicator; } return box; }, _set_photo_size : function() { var self = this; if (self._open == -1) return; var targ = { w:self._page.win.w - 30, h:self._page.win.h - 30 }; var zoom = { x:15, y:15 }; var navi = { p:9, n:9, y:0 }; if (!self._expanded) { // shrink image with the same aspect var orig = { w:self._imgs[self._open].w, h:self._imgs[self._open].h }; var ratio = 1.0; if ((orig.w >= targ.w || orig.h >= targ.h) && orig.h && orig.w) ratio = ((targ.w / orig.w) < (targ.h / orig.h)) ? targ.w / orig.w : targ.h / orig.h; self._img.width = Math.floor(orig.w * ratio); self._img.height = Math.floor(orig.h * ratio); self._expandable = (ratio < 1.0) ? true : false; if (self._resizable) self._expandable = true; if (Browser.isWinIE) self._box.style.display = "block"; self._imgpos.x = self._pos.x + (targ.w - self._img.width) / 2; self._imgpos.y = self._pos.y + (targ.h - self._img.height) / 2; navi.y = Math.floor(self._img.height / 2) - 10; self._show_caption(true); self._show_overall(false); } else { // zoomed or actual sized image var width = parseInt(self._imgs[self._open].w * self._level); var height = parseInt(self._imgs[self._open].h * self._level); self._minpos.x = self._pos.x + targ.w - width; self._minpos.y = self._pos.y + targ.h - height; if (width <= targ.w) self._imgpos.x = self._pos.x + (targ.w - width) / 2; else { if (self._imgpos.x > self._pos.x) self._imgpos.x = self._pos.x; else if (self._imgpos.x < self._minpos.x) self._imgpos.x = self._minpos.x; zoom.x = 15 + self._pos.x - self._imgpos.x; navi.p = self._pos.x - self._imgpos.x - 5; navi.n = width - self._page.win.w + self._imgpos.x + 25; if (Browser.isWinIE) navi.n -= 10; } if (height <= targ.h) { self._imgpos.y = self._pos.y + (targ.h - height) / 2; navi.y = Math.floor(self._img.height / 2) - 10; } else { if (self._imgpos.y > self._pos.y) self._imgpos.y = self._pos.y; else if (self._imgpos.y < self._minpos.y) self._imgpos.y = self._minpos.y; zoom.y = 15 + self._pos.y - self._imgpos.y; navi.y = Math.floor(targ.h / 2) - 10 + self._pos.y - self._imgpos.y; } self._img.width = width; self._img.height = height; self._show_caption(false); self._show_overall(true); } self._box.style.left = [self._imgpos.x,'px'].join(''); self._box.style.top = [self._imgpos.y,'px'].join(''); self._zoomimg.style.left = [zoom.x,'px'].join(''); self._zoomimg.style.top = [zoom.y,'px'].join(''); self._wrap.style.left = self._pos.x; if (self._prev && self._next) { self._prev.style.left = [navi.p,'px'].join(''); self._next.style.right = [navi.n,'px'].join(''); self._prev.style.top = self._next.style.top = [navi.y,'px'].join(''); } }, _show_overall : function(visible) { var self = this; if (self._overall == null) return; if (visible) { if (self._open == -1) return; var base = 100; var outer = { w:0, h:0, x:0, y:0 }; var inner = { w:0, h:0, x:0, y:0 }; var orig = { w:self._img.width , h:self._img.height }; var targ = { w:self._page.win.w - 30, h:self._page.win.h - 30 }; var max = orig.w; if (max < orig.h) max = orig.h; if (max < targ.w) max = targ.w; if (max < targ.h) max = targ.h; if (max < 1) return; outer.w = parseInt(orig.w / max * base); outer.h = parseInt(orig.h / max * base); inner.w = parseInt(targ.w / max * base); inner.h = parseInt(targ.h / max * base); outer.x = self._pos.x + targ.w - base - 20; outer.y = self._pos.y + targ.h - base - 20; inner.x = outer.x - parseInt((self._imgpos.x - self._pos.x) / max * base); inner.y = outer.y - parseInt((self._imgpos.y - self._pos.y) / max * base); self._overall.style.left = [outer.x,'px'].join(''); self._overall.style.top = [outer.y,'px'].join(''); self._overall.style.width = [outer.w,'px'].join(''); self._overall.style.height = [outer.h,'px'].join(''); self._indicator.style.left = [inner.x,'px'].join(''); self._indicator.style.top = [inner.y,'px'].join(''); self._indicator.style.width = [inner.w,'px'].join(''); self._indicator.style.height = [inner.h,'px'].join(''); self._overall.style.display = 'block' self._indicator.style.display = 'block'; } else { self._overall.style.display = 'none'; self._indicator.style.display = 'none'; } }, _set_size : function(onResize) { var self = this; if (self._open == -1) return; self._page.update(); self._pos.update(); var spin = self._wrap.firstChild; if (spin) { var top = (self._page.win.h - spin.height) / 2; if (self._wrap.style.position == 'absolute') top += self._pos.y; spin.style.top = [top,'px'].join(''); spin.style.left = [(self._page.win.w - spin.width - 30) / 2,'px'].join(''); } if (Browser.isWinIE) { self._wrap.style.width = [self._page.win.w,'px'].join(''); self._wrap.style.height = [self._page.win.h,'px'].join(''); self._wrap.style.top = [self._pos.y,'px'].join(''); } if (onResize) self._set_photo_size(); }, _set_cursor : function(obj) { var self = this; if (Browser.isWinIE && !Browser.isNewIE) return; obj.style.cursor = 'pointer'; }, _current_setindex : function() { var self = this; if (!self._openedset) return -1; var list = self._sets[self._openedset]; for (var i = 0,n = list.length; i < n; i++) { if (list[i] == self._open) return i; } return -1; }, _get_setlength : function() { var self = this; if (!self._openedset) return -1; return self._sets[self._openedset].length; }, _show_action : function() { var self = this; if (self._open == -1 || !self._expandable) return; if (!self._zoomimg) return; self._zoomimg.src = (self._expanded) ? self._shrink : self._expand;<|fim▁hole|> if (check > -1) { if (check > 0) self._prev.style.display = 'inline'; if (check < self._get_setlength() - 1) self._next.style.display = 'inline'; } }, _hide_action : function() { var self = this; if (self._zoomimg) self._zoomimg.style.display = 'none'; if (self._open > -1 && self._expanded) self._dragstop(null); if (self._prev) self._prev.style.display = 'none'; if (self._next) self._next.style.display = 'none'; }, _zoom : function() { var self = this; var closeBtn = document.getElementById('closeButton'); if (self._expanded) { self._reset_func(); self._expanded = false; if (closeBtn) closeBtn.style.display = 'inline'; } else if (self._open > -1) { self._level = 1; self._imgpos.x = self._pos.x; self._imgpos.y = self._pos.y; self._expanded = true; self._funcs.drag = function(evt) { self._dragstart(evt) }; self._funcs.dbl = function(evt) { self._close(null) }; if (self._resizable) { self._funcs.wheel = function(evt) { self._onwheel(evt) }; Event.register(self._box, 'mousewheel', self._funcs.wheel); } Event.register(self._img, 'mousedown', self._funcs.drag); Event.register(self._img, 'dblclick', self._funcs.dbl); if (closeBtn) closeBtn.style.display = 'none'; } self._set_photo_size(); self._show_action(); }, _reset_func : function() { var self = this; if (self._funcs.wheel != null) Event.deregister(self._box, 'mousewheel', self._funcs.wheel); if (self._funcs.move != null) Event.deregister(self._img, 'mousemove', self._funcs.move); if (self._funcs.up != null) Event.deregister(self._img, 'mouseup', self._funcs.up); if (self._funcs.drag != null) Event.deregister(self._img, 'mousedown', self._funcs.drag); if (self._funcs.dbl != null) Event.deregister(self._img, 'dblclick', self._funcs.dbl); self._funcs = {'move':null,'up':null,'drag':null,'wheel':null,'dbl':null}; }, _onwheel : function(evt) { var self = this; var delta = 0; evt = Event.getEvent(evt); if (evt.wheelDelta) delta = event.wheelDelta / -120; else if (evt.detail) delta = evt.detail / 3; if (Browser.isOpera) delta = - delta; var step = (self._level < 1) ? 0.1 : (self._level < 2) ? 0.25 : (self._level < 4) ? 0.5 : 1; self._level = (delta > 0) ? self._level + step : self._level - step; if (self._level > 8) self._level = 8; else if (self._level < 0.5) self._level = 0.5; self._set_photo_size(); return Event.stop(evt); }, _dragstart : function(evt) { var self = this; evt = Event.getEvent(evt); self._curpos.x = evt.screenX; self._curpos.y = evt.screenY; self._funcs.move = function(evnt) { self._dragging(evnt); }; self._funcs.up = function(evnt) { self._dragstop(evnt); }; Event.register(self._img, 'mousemove', self._funcs.move); Event.register(self._img, 'mouseup', self._funcs.up); return Event.stop(evt); }, _dragging : function(evt) { var self = this; evt = Event.getEvent(evt); self._imgpos.x += evt.screenX - self._curpos.x; self._imgpos.y += evt.screenY - self._curpos.y; self._curpos.x = evt.screenX; self._curpos.y = evt.screenY; self._set_photo_size(); return Event.stop(evt); }, _dragstop : function(evt) { var self = this; evt = Event.getEvent(evt); if (self._funcs.move != null) Event.deregister(self._img, 'mousemove', self._funcs.move); if (self._funcs.up != null) Event.deregister(self._img, 'mouseup', self._funcs.up); self._funcs.move = null; self._funcs.up = null; self._set_photo_size(); return (evt) ? Event.stop(evt) : false; }, _show_caption : function(enable) { var self = this; var caption = document.getElementById('lightboxCaption'); if (!caption) return; if (caption.innerHTML.length == 0 || !enable) { caption.style.display = 'none'; } else { // now display caption caption.style.top = [self._img.height + 10,'px'].join(''); // 10 is top margin of lightbox caption.style.left = '0px'; caption.style.width = [self._img.width + 20,'px'].join(''); // 20 is total side margin of lightbox caption.style.display = 'block'; } }, _toggle_wrap : function(flag) { var self = this; self._wrap.style.display = flag ? "block" : "none"; if (self._hiding.length == 0 && !self._first) { // some objects may overlap on overlay, so we hide them temporarily. var tags = ['select','embed','object']; for (var i = 0,n = tags.length; i < n; i++) { var elem = document.getElementsByTagName(tags[i]); for (var j = 0,m = elem.length; j < m; j++) { // check the original value at first. when alredy hidden, dont touch them var check = elem[j].style.visibility; if (!check) { if (elem[j].currentStyle) check = elem[j].currentStyle['visibility']; else if (document.defaultView) check = document.defaultView.getComputedStyle(elem[j], '').getPropertyValue('visibility'); } if (check == 'hidden') continue; self._hiding.push(elem[j]); } } self._first = true; } for (var i = 0,n = self._hiding.length; i < n; i++) self._hiding[i].style.visibility = flag ? "hidden" : "visible"; }, _show : function(num) { var self = this; var imag = new Image; if (num < 0 || num >= self._imgs.length) return; var loading = document.getElementById('loadingImage'); var caption = document.getElementById('lightboxCaption'); // var effect = document.getElementById('effectImage'); self._open = num; // set opened image number self._set_size(false); // calc and set wrapper size self._toggle_wrap(true); if (loading) loading.style.display = 'inline'; imag.onload = function() { if (self._imgs[self._open].w == -1) { // store original image width and height self._imgs[self._open].w = imag.width; self._imgs[self._open].h = imag.height; } /* if (effect) { effect.style.display = (!effect.className || self._imgs[self._open].cls == effect.className) ? 'block' : 'none'; }*/ if (caption) try { caption.innerHTML = self._imgs[self._open].title; } catch(e) { } self._set_photo_size(); // calc and set lightbox size self._hide_action(); self._box.style.display = "block"; self._img.src = imag.src; self._img.setAttribute('title', self._imgs[self._open].title); self._timer = window.setInterval(function() { self._set_size(true) }, 100); if (loading) loading.style.display = 'none'; if (self._imgs[self._open].set != 'lightbox') { var set = self._imgs[self._open].set; if (self._sets[set].length > 1) self._openedset = set; if (!self._prev || !self._next) self._openedset = null; } }; self._expandable = false; self._expanded = false; imag.src = self._imgs[self._open].src; }, _close_box : function() { var self = this; self._open = -1; self._openedset = null; self._hide_action(); self._hide_action(); self._reset_func(); self._show_overall(false); self._box.style.display = "none"; if (self._timer != null) { window.clearInterval(self._timer); self._timer = null; } }, _show_next : function(direction) { var self = this; if (!self._openedset) return self._close(null); var index = self._current_setindex() + direction; var targ = self._sets[self._openedset][index]; self._close_box(); self._show(targ); }, _close : function(evt) { var self = this; if (evt != null) { evt = Event.getEvent(evt); var targ = evt.target || evt.srcElement; if (targ && targ.getAttribute('id') == 'lightboxImage' && self._expanded) return; } self._close_box(); self._toggle_wrap(false); } }; Event.register(window, "load", function() { var lightbox = new LightBox({ loadingimg:'lightbox/loading.gif', expandimg:'lightbox/expand.gif', shrinkimg:'lightbox/shrink.gif', previmg:'lightbox/prev.gif', nextimg:'lightbox/next.gif', /* effectpos:{x:-40,y:-20}, effectclass:'effectable',*/ closeimg:'lightbox/close.gif', resizable:true }); });<|fim▁end|>
self._zoomimg.style.display = 'inline'; var check = self._current_setindex();
<|file_name|>ClusterLauncher.py<|end_file_name|><|fim▁begin|>import os import sys import time import subprocess import yaml import pathlib class ClusterLauncher: def __init__(self, config_yaml): self.Config = config_yaml <|fim▁hole|> #launch head node head = ClusterLauncher.LaunchUniCAVEWindow(config["build-path"], config["head-node"]) #wait a bit before launching child nodes if config["head-wait"] is not None: time.sleep(config["head-wait"]) #launch child nodes children = [] for child_node in config["child-nodes"]: children.append(ClusterLauncher.LaunchUniCAVEWindow(config["build-path"], child_node)) #wait a bit between launching each child if config["child-wait"] is not None: time.sleep(config["child-wait"]) #poll head node process done = False while not done: if head.poll() is not None: done = True time.sleep(config["sleep-time"]) #when done, close child processes and exit for child in children: child.kill() @staticmethod def LaunchUniCAVEWindow(path, machine_name=None): args = [path, "-popupWindow"] if machine_name is not None: args = args + ["overrideMachineName", machine_name] return subprocess.Popen(args) if __name__ == "__main__": ClusterLauncher(os.path.join(pathlib.Path(__file__).parent.absolute(), "config/input_test.yaml")).Launch()<|fim▁end|>
def Launch(self): #read config with open(self.Config, 'r') as yf: config = yaml.safe_load(yf)
<|file_name|>plotarea.py<|end_file_name|><|fim▁begin|>from __future__ import absolute_import # Copyright (c) 2010-2019 openpyxl <|fim▁hole|> Alias, ) from openpyxl.descriptors.excel import ( ExtensionList, ) from openpyxl.descriptors.sequence import ( MultiSequence, MultiSequencePart, ) from openpyxl.descriptors.nested import ( NestedBool, ) from ._3d import _3DBase from .area_chart import AreaChart, AreaChart3D from .bar_chart import BarChart, BarChart3D from .bubble_chart import BubbleChart from .line_chart import LineChart, LineChart3D from .pie_chart import PieChart, PieChart3D, ProjectedPieChart, DoughnutChart from .radar_chart import RadarChart from .scatter_chart import ScatterChart from .stock_chart import StockChart from .surface_chart import SurfaceChart, SurfaceChart3D from .layout import Layout from .shapes import GraphicalProperties from .text import RichText from .axis import ( NumericAxis, TextAxis, SeriesAxis, DateAxis, ) class DataTable(Serialisable): tagname = "dTable" showHorzBorder = NestedBool(allow_none=True) showVertBorder = NestedBool(allow_none=True) showOutline = NestedBool(allow_none=True) showKeys = NestedBool(allow_none=True) spPr = Typed(expected_type=GraphicalProperties, allow_none=True) graphicalProperties = Alias('spPr') txPr = Typed(expected_type=RichText, allow_none=True) extLst = Typed(expected_type=ExtensionList, allow_none=True) __elements__ = ('showHorzBorder', 'showVertBorder', 'showOutline', 'showKeys', 'spPr', 'txPr') def __init__(self, showHorzBorder=None, showVertBorder=None, showOutline=None, showKeys=None, spPr=None, txPr=None, extLst=None, ): self.showHorzBorder = showHorzBorder self.showVertBorder = showVertBorder self.showOutline = showOutline self.showKeys = showKeys self.spPr = spPr self.txPr = txPr class PlotArea(Serialisable): tagname = "plotArea" layout = Typed(expected_type=Layout, allow_none=True) dTable = Typed(expected_type=DataTable, allow_none=True) spPr = Typed(expected_type=GraphicalProperties, allow_none=True) graphicalProperties = Alias("spPr") extLst = Typed(expected_type=ExtensionList, allow_none=True) # at least one chart _charts = MultiSequence() areaChart = MultiSequencePart(expected_type=AreaChart, store="_charts") area3DChart = MultiSequencePart(expected_type=AreaChart3D, store="_charts") lineChart = MultiSequencePart(expected_type=LineChart, store="_charts") line3DChart = MultiSequencePart(expected_type=LineChart3D, store="_charts") stockChart = MultiSequencePart(expected_type=StockChart, store="_charts") radarChart = MultiSequencePart(expected_type=RadarChart, store="_charts") scatterChart = MultiSequencePart(expected_type=ScatterChart, store="_charts") pieChart = MultiSequencePart(expected_type=PieChart, store="_charts") pie3DChart = MultiSequencePart(expected_type=PieChart3D, store="_charts") doughnutChart = MultiSequencePart(expected_type=DoughnutChart, store="_charts") barChart = MultiSequencePart(expected_type=BarChart, store="_charts") bar3DChart = MultiSequencePart(expected_type=BarChart3D, store="_charts") ofPieChart = MultiSequencePart(expected_type=ProjectedPieChart, store="_charts") surfaceChart = MultiSequencePart(expected_type=SurfaceChart, store="_charts") surface3DChart = MultiSequencePart(expected_type=SurfaceChart3D, store="_charts") bubbleChart = MultiSequencePart(expected_type=BubbleChart, store="_charts") # axes _axes = MultiSequence() valAx = MultiSequencePart(expected_type=NumericAxis, store="_axes") catAx = MultiSequencePart(expected_type=TextAxis, store="_axes") dateAx = MultiSequencePart(expected_type=DateAxis, store="_axes") serAx = MultiSequencePart(expected_type=SeriesAxis, store="_axes") __elements__ = ('layout', '_charts', '_axes', 'dTable', 'spPr') def __init__(self, layout=None, dTable=None, spPr=None, _charts=(), _axes=(), extLst=None, ): self.layout = layout self.dTable = dTable self.spPr = spPr self._charts = _charts self._axes = _axes def to_tree(self, tagname=None, idx=None, namespace=None): axIds = set((ax.axId for ax in self._axes)) for chart in self._charts: for id, axis in chart._axes.items(): if id not in axIds: setattr(self, axis.tagname, axis) axIds.add(id) return super(PlotArea, self).to_tree(tagname) @classmethod def from_tree(cls, node): self = super(PlotArea, cls).from_tree(node) axes = dict((axis.axId, axis) for axis in self._axes) for chart in self._charts: if isinstance(chart, ScatterChart): x, y = (axes[axId] for axId in chart.axId) chart.x_axis = x chart.y_axis = y continue for axId in chart.axId: axis = axes.get(axId) if axis is None and isinstance(chart, _3DBase): # Series Axis can be optional chart.z_axis = None continue if axis.tagname in ("catAx", "dateAx"): chart.x_axis = axis elif axis.tagname == "valAx": chart.y_axis = axis elif axis.tagname == "serAx": chart.z_axis = axis return self<|fim▁end|>
from openpyxl.descriptors.serialisable import Serialisable from openpyxl.descriptors import ( Typed,
<|file_name|>MentionExtractor.java<|end_file_name|><|fim▁begin|>package cmucoref.mention.extractor; import cmucoref.document.Document; import cmucoref.document.Lexicon; import cmucoref.document.Sentence; import cmucoref.exception.MentionException; import cmucoref.mention.Mention; import cmucoref.mention.eventextractor.EventExtractor; import cmucoref.mention.extractor.relationextractor.*; import cmucoref.model.Options; import cmucoref.util.Pair; import cmucoref.mention.SpeakerInfo; import cmucoref.mention.WordNet; import cmucoref.mention.Dictionaries; import java.io.FileNotFoundException; import java.io.IOException; import java.io.InputStream; import java.io.InputStreamReader; import java.io.PrintStream; import java.util.ArrayList; import java.util.Arrays; import java.util.Collections; import java.util.HashMap; import java.util.HashSet; import java.util.List; import java.util.Map; import java.util.Set; import java.util.Properties; public abstract class MentionExtractor { protected Dictionaries dict; protected WordNet wordNet; protected EventExtractor eventExtractor; public MentionExtractor(){} public void createDict(String propfile) throws FileNotFoundException, IOException { Properties props = new Properties(); InputStream in = MentionExtractor.class.getClassLoader().getResourceAsStream(propfile); props.load(new InputStreamReader(in)); this.dict = new Dictionaries(props); } public void createWordNet(String wnDir) throws IOException { wordNet = new WordNet(wnDir); } public void closeWordNet() { wordNet.close(); } public void setEventExtractor(EventExtractor eventExtractor) { this.eventExtractor = eventExtractor; } public Dictionaries getDict() { return dict; } public WordNet getWordNet() { return wordNet; } public int sizeOfEvent() { return eventExtractor.sizeOfEvent(); } public static MentionExtractor createExtractor(String extractorClassName) throws InstantiationException, IllegalAccessException, ClassNotFoundException { return (MentionExtractor) Class.forName(extractorClassName).newInstance(); } public abstract List<List<Mention>> extractPredictedMentions(Document doc, Options options) throws IOException; protected void deleteDuplicatedMentions(List<Mention> mentions, Sentence sent) { //remove duplicated mentions Set<Mention> remove = new HashSet<Mention>(); for(int i = 0; i < mentions.size(); ++i) { Mention mention1 = mentions.get(i); for(int j = i + 1; j < mentions.size(); ++j) { Mention mention2 = mentions.get(j); if(mention1.equals(mention2)) { remove.add(mention2); } } } mentions.removeAll(remove); } protected void deleteSpuriousNamedEntityMentions(List<Mention> mentions, Sentence sent) { //remove overlap mentions Set<Mention> remove = new HashSet<Mention>(); for(Mention mention1 : mentions) { if(mention1.isPureNerMention(sent, dict)) { for(Mention mention2 : mentions) { if(mention1.overlap(mention2)) { remove.add(mention1); } } } } mentions.removeAll(remove); //remove single number named entity mentions remove.clear(); String[] NUMBERS = {"NUMBER", "ORDINAL", "CARDINAL", "MONEY", "QUANTITY"}; HashSet<String> numberNER = new HashSet<String>(Arrays.asList(NUMBERS)); for(Mention mention : mentions) { if(mention.endIndex - mention.startIndex == 1) { if(numberNER.contains(mention.headword.ner)) { remove.add(mention); } } } mentions.removeAll(remove); //remove NORP mentions as modifiers remove.clear(); for(Mention mention : mentions) { if((dict.isAdjectivalDemonym(mention.getSpan(sent)) || mention.headword.ner.equals("NORP")) && (mention.headword.postag.equals("JJ") || !dict.rolesofNoun.contains(mention.headword.basic_deprel))) { remove.add(mention); } } mentions.removeAll(remove); //remove mentions with non-noun head //TODO } protected void deleteSpuriousPronominalMentions(List<Mention> mentions, Sentence sent) { //remove "you know" mentions Set<Mention> remove = new HashSet<Mention>(); for(Mention mention : mentions) { if(mention.isPronominal() && (mention.endIndex - mention.startIndex == 1) && mention.headString.equals("you")) { if(mention.headIndex + 1 < sent.length()) { Lexicon lex = sent.getLexicon(mention.headIndex + 1); if(lex.form.equals("know")) { remove.add(mention); } } } } mentions.removeAll(remove); //remove "you know" part in a mention remove.clear(); for(Mention mention : mentions) { if(mention.endIndex - mention.startIndex > 2) { if(sent.getLexicon(mention.endIndex - 2).form.toLowerCase().equals("you") && sent.getLexicon(mention.endIndex - 1).form.toLowerCase().equals("know")) { mention.endIndex = mention.endIndex - 2; boolean duplicated = false; for(Mention m2 : mentions) { if(mention == m2) { continue; } if(mention.equals(m2)) { duplicated = true; break; } } if(duplicated) { remove.add(mention); } else { mention.process(sent, mentions, dict, wordNet, remove); } } } } mentions.removeAll(remove); } public List<Mention> getSingleMentionList(Document doc, List<List<Mention>> mentionList, Options options) throws InstantiationException, IllegalAccessException, ClassNotFoundException { List<Mention> allMentions = new ArrayList<Mention>(); for(List<Mention> mentions : mentionList) { allMentions.addAll(mentions); } //extract events for mentions if(options.useEventFeature()) { extractEvents(mentionList, doc, options); } //find speaker for each mention findSpeakers(doc, allMentions, mentionList); //Collections.sort(allMentions, Mention.syntacticOrderComparator); //re-assign mention ID; for(int i = 0; i < allMentions.size(); ++i) { Mention mention = allMentions.get(i); mention.mentionID = i; } if(options.usePreciseMatch()) { findPreciseMatchRelation(doc, allMentions); } return allMentions; } /** * * @param doc * @param allMentions */ protected void findSpeakers(Document doc, List<Mention> allMentions, List<List<Mention>> mentionList) { Map<String, SpeakerInfo> speakersMap = new HashMap<String, SpeakerInfo>(); speakersMap.put("<DEFAULT_SPEAKER>", new SpeakerInfo(0, "<DEFAULT_SPEAKER>", false)); // find default speakers from the speaker tags of document doc findDefaultSpeakers(doc, allMentions, speakersMap); //makr quotations markQuotaions(doc, false); findQuotationSpeakers(doc, allMentions, mentionList, dict, speakersMap); Collections.sort(allMentions, Mention.headIndexWithSpeakerOrderComparator); //find previous speakerinfo SpeakerInfo defaultSpeakerInfo = speakersMap.get("<DEFAULT_SPEAKER>"); SpeakerInfo preSpeakerInfo = defaultSpeakerInfo; Mention preMention = null; for(Mention mention : allMentions) { if(mention.speakerInfo.isQuotationSpeaker()) { continue; } if(preMention != null && !preMention.speakerInfo.equals(mention.speakerInfo)) { preSpeakerInfo = preMention.speakerInfo; } if(preSpeakerInfo.equals(defaultSpeakerInfo)) { mention.preSpeakerInfo = null; } else { mention.preSpeakerInfo = preSpeakerInfo; } preMention = mention; } } protected void findQuotationSpeakers(Document doc, List<Mention> allMentions, List<List<Mention>> mentionList, Dictionaries dict, Map<String, SpeakerInfo> speakersMap) { Pair<Integer, Integer> beginQuotation = new Pair<Integer, Integer>(); Pair<Integer, Integer> endQuotation = new Pair<Integer, Integer>(); boolean insideQuotation = false; int sizeOfDoc = doc.size(); for(int i = 0; i < sizeOfDoc; ++i) { Sentence sent = doc.getSentence(i); for(int j = 1; j < sent.length(); ++j) { int utterIndex = sent.getLexicon(j).utterance; if(utterIndex != 0 && !insideQuotation) { insideQuotation = true; beginQuotation.first = i; beginQuotation.second = j; } else if(utterIndex == 0 && insideQuotation) { insideQuotation = false; endQuotation.first = i; endQuotation.second = j; findQuotationSpeakers(doc, allMentions, mentionList, dict, speakersMap, beginQuotation, endQuotation); } } } if(insideQuotation) { endQuotation.first = sizeOfDoc - 1; endQuotation.second = doc.getSentence(endQuotation.first).length(); findQuotationSpeakers(doc, allMentions, mentionList, dict, speakersMap, beginQuotation, endQuotation); } } protected void findQuotationSpeakers(Document doc, List<Mention> allMentions, List<List<Mention>> mentionList, Dictionaries dict, Map<String, SpeakerInfo> speakersMap, Pair<Integer, Integer> beginQuotation, Pair<Integer, Integer> endQuotation) { Sentence sent = doc.getSentence(beginQuotation.first); List<Mention> mentions = mentionList.get(beginQuotation.first); SpeakerInfo speakerInfo = findQuotationSpeaker(sent, mentions, 1, beginQuotation.second, dict, speakersMap); if(speakerInfo != null) { assignUtterancetoSpeaker(doc, mentionList, dict, beginQuotation, endQuotation, speakerInfo); return; } sent = doc.getSentence(endQuotation.first); mentions = mentionList.get(endQuotation.first); speakerInfo = findQuotationSpeaker(sent, mentions, endQuotation.second, sent.length(), dict, speakersMap); if(speakerInfo != null) { assignUtterancetoSpeaker(doc, mentionList, dict, beginQuotation, endQuotation, speakerInfo); return; } if(beginQuotation.second <= 2 && beginQuotation.first > 0) { sent = doc.getSentence(beginQuotation.first - 1); mentions = mentionList.get(beginQuotation.first - 1); speakerInfo = findQuotationSpeaker(sent, mentions, 1, sent.length(), dict, speakersMap); if(speakerInfo != null) { assignUtterancetoSpeaker(doc, mentionList, dict, beginQuotation, endQuotation, speakerInfo); return; } } if(endQuotation.second == doc.getSentence(endQuotation.first).length() - 1 && doc.size() > endQuotation.first + 1) { sent = doc.getSentence(endQuotation.first + 1); mentions = mentionList.get(endQuotation.first + 1); speakerInfo = findQuotationSpeaker(sent, mentions, 1, sent.length(), dict, speakersMap); if(speakerInfo != null) { assignUtterancetoSpeaker(doc, mentionList, dict, beginQuotation, endQuotation, speakerInfo); return; } } } private void assignUtterancetoSpeaker(Document doc, List<List<Mention>> mentionList, Dictionaries dict, Pair<Integer, Integer> beginQuotation, Pair<Integer, Integer> endQuotation, SpeakerInfo speakerInfo) { for(int i = beginQuotation.first; i <= endQuotation.first; ++i) { Sentence sent = doc.getSentence(i); int start = i == beginQuotation.first ? beginQuotation.second : 1; int end = i == endQuotation.first ? endQuotation.second : sent.length() - 1; List<Mention> mentions = mentionList.get(i); for(Mention mention : mentions) { if(mention.startIndex >= start && mention.endIndex <= end) { mention.setSpeakerInfo(speakerInfo); } } } } protected SpeakerInfo findQuotationSpeaker(Sentence sent, List<Mention> mentions, int startIndex, int endIndex, Dictionaries dict, Map<String, SpeakerInfo> speakersMap) { for(int i = endIndex - 1; i >= startIndex; --i) { if(sent.getLexicon(i).utterance != 0) { continue; } String lemma = sent.getLexicon(i).lemma; if(dict.reportVerb.contains(lemma)) { int reportVerbPos = i; Lexicon reportVerb = sent.getLexicon(reportVerbPos); for(int j = startIndex; j < endIndex; ++j) { Lexicon lex = sent.getLexicon(j); if(lex.collapsed_head == reportVerbPos && (lex.collapsed_deprel.equals("nsubj") || lex.collapsed_deprel.equals("xsubj")) || reportVerb.collapsed_deprel.startsWith("conj_") && lex.collapsed_head == reportVerb.collapsed_head && (lex.collapsed_deprel.equals("nsubj") || lex.collapsed_deprel.equals("xsubj"))) { int speakerHeadIndex = j; for(Mention mention : mentions) { if(mention.getBelognTo() == null && mention.headIndex == speakerHeadIndex && mention.startIndex >= startIndex && mention.endIndex < endIndex) { if(mention.utteranceInfo == null) { String speakerName = mention.getSpan(sent); SpeakerInfo speakerInfo = new SpeakerInfo(speakersMap.size(), speakerName, true); speakersMap.put(speakerInfo.toString(), speakerInfo); speakerInfo.setSpeaker(mention); mention.utteranceInfo = speakerInfo; } return mention.utteranceInfo; } } String speakerName = sent.getLexicon(speakerHeadIndex).form; SpeakerInfo speakerInfo = new SpeakerInfo(speakersMap.size(), speakerName, true); speakersMap.put(speakerInfo.toString(), speakerInfo); return speakerInfo; } } } } return null; } /** * mark quotations for a document * @param doc * @param normalQuotationType */ private void markQuotaions(Document doc, boolean normalQuotationType) { int utteranceIndex = 0; boolean insideQuotation = false; boolean hasQuotation = false; for(Sentence sent : doc.getSentences()) { for(Lexicon lex : sent.getLexicons()) { lex.utterance = utteranceIndex; if(lex.form.equals("``") || (!insideQuotation && normalQuotationType && lex.form.equals("\""))) { utteranceIndex++; lex.utterance = utteranceIndex; insideQuotation = true; hasQuotation = true; } else if((utteranceIndex > 0 && lex.form.equals("''")) || (insideQuotation && normalQuotationType && lex.form.equals("\""))) { insideQuotation = false; utteranceIndex--; } } } if(!hasQuotation && !normalQuotationType) { markQuotaions(doc, true); } } /** * find default speakers from the speaker tags of document * @param doc * @param allMentions * @param speakersMap */ protected void findDefaultSpeakers(Document doc, List<Mention> allMentions, Map<String, SpeakerInfo> speakersMap) { for(Mention mention : allMentions) { Sentence sent = doc.getSentence(mention.sentID); String speaker = sent.getSpeaker().equals("-") ? "<DEFAULT_SPEAKER>" : sent.getSpeaker(); SpeakerInfo speakerInfo = speakersMap.get(speaker); if(speakerInfo == null) { speakerInfo = new SpeakerInfo(speakersMap.size(), speaker, false); speakersMap.put(speaker, speakerInfo); } mention.setSpeakerInfo(speakerInfo); } } protected void findPreciseMatchRelation(Document doc, List<Mention> allMentions) { for(int i = 1; i < allMentions.size(); ++i) { Mention anaph = allMentions.get(i); //find precise match for(int j = 0; j < i; ++j) { Mention antec = allMentions.get(j); if(anaph.preciseMatch(doc.getSentence(anaph.sentID), antec, doc.getSentence(antec.sentID), dict)) { anaph.addPreciseMatch(antec); } } //find string match for(int j = i - 1; j >= 0; --j) { Mention antec = allMentions.get(j); if(anaph.stringMatch(doc.getSentence(anaph.sentID), antec, doc.getSentence(antec.sentID), dict)) { anaph.addStringMatch(antec); } } } } protected void extractEvents(List<List<Mention>> mentionList, Document doc, Options options) throws InstantiationException, IllegalAccessException, ClassNotFoundException { eventExtractor.extractEvents(doc, mentionList, options); } protected void findSyntacticRelation(List<Mention> mentions, Sentence sent, Options options) throws InstantiationException, IllegalAccessException, ClassNotFoundException, MentionException{ markListMemberRelation(mentions, sent, RelationExtractor.createExtractor(options.getListMemberRelationExtractor())); deleteSpuriousListMentions(mentions, sent); correctHeadIndexforNERMentions(mentions, sent); markAppositionRelation(mentions, sent, RelationExtractor.createExtractor(options.getAppositionRelationExtractor())); markRoleAppositionRelation(mentions, sent, RelationExtractor.createExtractor(options.getRoleAppositionRelationExtractor())); markPredicateNominativeRelation(mentions, sent, RelationExtractor.createExtractor(options.getPredicateNominativeRelationExtractor())); deletePleonasticItwithTemproal(mentions, sent); //markRelativePronounRelation(mentions, sent, RelationExtractor.createExtractor(options.getRelativePronounRelationExtractor())); } /** * remove nested mention with shared headword (except enumeration/list): pick larger one * @param mentions * @param sent */ protected void deleteSpuriousListMentions(List<Mention> mentions, Sentence sent) { Set<Mention> remove = new HashSet<Mention>(); for(Mention mention1 : mentions) { for(Mention mention2 : mentions) { if(mention1.headIndex == mention2.headIndex && mention2.cover(mention1) && mention1.getBelognTo() == null) { remove.add(mention1); } } } mentions.removeAll(remove); } /** * remove pleonastic it with Temporal mentions (e.g. it is summer) * @param mentions * @param sent */ protected void deletePleonasticItwithTemproal(List<Mention> mentions, Sentence sent) { Set<Mention> remove = new HashSet<Mention>(); for(Mention mention : mentions) { if(mention.isPronominal() && mention.headString.equals("it") && mention.getPredicateNominatives() != null) { for(Mention predN : mention.getPredicateNominatives()) { if(!mentions.contains(predN)) { continue; } if(predN.isProper() && predN.headword.ner.equals("DATE") || predN.isNominative() && dict.temporals.contains(predN.headString)) { remove.add(mention); break; } } } else if(mention.isPronominal() && mention.headString.equals("it")) { Lexicon headword = sent.getLexicon(mention.originalHeadIndex); int head = headword.collapsed_head; if(sent.getLexicon(head).lemma.equals("be") && headword.collapsed_deprel.equals("nsubj")) { for(Mention mention2 : mentions) { Lexicon headword2 = sent.getLexicon(mention2.originalHeadIndex); if(headword2.id > head && headword2.collapsed_head == head && headword2.collapsed_deprel.startsWith("prep_") && (mention2.isProper() && mention2.headword.ner.equals("DATE") || mention2.isNominative() && dict.temporals.contains(mention2.headString))) { remove.add(mention); } } } } } mentions.removeAll(remove); } protected void correctHeadIndexforNERMentions(List<Mention> mentions, Sentence sent) { for(Mention mention : mentions) { mention.correctHeadIndex(sent, dict, wordNet); } } protected void markListMemberRelation(List<Mention> mentions, Sentence sent, RelationExtractor extractor) throws MentionException { Set<Pair<Integer, Integer>> foundPairs = extractor.extractRelation(sent, mentions); markMentionRelation(mentions, sent, foundPairs, "LISTMEMBER"); } protected void markAppositionRelation(List<Mention> mentions, Sentence sent, RelationExtractor extractor) throws MentionException { Set<Pair<Integer, Integer>> foundPairs = extractor.extractRelation(sent, mentions); markMentionRelation(mentions, sent, foundPairs, "APPOSITION"); } protected void markRoleAppositionRelation(List<Mention> mentions, Sentence sent, RelationExtractor extractor) throws MentionException { Set<Pair<Integer, Integer>> foundPairs = extractor.extractRelation(sent, mentions); markMentionRelation(mentions, sent, foundPairs, "ROLE_APPOSITION"); } <|fim▁hole|> markMentionRelation(mentions, sent, foundPairs, "PREDICATE_NOMINATIVE"); } protected void markRelativePronounRelation(List<Mention> mentions, Sentence sent, RelationExtractor extractor) throws MentionException { Set<Pair<Integer, Integer>> foundPairs = extractor.extractRelation(sent, mentions); markMentionRelation(mentions, sent, foundPairs, "RELATIVE_PRONOUN"); } protected void markMentionRelation(List<Mention> mentions, Sentence sent, Set<Pair<Integer, Integer>> foundPairs, String relation) throws MentionException { for(Mention mention1 : mentions) { for(Mention mention2 : mentions) { if(mention1.equals(mention2)) { continue; } if(relation.equals("LISTMEMBER")) { for(Pair<Integer, Integer> pair : foundPairs) { if(pair.first == mention1.mentionID && pair.second == mention2.mentionID) { mention2.addListMember(mention1, sent); } } } else if(relation.equals("PREDICATE_NOMINATIVE")) { for(Pair<Integer, Integer> pair : foundPairs) { if(pair.first == mention1.mentionID && pair.second == mention2.mentionID) { mention2.addPredicativeNominative(mention1, dict); } } } else if(relation.equals("ROLE_APPOSITION")) { for(Pair<Integer, Integer> pair : foundPairs) { if(pair.first == mention1.mentionID && pair.second == mention2.mentionID) { mention2.addRoleApposition(mention1, sent, dict); } } } else{ for(Pair<Integer, Integer> pair : foundPairs) { if(pair.first == mention1.originalHeadIndex && pair.second == mention2.originalHeadIndex) { if(relation.equals("APPOSITION")) { mention2.addApposition(mention1, dict); } else if(relation.equals("RELATIVE_PRONOUN")) { mention2.addRelativePronoun(mention1); } else { throw new MentionException("Unknown mention relation: " + relation); } } } } } } } public void displayMentions(Document doc, List<List<Mention>> mentionList, PrintStream printer){ printer.println("#begin document " + doc.getFileName() + " docId " + doc.getDocId()); int sentId = 0; for(List<Mention> mentions : mentionList){ printer.println("sent Id: " + sentId); for(Mention mention : mentions){ displayMention(doc.getSentence(sentId), mention, printer); } sentId++; printer.println("----------------------------------------"); } printer.println("end document"); printer.flush(); } public void displayMention(Sentence sent, Mention mention, PrintStream printer){ mention.display(sent, printer); } }<|fim▁end|>
protected void markPredicateNominativeRelation(List<Mention> mentions, Sentence sent, RelationExtractor extractor) throws MentionException { Set<Pair<Integer, Integer>> foundPairs = extractor.extractRelation(sent, mentions);
<|file_name|>timeout.rs<|end_file_name|><|fim▁begin|>extern crate futures; use futures::{Future, Poll}; use std::io; use std::marker::PhantomData; use std::time::{Duration, Instant}; pub struct Timeout<T,E> { timestamp: Instant, duration: Duration, error: E, phantom: PhantomData<T>, } impl<T,E> Timeout<T,E> where E: Fn() -> io::Error { pub fn new(duration: Duration, e: E) -> Timeout<T, E> { Timeout { timestamp: Instant::now(), duration: duration, phantom: PhantomData, error: e, } } pub fn is_elapsed(&self) -> bool { self.timestamp.elapsed() >= self.duration } } impl<T,E> Future for Timeout<T,E> where E: Fn() -> io::Error { type Item = T; type Error = io::Error; // Return type of the Future::poll method, indicates whether a future's value is ready or not. // // Ok(Async::Ready(t)) means that a future has successfully resolved // Ok(Async::NotReady) means that a future is not ready to complete yet // Err(e) means that a future has completed with the given failure fn poll(&mut self) -> Poll<Self::Item, Self::Error> { use futures::{Async, task}; if self.is_elapsed() { Err((self.error)())<|fim▁hole|> // in the future. Otherwise poll will only be run once. Ok(Async::NotReady) } } }<|fim▁end|>
} else { task::park().unpark(); // this tells the task driving the future to recheck this again
<|file_name|>jslint-check.js<|end_file_name|><|fim▁begin|>load("build/jslint.js"); var src = readFile("dist/jquery.ImageColorPicker.js"); JSLINT(src, { evil: true, forin: true }); // All of the following are known issues that we think are 'ok'<|fim▁hole|>// http://docs.jquery.com/JQuery_Core_Style_Guidelines var ok = { "Expected an identifier and instead saw 'undefined' (a reserved word).": true, "Use '===' to compare with 'null'.": true, "Use '!==' to compare with 'null'.": true, "Expected an assignment or function call and instead saw an expression.": true, "Expected a 'break' statement before 'case'.": true }; var e = JSLINT.errors, found = 0, w; for ( var i = 0; i < e.length; i++ ) { w = e[i]; if ( !ok[ w.reason ] ) { found++; print( "\n" + w.evidence + "\n" ); print( " Problem at line " + w.line + " character " + w.character + ": " + w.reason ); } } if ( found > 0 ) { print( "\n" + found + " Error(s) found." ); } else { print( "JSLint check passed." ); }<|fim▁end|>
// (in contradiction with JSLint) more information here:
<|file_name|>js_JIzlcM6aM0LMxxWAo8iNMOVxFXtNamiOTTx6GLKu6cc.js<|end_file_name|><|fim▁begin|>/** * @file * Javascript for Goole Map widget of Geolocation field. */ (function ($) { var geocoder; Drupal.geolocation = Drupal.geolocation || {}; Drupal.geolocation.maps = Drupal.geolocation.maps || {}; Drupal.geolocation.markers = Drupal.geolocation.markers || {}; /** * Set the latitude and longitude values to the input fields * And optionaly update the address field * * @param latLng * a location (latLng) object from google maps api * @param i * the index from the maps array we are working on * @param op * the op that was performed */ Drupal.geolocation.codeLatLng = function(latLng, i, op) { // Update the lat and lng input fields $('#geolocation-lat-' + i + ' input').attr('value', latLng.lat()); $('#geolocation-lat-item-' + i + ' .geolocation-lat-item-value').html(latLng.lat()); $('#geolocation-lng-' + i + ' input').attr('value', latLng.lng()); $('#geolocation-lng-item-' + i + ' .geolocation-lat-item-value').html(latLng.lng()); // Update the address field if ((op == 'marker' || op == 'geocoder') && geocoder) { geocoder.geocode({'latLng': latLng}, function(results, status) { if (status == google.maps.GeocoderStatus.OK) { $('#geolocation-address-' + i + ' input').val(results[0].formatted_address); if (op == 'geocoder') { Drupal.geolocation.setZoom(i, results[0].geometry.location_type); } } else { $('#geolocation-address-' + i + ' input').val(''); if (status != google.maps.GeocoderStatus.ZERO_RESULTS) { alert(Drupal.t('Geocoder failed due to: ') + status); } } }); } } /** * Get the location from the address field * * @param i * the index from the maps array we are working on */ Drupal.geolocation.codeAddress = function(i) { var address = $('#geolocation-address-' + i + ' input').val(); geocoder.geocode( { 'address': address }, function(results, status) { if (status == google.maps.GeocoderStatus.OK) { Drupal.geolocation.maps[i].setCenter(results[0].geometry.location); Drupal.geolocation.setMapMarker(results[0].geometry.location, i); Drupal.geolocation.codeLatLng(results[0].geometry.location, i, 'textinput'); Drupal.geolocation.setZoom(i, results[0].geometry.location_type); } else { alert(Drupal.t('Geocode was not successful for the following reason: ') + status); } }); } /** * Set zoom level depending on accuracy (location_type) * * @param location_type * location type as provided by google maps after geocoding a location */ Drupal.geolocation.setZoom = function(i, location_type) { if (location_type == 'APPROXIMATE') { Drupal.geolocation.maps[i].setZoom(10); } else if (location_type == 'GEOMETRIC_CENTER') { Drupal.geolocation.maps[i].setZoom(12); } else if (location_type == 'RANGE_INTERPOLATED' || location_type == 'ROOFTOP') { Drupal.geolocation.maps[i].setZoom(16); } } /** * Set/Update a marker on a map * * @param latLng * a location (latLng) object from google maps api * @param i * the index from the maps array we are working on */ Drupal.geolocation.setMapMarker = function(latLng, i) { // remove old marker if (Drupal.geolocation.markers[i]) { Drupal.geolocation.markers[i].setMap(null); } Drupal.geolocation.markers[i] = new google.maps.Marker({ map: Drupal.geolocation.maps[i], draggable: Drupal.settings.geolocation.settings.marker_draggable ? true : false, // I dont like this much, rather have no effect // Will leave it to see if someone notice and shouts at me! // If so, will see consider enabling it again // animation: google.maps.Animation.DROP, position: latLng }); google.maps.event.addListener(Drupal.geolocation.markers[i], 'dragend', function(me) { Drupal.geolocation.codeLatLng(me.latLng, i, 'marker'); Drupal.geolocation.setMapMarker(me.latLng, i); }); return false; // if called from <a>-Tag } /** * Get the current user location if one is given * @return * Formatted location */ Drupal.geolocation.getFormattedLocation = function() { if (google.loader.ClientLocation.address.country_code == "US" && google.loader.ClientLocation.address.region) { return google.loader.ClientLocation.address.city + ", " + google.loader.ClientLocation.address.region.toUpperCase(); } else { return google.loader.ClientLocation.address.city + ", " + google.loader.ClientLocation.address.country_code; } } /** * Clear/Remove the values and the marker * * @param i * the index from the maps array we are working on */ Drupal.geolocation.clearLocation = function(i) { $('#geolocation-lat-' + i + ' input').attr('value', ''); $('#geolocation-lat-item-' + i + ' .geolocation-lat-item-value').html(''); $('#geolocation-lng-' + i + ' input').attr('value', ''); $('#geolocation-lng-item-' + i + ' .geolocation-lat-item-value').html(''); $('#geolocation-address-' + i + ' input').attr('value', ''); Drupal.geolocation.markers[i].setMap(); } /** * Do something when no location can be found * * @param supportFlag * Whether the browser supports geolocation or not * @param i * the index from the maps array we are working on */ Drupal.geolocation.handleNoGeolocation = function(supportFlag, i) { var siberia = new google.maps.LatLng(60, 105); var newyork = new google.maps.LatLng(40.69847032728747, -73.9514422416687); if (supportFlag == true) { alert(Drupal.t("Geolocation service failed. We've placed you in NewYork.")); initialLocation = newyork; } else { alert(Drupal.t("Your browser doesn't support geolocation. We've placed you in Siberia.")); initialLocation = siberia; } Drupal.geolocation.maps[i].setCenter(initialLocation); Drupal.geolocation.setMapMarker(initialLocation, i); } Drupal.behaviors.geolocationGooglemaps = { attach: function(context, settings) { geocoder = new google.maps.Geocoder(); var lat; var lng; var latLng; var mapOptions; var browserSupportFlag = new Boolean(); var singleClick; // Work on each map $.each(Drupal.settings.geolocation.defaults, function(i, mapDefaults) { // Only make this once ;) $("#geolocation-map-" + i).once('geolocation-googlemaps', function(){ // Attach listeners $('#geolocation-address-' + i + ' input').keypress(function(ev){ if(ev.which == 13){ ev.preventDefault(); Drupal.geolocation.codeAddress(i); } }); $('#geolocation-address-geocode-' + i).click(function(e) { Drupal.geolocation.codeAddress(i); }); $('#geolocation-remove-' + i).click(function(e) { Drupal.geolocation.clearLocation(i);<|fim▁hole|> }); // START: Autodetect clientlocation. // First use browser geolocation if (navigator.geolocation) { browserSupportFlag = true; $('#geolocation-help-' + i + ':not(.geolocation-googlemaps-processed)').addClass('geolocation-googlemaps-processed').append(Drupal.t(', or use your browser geolocation system by clicking this link') +': <span id="geolocation-client-location-' + i + '" class="geolocation-client-location">' + Drupal.t('My Location') + '</span>'); // Set current user location, if available $('#geolocation-client-location-' + i + ':not(.geolocation-googlemaps-processed)').addClass('geolocation-googlemaps-processed').click(function() { navigator.geolocation.getCurrentPosition(function(position) { latLng = new google.maps.LatLng(position.coords.latitude, position.coords.longitude); Drupal.geolocation.maps[i].setCenter(latLng); Drupal.geolocation.setMapMarker(latLng, i); Drupal.geolocation.codeLatLng(latLng, i, 'geocoder'); }, function() { Drupal.geolocation.handleNoGeolocation(browserSupportFlag, i); }); }); } // If browser geolication is not supoprted, try ip location else if (google.loader.ClientLocation) { latLng = new google.maps.LatLng(google.loader.ClientLocation.latitude, google.loader.ClientLocation.longitude); $('#geolocation-help-' + i + ':not(.geolocation-googlemaps-processed)').addClass('geolocation-googlemaps-processed').append(Drupal.t(', or use the IP-based location by clicking this link') +': <span id="geolocation-client-location-' + i + '" class="geolocation-client-location">' + Drupal.geolocation.getFormattedLocation() + '</span>'); // Set current user location, if available $('#geolocation-client-location-' + i + ':not(.geolocation-googlemaps-processed)').addClass('geolocation-googlemaps-processed').click(function() { latLng = new google.maps.LatLng(google.loader.ClientLocation.latitude, google.loader.ClientLocation.longitude); Drupal.geolocation.maps[i].setCenter(latLng); Drupal.geolocation.setMapMarker(latLng, i); Drupal.geolocation.codeLatLng(latLng, i, 'geocoder'); }); } // END: Autodetect clientlocation. // Get current/default values // Get default values // This might not be necesarry // It can always come from e lat = $('#geolocation-lat-' + i + ' input').attr('value') == false ? mapDefaults.lat : $('#geolocation-lat-' + i + ' input').attr('value'); lng = $('#geolocation-lng-' + i + ' input').attr('value') == false ? mapDefaults.lng : $('#geolocation-lng-' + i + ' input').attr('value'); latLng = new google.maps.LatLng(lat, lng); // Set map options mapOptions = { zoom: 2, center: latLng, mapTypeId: google.maps.MapTypeId.ROADMAP, scrollwheel: (Drupal.settings.geolocation.settings.scrollwheel != undefined) ? Drupal.settings.geolocation.settings.scrollwheel : false } // Create map Drupal.geolocation.maps[i] = new google.maps.Map(document.getElementById("geolocation-map-" + i), mapOptions); if (lat && lng) { // Set initial marker Drupal.geolocation.codeLatLng(latLng, i, 'geocoder'); Drupal.geolocation.setMapMarker(latLng, i); } // Listener to set marker google.maps.event.addListener(Drupal.geolocation.maps[i], 'click', function(me) { // Set a timeOut so that it doesn't execute if dbclick is detected singleClick = setTimeout(function() { Drupal.geolocation.codeLatLng(me.latLng, i, 'marker'); Drupal.geolocation.setMapMarker(me.latLng, i); }, 500); }); // Detect double click to avoid setting marker google.maps.event.addListener(Drupal.geolocation.maps[i], 'dblclick', function(me) { clearTimeout(singleClick); }); }) }); } }; })(jQuery); ;<|fim▁end|>
<|file_name|>inherit_stock_location.py<|end_file_name|><|fim▁begin|># -*- coding: utf-8 -*- ############################################################################## # # Copyright (C) 2010-2012 Associazione OpenERP Italia # (<http://www.openerp-italia.org>). # Copyright (C) 2014 Didotech srl # (<http://www.didotech.com>). # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as published # by the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. #<|fim▁hole|># but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import logging from datetime import datetime from openerp import SUPERUSER_ID from openerp.osv import orm, fields from openerp.tools import DEFAULT_SERVER_DATETIME_FORMAT _logger = logging.getLogger(__name__) _logger.setLevel(logging.DEBUG) class stock_location(orm.Model): _inherit = "stock.location" _columns = { 'update_product_bylocation': fields.boolean('Show Product location quantity on db', help='If check create a columns on product_product table for get product for this location'), 'product_related_columns': fields.char('Columns Name on product_product') } def update_product_by_location(self, cr, uid, context=None): context = context or self.pool['res.users'].context_get(cr, uid) location_ids = self.search(cr, uid, [('update_product_bylocation', '=', True)], context=context) location_vals = {} start_time = datetime.now() date_product_by_location_update = start_time.strftime(DEFAULT_SERVER_DATETIME_FORMAT) if location_ids: product_obj = self.pool['product.product'] for location in self.browse(cr, uid, location_ids, context): location_vals[location.id] = location.product_related_columns product_ids = product_obj.search(cr, uid, [('type', '!=', 'service')], context=context) product_context = context.copy() product_vals = {} for product_id in product_ids: product_vals[product_id] = {} for location_keys in location_vals.keys(): product_context['location'] = location_keys for product in product_obj.browse(cr, uid, product_ids, product_context): if location_vals[location_keys] and (product[location_vals[location_keys]] != product.qty_available): product_vals[product.id][location_vals[location_keys]] = product.qty_available if product_vals: for product_id in product_vals.keys(): product_val = product_vals[product_id] if product_val: product_val['date_product_by_location_update'] = date_product_by_location_update product_obj.write(cr, uid, product_id, product_val, context) end_time = datetime.now() duration_seconds = (end_time - start_time) duration = '{sec}'.format(sec=duration_seconds) _logger.info(u'update_product_by_location get in {duration}'.format(duration=duration)) return True def create_product_by_location(self, cr, location_name, context): model_id = self.pool['ir.model.data'].get_object_reference(cr, SUPERUSER_ID, 'product', 'model_product_product')[1] fields_value = { 'field_description': location_name, 'groups': [[6, False, []]], 'model_id': model_id, 'name': 'x_{location_name}'.format(location_name=location_name).lower().replace(' ', '_'), 'readonly': False, 'required': False, 'select_level': '0', 'serialization_field_id': False, 'translate': False, 'ttype': 'float', } context_field = context.copy() context_field.update( { 'department_id': False, 'lang': 'it_IT', 'manual': True, # required for create columns on table 'uid': 1 } ) fields_id = self.pool['ir.model.fields'].create(cr, SUPERUSER_ID, fields_value, context_field) return fields_id, fields_value['name'] def write(self, cr, uid, ids, vals, context=None): context = context or self.pool['res.users'].context_get(cr, uid) if vals.get('update_product_bylocation', False): for location in self.browse(cr, uid, ids, context): field_id, field_name = self.create_product_by_location(cr, location.name, context) vals['product_related_columns'] = field_name return super(stock_location, self).write(cr, uid, ids, vals, context)<|fim▁end|>
# This program is distributed in the hope that it will be useful,
<|file_name|>urls.py<|end_file_name|><|fim▁begin|>from django.conf.urls import url, include from django.contrib import admin from django.conf import settings from django.conf.urls.static import static from api.views import router urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'^', include(router.urls)), url(r'^ui/', include('ui.urls', namespace='build_ui')), url(r'^api-auth/', include('rest_framework.urls', namespace='rest_framework')), url(r'^explorer/', include('rest_framework_swagger.urls', namespace='swagger')), ] # Setting up static files for development: if settings.DEBUG is True: urlpatterns = urlpatterns + \<|fim▁hole|><|fim▁end|>
static(settings.STATIC_URL, document_root=settings.STATIC_ROOT)
<|file_name|>twitter.py<|end_file_name|><|fim▁begin|><|fim▁hole|>import bitlyapi import urllib2 import json import re from hasjob import app @job('hasjob') def tweet(title, url, location=None, parsed_location=None, username=None): auth = OAuthHandler(app.config['TWITTER_CONSUMER_KEY'], app.config['TWITTER_CONSUMER_SECRET']) auth.set_access_token(app.config['TWITTER_ACCESS_KEY'], app.config['TWITTER_ACCESS_SECRET']) api = API(auth) urllength = 23 # Current Twitter standard for HTTPS (as of Oct 2014) maxlength = 140 - urllength - 1 # == 116 if username: maxlength -= len(username) + 2 locationtag = u'' if parsed_location: locationtags = [] for token in parsed_location.get('tokens', []): if 'geoname' in token and 'token' in token: locname = token['token'].strip() if locname: locationtags.append(u'#' + locname.title().replace(u' ', '')) locationtag = u' '.join(locationtags) if locationtag: maxlength -= len(locationtag) + 1 if not locationtag and location: # Make a hashtag from the first word in the location. This catches # locations like 'Anywhere' which have no geonameid but are still valid locationtag = u'#' + re.split('\W+', location)[0] maxlength -= len(locationtag) + 1 if len(title) > maxlength: text = title[:maxlength - 1] + u'…' else: text = title[:maxlength] text = text + ' ' + url # Don't shorten URLs, now that there's t.co if locationtag: text = text + ' ' + locationtag if username: text = text + ' @' + username api.update_status(text) def shorten(url): if app.config['BITLY_KEY']: b = bitlyapi.BitLy(app.config['BITLY_USER'], app.config['BITLY_KEY']) res = b.shorten(longUrl=url) return res['url'] else: req = urllib2.Request("https://www.googleapis.com/urlshortener/v1/url", headers={"Content-Type": "application/json"}, data=json.dumps({'longUrl': url})) request_result = urllib2.urlopen(req) result = request_result.read() result_json = json.loads(result) return result_json['id']<|fim▁end|>
# -*- coding: utf-8 -*- from flask.ext.rq import job from tweepy import OAuthHandler, API
<|file_name|>string_converter_python3.py<|end_file_name|><|fim▁begin|>from xcrawler.compatibility.string_converter.compatible_string_converter import CompatibleStringConverter <|fim▁hole|> class StringConverterPython3(CompatibleStringConverter): """A Python 3 compatible class for converting a string to a specified type. """ def convert_to_string(self, string): string = self.try_convert_to_unicode_string(string) return string def list_convert_to_string(self, list_strings): return [self.try_convert_to_unicode_string(s) for s in list_strings]<|fim▁end|>
<|file_name|>svnrepo.py<|end_file_name|><|fim▁begin|># # # Copyright 2009-2016 Ghent University # # This file is part of EasyBuild, # originally created by the HPC team of Ghent University (http://ugent.be/hpc/en), # with support of Ghent University (http://ugent.be/hpc), # the Flemish Supercomputer Centre (VSC) (https://www.vscentrum.be), # Flemish Research Foundation (FWO) (http://www.fwo.be/en) # and the Department of Economy, Science and Innovation (EWI) (http://www.ewi-vlaanderen.be/en). # # http://github.com/hpcugent/easybuild # # EasyBuild is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation v2. # # EasyBuild is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with EasyBuild. If not, see <http://www.gnu.org/licenses/>. # # """ Repository tools Svn repository :author: Stijn De Weirdt (Ghent University) :author: Dries Verdegem (Ghent University) :author: Kenneth Hoste (Ghent University) :author: Pieter De Baets (Ghent University) :author: Jens Timmerman (Ghent University) :author: Toon Willems (Ghent University) :author: Ward Poelmans (Ghent University) :author: Fotis Georgatos (Uni.Lu, NTUA) """ import getpass import os import socket import tempfile import time from vsc.utils import fancylogger from easybuild.tools.build_log import EasyBuildError from easybuild.tools.filetools import rmtree2 from easybuild.tools.repository.filerepo import FileRepository from easybuild.tools.utilities import only_if_module_is_available _log = fancylogger.getLogger('svnrepo', fname=False) # optional Python packages, these might be missing # failing imports are just ignored # PySVN try: import pysvn # @UnusedImport from pysvn import ClientError # IGNORE:E0611 pysvn fails to recognize ClientError is available HAVE_PYSVN = True except ImportError: _log.debug("Failed to import pysvn module") HAVE_PYSVN = False class SvnRepository(FileRepository): """ Class for svn repositories """ DESCRIPTION = ("An SVN repository. The 1st argument contains the " "subversion repository location, this can be a directory or an URL. " "The 2nd argument is a path inside the repository where to save the files.") USABLE = HAVE_PYSVN @only_if_module_is_available('pysvn', url='http://pysvn.tigris.org/') def __init__(self, *args): """ Set self.client to None. Real logic is in setup_repo and create_working_copy """ self.client = None FileRepository.__init__(self, *args) def setup_repo(self): """ Set up SVN repository. """ self.repo = os.path.join(self.repo, self.subdir) # try to connect to the repository self.log.debug("Try to connect to repository %s" % self.repo) try: self.client = pysvn.Client() self.client.exception_style = 0 except ClientError:<|fim▁hole|> if not self.client.is_url(self.repo): raise EasyBuildError("Provided repository %s is not a valid svn url", self.repo) except ClientError: raise EasyBuildError("Can't connect to svn repository %s", self.repo) def create_working_copy(self): """ Create SVN working copy. """ self.wc = tempfile.mkdtemp(prefix='svn-wc-') # check if tmppath exists # this will trigger an error if it does not exist try: self.client.info2(self.repo, recurse=False) except ClientError: raise EasyBuildError("Getting info from %s failed.", self.wc) try: res = self.client.update(self.wc) self.log.debug("Updated to revision %s in %s" % (res, self.wc)) except ClientError: raise EasyBuildError("Update in wc %s went wrong", self.wc) if len(res) == 0: raise EasyBuildError("Update returned empy list (working copy: %s)", self.wc) if res[0].number == -1: # revision number of update is -1 # means nothing has been checked out try: res = self.client.checkout(self.repo, self.wc) self.log.debug("Checked out revision %s in %s" % (res.number, self.wc)) except ClientError, err: raise EasyBuildError("Checkout of path / in working copy %s went wrong: %s", self.wc, err) def add_easyconfig(self, cfg, name, version, stats, append): """ Add easyconfig to SVN repository. """ dest = FileRepository.add_easyconfig(self, cfg, name, version, stats, append) self.log.debug("destination = %s" % dest) if dest: self.log.debug("destination status: %s" % self.client.status(dest)) if self.client and not self.client.status(dest)[0].is_versioned: # add it to version control self.log.debug("Going to add %s (working copy: %s, cwd %s)" % (dest, self.wc, os.getcwd())) self.client.add(dest) def commit(self, msg=None): """ Commit working copy to SVN repository """ tup = (socket.gethostname(), time.strftime("%Y-%m-%d_%H-%M-%S"), getpass.getuser(), msg) completemsg = "EasyBuild-commit from %s (time: %s, user: %s) \n%s" % tup try: self.client.checkin(self.wc, completemsg, recurse=True) except ClientError, err: raise EasyBuildError("Commit from working copy %s (msg: %s) failed: %s", self.wc, msg, err) def cleanup(self): """ Clean up SVN working copy. """ try: rmtree2(self.wc) except OSError, err: raise EasyBuildError("Can't remove working copy %s: %s", self.wc, err)<|fim▁end|>
raise EasyBuildError("Svn Client initialization failed.") try:
<|file_name|>15.4.4.17-7-c-iii-4.js<|end_file_name|><|fim▁begin|>// Copyright (c) 2012 Ecma International. All rights reserved. // Ecma International makes this code available under the terms and conditions set<|fim▁hole|>/*--- es5id: 15.4.4.17-7-c-iii-4 description: > Array.prototype.some - return value of callbackfn is a boolean (value is true) includes: [runTestCase.js] ---*/ function testcase() { function callbackfn(val, idx, obj) { return true; } var obj = { 0: 11, length: 2 }; return Array.prototype.some.call(obj, callbackfn); } runTestCase(testcase);<|fim▁end|>
// forth on http://hg.ecmascript.org/tests/test262/raw-file/tip/LICENSE (the // "Use Terms"). Any redistribution of this code must retain the above // copyright and this notice and otherwise comply with the Use Terms.
<|file_name|>test_moveby.py<|end_file_name|><|fim▁begin|># This code is so you can run the samples without installing the package import sys import os sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..')) # testinfo = "s, t 3, s, t 6.1, s, q" tags = "MoveBy" import cocos from cocos.director import director from cocos.actions import MoveBy from cocos.sprite import Sprite import pyglet class TestLayer(cocos.layer.Layer):<|fim▁hole|> x,y = director.get_window_size() self.sprite = Sprite( 'grossini.png', (x/2, y/2) ) self.add( self.sprite, name='sprite' ) self.sprite.do( MoveBy( (x/2,y/2), 6 ) ) def main(): director.init() test_layer = TestLayer () main_scene = cocos.scene.Scene() main_scene.add(test_layer, name='test_layer') director.run (main_scene) if __name__ == '__main__': main()<|fim▁end|>
def __init__(self): super( TestLayer, self ).__init__()
<|file_name|>util.py<|end_file_name|><|fim▁begin|>import tornado.web import json from tornado_cors import CorsMixin from common import ParameterFormat, EnumEncoder class DefaultRequestHandler(CorsMixin, tornado.web.RequestHandler): CORS_ORIGIN = '*' def initialize(self): self.default_format = self.get_argument("format", "json", True) self.show_about = self.get_argument("show_about", True, True) self.pg_version = self.get_argument("pg_version", 9.6, True) self.version = "2.0 beta" def write_about_stuff(self, format_type="alter_system"): default_comment = "--" if format_type == "conf": default_comment = "#" self.write("{} Generated by PGConfig {}\n".format(default_comment, self.version)) self.write("{} http://pgconfig.org\n\n".format(default_comment * 2)) def write_comment(self, format_type, comment): default_comment = "--" if format_type == "conf": default_comment = "#" if comment != "NONE": self.write("\n{} {}\n".format(default_comment, comment)) def write_config(self, output_data): if self.show_about is True: self.write_about_stuff("conf") for category in output_data: self.write("# {}\n".format(category["description"])) for parameter in category["parameters"]: config_value = parameter.get("config_value", "NI") value_format = parameter.get("format", ParameterFormat.NONE) if value_format in (ParameterFormat.String, ParameterFormat.Time): config_value = "'{}'".format(config_value) parameter_comment = parameter.get("comment", "NONE") if parameter_comment != "NONE": self.write_comment("conf", parameter_comment) self.write("{} = {}\n".format(parameter["name"], config_value)) self.write("\n") def write_alter_system(self, output_data): if float(self.pg_version) <= 9.3: self.write("-- ALTER SYSTEM format it's only supported on version 9.4 and higher. Use 'conf' format instead.") else: if self.show_about is True: self.write_about_stuff() for category in output_data: self.write("-- {}\n".format(category["description"])) for parameter in category["parameters"]: config_value = parameter.get("config_value", "NI") parameter_comment = parameter.get("comment", "NONE") self.write_comment("alter_system", parameter_comment) self.write("ALTER SYSTEM SET {} TO '{}';\n".format(parameter[ "name"], config_value)) self.write("\n") def write_plain(self, message=list()): if len(message) == 1: self.write(message[0]) else: for line in message: self.write(line + '\n') def write_bash(self, message=list()): bash_script = """ #!/bin/bash """ self.write(bash_script) if len(message) == 1: self.write('SQL_QUERY="{}"\n'.format(message[0])) self.write('psql -c "${SQL_QUERY}"\n') else: for line in message: self.write('SQL_QUERY="{}"\n'.format(line)) self.write('psql -c "${SQL_QUERY}"\n\n') def write_json_api(self, message): self.set_header('Content-Type', 'application/vnd.api+json') _document = {} _document["data"] = message _meta = {} _meta["copyright"] = "PGConfig API" _meta["version"] = self.version _meta["arguments"] = self.request.arguments _document["meta"] = _meta _document["jsonapi"] = {"version": "1.0"} full_url = self.request.protocol + "://" + self.request.host + self.request.uri _document["links"] = {"self": full_url} self.write( json.dumps( _document, sort_keys=True, separators=(',', ': '), cls=EnumEncoder)) def write_json(self, message=list()): self.set_header('Content-Type', 'application/json') if len(message) == 1: self.write("{ \"output\": \"" + message[0] + "\"}") else: new_output = "{ \"output\": [" first_line = True for line in message: if not first_line: new_output += "," else: first_line = False new_output += "\"{}\"".format(line) new_output += "] } "<|fim▁hole|> def return_output(self, message=list()): # default_format=self.get_argument("format", "json", True) # converting string input into a list (for solve issue with multiline strings) process_data = [] if not isinstance(message, list): process_data.insert(0, message) else: process_data = message if self.default_format == "json": self.write_json_api(message) elif self.default_format == "bash": self.write_bash(message) elif self.default_format == "conf": self.write_config(message) elif self.default_format == "alter_system": self.write_alter_system(message) else: self.write_plain(message) class GeneratorRequestHandler(DefaultRequestHandler): pass<|fim▁end|>
self.write(new_output)
<|file_name|>dom-debug.js<|end_file_name|><|fim▁begin|>/* Copyright (c) 2008, Yahoo! Inc. All rights reserved. Code licensed under the BSD License: http://developer.yahoo.net/yui/license.txt version: 2.5.2 */ /** * The dom module provides helper methods for manipulating Dom elements. * @module dom *<|fim▁hole|> var Y = YAHOO.util, // internal shorthand getStyle, // for load time browser branching setStyle, // ditto propertyCache = {}, // for faster hyphen converts reClassNameCache = {}, // cache regexes for className document = window.document; // cache for faster lookups YAHOO.env._id_counter = YAHOO.env._id_counter || 0; // for use with generateId (global to save state if Dom is overwritten) // brower detection var isOpera = YAHOO.env.ua.opera, isSafari = YAHOO.env.ua.webkit, isGecko = YAHOO.env.ua.gecko, isIE = YAHOO.env.ua.ie; // regex cache var patterns = { HYPHEN: /(-[a-z])/i, // to normalize get/setStyle ROOT_TAG: /^body|html$/i, // body for quirks mode, html for standards, OP_SCROLL:/^(?:inline|table-row)$/i }; var toCamel = function(property) { if ( !patterns.HYPHEN.test(property) ) { return property; // no hyphens } if (propertyCache[property]) { // already converted return propertyCache[property]; } var converted = property; while( patterns.HYPHEN.exec(converted) ) { converted = converted.replace(RegExp.$1, RegExp.$1.substr(1).toUpperCase()); } propertyCache[property] = converted; return converted; //return property.replace(/-([a-z])/gi, function(m0, m1) {return m1.toUpperCase()}) // cant use function as 2nd arg yet due to safari bug }; var getClassRegEx = function(className) { var re = reClassNameCache[className]; if (!re) { re = new RegExp('(?:^|\\s+)' + className + '(?:\\s+|$)'); reClassNameCache[className] = re; } return re; }; // branching at load instead of runtime if (document.defaultView && document.defaultView.getComputedStyle) { // W3C DOM method getStyle = function(el, property) { var value = null; if (property == 'float') { // fix reserved word property = 'cssFloat'; } var computed = el.ownerDocument.defaultView.getComputedStyle(el, ''); if (computed) { // test computed before touching for safari value = computed[toCamel(property)]; } return el.style[property] || value; }; } else if (document.documentElement.currentStyle && isIE) { // IE method getStyle = function(el, property) { switch( toCamel(property) ) { case 'opacity' :// IE opacity uses filter var val = 100; try { // will error if no DXImageTransform val = el.filters['DXImageTransform.Microsoft.Alpha'].opacity; } catch(e) { try { // make sure its in the document val = el.filters('alpha').opacity; } catch(e) { YAHOO.log('getStyle: IE filter failed', 'error', 'Dom'); } } return val / 100; case 'float': // fix reserved word property = 'styleFloat'; // fall through default: // test currentStyle before touching var value = el.currentStyle ? el.currentStyle[property] : null; return ( el.style[property] || value ); } }; } else { // default to inline only getStyle = function(el, property) { return el.style[property]; }; } if (isIE) { setStyle = function(el, property, val) { switch (property) { case 'opacity': if ( YAHOO.lang.isString(el.style.filter) ) { // in case not appended el.style.filter = 'alpha(opacity=' + val * 100 + ')'; if (!el.currentStyle || !el.currentStyle.hasLayout) { el.style.zoom = 1; // when no layout or cant tell } } break; case 'float': property = 'styleFloat'; default: el.style[property] = val; } }; } else { setStyle = function(el, property, val) { if (property == 'float') { property = 'cssFloat'; } el.style[property] = val; }; } var testElement = function(node, method) { return node && node.nodeType == 1 && ( !method || method(node) ); }; /** * Provides helper methods for DOM elements. * @namespace YAHOO.util * @class Dom */ YAHOO.util.Dom = { /** * Returns an HTMLElement reference. * @method get * @param {String | HTMLElement |Array} el Accepts a string to use as an ID for getting a DOM reference, an actual DOM reference, or an Array of IDs and/or HTMLElements. * @return {HTMLElement | Array} A DOM reference to an HTML element or an array of HTMLElements. */ get: function(el) { if (el && (el.nodeType || el.item)) { // Node, or NodeList return el; } if (YAHOO.lang.isString(el) || !el) { // id or null return document.getElementById(el); } if (el.length !== undefined) { // array-like var c = []; for (var i = 0, len = el.length; i < len; ++i) { c[c.length] = Y.Dom.get(el[i]); } return c; } return el; // some other object, just pass it back }, /** * Normalizes currentStyle and ComputedStyle. * @method getStyle * @param {String | HTMLElement |Array} el Accepts a string to use as an ID, an actual DOM reference, or an Array of IDs and/or HTMLElements. * @param {String} property The style property whose value is returned. * @return {String | Array} The current value of the style property for the element(s). */ getStyle: function(el, property) { property = toCamel(property); var f = function(element) { return getStyle(element, property); }; return Y.Dom.batch(el, f, Y.Dom, true); }, /** * Wrapper for setting style properties of HTMLElements. Normalizes "opacity" across modern browsers. * @method setStyle * @param {String | HTMLElement | Array} el Accepts a string to use as an ID, an actual DOM reference, or an Array of IDs and/or HTMLElements. * @param {String} property The style property to be set. * @param {String} val The value to apply to the given property. */ setStyle: function(el, property, val) { property = toCamel(property); var f = function(element) { setStyle(element, property, val); YAHOO.log('setStyle setting ' + property + ' to ' + val, 'info', 'Dom'); }; Y.Dom.batch(el, f, Y.Dom, true); }, /** * Gets the current position of an element based on page coordinates. Element must be part of the DOM tree to have page coordinates (display:none or elements not appended return false). * @method getXY * @param {String | HTMLElement | Array} el Accepts a string to use as an ID, an actual DOM reference, or an Array of IDs and/or HTMLElements * @return {Array} The XY position of the element(s) */ getXY: function(el) { var f = function(el) { // has to be part of document to have pageXY if ( (el.parentNode === null || el.offsetParent === null || this.getStyle(el, 'display') == 'none') && el != el.ownerDocument.body) { YAHOO.log('getXY failed: element not available', 'error', 'Dom'); return false; } YAHOO.log('getXY returning ' + getXY(el), 'info', 'Dom'); return getXY(el); }; return Y.Dom.batch(el, f, Y.Dom, true); }, /** * Gets the current X position of an element based on page coordinates. The element must be part of the DOM tree to have page coordinates (display:none or elements not appended return false). * @method getX * @param {String | HTMLElement | Array} el Accepts a string to use as an ID, an actual DOM reference, or an Array of IDs and/or HTMLElements * @return {Number | Array} The X position of the element(s) */ getX: function(el) { var f = function(el) { return Y.Dom.getXY(el)[0]; }; return Y.Dom.batch(el, f, Y.Dom, true); }, /** * Gets the current Y position of an element based on page coordinates. Element must be part of the DOM tree to have page coordinates (display:none or elements not appended return false). * @method getY * @param {String | HTMLElement | Array} el Accepts a string to use as an ID, an actual DOM reference, or an Array of IDs and/or HTMLElements * @return {Number | Array} The Y position of the element(s) */ getY: function(el) { var f = function(el) { return Y.Dom.getXY(el)[1]; }; return Y.Dom.batch(el, f, Y.Dom, true); }, /** * Set the position of an html element in page coordinates, regardless of how the element is positioned. * The element(s) must be part of the DOM tree to have page coordinates (display:none or elements not appended return false). * @method setXY * @param {String | HTMLElement | Array} el Accepts a string to use as an ID, an actual DOM reference, or an Array of IDs and/or HTMLElements * @param {Array} pos Contains X & Y values for new position (coordinates are page-based) * @param {Boolean} noRetry By default we try and set the position a second time if the first fails */ setXY: function(el, pos, noRetry) { var f = function(el) { var style_pos = this.getStyle(el, 'position'); if (style_pos == 'static') { // default to relative this.setStyle(el, 'position', 'relative'); style_pos = 'relative'; } var pageXY = this.getXY(el); if (pageXY === false) { // has to be part of doc to have pageXY YAHOO.log('setXY failed: element not available', 'error', 'Dom'); return false; } var delta = [ // assuming pixels; if not we will have to retry parseInt( this.getStyle(el, 'left'), 10 ), parseInt( this.getStyle(el, 'top'), 10 ) ]; if ( isNaN(delta[0]) ) {// in case of 'auto' delta[0] = (style_pos == 'relative') ? 0 : el.offsetLeft; } if ( isNaN(delta[1]) ) { // in case of 'auto' delta[1] = (style_pos == 'relative') ? 0 : el.offsetTop; } if (pos[0] !== null) { el.style.left = pos[0] - pageXY[0] + delta[0] + 'px'; } if (pos[1] !== null) { el.style.top = pos[1] - pageXY[1] + delta[1] + 'px'; } if (!noRetry) { var newXY = this.getXY(el); // if retry is true, try one more time if we miss if ( (pos[0] !== null && newXY[0] != pos[0]) || (pos[1] !== null && newXY[1] != pos[1]) ) { this.setXY(el, pos, true); } } YAHOO.log('setXY setting position to ' + pos, 'info', 'Dom'); }; Y.Dom.batch(el, f, Y.Dom, true); }, /** * Set the X position of an html element in page coordinates, regardless of how the element is positioned. * The element must be part of the DOM tree to have page coordinates (display:none or elements not appended return false). * @method setX * @param {String | HTMLElement | Array} el Accepts a string to use as an ID, an actual DOM reference, or an Array of IDs and/or HTMLElements. * @param {Int} x The value to use as the X coordinate for the element(s). */ setX: function(el, x) { Y.Dom.setXY(el, [x, null]); }, /** * Set the Y position of an html element in page coordinates, regardless of how the element is positioned. * The element must be part of the DOM tree to have page coordinates (display:none or elements not appended return false). * @method setY * @param {String | HTMLElement | Array} el Accepts a string to use as an ID, an actual DOM reference, or an Array of IDs and/or HTMLElements. * @param {Int} x To use as the Y coordinate for the element(s). */ setY: function(el, y) { Y.Dom.setXY(el, [null, y]); }, /** * Returns the region position of the given element. * The element must be part of the DOM tree to have a region (display:none or elements not appended return false). * @method getRegion * @param {String | HTMLElement | Array} el Accepts a string to use as an ID, an actual DOM reference, or an Array of IDs and/or HTMLElements. * @return {Region | Array} A Region or array of Region instances containing "top, left, bottom, right" member data. */ getRegion: function(el) { var f = function(el) { if ( (el.parentNode === null || el.offsetParent === null || this.getStyle(el, 'display') == 'none') && el != el.ownerDocument.body) { YAHOO.log('getRegion failed: element not available', 'error', 'Dom'); return false; } var region = Y.Region.getRegion(el); YAHOO.log('getRegion returning ' + region, 'info', 'Dom'); return region; }; return Y.Dom.batch(el, f, Y.Dom, true); }, /** * Returns the width of the client (viewport). * @method getClientWidth * @deprecated Now using getViewportWidth. This interface left intact for back compat. * @return {Int} The width of the viewable area of the page. */ getClientWidth: function() { return Y.Dom.getViewportWidth(); }, /** * Returns the height of the client (viewport). * @method getClientHeight * @deprecated Now using getViewportHeight. This interface left intact for back compat. * @return {Int} The height of the viewable area of the page. */ getClientHeight: function() { return Y.Dom.getViewportHeight(); }, /** * Returns a array of HTMLElements with the given class. * For optimized performance, include a tag and/or root node when possible. * @method getElementsByClassName * @param {String} className The class name to match against * @param {String} tag (optional) The tag name of the elements being collected * @param {String | HTMLElement} root (optional) The HTMLElement or an ID to use as the starting point * @param {Function} apply (optional) A function to apply to each element when found * @return {Array} An array of elements that have the given class name */ getElementsByClassName: function(className, tag, root, apply) { tag = tag || '*'; root = (root) ? Y.Dom.get(root) : null || document; if (!root) { return []; } var nodes = [], elements = root.getElementsByTagName(tag), re = getClassRegEx(className); for (var i = 0, len = elements.length; i < len; ++i) { if ( re.test(elements[i].className) ) { nodes[nodes.length] = elements[i]; if (apply) { apply.call(elements[i], elements[i]); } } } return nodes; }, /** * Determines whether an HTMLElement has the given className. * @method hasClass * @param {String | HTMLElement | Array} el The element or collection to test * @param {String} className the class name to search for * @return {Boolean | Array} A boolean value or array of boolean values */ hasClass: function(el, className) { var re = getClassRegEx(className); var f = function(el) { YAHOO.log('hasClass returning ' + re.test(el.className), 'info', 'Dom'); return re.test(el.className); }; return Y.Dom.batch(el, f, Y.Dom, true); }, /** * Adds a class name to a given element or collection of elements. * @method addClass * @param {String | HTMLElement | Array} el The element or collection to add the class to * @param {String} className the class name to add to the class attribute * @return {Boolean | Array} A pass/fail boolean or array of booleans */ addClass: function(el, className) { var f = function(el) { if (this.hasClass(el, className)) { return false; // already present } YAHOO.log('addClass adding ' + className, 'info', 'Dom'); el.className = YAHOO.lang.trim([el.className, className].join(' ')); return true; }; return Y.Dom.batch(el, f, Y.Dom, true); }, /** * Removes a class name from a given element or collection of elements. * @method removeClass * @param {String | HTMLElement | Array} el The element or collection to remove the class from * @param {String} className the class name to remove from the class attribute * @return {Boolean | Array} A pass/fail boolean or array of booleans */ removeClass: function(el, className) { var re = getClassRegEx(className); var f = function(el) { if (!className || !this.hasClass(el, className)) { return false; // not present } YAHOO.log('removeClass removing ' + className, 'info', 'Dom'); var c = el.className; el.className = c.replace(re, ' '); if ( this.hasClass(el, className) ) { // in case of multiple adjacent this.removeClass(el, className); } el.className = YAHOO.lang.trim(el.className); // remove any trailing spaces return true; }; return Y.Dom.batch(el, f, Y.Dom, true); }, /** * Replace a class with another class for a given element or collection of elements. * If no oldClassName is present, the newClassName is simply added. * @method replaceClass * @param {String | HTMLElement | Array} el The element or collection to remove the class from * @param {String} oldClassName the class name to be replaced * @param {String} newClassName the class name that will be replacing the old class name * @return {Boolean | Array} A pass/fail boolean or array of booleans */ replaceClass: function(el, oldClassName, newClassName) { if (!newClassName || oldClassName === newClassName) { // avoid infinite loop return false; } var re = getClassRegEx(oldClassName); var f = function(el) { YAHOO.log('replaceClass replacing ' + oldClassName + ' with ' + newClassName, 'info', 'Dom'); if ( !this.hasClass(el, oldClassName) ) { this.addClass(el, newClassName); // just add it if nothing to replace return true; // NOTE: return } el.className = el.className.replace(re, ' ' + newClassName + ' '); if ( this.hasClass(el, oldClassName) ) { // in case of multiple adjacent this.replaceClass(el, oldClassName, newClassName); } el.className = YAHOO.lang.trim(el.className); // remove any trailing spaces return true; }; return Y.Dom.batch(el, f, Y.Dom, true); }, /** * Returns an ID and applies it to the element "el", if provided. * @method generateId * @param {String | HTMLElement | Array} el (optional) An optional element array of elements to add an ID to (no ID is added if one is already present). * @param {String} prefix (optional) an optional prefix to use (defaults to "yui-gen"). * @return {String | Array} The generated ID, or array of generated IDs (or original ID if already present on an element) */ generateId: function(el, prefix) { prefix = prefix || 'yui-gen'; var f = function(el) { if (el && el.id) { // do not override existing ID YAHOO.log('generateId returning existing id ' + el.id, 'info', 'Dom'); return el.id; } var id = prefix + YAHOO.env._id_counter++; YAHOO.log('generateId generating ' + id, 'info', 'Dom'); if (el) { el.id = id; } return id; }; // batch fails when no element, so just generate and return single ID return Y.Dom.batch(el, f, Y.Dom, true) || f.apply(Y.Dom, arguments); }, /** * Determines whether an HTMLElement is an ancestor of another HTML element in the DOM hierarchy. * @method isAncestor * @param {String | HTMLElement} haystack The possible ancestor * @param {String | HTMLElement} needle The possible descendent * @return {Boolean} Whether or not the haystack is an ancestor of needle */ isAncestor: function(haystack, needle) { haystack = Y.Dom.get(haystack); needle = Y.Dom.get(needle); if (!haystack || !needle) { return false; } if (haystack.contains && needle.nodeType && !isSafari) { // safari contains is broken YAHOO.log('isAncestor returning ' + haystack.contains(needle), 'info', 'Dom'); return haystack.contains(needle); } else if ( haystack.compareDocumentPosition && needle.nodeType ) { YAHOO.log('isAncestor returning ' + !!(haystack.compareDocumentPosition(needle) & 16), 'info', 'Dom'); return !!(haystack.compareDocumentPosition(needle) & 16); } else if (needle.nodeType) { // fallback to crawling up (safari) return !!this.getAncestorBy(needle, function(el) { return el == haystack; }); } YAHOO.log('isAncestor failed; most likely needle is not an HTMLElement', 'error', 'Dom'); return false; }, /** * Determines whether an HTMLElement is present in the current document. * @method inDocument * @param {String | HTMLElement} el The element to search for * @return {Boolean} Whether or not the element is present in the current document */ inDocument: function(el) { return this.isAncestor(document.documentElement, el); }, /** * Returns a array of HTMLElements that pass the test applied by supplied boolean method. * For optimized performance, include a tag and/or root node when possible. * @method getElementsBy * @param {Function} method - A boolean method for testing elements which receives the element as its only argument. * @param {String} tag (optional) The tag name of the elements being collected * @param {String | HTMLElement} root (optional) The HTMLElement or an ID to use as the starting point * @param {Function} apply (optional) A function to apply to each element when found * @return {Array} Array of HTMLElements */ getElementsBy: function(method, tag, root, apply) { tag = tag || '*'; root = (root) ? Y.Dom.get(root) : null || document; if (!root) { return []; } var nodes = [], elements = root.getElementsByTagName(tag); for (var i = 0, len = elements.length; i < len; ++i) { if ( method(elements[i]) ) { nodes[nodes.length] = elements[i]; if (apply) { apply(elements[i]); } } } YAHOO.log('getElementsBy returning ' + nodes, 'info', 'Dom'); return nodes; }, /** * Runs the supplied method against each item in the Collection/Array. * The method is called with the element(s) as the first arg, and the optional param as the second ( method(el, o) ). * @method batch * @param {String | HTMLElement | Array} el (optional) An element or array of elements to apply the method to * @param {Function} method The method to apply to the element(s) * @param {Any} o (optional) An optional arg that is passed to the supplied method * @param {Boolean} override (optional) Whether or not to override the scope of "method" with "o" * @return {Any | Array} The return value(s) from the supplied method */ batch: function(el, method, o, override) { el = (el && (el.tagName || el.item)) ? el : Y.Dom.get(el); // skip get() when possible if (!el || !method) { YAHOO.log('batch failed: invalid arguments', 'error', 'Dom'); return false; } var scope = (override) ? o : window; if (el.tagName || el.length === undefined) { // element or not array-like return method.call(scope, el, o); } var collection = []; for (var i = 0, len = el.length; i < len; ++i) { collection[collection.length] = method.call(scope, el[i], o); } return collection; }, /** * Returns the height of the document. * @method getDocumentHeight * @return {Int} The height of the actual document (which includes the body and its margin). */ getDocumentHeight: function() { var scrollHeight = (document.compatMode != 'CSS1Compat') ? document.body.scrollHeight : document.documentElement.scrollHeight; var h = Math.max(scrollHeight, Y.Dom.getViewportHeight()); YAHOO.log('getDocumentHeight returning ' + h, 'info', 'Dom'); return h; }, /** * Returns the width of the document. * @method getDocumentWidth * @return {Int} The width of the actual document (which includes the body and its margin). */ getDocumentWidth: function() { var scrollWidth = (document.compatMode != 'CSS1Compat') ? document.body.scrollWidth : document.documentElement.scrollWidth; var w = Math.max(scrollWidth, Y.Dom.getViewportWidth()); YAHOO.log('getDocumentWidth returning ' + w, 'info', 'Dom'); return w; }, /** * Returns the current height of the viewport. * @method getViewportHeight * @return {Int} The height of the viewable area of the page (excludes scrollbars). */ getViewportHeight: function() { var height = self.innerHeight; // Safari, Opera var mode = document.compatMode; if ( (mode || isIE) && !isOpera ) { // IE, Gecko height = (mode == 'CSS1Compat') ? document.documentElement.clientHeight : // Standards document.body.clientHeight; // Quirks } YAHOO.log('getViewportHeight returning ' + height, 'info', 'Dom'); return height; }, /** * Returns the current width of the viewport. * @method getViewportWidth * @return {Int} The width of the viewable area of the page (excludes scrollbars). */ getViewportWidth: function() { var width = self.innerWidth; // Safari var mode = document.compatMode; if (mode || isIE) { // IE, Gecko, Opera width = (mode == 'CSS1Compat') ? document.documentElement.clientWidth : // Standards document.body.clientWidth; // Quirks } YAHOO.log('getViewportWidth returning ' + width, 'info', 'Dom'); return width; }, /** * Returns the nearest ancestor that passes the test applied by supplied boolean method. * For performance reasons, IDs are not accepted and argument validation omitted. * @method getAncestorBy * @param {HTMLElement} node The HTMLElement to use as the starting point * @param {Function} method - A boolean method for testing elements which receives the element as its only argument. * @return {Object} HTMLElement or null if not found */ getAncestorBy: function(node, method) { while (node = node.parentNode) { // NOTE: assignment if ( testElement(node, method) ) { YAHOO.log('getAncestorBy returning ' + node, 'info', 'Dom'); return node; } } YAHOO.log('getAncestorBy returning null (no ancestor passed test)', 'error', 'Dom'); return null; }, /** * Returns the nearest ancestor with the given className. * @method getAncestorByClassName * @param {String | HTMLElement} node The HTMLElement or an ID to use as the starting point * @param {String} className * @return {Object} HTMLElement */ getAncestorByClassName: function(node, className) { node = Y.Dom.get(node); if (!node) { YAHOO.log('getAncestorByClassName failed: invalid node argument', 'error', 'Dom'); return null; } var method = function(el) { return Y.Dom.hasClass(el, className); }; return Y.Dom.getAncestorBy(node, method); }, /** * Returns the nearest ancestor with the given tagName. * @method getAncestorByTagName * @param {String | HTMLElement} node The HTMLElement or an ID to use as the starting point * @param {String} tagName * @return {Object} HTMLElement */ getAncestorByTagName: function(node, tagName) { node = Y.Dom.get(node); if (!node) { YAHOO.log('getAncestorByTagName failed: invalid node argument', 'error', 'Dom'); return null; } var method = function(el) { return el.tagName && el.tagName.toUpperCase() == tagName.toUpperCase(); }; return Y.Dom.getAncestorBy(node, method); }, /** * Returns the previous sibling that is an HTMLElement. * For performance reasons, IDs are not accepted and argument validation omitted. * Returns the nearest HTMLElement sibling if no method provided. * @method getPreviousSiblingBy * @param {HTMLElement} node The HTMLElement to use as the starting point * @param {Function} method A boolean function used to test siblings * that receives the sibling node being tested as its only argument * @return {Object} HTMLElement or null if not found */ getPreviousSiblingBy: function(node, method) { while (node) { node = node.previousSibling; if ( testElement(node, method) ) { return node; } } return null; }, /** * Returns the previous sibling that is an HTMLElement * @method getPreviousSibling * @param {String | HTMLElement} node The HTMLElement or an ID to use as the starting point * @return {Object} HTMLElement or null if not found */ getPreviousSibling: function(node) { node = Y.Dom.get(node); if (!node) { YAHOO.log('getPreviousSibling failed: invalid node argument', 'error', 'Dom'); return null; } return Y.Dom.getPreviousSiblingBy(node); }, /** * Returns the next HTMLElement sibling that passes the boolean method. * For performance reasons, IDs are not accepted and argument validation omitted. * Returns the nearest HTMLElement sibling if no method provided. * @method getNextSiblingBy * @param {HTMLElement} node The HTMLElement to use as the starting point * @param {Function} method A boolean function used to test siblings * that receives the sibling node being tested as its only argument * @return {Object} HTMLElement or null if not found */ getNextSiblingBy: function(node, method) { while (node) { node = node.nextSibling; if ( testElement(node, method) ) { return node; } } return null; }, /** * Returns the next sibling that is an HTMLElement * @method getNextSibling * @param {String | HTMLElement} node The HTMLElement or an ID to use as the starting point * @return {Object} HTMLElement or null if not found */ getNextSibling: function(node) { node = Y.Dom.get(node); if (!node) { YAHOO.log('getNextSibling failed: invalid node argument', 'error', 'Dom'); return null; } return Y.Dom.getNextSiblingBy(node); }, /** * Returns the first HTMLElement child that passes the test method. * @method getFirstChildBy * @param {HTMLElement} node The HTMLElement to use as the starting point * @param {Function} method A boolean function used to test children * that receives the node being tested as its only argument * @return {Object} HTMLElement or null if not found */ getFirstChildBy: function(node, method) { var child = ( testElement(node.firstChild, method) ) ? node.firstChild : null; return child || Y.Dom.getNextSiblingBy(node.firstChild, method); }, /** * Returns the first HTMLElement child. * @method getFirstChild * @param {String | HTMLElement} node The HTMLElement or an ID to use as the starting point * @return {Object} HTMLElement or null if not found */ getFirstChild: function(node, method) { node = Y.Dom.get(node); if (!node) { YAHOO.log('getFirstChild failed: invalid node argument', 'error', 'Dom'); return null; } return Y.Dom.getFirstChildBy(node); }, /** * Returns the last HTMLElement child that passes the test method. * @method getLastChildBy * @param {HTMLElement} node The HTMLElement to use as the starting point * @param {Function} method A boolean function used to test children * that receives the node being tested as its only argument * @return {Object} HTMLElement or null if not found */ getLastChildBy: function(node, method) { if (!node) { YAHOO.log('getLastChild failed: invalid node argument', 'error', 'Dom'); return null; } var child = ( testElement(node.lastChild, method) ) ? node.lastChild : null; return child || Y.Dom.getPreviousSiblingBy(node.lastChild, method); }, /** * Returns the last HTMLElement child. * @method getLastChild * @param {String | HTMLElement} node The HTMLElement or an ID to use as the starting point * @return {Object} HTMLElement or null if not found */ getLastChild: function(node) { node = Y.Dom.get(node); return Y.Dom.getLastChildBy(node); }, /** * Returns an array of HTMLElement childNodes that pass the test method. * @method getChildrenBy * @param {HTMLElement} node The HTMLElement to start from * @param {Function} method A boolean function used to test children * that receives the node being tested as its only argument * @return {Array} A static array of HTMLElements */ getChildrenBy: function(node, method) { var child = Y.Dom.getFirstChildBy(node, method); var children = child ? [child] : []; Y.Dom.getNextSiblingBy(child, function(node) { if ( !method || method(node) ) { children[children.length] = node; } return false; // fail test to collect all children }); return children; }, /** * Returns an array of HTMLElement childNodes. * @method getChildren * @param {String | HTMLElement} node The HTMLElement or an ID to use as the starting point * @return {Array} A static array of HTMLElements */ getChildren: function(node) { node = Y.Dom.get(node); if (!node) { YAHOO.log('getChildren failed: invalid node argument', 'error', 'Dom'); } return Y.Dom.getChildrenBy(node); }, /** * Returns the left scroll value of the document * @method getDocumentScrollLeft * @param {HTMLDocument} document (optional) The document to get the scroll value of * @return {Int} The amount that the document is scrolled to the left */ getDocumentScrollLeft: function(doc) { doc = doc || document; return Math.max(doc.documentElement.scrollLeft, doc.body.scrollLeft); }, /** * Returns the top scroll value of the document * @method getDocumentScrollTop * @param {HTMLDocument} document (optional) The document to get the scroll value of * @return {Int} The amount that the document is scrolled to the top */ getDocumentScrollTop: function(doc) { doc = doc || document; return Math.max(doc.documentElement.scrollTop, doc.body.scrollTop); }, /** * Inserts the new node as the previous sibling of the reference node * @method insertBefore * @param {String | HTMLElement} newNode The node to be inserted * @param {String | HTMLElement} referenceNode The node to insert the new node before * @return {HTMLElement} The node that was inserted (or null if insert fails) */ insertBefore: function(newNode, referenceNode) { newNode = Y.Dom.get(newNode); referenceNode = Y.Dom.get(referenceNode); if (!newNode || !referenceNode || !referenceNode.parentNode) { YAHOO.log('insertAfter failed: missing or invalid arg(s)', 'error', 'Dom'); return null; } return referenceNode.parentNode.insertBefore(newNode, referenceNode); }, /** * Inserts the new node as the next sibling of the reference node * @method insertAfter * @param {String | HTMLElement} newNode The node to be inserted * @param {String | HTMLElement} referenceNode The node to insert the new node after * @return {HTMLElement} The node that was inserted (or null if insert fails) */ insertAfter: function(newNode, referenceNode) { newNode = Y.Dom.get(newNode); referenceNode = Y.Dom.get(referenceNode); if (!newNode || !referenceNode || !referenceNode.parentNode) { YAHOO.log('insertAfter failed: missing or invalid arg(s)', 'error', 'Dom'); return null; } if (referenceNode.nextSibling) { return referenceNode.parentNode.insertBefore(newNode, referenceNode.nextSibling); } else { return referenceNode.parentNode.appendChild(newNode); } }, /** * Creates a Region based on the viewport relative to the document. * @method getClientRegion * @return {Region} A Region object representing the viewport which accounts for document scroll */ getClientRegion: function() { var t = Y.Dom.getDocumentScrollTop(), l = Y.Dom.getDocumentScrollLeft(), r = Y.Dom.getViewportWidth() + l, b = Y.Dom.getViewportHeight() + t; return new Y.Region(t, r, b, l); } }; var getXY = function() { if (document.documentElement.getBoundingClientRect) { // IE return function(el) { var box = el.getBoundingClientRect(); var rootNode = el.ownerDocument; return [box.left + Y.Dom.getDocumentScrollLeft(rootNode), box.top + Y.Dom.getDocumentScrollTop(rootNode)]; }; } else { return function(el) { // manually calculate by crawling up offsetParents var pos = [el.offsetLeft, el.offsetTop]; var parentNode = el.offsetParent; // safari: subtract body offsets if el is abs (or any offsetParent), unless body is offsetParent var accountForBody = (isSafari && Y.Dom.getStyle(el, 'position') == 'absolute' && el.offsetParent == el.ownerDocument.body); if (parentNode != el) { while (parentNode) { pos[0] += parentNode.offsetLeft; pos[1] += parentNode.offsetTop; if (!accountForBody && isSafari && Y.Dom.getStyle(parentNode,'position') == 'absolute' ) { accountForBody = true; } parentNode = parentNode.offsetParent; } } if (accountForBody) { //safari doubles in this case pos[0] -= el.ownerDocument.body.offsetLeft; pos[1] -= el.ownerDocument.body.offsetTop; } parentNode = el.parentNode; // account for any scrolled ancestors while ( parentNode.tagName && !patterns.ROOT_TAG.test(parentNode.tagName) ) { if (parentNode.scrollTop || parentNode.scrollLeft) { // work around opera inline/table scrollLeft/Top bug (false reports offset as scroll) if (!patterns.OP_SCROLL.test(Y.Dom.getStyle(parentNode, 'display'))) { if (!isOpera || Y.Dom.getStyle(parentNode, 'overflow') !== 'visible') { // opera inline-block misreports when visible pos[0] -= parentNode.scrollLeft; pos[1] -= parentNode.scrollTop; } } } parentNode = parentNode.parentNode; } return pos; }; } }() // NOTE: Executing for loadtime branching })(); /** * A region is a representation of an object on a grid. It is defined * by the top, right, bottom, left extents, so is rectangular by default. If * other shapes are required, this class could be extended to support it. * @namespace YAHOO.util * @class Region * @param {Int} t the top extent * @param {Int} r the right extent * @param {Int} b the bottom extent * @param {Int} l the left extent * @constructor */ YAHOO.util.Region = function(t, r, b, l) { /** * The region's top extent * @property top * @type Int */ this.top = t; /** * The region's top extent as index, for symmetry with set/getXY * @property 1 * @type Int */ this[1] = t; /** * The region's right extent * @property right * @type int */ this.right = r; /** * The region's bottom extent * @property bottom * @type Int */ this.bottom = b; /** * The region's left extent * @property left * @type Int */ this.left = l; /** * The region's left extent as index, for symmetry with set/getXY * @property 0 * @type Int */ this[0] = l; }; /** * Returns true if this region contains the region passed in * @method contains * @param {Region} region The region to evaluate * @return {Boolean} True if the region is contained with this region, * else false */ YAHOO.util.Region.prototype.contains = function(region) { return ( region.left >= this.left && region.right <= this.right && region.top >= this.top && region.bottom <= this.bottom ); // this.logger.debug("does " + this + " contain " + region + " ... " + ret); }; /** * Returns the area of the region * @method getArea * @return {Int} the region's area */ YAHOO.util.Region.prototype.getArea = function() { return ( (this.bottom - this.top) * (this.right - this.left) ); }; /** * Returns the region where the passed in region overlaps with this one * @method intersect * @param {Region} region The region that intersects * @return {Region} The overlap region, or null if there is no overlap */ YAHOO.util.Region.prototype.intersect = function(region) { var t = Math.max( this.top, region.top ); var r = Math.min( this.right, region.right ); var b = Math.min( this.bottom, region.bottom ); var l = Math.max( this.left, region.left ); if (b >= t && r >= l) { return new YAHOO.util.Region(t, r, b, l); } else { return null; } }; /** * Returns the region representing the smallest region that can contain both * the passed in region and this region. * @method union * @param {Region} region The region that to create the union with * @return {Region} The union region */ YAHOO.util.Region.prototype.union = function(region) { var t = Math.min( this.top, region.top ); var r = Math.max( this.right, region.right ); var b = Math.max( this.bottom, region.bottom ); var l = Math.min( this.left, region.left ); return new YAHOO.util.Region(t, r, b, l); }; /** * toString * @method toString * @return string the region properties */ YAHOO.util.Region.prototype.toString = function() { return ( "Region {" + "top: " + this.top + ", right: " + this.right + ", bottom: " + this.bottom + ", left: " + this.left + "}" ); }; /** * Returns a region that is occupied by the DOM element * @method getRegion * @param {HTMLElement} el The element * @return {Region} The region that the element occupies * @static */ YAHOO.util.Region.getRegion = function(el) { var p = YAHOO.util.Dom.getXY(el); var t = p[1]; var r = p[0] + el.offsetWidth; var b = p[1] + el.offsetHeight; var l = p[0]; return new YAHOO.util.Region(t, r, b, l); }; ///////////////////////////////////////////////////////////////////////////// /** * A point is a region that is special in that it represents a single point on * the grid. * @namespace YAHOO.util * @class Point * @param {Int} x The X position of the point * @param {Int} y The Y position of the point * @constructor * @extends YAHOO.util.Region */ YAHOO.util.Point = function(x, y) { if (YAHOO.lang.isArray(x)) { // accept input from Dom.getXY, Event.getXY, etc. y = x[1]; // dont blow away x yet x = x[0]; } /** * The X position of the point, which is also the right, left and index zero (for Dom.getXY symmetry) * @property x * @type Int */ this.x = this.right = this.left = this[0] = x; /** * The Y position of the point, which is also the top, bottom and index one (for Dom.getXY symmetry) * @property y * @type Int */ this.y = this.top = this.bottom = this[1] = y; }; YAHOO.util.Point.prototype = new YAHOO.util.Region(); YAHOO.register("dom", YAHOO.util.Dom, {version: "2.5.2", build: "1076"});<|fim▁end|>
*/ (function() {
<|file_name|>apps.py<|end_file_name|><|fim▁begin|>from django.apps import AppConfig class JcvrbaseappConfig(AppConfig):<|fim▁hole|><|fim▁end|>
name = 'jcvrbaseapp'
<|file_name|>derive_input_object.rs<|end_file_name|><|fim▁begin|>use fnv::FnvHashMap; use juniper::{ marker, DefaultScalarValue, FromInputValue, GraphQLInputObject, GraphQLType, GraphQLValue, InputValue, Registry, ToInputValue, };<|fim▁hole|>#[graphql( name = "MyInput", description = "input descr", scalar = DefaultScalarValue )] struct Input { regular_field: String, #[graphql(name = "haha", default = "33", description = "haha descr")] c: i32, #[graphql(default)] other: Option<bool>, } #[derive(GraphQLInputObject, Debug, PartialEq)] #[graphql(rename = "none")] struct NoRenameInput { regular_field: String, } /// Object comment. #[derive(GraphQLInputObject, Debug, PartialEq)] struct DocComment { /// Field comment. regular_field: bool, } /// Doc 1.\ /// Doc 2. /// /// Doc 4. #[derive(GraphQLInputObject, Debug, PartialEq)] struct MultiDocComment { /// Field 1. /// Field 2. regular_field: bool, } /// This is not used as the description. #[derive(GraphQLInputObject, Debug, PartialEq)] #[graphql(description = "obj override")] struct OverrideDocComment { /// This is not used as the description. #[graphql(description = "field override")] regular_field: bool, } #[derive(Debug, PartialEq)] struct Fake; impl<'a> marker::IsInputType<DefaultScalarValue> for &'a Fake {} impl<'a> FromInputValue for &'a Fake { fn from_input_value(_v: &InputValue) -> Option<&'a Fake> { None } } impl<'a> ToInputValue for &'a Fake { fn to_input_value(&self) -> InputValue { InputValue::scalar("this is fake") } } impl<'a> GraphQLType<DefaultScalarValue> for &'a Fake { fn name(_: &()) -> Option<&'static str> { None } fn meta<'r>(_: &(), registry: &mut Registry<'r>) -> juniper::meta::MetaType<'r> where DefaultScalarValue: 'r, { let meta = registry.build_enum_type::<&'a Fake>( &(), &[juniper::meta::EnumValue { name: "fake".to_string(), description: None, deprecation_status: juniper::meta::DeprecationStatus::Current, }], ); meta.into_meta() } } impl<'a> GraphQLValue<DefaultScalarValue> for &'a Fake { type Context = (); type TypeInfo = (); fn type_name<'i>(&self, info: &'i Self::TypeInfo) -> Option<&'i str> { <Self as GraphQLType>::name(info) } } #[derive(GraphQLInputObject, Debug, PartialEq)] #[graphql(scalar = DefaultScalarValue)] struct WithLifetime<'a> { regular_field: &'a Fake, } #[test] fn test_derived_input_object() { assert_eq!( <Input as GraphQLType<DefaultScalarValue>>::name(&()), Some("MyInput") ); // Validate meta info. let mut registry: Registry = Registry::new(FnvHashMap::default()); let meta = Input::meta(&(), &mut registry); assert_eq!(meta.name(), Some("MyInput")); assert_eq!(meta.description(), Some(&"input descr".to_string())); // Test default value injection. let input_no_defaults: InputValue = ::serde_json::from_value(serde_json::json!({ "regularField": "a", })) .unwrap(); let output_no_defaults: Input = FromInputValue::from_input_value(&input_no_defaults).unwrap(); assert_eq!( output_no_defaults, Input { regular_field: "a".into(), c: 33, other: None, } ); // Test with all values supplied. let input: InputValue = ::serde_json::from_value(serde_json::json!({ "regularField": "a", "haha": 55, "other": true, })) .unwrap(); let output: Input = FromInputValue::from_input_value(&input).unwrap(); assert_eq!( output, Input { regular_field: "a".into(), c: 55, other: Some(true), } ); // Test disable renaming let input: InputValue = ::serde_json::from_value(serde_json::json!({ "regular_field": "hello", })) .unwrap(); let output: NoRenameInput = FromInputValue::from_input_value(&input).unwrap(); assert_eq!( output, NoRenameInput { regular_field: "hello".into(), } ); } #[test] fn test_doc_comment() { let mut registry: Registry = Registry::new(FnvHashMap::default()); let meta = DocComment::meta(&(), &mut registry); assert_eq!(meta.description(), Some(&"Object comment.".to_string())); } #[test] fn test_multi_doc_comment() { let mut registry: Registry = Registry::new(FnvHashMap::default()); let meta = MultiDocComment::meta(&(), &mut registry); assert_eq!( meta.description(), Some(&"Doc 1. Doc 2.\n\nDoc 4.".to_string()) ); } #[test] fn test_doc_comment_override() { let mut registry: Registry = Registry::new(FnvHashMap::default()); let meta = OverrideDocComment::meta(&(), &mut registry); assert_eq!(meta.description(), Some(&"obj override".to_string())); }<|fim▁end|>
#[derive(GraphQLInputObject, Debug, PartialEq)]
<|file_name|>fusiontables_v1_messages.py<|end_file_name|><|fim▁begin|># # Copyright 2015 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Generated message classes for fusiontables version v1. API for working with Fusion Tables data. """ # NOTE: This file is autogenerated and should not be edited by hand. from apitools.base.protorpclite import messages as _messages package = 'fusiontables' class Column(_messages.Message): """Specifies the id, name and type of a column in a table. Messages: BaseColumnValue: Optional identifier of the base column. If present, this column is derived from the specified base column. Fields: baseColumn: Optional identifier of the base column. If present, this<|fim▁hole|> column is derived from the specified base column. columnId: Identifier for the column. description: Optional column description. graph_predicate: Optional column predicate. Used to map table to graph data model (subject,predicate,object) See http://www.w3.org/TR/2014/REC- rdf11-concepts-20140225/#data-model kind: Type name: a template for an individual column. name: Required name of the column. type: Required type of the column. """ class BaseColumnValue(_messages.Message): """Optional identifier of the base column. If present, this column is derived from the specified base column. Fields: columnId: The id of the column in the base table from which this column is derived. tableIndex: Offset to the entry in the list of base tables in the table definition. """ columnId = _messages.IntegerField(1, variant=_messages.Variant.INT32) tableIndex = _messages.IntegerField(2, variant=_messages.Variant.INT32) baseColumn = _messages.MessageField('BaseColumnValue', 1) columnId = _messages.IntegerField(2, variant=_messages.Variant.INT32) description = _messages.StringField(3) graph_predicate = _messages.StringField(4) kind = _messages.StringField(5, default=u'fusiontables#column') name = _messages.StringField(6) type = _messages.StringField(7) class ColumnList(_messages.Message): """Represents a list of columns in a table. Fields: items: List of all requested columns. kind: Type name: a list of all columns. nextPageToken: Token used to access the next page of this result. No token is displayed if there are no more pages left. totalItems: Total number of columns for the table. """ items = _messages.MessageField('Column', 1, repeated=True) kind = _messages.StringField(2, default=u'fusiontables#columnList') nextPageToken = _messages.StringField(3) totalItems = _messages.IntegerField(4, variant=_messages.Variant.INT32) class FusiontablesColumnListRequest(_messages.Message): """A FusiontablesColumnListRequest object. Fields: maxResults: Maximum number of columns to return. Optional. Default is 5. pageToken: Continuation token specifying which result page to return. Optional. tableId: Table whose columns are being listed. """ maxResults = _messages.IntegerField(1, variant=_messages.Variant.UINT32) pageToken = _messages.StringField(2) tableId = _messages.StringField(3, required=True) class FusiontablesColumnListAlternateRequest(_messages.Message): """A FusiontablesColumnListRequest object. Fields: pageSize: Maximum number of columns to return. Optional. Default is 5. pageToken: Continuation token specifying which result page to return. Optional. tableId: Table whose columns are being listed. """ pageSize = _messages.IntegerField(1, variant=_messages.Variant.UINT32) pageToken = _messages.StringField(2) tableId = _messages.StringField(3, required=True) class ColumnListAlternate(_messages.Message): """Represents a list of columns in a table. Fields: items: List of all requested columns. kind: Type name: a list of all columns. nextPageToken: Token used to access the next page of this result. No token is displayed if there are no more pages left. totalItems: Total number of columns for the table. """ columns = _messages.MessageField('Column', 1, repeated=True) kind = _messages.StringField(2, default=u'fusiontables#columnList') nextPageToken = _messages.StringField(3) totalItems = _messages.IntegerField(4, variant=_messages.Variant.INT32)<|fim▁end|>
<|file_name|>2-17-a.cpp<|end_file_name|><|fim▁begin|>//<|fim▁hole|>using std::vector; #include <iostream> #include <cassert> int min3(int a, int b, int c); int minSubSum(const vector<int>& a); int minSubRec(const vector<int>& a, int left, int right); int main() { vector<int> v{-2, -1, -2, 3, -4}; assert(minSubSum(v) == -6); } int minSubSum(const vector<int>& a) { return minSubRec(a, 0, a.size() - 1); } int minSubRec(const vector<int>& a, int left, int right) { if (left == right) if (a[left] < 0) return a[left]; else return 0; int mid = (left + right) / 2; int minLeftSum = minSubRec(a, left, mid); int minRightSum = minSubRec(a, mid + 1, right); int minLeftBorderSum = 0, leftBorderSum = 0; for (int i = mid; i >= left; --i) { leftBorderSum += a[i]; if (leftBorderSum < minLeftBorderSum) minLeftBorderSum = leftBorderSum; } int minRightBorderSum = 0, rightBorderSum = 0; for (int i = mid + 1; i <= right; ++i) { rightBorderSum += a[i]; if (rightBorderSum < minRightBorderSum) minRightBorderSum = rightBorderSum; } return min3(minLeftSum, minRightSum, minLeftBorderSum + minRightBorderSum); } int min3(int a, int b, int c) { int min2 = a < b ? a : b; return min2 < c ? min2 : c; }<|fim▁end|>
// Created by paysonl on 16-10-20. // #include <vector>
<|file_name|>setup.py<|end_file_name|><|fim▁begin|># Copyright 2016 Adler Brediks Medrado # # Licensed under the Apache License, Version 2.0 (the "License");<|fim▁hole|># You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from setuptools import setup, find_packages with open("requirements.txt") as reqs: install_requires = reqs.readlines() setup( name="abbr", version="0.0.1", url="https://github.com/adlermedrado/abbr", author="Adler Brediks Medrado", author_email="[email protected]", license="Apache-2.0", description="A client library to abbreviate string contents", long_description=open('README.rst').read(), packages=find_packages(), install_requires=install_requires, include_package_data=True, classifiers=[ 'Development Status :: 2 - Pre-Alpha', 'Programming Language :: Python', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.5', ], )<|fim▁end|>
# you may not use this file except in compliance with the License.
<|file_name|>OneDirectMappingProject.java<|end_file_name|><|fim▁begin|>/******************************************************************************* * Copyright (c) 1998, 2012 Oracle and/or its affiliates. All rights reserved. * This program and the accompanying materials are made available under the * terms of the Eclipse Public License v1.0 and Eclipse Distribution License v. 1.0 * which accompanies this distribution. * The Eclipse Public License is available at http://www.eclipse.org/legal/epl-v10.html * and the Eclipse Distribution License is available at * http://www.eclipse.org/org/documents/edl-v10.php. * * Contributors: * Oracle - initial API and implementation from Oracle TopLink <|fim▁hole|>import org.eclipse.persistence.sessions.Project; import org.eclipse.persistence.oxm.*; import org.eclipse.persistence.oxm.mappings.XMLDirectMapping; public class OneDirectMappingProject extends Project { public OneDirectMappingProject() { super(); addEmployeeDescriptor(); } public void addEmployeeDescriptor() { XMLDescriptor descriptor = new XMLDescriptor(); descriptor.setDefaultRootElement("employee"); descriptor.setJavaClass(Employee.class); XMLDirectMapping firstNameMapping = new XMLDirectMapping(); firstNameMapping.setAttributeName("firstName"); firstNameMapping.setXPath("first-name/text()"); firstNameMapping.readOnly(); descriptor.addMapping(firstNameMapping); this.addDescriptor(descriptor); } }<|fim▁end|>
******************************************************************************/ package org.eclipse.persistence.testing.oxm.readonly;
<|file_name|>module.py<|end_file_name|><|fim▁begin|># -*- coding: utf-8 -*- # Copyright(C) 2017 Vincent A # # This file is part of a weboob module. # # This weboob module is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This weboob module is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this weboob module. If not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from weboob.tools.backend import Module from weboob.capabilities.weather import CapWeather <|fim▁hole|> __all__ = ['LameteoagricoleModule'] class LameteoagricoleModule(Module, CapWeather): NAME = 'lameteoagricole' DESCRIPTION = u'lameteoagricole website' MAINTAINER = u'Vincent A' EMAIL = '[email protected]' LICENSE = 'AGPLv3+' VERSION = '2.1' BROWSER = LameteoagricoleBrowser def iter_city_search(self, pattern): return self.browser.iter_cities(pattern) def get_current(self, city_id): return self.browser.get_current(city_id) def iter_forecast(self, city_id): return self.browser.iter_forecast(city_id)<|fim▁end|>
from .browser import LameteoagricoleBrowser
<|file_name|>setup.py<|end_file_name|><|fim▁begin|>from setuptools import setup import os execfile(os.path.join('sheetsync','version.py')) with open('README.rst') as fh: long_description = fh.read() with open('requirements.txt') as fh: requirements = [line.strip() for line in fh.readlines()] setup( name='sheetsync', version=__version__, description="Synchronize rows of data with a google spreadsheet", long_description=long_description, author='Mark Brenig-Jones', author_email='[email protected]', url='https://github.com/mbrenig/SheetSync/', packages=['sheetsync'], platforms='any', install_requires=requirements, classifiers=[ "Development Status :: 4 - Beta",<|fim▁hole|><|fim▁end|>
"License :: OSI Approved :: MIT License", "Programming Language :: Python :: 2.7", ], )
<|file_name|>Components.js<|end_file_name|><|fim▁begin|>/** * Most of this code is adapated from https://github.com/qimingweng/react-modal-dialog */ import React from 'react'; import ReactDOM from 'react-dom'; import EventStack from './EventStack'; import PropTypes from 'prop-types'; const ESCAPE = 27; /** * Get the keycode from an event * @param {Event} event The browser event from which we want the keycode from * @returns {Number} The number of the keycode */ const getKeyCodeFromEvent = event => event.which || event.keyCode || event.charCode; // Render into subtree is necessary for parent contexts to transfer over // For example, for react-router const renderSubtreeIntoContainer = ReactDOM.unstable_renderSubtreeIntoContainer; export class ModalContent extends React.Component { constructor(props) { super(props); this.handleGlobalClick = this.handleGlobalClick.bind(this); this.handleGlobalKeydown = this.handleGlobalKeydown.bind(this); } componentWillMount() { /** * This is done in the componentWillMount instead of the componentDidMount * because this way, a modal that is a child of another will have register * for events after its parent */ this.eventToken = EventStack.addListenable([ ['click', this.handleGlobalClick], ['keydown', this.handleGlobalKeydown] ]); } componentWillUnmount() { EventStack.removeListenable(this.eventToken); } shouldClickDismiss(event) { const { target } = event; // This piece of code isolates targets which are fake clicked by things // like file-drop handlers if (target.tagName === 'INPUT' && target.type === 'file') { return false; } if (!this.props.dismissOnBackgroundClick) { if (target !== this.refs.self || this.refs.self.contains(target)) { return false; } } else { if (target === this.refs.self || this.refs.self.contains(target)) { return false; } } return true; } handleGlobalClick(event) { if (this.shouldClickDismiss(event)) { if (typeof this.props.onClose === 'function') { this.props.onClose(event); } } } handleGlobalKeydown(event) { if (getKeyCodeFromEvent(event) === ESCAPE) { if (typeof this.props.onClose === 'function') { this.props.onClose(event); } } } render() { return ( <div ref="self" className={'liq_modal-content'}> {this.props.showButton && this.props.onClose ? ( <button onClick={this.props.onClose} className={'liq_modal-close'} data-test="close_modal_button"> <span aria-hidden="true">{'×'}</span> </button> ) : null} {this.props.children} </div> ); } } ModalContent.propTypes = { showButton: PropTypes.bool, onClose: PropTypes.func, // required for the close button className: PropTypes.string, children: PropTypes.node, dismissOnBackgroundClick: PropTypes.bool }; ModalContent.defaultProps = { dismissOnBackgroundClick: true, showButton: true }; export class ModalPortal extends React.Component { componentDidMount() { // disable scrolling on body document.body.classList.add('liq_modal-open'); // Create a div and append it to the body this._target = document.body.appendChild(document.createElement('div')); // Mount a component on that div this._component = renderSubtreeIntoContainer( this, this.props.children, this._target ); // A handler call in case you want to do something when a modal opens, like add a class to the body or something if (typeof this.props.onModalDidMount === 'function') { this.props.onModalDidMount(); } } componentDidUpdate() { // When the child component updates, we have to make sure the content rendered to the DOM is updated to this._component = renderSubtreeIntoContainer( this, this.props.children, this._target ); } componentWillUnmount() { /** * Let this be some discussion about fading out the components on unmount. * Right now, there is the issue that if a stack of components are layered * on top of each other, and you programmatically dismiss the bottom one, * it actually takes some time for the animation to catch up to the top one, * because each modal doesn't send a dismiss signal to its children until * it itself is totally gone... */ // TODO: REMOVE THIS - THINK OUT OF THE BOX const done = () => { // Modal will unmount now // Call a handler, like onModalDidMount if (typeof this.props.onModalWillUnmount === 'function') { this.props.onModalWillUnmount(); } // Remove the node and clean up after the target ReactDOM.unmountComponentAtNode(this._target); document.body.removeChild(this._target); document.body.classList.remove('liq_modal-open'); }; // A similar API to react-transition-group if ( this._component && typeof this._component.componentWillLeave === 'function' ) { // Pass the callback to be called on completion this._component.componentWillLeave(done); } else { // Call completion immediately done(); } } render() { return null; } // This doesn't actually return anything to render } ModalPortal.propTypes = { onClose: PropTypes.func, // This is called when the dialog should close<|fim▁hole|> onModalDidMount: PropTypes.func, // optional, called on mount onModalWillUnmount: PropTypes.func // optional, called on unmount }; export class ModalBackground extends React.Component { constructor(props) { super(props); this.state = { // This is set to false as soon as the component has mounted // This allows the component to change its css and animate in transparent: true }; } componentDidMount() { // Create a delay so CSS will animate requestAnimationFrame(() => this.setState({ transparent: false })); } componentWillLeave(callback) { this.setState({ transparent: true, componentIsLeaving: true }); // There isn't a good way to figure out what the duration is exactly, // because parts of the animation are carried out in CSS... setTimeout(() => { callback(); }, this.props.duration); } render() { const { transparent } = this.state; const overlayStyle = { opacity: transparent ? 0 : 0.6 }; const containerStyle = { opacity: transparent ? 0 : 1 }; return ( <div className={'liq_modal-background'} style={{ zIndex: this.props.zIndex }}> <div style={overlayStyle} className={'liq_modal-background__overlay'} /> <div style={containerStyle} className={'liq_modal-background__container'}> {this.props.children} </div> </div> ); } } ModalBackground.defaultProps = { duration: 300, zIndex: 1100 // to lay above tooltips and the website header }; ModalBackground.propTypes = { onClose: PropTypes.func, duration: PropTypes.number.isRequired, zIndex: PropTypes.number.isRequired, children: PropTypes.node }; export default { ModalPortal, ModalBackground, ModalContent };<|fim▁end|>
children: PropTypes.node,
<|file_name|>TestAnalogInUsingJoistick.cpp<|end_file_name|><|fim▁begin|>/* * The basic test codes to check the AnglogIn. */ #include "mbed.h" static AnalogIn ain_x(A6); // connect joistics' x axis to A6 pin of mbed static AnalogIn ain_y(A5); // connect joistics' y axis to A5 pin of mbed static void printAnalogInput(AnalogIn *xin, AnalogIn *yin) { uint8_t point[2] = { 0 }; point[0] = (uint8_t)(xin->read()*100.0f + 0.5); point[1] = (uint8_t)(yin->read()*100.0f + 0.5); printf("x: %d, y: %d\n", point[0], point[1]);<|fim▁hole|>int main() { while(1) { printAnalogInput(&ain_x, &ain_y); wait(0.2f); } }<|fim▁end|>
}
<|file_name|>hattrie.go<|end_file_name|><|fim▁begin|>package safebrowsing /* #include <stdio.h> #include <stdlib.h> #include <string.h> #include "hat-trie.h" hattrie_t* start() { hattrie_t* trie; trie = hattrie_create(); return trie; } void set(hattrie_t* h, char* key, size_t len) { value_t* val; val = hattrie_get(h, key, len); *val = 1; } int get(hattrie_t* h, char* key, size_t len) { value_t* val; val = hattrie_tryget(h, key, len); if (val != 0) { return *val; } return 0; } void delete(hattrie_t* h, char* key, size_t len) { value_t* val; val = hattrie_tryget(h, key, len); if (val != 0) { *val = 0; } } char* hattrie_iter_key_string(hattrie_iter_t* i, size_t* len) { const char* in_key; char* out_key; in_key = hattrie_iter_key(i, len); out_key = malloc((*len) * sizeof(char)); memcpy(out_key, in_key, *len); return out_key; } */ import "C" import ( "runtime" "sync" "unsafe" ) type HatTrie struct { trie *C.hattrie_t l sync.RWMutex // we name it because we don't want to expose it } func finalizeHatTrie(c *HatTrie) { C.hattrie_free(c.trie) } func NewTrie() *HatTrie { trie := C.start() out := &HatTrie{ trie: trie, } runtime.SetFinalizer(out, finalizeHatTrie) return out } func (h *HatTrie) Delete(key string) { h.l.Lock() defer h.l.Unlock() ckey := C.CString(key) defer C.free(unsafe.Pointer(ckey)) C.delete(h.trie, ckey, C.size_t(len(key))) } func (h *HatTrie) Set(key string) { h.l.Lock() defer h.l.Unlock() ckey := C.CString(key) defer C.free(unsafe.Pointer(ckey)) C.set(h.trie, ckey, C.size_t(len(key))) } func (h *HatTrie) Get(key string) bool { h.l.RLock() defer h.l.RUnlock() <|fim▁hole|> return val == 1 } type HatTrieIterator struct { iterator *C.hattrie_iter_t } func finalizeHatTrieIterator(i *HatTrieIterator) { C.hattrie_iter_free(i.iterator) } func (h *HatTrie) Iterator() *HatTrieIterator { out := C.hattrie_iter_begin(h.trie, true) hi := &HatTrieIterator{ iterator: out, } runtime.SetFinalizer(hi, finalizeHatTrieIterator) return hi } func (i *HatTrieIterator) Next() string { if C.hattrie_iter_finished(i.iterator) { return "" } keylen := C.size_t(0) ckey := C.hattrie_iter_key_string(i.iterator, &keylen) defer C.free(unsafe.Pointer(ckey)) key := C.GoStringN(ckey, C.int(keylen)) C.hattrie_iter_next(i.iterator) return key }<|fim▁end|>
ckey := C.CString(key) defer C.free(unsafe.Pointer(ckey)) val := C.get(h.trie, ckey, C.size_t(len(key)))
<|file_name|>Triangle.java<|end_file_name|><|fim▁begin|>package dkalymbaev.triangle; class Triangle { /** * This class defines points a b and c as peacks of triangle. */ public Point a; public Point b; public Point c; /** * Creating of a new objects. * @param a is the length of the first side. * @param b is the length of the second side. * @param c is the length of the third side. */ public Triangle(Point a, Point b, Point c) { this.a = a; this.b = b; this.c = c; } public double area() { double ab = a.distanceTo(b);<|fim▁hole|> double bc = b.distanceTo(c); double ac = a.distanceTo(c); double halfperim = ((ab + bc + ac) / 2); double area = Math.sqrt(halfperim * (halfperim - ab) * (halfperim - bc) * (halfperim - ac)); if (ab > 0 && bc > 0 && ac > 0) { return area; } else { return 0; } } }<|fim▁end|>
<|file_name|>config.go<|end_file_name|><|fim▁begin|>package logger // FileConfig represents a file backend configuration. type FileConfig struct { // Logging output severity level. Messages with higher severity value will be discarded. Level string // File path of the logging output. If path is either empty or "-", logging will be output to os.Stderr. Path string } // SyslogConfig represents a syslog backend configuration. type SyslogConfig struct { // Logging output severity level. Messages with higher severity value will be discarded. Level string // syslog facility to send messages to. Facility string // syslog tag to specify in messages. Tag string // syslog service address and transport type (either "udp", "tcp" or "unix"). If not sepcified, local syslog will // be used.<|fim▁hole|>}<|fim▁end|>
Address string Transport string
<|file_name|>urls.py<|end_file_name|><|fim▁begin|>"""kaxabu URL Configuration The `urlpatterns` list routes URLs to views. For more information please see:<|fim▁hole|> 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: url(r'^$', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.conf.urls import url, include 2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls')) """ from django.conf.urls import url, include from django.contrib import admin urlpatterns = [ url(r'^', include('族語辭典.網址')), url(r'^admin/', admin.site.urls), ]<|fim▁end|>
ERROR: type should be large_string, got " https://docs.djangoproject.com/en/1.9/topics/http/urls/\nExamples:\nFunction views"
<|file_name|>lib.rs<|end_file_name|><|fim▁begin|>#[macro_use] #[cfg(test)] extern crate lazy_static; #[cfg(test)] extern crate itertools; extern crate r68k_common; #[cfg(test)] extern crate r68k_tools; pub mod cpu; pub mod ram; pub mod interrupts; pub mod musashi; #[cfg(test)] mod tests { use cpu::TestCore; use r68k_tools::memory::MemoryVec; use r68k_tools::PC; use r68k_tools::disassembler::disassemble; use r68k_tools::Exception; use cpu::ops::handlers::InstructionSetGenerator; use cpu::ops::handlers::OpcodeHandler; use r68k_tools::disassembler::Disassembler; #[test] // #[ignore] fn roundtrips() { let mut over = 0; let mut under = 0; let mut wrong = 0; let gen = InstructionSetGenerator::<TestCore>::new(); let optable: Vec<&str> = gen.generate_with("???", |ref op| op.name); let d = Disassembler::new(); for opcode in 0x0000..0xffff { let op = optable[opcode]; let parts:Vec<&str> = op.split('_').collect(); let mnemonic = parts[0]; let mut pc = PC(0); let extension_word_mask = 0b1111_1000_1111_1111; // bits 8-10 should always be zero in the ea extension word // as we don't know which word will be seen as the ea extension word // (as opposed to immediate operand values) just make sure these aren't set. let dasm_mem = &mut MemoryVec::new16(pc, vec![opcode as u16, 0x001f, 0x00a4, 0x1234 & extension_word_mask, 0x5678 & extension_word_mask]); // println!("PREDASM {:04x}", opcode); match d.disassemble(pc, dasm_mem) { Err(Exception::IllegalInstruction(_opcode, _)) => if op != "???" && op != "unimplemented_1111" && op != "unimplemented_1010" && op != "illegal" { under += 1; println!("{:04x}: {} disasm under", opcode, op); } , //println!("{:04x}:\t\tover", opcode), Ok((new_pc, dis_inst)) => if op == "???" || op == "unimplemented_1111" || op == "unimplemented_1010" || op == "illegal" { over += 1;<|fim▁hole|> println!("{:04x}: {} disasm over, {}", opcode, op, dis_inst); } else if dis_inst.mnemonic.to_lowercase() != mnemonic && mnemonic != "real" { // ILLEGAL == real_illegal wrong += 1; println!("{:04x}: {} disasm different {}", opcode, op, dis_inst); }, } }; println!("{} opcodes over, {} under, {} wrong", over, under, wrong); } }<|fim▁end|>
<|file_name|>nn_cuda.py<|end_file_name|><|fim▁begin|>from __future__ import print_function, division import os import sys import time import h5py import torch from torch import nn from torch.autograd import Variable import torch.nn.functional as F import torch.utils.data as Data import numpy as np import json from sklearn.metrics import mean_squared_error from obspy.signal.filter import envelope NPTS = 60 input_size = 3 * NPTS + 1 hidden_size = 90 num_layers = 10 LR = 0.001 weight_decay=0.005 nepochs = 20 DROPOUT=0.01 torch.manual_seed(1) # reproducible CUDA_FLAG = torch.cuda.is_available() def dump_json(data, fn): with open(fn, 'w') as fh: json.dump(data, fh, indent=2, sort_keys=True) class Net(nn.Module): def __init__(self, input_size, hidden_size, num_layers=1): super(Net, self).__init__() self.input_size = input_size self.hidden_size = hidden_size self.num_layers = num_layers self.cuda_flag = torch.cuda.is_available() self.hidden0 = nn.Linear(input_size, hidden_size) self.hidden1 = nn.Linear(hidden_size, hidden_size) self.hidden2 = nn.Linear(hidden_size, hidden_size) self.predict = nn.Linear(hidden_size, 1) def forward(self, x): # input layer x = F.relu(self.hidden0(x)) x = F.relu(self.hidden1(x)) x = F.relu(self.hidden2(x)) # predict layer x = self.predict(x) return x def construct_nn(input_size, hidden_size, num_layers): model = nn.Sequential() # input layer model.add_module("input", nn.Linear(input_size, hidden_size)) model.add_module("ReLU_input", nn.ReLU()) #model.add_module("Tanh", nn.Tanh()) # add hidden layer for idx in range(num_layers): model.add_module("hidden-%d" % idx, nn.Linear(hidden_size, hidden_size)) model.add_module("ReLU-%d" % idx, nn.ReLU()) #model.add_module("Tanh-%d" % idx, nn.Tanh()) model.add_module("output", nn.Linear(hidden_size, 1)) return model def load_data(npts=60): print("Loading waveform npts: %d" % npts) t1 = time.time() f = h5py.File("./data/input.h5") #waveforms = np.array(f["waveform"])[:2000, :, :] #magnitudes = np.array(f["magnitude"])[:2000] train_x = np.array(f["train_x"])[:, :, 0:npts] train_y = np.array(f["train_y"]) test_x = np.array(f["test_x"])[:, :, 0:npts] test_y = np.array(f["test_y"]) train_d = np.array(f["train_distance"]) test_d = np.array(f["test_distance"]) t2 = time.time() print("Time used in reading data: %.2f sec" % (t2 - t1)) print("train x and y shape: ", train_x.shape, train_y.shape) print("test x and y shape: ", test_x.shape, test_y.shape) print("train d and test d shape: ", train_d.shape, test_d.shape) return {"train_x": train_x, "train_y": train_y, "test_x": test_x, "test_y": test_y, "train_d": train_d, "test_d": test_d} def make_dataloader(xs, ys): xs = torch.Tensor(xs).cuda() ys = torch.Tensor(ys).cuda() torch_dataset = Data.TensorDataset(data_tensor=xs, target_tensor=ys) loader = Data.DataLoader(dataset=torch_dataset, batch_size=1, shuffle=True) return loader def predict_on_test(net, test_x): print("Predict...") pred_y = [] for idx in range(len(test_x)): x = test_x[idx, :] x = Variable(torch.unsqueeze(torch.Tensor(x), dim=0)).cuda() y_p = net(x) _y = float(y_p.cpu().data.numpy()[0]) # print("pred %d: %f | true y: %f" % (idx, _y, test_y[idx]))<|fim▁hole|> pred_y.append(_y) return pred_y def transfer_data_into_envelope(data): t1 = time.time() data_env = np.zeros(data.shape) for idx1 in range(data.shape[0]): for idx2 in range(data.shape[1]): data_env[idx1, idx2, :] = envelope(data[idx1, idx2, :]) t2 = time.time() print("Time used to convert envelope: %.2f sec" % (t2 - t1)) return data_env def transform_features(input_data, dtype="disp", dt=0.05, envelope_flag=False): """ Transform from displacement to a certrain data type, such as accelaration, velocity, or displacement itself. """ print("[Transform]Input data shape before transform: ", input_data.shape) t1 = time.time() if dtype == "disp": data = input_data elif dtype == "vel": vel = np.gradient(input_data, dt, axis=2) data = vel elif dtype == "acc": vel = np.gradient(input_data, dt, axis=2) acc = np.gradient(vel, dt, axis=2) data = acc elif dtype == "acc_cumul_log_sum": vel = np.gradient(input_data, dt, axis=2) acc = np.gradient(vel, dt, axis=2) data = np.log(np.cumsum(np.abs(acc) * dt, axis=2) + 1) else: raise ValueError("unkonw dtype: %s" % dtype) if envelope_flag: data = transfer_data_into_envelope(data) t2 = time.time() print("time used in transform: %.2f sec" % (t2 - t1)) return data def add_distance_to_features(x, d): nlen = x.shape[1] x_new = np.zeros([x.shape[0], x.shape[1]+1]) x_new[:, 0:nlen] = x[:, :] x_new[:, nlen] = np.log(d) print("[Add distance]shape change after adding distance as feature: ", x.shape, "-->", x_new.shape) return x_new def combine_components_waveform(x): time_step = x.shape[2] print("time step in waveform: %d" % time_step) x_new = np.zeros([x.shape[0], time_step*3]) for idx in range(len(x)): x_new[idx, 0:time_step] = x[idx, 0, :] x_new[idx, time_step:(2*time_step)] = x[idx, 1, :] x_new[idx, (2*time_step):(3*time_step)] = x[idx, 2, :] print("[Combine]shape change after combining components: ", x.shape, "-->", x_new.shape) return x_new def standarize_features(train_x, test_x): vmax = np.max(np.abs(train_x)) print("[Norm]max value of input waveform: %f" % vmax) return train_x / vmax, test_x / vmax def load_and_process_features(data_split, dtype, envelope_flag): train_x = transform_features(data_split["train_x"], dtype=dtype, envelope_flag=envelope_flag) test_x = transform_features(data_split["test_x"], dtype=dtype, envelope_flag=envelope_flag) train_x = combine_components_waveform(train_x) test_x = combine_components_waveform(test_x) train_x, test_x = standarize_features(train_x, test_x) train_x = add_distance_to_features(train_x, data_split["train_d"]) test_x = add_distance_to_features(test_x, data_split["test_d"]) return train_x, test_x def main(outputdir, dtype, npts=60, envelope_flag=False): print("Working on dtype(%s) --- outputdir(%s)" % (dtype, outputdir)) print("Envelope flag: %s" % envelope_flag) data_split = load_data(npts=npts) train_x, test_x = load_and_process_features( data_split, dtype, envelope_flag) train_loader = make_dataloader(train_x, data_split["train_y"]) #net = Net(input_size, hidden_size, num_layers) net = construct_nn(input_size, hidden_size, num_layers) net.cuda() print(net) optimizer = torch.optim.Adam(net.parameters(), lr=LR, weight_decay=0.0005) loss_func = nn.MSELoss() # train ntest = data_split["train_x"].shape[0] all_loss = {} for epoch in range(nepochs): loss_epoch = [] for step, (batch_x, batch_y) in enumerate(train_loader): if step % int((ntest/10) + 1) == 1: print('Epoch: ', epoch, '| Step: %d/%d' % (step, ntest), "| Loss: %f" % np.mean(loss_epoch)) if CUDA_FLAG: x = Variable(batch_x).cuda() y = Variable(torch.Tensor([batch_y.numpy(), ])).cuda() else: x = Variable(x) y = Variable(torch.Tensor([batch_y.numpy(), ])) prediction = net(x) loss = loss_func(prediction, y) optimizer.zero_grad() # clear gradients for this training step loss.backward() # backpropagation, compute gradients optimizer.step() loss_epoch.append(loss.data[0]) all_loss["epoch_%d" % epoch] = loss_epoch outputfn = os.path.join(outputdir, "loss.epoch_%d.json" % epoch) print("=== Mean loss in epoch(%d): %f(log: %s) ===" % (epoch, np.mean(loss_epoch), outputfn)) dump_json(loss_epoch, outputfn) # test pred_y = predict_on_test(net, test_x) test_y = data_split["test_y"] _mse = mean_squared_error(test_y, pred_y) _std = np.std(test_y - pred_y) print("MSE and error std: %f, %f" % (_mse, _std)) outputfn = os.path.join(outputdir, "prediction.json") print("output file: %s" % outputfn) data = {"test_y": list(test_y), "test_y_pred": list(pred_y), "epoch_loss": all_loss, "mse": _mse, "err_std": _std} dump_json(data, outputfn) if __name__ == "__main__": if len(sys.argv) != 2: print("input dtype please...exit") sys.exit() dtype = sys.argv[1] outputdir = "output.%s" % dtype if not os.path.exists(outputdir): os.makedirs(outputdir) envelope_flag=True if dtype == "acc_cumul_log_sum": envelope_flag=False main(outputdir, dtype, npts=NPTS, envelope_flag=envelope_flag)<|fim▁end|>
<|file_name|>range.cc<|end_file_name|><|fim▁begin|>// Copyright (C) 2014 Free Software Foundation, Inc. // // This file is part of the GNU ISO C++ Library. This library is free // software; you can redistribute it and/or modify it under the // terms of the GNU General Public License as published by the // Free Software Foundation; either version 3, or (at your option) // any later version. // This library is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // You should have received a copy of the GNU General Public License along // with this library; see the file COPYING3. If not see // <http://www.gnu.org/licenses/>. #include <regex> #include <testsuite_performance.h> using namespace __gnu_test;<|fim▁hole|> resource_counter resource; start_counters(time, resource); // this should get compiled to just L"[abcd]" auto re = std::wregex(L'[' + std::wstring(300, L'a') + L"bc" + std::wstring(1000, 'a') + L"d]"); bool ok = true; for (int i = 0; i < 100000; ++i) ok = ok && (std::regex_match(L"b", re) && std::regex_match(L"d", re)); stop_counters(time, resource); report_performance(__FILE__, "", time, resource); return ok ? 0 : 1; }<|fim▁end|>
int main() { time_counter time;
<|file_name|>fallback_chooser.py<|end_file_name|><|fim▁begin|>"""Pure TTY chooser UI""" from __future__ import print_function, absolute_import __author__ = "Stephan Sokolow (deitarion/SSokolow)" __license__ = "GNU GPL 2 or later" import os # Use readline if available but don't depend on it try: import readline # Shut PyFlakes up readline # pylint: disable=pointless-statement except ImportError: pass def parse_choice(in_str): """Parse a string containing one or more integers or Python ranges separated by commas. @returns: A list of integers @attention: Unlike Python, this treats ranges as inclusive of the upper bound. """<|fim▁hole|> for x in in_str.replace(',', ' ').split(): try: choices.append(int(x)) except ValueError: try: first, last = [int(y) for y in x.split(':', 1)] choices.extend(range(first, last + 1)) except ValueError: print("Not an integer or range: %s" % x) return choices # TODO: Document and, if necessary, refactor def choose(results, strip_path, enqueue): # Draw the menu for pos, val in enumerate(results): val = strip_path and os.path.basename(val) or val print("%3d) %s" % (pos + 1, val)) choices = raw_input("Choice(s) (Ctrl+C to cancel): ") if 'q' in choices.lower(): enqueue = True choices = choices.replace('q', '') # FIXME: This will distort # the "Not an integer" message for values containing "q". output = [] for index in parse_choice(choices): if index > 0 and index <= len(results): output.append(results[index - 1]) else: print("Invalid result index: %d" % index) return output, enqueue<|fim▁end|>
try: return [int(in_str)] except ValueError: choices = []
<|file_name|>bitcoin_sk.ts<|end_file_name|><|fim▁begin|><?xml version="1.0" ?><!DOCTYPE TS><TS language="sk" version="2.0"> <defaultcodec>UTF-8</defaultcodec> <context> <name>AboutDialog</name> <message> <location filename="../forms/aboutdialog.ui" line="+14"/> <source>About StreamCoin</source> <translation>O StreamCoin</translation> </message> <message> <location line="+39"/> <source>&lt;b&gt;StreamCoin&lt;/b&gt; version</source> <translation>&lt;b&gt;StreamCoin&lt;/b&gt; verzia</translation> </message> <message> <location line="+57"/> <source> This is experimental software. Distributed under the MIT/X11 software license, see the accompanying file COPYING or http://www.opensource.org/licenses/mit-license.php. This product includes software developed by the OpenSSL Project for use in the OpenSSL Toolkit (http://www.openssl.org/) and cryptographic software written by Eric Young ([email protected]) and UPnP software written by Thomas Bernard.</source> <translation type="unfinished"/> </message> <message> <location filename="../aboutdialog.cpp" line="+14"/> <source>Copyright</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>The StreamCoin developers</source> <translation type="unfinished"/> </message> </context> <context> <name>AddressBookPage</name> <message> <location filename="../forms/addressbookpage.ui" line="+14"/> <source>Address Book</source> <translation>Adresár</translation> </message> <message> <location line="+19"/> <source>Double-click to edit address or label</source> <translation>Dvojklikom editovať adresu alebo popis</translation> </message> <message> <location line="+27"/> <source>Create a new address</source> <translation>Vytvoriť novú adresu</translation> </message> <message> <location line="+14"/> <source>Copy the currently selected address to the system clipboard</source> <translation>Kopírovať práve zvolenú adresu do systémového klipbordu</translation> </message> <message> <location line="-11"/> <source>&amp;New Address</source> <translation>&amp;Nová adresa</translation> </message> <message> <location filename="../addressbookpage.cpp" line="+63"/> <source>These are your StreamCoin addresses for receiving payments. You may want to give a different one to each sender so you can keep track of who is paying you.</source> <translation>Toto sú Vaše StreamCoin adresy pre prijímanie platieb. Môžete dať každému odosielateľovi inú rôznu adresu a tak udržiavať prehľad o platbách.</translation> </message> <message> <location filename="../forms/addressbookpage.ui" line="+14"/> <source>&amp;Copy Address</source> <translation>&amp;Kopírovať adresu</translation> </message> <message> <location line="+11"/> <source>Show &amp;QR Code</source> <translation>Zobraz &amp;QR Kód</translation> </message> <message> <location line="+11"/> <source>Sign a message to prove you own a StreamCoin address</source> <translation>Podpísať správu a dokázať že vlastníte túto adresu</translation> </message> <message> <location line="+3"/> <source>Sign &amp;Message</source> <translation>Podpísať &amp;správu</translation> </message> <message> <location line="+25"/> <source>Delete the currently selected address from the list</source> <translation type="unfinished"/> </message> <message> <location line="+27"/> <source>Export the data in the current tab to a file</source> <translation>Exportovať tento náhľad do súboru</translation> </message> <message> <location line="+3"/> <source>&amp;Export</source> <translation type="unfinished"/> </message> <message> <location line="-44"/> <source>Verify a message to ensure it was signed with a specified StreamCoin address</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>&amp;Verify Message</source> <translation type="unfinished"/> </message><|fim▁hole|> <translation>&amp;Zmazať</translation> </message> <message> <location filename="../addressbookpage.cpp" line="-5"/> <source>These are your StreamCoin addresses for sending payments. Always check the amount and the receiving address before sending coins.</source> <translation type="unfinished"/> </message> <message> <location line="+13"/> <source>Copy &amp;Label</source> <translation>Kopírovať &amp;popis</translation> </message> <message> <location line="+1"/> <source>&amp;Edit</source> <translation>&amp;Upraviť</translation> </message> <message> <location line="+1"/> <source>Send &amp;Coins</source> <translation type="unfinished"/> </message> <message> <location line="+260"/> <source>Export Address Book Data</source> <translation>Exportovať dáta z adresára</translation> </message> <message> <location line="+1"/> <source>Comma separated file (*.csv)</source> <translation>Čiarkou oddelený súbor (*.csv)</translation> </message> <message> <location line="+13"/> <source>Error exporting</source> <translation>Chyba exportu.</translation> </message> <message> <location line="+0"/> <source>Could not write to file %1.</source> <translation>Nedalo sa zapisovať do súboru %1.</translation> </message> </context> <context> <name>AddressTableModel</name> <message> <location filename="../addresstablemodel.cpp" line="+144"/> <source>Label</source> <translation>Popis</translation> </message> <message> <location line="+0"/> <source>Address</source> <translation>Adresa</translation> </message> <message> <location line="+36"/> <source>(no label)</source> <translation>(bez popisu)</translation> </message> </context> <context> <name>AskPassphraseDialog</name> <message> <location filename="../forms/askpassphrasedialog.ui" line="+26"/> <source>Passphrase Dialog</source> <translation type="unfinished"/> </message> <message> <location line="+21"/> <source>Enter passphrase</source> <translation>Zadajte heslo</translation> </message> <message> <location line="+14"/> <source>New passphrase</source> <translation>Nové heslo</translation> </message> <message> <location line="+14"/> <source>Repeat new passphrase</source> <translation>Zopakujte nové heslo</translation> </message> <message> <location filename="../askpassphrasedialog.cpp" line="+33"/> <source>Enter the new passphrase to the wallet.&lt;br/&gt;Please use a passphrase of &lt;b&gt;10 or more random characters&lt;/b&gt;, or &lt;b&gt;eight or more words&lt;/b&gt;.</source> <translation>Zadajte nové heslo k peňaženke.&lt;br/&gt;Prosím použite heslo s dĺžkou aspon &lt;b&gt;10 alebo viac náhodných znakov&lt;/b&gt;, alebo &lt;b&gt;8 alebo viac slov&lt;/b&gt;.</translation> </message> <message> <location line="+1"/> <source>Encrypt wallet</source> <translation>Zašifrovať peňaženku</translation> </message> <message> <location line="+3"/> <source>This operation needs your wallet passphrase to unlock the wallet.</source> <translation>Táto operácia potrebuje heslo k vašej peňaženke aby ju mohla dešifrovať.</translation> </message> <message> <location line="+5"/> <source>Unlock wallet</source> <translation>Odomknúť peňaženku</translation> </message> <message> <location line="+3"/> <source>This operation needs your wallet passphrase to decrypt the wallet.</source> <translation>Táto operácia potrebuje heslo k vašej peňaženke na dešifrovanie peňaženky.</translation> </message> <message> <location line="+5"/> <source>Decrypt wallet</source> <translation>Dešifrovať peňaženku</translation> </message> <message> <location line="+3"/> <source>Change passphrase</source> <translation>Zmena hesla</translation> </message> <message> <location line="+1"/> <source>Enter the old and new passphrase to the wallet.</source> <translation>Zadajte staré a nové heslo k peňaženke.</translation> </message> <message> <location line="+46"/> <source>Confirm wallet encryption</source> <translation>Potvrďte šifrovanie peňaženky</translation> </message> <message> <location line="+1"/> <source>Warning: If you encrypt your wallet and lose your passphrase, you will &lt;b&gt;LOSE ALL OF YOUR LITECOINS&lt;/b&gt;!</source> <translation>Varovanie: Ak zašifrujete peňaženku a stratíte heslo, &lt;b&gt;STRATÍTE VŠETKY VAŠE LITECOINY&lt;/b&gt;!⏎</translation> </message> <message> <location line="+0"/> <source>Are you sure you wish to encrypt your wallet?</source> <translation>Ste si istí, že si želáte zašifrovať peňaženku?</translation> </message> <message> <location line="+15"/> <source>IMPORTANT: Any previous backups you have made of your wallet file should be replaced with the newly generated, encrypted wallet file. For security reasons, previous backups of the unencrypted wallet file will become useless as soon as you start using the new, encrypted wallet.</source> <translation type="unfinished"/> </message> <message> <location line="+100"/> <location line="+24"/> <source>Warning: The Caps Lock key is on!</source> <translation>Varovanie: Caps Lock je zapnutý</translation> </message> <message> <location line="-130"/> <location line="+58"/> <source>Wallet encrypted</source> <translation>Peňaženka zašifrovaná</translation> </message> <message> <location line="-56"/> <source>StreamCoin will close now to finish the encryption process. Remember that encrypting your wallet cannot fully protect your streamcoins from being stolen by malware infecting your computer.</source> <translation>StreamCoin sa teraz ukončí pre dokončenie procesu šifrovania. Pamätaj že šifrovanie peňaženky Ťa nemôže úplne ochrániť pred kráďežou streamcoinov pomocou škodlivého software.</translation> </message> <message> <location line="+13"/> <location line="+7"/> <location line="+42"/> <location line="+6"/> <source>Wallet encryption failed</source> <translation>Šifrovanie peňaženky zlyhalo</translation> </message> <message> <location line="-54"/> <source>Wallet encryption failed due to an internal error. Your wallet was not encrypted.</source> <translation>Šifrovanie peňaženky zlyhalo kôli internej chybe. Vaša peňaženka nebola zašifrovaná.</translation> </message> <message> <location line="+7"/> <location line="+48"/> <source>The supplied passphrases do not match.</source> <translation>Zadané heslá nesúhlasia.</translation> </message> <message> <location line="-37"/> <source>Wallet unlock failed</source> <translation>Odomykanie peňaženky zlyhalo</translation> </message> <message> <location line="+1"/> <location line="+11"/> <location line="+19"/> <source>The passphrase entered for the wallet decryption was incorrect.</source> <translation>Zadané heslo pre dešifrovanie peňaženky bolo nesprávne.</translation> </message> <message> <location line="-20"/> <source>Wallet decryption failed</source> <translation>Zlyhalo šifrovanie peňaženky.</translation> </message> <message> <location line="+14"/> <source>Wallet passphrase was successfully changed.</source> <translation>Heslo k peňaženke bolo úspešne zmenené.</translation> </message> </context> <context> <name>BitcoinGUI</name> <message> <location filename="../bitcoingui.cpp" line="+233"/> <source>Sign &amp;message...</source> <translation>Podpísať &amp;správu...</translation> </message> <message> <location line="+280"/> <source>Synchronizing with network...</source> <translation>Synchronizácia so sieťou...</translation> </message> <message> <location line="-349"/> <source>&amp;Overview</source> <translation>&amp;Prehľad</translation> </message> <message> <location line="+1"/> <source>Show general overview of wallet</source> <translation>Zobraziť celkový prehľad o peňaženke</translation> </message> <message> <location line="+20"/> <source>&amp;Transactions</source> <translation>&amp;Transakcie</translation> </message> <message> <location line="+1"/> <source>Browse transaction history</source> <translation>Prechádzať históriu transakcií</translation> </message> <message> <location line="+7"/> <source>Edit the list of stored addresses and labels</source> <translation>Editovať zoznam uložených adries a popisov</translation> </message> <message> <location line="-14"/> <source>Show the list of addresses for receiving payments</source> <translation>Zobraziť zoznam adries pre prijímanie platieb.</translation> </message> <message> <location line="+31"/> <source>E&amp;xit</source> <translation>U&amp;končiť</translation> </message> <message> <location line="+1"/> <source>Quit application</source> <translation>Ukončiť program</translation> </message> <message> <location line="+4"/> <source>Show information about StreamCoin</source> <translation>Zobraziť informácie o StreamCoin</translation> </message> <message> <location line="+2"/> <source>About &amp;Qt</source> <translation>O &amp;Qt</translation> </message> <message> <location line="+1"/> <source>Show information about Qt</source> <translation>Zobrazit informácie o Qt</translation> </message> <message> <location line="+2"/> <source>&amp;Options...</source> <translation>&amp;Možnosti...</translation> </message> <message> <location line="+6"/> <source>&amp;Encrypt Wallet...</source> <translation>&amp;Zašifrovať Peňaženku...</translation> </message> <message> <location line="+3"/> <source>&amp;Backup Wallet...</source> <translation>&amp;Backup peňaženku...</translation> </message> <message> <location line="+2"/> <source>&amp;Change Passphrase...</source> <translation>&amp;Zmena Hesla...</translation> </message> <message> <location line="+285"/> <source>Importing blocks from disk...</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Reindexing blocks on disk...</source> <translation type="unfinished"/> </message> <message> <location line="-347"/> <source>Send coins to a StreamCoin address</source> <translation>Poslať streamcoins na adresu</translation> </message> <message> <location line="+49"/> <source>Modify configuration options for StreamCoin</source> <translation>Upraviť možnosti nastavenia pre streamcoin</translation> </message> <message> <location line="+9"/> <source>Backup wallet to another location</source> <translation>Zálohovať peňaženku na iné miesto</translation> </message> <message> <location line="+2"/> <source>Change the passphrase used for wallet encryption</source> <translation>Zmeniť heslo použité na šifrovanie peňaženky</translation> </message> <message> <location line="+6"/> <source>&amp;Debug window</source> <translation>&amp;Okno pre ladenie</translation> </message> <message> <location line="+1"/> <source>Open debugging and diagnostic console</source> <translation>Otvor konzolu pre ladenie a diagnostiku</translation> </message> <message> <location line="-4"/> <source>&amp;Verify message...</source> <translation type="unfinished"/> </message> <message> <location line="-165"/> <location line="+530"/> <source>StreamCoin</source> <translation type="unfinished"/> </message> <message> <location line="-530"/> <source>Wallet</source> <translation>Peňaženka</translation> </message> <message> <location line="+101"/> <source>&amp;Send</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>&amp;Receive</source> <translation type="unfinished"/> </message> <message> <location line="+14"/> <source>&amp;Addresses</source> <translation type="unfinished"/> </message> <message> <location line="+22"/> <source>&amp;About StreamCoin</source> <translation>&amp;O StreamCoin</translation> </message> <message> <location line="+9"/> <source>&amp;Show / Hide</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Show or hide the main Window</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Encrypt the private keys that belong to your wallet</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>Sign messages with your StreamCoin addresses to prove you own them</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Verify messages to ensure they were signed with specified StreamCoin addresses</source> <translation type="unfinished"/> </message> <message> <location line="+28"/> <source>&amp;File</source> <translation>&amp;Súbor</translation> </message> <message> <location line="+7"/> <source>&amp;Settings</source> <translation>&amp;Nastavenia</translation> </message> <message> <location line="+6"/> <source>&amp;Help</source> <translation>&amp;Pomoc</translation> </message> <message> <location line="+9"/> <source>Tabs toolbar</source> <translation>Lišta záložiek</translation> </message> <message> <location line="+17"/> <location line="+10"/> <source>[testnet]</source> <translation>[testovacia sieť]</translation> </message> <message> <location line="+47"/> <source>StreamCoin client</source> <translation>StreamCoin klient</translation> </message> <message numerus="yes"> <location line="+141"/> <source>%n active connection(s) to StreamCoin network</source> <translation><numerusform>%n aktívne spojenie v StreamCoin sieti</numerusform><numerusform>%n aktívne spojenia v StreamCoin sieti</numerusform><numerusform>%n aktívnych spojení v Bitconi sieti</numerusform></translation> </message> <message> <location line="+22"/> <source>No block source available...</source> <translation type="unfinished"/> </message> <message> <location line="+12"/> <source>Processed %1 of %2 (estimated) blocks of transaction history.</source> <translation type="unfinished"/> </message> <message> <location line="+4"/> <source>Processed %1 blocks of transaction history.</source> <translation type="unfinished"/> </message> <message numerus="yes"> <location line="+20"/> <source>%n hour(s)</source> <translation type="unfinished"><numerusform></numerusform><numerusform></numerusform><numerusform></numerusform></translation> </message> <message numerus="yes"> <location line="+4"/> <source>%n day(s)</source> <translation type="unfinished"><numerusform></numerusform><numerusform></numerusform><numerusform></numerusform></translation> </message> <message numerus="yes"> <location line="+4"/> <source>%n week(s)</source> <translation type="unfinished"><numerusform></numerusform><numerusform></numerusform><numerusform></numerusform></translation> </message> <message> <location line="+4"/> <source>%1 behind</source> <translation type="unfinished"/> </message> <message> <location line="+14"/> <source>Last received block was generated %1 ago.</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Transactions after this will not yet be visible.</source> <translation type="unfinished"/> </message> <message> <location line="+22"/> <source>Error</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Warning</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Information</source> <translation type="unfinished"/> </message> <message> <location line="+70"/> <source>This transaction is over the size limit. You can still send it for a fee of %1, which goes to the nodes that process your transaction and helps to support the network. Do you want to pay the fee?</source> <translation type="unfinished"/> </message> <message> <location line="-140"/> <source>Up to date</source> <translation>Aktualizovaný</translation> </message> <message> <location line="+31"/> <source>Catching up...</source> <translation>Sťahujem...</translation> </message> <message> <location line="+113"/> <source>Confirm transaction fee</source> <translation>Potvrď poplatok za transakciu.</translation> </message> <message> <location line="+8"/> <source>Sent transaction</source> <translation>Odoslané transakcie</translation> </message> <message> <location line="+0"/> <source>Incoming transaction</source> <translation>Prijaté transakcie</translation> </message> <message> <location line="+1"/> <source>Date: %1 Amount: %2 Type: %3 Address: %4 </source> <translation>Dátum: %1 Suma: %2 Typ: %3 Adresa: %4</translation> </message> <message> <location line="+33"/> <location line="+23"/> <source>URI handling</source> <translation type="unfinished"/> </message> <message> <location line="-23"/> <location line="+23"/> <source>URI can not be parsed! This can be caused by an invalid StreamCoin address or malformed URI parameters.</source> <translation type="unfinished"/> </message> <message> <location line="+17"/> <source>Wallet is &lt;b&gt;encrypted&lt;/b&gt; and currently &lt;b&gt;unlocked&lt;/b&gt;</source> <translation>Peňaženka je &lt;b&gt;zašifrovaná&lt;/b&gt; a momentálne &lt;b&gt;odomknutá&lt;/b&gt;</translation> </message> <message> <location line="+8"/> <source>Wallet is &lt;b&gt;encrypted&lt;/b&gt; and currently &lt;b&gt;locked&lt;/b&gt;</source> <translation>Peňaženka je &lt;b&gt;zašifrovaná&lt;/b&gt; a momentálne &lt;b&gt;zamknutá&lt;/b&gt;</translation> </message> <message> <location filename="../bitcoin.cpp" line="+111"/> <source>A fatal error occurred. StreamCoin can no longer continue safely and will quit.</source> <translation type="unfinished"/> </message> </context> <context> <name>ClientModel</name> <message> <location filename="../clientmodel.cpp" line="+104"/> <source>Network Alert</source> <translation type="unfinished"/> </message> </context> <context> <name>EditAddressDialog</name> <message> <location filename="../forms/editaddressdialog.ui" line="+14"/> <source>Edit Address</source> <translation>Upraviť adresu</translation> </message> <message> <location line="+11"/> <source>&amp;Label</source> <translation>&amp;Popis</translation> </message> <message> <location line="+10"/> <source>The label associated with this address book entry</source> <translation>Popis priradený k tomuto záznamu v adresári</translation> </message> <message> <location line="+7"/> <source>&amp;Address</source> <translation>&amp;Adresa</translation> </message> <message> <location line="+10"/> <source>The address associated with this address book entry. This can only be modified for sending addresses.</source> <translation>Adresa spojená s týmto záznamom v adresári. Možno upravovať len pre odosielajúce adresy.</translation> </message> <message> <location filename="../editaddressdialog.cpp" line="+21"/> <source>New receiving address</source> <translation>Nová adresa pre prijímanie</translation> </message> <message> <location line="+4"/> <source>New sending address</source> <translation>Nová adresa pre odoslanie</translation> </message> <message> <location line="+3"/> <source>Edit receiving address</source> <translation>Upraviť prijímacie adresy</translation> </message> <message> <location line="+4"/> <source>Edit sending address</source> <translation>Upraviť odosielaciu adresu</translation> </message> <message> <location line="+76"/> <source>The entered address &quot;%1&quot; is already in the address book.</source> <translation>Vložená adresa &quot;%1&quot; sa už nachádza v adresári.</translation> </message> <message> <location line="-5"/> <source>The entered address &quot;%1&quot; is not a valid StreamCoin address.</source> <translation>Vložená adresa &quot;%1&quot; nieje platnou adresou streamcoin.</translation> </message> <message> <location line="+10"/> <source>Could not unlock wallet.</source> <translation>Nepodarilo sa odomknúť peňaženku.</translation> </message> <message> <location line="+5"/> <source>New key generation failed.</source> <translation>Generovanie nového kľúča zlyhalo.</translation> </message> </context> <context> <name>GUIUtil::HelpMessageBox</name> <message> <location filename="../guiutil.cpp" line="+424"/> <location line="+12"/> <source>StreamCoin-Qt</source> <translation type="unfinished"/> </message> <message> <location line="-12"/> <source>version</source> <translation>verzia</translation> </message> <message> <location line="+2"/> <source>Usage:</source> <translation>Použitie:</translation> </message> <message> <location line="+1"/> <source>command-line options</source> <translation type="unfinished"/> </message> <message> <location line="+4"/> <source>UI options</source> <translation>UI možnosti</translation> </message> <message> <location line="+1"/> <source>Set language, for example &quot;de_DE&quot; (default: system locale)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Start minimized</source> <translation>Spustiť minimalizované</translation> </message> <message> <location line="+1"/> <source>Show splash screen on startup (default: 1)</source> <translation type="unfinished"/> </message> </context> <context> <name>OptionsDialog</name> <message> <location filename="../forms/optionsdialog.ui" line="+14"/> <source>Options</source> <translation>Možnosti</translation> </message> <message> <location line="+16"/> <source>&amp;Main</source> <translation>&amp;Hlavné</translation> </message> <message> <location line="+6"/> <source>Optional transaction fee per kB that helps make sure your transactions are processed quickly. Most transactions are 1 kB.</source> <translation type="unfinished"/> </message> <message> <location line="+15"/> <source>Pay transaction &amp;fee</source> <translation>Zaplatiť transakčné &amp;poplatky</translation> </message> <message> <location line="+31"/> <source>Automatically start StreamCoin after logging in to the system.</source> <translation>Automaticky spustiť StreamCoin po zapnutí počítača</translation> </message> <message> <location line="+3"/> <source>&amp;Start StreamCoin on system login</source> <translation>&amp;Spustiť StreamCoin pri spustení systému správy okien</translation> </message> <message> <location line="+35"/> <source>Reset all client options to default.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>&amp;Reset Options</source> <translation type="unfinished"/> </message> <message> <location line="+13"/> <source>&amp;Network</source> <translation type="unfinished"/> </message> <message> <location line="+6"/> <source>Automatically open the StreamCoin client port on the router. This only works when your router supports UPnP and it is enabled.</source> <translation>Automaticky otvorit port pre StreamCoin na routeri. Toto funguje len ak router podporuje UPnP a je táto podpora aktivovaná.</translation> </message> <message> <location line="+3"/> <source>Map port using &amp;UPnP</source> <translation>Mapovať port pomocou &amp;UPnP</translation> </message> <message> <location line="+7"/> <source>Connect to the StreamCoin network through a SOCKS proxy (e.g. when connecting through Tor).</source> <translation>Pripojiť do siete StreamCoin cez SOCKS proxy (napr. keď sa pripájate cez Tor)</translation> </message> <message> <location line="+3"/> <source>&amp;Connect through SOCKS proxy:</source> <translation>&amp;Pripojiť cez SOCKS proxy:</translation> </message> <message> <location line="+9"/> <source>Proxy &amp;IP:</source> <translation type="unfinished"/> </message> <message> <location line="+19"/> <source>IP address of the proxy (e.g. 127.0.0.1)</source> <translation>IP addresa proxy (napr. 127.0.0.1)</translation> </message> <message> <location line="+7"/> <source>&amp;Port:</source> <translation type="unfinished"/> </message> <message> <location line="+19"/> <source>Port of the proxy (e.g. 9050)</source> <translation>Port proxy (napr. 9050)</translation> </message> <message> <location line="+7"/> <source>SOCKS &amp;Version:</source> <translation type="unfinished"/> </message> <message> <location line="+13"/> <source>SOCKS version of the proxy (e.g. 5)</source> <translation type="unfinished"/> </message> <message> <location line="+36"/> <source>&amp;Window</source> <translation type="unfinished"/> </message> <message> <location line="+6"/> <source>Show only a tray icon after minimizing the window.</source> <translation>Zobraziť len ikonu na lište po minimalizovaní okna.</translation> </message> <message> <location line="+3"/> <source>&amp;Minimize to the tray instead of the taskbar</source> <translation>Zobraziť len ikonu na lište po minimalizovaní okna.</translation> </message> <message> <location line="+7"/> <source>Minimize instead of exit the application when the window is closed. When this option is enabled, the application will be closed only after selecting Quit in the menu.</source> <translation>Minimalizovat namiesto ukončenia aplikácie keď sa okno zavrie. Keď je zvolená táto možnosť, aplikácia sa zavrie len po zvolení Ukončiť v menu.</translation> </message> <message> <location line="+3"/> <source>M&amp;inimize on close</source> <translation>M&amp;inimalizovať pri zavretí</translation> </message> <message> <location line="+21"/> <source>&amp;Display</source> <translation>&amp;Displej</translation> </message> <message> <location line="+8"/> <source>User Interface &amp;language:</source> <translation type="unfinished"/> </message> <message> <location line="+13"/> <source>The user interface language can be set here. This setting will take effect after restarting StreamCoin.</source> <translation type="unfinished"/> </message> <message> <location line="+11"/> <source>&amp;Unit to show amounts in:</source> <translation>&amp;Zobrazovať hodnoty v jednotkách:</translation> </message> <message> <location line="+13"/> <source>Choose the default subdivision unit to show in the interface and when sending coins.</source> <translation type="unfinished"/> </message> <message> <location line="+9"/> <source>Whether to show StreamCoin addresses in the transaction list or not.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>&amp;Display addresses in transaction list</source> <translation>&amp;Zobraziť adresy zo zoznamu transakcií</translation> </message> <message> <location line="+71"/> <source>&amp;OK</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>&amp;Cancel</source> <translation type="unfinished"/> </message> <message> <location line="+10"/> <source>&amp;Apply</source> <translation type="unfinished"/> </message> <message> <location filename="../optionsdialog.cpp" line="+53"/> <source>default</source> <translation type="unfinished"/> </message> <message> <location line="+130"/> <source>Confirm options reset</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Some settings may require a client restart to take effect.</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>Do you want to proceed?</source> <translation type="unfinished"/> </message> <message> <location line="+42"/> <location line="+9"/> <source>Warning</source> <translation>Varovanie</translation> </message> <message> <location line="-9"/> <location line="+9"/> <source>This setting will take effect after restarting StreamCoin.</source> <translation type="unfinished"/> </message> <message> <location line="+29"/> <source>The supplied proxy address is invalid.</source> <translation type="unfinished"/> </message> </context> <context> <name>OverviewPage</name> <message> <location filename="../forms/overviewpage.ui" line="+14"/> <source>Form</source> <translation>Forma</translation> </message> <message> <location line="+50"/> <location line="+166"/> <source>The displayed information may be out of date. Your wallet automatically synchronizes with the StreamCoin network after a connection is established, but this process has not completed yet.</source> <translation type="unfinished"/> </message> <message> <location line="-124"/> <source>Balance:</source> <translation>Zostatok:</translation> </message> <message> <location line="+29"/> <source>Unconfirmed:</source> <translation>Nepotvrdené:</translation> </message> <message> <location line="-78"/> <source>Wallet</source> <translation>Peňaženka</translation> </message> <message> <location line="+107"/> <source>Immature:</source> <translation type="unfinished"/> </message> <message> <location line="+13"/> <source>Mined balance that has not yet matured</source> <translation type="unfinished"/> </message> <message> <location line="+46"/> <source>&lt;b&gt;Recent transactions&lt;/b&gt;</source> <translation>&lt;b&gt;Nedávne transakcie&lt;/b&gt;</translation> </message> <message> <location line="-101"/> <source>Your current balance</source> <translation>Váš súčasný zostatok</translation> </message> <message> <location line="+29"/> <source>Total of transactions that have yet to be confirmed, and do not yet count toward the current balance</source> <translation>Suma transakcií ktoré ešte neboli potvrdené a nezapočítavaju sa do celkového zostatku.</translation> </message> <message> <location filename="../overviewpage.cpp" line="+116"/> <location line="+1"/> <source>out of sync</source> <translation type="unfinished"/> </message> </context> <context> <name>PaymentServer</name> <message> <location filename="../paymentserver.cpp" line="+107"/> <source>Cannot start streamcoin: click-to-pay handler</source> <translation type="unfinished"/> </message> </context> <context> <name>QRCodeDialog</name> <message> <location filename="../forms/qrcodedialog.ui" line="+14"/> <source>QR Code Dialog</source> <translation type="unfinished"/> </message> <message> <location line="+59"/> <source>Request Payment</source> <translation>Vyžiadať platbu</translation> </message> <message> <location line="+56"/> <source>Amount:</source> <translation>Suma:</translation> </message> <message> <location line="-44"/> <source>Label:</source> <translation>Popis:</translation> </message> <message> <location line="+19"/> <source>Message:</source> <translation>Správa:</translation> </message> <message> <location line="+71"/> <source>&amp;Save As...</source> <translation>&amp;Uložiť ako...</translation> </message> <message> <location filename="../qrcodedialog.cpp" line="+62"/> <source>Error encoding URI into QR Code.</source> <translation>Chyba v zakódovaní URI do QR kódu</translation> </message> <message> <location line="+40"/> <source>The entered amount is invalid, please check.</source> <translation type="unfinished"/> </message> <message> <location line="+23"/> <source>Resulting URI too long, try to reduce the text for label / message.</source> <translation>Výsledné URI príliš dlhé, skráť text pre názov / správu.</translation> </message> <message> <location line="+25"/> <source>Save QR Code</source> <translation>Ukladanie QR kódu</translation> </message> <message> <location line="+0"/> <source>PNG Images (*.png)</source> <translation>PNG obrázky (*.png)</translation> </message> </context> <context> <name>RPCConsole</name> <message> <location filename="../forms/rpcconsole.ui" line="+46"/> <source>Client name</source> <translation>Meno klienta</translation> </message> <message> <location line="+10"/> <location line="+23"/> <location line="+26"/> <location line="+23"/> <location line="+23"/> <location line="+36"/> <location line="+53"/> <location line="+23"/> <location line="+23"/> <location filename="../rpcconsole.cpp" line="+339"/> <source>N/A</source> <translation>nie je k dispozícii</translation> </message> <message> <location line="-217"/> <source>Client version</source> <translation>Verzia klienta</translation> </message> <message> <location line="-45"/> <source>&amp;Information</source> <translation type="unfinished"/> </message> <message> <location line="+68"/> <source>Using OpenSSL version</source> <translation type="unfinished"/> </message> <message> <location line="+49"/> <source>Startup time</source> <translation type="unfinished"/> </message> <message> <location line="+29"/> <source>Network</source> <translation>Sieť</translation> </message> <message> <location line="+7"/> <source>Number of connections</source> <translation>Počet pripojení</translation> </message> <message> <location line="+23"/> <source>On testnet</source> <translation>Na testovacej sieti</translation> </message> <message> <location line="+23"/> <source>Block chain</source> <translation>Reťazec blokov</translation> </message> <message> <location line="+7"/> <source>Current number of blocks</source> <translation>Aktuálny počet blokov</translation> </message> <message> <location line="+23"/> <source>Estimated total blocks</source> <translation type="unfinished"/> </message> <message> <location line="+23"/> <source>Last block time</source> <translation type="unfinished"/> </message> <message> <location line="+52"/> <source>&amp;Open</source> <translation type="unfinished"/> </message> <message> <location line="+16"/> <source>Command-line options</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>Show the StreamCoin-Qt help message to get a list with possible StreamCoin command-line options.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>&amp;Show</source> <translation type="unfinished"/> </message> <message> <location line="+24"/> <source>&amp;Console</source> <translation type="unfinished"/> </message> <message> <location line="-260"/> <source>Build date</source> <translation type="unfinished"/> </message> <message> <location line="-104"/> <source>StreamCoin - Debug window</source> <translation type="unfinished"/> </message> <message> <location line="+25"/> <source>StreamCoin Core</source> <translation type="unfinished"/> </message> <message> <location line="+279"/> <source>Debug log file</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>Open the StreamCoin debug log file from the current data directory. This can take a few seconds for large log files.</source> <translation type="unfinished"/> </message> <message> <location line="+102"/> <source>Clear console</source> <translation type="unfinished"/> </message> <message> <location filename="../rpcconsole.cpp" line="-30"/> <source>Welcome to the StreamCoin RPC console.</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Use up and down arrows to navigate history, and &lt;b&gt;Ctrl-L&lt;/b&gt; to clear screen.</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Type &lt;b&gt;help&lt;/b&gt; for an overview of available commands.</source> <translation type="unfinished"/> </message> </context> <context> <name>SendCoinsDialog</name> <message> <location filename="../forms/sendcoinsdialog.ui" line="+14"/> <location filename="../sendcoinsdialog.cpp" line="+124"/> <location line="+5"/> <location line="+5"/> <location line="+5"/> <location line="+6"/> <location line="+5"/> <location line="+5"/> <source>Send Coins</source> <translation>Poslať StreamCoins</translation> </message> <message> <location line="+50"/> <source>Send to multiple recipients at once</source> <translation>Poslať viacerým príjemcom naraz</translation> </message> <message> <location line="+3"/> <source>Add &amp;Recipient</source> <translation>&amp;Pridať príjemcu</translation> </message> <message> <location line="+20"/> <source>Remove all transaction fields</source> <translation>Odobrať všetky políčka transakcie</translation> </message> <message> <location line="+3"/> <source>Clear &amp;All</source> <translation>Zmazať &amp;všetko</translation> </message> <message> <location line="+22"/> <source>Balance:</source> <translation>Zostatok:</translation> </message> <message> <location line="+10"/> <source>123.456 BTC</source> <translation>123.456 BTC</translation> </message> <message> <location line="+31"/> <source>Confirm the send action</source> <translation>Potvrďte odoslanie</translation> </message> <message> <location line="+3"/> <source>S&amp;end</source> <translation>&amp;Odoslať</translation> </message> <message> <location filename="../sendcoinsdialog.cpp" line="-59"/> <source>&lt;b&gt;%1&lt;/b&gt; to %2 (%3)</source> <translation>&lt;b&gt;%1&lt;/b&gt; do %2 (%3)</translation> </message> <message> <location line="+5"/> <source>Confirm send coins</source> <translation>Potvrdiť odoslanie streamcoins</translation> </message> <message> <location line="+1"/> <source>Are you sure you want to send %1?</source> <translation>Ste si istí, že chcete odoslať %1?</translation> </message> <message> <location line="+0"/> <source> and </source> <translation> a</translation> </message> <message> <location line="+23"/> <source>The recipient address is not valid, please recheck.</source> <translation>Adresa príjemcu je neplatná, prosím, overte ju.</translation> </message> <message> <location line="+5"/> <source>The amount to pay must be larger than 0.</source> <translation>Suma na úhradu musí byť väčšia ako 0.</translation> </message> <message> <location line="+5"/> <source>The amount exceeds your balance.</source> <translation>Suma je vyššia ako Váš zostatok.</translation> </message> <message> <location line="+5"/> <source>The total exceeds your balance when the %1 transaction fee is included.</source> <translation>Suma celkom prevyšuje Váš zostatok ak sú započítané %1 transakčné poplatky.</translation> </message> <message> <location line="+6"/> <source>Duplicate address found, can only send to each address once per send operation.</source> <translation>Duplikát adresy objavený, je možné poslať na každú adresu len raz v jednej odchádzajúcej transakcii.</translation> </message> <message> <location line="+5"/> <source>Error: Transaction creation failed!</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Error: The transaction was rejected. This might happen if some of the coins in your wallet were already spent, such as if you used a copy of wallet.dat and coins were spent in the copy but not marked as spent here.</source> <translation>Chyba: Transakcia bola odmietnutá. Toto sa môže stať ak niektoré z mincí vo vašej peňaženke boli už utratené, napríklad ak používaš kópiu wallet.dat a mince označené v druhej kópií neboli označené ako utratené v tejto.</translation> </message> </context> <context> <name>SendCoinsEntry</name> <message> <location filename="../forms/sendcoinsentry.ui" line="+14"/> <source>Form</source> <translation>Forma</translation> </message> <message> <location line="+15"/> <source>A&amp;mount:</source> <translation>Su&amp;ma:</translation> </message> <message> <location line="+13"/> <source>Pay &amp;To:</source> <translation>Zapla&amp;tiť:</translation> </message> <message> <location line="+34"/> <source>The address to send the payment to (e.g. Ser4Ac2awZmCQrGFP2W34PM2UAlC8ax3)</source> <translation type="unfinished"/> </message> <message> <location line="+60"/> <location filename="../sendcoinsentry.cpp" line="+26"/> <source>Enter a label for this address to add it to your address book</source> <translation>Vložte popis pre túto adresu aby sa pridala do adresára</translation> </message> <message> <location line="-78"/> <source>&amp;Label:</source> <translation>&amp;Popis:</translation> </message> <message> <location line="+28"/> <source>Choose address from address book</source> <translation>Zvoľte adresu z adresára</translation> </message> <message> <location line="+10"/> <source>Alt+A</source> <translation>Alt+A</translation> </message> <message> <location line="+7"/> <source>Paste address from clipboard</source> <translation>Vložiť adresu z klipbordu</translation> </message> <message> <location line="+10"/> <source>Alt+P</source> <translation>Alt+P</translation> </message> <message> <location line="+7"/> <source>Remove this recipient</source> <translation>Odstrániť tohto príjemcu</translation> </message> <message> <location filename="../sendcoinsentry.cpp" line="+1"/> <source>Enter a StreamCoin address (e.g. Ser4Ac2awZmCQrGFP2W34PM2UAlC8ax3)</source> <translation>Zadajte StreamCoin adresu (napr. Ser4Ac2awZmCQrGFP2W34PM2UAlC8ax3)</translation> </message> </context> <context> <name>SignVerifyMessageDialog</name> <message> <location filename="../forms/signverifymessagedialog.ui" line="+14"/> <source>Signatures - Sign / Verify a Message</source> <translation type="unfinished"/> </message> <message> <location line="+13"/> <source>&amp;Sign Message</source> <translation>&amp;Podpísať Správu</translation> </message> <message> <location line="+6"/> <source>You can sign messages with your addresses to prove you own them. Be careful not to sign anything vague, as phishing attacks may try to trick you into signing your identity over to them. Only sign fully-detailed statements you agree to.</source> <translation>Môžete podpísať správy svojou adresou a dokázať, že ju vlastníte. Buďte opatrní a podpíšte len prehlásenia s ktorými plne súhlasíte, nakoľko útoky typu &quot;phishing&quot; Vás môžu lákať k ich podpísaniu.</translation> </message> <message> <location line="+18"/> <source>The address to sign the message with (e.g. Ser4Ac2awZmCQrGFP2W34PM2UAlC8ax3)</source> <translation>Zadajte StreamCoin adresu (napr. Ser4Ac2awZmCQrGFP2W34PM2UAlC8ax3)</translation> </message> <message> <location line="+10"/> <location line="+213"/> <source>Choose an address from the address book</source> <translation>Zvoľte adresu z adresára</translation> </message> <message> <location line="-203"/> <location line="+213"/> <source>Alt+A</source> <translation>Alt+A</translation> </message> <message> <location line="-203"/> <source>Paste address from clipboard</source> <translation>Vložte adresu z klipbordu</translation> </message> <message> <location line="+10"/> <source>Alt+P</source> <translation>Alt+P</translation> </message> <message> <location line="+12"/> <source>Enter the message you want to sign here</source> <translation>Sem vložte správu ktorú chcete podpísať</translation> </message> <message> <location line="+7"/> <source>Signature</source> <translation type="unfinished"/> </message> <message> <location line="+27"/> <source>Copy the current signature to the system clipboard</source> <translation type="unfinished"/> </message> <message> <location line="+21"/> <source>Sign the message to prove you own this StreamCoin address</source> <translation>Podpíšte správu aby ste dokázali že vlastníte túto adresu</translation> </message> <message> <location line="+3"/> <source>Sign &amp;Message</source> <translation type="unfinished"/> </message> <message> <location line="+14"/> <source>Reset all sign message fields</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <location line="+146"/> <source>Clear &amp;All</source> <translation>Zmazať &amp;všetko</translation> </message> <message> <location line="-87"/> <source>&amp;Verify Message</source> <translation type="unfinished"/> </message> <message> <location line="+6"/> <source>Enter the signing address, message (ensure you copy line breaks, spaces, tabs, etc. exactly) and signature below to verify the message. Be careful not to read more into the signature than what is in the signed message itself, to avoid being tricked by a man-in-the-middle attack.</source> <translation type="unfinished"/> </message> <message> <location line="+21"/> <source>The address the message was signed with (e.g. Ser4Ac2awZmCQrGFP2W34PM2UAlC8ax3)</source> <translation>Zadajte StreamCoin adresu (napr. Ser4Ac2awZmCQrGFP2W34PM2UAlC8ax3)</translation> </message> <message> <location line="+40"/> <source>Verify the message to ensure it was signed with the specified StreamCoin address</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Verify &amp;Message</source> <translation type="unfinished"/> </message> <message> <location line="+14"/> <source>Reset all verify message fields</source> <translation type="unfinished"/> </message> <message> <location filename="../signverifymessagedialog.cpp" line="+27"/> <location line="+3"/> <source>Enter a StreamCoin address (e.g. Ser4Ac2awZmCQrGFP2W34PM2UAlC8ax3)</source> <translation>Zadajte StreamCoin adresu (napr. Ser4Ac2awZmCQrGFP2W34PM2UAlC8ax3)</translation> </message> <message> <location line="-2"/> <source>Click &quot;Sign Message&quot; to generate signature</source> <translation>Kliknite &quot;Podpísať Správu&quot; na získanie podpisu</translation> </message> <message> <location line="+3"/> <source>Enter StreamCoin signature</source> <translation type="unfinished"/> </message> <message> <location line="+82"/> <location line="+81"/> <source>The entered address is invalid.</source> <translation type="unfinished"/> </message> <message> <location line="-81"/> <location line="+8"/> <location line="+73"/> <location line="+8"/> <source>Please check the address and try again.</source> <translation type="unfinished"/> </message> <message> <location line="-81"/> <location line="+81"/> <source>The entered address does not refer to a key.</source> <translation type="unfinished"/> </message> <message> <location line="-73"/> <source>Wallet unlock was cancelled.</source> <translation type="unfinished"/> </message> <message> <location line="+8"/> <source>Private key for the entered address is not available.</source> <translation type="unfinished"/> </message> <message> <location line="+12"/> <source>Message signing failed.</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Message signed.</source> <translation type="unfinished"/> </message> <message> <location line="+59"/> <source>The signature could not be decoded.</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <location line="+13"/> <source>Please check the signature and try again.</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>The signature did not match the message digest.</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>Message verification failed.</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Message verified.</source> <translation type="unfinished"/> </message> </context> <context> <name>SplashScreen</name> <message> <location filename="../splashscreen.cpp" line="+22"/> <source>The StreamCoin developers</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>[testnet]</source> <translation>[testovacia sieť]</translation> </message> </context> <context> <name>TransactionDesc</name> <message> <location filename="../transactiondesc.cpp" line="+20"/> <source>Open until %1</source> <translation>Otvorené do %1</translation> </message> <message> <location line="+6"/> <source>%1/offline</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>%1/unconfirmed</source> <translation>%1/nepotvrdené</translation> </message> <message> <location line="+2"/> <source>%1 confirmations</source> <translation>%1 potvrdení</translation> </message> <message> <location line="+18"/> <source>Status</source> <translation>Stav</translation> </message> <message numerus="yes"> <location line="+7"/> <source>, broadcast through %n node(s)</source> <translation type="unfinished"><numerusform></numerusform><numerusform></numerusform><numerusform></numerusform></translation> </message> <message> <location line="+4"/> <source>Date</source> <translation>Dátum</translation> </message> <message> <location line="+7"/> <source>Source</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>Generated</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <location line="+17"/> <source>From</source> <translation>od</translation> </message> <message> <location line="+1"/> <location line="+22"/> <location line="+58"/> <source>To</source> <translation type="unfinished"/> </message> <message> <location line="-77"/> <location line="+2"/> <source>own address</source> <translation type="unfinished"/> </message> <message> <location line="-2"/> <source>label</source> <translation>popis</translation> </message> <message> <location line="+37"/> <location line="+12"/> <location line="+45"/> <location line="+17"/> <location line="+30"/> <source>Credit</source> <translation>Kredit</translation> </message> <message numerus="yes"> <location line="-102"/> <source>matures in %n more block(s)</source> <translation type="unfinished"><numerusform></numerusform><numerusform></numerusform><numerusform></numerusform></translation> </message> <message> <location line="+2"/> <source>not accepted</source> <translation>neprijaté</translation> </message> <message> <location line="+44"/> <location line="+8"/> <location line="+15"/> <location line="+30"/> <source>Debit</source> <translation>Debet</translation> </message> <message> <location line="-39"/> <source>Transaction fee</source> <translation>Transakčný poplatok</translation> </message> <message> <location line="+16"/> <source>Net amount</source> <translation>Suma netto</translation> </message> <message> <location line="+6"/> <source>Message</source> <translation>Správa</translation> </message> <message> <location line="+2"/> <source>Comment</source> <translation>Komentár</translation> </message> <message> <location line="+2"/> <source>Transaction ID</source> <translation>ID transakcie</translation> </message> <message> <location line="+3"/> <source>Generated coins must mature 120 blocks before they can be spent. When you generated this block, it was broadcast to the network to be added to the block chain. If it fails to get into the chain, its state will change to &quot;not accepted&quot; and it won&apos;t be spendable. This may occasionally happen if another node generates a block within a few seconds of yours.</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>Debug information</source> <translation type="unfinished"/> </message> <message> <location line="+8"/> <source>Transaction</source> <translation>Transakcie</translation> </message> <message> <location line="+3"/> <source>Inputs</source> <translation type="unfinished"/> </message> <message> <location line="+23"/> <source>Amount</source> <translation>Suma</translation> </message> <message> <location line="+1"/> <source>true</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>false</source> <translation type="unfinished"/> </message> <message> <location line="-209"/> <source>, has not been successfully broadcast yet</source> <translation>, ešte nebola úspešne odoslaná</translation> </message> <message numerus="yes"> <location line="-35"/> <source>Open for %n more block(s)</source> <translation type="unfinished"><numerusform></numerusform><numerusform></numerusform><numerusform></numerusform></translation> </message> <message> <location line="+70"/> <source>unknown</source> <translation>neznámy</translation> </message> </context> <context> <name>TransactionDescDialog</name> <message> <location filename="../forms/transactiondescdialog.ui" line="+14"/> <source>Transaction details</source> <translation>Detaily transakcie</translation> </message> <message> <location line="+6"/> <source>This pane shows a detailed description of the transaction</source> <translation>Táto časť obrazovky zobrazuje detailný popis transakcie</translation> </message> </context> <context> <name>TransactionTableModel</name> <message> <location filename="../transactiontablemodel.cpp" line="+225"/> <source>Date</source> <translation>Dátum</translation> </message> <message> <location line="+0"/> <source>Type</source> <translation>Typ</translation> </message> <message> <location line="+0"/> <source>Address</source> <translation>Adresa</translation> </message> <message> <location line="+0"/> <source>Amount</source> <translation>Hodnota</translation> </message> <message numerus="yes"> <location line="+57"/> <source>Open for %n more block(s)</source> <translation type="unfinished"><numerusform></numerusform><numerusform></numerusform><numerusform></numerusform></translation> </message> <message> <location line="+3"/> <source>Open until %1</source> <translation>Otvorené do %1</translation> </message> <message> <location line="+3"/> <source>Offline (%1 confirmations)</source> <translation>Offline (%1 potvrdení)</translation> </message> <message> <location line="+3"/> <source>Unconfirmed (%1 of %2 confirmations)</source> <translation>Nepotvrdené (%1 z %2 potvrdení)</translation> </message> <message> <location line="+3"/> <source>Confirmed (%1 confirmations)</source> <translation>Potvrdené (%1 potvrdení)</translation> </message> <message numerus="yes"> <location line="+8"/> <source>Mined balance will be available when it matures in %n more block(s)</source> <translation type="unfinished"><numerusform></numerusform><numerusform></numerusform><numerusform></numerusform></translation> </message> <message> <location line="+5"/> <source>This block was not received by any other nodes and will probably not be accepted!</source> <translation>Ten blok nebol prijatý žiadnou inou nódou a pravdepodobne nebude akceptovaný!</translation> </message> <message> <location line="+3"/> <source>Generated but not accepted</source> <translation>Vypočítané ale neakceptované</translation> </message> <message> <location line="+43"/> <source>Received with</source> <translation>Prijaté s</translation> </message> <message> <location line="+2"/> <source>Received from</source> <translation>Prijaté od:</translation> </message> <message> <location line="+3"/> <source>Sent to</source> <translation>Odoslané na</translation> </message> <message> <location line="+2"/> <source>Payment to yourself</source> <translation>Platba sebe samému</translation> </message> <message> <location line="+2"/> <source>Mined</source> <translation>Vyfárané</translation> </message> <message> <location line="+38"/> <source>(n/a)</source> <translation>(n/a)</translation> </message> <message> <location line="+199"/> <source>Transaction status. Hover over this field to show number of confirmations.</source> <translation>Status transakcie. Pohybujte myšou nad týmto poľom a zjaví sa počet potvrdení.</translation> </message> <message> <location line="+2"/> <source>Date and time that the transaction was received.</source> <translation>Dátum a čas prijatia transakcie.</translation> </message> <message> <location line="+2"/> <source>Type of transaction.</source> <translation>Typ transakcie.</translation> </message> <message> <location line="+2"/> <source>Destination address of transaction.</source> <translation>Cieľová adresa transakcie.</translation> </message> <message> <location line="+2"/> <source>Amount removed from or added to balance.</source> <translation>Suma pridaná alebo odobraná k zostatku.</translation> </message> </context> <context> <name>TransactionView</name> <message> <location filename="../transactionview.cpp" line="+52"/> <location line="+16"/> <source>All</source> <translation>Všetko</translation> </message> <message> <location line="-15"/> <source>Today</source> <translation>Dnes</translation> </message> <message> <location line="+1"/> <source>This week</source> <translation>Tento týždeň</translation> </message> <message> <location line="+1"/> <source>This month</source> <translation>Tento mesiac</translation> </message> <message> <location line="+1"/> <source>Last month</source> <translation>Minulý mesiac</translation> </message> <message> <location line="+1"/> <source>This year</source> <translation>Tento rok</translation> </message> <message> <location line="+1"/> <source>Range...</source> <translation>Rozsah...</translation> </message> <message> <location line="+11"/> <source>Received with</source> <translation>Prijaté s</translation> </message> <message> <location line="+2"/> <source>Sent to</source> <translation>Odoslané na</translation> </message> <message> <location line="+2"/> <source>To yourself</source> <translation>Samému sebe</translation> </message> <message> <location line="+1"/> <source>Mined</source> <translation>Vyfárané</translation> </message> <message> <location line="+1"/> <source>Other</source> <translation>Iné</translation> </message> <message> <location line="+7"/> <source>Enter address or label to search</source> <translation>Vložte adresu alebo popis pre vyhľadávanie</translation> </message> <message> <location line="+7"/> <source>Min amount</source> <translation>Min množstvo</translation> </message> <message> <location line="+34"/> <source>Copy address</source> <translation>Kopírovať adresu</translation> </message> <message> <location line="+1"/> <source>Copy label</source> <translation>Kopírovať popis</translation> </message> <message> <location line="+1"/> <source>Copy amount</source> <translation>Kopírovať sumu</translation> </message> <message> <location line="+1"/> <source>Copy transaction ID</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Edit label</source> <translation>Editovať popis</translation> </message> <message> <location line="+1"/> <source>Show transaction details</source> <translation type="unfinished"/> </message> <message> <location line="+139"/> <source>Export Transaction Data</source> <translation>Exportovať transakčné dáta</translation> </message> <message> <location line="+1"/> <source>Comma separated file (*.csv)</source> <translation>Čiarkou oddelovaný súbor (*.csv)</translation> </message> <message> <location line="+8"/> <source>Confirmed</source> <translation>Potvrdené</translation> </message> <message> <location line="+1"/> <source>Date</source> <translation>Dátum</translation> </message> <message> <location line="+1"/> <source>Type</source> <translation>Typ</translation> </message> <message> <location line="+1"/> <source>Label</source> <translation>Popis</translation> </message> <message> <location line="+1"/> <source>Address</source> <translation>Adresa</translation> </message> <message> <location line="+1"/> <source>Amount</source> <translation>Suma</translation> </message> <message> <location line="+1"/> <source>ID</source> <translation>ID</translation> </message> <message> <location line="+4"/> <source>Error exporting</source> <translation>Chyba exportu</translation> </message> <message> <location line="+0"/> <source>Could not write to file %1.</source> <translation>Nedalo sa zapisovať do súboru %1.</translation> </message> <message> <location line="+100"/> <source>Range:</source> <translation>Rozsah:</translation> </message> <message> <location line="+8"/> <source>to</source> <translation>do</translation> </message> </context> <context> <name>WalletModel</name> <message> <location filename="../walletmodel.cpp" line="+193"/> <source>Send Coins</source> <translation>Poslať StreamCoins</translation> </message> </context> <context> <name>WalletView</name> <message> <location filename="../walletview.cpp" line="+42"/> <source>&amp;Export</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Export the data in the current tab to a file</source> <translation>Exportovať tento náhľad do súboru</translation> </message> <message> <location line="+193"/> <source>Backup Wallet</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>Wallet Data (*.dat)</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Backup Failed</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>There was an error trying to save the wallet data to the new location.</source> <translation type="unfinished"/> </message> <message> <location line="+4"/> <source>Backup Successful</source> <translation type="unfinished"/> </message> <message> <location line="+0"/> <source>The wallet data was successfully saved to the new location.</source> <translation type="unfinished"/> </message> </context> <context> <name>bitcoin-core</name> <message> <location filename="../bitcoinstrings.cpp" line="+94"/> <source>StreamCoin version</source> <translation>StreamCoin verzia</translation> </message> <message> <location line="+102"/> <source>Usage:</source> <translation>Použitie:</translation> </message> <message> <location line="-29"/> <source>Send command to -server or streamcoind</source> <translation>Odoslať príkaz -server alebo streamcoind</translation> </message> <message> <location line="-23"/> <source>List commands</source> <translation>Zoznam príkazov</translation> </message> <message> <location line="-12"/> <source>Get help for a command</source> <translation>Dostať pomoc pre príkaz</translation> </message> <message> <location line="+24"/> <source>Options:</source> <translation>Možnosti:</translation> </message> <message> <location line="+24"/> <source>Specify configuration file (default: streamcoin.conf)</source> <translation>Určiť súbor s nastaveniami (predvolené: streamcoin.conf)</translation> </message> <message> <location line="+3"/> <source>Specify pid file (default: streamcoind.pid)</source> <translation>Určiť súbor pid (predvolené: streamcoind.pid)</translation> </message> <message> <location line="-1"/> <source>Specify data directory</source> <translation>Určiť priečinok s dátami</translation> </message> <message> <location line="-9"/> <source>Set database cache size in megabytes (default: 25)</source> <translation>Veľkosť vyrovnávajúcej pamäte pre databázu v megabytoch (predvolené:25)</translation> </message> <message> <location line="-28"/> <source>Listen for connections on &lt;port&gt; (default: 9333 or testnet: 19333)</source> <translation>Načúvať spojeniam na &lt;port&gt; (prednastavené: 9333 alebo testovacia sieť: 19333)</translation> </message> <message> <location line="+5"/> <source>Maintain at most &lt;n&gt; connections to peers (default: 125)</source> <translation>Udržiavať maximálne &lt;n&gt; spojení (predvolené: 125)</translation> </message> <message> <location line="-48"/> <source>Connect to a node to retrieve peer addresses, and disconnect</source> <translation type="unfinished"/> </message> <message> <location line="+82"/> <source>Specify your own public address</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Threshold for disconnecting misbehaving peers (default: 100)</source> <translation>Hranica pre odpojenie zle sa správajúcich peerov (predvolené: 100)</translation> </message> <message> <location line="-134"/> <source>Number of seconds to keep misbehaving peers from reconnecting (default: 86400)</source> <translation>Počet sekúnd kedy sa zabráni zle sa správajúcim peerom znovupripojenie (predvolené: 86400)</translation> </message> <message> <location line="-29"/> <source>An error occurred while setting up the RPC port %u for listening on IPv4: %s</source> <translation type="unfinished"/> </message> <message> <location line="+27"/> <source>Listen for JSON-RPC connections on &lt;port&gt; (default: 9332 or testnet: 19332)</source> <translation>Počúvať JSON-RPC spojeniam na &lt;port&gt; (predvolené: 9332 or testnet: 19332)</translation> </message> <message> <location line="+37"/> <source>Accept command line and JSON-RPC commands</source> <translation>Prijímať príkazy z príkazového riadku a JSON-RPC</translation> </message> <message> <location line="+76"/> <source>Run in the background as a daemon and accept commands</source> <translation>Bežať na pozadí ako démon a prijímať príkazy</translation> </message> <message> <location line="+37"/> <source>Use the test network</source> <translation>Použiť testovaciu sieť</translation> </message> <message> <location line="-112"/> <source>Accept connections from outside (default: 1 if no -proxy or -connect)</source> <translation type="unfinished"/> </message> <message> <location line="-80"/> <source>%s, you must set a rpcpassword in the configuration file: %s It is recommended you use the following random password: rpcuser=streamcoinrpc rpcpassword=%s (you do not need to remember this password) The username and password MUST NOT be the same. If the file does not exist, create it with owner-readable-only file permissions. It is also recommended to set alertnotify so you are notified of problems; for example: alertnotify=echo %%s | mail -s &quot;StreamCoin Alert&quot; [email protected] </source> <translation type="unfinished"/> </message> <message> <location line="+17"/> <source>An error occurred while setting up the RPC port %u for listening on IPv6, falling back to IPv4: %s</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Bind to given address and always listen on it. Use [host]:port notation for IPv6</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Cannot obtain a lock on data directory %s. StreamCoin is probably already running.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Error: The transaction was rejected! This might happen if some of the coins in your wallet were already spent, such as if you used a copy of wallet.dat and coins were spent in the copy but not marked as spent here.</source> <translation type="unfinished"/> </message> <message> <location line="+4"/> <source>Error: This transaction requires a transaction fee of at least %s because of its amount, complexity, or use of recently received funds!</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Execute command when a relevant alert is received (%s in cmd is replaced by message)</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Execute command when a wallet transaction changes (%s in cmd is replaced by TxID)</source> <translation type="unfinished"/> </message> <message> <location line="+11"/> <source>Set maximum size of high-priority/low-fee transactions in bytes (default: 27000)</source> <translation type="unfinished"/> </message> <message> <location line="+6"/> <source>This is a pre-release test build - use at your own risk - do not use for mining or merchant applications</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Warning: -paytxfee is set very high! This is the transaction fee you will pay if you send a transaction.</source> <translation>Varovanie: -paytxfee je nastavené veľmi vysoko. Toto sú transakčné poplatky ktoré zaplatíte ak odošlete transakciu.</translation> </message> <message> <location line="+3"/> <source>Warning: Displayed transactions may not be correct! You may need to upgrade, or other nodes may need to upgrade.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Warning: Please check that your computer&apos;s date and time are correct! If your clock is wrong StreamCoin will not work properly.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Warning: error reading wallet.dat! All keys read correctly, but transaction data or address book entries might be missing or incorrect.</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Warning: wallet.dat corrupt, data salvaged! Original wallet.dat saved as wallet.{timestamp}.bak in %s; if your balance or transactions are incorrect you should restore from a backup.</source> <translation type="unfinished"/> </message> <message> <location line="+14"/> <source>Attempt to recover private keys from a corrupt wallet.dat</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Block creation options:</source> <translation type="unfinished"/> </message> <message> <location line="+5"/> <source>Connect only to the specified node(s)</source> <translation>Pripojiť sa len k určenej nóde</translation> </message> <message> <location line="+3"/> <source>Corrupted block database detected</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Discover own IP address (default: 1 when listening and no -externalip)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Do you want to rebuild the block database now?</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Error initializing block database</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Error initializing wallet database environment %s!</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Error loading block database</source> <translation type="unfinished"/> </message> <message> <location line="+4"/> <source>Error opening block database</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Error: Disk space is low!</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Error: Wallet locked, unable to create transaction!</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Error: system error: </source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Failed to listen on any port. Use -listen=0 if you want this.</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Failed to read block info</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Failed to read block</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Failed to sync block index</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Failed to write block index</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Failed to write block info</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Failed to write block</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Failed to write file info</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Failed to write to coin database</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Failed to write transaction index</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Failed to write undo data</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Find peers using DNS lookup (default: 1 unless -connect)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Generate coins (default: 0)</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>How many blocks to check at startup (default: 288, 0 = all)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>How thorough the block verification is (0-4, default: 3)</source> <translation type="unfinished"/> </message> <message> <location line="+19"/> <source>Not enough file descriptors available.</source> <translation type="unfinished"/> </message> <message> <location line="+8"/> <source>Rebuild block chain index from current blk000??.dat files</source> <translation type="unfinished"/> </message> <message> <location line="+16"/> <source>Set the number of threads to service RPC calls (default: 4)</source> <translation type="unfinished"/> </message> <message> <location line="+26"/> <source>Verifying blocks...</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Verifying wallet...</source> <translation type="unfinished"/> </message> <message> <location line="-69"/> <source>Imports blocks from external blk000??.dat file</source> <translation type="unfinished"/> </message> <message> <location line="-76"/> <source>Set the number of script verification threads (up to 16, 0 = auto, &lt;0 = leave that many cores free, default: 0)</source> <translation type="unfinished"/> </message> <message> <location line="+77"/> <source>Information</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Invalid -tor address: &apos;%s&apos;</source> <translation>Neplatná adresa tor: &apos;%s&apos;</translation> </message> <message> <location line="+1"/> <source>Invalid amount for -minrelaytxfee=&lt;amount&gt;: &apos;%s&apos;</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Invalid amount for -mintxfee=&lt;amount&gt;: &apos;%s&apos;</source> <translation type="unfinished"/> </message> <message> <location line="+8"/> <source>Maintain a full transaction index (default: 0)</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Maximum per-connection receive buffer, &lt;n&gt;*1000 bytes (default: 5000)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Maximum per-connection send buffer, &lt;n&gt;*1000 bytes (default: 1000)</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Only accept block chain matching built-in checkpoints (default: 1)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Only connect to nodes in network &lt;net&gt; (IPv4, IPv6 or Tor)</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Output extra debugging information. Implies all other -debug* options</source> <translation>Produkovať extra ladiace informácie. Implies all other -debug* options</translation> </message> <message> <location line="+1"/> <source>Output extra network debugging information</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Prepend debug output with timestamp</source> <translation>Pridať na začiatok ladiaceho výstupu časový údaj</translation> </message> <message> <location line="+5"/> <source>SSL options: (see the StreamCoin Wiki for SSL setup instructions)</source> <translation>SSL možnosť: (pozrite StreamCoin Wiki pre návod na nastavenie SSL)</translation> </message> <message> <location line="+1"/> <source>Select the version of socks proxy to use (4-5, default: 5)</source> <translation type="unfinished"/> </message> <message> <location line="+3"/> <source>Send trace/debug info to console instead of debug.log file</source> <translation>Odoslať trace/debug informácie na konzolu namiesto debug.info žurnálu</translation> </message> <message> <location line="+1"/> <source>Send trace/debug info to debugger</source> <translation>Odoslať trace/debug informácie do ladiaceho programu</translation> </message> <message> <location line="+5"/> <source>Set maximum block size in bytes (default: 250000)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Set minimum block size in bytes (default: 0)</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Shrink debug.log file on client startup (default: 1 when no -debug)</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Signing transaction failed</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Specify connection timeout in milliseconds (default: 5000)</source> <translation>Určiť aut spojenia v milisekundách (predvolené: 5000)</translation> </message> <message> <location line="+4"/> <source>System error: </source> <translation type="unfinished"/> </message> <message> <location line="+4"/> <source>Transaction amount too small</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Transaction amounts must be positive</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Transaction too large</source> <translation type="unfinished"/> </message> <message> <location line="+7"/> <source>Use UPnP to map the listening port (default: 0)</source> <translation>Skúsiť použiť UPnP pre mapovanie počúvajúceho portu (default: 0)</translation> </message> <message> <location line="+1"/> <source>Use UPnP to map the listening port (default: 1 when listening)</source> <translation>Skúsiť použiť UPnP pre mapovanie počúvajúceho portu (default: 1 when listening)</translation> </message> <message> <location line="+1"/> <source>Use proxy to reach tor hidden services (default: same as -proxy)</source> <translation type="unfinished"/> </message> <message> <location line="+2"/> <source>Username for JSON-RPC connections</source> <translation>Užívateľské meno pre JSON-RPC spojenia</translation> </message> <message> <location line="+4"/> <source>Warning</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Warning: This version is obsolete, upgrade required!</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>You need to rebuild the databases using -reindex to change -txindex</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>wallet.dat corrupt, salvage failed</source> <translation type="unfinished"/> </message> <message> <location line="-50"/> <source>Password for JSON-RPC connections</source> <translation>Heslo pre JSON-rPC spojenia</translation> </message> <message> <location line="-67"/> <source>Allow JSON-RPC connections from specified IP address</source> <translation>Povoliť JSON-RPC spojenia z určenej IP adresy.</translation> </message> <message> <location line="+76"/> <source>Send commands to node running on &lt;ip&gt; (default: 127.0.0.1)</source> <translation>Poslať príkaz nóde bežiacej na &lt;ip&gt; (predvolené: 127.0.0.1)</translation> </message> <message> <location line="-120"/> <source>Execute command when the best block changes (%s in cmd is replaced by block hash)</source> <translation>Vykonaj príkaz, ak zmeny v najlepšom bloku (%s v príkaze nahradí blok hash)</translation> </message> <message> <location line="+147"/> <source>Upgrade wallet to latest format</source> <translation>Aktualizuj peňaženku na najnovší formát.</translation> </message> <message> <location line="-21"/> <source>Set key pool size to &lt;n&gt; (default: 100)</source> <translation>Nastaviť zásobu adries na &lt;n&gt; (predvolené: 100)</translation> </message> <message> <location line="-12"/> <source>Rescan the block chain for missing wallet transactions</source> <translation>Znovu skenovať reťaz blokov pre chýbajúce transakcie</translation> </message> <message> <location line="+35"/> <source>Use OpenSSL (https) for JSON-RPC connections</source> <translation>Použiť OpenSSL (https) pre JSON-RPC spojenia</translation> </message> <message> <location line="-26"/> <source>Server certificate file (default: server.cert)</source> <translation>Súbor s certifikátom servra (predvolené: server.cert)</translation> </message> <message> <location line="+1"/> <source>Server private key (default: server.pem)</source> <translation>Súkromný kľúč servra (predvolené: server.pem)</translation> </message> <message> <location line="-151"/> <source>Acceptable ciphers (default: TLSv1+HIGH:!SSLv2:!aNULL:!eNULL:!AH:!3DES:@STRENGTH)</source> <translation>Prijateľné šifry (predvolené: TLSv1+HIGH:!SSLv2:!aNULL:!eNULL:!AH:!3DES:@STRENGTH)</translation> </message> <message> <location line="+165"/> <source>This help message</source> <translation>Táto pomocná správa</translation> </message> <message> <location line="+6"/> <source>Unable to bind to %s on this computer (bind returned error %d, %s)</source> <translation type="unfinished"/> </message> <message> <location line="-91"/> <source>Connect through socks proxy</source> <translation>Pripojenie cez socks proxy</translation> </message> <message> <location line="-10"/> <source>Allow DNS lookups for -addnode, -seednode and -connect</source> <translation>Povoliť vyhľadávanie DNS pre pridanie nódy a spojenie</translation> </message> <message> <location line="+55"/> <source>Loading addresses...</source> <translation>Načítavanie adries...</translation> </message> <message> <location line="-35"/> <source>Error loading wallet.dat: Wallet corrupted</source> <translation>Chyba načítania wallet.dat: Peňaženka je poškodená</translation> </message> <message> <location line="+1"/> <source>Error loading wallet.dat: Wallet requires newer version of StreamCoin</source> <translation>Chyba načítania wallet.dat: Peňaženka vyžaduje novšiu verziu StreamCoin</translation> </message> <message> <location line="+93"/> <source>Wallet needed to be rewritten: restart StreamCoin to complete</source> <translation>Bolo potrebné prepísať peňaženku: dokončite reštartovaním StreamCoin</translation> </message> <message> <location line="-95"/> <source>Error loading wallet.dat</source> <translation>Chyba načítania wallet.dat</translation> </message> <message> <location line="+28"/> <source>Invalid -proxy address: &apos;%s&apos;</source> <translation>Neplatná adresa proxy: &apos;%s&apos;</translation> </message> <message> <location line="+56"/> <source>Unknown network specified in -onlynet: &apos;%s&apos;</source> <translation type="unfinished"/> </message> <message> <location line="-1"/> <source>Unknown -socks proxy version requested: %i</source> <translation type="unfinished"/> </message> <message> <location line="-96"/> <source>Cannot resolve -bind address: &apos;%s&apos;</source> <translation type="unfinished"/> </message> <message> <location line="+1"/> <source>Cannot resolve -externalip address: &apos;%s&apos;</source> <translation type="unfinished"/> </message> <message> <location line="+44"/> <source>Invalid amount for -paytxfee=&lt;amount&gt;: &apos;%s&apos;</source> <translation>Neplatná suma pre -paytxfee=&lt;amount&gt;: &apos;%s&apos;</translation> </message> <message> <location line="+1"/> <source>Invalid amount</source> <translation>Neplatná suma</translation> </message> <message> <location line="-6"/> <source>Insufficient funds</source> <translation>Nedostatok prostriedkov</translation> </message> <message> <location line="+10"/> <source>Loading block index...</source> <translation>Načítavanie zoznamu blokov...</translation> </message> <message> <location line="-57"/> <source>Add a node to connect to and attempt to keep the connection open</source> <translation>Pridať nód na pripojenie a pokus o udržanie pripojenia otvoreného</translation> </message> <message> <location line="-25"/> <source>Unable to bind to %s on this computer. StreamCoin is probably already running.</source> <translation type="unfinished"/> </message> <message> <location line="+64"/> <source>Fee per KB to add to transactions you send</source> <translation>Poplatok za kB ktorý treba pridať k odoslanej transakcii</translation> </message> <message> <location line="+19"/> <source>Loading wallet...</source> <translation>Načítavam peňaženku...</translation> </message> <message> <location line="-52"/> <source>Cannot downgrade wallet</source> <translation>Nie je možné prejsť na nižšiu verziu peňaženky</translation> </message> <message> <location line="+3"/> <source>Cannot write default address</source> <translation>Nie je možné zapísať predvolenú adresu.</translation> </message> <message> <location line="+64"/> <source>Rescanning...</source> <translation type="unfinished"/> </message> <message> <location line="-57"/> <source>Done loading</source> <translation>Dokončené načítavanie</translation> </message> <message> <location line="+82"/> <source>To use the %s option</source> <translation>Použiť %s možnosť.</translation> </message> <message> <location line="-74"/> <source>Error</source> <translation>Chyba</translation> </message> <message> <location line="-31"/> <source>You must set rpcpassword=&lt;password&gt; in the configuration file: %s If the file does not exist, create it with owner-readable-only file permissions.</source> <translation>Musíš nastaviť rpcpassword=&lt;heslo&gt; v konfiguračnom súbore: %s Ak súbor neexistuje, vytvor ho s oprávnením pre čítanie len vlastníkom (owner-readable-only)</translation> </message> </context> </TS><|fim▁end|>
<message> <location line="+14"/> <source>&amp;Delete</source>
<|file_name|>_9_1_geometry_shader_houses.rs<|end_file_name|><|fim▁begin|>#![allow(non_upper_case_globals)] #![allow(non_snake_case)] use std::ptr; use std::mem; use std::os::raw::c_void; extern crate glfw; use self::glfw::Context; extern crate gl; use self::gl::types::*; use cgmath::{Point3}; use common::{process_events, processInput}; use shader::Shader; use camera::Camera; // settings const SCR_WIDTH: u32 = 1280; const SCR_HEIGHT: u32 = 720; pub fn main_4_9_1() { let mut camera = Camera { Position: Point3::new(0.0, 0.0, 3.0), ..Camera::default() }; let mut firstMouse = true; let mut lastX: f32 = SCR_WIDTH as f32 / 2.0; let mut lastY: f32 = SCR_HEIGHT as f32 / 2.0; // timing let mut deltaTime: f32; // time between current frame and last frame let mut lastFrame: f32 = 0.0; // glfw: initialize and configure // ------------------------------ let mut glfw = glfw::init(glfw::FAIL_ON_ERRORS).unwrap(); glfw.window_hint(glfw::WindowHint::ContextVersion(3, 3)); glfw.window_hint(glfw::WindowHint::OpenGlProfile(glfw::OpenGlProfileHint::Core));<|fim▁hole|> // glfw window creation // -------------------- let (mut window, events) = glfw.create_window(SCR_WIDTH, SCR_HEIGHT, "LearnOpenGL", glfw::WindowMode::Windowed) .expect("Failed to create GLFW window"); window.make_current(); window.set_framebuffer_size_polling(true); window.set_cursor_pos_polling(true); window.set_scroll_polling(true); // tell GLFW to capture our mouse window.set_cursor_mode(glfw::CursorMode::Disabled); // gl: load all OpenGL function pointers // --------------------------------------- gl::load_with(|symbol| window.get_proc_address(symbol) as *const _); let (shader, VBO, VAO) = unsafe { // configure global opengl state // ----------------------------- gl::Enable(gl::DEPTH_TEST); // build and compile shaders // ------------------------- let shader = Shader::with_geometry_shader( "src/_4_advanced_opengl/shaders/9.1.geometry_shader.vs", "src/_4_advanced_opengl/shaders/9.1.geometry_shader.fs", "src/_4_advanced_opengl/shaders/9.1.geometry_shader.gs" ); // set up vertex data (and buffer(s)) and configure vertex attributes // ------------------------------------------------------------------ let points: [f32; 20] = [ -0.5, 0.5, 1.0, 0.0, 0.0, // top-left 0.5, 0.5, 0.0, 1.0, 0.0, // top-right 0.5, -0.5, 0.0, 0.0, 1.0, // bottom-right -0.5, -0.5, 1.0, 1.0, 0.0 // bottom-left ]; // cube VAO let (mut VAO, mut VBO) = (0, 0); gl::GenVertexArrays(1, &mut VAO); gl::GenBuffers(1, &mut VBO); gl::BindVertexArray(VAO); gl::BindBuffer(gl::ARRAY_BUFFER, VBO); gl::BufferData(gl::ARRAY_BUFFER, (points.len() * mem::size_of::<GLfloat>()) as GLsizeiptr, &points[0] as *const f32 as *const c_void, gl::STATIC_DRAW); let stride = 5 * mem::size_of::<GLfloat>() as GLsizei; gl::EnableVertexAttribArray(0); gl::VertexAttribPointer(0, 2, gl::FLOAT, gl::FALSE, stride, ptr::null()); gl::EnableVertexAttribArray(1); gl::VertexAttribPointer(1, 3, gl::FLOAT, gl::FALSE, stride, (2 * mem::size_of::<GLfloat>()) as *const f32 as *const c_void); gl::BindVertexArray(0); (shader, VBO, VAO) }; // render loop // ----------- while !window.should_close() { // per-frame time logic // -------------------- let currentFrame = glfw.get_time() as f32; deltaTime = currentFrame - lastFrame; lastFrame = currentFrame; // events // ----- process_events(&events, &mut firstMouse, &mut lastX, &mut lastY, &mut camera); // input // ----- processInput(&mut window, deltaTime, &mut camera); // render // ------ unsafe { gl::ClearColor(0.1, 0.1, 0.1, 1.0); gl::Clear(gl::COLOR_BUFFER_BIT | gl::DEPTH_BUFFER_BIT); shader.useProgram(); gl::BindVertexArray(VAO); gl::DrawArrays(gl::POINTS, 0, 4); } // glfw: swap buffers and poll IO events (keys pressed/released, mouse moved etc.) // ------------------------------------------------------------------------------- window.swap_buffers(); glfw.poll_events(); } // optional: de-allocate all resources once they've outlived their purpose: // ------------------------------------------------------------------------ unsafe { gl::DeleteVertexArrays(1, &VAO); gl::DeleteBuffers(1, &VBO); } }<|fim▁end|>
#[cfg(target_os = "macos")] glfw.window_hint(glfw::WindowHint::OpenGlForwardCompat(true));
<|file_name|>h_third_party_apps.py<|end_file_name|><|fim▁begin|>try:<|fim▁hole|> INSTALLED_APPS except NameError: INSTALLED_APPS=() #Generated Config - Don't modify above this line<|fim▁end|>
<|file_name|>memory.rs<|end_file_name|><|fim▁begin|>// This file is part of zinc64. // Copyright (c) 2016-2019 Sebastian Jastrzebski. All rights reserved. // Licensed under the GPLv3. See LICENSE file in the project root for full license text. #[cfg(not(feature = "std"))] use alloc::prelude::*; use zinc64_core::{Addressable, AddressableFaded, Bank, Mmu, Ram, Rom, Shared}; use crate::mem::{BaseAddr, Mmio}; // Spec: COMMODORE 64 MEMORY MAPS p. 263 // Design: // Inspired by UAE memory address64k/bank concepts. // We define Addressable trait to represent a bank of memory and use memory configuration // based on zones that can be mapped to different banks. CPU uses IoPort @ 0x0001 to reconfigure // memory layout. pub struct Memory { mmu: Shared<dyn Mmu>, expansion_port: Shared<dyn AddressableFaded>, io: Mmio, ram: Shared<Ram>, basic: Shared<Rom>, charset: Shared<Rom>,<|fim▁hole|>} impl Memory { pub fn new( mmu: Shared<dyn Mmu>, expansion_port: Shared<dyn AddressableFaded>, io: Mmio, ram: Shared<Ram>, rom_basic: Shared<Rom>, rom_charset: Shared<Rom>, rom_kernal: Shared<Rom>, ) -> Self { Memory { mmu, expansion_port, io, ram, basic: rom_basic, charset: rom_charset, kernal: rom_kernal, } } } impl Addressable for Memory { fn read(&self, address: u16) -> u8 { let bank = self.mmu.borrow().map(address); match bank { Bank::Ram => self.ram.borrow().read(address), Bank::Basic => self.basic.borrow().read(address), Bank::Charset => self .charset .borrow() .read(address - BaseAddr::Charset.addr()), Bank::Kernal => self.kernal.borrow().read(address), Bank::RomL => self .expansion_port .borrow_mut() .read(address) .unwrap_or(self.ram.borrow().read(address)), Bank::RomH => self .expansion_port .borrow_mut() .read(address) .unwrap_or(self.ram.borrow().read(address)), Bank::Io => self.io.read(address), Bank::Disabled => 0, } } fn write(&mut self, address: u16, value: u8) { let bank = self.mmu.borrow().map(address); match bank { Bank::Ram => self.ram.borrow_mut().write(address, value), Bank::Basic => self.ram.borrow_mut().write(address, value), Bank::Charset => self.ram.borrow_mut().write(address, value), Bank::Kernal => self.ram.borrow_mut().write(address, value), Bank::RomL => self.ram.borrow_mut().write(address, value), Bank::RomH => self.ram.borrow_mut().write(address, value), Bank::Io => self.io.write(address, value), Bank::Disabled => {} } } } #[cfg(test)] mod tests { /* FIXME nostd: enable test use super::*; use zinc64_core::{new_shared, Addressable, Ram, Rom}; impl Addressable for Ram { fn read(&self, address: u16) -> u8 { self.read(address) } fn write(&mut self, address: u16, value: u8) { self.write(address, value); } } fn setup_memory() -> Memory { let basic = new_shared(Rom::new(0x1000, BaseAddr::Basic.addr(), 0x10)); let charset = new_shared(Rom::new(0x1000, 0x0000, 0x11)); let kernal = new_shared(Rom::new(0x1000, BaseAddr::Kernal.addr(), 0x12)); let mut mmio = Box::new(Ram::new(0x10000)); mmio.fill(0x22); let expansion_port = new_shared(Ram::new(0x1000)); expansion_port.borrow_mut().fill(0x33); let ram = new_shared(Ram::new(0x10000)); ram.borrow_mut().fill(0x44); Memory::new(expansion_port, mmio, ram, basic, charset, kernal) } #[test] fn read_basic() { let mut mem = setup_memory(); mem.switch_banks(31); assert_eq!(0x10, mem.read(BaseAddr::Basic.addr())); } #[test] fn read_charset() { let mut mem = setup_memory(); mem.switch_banks(27); assert_eq!(0x11, mem.read(BaseAddr::Charset.addr())); } #[test] fn read_io() { let mut mem = setup_memory(); mem.switch_banks(31); assert_eq!(0x22, mem.read(0xd000)); } #[test] fn read_kernal() { let mut mem = setup_memory(); mem.switch_banks(31); assert_eq!(0x12, mem.read(BaseAddr::Kernal.addr())); } #[test] fn write_page_0() { let mut mem = setup_memory(); mem.write(0x00f0, 0xff); assert_eq!(0xff, mem.ram.borrow().read(0x00f0)); } #[test] fn write_page_1() { let mut mem = setup_memory(); mem.write(0x0100, 0xff); assert_eq!(0xff, mem.ram.borrow().read(0x0100)); } */ }<|fim▁end|>
kernal: Shared<Rom>,
<|file_name|>ShortTimeProcessImpl.cpp<|end_file_name|><|fim▁begin|>/* * ShortTimeProcess.cpp * Copyright 2016 (c) Jordi Adell * Created on: 2015 * Author: Jordi Adell - [email protected] * * This file is part of DSPONE * * DSPONE is free software: you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * DSPONE is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * alogn with DSPONE. If not, see <http://www.gnu.org/licenses/>. */ #include <dspone/rt/ShortTimeProcessImpl.h> #include <dspone/rt/ShortTimeProcess.h> namespace dsp { ShortTimeProcessImpl::ShortTimeProcessImpl(ShortTimeProcess *frameProcessor, int windowSize, int analysisLength, int nchannels, Mode mode) : _frameProcessor(frameProcessor), _windowSize (windowSize), _windowShift (_windowSize/2), _halfWindowSize (_windowSize/2), _nchannels(nchannels), _nDataChannels(10), _analysisLength((analysisLength == 0) ? _windowSize + 2 : analysisLength), _window(new BaseType[_windowSize+1]), // Needs to be of odd length in order to preserve the original energy. _iwindow(new BaseType[_windowSize]), _mode(mode), _doSynthesis(_mode == ShortTimeProcess::ANALYSIS_SYNTHESIS) { initVariableMembers(); initWindowBuffers(); initBuffers(); } double ShortTimeProcessImpl::getRate() const { return _windowShift; } int ShortTimeProcessImpl::getNumberOfChannels() const { return _nchannels; } int ShortTimeProcessImpl::getNumberOfDataChannels() const { return _nDataChannels; } ShortTimeProcessImpl::~ShortTimeProcessImpl() { } void ShortTimeProcessImpl::initVariableMembers() { _firstCall = true; } void ShortTimeProcessImpl::initWindowBuffers() { // Create auxiliar buffers _frame.reset(new BaseType[_windowSize]); // Setting window vector wipp::set(1.0,_window.get(),_windowSize+1); wipp::window(_window.get(), _windowSize+1, wipp::wippHANN); wipp::sqrt(&_window[1],_windowSize-1); _window[0]=0; // Setting inverse window vector BaseType ones[_windowSize]; wipp::set(1.0, ones, _windowSize); wipp::copyBuffer(_window.get(),_iwindow.get(),_windowSize); wipp::div(&_window[1], ones, &_iwindow[1], _windowSize-1); } void ShortTimeProcessImpl::initBuffers() { wipp::setZeros(_frame.get(),_windowSize); for (unsigned int i = 0; i<_nchannels; ++i) { _latencyBufferProcessed.push_back(SignalPtr(new BaseType[_windowSize])); wipp::wipp_circular_buffer_t *cb; wipp::init_cirular_buffer(&cb, _maximumLatencyBufferSize, _frame.get(), _windowSize); _latencyBufferSignal.push_back(cb); wipp::setZeros(_latencyBufferProcessed.back().get(), _windowSize); _analysisFramesPtr.push_back(new BaseType[_analysisLength]); _analysisFrames.push_back(SignalPtr(_analysisFramesPtr.back())); wipp::setZeros(_analysisFrames.back().get(), _analysisLength); } allocateNDataChannels(_nDataChannels); } void ShortTimeProcessImpl::allocateNDataChannels(int nDataChannels) { for (unsigned int i = _dataFrames.size(); i < nDataChannels; ++i) { _dataFramesPtr.push_back(new BaseType[_windowSize]); _dataFrames.push_back(SignalPtr(_dataFramesPtr.back())); wipp::setZeros(_dataFrames.back().get(), _windowSize); } for (unsigned int i = _latencyBufferSignal.size(); i < nDataChannels + _nchannels; ++i) { BaseType zeros[_windowSize]; wipp::setZeros(zeros, _windowSize); // _latencyBufferSignal.push_back(container::CircularBuffer<BaseType, _maximumLatencyBufferSize>()); _latencyBufferProcessed.push_back(SignalPtr(new BaseType[_windowSize])); wipp::setZeros(_latencyBufferProcessed.back().get(), _windowSize); wipp::wipp_circular_buffer_t *cb; wipp::init_cirular_buffer(&cb, _maximumLatencyBufferSize, zeros, _windowSize); _latencyBufferSignal.push_back(cb); } } int ShortTimeProcessImpl::getAmountOfRemainingSamples() { if (_firstCall) return 0; else return _windowShift; } void ShortTimeProcessImpl::clear() { _latencyBufferSignal.clear(); _analysisFrames.clear(); _analysisFramesPtr.clear(); for (size_t i = 0; i < _latencyBufferSignal.size(); ++i) { wipp::wipp_circular_buffer_t *f = _latencyBufferSignal.at(i); wipp::delete_circular_buffer(&f); } _dataFrames.clear(); _dataFramesPtr.clear(); _latencyBufferProcessed.clear(); initBuffers(); } void ShortTimeProcessImpl::ShortTimeProcessing(const SignalVector &signal, const SignalVector &output, int length) { std::string msg = "ShortTimeProcess process frame of "; msg += std::to_string(_windowSize) + " samples"; int lastFrame = length - _windowSize; int sample = 0; SignalVector::const_iterator it; for (it = output.begin(); it != output.end(); ++it) wipp::setZeros(it->get(), length); for (sample=0; sample <= lastFrame; sample = sample + _windowShift) { unsigned int channel; for (channel = 0, it = signal.begin(); // Foreach signal channel it != signal.end() && channel < _nchannels; ++it, ++channel) { BaseType *ptr = it->get(); wipp::copyBuffer(&ptr[sample],_frame.get(),_windowSize); wipp::mult(_window.get(),_frame.get(),_windowSize); frameAnalysis(_frame.get(), _analysisFrames[channel].get(), _windowSize, _analysisLength, channel); } for(channel = 0; // Foreach data channel it != signal.end() && channel < _nDataChannels; ++it, ++channel) { BaseType *ptr = it->get(); wipp::copyBuffer(&ptr[sample], _dataFrames[channel].get(), _windowSize); } /// Set data channels point to the actual signal channels. processParametrisation(); if (!_doSynthesis) continue; for (channel = 0, it = output.begin(); // Foreach signal channel it != output.end() && channel < _nchannels; ++it, ++channel) { BaseType *ptr =it->get(); frameSynthesis(_frame.get(), _analysisFrames[channel].get(), _windowSize, _analysisLength, channel); wipp::mult(_window.get(),_frame.get(),_windowSize); wipp::add(_frame.get(),&ptr[sample],_windowSize); } for(channel = 0; // Foreach data channel it != output.end() && channel < _nDataChannels; ++it, ++channel) { BaseType *ptr = it->get(); wipp::copyBuffer(_dataFrames[channel].get(), &ptr[sample], _windowSize); } } } int ShortTimeProcessImpl::calculateOrderFromSampleRate(int sampleRate, double frameRate) { int order = static_cast<int>(log(2*frameRate*sampleRate)/log(2.0F) + 0.5); if (0 < order && order < 4) { ERROR_STREAM("Too low order for FFT: " << order << " this might cose undefined errors."); } else if(order <= 0) { std::stringstream oss; oss << "Negative or zero FFT order value: " << order << ". " << "Order is optained from frame rate: " << frameRate << " and sample rate: " << sampleRate; throw(DspException(oss.str())); } return order; } void ShortTimeProcessImpl::unwindowFrame(double *frame, size_t length) const { wipp::mult(_iwindow.get(), frame, ((length < _windowSize) ? length : _windowSize)); } void ShortTimeProcessImpl::unwindowFrame(double *frame, double *unwindowed, size_t length) const { wipp::setZeros(unwindowed, length); wipp::copyBuffer(frame, unwindowed, length); wipp::mult(_iwindow.get(), frame, ((length < _windowSize) ? length : _windowSize)); } int ShortTimeProcessImpl::getLatency() const { size_t occupancy; // return _latencyBufferSignal[0].getCountUsed(); wipp::cf_occupancy(_latencyBufferSignal[0], &occupancy); return occupancy; } int ShortTimeProcessImpl::getMaxLatency() const { return getWindowSize() + _windowShift; } int ShortTimeProcessImpl::getMinLatency() const<|fim▁hole|>int ShortTimeProcessImpl::getAnalysisLength() const { return _analysisLength; } int ShortTimeProcessImpl::getWindowSize() const { return _windowSize; } int ShortTimeProcessImpl::getFrameSize() const { return getWindowSize(); } int ShortTimeProcessImpl::getFrameRate() const { return _windowShift; } void ShortTimeProcessImpl::frameAnalysis(BaseType *inFrame, BaseType *analysis, int frameLength, int analysisLength, int channel) { _qtdebug.plot(inFrame, frameLength, QtDebug::IN_FRAME, channel); _frameProcessor->frameAnalysis(inFrame, analysis, frameLength, analysisLength, channel); } void ShortTimeProcessImpl::processParametrisation() { _qtdebug.plot(_analysisFrames, _analysisLength, QtDebug::IN_ANALYSIS); _frameProcessor->processParametrisation(_analysisFramesPtr, _analysisLength, _dataFramesPtr, _windowSize); _qtdebug.plot(_analysisFrames, _analysisLength, QtDebug::OUT_ANALYSIS); } void ShortTimeProcessImpl::frameSynthesis(BaseType *outFrame, BaseType *analysis, int frameLength, int analysisLength, int channel) { _frameProcessor->frameSynthesis(outFrame, analysis, frameLength, analysisLength, channel); _qtdebug.plot(outFrame, frameLength, QtDebug::OUT_FRAME, channel); } }<|fim▁end|>
{ return _windowShift; }
<|file_name|>digits_test.go<|end_file_name|><|fim▁begin|>package digits import ( "net/http" "net/http/httptest" "net/url" "testing" "github.com/stretchr/testify/assert" ) // testServer returns an http Client, ServeMux, and Server. The client proxies // requests to the server and handlers can be registered on the mux to handle // requests. The caller must close the test server. func testServer() (*http.Client, *http.ServeMux, *httptest.Server) { mux := http.NewServeMux() server := httptest.NewServer(mux) transport := &RewriteTransport{&http.Transport{ Proxy: func(req *http.Request) (*url.URL, error) { return url.Parse(server.URL) }, }} client := &http.Client{Transport: transport} return client, mux, server } // RewriteTransport rewrites https requests to http to avoid TLS cert issues // during testing. type RewriteTransport struct { Transport http.RoundTripper } // RoundTrip rewrites the request scheme to http and calls through to the // composed RoundTripper or if it is nil, to the http.DefaultTransport. func (t *RewriteTransport) RoundTrip(req *http.Request) (*http.Response, error) { req.URL.Scheme = "http" if t.Transport == nil {<|fim▁hole|>} // assertMethod tests that the Request has the expected HTTP method. func assertMethod(t *testing.T, expectedMethod string, req *http.Request) { assert.Equal(t, expectedMethod, req.Method) } // assertQuery tests that the Request has the expected url query key/val pairs func assertQuery(t *testing.T, expected map[string]string, req *http.Request) { queryValues := req.URL.Query() expectedValues := url.Values{} for key, value := range expected { expectedValues.Add(key, value) } assert.Equal(t, expectedValues, queryValues) }<|fim▁end|>
return http.DefaultTransport.RoundTrip(req) } return t.Transport.RoundTrip(req)
<|file_name|>imagenet.py<|end_file_name|><|fim▁begin|># ---------------------------------------------------------------------- # Numenta Platform for Intelligent Computing (NuPIC) # Copyright (C) 2020, Numenta, Inc. Unless you have an agreement # with Numenta, Inc., for a separate license for this software code, the # following terms and conditions apply: # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero Public License version 3 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU Affero Public License for more details. # # You should have received a copy of the GNU Affero Public License # along with this program. If not, see http://www.gnu.org/licenses. # # http://numenta.org/licenses/ # ---------------------------------------------------------------------- import os from torchvision import transforms from torchvision.transforms import RandomResizedCrop import nupic.research.frameworks.pytorch.dataset_utils.auto_augment as aa from nupic.research.frameworks.pytorch.dataset_utils import HDF5Dataset from nupic.research.frameworks.pytorch.datasets.imagenet_factory import ( IMAGENET_NUM_CLASSES as IMAGENET_CLASS_SUBSETS, ) class ImageNet100(object): def __init__(self, use_auto_augment=False): self.use_auto_augment = use_auto_augment self.train_dataset = None self.test_dataset = None def get_train_dataset(self, iteration): if self.train_dataset is None: if self.use_auto_augment: transform = transforms.Compose( transforms=[ RandomResizedCrop(224), transforms.RandomHorizontalFlip(), aa.ImageNetPolicy(), transforms.ToTensor(), transforms.Normalize( mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225], inplace=True ), ], ) else: transform = transforms.Compose( transforms=[ RandomResizedCrop(224), transforms.RandomHorizontalFlip(), transforms.ToTensor(), transforms.Normalize( mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225],<|fim▁hole|> ) self.train_dataset = HDF5Dataset( hdf5_file=os.path.expanduser("~/nta/data/imagenet/imagenet.hdf5"), root="train", classes=IMAGENET_CLASS_SUBSETS[100], transform=transform) return self.train_dataset def get_test_dataset(self, noise_level=0.0): assert noise_level == 0.0 if self.test_dataset is None: transform = transforms.Compose( [ transforms.Resize(256), transforms.CenterCrop(224), transforms.ToTensor(), transforms.Normalize( mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225], inplace=True ), ] ) self.test_dataset = HDF5Dataset( hdf5_file=os.path.expanduser("~/nta/data/imagenet/imagenet.hdf5"), root="val", classes=IMAGENET_CLASS_SUBSETS[100], transform=transform) return self.test_dataset<|fim▁end|>
inplace=True ), ],
<|file_name|>s3query.py<|end_file_name|><|fim▁begin|># -*- coding: utf-8 -*- """ S3 Query Construction @copyright: 2009-2015 (c) Sahana Software Foundation @license: MIT Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. """ __all__ = ("FS", "S3FieldSelector", "S3Joins", "S3ResourceField", "S3ResourceQuery", "S3URLQuery", "S3URLQueryParser", ) import datetime import re import sys from gluon import current from gluon.storage import Storage from s3dal import Field, Row from s3fields import S3RepresentLazy from s3utils import s3_get_foreign_key, s3_unicode, S3TypeConverter ogetattr = object.__getattribute__ TEXTTYPES = ("string", "text") # ============================================================================= class S3FieldSelector(object): """ Helper class to construct a resource query """ LOWER = "lower" UPPER = "upper" OPERATORS = [LOWER, UPPER] def __init__(self, name, type=None): """ Constructor """ if not isinstance(name, basestring) or not name: raise SyntaxError("name required") self.name = str(name) self.type = type self.op = None # ------------------------------------------------------------------------- def __lt__(self, value): return S3ResourceQuery(S3ResourceQuery.LT, self, value) # ------------------------------------------------------------------------- def __le__(self, value): return S3ResourceQuery(S3ResourceQuery.LE, self, value) # ------------------------------------------------------------------------- def __eq__(self, value): return S3ResourceQuery(S3ResourceQuery.EQ, self, value) # ------------------------------------------------------------------------- def __ne__(self, value): return S3ResourceQuery(S3ResourceQuery.NE, self, value) # ------------------------------------------------------------------------- def __ge__(self, value): return S3ResourceQuery(S3ResourceQuery.GE, self, value) # ------------------------------------------------------------------------- def __gt__(self, value): return S3ResourceQuery(S3ResourceQuery.GT, self, value) # ------------------------------------------------------------------------- def like(self, value): return S3ResourceQuery(S3ResourceQuery.LIKE, self, value) # ------------------------------------------------------------------------- def belongs(self, value): return S3ResourceQuery(S3ResourceQuery.BELONGS, self, value) # ------------------------------------------------------------------------- def contains(self, value): return S3ResourceQuery(S3ResourceQuery.CONTAINS, self, value) # ------------------------------------------------------------------------- def anyof(self, value): return S3ResourceQuery(S3ResourceQuery.ANYOF, self, value) # ------------------------------------------------------------------------- def typeof(self, value): return S3ResourceQuery(S3ResourceQuery.TYPEOF, self, value) # ------------------------------------------------------------------------- def lower(self): self.op = self.LOWER return self # ------------------------------------------------------------------------- def upper(self): self.op = self.UPPER return self # ------------------------------------------------------------------------- def expr(self, val): if self.op and val is not None: if self.op == self.LOWER and \ hasattr(val, "lower") and callable(val.lower) and \ (not isinstance(val, Field) or val.type in TEXTTYPES): return val.lower() elif self.op == self.UPPER and \ hasattr(val, "upper") and callable(val.upper) and \ (not isinstance(val, Field) or val.type in TEXTTYPES): return val.upper() return val # ------------------------------------------------------------------------- def represent(self, resource): try: rfield = S3ResourceField(resource, self.name) except: colname = None else: colname = rfield.colname if colname: if self.op is not None: return "%s.%s()" % (colname, self.op) else: return colname else: return "(%s?)" % self.name # ------------------------------------------------------------------------- @classmethod def extract(cls, resource, row, field): """ Extract a value from a Row @param resource: the resource @param row: the Row @param field: the field @return: field if field is not a Field/S3FieldSelector instance, the value from the row otherwise """ error = lambda fn: KeyError("Field not found: %s" % fn) t = type(field) if isinstance(field, Field): colname = str(field) tname, fname = colname.split(".", 1) elif t is S3FieldSelector: rfield = S3ResourceField(resource, field.name) colname = rfield.colname if not colname: # unresolvable selector raise error(field.name) fname = rfield.fname tname = rfield.tname elif t is S3ResourceField: colname = field.colname if not colname: # unresolved selector return None fname = field.fname tname = field.tname else: return field if type(row) is Row: try: if tname in row.__dict__: value = ogetattr(ogetattr(row, tname), fname) else: value = ogetattr(row, fname) except: try: value = row[colname] except (KeyError, AttributeError): raise error(colname) elif fname in row: value = row[fname] elif colname in row: value = row[colname] elif tname is not None and \ tname in row and fname in row[tname]: value = row[tname][fname] else: raise error(colname) if callable(value): # Lazy virtual field try: value = value() except: current.log.error(sys.exc_info()[1]) value = None if hasattr(field, "expr"): return field.expr(value) return value # ------------------------------------------------------------------------- def resolve(self, resource): """ Resolve this field against a resource @param resource: the resource """ return S3ResourceField(resource, self.name) # ============================================================================= # Short name for the S3FieldSelector class # FS = S3FieldSelector # ============================================================================= class S3FieldPath(object): """ Helper class to parse field selectors """ # ------------------------------------------------------------------------- @classmethod def resolve(cls, resource, selector, tail=None): """ Resolve a selector (=field path) against a resource @param resource: the S3Resource to resolve against @param selector: the field selector string @param tail: tokens to append to the selector The general syntax for a selector is: selector = {[alias].}{[key]$}[field|selector] (Parts in {} are optional, | indicates alternatives) * Alias can be: ~ refers to the resource addressed by the preceding parts of the selector (=last resource) component alias of a component of the last resource linktable alias of a link table of the last resource table name of a table that has a foreign key for the last resource (auto-detect the key) key:table same as above, but specifying the foreign key * Key can be: key the name of a foreign key in the last resource context a context expression * Field can be: fieldname the name of a field or virtual field of the last resource context a context expression A "context expression" is a name enclosed in parentheses: (context) During parsing, context expressions get replaced by the string which has been configured for this name for the last resource with: s3db.configure(tablename, context = dict(name = "string")) With context expressions, the same selector can be used for different resources, each time resolving into the specific field path. However, the field addressed must be of the same type in all resources to form valid queries. If a context name can not be resolved, resolve() will still succeed - but the S3FieldPath returned will have colname=None and ftype="context" (=unresolvable context). """ if not selector: raise SyntaxError("Invalid selector: %s" % selector) tokens = re.split("(\.|\$)", selector) if tail: tokens.extend(tail) parser = cls(resource, None, tokens) parser.original = selector return parser # ------------------------------------------------------------------------- def __init__(self, resource, table, tokens): """ Constructor - not to be called directly, use resolve() instead @param resource: the S3Resource @param table: the table @param tokens: the tokens as list """ s3db = current.s3db if table is None: table = resource.table # Initialize self.original = None self.tname = table._tablename self.fname = None self.field = None self.ftype = None self.virtual = False self.colname = None self.joins = {} self.distinct = False self.multiple = True head = tokens.pop(0) tail = None if head and head[0] == "(" and head[-1] == ")": # Context expression head = head.strip("()") self.fname = head self.ftype = "context" if not resource: resource = s3db.resource(table, components=[]) context = resource.get_config("context") if context and head in context: tail = self.resolve(resource, context[head], tail=tokens) else: # unresolvable pass elif tokens: # Resolve the tail op = tokens.pop(0) if tokens: if op == ".": # head is a component or linktable alias, and tokens is # a field expression in the component/linked table if not resource: resource = s3db.resource(table, components=[]) ktable, join, m, d = self._resolve_alias(resource, head) self.multiple = m self.distinct = d else: # head is a foreign key in the current table and tokens is # a field expression in the referenced table ktable, join = self._resolve_key(table, head) self.distinct = True if join is not None: self.joins[ktable._tablename] = join tail = S3FieldPath(None, ktable, tokens) else: raise SyntaxError("trailing operator") if tail is None: # End of the expression if self.ftype != "context": # Expression is resolved, head is a field name: self.field = self._resolve_field(table, head) if not self.field: self.virtual = True self.ftype = "virtual" else: self.virtual = False self.ftype = str(self.field.type) self.fname = head self.colname = "%s.%s" % (self.tname, self.fname) else: # Read field data from tail self.tname = tail.tname self.fname = tail.fname self.field = tail.field self.ftype = tail.ftype self.virtual = tail.virtual self.colname = tail.colname self.distinct |= tail.distinct self.multiple |= tail.multiple self.joins.update(tail.joins) # ------------------------------------------------------------------------- @staticmethod def _resolve_field(table, fieldname): """ Resolve a field name against the table, recognizes "id" as table._id.name, and "uid" as current.xml.UID. @param table: the Table @param fieldname: the field name @return: the Field """ if fieldname == "uid": fieldname = current.xml.UID if fieldname == "id": field = table._id elif fieldname in table.fields: field = ogetattr(table, fieldname) else: field = None return field # ------------------------------------------------------------------------- @staticmethod def _resolve_key(table, fieldname): """ Resolve a foreign key into the referenced table and the join and left join between the current table and the referenced table @param table: the current Table @param fieldname: the fieldname of the foreign key @return: tuple of (referenced table, join, left join) @raise: AttributeError is either the field or the referended table are not found @raise: SyntaxError if the field is not a foreign key """ if fieldname in table.fields: f = table[fieldname] else: raise AttributeError("key not found: %s" % fieldname) ktablename, pkey, multiple = s3_get_foreign_key(f, m2m=False) if not ktablename: raise SyntaxError("%s is not a foreign key" % f) ktable = current.s3db.table(ktablename, AttributeError("undefined table %s" % ktablename), db_only=True) pkey = ktable[pkey] if pkey else ktable._id join = [ktable.on(f == pkey)] return ktable, join # ------------------------------------------------------------------------- @staticmethod def _resolve_alias(resource, alias): """ Resolve a table alias into the linked table (component, linktable or free join), and the joins and left joins between the current resource and the linked table. @param resource: the current S3Resource @param alias: the alias @return: tuple of (linked table, joins, left joins, multiple, distinct), the two latter being flags to indicate possible ambiguous query results (needed by the query builder) @raise: AttributeError if one of the key fields or tables can not be found @raise: SyntaxError if the alias can not be resolved (e.g. because on of the keys isn't a foreign key, points to the wrong table or is ambiguous) """ # Alias for this resource? if alias in ("~", resource.alias): return resource.table, None, False, False multiple = True linked = resource.linked if linked and linked.alias == alias: # It's the linked table linktable = resource.table ktable = linked.table join = [ktable.on(ktable[linked.fkey] == linktable[linked.rkey])] return ktable, join, multiple, True s3db = current.s3db tablename = resource.tablename # Try to attach the component if alias not in resource.components and \ alias not in resource.links: _alias = alias hook = s3db.get_component(tablename, alias) if not hook: _alias = s3db.get_alias(tablename, alias) if _alias: hook = s3db.get_component(tablename, _alias) if hook: resource._attach(_alias, hook) components = resource.components links = resource.links if alias in components: # Is a component component = components[alias] ktable = component.table join = component._join() multiple = component.multiple elif alias in links: # Is a linktable link = links[alias] ktable = link.table join = link._join() elif "_" in alias: # Is a free join DELETED = current.xml.DELETED table = resource.table tablename = resource.tablename pkey = fkey = None # Find the table fkey, kname = (alias.split(":") + [None])[:2] if not kname: fkey, kname = kname, fkey ktable = s3db.table(kname, AttributeError("table not found: %s" % kname), db_only=True) if fkey is None: # Autodetect left key for fname in ktable.fields: tn, key, m = s3_get_foreign_key(ktable[fname], m2m=False) if not tn: continue if tn == tablename: if fkey is not None: raise SyntaxError("ambiguous foreign key in %s" % alias) else: fkey = fname if key: pkey = key if fkey is None: raise SyntaxError("no foreign key for %s in %s" % (tablename, kname)) else: # Check left key if fkey not in ktable.fields: raise AttributeError("no field %s in %s" % (fkey, kname)) tn, pkey, m = s3_get_foreign_key(ktable[fkey], m2m=False) if tn and tn != tablename: raise SyntaxError("%s.%s is not a foreign key for %s" % (kname, fkey, tablename)) elif not tn: raise SyntaxError("%s.%s is not a foreign key" % (kname, fkey)) # Default primary key if pkey is None: pkey = table._id.name # Build join query = (table[pkey] == ktable[fkey]) if DELETED in ktable.fields: query &= ktable[DELETED] != True join = [ktable.on(query)] else: raise SyntaxError("Invalid tablename: %s" % alias) return ktable, join, multiple, True # ============================================================================= class S3ResourceField(object): """ Helper class to resolve a field selector against a resource """ # ------------------------------------------------------------------------- def __init__(self, resource, selector, label=None): """ Constructor @param resource: the resource @param selector: the field selector (string) """ self.resource = resource self.selector = selector lf = S3FieldPath.resolve(resource, selector) self.tname = lf.tname self.fname = lf.fname self.colname = lf.colname self._joins = lf.joins self.distinct = lf.distinct self.multiple = lf.multiple self._join = None self.field = lf.field self.virtual = False self.represent = s3_unicode self.requires = None if self.field is not None: field = self.field self.ftype = str(field.type) if resource.linked is not None and self.ftype == "id": # Always represent the link-table's ID as the # linked record's ID => needed for data tables self.represent = lambda i, resource=resource: \ resource.component_id(None, i) else: self.represent = field.represent self.requires = field.requires elif self.colname: self.virtual = True self.ftype = "virtual" else: self.ftype = "context" # Fall back to the field label if label is None: fname = self.fname if fname in ["L1", "L2", "L3", "L3", "L4", "L5"]: try: label = current.gis.get_location_hierarchy(fname) except: label = None elif fname == "L0": label = current.messages.COUNTRY if label is None: f = self.field if f: label = f.label elif fname: label = " ".join([s.strip().capitalize() for s in fname.split("_") if s]) else: label = None self.label = label self.show = True # ------------------------------------------------------------------------- def __repr__(self): """ String representation of this instance """ return "<S3ResourceField " \ "selector='%s' " \ "label='%s' " \ "table='%s' " \ "field='%s' " \ "type='%s'>" % \ (self.selector, self.label, self.tname, self.fname, self.ftype) # ------------------------------------------------------------------------- @property def join(self): """ Implicit join (Query) for this field, for backwards-compatibility """ if self._join is not None: return self._join join = self._join = {} for tablename, joins in self._joins.items(): query = None for expression in joins: if query is None: query = expression.second else: query &= expression.second if query: join[tablename] = query return join # ------------------------------------------------------------------------- @property def left(self): """ The left joins for this field, for backwards-compability """ return self._joins # ------------------------------------------------------------------------- def extract(self, row, represent=False, lazy=False): """ Extract the value for this field from a row @param row: the Row @param represent: render a text representation for the value @param lazy: return a lazy representation handle if available """ tname = self.tname fname = self.fname colname = self.colname error = "Field not found in Row: %s" % colname if type(row) is Row: try: if tname in row.__dict__: value = ogetattr(ogetattr(row, tname), fname) else: value = ogetattr(row, fname) except: try: value = row[colname] except (KeyError, AttributeError): raise KeyError(error) elif fname in row: value = row[fname] elif colname in row: value = row[colname] elif tname is not None and \ tname in row and fname in row[tname]: value = row[tname][fname] else: raise KeyError(error) if callable(value): # Lazy virtual field try: value = value() except: current.log.error(sys.exc_info()[1]) value = None if represent: renderer = self.represent if callable(renderer): if lazy and hasattr(renderer, "bulk"): return S3RepresentLazy(value, renderer) else: return renderer(value) else: return s3_unicode(value) else: return value # ============================================================================= class S3Joins(object): """ A collection of joins """ def __init__(self, tablename, joins=None): """ Constructor @param tablename: the name of the master table @param joins: list of joins """ self.tablename = tablename self.joins = {} self.tables = set() self.add(joins) # ------------------------------------------------------------------------- def __iter__(self): """ Iterate over the names of all joined tables in the collection """ return self.joins.__iter__() # ------------------------------------------------------------------------- def __getitem__(self, tablename): """ Get the list of joins for a table @param tablename: the tablename """ return self.joins.__getitem__(tablename) # ------------------------------------------------------------------------- def __setitem__(self, tablename, joins): """ Update the joins for a table @param tablename: the tablename @param joins: the list of joins for this table """ master = self.tablename joins_dict = self.joins tables = current.db._adapter.tables joins_dict[tablename] = joins if len(joins) > 1: for join in joins: try: tname = join.first._tablename except AttributeError: tname = str(join.first) if tname not in joins_dict and \ master in tables(join.second): joins_dict[tname] = [join] self.tables.add(tablename) return # ------------------------------------------------------------------------- def keys(self): """ Get a list of names of all joined tables """ return self.joins.keys() # ------------------------------------------------------------------------- def items(self): """ Get a list of tuples (tablename, [joins]) for all joined tables """ return self.joins.items() # ------------------------------------------------------------------------- def values(self): """ Get a list of joins for all joined tables @return: a nested list like [[join, join, ...], ...] """ return self.joins.values() # ------------------------------------------------------------------------- def add(self, joins): """ Add joins to this collection @param joins: a join or a list/tuple of joins @return: the list of names of all tables for which joins have been added to the collection """ tablenames = set() if joins: if not isinstance(joins, (list, tuple)): joins = [joins] for join in joins: tablename = join.first._tablename self[tablename] = [join] tablenames.add(tablename) return list(tablenames) # ------------------------------------------------------------------------- def extend(self, other): """ Extend this collection with the joins from another collection @param other: the other collection (S3Joins), or a dict like {tablename: [join, join]} @return: the list of names of all tables for which joins have been added to the collection """ if type(other) is S3Joins: add = self.tables.add else: add = None joins = self.joins if type(other) is S3Joins else self for tablename in other: if tablename not in self.joins: joins[tablename] = other[tablename] if add: add(tablename) return other.keys() # ------------------------------------------------------------------------- def __repr__(self): """ String representation of this collection """ return "<S3Joins %s>" % str([str(j) for j in self.as_list()]) # ------------------------------------------------------------------------- def as_list(self, tablenames=None, aqueries=None, prefer=None): """ Return joins from this collection as list @param tablenames: the names of the tables for which joins shall be returned, defaults to all tables in the collection. Dependencies will be included automatically (if available) @param aqueries: dict of accessible-queries {tablename: query} to include in the joins; if there is no entry for a particular table, then it will be looked up from current.auth and added to the dict. To prevent differential authorization of a particular joined table, set {<tablename>: None} in the dict @param prefer: If any table or any of its dependencies would be joined by this S3Joins collection, then skip this table here (and enforce it to be joined by the preferred collection), to prevent duplication of left joins as inner joins: join = inner_joins.as_list(prefer=left_joins) left = left_joins.as_list() @return: a list of joins, ordered by their interdependency, which can be used as join/left parameter of Set.select() """ accessible_query = current.auth.s3_accessible_query if tablenames is None: tablenames = self.tables else: tablenames = set(tablenames) skip = set() if prefer: preferred_joins = prefer.as_list(tablenames=tablenames) for join in preferred_joins: try: tname = join.first._tablename except AttributeError: tname = str(join.first) skip.add(tname) tablenames -= skip joins = self.joins # Resolve dependencies required_tables = set() get_tables = current.db._adapter.tables for tablename in tablenames: if tablename not in joins or \ tablename == self.tablename or \ tablename in skip: continue join_list = joins[tablename] preferred = False dependencies = set() for join in join_list: join_tables = set(get_tables(join.second)) if join_tables: if any((tname in skip for tname in join_tables)): preferred = True dependencies |= join_tables if preferred: skip.add(tablename) skip |= dependencies prefer.extend({tablename: join_list}) else: required_tables.add(tablename) required_tables |= dependencies # Collect joins joins_dict = {} for tablename in required_tables: if tablename not in joins or tablename == self.tablename: continue for join in joins[tablename]: j = join table = j.first tname = table._tablename if aqueries is not None and tname in tablenames: if tname not in aqueries: aquery = accessible_query("read", table) aqueries[tname] = aquery else: aquery = aqueries[tname] if aquery is not None: j = join.first.on(join.second & aquery) joins_dict[tname] = j # Sort joins (if possible) try: return self.sort(joins_dict.values()) except RuntimeError: return joins_dict.values() # ------------------------------------------------------------------------- @classmethod def sort(cls, joins): """ Sort a list of left-joins by their interdependency @param joins: the list of joins """ if len(joins) <= 1: return joins r = list(joins) tables = current.db._adapter.tables append = r.append head = None for i in xrange(len(joins)): join = r.pop(0) head = join tablenames = tables(join.second) for j in r: try: tn = j.first._tablename except AttributeError: tn = str(j.first) if tn in tablenames: head = None break if head is not None: break else: append(join) if head is not None: return [head] + cls.sort(r) else: raise RuntimeError("circular join dependency")<|fim▁hole|>class S3ResourceQuery(object): """ Helper class representing a resource query - unlike DAL Query objects, these can be converted to/from URL filters """ # Supported operators NOT = "not" AND = "and" OR = "or" LT = "lt" LE = "le" EQ = "eq" NE = "ne" GE = "ge" GT = "gt" LIKE = "like" BELONGS = "belongs" CONTAINS = "contains" ANYOF = "anyof" TYPEOF = "typeof" COMPARISON = [LT, LE, EQ, NE, GE, GT, LIKE, BELONGS, CONTAINS, ANYOF, TYPEOF] OPERATORS = [NOT, AND, OR] + COMPARISON # ------------------------------------------------------------------------- def __init__(self, op, left=None, right=None): """ Constructor """ if op not in self.OPERATORS: raise SyntaxError("Invalid operator: %s" % op) self.op = op self.left = left self.right = right # ------------------------------------------------------------------------- def __and__(self, other): """ AND """ return S3ResourceQuery(self.AND, self, other) # ------------------------------------------------------------------------- def __or__(self, other): """ OR """ return S3ResourceQuery(self.OR, self, other) # ------------------------------------------------------------------------- def __invert__(self): """ NOT """ if self.op == self.NOT: return self.left else: return S3ResourceQuery(self.NOT, self) # ------------------------------------------------------------------------- def _joins(self, resource, left=False): op = self.op l = self.left r = self.right if op in (self.AND, self.OR): if isinstance(l, S3ResourceQuery): ljoins, ld = l._joins(resource, left=left) else: ljoins, ld = {}, False if isinstance(r, S3ResourceQuery): rjoins, rd = r._joins(resource, left=left) else: rjoins, rd = {}, False ljoins = dict(ljoins) ljoins.update(rjoins) return (ljoins, ld or rd) elif op == self.NOT: if isinstance(l, S3ResourceQuery): return l._joins(resource, left=left) else: return {}, False joins, distinct = {}, False if isinstance(l, S3FieldSelector): try: rfield = l.resolve(resource) except (SyntaxError, AttributeError): pass else: distinct = rfield.distinct if distinct and left or not distinct and not left: joins = rfield._joins return (joins, distinct) # ------------------------------------------------------------------------- def fields(self): """ Get all field selectors involved with this query """ op = self.op l = self.left r = self.right if op in (self.AND, self.OR): lf = l.fields() rf = r.fields() return lf + rf elif op == self.NOT: return l.fields() elif isinstance(l, S3FieldSelector): return [l.name] else: return [] # ------------------------------------------------------------------------- def split(self, resource): """ Split this query into a real query and a virtual one (AND) @param resource: the S3Resource @return: tuple (DAL-translatable sub-query, virtual filter), both S3ResourceQuery instances """ op = self.op l = self.left r = self.right if op == self.AND: lq, lf = l.split(resource) \ if isinstance(l, S3ResourceQuery) else (l, None) rq, rf = r.split(resource) \ if isinstance(r, S3ResourceQuery) else (r, None) q = lq if rq is not None: if q is not None: q &= rq else: q = rq f = lf if rf is not None: if f is not None: f &= rf else: f = rf return q, f elif op == self.OR: lq, lf = l.split(resource) \ if isinstance(l, S3ResourceQuery) else (l, None) rq, rf = r.split(resource) \ if isinstance(r, S3ResourceQuery) else (r, None) if lf is not None or rf is not None: return None, self else: q = lq if rq is not None: if q is not None: q |= rq else: q = rq return q, None elif op == self.NOT: if isinstance(l, S3ResourceQuery): if l.op == self.OR: i = (~(l.left)) & (~(l.right)) return i.split(resource) else: q, f = l.split(resource) if q is not None and f is not None: return None, self elif q is not None: return ~q, None elif f is not None: return None, ~f else: return ~l, None l = self.left try: if isinstance(l, S3FieldSelector): lfield = l.resolve(resource) else: lfield = S3ResourceField(resource, l) except: lfield = None if not lfield or lfield.field is None: return None, self else: return self, None # ------------------------------------------------------------------------- def transform(self, resource): """ Placeholder for transformation method @param resource: the S3Resource """ # @todo: implement return self # ------------------------------------------------------------------------- def query(self, resource): """ Convert this S3ResourceQuery into a DAL query, ignoring virtual fields (the necessary joins for this query can be constructed with the joins() method) @param resource: the resource to resolve the query against """ op = self.op l = self.left r = self.right # Resolve query components if op == self.AND: l = l.query(resource) if isinstance(l, S3ResourceQuery) else l r = r.query(resource) if isinstance(r, S3ResourceQuery) else r if l is None or r is None: return None elif l is False or r is False: return l if r is False else r if l is False else False else: return l & r elif op == self.OR: l = l.query(resource) if isinstance(l, S3ResourceQuery) else l r = r.query(resource) if isinstance(r, S3ResourceQuery) else r if l is None or r is None: return None elif l is False or r is False: return l if r is False else r if l is False else False else: return l | r elif op == self.NOT: l = l.query(resource) if isinstance(l, S3ResourceQuery) else l if l is None: return None elif l is False: return False else: return ~l # Resolve the fields if isinstance(l, S3FieldSelector): try: rfield = S3ResourceField(resource, l.name) except: return None if rfield.virtual: return None elif not rfield.field: return False lfield = l.expr(rfield.field) elif isinstance(l, Field): lfield = l else: return None # not a field at all if isinstance(r, S3FieldSelector): try: rfield = S3ResourceField(resource, r.name) except: return None rfield = rfield.field if rfield.virtual: return None elif not rfield.field: return False rfield = r.expr(rfield.field) else: rfield = r # Resolve the operator invert = False query_bare = self._query_bare ftype = str(lfield.type) if isinstance(rfield, (list, tuple)) and ftype[:4] != "list": if op == self.EQ: op = self.BELONGS elif op == self.NE: op = self.BELONGS invert = True elif op not in (self.BELONGS, self.TYPEOF): query = None for v in rfield: q = query_bare(op, lfield, v) if q is not None: if query is None: query = q else: query |= q return query # Convert date(time) strings if ftype == "datetime" and \ isinstance(rfield, basestring): rfield = S3TypeConverter.convert(datetime.datetime, rfield) elif ftype == "date" and \ isinstance(rfield, basestring): rfield = S3TypeConverter.convert(datetime.date, rfield) query = query_bare(op, lfield, rfield) if invert: query = ~(query) return query # ------------------------------------------------------------------------- def _query_bare(self, op, l, r): """ Translate a filter expression into a DAL query @param op: the operator @param l: the left operand @param r: the right operand """ if op == self.CONTAINS: q = l.contains(r, all=True) elif op == self.ANYOF: # NB str/int doesn't matter here q = l.contains(r, all=False) elif op == self.BELONGS: q = self._query_belongs(l, r) elif op == self.TYPEOF: q = self._query_typeof(l, r) elif op == self.LIKE: q = l.like(s3_unicode(r)) elif op == self.LT: q = l < r elif op == self.LE: q = l <= r elif op == self.EQ: q = l == r elif op == self.NE: q = l != r elif op == self.GE: q = l >= r elif op == self.GT: q = l > r else: q = None return q # ------------------------------------------------------------------------- def _query_typeof(self, l, r): """ Translate TYPEOF into DAL expression @param l: the left operator @param r: the right operator """ hierarchy, field, nodeset, none = self._resolve_hierarchy(l, r) if not hierarchy: # Not a hierarchical query => use simple belongs return self._query_belongs(l, r) if not field: # Field does not exist (=>skip subquery) return None # Construct the subquery list_type = str(field.type)[:5] == "list:" if nodeset: if list_type: q = (field.contains(list(nodeset))) elif len(nodeset) > 1: q = (field.belongs(nodeset)) else: q = (field == tuple(nodeset)[0]) else: q = None if none: # None needs special handling with older DAL versions if not list_type: if q is None: q = (field == None) else: q |= (field == None) if q is None: # Values not resolvable (=subquery always fails) q = field.belongs(set()) return q # ------------------------------------------------------------------------- @classmethod def _resolve_hierarchy(cls, l, r): """ Resolve the hierarchical lookup in a typeof-query @param l: the left operator @param r: the right operator """ from s3hierarchy import S3Hierarchy tablename = l.tablename # Connect to the hierarchy hierarchy = S3Hierarchy(tablename) if hierarchy.config is None: # Reference to a hierarchical table? ktablename, key = s3_get_foreign_key(l)[:2] if ktablename: hierarchy = S3Hierarchy(ktablename) else: key = None list_type = str(l.type)[:5] == "list:" if hierarchy.config is None and not list_type: # No hierarchy configured and no list:reference return False, None, None, None field, keys = l, r if not key: s3db = current.s3db table = s3db[tablename] if l.name != table._id.name: # Lookup-field rather than primary key => resolve it # Build a filter expression for the lookup table fs = S3FieldSelector(l.name) if list_type: expr = fs.contains(r) else: expr = cls._query_belongs(l, r, field = fs) # Resolve filter expression into subquery resource = s3db.resource(tablename) if expr is not None: subquery = expr.query(resource) else: subquery = None if not subquery: # Field doesn't exist return True, None, None, None # Execute query and retrieve the lookup table IDs DELETED = current.xml.DELETED if DELETED in table.fields: subquery &= table[DELETED] != True rows = current.db(subquery).select(table._id) # Override field/keys field = table[hierarchy.pkey.name] keys = set([row[table._id.name] for row in rows]) nodeset, none = None, False if keys: # Lookup all descendant types from the hierarchy none = False if not isinstance(keys, (list, tuple, set)): keys = set([keys]) nodes = set() for node in keys: if node is None: none = True else: try: node_id = long(node) except ValueError: continue nodes.add(node_id) if hierarchy.config is not None: nodeset = hierarchy.findall(nodes, inclusive=True) else: nodeset = nodes elif keys is None: none = True return True, field, nodeset, none # ------------------------------------------------------------------------- @staticmethod def _query_belongs(l, r, field=None): """ Resolve BELONGS into a DAL expression (or S3ResourceQuery if field is an S3FieldSelector) @param l: the left operator @param r: the right operator @param field: alternative left operator """ if field is None: field = l expr = None none = False if not isinstance(r, (list, tuple, set)): items = [r] else: items = r if None in items: none = True items = [item for item in items if item is not None] wildcard = False if str(l.type) in ("string", "text"): for item in items: if isinstance(item, basestring): if "*" in item and "%" not in item: s = item.replace("*", "%") else: s = item else: try: s = str(item) except: continue if "%" in s: wildcard = True _expr = (field.like(s)) else: _expr = (field == s) if expr is None: expr = _expr else: expr |= _expr if not wildcard: if len(items) == 1: # Don't use belongs() for single value expr = (field == tuple(items)[0]) elif items: expr = (field.belongs(items)) if none: # None needs special handling with older DAL versions if expr is None: expr = (field == None) else: expr |= (field == None) elif expr is None: expr = field.belongs(set()) return expr # ------------------------------------------------------------------------- def __call__(self, resource, row, virtual=True): """ Probe whether the row matches the query @param resource: the resource to resolve the query against @param row: the DB row @param virtual: execute only virtual queries """ if self.op == self.AND: l = self.left(resource, row, virtual=False) r = self.right(resource, row, virtual=False) if l is None: return r if r is None: return l return l and r elif self.op == self.OR: l = self.left(resource, row, virtual=False) r = self.right(resource, row, virtual=False) if l is None: return r if r is None: return l return l or r elif self.op == self.NOT: l = self.left(resource, row) if l is None: return None else: return not l real = False left = self.left if isinstance(left, S3FieldSelector): try: lfield = left.resolve(resource) except (AttributeError, KeyError, SyntaxError): return None if lfield.field is not None: real = True elif not lfield.virtual: # Unresolvable expression => skip return None else: lfield = left if isinstance(left, Field): real = True right = self.right if isinstance(right, S3FieldSelector): try: rfield = right.resolve(resource) except (AttributeError, KeyError, SyntaxError): return None if rfield.virtual: real = False elif rfield.field is None: # Unresolvable expression => skip return None else: rfield = right if virtual and real: return None extract = lambda f: S3FieldSelector.extract(resource, row, f) try: l = extract(lfield) r = extract(rfield) except (KeyError, SyntaxError): current.log.error(sys.exc_info()[1]) return None if isinstance(left, S3FieldSelector): l = left.expr(l) if isinstance(right, S3FieldSelector): r = right.expr(r) op = self.op invert = False probe = self._probe if isinstance(rfield, (list, tuple)) and \ not isinstance(lfield, (list, tuple)): if op == self.EQ: op = self.BELONGS elif op == self.NE: op = self.BELONGS invert = True elif op != self.BELONGS: for v in r: try: r = probe(op, l, v) except (TypeError, ValueError): r = False if r: return True return False try: r = probe(op, l, r) except (TypeError, ValueError): return False if invert and r is not None: return not r else: return r # ------------------------------------------------------------------------- def _probe(self, op, l, r): """ Probe whether the value pair matches the query @param l: the left value @param r: the right value """ result = False convert = S3TypeConverter.convert # Fallbacks for TYPEOF if op == self.TYPEOF: if isinstance(l, (list, tuple, set)): op = self.ANYOF elif isinstance(r, (list, tuple, set)): op = self.BELONGS else: op = self.EQ if op == self.CONTAINS: r = convert(l, r) result = self._probe_contains(l, r) elif op == self.ANYOF: if not isinstance(r, (list, tuple, set)): r = [r] for v in r: if isinstance(l, (list, tuple, set, basestring)): if self._probe_contains(l, v): return True elif l == v: return True return False elif op == self.BELONGS: if not isinstance(r, (list, tuple, set)): r = [r] r = convert(l, r) result = self._probe_contains(r, l) elif op == self.LIKE: pattern = re.escape(str(r)).replace("\\%", ".*").replace(".*.*", "\\%") return re.match(pattern, str(l)) is not None else: r = convert(l, r) if op == self.LT: result = l < r elif op == self.LE: result = l <= r elif op == self.EQ: result = l == r elif op == self.NE: result = l != r elif op == self.GE: result = l >= r elif op == self.GT: result = l > r return result # ------------------------------------------------------------------------- @staticmethod def _probe_contains(a, b): """ Probe whether a contains b """ if a is None: return False try: if isinstance(a, basestring): return str(b) in a elif isinstance(a, (list, tuple, set)): if isinstance(b, (list, tuple, set)): convert = S3TypeConverter.convert found = True for _b in b: if _b not in a: found = False for _a in a: try: if convert(_a, _b) == _a: found = True break except (TypeError, ValueError): continue if not found: break return found else: return b in a else: return str(b) in str(a) except: return False # ------------------------------------------------------------------------- def represent(self, resource): """ Represent this query as a human-readable string. @param resource: the resource to resolve the query against """ op = self.op l = self.left r = self.right if op == self.AND: l = l.represent(resource) \ if isinstance(l, S3ResourceQuery) else str(l) r = r.represent(resource) \ if isinstance(r, S3ResourceQuery) else str(r) return "(%s and %s)" % (l, r) elif op == self.OR: l = l.represent(resource) \ if isinstance(l, S3ResourceQuery) else str(l) r = r.represent(resource) \ if isinstance(r, S3ResourceQuery) else str(r) return "(%s or %s)" % (l, r) elif op == self.NOT: l = l.represent(resource) \ if isinstance(l, S3ResourceQuery) else str(l) return "(not %s)" % l else: if isinstance(l, S3FieldSelector): l = l.represent(resource) elif isinstance(l, basestring): l = '"%s"' % l if isinstance(r, S3FieldSelector): r = r.represent(resource) elif isinstance(r, basestring): r = '"%s"' % r if op == self.CONTAINS: return "(%s in %s)" % (r, l) elif op == self.BELONGS: return "(%s in %s)" % (l, r) elif op == self.ANYOF: return "(%s contains any of %s)" % (l, r) elif op == self.TYPEOF: return "(%s is a type of %s)" % (l, r) elif op == self.LIKE: return "(%s like %s)" % (l, r) elif op == self.LT: return "(%s < %s)" % (l, r) elif op == self.LE: return "(%s <= %s)" % (l, r) elif op == self.EQ: return "(%s == %s)" % (l, r) elif op == self.NE: return "(%s != %s)" % (l, r) elif op == self.GE: return "(%s >= %s)" % (l, r) elif op == self.GT: return "(%s > %s)" % (l, r) else: return "(%s ?%s? %s)" % (l, op, r) # ------------------------------------------------------------------------- def serialize_url(self, resource=None): """ Serialize this query as URL query @return: a Storage of URL variables """ op = self.op l = self.left r = self.right url_query = Storage() def _serialize(n, o, v, invert): try: quote = lambda s: s if "," not in s else '"%s"' % s if isinstance(v, list): v = ",".join([quote(S3TypeConverter.convert(str, val)) for val in v]) else: v = quote(S3TypeConverter.convert(str, v)) except: return if "." not in n: if resource is not None: n = "~.%s" % n else: return url_query if o == self.LIKE: v = v.replace("%", "*") if o == self.EQ: operator = "" else: operator = "__%s" % o if invert: operator = "%s!" % operator key = "%s%s" % (n, operator) if key in url_query: url_query[key] = "%s,%s" % (url_query[key], v) else: url_query[key] = v return url_query if op == self.AND: lu = l.serialize_url(resource=resource) url_query.update(lu) ru = r.serialize_url(resource=resource) url_query.update(ru) elif op == self.OR: sub = self._or() if sub is None: # This OR-subtree is not serializable return url_query n, o, v, invert = sub _serialize(n, o, v, invert) elif op == self.NOT: lu = l.serialize_url(resource=resource) for k in lu: url_query["%s!" % k] = lu[k] elif isinstance(l, S3FieldSelector): _serialize(l.name, op, r, False) return url_query # ------------------------------------------------------------------------- def _or(self): """ Helper method to URL-serialize an OR-subtree in a query in alternative field selector syntax if they all use the same operator and value (this is needed to URL-serialize an S3SearchSimpleWidget query). """ op = self.op l = self.left r = self.right if op == self.AND: return None elif op == self.NOT: lname, lop, lval, linv = l._or() return (lname, lop, lval, not linv) elif op == self.OR: lvars = l._or() rvars = r._or() if lvars is None or rvars is None: return None lname, lop, lval, linv = lvars rname, rop, rval, rinv = rvars if lop != rop or linv != rinv: return None if lname == rname: return (lname, lop, [lval, rval], linv) elif lval == rval: return ("%s|%s" % (lname, rname), lop, lval, linv) else: return None else: return (l.name, op, r, False) # ============================================================================= class S3URLQuery(object): """ URL Query Parser """ # ------------------------------------------------------------------------- @classmethod def parse(cls, resource, vars): """ Construct a Storage of S3ResourceQuery from a Storage of get_vars @param resource: the S3Resource @param vars: the get_vars @return: Storage of S3ResourceQuery like {alias: query}, where alias is the alias of the component the query concerns """ query = Storage() if resource is None: return query if not vars: return query subquery = cls._subquery allof = lambda l, r: l if r is None else r if l is None else r & l for key, value in vars.iteritems(): if key == "$filter": # Instantiate the advanced filter parser parser = S3URLQueryParser() if parser.parser is None: # not available continue # Multiple $filter expressions? expressions = value if type(value) is list else [value] # Default alias (=master) default_alias = resource.alias # Parse all expressions for expression in expressions: parsed = parser.parse(expression) for alias in parsed: q = parsed[alias] qalias = alias if alias is not None else default_alias if qalias not in query: query[qalias] = [q] else: query[qalias].append(q) # Stop here continue elif not("." in key or key[0] == "(" and ")" in key): # Not a filter expression continue # Process old-style filters selectors, op, invert = cls.parse_expression(key) if type(value) is list: # Multiple queries with the same selector (AND) q = reduce(allof, [subquery(selectors, op, invert, v) for v in value], None) else: q = subquery(selectors, op, invert, value) if q is None: continue # Append to query if len(selectors) > 1: aliases = [s.split(".", 1)[0] for s in selectors] if len(set(aliases)) == 1: alias = aliases[0] else: alias = resource.alias #alias = resource.alias else: alias = selectors[0].split(".", 1)[0] if alias == "~": alias = resource.alias if alias not in query: query[alias] = [q] else: query[alias].append(q) return query # ------------------------------------------------------------------------- @staticmethod def parse_url(url): """ Parse a URL query into get_vars @param query: the URL query string @return: the get_vars (Storage) """ if not url: return Storage() elif "?" in url: query = url.split("?", 1)[1] elif "=" in url: query = url else: return Storage() import cgi dget = cgi.parse_qsl(query, keep_blank_values=1) get_vars = Storage() for (key, value) in dget: if key in get_vars: if type(get_vars[key]) is list: get_vars[key].append(value) else: get_vars[key] = [get_vars[key], value] else: get_vars[key] = value return get_vars # ------------------------------------------------------------------------- @staticmethod def parse_expression(key): """ Parse a URL expression @param key: the key for the URL variable @return: tuple (selectors, operator, invert) """ if key[-1] == "!": invert = True else: invert = False fs = key.rstrip("!") op = None if "__" in fs: fs, op = fs.split("__", 1) op = op.strip("_") if not op: op = "eq" if "|" in fs: selectors = [s for s in fs.split("|") if s] else: selectors = [fs] return selectors, op, invert # ------------------------------------------------------------------------- @staticmethod def parse_value(value): """ Parse a URL query value @param value: the value @return: the parsed value """ uquote = lambda w: w.replace('\\"', '\\"\\') \ .strip('"') \ .replace('\\"\\', '"') NONE = ("NONE", "None") if type(value) is not list: value = [value] vlist = [] for item in value: w = "" quote = False ignore_quote = False for c in s3_unicode(item): if c == '"' and not ignore_quote: w += c quote = not quote elif c == "," and not quote: if w in NONE: w = None else: w = uquote(w).encode("utf-8") vlist.append(w) w = "" else: w += c if c == "\\": ignore_quote = True else: ignore_quote = False if w in NONE: w = None else: w = uquote(w).encode("utf-8") vlist.append(w) if len(vlist) == 1: return vlist[0] return vlist # ------------------------------------------------------------------------- @classmethod def _subquery(cls, selectors, op, invert, value): """ Construct a sub-query from URL selectors, operator and value @param selectors: the selector(s) @param op: the operator @param invert: invert the query @param value: the value """ v = cls.parse_value(value) q = None for fs in selectors: if op == S3ResourceQuery.LIKE: # Auto-lowercase and replace wildcard f = S3FieldSelector(fs).lower() if isinstance(v, basestring): v = v.replace("*", "%").lower() elif isinstance(v, list): v = [x.replace("*", "%").lower() for x in v if x is not None] else: f = S3FieldSelector(fs) rquery = None try: rquery = S3ResourceQuery(op, f, v) except SyntaxError: current.log.error("Invalid URL query operator: %s (sub-query ignored)" % op) q = None break # Invert operation if invert: rquery = ~rquery # Add to subquery if q is None: q = rquery elif invert: q &= rquery else: q |= rquery return q # ============================================================================= # Helper to combine multiple queries using AND # combine = lambda x, y: x & y if x is not None else y # ============================================================================= class S3URLQueryParser(object): """ New-style URL Filter Parser """ def __init__(self): """ Constructor """ self.parser = None self.ParseResults = None self.ParseException = None self._parser() # ------------------------------------------------------------------------- def _parser(self): """ Import PyParsing and define the syntax for filter expressions """ # PyParsing available? try: import pyparsing as pp except ImportError: current.log.error("Advanced filter syntax requires pyparsing, $filter ignored") return False # Selector Syntax context = lambda s, l, t: t[0].replace("[", "(").replace("]", ")") selector = pp.Word(pp.alphas + "[]~", pp.alphanums + "_.$:[]") selector.setParseAction(context) keyword = lambda x, y: x | pp.Keyword(y) if x else pp.Keyword(y) # Expression Syntax function = reduce(keyword, S3FieldSelector.OPERATORS) expression = function + \ pp.Literal("(").suppress() + \ selector + \ pp.Literal(")").suppress() # Comparison Syntax comparison = reduce(keyword, S3ResourceQuery.COMPARISON) # Value Syntax number = pp.Regex(r"[+-]?\d+(:?\.\d*)?(:?[eE][+-]?\d+)?") value = number | \ pp.Keyword("NONE") | \ pp.quotedString | \ pp.Word(pp.alphanums + pp.printables) qe = pp.Group(pp.Group(expression | selector) + comparison + pp.originalTextFor(pp.delimitedList(value, combine=True))) parser = pp.operatorPrecedence(qe, [("not", 1, pp.opAssoc.RIGHT, ), ("and", 2, pp.opAssoc.LEFT, ), ("or", 2, pp.opAssoc.LEFT, ), ]) self.parser = parser self.ParseResults = pp.ParseResults self.ParseException = pp.ParseException return True # ------------------------------------------------------------------------- def parse(self, expression): """ Parse a string expression and convert it into a dict of filters (S3ResourceQueries). @parameter expression: the filter expression as string @return: a dict of {component_alias: filter_query} """ query = {} parser = self.parser if not expression or parser is None: return query try: parsed = parser.parseString(expression) except self.ParseException: current.log.error("Invalid URL Filter Expression: '%s'" % expression) else: if parsed: query = self.convert_expression(parsed[0]) return query # ------------------------------------------------------------------------- def convert_expression(self, expression): """ Convert a parsed filter expression into a dict of filters (S3ResourceQueries) @param expression: the parsed filter expression (ParseResults) @returns: a dict of {component_alias: filter_query} """ ParseResults = self.ParseResults convert = self.convert_expression if isinstance(expression, ParseResults): first, op, second = ([None, None, None] + list(expression))[-3:] if isinstance(first, ParseResults): first = convert(first) if isinstance(second, ParseResults): second = convert(second) if op == "not": return self._not(second) elif op == "and": return self._and(first, second) elif op == "or": return self._or(first, second) elif op in S3ResourceQuery.COMPARISON: return self._query(op, first, second) elif op in S3FieldSelector.OPERATORS and second: selector = S3FieldSelector(second) selector.op = op return selector elif op is None and second: return S3FieldSelector(second) else: return None # ------------------------------------------------------------------------- def _and(self, first, second): """ Conjunction of two query {component_alias: filter_query} (AND) @param first: the first dict @param second: the second dict @return: the combined dict """ if not first: return second if not second: return first result = dict(first) for alias, subquery in second.items(): if alias not in result: result[alias] = subquery else: result[alias] &= subquery return result # ------------------------------------------------------------------------- def _or(self, first, second): """ Disjunction of two query dicts {component_alias: filter_query} (OR) @param first: the first query dict @param second: the second query dict @return: the combined dict """ if not first: return second if not second: return first if len(first) > 1: first = {None: reduce(combine, first.values())} if len(second) > 1: second = {None: reduce(combine, second.values())} falias = first.keys()[0] salias = second.keys()[0] alias = falias if falias == salias else None return {alias: first[falias] | second[salias]} # ------------------------------------------------------------------------- def _not(self, query): """ Negation of a query dict @param query: the query dict {component_alias: filter_query} """ if query is None: return None if len(query) == 1: alias, sub = query.items()[0] if sub.op == S3ResourceQuery.OR and alias is None: l = sub.left r = sub.right lalias = self._alias(sub.left.left) ralias = self._alias(sub.right.left) if lalias == ralias: return {alias: ~sub} else: # not(A or B) => not(A) and not(B) return {lalias: ~sub.left, ralias: ~sub.right} else: if sub.op == S3ResourceQuery.NOT: return {alias: sub.left} else: return {alias: ~sub} else: return {None: ~reduce(combine, query.values())} # ------------------------------------------------------------------------- def _query(self, op, first, second): """ Create an S3ResourceQuery @param op: the operator @param first: the first operand (=S3FieldSelector) @param second: the second operand (=value) """ if not isinstance(first, S3FieldSelector): return {} selector = first alias = self._alias(selector) value = S3URLQuery.parse_value(second.strip()) if op == S3ResourceQuery.LIKE: if isinstance(value, basestring): value = value.replace("*", "%").lower() elif isinstance(value, list): value = [x.replace("*", "%").lower() for x in value if x is not None] return {alias: S3ResourceQuery(op, selector, value)} # ------------------------------------------------------------------------- @staticmethod def _alias(selector): """ Get the component alias from an S3FieldSelector (DRY Helper) @param selector: the S3FieldSelector @return: the alias as string or None for the master resource """ alias = None if selector and isinstance(selector, S3FieldSelector): prefix = selector.name.split("$", 1)[0] if "." in prefix: alias = prefix.split(".", 1)[0] if alias in ("~", ""): alias = None return alias # END =========================================================================<|fim▁end|>
# =============================================================================
<|file_name|>ImageViewerQt.py<|end_file_name|><|fim▁begin|># -*- coding: utf-8 -*- """ ImageViewer.py: PyQt image viewer widget for a QPixmap in a QGraphicsView scene with mouse zooming and panning. """ import os.path from PyQt5.QtCore import Qt, QRectF, pyqtSignal, QT_VERSION_STR from PyQt5.QtGui import QImage, QPixmap, QPainterPath, QWheelEvent from PyQt5.QtWidgets import QGraphicsView, QGraphicsScene, QFileDialog __author__ = "Marcel Goldschen-Ohm <[email protected]>" __version__ = '0.9.0' class ImageViewerQt(QGraphicsView): """ PyQt image viewer widget for a QPixmap in a QGraphicsView scene with mouse zooming and panning. Displays a QImage or QPixmap (QImage is internally converted to a QPixmap). To display any other image format, you must first convert it to a QImage or QPixmap. Some useful image format conversion utilities: qimage2ndarray: NumPy ndarray <==> QImage (https://github.com/hmeine/qimage2ndarray) ImageQt: PIL Image <==> QImage (https://github.com/python-pillow/Pillow/blob/master/PIL/ImageQt.py) Mouse interaction: Left mouse button drag: Pan image. Right mouse button drag: Zoom box. Right mouse button doubleclick: Zoom to show entire image. """ # Mouse button signals emit image scene (x, y) coordinates. # !!! For image (row, column) matrix indexing, row = y and column = x. leftMouseButtonPressed = pyqtSignal(float, float) rightMouseButtonPressed = pyqtSignal(float, float) leftMouseButtonReleased = pyqtSignal(float, float) rightMouseButtonReleased = pyqtSignal(float, float) leftMouseButtonDoubleClicked = pyqtSignal(float, float) rightMouseButtonDoubleClicked = pyqtSignal(float, float) def __init__(self): QGraphicsView.__init__(self) # Image is displayed as a QPixmap in a QGraphicsScene attached to this QGraphicsView. self.scene = QGraphicsScene() self.setScene(self.scene) # Store a local handle to the scene's current image pixmap. self._pixmapHandle = None # Image aspect ratio mode. # !!! ONLY applies to full image. Aspect ratio is always ignored when zooming. # Qt.IgnoreAspectRatio: Scale image to fit viewport. # Qt.KeepAspectRatio: Scale image to fit inside viewport, preserving aspect ratio. # Qt.KeepAspectRatioByExpanding: Scale image to fill the viewport, preserving aspect ratio. self.aspectRatioMode = Qt.KeepAspectRatio # Scroll bar behaviour. # Qt.ScrollBarAlwaysOff: Never shows a scroll bar. # Qt.ScrollBarAlwaysOn: Always shows a scroll bar. # Qt.ScrollBarAsNeeded: Shows a scroll bar only when zoomed. self.setHorizontalScrollBarPolicy(Qt.ScrollBarAsNeeded) self.setVerticalScrollBarPolicy(Qt.ScrollBarAsNeeded) # Stack of QRectF zoom boxes in scene coordinates. self.zoomStack = [] # Flags for enabling/disabling mouse interaction. self.canZoom = True self.canPan = True def hasImage(self): """ Returns whether or not the scene contains an image pixmap. """ return self._pixmapHandle is not None def clearImage(self): """ Removes the current image pixmap from the scene if it exists. """ if self.hasImage(): self.scene.removeItem(self._pixmapHandle) self._pixmapHandle = None def pixmap(self): """ Returns the scene's current image pixmap as a QPixmap, or else None if no image exists. :rtype: QPixmap | None """ if self.hasImage(): return self._pixmapHandle.pixmap() return None def image(self): """ Returns the scene's current image pixmap as a QImage, or else None if no image exists. :rtype: QImage | None """ if self.hasImage(): return self._pixmapHandle.pixmap().toImage() return None def setImage(self, image): """ Set the scene's current image pixmap to the input QImage or QPixmap. Raises a RuntimeError if the input image has type other than QImage or QPixmap. :type image: QImage | QPixmap """ if type(image) is QPixmap: pixmap = image elif type(image) is QImage: pixmap = QPixmap.fromImage(image) elif image is None: pixmap = QPixmap() else: raise RuntimeError("ImageViewer.setImage: Argument must be a QImage or QPixmap.") if self.hasImage(): self._pixmapHandle.setPixmap(pixmap) else: self._pixmapHandle = self.scene.addPixmap(pixmap) self.setSceneRect(QRectF(pixmap.rect())) # Set scene size to image size. self.zoomStack = [] self.updateViewer() def loadImageFromFile(self, fileName=""): """ Load an image from file. Without any arguments, loadImageFromFile() will popup a file dialog to choose the image file. With a fileName argument, loadImageFromFile(fileName) will attempt to load the specified image file directly. """ if len(fileName) == 0: if QT_VERSION_STR[0] == '4': fileName = QFileDialog.getOpenFileName(self, "Open image file.") elif QT_VERSION_STR[0] == '5': fileName, dummy = QFileDialog.getOpenFileName(self, "Open image file.") if len(fileName) and os.path.isfile(fileName): image = QImage(fileName) self.setImage(image) def updateViewer(self): """ Show current zoom (if showing entire image, apply current aspect ratio mode). """ if not self.hasImage(): return if len(self.zoomStack) and self.sceneRect().contains(self.zoomStack[-1]): #self.fitInView(self.zoomStack[-1], Qt.IgnoreAspectRatio) # Show zoomed rect (ignore aspect ratio). self.fitInView(self.zoomStack[-1], self.aspectRatioMode) # Show zoomed rect (ignore aspect ratio). else: self.zoomStack = [] # Clear the zoom stack (in case we got here because of an invalid zoom). self.fitInView(self.sceneRect(), self.aspectRatioMode) # Show entire image (use current aspect ratio mode). def resizeEvent(self, event): """ Maintain current zoom on resize. """ self.updateViewer() def mousePressEvent(self, event): """ Start mouse pan or zoom mode. """ scenePos = self.mapToScene(event.pos()) if event.button() == Qt.LeftButton: if self.canPan: self.setDragMode(QGraphicsView.ScrollHandDrag) self.leftMouseButtonPressed.emit(scenePos.x(), scenePos.y()) elif event.button() == Qt.RightButton: if self.canZoom: self.setDragMode(QGraphicsView.RubberBandDrag) self.rightMouseButtonPressed.emit(scenePos.x(), scenePos.y()) QGraphicsView.mousePressEvent(self, event) def mouseReleaseEvent(self, event): """ Stop mouse pan or zoom mode (apply zoom if valid). """ QGraphicsView.mouseReleaseEvent(self, event) scenePos = self.mapToScene(event.pos()) if event.button() == Qt.LeftButton: self.setDragMode(QGraphicsView.NoDrag) self.leftMouseButtonReleased.emit(scenePos.x(), scenePos.y()) elif event.button() == Qt.RightButton: if self.canZoom: #viewBBox = self.zoomStack[-1] if len(self.zoomStack) else self.sceneRect() viewBBox = self.sceneRect() selectionBBox = self.scene.selectionArea().boundingRect().intersected(viewBBox) self.scene.setSelectionArea(QPainterPath()) # Clear current selection area. if selectionBBox.isValid() and (selectionBBox != viewBBox): self.zoomStack.append(selectionBBox)<|fim▁hole|> def mouseDoubleClickEvent(self, event): """ Show entire image. """ scenePos = self.mapToScene(event.pos()) if event.button() == Qt.LeftButton: self.leftMouseButtonDoubleClicked.emit(scenePos.x(), scenePos.y()) elif event.button() == Qt.RightButton: if self.canZoom: self.fitZoom() self.rightMouseButtonDoubleClicked.emit(scenePos.x(), scenePos.y()) QGraphicsView.mouseDoubleClickEvent(self, event) def fitZoom(self): # Clear zoom stack. self.zoomStack = [] self.updateViewer() def zoom( self, zoomIn, pos=None ): if pos is None: pos = self.sceneRect().center().toPoint() rect = QRectF( self.mapToScene(self.viewport().geometry()).boundingRect() ) if zoomIn: W = rect.width()/2 H = rect.height()/2 else: W = rect.width()*2 H = rect.height()*2 rect.setWidth( W ) rect.setHeight( H ) rect.moveCenter( self.mapToScene( pos ) ) rect = rect.intersected( self.sceneRect() ) self.zoomStack.append(rect) self.updateViewer() def wheelEvent(self, event): if not self.canZoom: return self.zoom( event.angleDelta().y()>0, event.pos() )<|fim▁end|>
self.updateViewer() self.setDragMode(QGraphicsView.NoDrag) self.rightMouseButtonReleased.emit(scenePos.x(), scenePos.y())
<|file_name|>SliderWnd.cpp<|end_file_name|><|fim▁begin|>#include "SliderWnd.h" #include <windowsx.h> #include "SliderKnob.h" #define TIMER_IGNORE_INPUT 50 #define IGNORE_DURATION 100 SliderWnd::SliderWnd(LPCWSTR className, LPCWSTR title, HINSTANCE hInstance) : MeterWnd(className, title, hInstance), _dragging(false) { long styles = GetWindowLongPtr(Window::Handle(), GWL_EXSTYLE); styles &= ~(WS_EX_NOACTIVATE | WS_EX_TRANSPARENT); SetWindowLongPtr(Window::Handle(), GWL_EXSTYLE, styles); } void SliderWnd::Show() { PositionWindow(); MeterWnd::Show(false); _ignoreInput = true; SetTimer(Window::Handle(), TIMER_IGNORE_INPUT, IGNORE_DURATION, NULL); } void SliderWnd::Knob(SliderKnob *knob) { _knob = knob; AddMeter(knob); } void SliderWnd::PositionWindow() { POINT p; GetCursorPos(&p); HMONITOR monitor = MonitorFromPoint(p, MONITOR_DEFAULTTONEAREST); MONITORINFO mInfo = {}; mInfo.cbSize = sizeof(mInfo); GetMonitorInfo(monitor, &mInfo); RECT mRect = mInfo.rcWork; /* Default location is the bottom of the window centered above cursor. */ POINT loc; loc.x = p.x - _size.cx / 2; loc.y = p.y - _size.cy; /* Reposition the window if it's falling off the monitor somewhere. */<|fim▁hole|> if (loc.x < mRect.left) { loc.x = mRect.left; } if (p.y > mRect.bottom) { loc.y = mRect.bottom - _size.cy; } if (p.x > mRect.right - _size.cx) { loc.x = mRect.right - _size.cx; } Position(loc.x, loc.y); } bool SliderWnd::MouseOverKnob(int x, int y) { if (x >= _knob->X() && x <= _knob->X() + _knob->Width() && y >= _knob->Y() && y <= _knob->Y() + _knob->Height()) { return true; } else { return false; } } bool SliderWnd::MouseOverTrack(int x, int y) { if (x >= _knob->TrackX() && x <= _knob->TrackX() + _knob->TrackWidth() && y >= _knob->TrackY() && y <= _knob->TrackY() + _knob->TrackHeight()) { return true; } else { return false; } } void SliderWnd::UpdateKnob(int x, int y) { int oldLoc, newLoc, drag, knobMax, knobMin; if (_vertical) { oldLoc = _knob->Y(); drag = y; knobMax = _knob->TrackY() + _knob->TrackHeight() - _knob->Height(); knobMin = _knob->TrackY(); } else { oldLoc = _knob->X(); drag = x; knobMax = _knob->TrackX() + _knob->TrackWidth() - _knob->Width(); knobMin = _knob->TrackX(); } if (drag - _dragOffset > knobMax) { newLoc = knobMax; } else if (drag - _dragOffset < knobMin) { newLoc = knobMin; } else { newLoc = drag - _dragOffset; } if (oldLoc != newLoc) { if (_vertical) { _knob->Y(newLoc); } else { _knob->X(newLoc); } MeterWnd::MeterLevels(_knob->Value()); SliderChanged(); Update(); } } LRESULT SliderWnd::WndProc( HWND hWnd, UINT message, WPARAM wParam, LPARAM lParam) { switch (message) { case WM_TIMER: if (wParam == TIMER_IGNORE_INPUT) { KillTimer(hWnd, TIMER_IGNORE_INPUT); _ignoreInput = false; } break; case WM_KILLFOCUS: Hide(); break; case WM_ACTIVATEAPP: if (wParam == 0) { /* We're being deactivated */ Hide(); } break; case WM_MOUSEMOVE: if (_dragging) { int x = GET_X_LPARAM(lParam); int y = GET_Y_LPARAM(lParam); UpdateKnob(x, y); } break; case WM_LBUTTONDOWN: { if (_knob == NULL) { break; } int x = GET_X_LPARAM(lParam); int y = GET_Y_LPARAM(lParam); if (MouseOverKnob(x, y)) { _dragging = true; if (_vertical) { _dragOffset = y - _knob->Y(); } else { _dragOffset = x - _knob->X(); } SetCapture(hWnd); } else if (MouseOverTrack(x, y)) { /* Simulate the mouse dragging to the clicked location: */ UpdateKnob(x, y); } break; } case WM_LBUTTONUP: _dragging = false; ReleaseCapture(); break; } /* Input events: arrow keys, right or middle mouse buttons, scrolling */ if (_ignoreInput == false) { switch (message) { case WM_MBUTTONUP: KeyPress(VK_MBUTTON); break; case WM_RBUTTONDOWN: KeyPress(VK_RBUTTON); break; case WM_MOUSEWHEEL: { if (_ignoreInput) { break; } short scroll = GET_WHEEL_DELTA_WPARAM(wParam); if (scroll > 0) { ScrollUp(); } else if (scroll < 0) { ScrollDown(); } break; } case WM_KEYUP: KeyPress(wParam); break; } } return MeterWnd::WndProc(hWnd, message, wParam, lParam); }<|fim▁end|>
if (loc.y < mRect.top) { loc.y = mRect.top; }
<|file_name|>FindCallVisitor.java<|end_file_name|><|fim▁begin|>/** * Copyright (c) 2005-2012 by Appcelerator, Inc. All Rights Reserved. * Licensed under the terms of the Eclipse Public License (EPL). * Please see the license.txt included with this distribution for details. * Any modifications to this file must keep this entire header intact. */ package com.python.pydev.refactoring.wizards.rename.visitors; import java.util.Stack; import org.python.pydev.parser.jython.SimpleNode; import org.python.pydev.parser.jython.Visitor; import org.python.pydev.parser.jython.ast.Call; import org.python.pydev.parser.jython.ast.Name; import org.python.pydev.parser.jython.ast.NameTok; /** * This visitor is used to find a call given its ast * * @author Fabio */ public class FindCallVisitor extends Visitor { private Name name; private NameTok nameTok; private Call call; private Stack<Call> lastCall = new Stack<Call>();<|fim▁hole|> } public FindCallVisitor(NameTok nameTok) { this.nameTok = nameTok; } public Call getCall() { return call; } @Override public Object visitCall(Call node) throws Exception { if (this.call != null) { return null; } if (node.func == name) { //check the name (direct) this.call = node; } else if (nameTok != null) { //check the name tok (inside of attribute) lastCall.push(node); Object r = super.visitCall(node); lastCall.pop(); if (this.call != null) { return null; } return r; } if (this.call != null) { return null; } return super.visitCall(node); } @Override public Object visitNameTok(NameTok node) throws Exception { if (node == nameTok) { if (lastCall.size() > 0) { call = lastCall.peek(); } return null; } return super.visitNameTok(node); } public static Call findCall(NameTok nametok, SimpleNode root) { FindCallVisitor visitor = new FindCallVisitor(nametok); try { visitor.traverse(root); } catch (Exception e) { throw new RuntimeException(e); } return visitor.call; } public static Call findCall(Name name, SimpleNode root) { FindCallVisitor visitor = new FindCallVisitor(name); try { visitor.traverse(root); } catch (Exception e) { throw new RuntimeException(e); } return visitor.call; } }<|fim▁end|>
public FindCallVisitor(Name name) { this.name = name;
<|file_name|>qa_burst_shaper.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright 2015 Free Software Foundation, Inc. # # This file is part of GNU Radio # # SPDX-License-Identifier: GPL-3.0-or-later # # from gnuradio import gr, gr_unittest from gnuradio import blocks, digital import pmt import numpy as np import sys def make_length_tag(offset, length): return gr.python_to_tag({'offset': offset, 'key': pmt.intern('packet_len'), 'value': pmt.from_long(length), 'srcid': pmt.intern('qa_burst_shaper')}) def make_tag(offset, key, value): return gr.python_to_tag({'offset': offset, 'key': pmt.intern(key), 'value': value, 'srcid': pmt.intern('qa_burst_shaper')}) def compare_tags(a, b): return a.offset == b.offset and pmt.equal(a.key, b.key) and \ pmt.equal(a.value, b.value) class qa_burst_shaper (gr_unittest.TestCase): def setUp(self): self.tb = gr.top_block() def tearDown(self): self.tb = None def test_ff(self): ''' test_ff: test with float values, even length window, zero padding, and no phasing ''' prepad = 10 postpad = 10 length = 20 data = np.ones(length) window = np.concatenate((-2.0 * np.ones(5), -4.0 * np.ones(5))) tags = (make_length_tag(0, length),) expected = np.concatenate((np.zeros(prepad), window[0:5], np.ones(length - len(window)), window[5:10], np.zeros(postpad))) etag = make_length_tag(0, length + prepad + postpad) # flowgraph source = blocks.vector_source_f(data, tags=tags) shaper = digital.burst_shaper_ff(window, pre_padding=prepad, post_padding=postpad) sink = blocks.vector_sink_f() self.tb.connect(source, shaper, sink) self.tb.run() # checks self.assertFloatTuplesAlmostEqual(sink.data(), expected, 6) self.assertTrue(compare_tags(sink.tags()[0], etag)) def test_cc(self): ''' test_cc: test with complex values, even length window, zero padding, and no phasing ''' prepad = 10 postpad = 10 length = 20 data = np.ones(length, dtype=complex) window = np.concatenate((-2.0 * np.ones(5, dtype=complex), -4.0 * np.ones(5, dtype=complex))) tags = (make_length_tag(0, length),) expected = np.concatenate((np.zeros(prepad, dtype=complex), window[0:5], np.ones(length - len(window), dtype=complex), window[5:10], np.zeros(postpad, dtype=complex))) etag = make_length_tag(0, length + prepad + postpad) # flowgraph source = blocks.vector_source_c(data, tags=tags) shaper = digital.burst_shaper_cc(window, pre_padding=prepad, post_padding=postpad) sink = blocks.vector_sink_c() self.tb.connect(source, shaper, sink) self.tb.run() # checks self.assertComplexTuplesAlmostEqual(sink.data(), expected, 6) self.assertTrue(compare_tags(sink.tags()[0], etag)) def test_ff_with_phasing(self): ''' test_ff_with_phasing: test with float values, even length window, zero padding, and phasing ''' prepad = 10 postpad = 10 length = 20 data = np.ones(length) window = np.concatenate((-2.0 * np.ones(5), -4.0 * np.ones(5))) tags = (make_length_tag(0, length),) phasing = np.zeros(5) for i in range(5): phasing[i] = ((-1.0)**i) expected = np.concatenate((np.zeros(prepad), phasing * window[0:5], np.ones(length), phasing * window[5:10], np.zeros(postpad))) etag = make_length_tag(0, length + prepad + postpad + len(window)) # flowgraph source = blocks.vector_source_f(data, tags=tags) shaper = digital.burst_shaper_ff(window, pre_padding=prepad, post_padding=postpad, insert_phasing=True) sink = blocks.vector_sink_f() self.tb.connect(source, shaper, sink) self.tb.run() # checks self.assertFloatTuplesAlmostEqual(sink.data(), expected, 6) self.assertTrue(compare_tags(sink.tags()[0], etag)) def test_cc_with_phasing(self): ''' test_cc_with_phasing: test with complex values, even length window, zero padding, and phasing ''' prepad = 10 postpad = 10 length = 20 data = np.ones(length, dtype=complex) window = np.concatenate((-2.0 * np.ones(5, dtype=complex), -4.0 * np.ones(5, dtype=complex))) tags = (make_length_tag(0, length),) phasing = np.zeros(5, dtype=complex) for i in range(5): phasing[i] = complex((-1.0)**i) expected = np.concatenate((np.zeros(prepad, dtype=complex), phasing * window[0:5], np.ones(length, dtype=complex), phasing * window[5:10], np.zeros(postpad, dtype=complex))) etag = make_length_tag(0, length + prepad + postpad + len(window)) # flowgraph source = blocks.vector_source_c(data, tags=tags) shaper = digital.burst_shaper_cc(window, pre_padding=prepad, post_padding=postpad, insert_phasing=True) sink = blocks.vector_sink_c() self.tb.connect(source, shaper, sink) self.tb.run() # checks self.assertComplexTuplesAlmostEqual(sink.data(), expected, 6) self.assertTrue(compare_tags(sink.tags()[0], etag)) def test_odd_window(self): ''' test_odd_window: test with odd length window; center sample should be applied at end of up flank and beginning of down flank ''' prepad = 10 postpad = 10 length = 20 data = np.ones(length) window = np.concatenate((-2.0 * np.ones(5), -3.0 * np.ones(1), -4.0 * np.ones(5))) tags = (make_length_tag(0, length),) expected = np.concatenate((np.zeros(prepad), window[0:6], np.ones(length - len(window) - 1), window[5:11], np.zeros(postpad))) etag = make_length_tag(0, length + prepad + postpad) # flowgraph source = blocks.vector_source_f(data, tags=tags) shaper = digital.burst_shaper_ff(window, pre_padding=prepad, post_padding=postpad) sink = blocks.vector_sink_f() self.tb.connect(source, shaper, sink) self.tb.run() # checks self.assertFloatTuplesAlmostEqual(sink.data(), expected, 6) self.assertTrue(compare_tags(sink.tags()[0], etag)) def test_short_burst(self): ''' test_short_burst: test with burst length shorter than window length; clips the window up and down flanks to FLOOR(length/2) samples ''' prepad = 10 postpad = 10 length = 9 data = np.ones(length) window = np.arange(length + 2, dtype=float) tags = (make_length_tag(0, length),) expected = np.concatenate((np.zeros(prepad), window[0:4], np.ones(1), window[5:9], np.zeros(postpad))) etag = make_length_tag(0, length + prepad + postpad) # flowgraph source = blocks.vector_source_f(data, tags=tags) shaper = digital.burst_shaper_ff(window, pre_padding=prepad, post_padding=postpad) sink = blocks.vector_sink_f() self.tb.connect(source, shaper, sink) self.tb.run() # checks self.assertFloatTuplesAlmostEqual(sink.data(), expected, 6) self.assertTrue(compare_tags(sink.tags()[0], etag)) def test_consecutive_bursts(self): ''' test_consecutive_bursts: test with consecutive bursts of different lengths ''' prepad = 10 postpad = 10 length1 = 15 length2 = 25 data = np.concatenate((np.ones(length1), -1.0 * np.ones(length2))) window = np.concatenate((-2.0 * np.ones(5), -4.0 * np.ones(5))) tags = (make_length_tag(0, length1), make_length_tag(length1, length2)) expected = np.concatenate((np.zeros(prepad), window[0:5], np.ones(length1 - len(window)), window[5:10], np.zeros(postpad + prepad), -1.0 * window[0:5], -1.0 * np.ones(length2 - len(window)), -1.0 * window[5:10], np.zeros(postpad))) etags = (make_length_tag(0, length1 + prepad + postpad), make_length_tag(length1 + prepad + postpad, length2 + prepad + postpad)) # flowgraph source = blocks.vector_source_f(data, tags=tags) shaper = digital.burst_shaper_ff(window, pre_padding=prepad, post_padding=postpad) sink = blocks.vector_sink_f() self.tb.connect(source, shaper, sink) self.tb.run() # checks self.assertFloatTuplesAlmostEqual(sink.data(), expected, 6) for i in range(len(etags)): self.assertTrue(compare_tags(sink.tags()[i], etags[i])) def test_tag_gap(self): ''' test_tag_gap: test with gap between tags; should drop samples that are between proper tagged streams ''' prepad = 10 postpad = 10 length = 20 gap_len = 5 data = np.arange(2 * length + gap_len, dtype=float) window = np.concatenate((-2.0 * np.ones(5), -4.0 * np.ones(5))) ewindow = window * \ np.array([1, -1, 1, -1, 1, 1, -1, 1, -1, 1], dtype=float) tags = (make_length_tag(0, length), make_length_tag(length + gap_len, length)) expected = np.concatenate((np.zeros(prepad), ewindow[0:5], np.arange(0, length, dtype=float), ewindow[5:10], np.zeros(postpad), np.zeros(prepad), ewindow[0:5], np.arange(length + gap_len, 2 * length + gap_len, dtype=float), ewindow[5:10], np.zeros(postpad))) burst_len = length + len(window) + prepad + postpad etags = (make_length_tag(0, burst_len), make_length_tag(burst_len, burst_len)) # flowgraph source = blocks.vector_source_f(data, tags=tags) shaper = digital.burst_shaper_ff(window, pre_padding=prepad, post_padding=postpad, insert_phasing=True) sink = blocks.vector_sink_f() self.tb.connect(source, shaper, sink) self.tb.run() # checks self.assertFloatTuplesAlmostEqual(sink.data(), expected, 6) for i in range(len(etags)): self.assertTrue(compare_tags(sink.tags()[i], etags[i])) def test_tag_propagation(self): ''' test_tag_propagation: test that non length tags are handled correctly ''' prepad = 10 postpad = 10 length1 = 15 length2 = 25 gap_len = 5 lentag1_offset = 0 lentag2_offset = length1 + gap_len tag1_offset = 0 # accompanies first length tag tag2_offset = length1 + gap_len # accompanies second length tag tag3_offset = 2 # in ramp-up state tag4_offset = length1 + 2 # in gap; tag will be dropped tag5_offset = length1 + gap_len + 7 # in copy state data = np.concatenate((np.ones(length1), np.zeros(gap_len), -1.0 * np.ones(length2))) window = np.concatenate((-2.0 * np.ones(5), -4.0 * np.ones(5))) tags = (make_length_tag(lentag1_offset, length1), make_length_tag(lentag2_offset, length2), make_tag(tag1_offset, 'head', pmt.intern('tag1')), make_tag(tag2_offset, 'head', pmt.intern('tag2')), make_tag(tag3_offset, 'body', pmt.intern('tag3')), make_tag(tag4_offset, 'body', pmt.intern('tag4')), make_tag(tag5_offset, 'body', pmt.intern('tag5'))) expected = np.concatenate((np.zeros(prepad), window[0:5], np.ones(length1 - len(window)), window[5:10], np.zeros(postpad + prepad), -1.0 * window[0:5],<|fim▁hole|> elentag2_offset = length1 + prepad + postpad etag1_offset = 0 etag2_offset = elentag2_offset etag3_offset = prepad + tag3_offset etag5_offset = 2 * prepad + postpad + tag5_offset - gap_len etags = (make_length_tag(elentag1_offset, length1 + prepad + postpad), make_length_tag(elentag2_offset, length2 + prepad + postpad), make_tag(etag1_offset, 'head', pmt.intern('tag1')), make_tag(etag2_offset, 'head', pmt.intern('tag2')), make_tag(etag3_offset, 'body', pmt.intern('tag3')), make_tag(etag5_offset, 'body', pmt.intern('tag5'))) # flowgraph source = blocks.vector_source_f(data, tags=tags) shaper = digital.burst_shaper_ff(window, pre_padding=prepad, post_padding=postpad) sink = blocks.vector_sink_f() self.tb.connect(source, shaper, sink) self.tb.run() # checks self.assertFloatTuplesAlmostEqual(sink.data(), expected, 6) for x, y in zip(sorted(sink.tags()), sorted(etags)): self.assertTrue(compare_tags(x, y)) if __name__ == '__main__': gr_unittest.run(qa_burst_shaper)<|fim▁end|>
-1.0 * np.ones(length2 - len(window)), -1.0 * window[5:10], np.zeros(postpad))) elentag1_offset = 0
<|file_name|>Attribute.java<|end_file_name|><|fim▁begin|>package com.thecodeinside.easyfactory.core; /** * A factory's attribute. * * @author Wellington Pinheiro <[email protected]> * * @param <T> type of the attribute */ public class Attribute<T> { private String id; private T value; public String getId() { return this.id; } public T getValue() { return this.value; } public Attribute() { } public Attribute(String id, T value) { this.id = id;<|fim▁hole|> @Override public String toString() { return "Attribute [id=" + id + ", value=" + value + "]"; } public boolean isReference() { return value instanceof FactoryReference; } public boolean isNotReference() { return !isReference(); } }<|fim▁end|>
this.value = value; }
<|file_name|>birdCLEF_test.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python print "HANDLING IMPORTS...",<|fim▁hole|>import random import operator import argparse import numpy as np import cv2 from sklearn.utils import shuffle import itertools import scipy.io.wavfile as wave from scipy import interpolate import python_speech_features as psf from pydub import AudioSegment import pickle import theano import theano.tensor as T from lasagne import random as lasagne_random from lasagne import layers as l from lasagne import nonlinearities from lasagne import init from lasagne import objectives from lasagne import updates from lasagne import regularization try: from lasagne.layers.dnn import BatchNormDNNLayer as BatchNormLayer except ImportError: from lasagne.layers import BatchNormLayer print "DONE!" ######################## CONFIG ######################### #Fixed random seed RANDOM_SEED = 1337 RANDOM = np.random.RandomState(RANDOM_SEED) lasagne_random.set_rng(RANDOM) #Image params IM_SIZE = (512, 256) #(width, height) IM_DIM = 1 #General model params MODEL_TYPE = 1 MULTI_LABEL = False NONLINEARITY = nonlinearities.elu #nonlinearities.rectify INIT_GAIN = 1.0 #1.0 if elu, sqrt(2) if rectify #Pre-trained model params MODEL_PATH = 'model/' PRETRAINED_MODEL = 'birdCLEF_TUCMI_Run1_Model.pkl' #We need to define the class labels our net has learned #but we use another file for that from birdCLEF_class_labels import CLASSES ################### ARGUMENT PARSER ##################### def parse_args(): parser = argparse.ArgumentParser(description='BirdCLEF bird sound classification') parser.add_argument('--filename', dest='filename', help='path to sample wav file for testing', type=str, default='') parser.add_argument('--overlap', dest='spec_overlap', help='spectrogram overlap in seconds', type=int, default=0) parser.add_argument('--results', dest='num_results', help='number of results', type=int, default=5) parser.add_argument('--confidence', dest='min_confidence', help='confidence threshold', type=float, default=0.01) args = parser.parse_args() return args ################ SPECTROGRAM EXTRACTION ################# #Change sample rate if not 44.1 kHz def changeSampleRate(sig, rate): duration = sig.shape[0] / rate time_old = np.linspace(0, duration, sig.shape[0]) time_new = np.linspace(0, duration, int(sig.shape[0] * 44100 / rate)) interpolator = interpolate.interp1d(time_old, sig.T) new_audio = interpolator(time_new).T sig = np.round(new_audio).astype(sig.dtype) return sig, 44100 #Get magnitude spec from signal split def getMagSpec(sig, rate, winlen, winstep, NFFT): #get frames winfunc = lambda x:np.ones((x,)) frames = psf.sigproc.framesig(sig, winlen*rate, winstep*rate, winfunc) #Magnitude Spectrogram magspec = np.rot90(psf.sigproc.magspec(frames, NFFT)) return magspec #Split signal into five-second chunks with overlap of 4 and minimum length of 1 second #Use these settings for other chunk lengths: #winlen, winstep, seconds #0.05, 0.0097, 5s #0.05, 0.0195, 10s #0.05, 0.0585, 30s def getMultiSpec(path, seconds=5, overlap=2, minlen=1, winlen=0.05, winstep=0.0097, NFFT=840): #open wav file (rate,sig) = wave.read(path) #adjust to different sample rates if rate != 44100: sig, rate = changeSampleRate(sig, rate) #split signal with overlap sig_splits = [] for i in xrange(0, len(sig), int((seconds - overlap) * rate)): split = sig[i:i + seconds * rate] if len(split) >= minlen * rate: sig_splits.append(split) #is signal too short for segmentation? if len(sig_splits) == 0: sig_splits.append(sig) #calculate spectrogram for every split for sig in sig_splits: #preemphasis sig = psf.sigproc.preemphasis(sig, coeff=0.95) #get spec magspec = getMagSpec(sig, rate, winlen, winstep, NFFT) #get rid of high frequencies h, w = magspec.shape[:2] magspec = magspec[h - 256:, :] #normalize in [0, 1] magspec -= magspec.min(axis=None) magspec /= magspec.max(axis=None) #fix shape to 512x256 pixels without distortion magspec = magspec[:256, :512] temp = np.zeros((256, 512), dtype="float32") temp[:magspec.shape[0], :magspec.shape[1]] = magspec magspec = temp.copy() magspec = cv2.resize(magspec, (512, 256)) #DEBUG: show spec #cv2.imshow('SPEC', magspec) #cv2.waitKey(-1) yield magspec ################## BUILDING THE MODEL ################### def buildModel(mtype=1): print "BUILDING MODEL TYPE", mtype, "..." #default settings (Model 1) filters = 64 first_stride = 2 last_filter_multiplier = 16 #specific model type settings (see working notes for details) if mtype == 2: first_stride = 1 elif mtype == 3: filters = 32 last_filter_multiplier = 8 #input layer net = l.InputLayer((None, IM_DIM, IM_SIZE[1], IM_SIZE[0])) #conv layers net = l.batch_norm(l.Conv2DLayer(net, num_filters=filters, filter_size=7, pad='same', stride=first_stride, W=init.HeNormal(gain=INIT_GAIN), nonlinearity=NONLINEARITY)) net = l.MaxPool2DLayer(net, pool_size=2) if mtype == 2: net = l.batch_norm(l.Conv2DLayer(net, num_filters=filters, filter_size=5, pad='same', stride=1, W=init.HeNormal(gain=INIT_GAIN), nonlinearity=NONLINEARITY)) net = l.MaxPool2DLayer(net, pool_size=2) net = l.batch_norm(l.Conv2DLayer(net, num_filters=filters * 2, filter_size=5, pad='same', stride=1, W=init.HeNormal(gain=INIT_GAIN), nonlinearity=NONLINEARITY)) net = l.MaxPool2DLayer(net, pool_size=2) net = l.batch_norm(l.Conv2DLayer(net, num_filters=filters * 4, filter_size=3, pad='same', stride=1, W=init.HeNormal(gain=INIT_GAIN), nonlinearity=NONLINEARITY)) net = l.MaxPool2DLayer(net, pool_size=2) net = l.batch_norm(l.Conv2DLayer(net, num_filters=filters * 8, filter_size=3, pad='same', stride=1, W=init.HeNormal(gain=INIT_GAIN), nonlinearity=NONLINEARITY)) net = l.MaxPool2DLayer(net, pool_size=2) net = l.batch_norm(l.Conv2DLayer(net, num_filters=filters * last_filter_multiplier, filter_size=3, pad='same', stride=1, W=init.HeNormal(gain=INIT_GAIN), nonlinearity=NONLINEARITY)) net = l.MaxPool2DLayer(net, pool_size=2) print "\tFINAL POOL OUT SHAPE:", l.get_output_shape(net) #dense layers net = l.batch_norm(l.DenseLayer(net, 512, W=init.HeNormal(gain=INIT_GAIN), nonlinearity=NONLINEARITY)) net = l.batch_norm(l.DenseLayer(net, 512, W=init.HeNormal(gain=INIT_GAIN), nonlinearity=NONLINEARITY)) #Classification Layer if MULTI_LABEL: net = l.DenseLayer(net, NUM_CLASSES, nonlinearity=nonlinearities.sigmoid, W=init.HeNormal(gain=1)) else: net = l.DenseLayer(net, NUM_CLASSES, nonlinearity=nonlinearities.softmax, W=init.HeNormal(gain=1)) print "...DONE!" #model stats print "MODEL HAS", (sum(hasattr(layer, 'W') for layer in l.get_all_layers(net))), "WEIGHTED LAYERS" print "MODEL HAS", l.count_params(net), "PARAMS" return net NUM_CLASSES = len(CLASSES) NET = buildModel(MODEL_TYPE) #################### MODEL LOAD ######################## def loadParams(epoch, filename=None): print "IMPORTING MODEL PARAMS...", net_filename = MODEL_PATH + filename with open(net_filename, 'rb') as f: params = pickle.load(f) l.set_all_param_values(NET, params) print "DONE!" #load params of trained model loadParams(-1, filename=PRETRAINED_MODEL) ################# PREDICTION FUNCTION #################### def getPredictionFuntion(net): net_output = l.get_output(net, deterministic=True) print "COMPILING THEANO TEST FUNCTION...", start = time.time() test_net = theano.function([l.get_all_layers(NET)[0].input_var], net_output, allow_input_downcast=True) print "DONE! (", int(time.time() - start), "s )" return test_net TEST_NET = getPredictionFuntion(NET) ################# PREDICTION POOLING #################### def predictionPooling(p): #You can test different prediction pooling strategies here #We only use average pooling p_pool = np.mean(p, axis=0) return p_pool ####################### PREDICT ######################### def predict(img): #transpose image if dim=3 try: img = np.transpose(img, (2, 0, 1)) except: pass #reshape image img = img.reshape(-1, IM_DIM, IM_SIZE[1], IM_SIZE[0]) #calling the test function returns the net output prediction = TEST_NET(img)[0] return prediction ####################### TESTING ######################### def testFile(path, spec_overlap=4, num_results=5, confidence_threshold=0.01): #time start = time.time() #extract spectrograms from wav-file and process them predictions = [] spec_cnt = 0 for spec in getMultiSpec(path, overlap=spec_overlap, minlen=1): #make prediction p = predict(spec) spec_cnt += 1 #stack predictions if len(predictions): predictions = np.vstack([predictions, p]) else: predictions = p #prediction pooling p_pool = predictionPooling(predictions) #get class labels for predictions p_labels = {} for i in range(p_pool.shape[0]): if p_pool[i] >= confidence_threshold: p_labels[CLASSES[i]] = p_pool[i] #sort by confidence and limit results (None returns all results) p_sorted = sorted(p_labels.items(), key=operator.itemgetter(1), reverse=True)[:num_results] #take time again dur = time.time() - start return p_sorted, spec_cnt, dur #################### EXAMPLE USAGE ###################### if __name__ == "__main__": #adjust config args = parse_args() #do testing print 'TESTING:', args.filename pred, cnt, dur = testFile(args.filename, args.spec_overlap, args.num_results, args.min_confidence) print 'TOP PREDICTION(S):' for p in pred: print '\t', p[0], int(p[1] * 100), '%' print 'PREDICTION FOR', cnt, 'SPECS TOOK', int(dur * 1000), 'ms (', int(dur / cnt * 1000) , 'ms/spec', ')'<|fim▁end|>
import os import time
<|file_name|>cinetorrent.py<|end_file_name|><|fim▁begin|># -*- coding: utf-8 -*- import sys PY3 = False if sys.version_info[0] >= 3: PY3 = True; unicode = str; unichr = chr; long = int if PY3: import urllib.parse as urlparse # Es muy lento en PY2. En PY3 es nativo else: import urlparse # Usamos el nativo de PY2 que es más rápido import re import time import traceback import base64 from channelselector import get_thumb from core import httptools from core import servertools from core import scrapertools from core import tmdb from core.item import Item from platformcode import config, logger from lib import generictools from channels import filtertools from channels import autoplay # Canal común con Cinetorrent, Magnetpelis, Pelispanda, Yestorrent IDIOMAS = {'Castellano': 'CAST', 'Latino': 'LAT', 'Version Original': 'VO'} list_language = list(IDIOMAS.values()) list_quality = [] list_servers = ['torrent'] canonical = { 'channel': 'cinetorrent', 'host': config.get_setting("current_host", 'cinetorrent', default=''), 'host_alt': ['https://cinetorrent.co/'], 'host_black_list': [], 'CF': False, 'CF_test': False, 'alfa_s': True } host = canonical['host'] or canonical['host_alt'][0] channel = canonical['channel'] categoria = channel.capitalize() host_torrent = host[:-1] patron_host = '((?:http.*\:)?\/\/(?:.*ww[^\.]*)?\.?(?:[^\.]+\.)?[\w|\-]+\.\w+)(?:\/|\?|$)' patron_domain = '(?:http.*\:)?\/\/(?:.*ww[^\.]*)?\.?(?:[^\.]+\.)?([\w|\-]+\.\w+)(?:\/|\?|$)' domain = scrapertools.find_single_match(host, patron_domain) __modo_grafico__ = config.get_setting('modo_grafico', channel) # TMDB? IDIOMAS_TMDB = {0: 'es', 1: 'en', 2: 'es,en'} idioma_busqueda = IDIOMAS_TMDB[config.get_setting('modo_grafico_lang', channel)] # Idioma base para TMDB idioma_busqueda_VO = IDIOMAS_TMDB[2] # Idioma para VO modo_ultima_temp = config.get_setting('seleccionar_ult_temporadda_activa', channel) #Actualización sólo últ. Temporada? timeout = config.get_setting('timeout_downloadpage', channel) season_colapse = config.get_setting('season_colapse', channel) # Season colapse? filter_languages = config.get_setting('filter_languages', channel) # Filtrado de idiomas? def mainlist(item): logger.info() itemlist = [] thumb_pelis = get_thumb("channels_movie.png") thumb_series = get_thumb("channels_tvshow.png") thumb_genero = get_thumb("genres.png") thumb_anno = get_thumb("years.png") thumb_calidad = get_thumb("top_rated.png") thumb_buscar = get_thumb("search.png") thumb_separador = get_thumb("next.png") thumb_settings = get_thumb("setting_0.png") autoplay.init(item.channel, list_servers, list_quality) itemlist.append(Item(channel=item.channel, title="Películas", action="submenu", url=host, thumbnail=thumb_pelis, extra="peliculas")) itemlist.append(Item(channel=item.channel, title=" - por Género", action="genero", url=host, thumbnail=thumb_genero, extra="peliculas")) itemlist.append(Item(channel=item.channel, title=" - por Año", action="anno", url=host, thumbnail=thumb_anno, extra="peliculas")) if channel not in ['magnetpelis']: itemlist.append(Item(channel=item.channel, title=" - por Calidad", action="calidad", url=host, thumbnail=thumb_calidad, extra="peliculas")) itemlist.append(Item(channel=item.channel, title="Series", action="submenu", url=host, thumbnail=thumb_series, extra="series")) itemlist.append(Item(channel=item.channel, title=" - por Año", action="anno", url=host, thumbnail=thumb_anno, extra="series")) itemlist.append(Item(channel=item.channel, title="Buscar...", action="search", url=host, thumbnail=thumb_buscar, extra="search")) itemlist.append(Item(channel=item.channel, url=host, title="[COLOR yellow]Configuración:[/COLOR]", folder=False, thumbnail=thumb_separador)) itemlist.append(Item(channel=item.channel, action="configuracion", title="Configurar canal", thumbnail=thumb_settings)) autoplay.show_option(item.channel, itemlist) #Activamos Autoplay return itemlist def configuracion(item): from platformcode import platformtools ret = platformtools.show_channel_settings() platformtools.itemlist_refresh() return def submenu(item): patron = '<li\s*class="header__nav-item">\s*<a\s*href="([^"]+)"\s*class="header__nav-link">([^<]+)<\/a>' data, response, item, itemlist = generictools.downloadpage(item.url, timeout=timeout, canonical=canonical, s2=False, patron=patron, item=item, itemlist=[]) # Descargamos la página #Verificamos si se ha cargado una página, y si además tiene la estructura correcta if not response.sucess or itemlist: # Si ERROR o lista de errores .. return itemlist # ... Salimos matches = re.compile(patron, re.DOTALL).findall(data) #logger.debug(patron) #logger.debug(matches) #logger.debug(data) if not matches: logger.error("ERROR 02: SUBMENU: Ha cambiado la estructura de la Web " + " / PATRON: " + patron + " / DATA: " + data) itemlist.append(item.clone(action='', title=item.category + ': ERROR 02: SUBMENU: Ha cambiado la estructura de la Web. ' + 'Reportar el error con el log')) return itemlist #si no hay más datos, algo no funciona, pintamos lo que tenemos for scrapedurl, scrapedtitle in matches: if scrapertools.slugify(scrapedtitle) in item.extra: item.url = urlparse.urljoin(host, scrapedurl.replace(scrapertools.find_single_match(scrapedurl, patron_host), '')) if not item.url.endswith('/'): item.url += '/' item.url += 'page/1' return listado(item) return itemlist def anno(item): logger.info() from platformcode import platformtools itemlist = [] patron = '(?i)<a\s*class="dropdown-toggle\s*header__nav-link"\s*href="#"\s*' patron += 'role="button"\s*data-toggle="dropdown">[^<]*A.O\s*<\/a>\s*' patron += '<ul\s*class="dropdown-menu\s*header__dropdown-menu">\s*(.*?)\s*<\/ul>\s*<\/li>' data, response, item, itemlist = generictools.downloadpage(item.url, timeout=timeout, canonical=canonical, s2=False, patron=patron, item=item, itemlist=[]) # Descargamos la página #Verificamos si se ha cargado una página, y si además tiene la estructura correcta if not response.sucess or itemlist: # Si ERROR o lista de errores ... return itemlist # ... Salimos data = scrapertools.find_single_match(data, patron) patron = '<li><a\s*href="([^"]+)"\s*target="[^"]*">([^<]+)\s*<\/a>\s*<\/li>' matches = re.compile(patron, re.DOTALL).findall(data) #logger.debug(patron) #logger.debug(matches) #logger.debug(data) if not matches: logger.error("ERROR 02: SUBMENU: Ha cambiado la estructura de la Web " + " / PATRON: " + patron + " / DATA: " + data) itemlist.append(item.clone(action='', title=item.category + ': ERROR 02: SUBMENU: Ha cambiado la estructura de la Web. ' + 'Reportar el error con el log')) return itemlist #si no hay más datos, algo no funciona, pintamos lo que tenemos year = platformtools.dialog_numeric(0, "Introduzca el Año de búsqueda", default="") item.url = re.sub(r'years/\d+', 'years/%s' % year, matches[0][0]) if not item.url.endswith('/'): item.url += '/' item.url += 'page/1' item.extra2 = 'anno' + str(year) return listado(item) def genero(item): logger.info() itemlist = [] patron = '(?i)<a\s*class="dropdown-toggle\s*header__nav-link"\s*href="#"\s*' patron += 'role="button"\s*data-toggle="dropdown">[^<]*G.nero\s*<\/a>\s*' patron += '<ul\s*class="dropdown-menu\s*header__dropdown-menu">\s*(.*?)\s*<\/ul>\s*<\/li>' data, response, item, itemlist = generictools.downloadpage(item.url, timeout=timeout, canonical=canonical, s2=False, patron=patron, item=item, itemlist=[]) # Descargamos la página #Verificamos si se ha cargado una página, y si además tiene la estructura correcta if not response.sucess or itemlist: # Si ERROR o lista de errores ... return itemlist # ... Salimos data = scrapertools.find_single_match(data, patron) patron = '<li><a\s*href="([^"]+)"\s*target="[^"]*">([^<]+)\s*<\/a>\s*<\/li>' matches = re.compile(patron, re.DOTALL).findall(data) #logger.debug(patron) #logger.debug(matches) #logger.debug(data) if not matches: logger.error("ERROR 02: SUBMENU: Ha cambiado la estructura de la Web " + " / PATRON: " + patron + " / DATA: " + data) itemlist.append(item.clone(action='', title=item.category + ': ERROR 02: SUBMENU: Ha cambiado la estructura de la Web. ' + 'Reportar el error con el log')) return itemlist #si no hay más datos, algo no funciona, pintamos lo que tenemos for scrapedurl, gen in matches: itemlist.append(item.clone(action="listado", title=gen.capitalize(), url=scrapedurl + 'page/1', extra2='genero')) return itemlist def calidad(item): logger.info() itemlist = [] patron = '(?i)<a\s*class="dropdown-toggle\s*header__nav-link"\s*href="#"\s*' patron += 'role="button"\s*data-toggle="dropdown">[^<]*calidad\s*<\/a>\s*' patron += '<ul\s*class="dropdown-menu\s*header__dropdown-menu">\s*(.*?)\s*<\/ul>\s*<\/li>' data, response, item, itemlist = generictools.downloadpage(item.url, timeout=timeout, canonical=canonical, s2=False, patron=patron, item=item, itemlist=[]) # Descargamos la página #Verificamos si se ha cargado una página, y si además tiene la estructura correcta if not response.sucess or itemlist: # Si ERROR o lista de errores ... return itemlist # ... Salimos data = scrapertools.find_single_match(data, patron) patron = '<li><a\s*href="([^"]+)"\s*target="[^"]*">([^<]+)\s*<\/a>\s*<\/li>' matches = re.compile(patron, re.DOTALL).findall(data) #logger.debug(patron) #logger.debug(matches) #logger.debug(data) if not matches: logger.error("ERROR 02: SUBMENU: Ha cambiado la estructura de la Web " + " / PATRON: " + patron + " / DATA: " + data) itemlist.append(item.clone(action='', title=item.category + ': ERROR 02: SUBMENU: Ha cambiado la estructura de la Web. ' + 'Reportar el error con el log')) return itemlist #si no hay más datos, algo no funciona, pintamos lo que tenemos for scrapedurl, cal in matches: if cal not in ['HD', '720p']: itemlist.append(item.clone(action="listado", title=cal.capitalize(), url=scrapedurl + 'page/1', extra2='calidad')) return itemlist def listado(item): # Listado principal y de búsquedas logger.info() itemlist = [] item.category = categoria thumb_pelis = get_thumb("channels_movie.png") thumb_series = get_thumb("channels_tvshow.png") #logger.debug(item) curr_page = 1 # Página inicial last_page = 99999 # Última página inicial last_page_print = 1 # Última página inicial, para píe de página page_factor = 1.0 # Factor de conversión de pag. web a pag. Alfa if item.curr_page: curr_page = int(item.curr_page) # Si viene de una pasada anterior, lo usamos del item.curr_page # ... y lo borramos if item.last_page: last_page = int(item.last_page) # Si viene de una pasada anterior, lo usamos del item.last_page # ... y lo borramos if item.page_factor: page_factor = float(item.page_factor) # Si viene de una pasada anterior, lo usamos del item.page_factor # ... y lo borramos if item.last_page_print: last_page_print = item.last_page_print # Si viene de una pasada anterior, lo usamos del item.last_page_print # ... y lo borramos cnt_tot = 30 # Poner el num. máximo de items por página cnt_title = 0 # Contador de líneas insertadas en Itemlist if item.cnt_tot_match: cnt_tot_match = float(item.cnt_tot_match) # restauramos el contador TOTAL de líneas procesadas de matches del item.cnt_tot_match else: cnt_tot_match = 0.0 # Contador TOTAL de líneas procesadas de matches inicio = time.time() # Controlaremos que el proceso no exceda de un tiempo razonable fin = inicio + 5 # Después de este tiempo pintamos (segundos) timeout_search = timeout * 2 # Timeout para descargas if item.extra == 'search' and item.extra2 == 'episodios': # Si viene de episodio que quitan los límites cnt_tot = 999 fin = inicio + 30 #Sistema de paginado para evitar páginas vacías o semi-vacías en casos de búsquedas con series con muchos episodios title_lista = [] # Guarda la lista de series que ya están en Itemlist, para no duplicar lineas if item.title_lista: # Si viene de una pasada anterior, la lista ya estará guardada title_lista.extend(item.title_lista) # Se usa la lista de páginas anteriores en Item del item.title_lista # ... limpiamos matches = [] if not item.extra2: # Si viene de Catálogo o de Alfabeto item.extra2 = '' post = None if item.post: # Rescatamos el Post, si lo hay post = item.post next_page_url = item.url # Máximo num. de líneas permitidas por TMDB. Máx de 5 segundos por Itemlist para no degradar el rendimiento while (cnt_title < cnt_tot and curr_page <= last_page and fin > time.time()) or item.matches: # Descarga la página data = '' cnt_match = 0 # Contador de líneas procesadas de matches if not item.matches: # si no viene de una pasada anterior, descargamos data, response, item, itemlist = generictools.downloadpage(next_page_url, canonical=canonical, timeout=timeout_search, post=post, s2=False, item=item, itemlist=itemlist) # Descargamos la página) # Verificamos si ha cambiado el Host if response.host: next_page_url = response.url_new # Verificamos si se ha cargado una página correcta curr_page += 1 # Apunto ya a la página siguiente if not data or not response.sucess: # Si la web está caída salimos sin dar error if len(itemlist) > 1: # Si hay algo que pintar lo pintamos last_page = 0 break return itemlist # Si no hay nada más, salimos directamente #Patrón para búsquedas, pelis y series patron = '<div\s*class="[^"]+">\s*<div\s*class="card">\s*<a\s*href="([^"]+)"' patron += '\s*class="card__cover">\s*<img[^>]+src="([^"]*)"\s*alt="[^"]*"\s*\/*>\s*' patron += '<div\s*class="card__play">.*?<\/div>\s*<ul\s*class="card__list">\s*' patron += '<li>([^<]+)<\/li>\s*<\/ul>\s*(?:<ul\s*class="card__list\s*right">\s*' patron += '<li>(\w*)<\/li>[^"]*<\/ul>\s*)?<\/a>\s*<div\s*class="card__content">\s*' patron += '<h3\s*class="card__title"><a\s*href="[^"]+">([^<]+)<\/a><\/h3>' patron += '.*?<\/div>\s*<\/div>\s*<\/div>' if not item.matches: # De pasada anterior? matches = re.compile(patron, re.DOTALL).findall(data) else: matches = item.matches del item.matches #logger.debug("PATRON: " + patron) #logger.debug(matches) #logger.debug(data) if not matches and item.extra != 'search' and not item.extra2: #error logger.error("ERROR 02: LISTADO: Ha cambiado la estructura de la Web " + " / PATRON: " + patron + " / DATA: " + data) itemlist.append(item.clone(action='', title=item.channel.capitalize() + ': ERROR 02: LISTADO: Ha cambiado la estructura de la Web. ' + 'Reportar el error con el log')) break #si no hay más datos, algo no funciona, pintamos lo que tenemos if not matches and item.extra == 'search': #búsqueda vacía if len(itemlist) > 0: # Si hay algo que pintar lo pintamos last_page = 0 break return itemlist #Salimos # Buscamos la próxima página next_page_url = re.sub(r'page\/(\d+)', 'page/%s' % str(curr_page), item.url) #logger.debug('curr_page: ' + str(curr_page) + ' / last_page: ' + str(last_page)) # Buscamos la última página if last_page == 99999: #Si es el valor inicial, buscamos patron_last = '<ul\s*class="pagination[^"]+">.*?<li>\s*<a\s*class="page-numbers"\s*href="[^"]+">' patron_last += '(\d+)<\/a><\/li>\s*<li>\s*<a\s*class="next page-numbers"\s*href="[^"]+">»<\/a><\/li>\s*<\/ul>' try: last_page = int(scrapertools.find_single_match(data, patron_last)) page_factor = float(len(matches)) / float(cnt_tot) except: #Si no lo encuentra, lo ponemos a 999 last_page = 1 last_page_print = int((float(len(matches)) / float(cnt_tot)) + 0.999999) #logger.debug('curr_page: ' + str(curr_page) + ' / last_page: ' + str(last_page)) #Empezamos el procesado de matches for scrapedurl, scrapedthumb, scrapedquality, scrapedlanguage, scrapedtitle in matches: cnt_match += 1 title = scrapedtitle title = scrapertools.remove_htmltags(title).rstrip('.') # Removemos Tags del título url = scrapedurl title_subs = [] #creamos una lista para guardar info importante # Slugify, pero más light title = title.replace("á", "a").replace("é", "e").replace("í", "i")\ .replace("ó", "o").replace("ú", "u").replace("ü", "u")\ .replace("�", "ñ").replace("ñ", "ñ") title = scrapertools.decode_utf8_error(title) # Se filtran las entradas para evitar duplicados de Temporadas url_list = url if url_list in title_lista: #Si ya hemos procesado el título, lo ignoramos continue else: title_lista += [url_list] #la añadimos a la lista de títulos # Si es una búsqueda por años, filtramos por tipo de contenido if item.extra == 'series' and '/serie' not in url: continue elif item.extra == 'peliculas' and '/serie' in url: continue cnt_title += 1 # Incrementamos el contador de entradas válidas item_local = item.clone() #Creamos copia de Item para trabajar if item_local.tipo: #... y limpiamos del item_local.tipo if item_local.totalItems: del item_local.totalItems if item_local.intervencion: del item_local.intervencion if item_local.viewmode: del item_local.viewmode item_local.extra2 = True del item_local.extra2 item_local.text_bold = True del item_local.text_bold item_local.text_color = True del item_local.text_color # Después de un Search se restablecen las categorías if item_local.extra == 'search': if '/serie' in url: item_local.extra = 'series' # Serie búsqueda else: item_local.extra = 'peliculas' # Película búsqueda # Procesamos idiomas item_local.language = [] #creamos lista para los idiomas if '[Subs. integrados]' in scrapedquality or '(Sub Forzados)' in scrapedquality \ or 'Sub' in scrapedquality or 'ing' in scrapedlanguage.lower(): item_local.language = ['VOS'] # añadimos VOS if 'lat' in scrapedlanguage.lower(): item_local.language += ['LAT'] # añadimos LAT if 'castellano' in scrapedquality.lower() or ('español' in scrapedquality.lower() \ and not 'latino' in scrapedquality.lower()) or 'cas' in scrapedlanguage.lower(): item_local.language += ['CAST'] # añadimos CAST if '[Dual' in title or 'dual' in scrapedquality.lower() or 'dual' in scrapedlanguage.lower(): title = re.sub(r'(?i)\[dual.*?\]', '', title) item_local.language += ['DUAL'] # añadimos DUAL if not item_local.language: item_local.language = ['LAT'] # [LAT] por defecto # Procesamos Calidad if scrapedquality: item_local.quality = scrapertools.remove_htmltags(scrapedquality) # iniciamos calidad if '[720p]' in scrapedquality.lower() or '720p' in scrapedquality.lower(): item_local.quality = '720p' if '[1080p]' in scrapedquality.lower() or '1080p' in scrapedquality.lower(): item_local.quality = '1080p' if '4k' in scrapedquality.lower(): item_local.quality = '4K' if '3d' in scrapedquality.lower() and not '3d' in item_local.quality.lower(): item_local.quality += ', 3D' if not item_local.quality or item_local.extra == 'series': item_local.quality = '720p' item_local.thumbnail = '' #iniciamos thumbnail item_local.url = urlparse.urljoin(host, url) #guardamos la url final item_local.context = "['buscar_trailer']" #... y el contexto # Guardamos los formatos para series if item_local.extra == 'series' or '/serie' in item_local.url: item_local.contentType = "tvshow" item_local.action = "episodios" item_local.season_colapse = season_colapse #Muestra las series agrupadas por temporadas? else: # Guardamos los formatos para películas item_local.contentType = "movie" item_local.action = "findvideos" #Limpiamos el título de la basura innecesaria if item_local.contentType == "tvshow": title = scrapertools.find_single_match(title, '(^.*?)\s*(?:$|\(|\[|-)') title = re.sub(r'(?i)TV|Online|(4k-hdr)|(fullbluray)|4k| - 4k|(3d)|miniserie', '', title).strip() item_local.quality = re.sub(r'(?i)proper|unrated|directors|cut|repack|internal|real|extended|masted|docu|super|duper|amzn|uncensored|hulu', '', item_local.quality).strip() #Analizamos el año. Si no está claro ponemos '-' item_local.infoLabels["year"] = '-' try: if 'anno' in item.extra2: item_local.infoLabels["year"] = int(item.extra2.replace('anno', '')) except: pass #Terminamos de limpiar el título title = re.sub(r'[\(|\[]\s+[\)|\]]', '', title) title = title.replace('()', '').replace('[]', '').replace('[4K]', '').replace('(4K)', '').strip().lower().title() item_local.from_title = title.strip().lower().title() #Guardamos esta etiqueta para posible desambiguación de título #Salvamos el título según el tipo de contenido if item_local.contentType == "movie": item_local.contentTitle = title else: item_local.contentSerieName = title.strip().lower().title() item_local.title = title.strip().lower().title() #Guarda la variable temporal que almacena la info adicional del título a ser restaurada después de TMDB item_local.title_subs = title_subs #Salvamos y borramos el número de temporadas porque TMDB a veces hace tonterias. Lo pasamos como serie completa if item_local.contentSeason and (item_local.contentType == "season" \ or item_local.contentType == "tvshow"): item_local.contentSeason_save = item_local.contentSeason del item_local.infoLabels['season'] #Ahora se filtra por idioma, si procede, y se pinta lo que vale if filter_languages > 0: #Si hay idioma seleccionado, se filtra itemlist = filtertools.get_link(itemlist, item_local, list_language) else: itemlist.append(item_local.clone()) #Si no, pintar pantalla cnt_title = len(itemlist) # Recalculamos los items después del filtrado if cnt_title >= cnt_tot and (len(matches) - cnt_match) + cnt_title > cnt_tot * 1.3: #Contador de líneas añadidas break #logger.debug(item_local) matches = matches[cnt_match:] # Salvamos la entradas no procesadas cnt_tot_match += cnt_match # Calcular el num. total de items mostrados #Pasamos a TMDB la lista completa Itemlist tmdb.set_infoLabels(itemlist, __modo_grafico__, idioma_busqueda=idioma_busqueda) #Llamamos al método para el maquillaje de los títulos obtenidos desde TMDB item, itemlist = generictools.post_tmdb_listado(item, itemlist) # Si es necesario añadir paginacion if curr_page <= last_page or len(matches) > 0: curr_page_print = int(cnt_tot_match / float(cnt_tot)) if curr_page_print < 1: curr_page_print = 1 if last_page: if last_page > 1: last_page_print = int((last_page * page_factor) + 0.999999) title = '%s de %s' % (curr_page_print, last_page_print) else: title = '%s' % curr_page_print itemlist.append(Item(channel=item.channel, action="listado", title=">> Página siguiente " + title, title_lista=title_lista, url=next_page_url, extra=item.extra, extra2=item.extra2, last_page=str(last_page), curr_page=str(curr_page), page_factor=str(page_factor), cnt_tot_match=str(cnt_tot_match), matches=matches, last_page_print=last_page_print, post=post)) return itemlist def findvideos(item): logger.info() itemlist = [] itemlist_t = [] #Itemlist total de enlaces itemlist_f = [] #Itemlist de enlaces filtrados matches = [] data = '' response = { 'data': data, 'sucess': False, 'code': 0 } response = type('HTTPResponse', (), response) #logger.debug(item) #Bajamos los datos de la página y seleccionamos el bloque patron = '\s*<th\s*class="hide-on-mobile">Total\s*Descargas<\/th>\s*<th>' patron += 'Descargar<\/th>\s*<\/thead>\s*<tbody>\s*(.*?<\/tr>)\s*<\/tbody>' patron += '\s*<\/table>\s*<\/div>' if not item.matches: data, response, item, itemlist = generictools.downloadpage(item.url, timeout=timeout, canonical=canonical, s2=False, patron=patron, item=item, itemlist=[]) # Descargamos la página) #Verificamos si se ha cargado una página, y si además tiene la estructura correcta if (not data and not item.matches) or response.code == 999: if item.emergency_urls and not item.videolibray_emergency_urls: #Hay urls de emergencia? if len(item.emergency_urls) > 1: matches = item.emergency_urls[1] #Restauramos matches de vídeos elif len(item.emergency_urls) == 1 and item.emergency_urls[0]: matches = item.emergency_urls[0] #Restauramos matches de vídeos - OLD FORMAT item.armagedon = True #Marcamos la situación como catastrófica else: if item.videolibray_emergency_urls: #Si es llamado desde creación de Videoteca... return item #Devolvemos el Item de la llamada else: return itemlist #si no hay más datos, algo no funciona, pintamos lo que tenemos elif data: # Seleccionamos el bloque y buscamos los apartados data = scrapertools.find_single_match(data, patron) patron = '<tr>(?:\s*<td>(\d+)<\/td>)?(?:\s*<td>([^<]*)<\/td>)?\s*<td>([^<]+)<\/td>' patron += '\s*<td>([^<]*)<\/td>\s*<td\s*class=[^<]+<\/td>(?:\s*<td>([^<]+)<\/td>)?' patron += '(?:\s*<td\s*class=[^<]+<\/td>)?\s*<td\s*class=[^<]+<\/td>\s*<td>\s*' patron += '<a\s*class="[^>]+href="([^"]+)"' if not item.armagedon: if not item.matches: matches = re.compile(patron, re.DOTALL).findall(data) else: matches = item.matches #logger.debug("PATRON: " + patron) #logger.debug(matches) #logger.debug(data) if not matches: #error return itemlist #Si es un lookup para cargar las urls de emergencia en la Videoteca... if item.videolibray_emergency_urls: item.emergency_urls = [] #Iniciamos emergency_urls item.emergency_urls.append([]) #Reservamos el espacio para los .torrents locales matches_list = [] # Convertimos matches-tuple a matches-list for tupla in matches: if isinstance(tupla, tuple): matches_list.append(list(tupla)) if matches_list: item.emergency_urls.append(matches_list) # Salvamnos matches de los vídeos... else: item.emergency_urls.append(matches) #Llamamos al método para crear el título general del vídeo, con toda la información obtenida de TMDB if not item.videolibray_emergency_urls: item, itemlist = generictools.post_tmdb_findvideos(item, itemlist) #Ahora tratamos los enlaces .torrent con las diferentes calidades for x, (episode_num, scrapedserver, scrapedquality, scrapedlanguage, scrapedsize, scrapedurl) in enumerate(matches): scrapedpassword = '' #Generamos una copia de Item para trabajar sobre ella item_local = item.clone() item_local.url = generictools.convert_url_base64(scrapedurl, host_torrent) if item.videolibray_emergency_urls and item_local.url != scrapedurl: item.emergency_urls[1][x][4] = item_local.url # Restauramos urls de emergencia si es necesario local_torr = '' if item.emergency_urls and not item.videolibray_emergency_urls: try: # Guardamos la url ALTERNATIVA if item.emergency_urls[0][0].startswith('http') or item.emergency_urls[0][0].startswith('//'): item_local.torrent_alt = generictools.convert_url_base64(item.emergency_urls[0][0], host_torrent) else: item_local.torrent_alt = generictools.convert_url_base64(item.emergency_urls[0][0]) except: item_local.torrent_alt = '' item.emergency_urls[0] = [] from core import filetools if item.contentType == 'movie': FOLDER = config.get_setting("folder_movies") else: FOLDER = config.get_setting("folder_tvshows") if item.armagedon and item_local.torrent_alt: item_local.url = item_local.torrent_alt # Restauramos la url if not item.torrent_alt.startswith('http'): local_torr = filetools.join(config.get_videolibrary_path(), FOLDER, item_local.url) if len(item.emergency_urls[0]) > 1: del item.emergency_urls[0][0] # Procesamos idiomas item_local.language = [] #creamos lista para los idiomas item_local.quality = scrapedquality # Copiamos la calidad if '[Subs. integrados]' in scrapedlanguage or '(Sub Forzados)' in scrapedlanguage \ or 'Subs integrados' in scrapedlanguage: item_local.language = ['VOS'] # añadimos VOS if 'castellano' in scrapedlanguage.lower() or ('español' in scrapedlanguage.lower() and not 'latino' in scrapedlanguage.lower()): item_local.language += ['CAST'] # añadimos CAST if 'dual' in item_local.quality.lower(): item_local.quality = re.sub(r'(?i)dual.*?', '', item_local.quality).strip() item_local.language += ['DUAL'] # añadimos DUAL if not item_local.language: item_local.language = ['LAT'] # [LAT] por defecto #Buscamos tamaño en el archivo .torrent size = '' if item_local.torrent_info: size = item_local.torrent_info elif scrapedsize: size = scrapedsize if not size and not item.videolibray_emergency_urls and not item_local.url.startswith('magnet:'): if not item.armagedon: size = generictools.get_torrent_size(item_local.url, local_torr=local_torr) #Buscamos el tamaño en el .torrent if 'ERROR' in size and item.emergency_urls and not item.videolibray_emergency_urls: item_local.armagedon = True try: # Restauramos la url if item.emergency_urls[0][0].startswith('http') or item.emergency_urls[0][0].startswith('//'): item_local.url = generictools.convert_url_base64(item.emergency_urls[0][0], host_torrent) else: item_local.url = generictools.convert_url_base64(item.emergency_urls[0][0]) if not item.url.startswith('http'): local_torr = filetools.join(config.get_videolibrary_path(), FOLDER, item_local.url) except: item_local.torrent_alt = '' item.emergency_urls[0] = [] size = generictools.get_torrent_size(item_local.url, local_torr=local_torr) if size: size = size.replace('GB', 'G·B').replace('Gb', 'G·b').replace('MB', 'M·B')\ .replace('Mb', 'M·b').replace('.', ',') item_local.torrent_info = '%s, ' % size #Agregamos size if item_local.url.startswith('magnet:') and not 'Magnet' in item_local.torrent_info: item_local.torrent_info += ' Magnet' if item_local.torrent_info: item_local.torrent_info = item_local.torrent_info.strip().strip(',') if item.videolibray_emergency_urls: item.torrent_info = item_local.torrent_info if not item.unify: item_local.torrent_info = '[%s]' % item_local.torrent_info # Guadamos la password del RAR password = scrapedpassword # Si tiene contraseña, la guardamos y la pintamos if password or item.password: if not item.password: item.password = password item_local.password = item.password itemlist.append(item.clone(action="", title="[COLOR magenta][B] Contraseña: [/B][/COLOR]'" + item_local.password + "'", folder=False)) # Guardamos urls de emergencia si se viene desde un Lookup de creación de Videoteca if item.videolibray_emergency_urls: item.emergency_urls[0].append(item_local.url) #guardamos la url y nos vamos continue if item_local.armagedon: item_local.quality = '[COLOR hotpink][E][/COLOR] [COLOR limegreen]%s[/COLOR]' % item_local.quality #Ahora pintamos el link del Torrent item_local.title = '[[COLOR yellow]?[/COLOR]] [COLOR yellow][Torrent][/COLOR] ' \ + '[COLOR limegreen][%s][/COLOR] [COLOR red]%s[/COLOR] %s' % \ (item_local.quality, str(item_local.language), \ item_local.torrent_info) #Preparamos título y calidad, quitando etiquetas vacías item_local.title = re.sub(r'\s?\[COLOR \w+\]\[\[?\s?\]?\]\[\/COLOR\]', '', item_local.title) item_local.title = re.sub(r'\s?\[COLOR \w+\]\s?\[\/COLOR\]', '', item_local.title) item_local.title = item_local.title.replace("--", "").replace("[]", "")\ .replace("()", "").replace("(/)", "").replace("[/]", "")\ .replace("|", "").strip() item_local.quality = re.sub(r'\s?\[COLOR \w+\]\[\[?\s?\]?\]\[\/COLOR\]', '', item_local.quality) item_local.quality = re.sub(r'\s?\[COLOR \w+\]\s?\[\/COLOR\]', '', item_local.quality) item_local.quality = item_local.quality.replace("--", "").replace("[]", "")\ .replace("()", "").replace("(/)", "").replace("[/]", "")\ .replace("|", "").strip() item_local.server = scrapedserver.lower() #Servidor if item_local.url.startswith('magnet:') or not item_local.server: item_local.server = 'torrent' if item_local.server != 'torrent': if config.get_setting("hidepremium"): #Si no se aceptan servidore premium, se ignoran if not servertools.is_server_enabled(item_local.server): continue devuelve = servertools.findvideosbyserver(item_local.url, item_local.server) #existe el link ? if not devuelve: continue item_local.url = devuelve[0][1]<|fim▁hole|> item_local.alive = servertools.check_video_link(item_local.url, item_local.server, timeout=timeout) #activo el link ? if 'NO' in item_local.alive: continue if not item_local.torrent_info or 'Magnet' in item_local.torrent_info: item_local.alive = "??" #Calidad del link sin verificar elif 'ERROR' in item_local.torrent_info and 'Pincha' in item_local.torrent_info: item_local.alive = "ok" #link en error, CF challenge, Chrome disponible elif 'ERROR' in item_local.torrent_info and 'Introduce' in item_local.torrent_info: item_local.alive = "??" #link en error, CF challenge, ruta de descarga no disponible item_local.channel = 'setting' item_local.action = 'setting_torrent' item_local.unify = False item_local.folder = False item_local.item_org = item.tourl() elif 'ERROR' in item_local.torrent_info: item_local.alive = "no" #Calidad del link en error, CF challenge? else: item_local.alive = "ok" #Calidad del link verificada if item_local.channel != 'setting': item_local.action = "play" #Visualizar vídeo itemlist_t.append(item_local.clone()) #Pintar pantalla, si no se filtran idiomas # Requerido para FilterTools if config.get_setting('filter_languages', channel) > 0: #Si hay idioma seleccionado, se filtra itemlist_f = filtertools.get_link(itemlist_f, item_local, list_language) #Pintar pantalla, si no está vacío #logger.debug("TORRENT: " + scrapedurl + " / title gen/torr: " + item.title + " / " + item_local.title + " / calidad: " + item_local.quality + " / content: " + item_local.contentTitle + " / " + item_local.contentSerieName) #logger.debug(item_local) #Si es un lookup para cargar las urls de emergencia en la Videoteca... if item.videolibray_emergency_urls: return item #... nos vamos if len(itemlist_f) > 0: #Si hay entradas filtradas... itemlist.extend(itemlist_f) #Pintamos pantalla filtrada else: if config.get_setting('filter_languages', channel) > 0 and len(itemlist_t) > 0: #Si no hay entradas filtradas ... thumb_separador = get_thumb("next.png") #... pintamos todo con aviso itemlist.append(Item(channel=item.channel, url=host, title="[COLOR red][B]NO hay elementos con el idioma seleccionado[/B][/COLOR]", thumbnail=thumb_separador, folder=False)) if len(itemlist_t) == 0: if len(itemlist) == 0 or (len(itemlist) > 0 and itemlist[-1].server != 'torrent'): return [] itemlist.extend(itemlist_t) #Pintar pantalla con todo si no hay filtrado # Requerido para AutoPlay autoplay.start(itemlist, item) #Lanzamos Autoplay return itemlist def episodios(item): logger.info() itemlist = [] item.category = categoria #logger.debug(item) if item.from_title: item.title = item.from_title #Limpiamos num. Temporada y Episodio que ha podido quedar por Novedades season_display = 0 if item.contentSeason: if item.season_colapse: #Si viene del menú de Temporadas... season_display = item.contentSeason #... salvamos el num de sesión a pintar item.from_num_season_colapse = season_display del item.season_colapse item.contentType = "tvshow" if item.from_title_season_colapse: item.title = item.from_title_season_colapse del item.from_title_season_colapse if item.infoLabels['title']: del item.infoLabels['title'] del item.infoLabels['season'] if item.contentEpisodeNumber: del item.infoLabels['episode'] if season_display == 0 and item.from_num_season_colapse: season_display = item.from_num_season_colapse # Obtener la información actualizada de la Serie. TMDB es imprescindible para Videoteca idioma = idioma_busqueda if 'VO' in str(item.language): idioma = idioma_busqueda_VO try: tmdb.set_infoLabels(item, True, idioma_busqueda=idioma) except: pass modo_ultima_temp_alt = modo_ultima_temp if item.ow_force == "1": #Si hay una migración de canal o url, se actualiza todo modo_ultima_temp_alt = False # Vemos la última temporada de TMDB y del .nfo max_temp = 1 if item.infoLabels['number_of_seasons']: max_temp = item.infoLabels['number_of_seasons'] y = [] if modo_ultima_temp_alt and item.library_playcounts: #Averiguar cuantas temporadas hay en Videoteca patron = 'season (\d+)' matches = re.compile(patron, re.DOTALL).findall(str(item.library_playcounts)) for x in matches: y += [int(x)] max_temp = max(y) # Si la series tiene solo una temporada, o se lista solo una temporada, guardamos la url y seguimos normalmente list_temp = [] list_temp.append(item.url) #logger.debug(list_temp) # Descarga las páginas for _url in list_temp: # Recorre todas las temporadas encontradas url = _url data, response, item, itemlist = generictools.downloadpage(url, timeout=timeout, canonical=canonical, s2=False, item=item, itemlist=itemlist) # Descargamos la página # Verificamos si ha cambiado el Host if response.host: for x, u in enumerate(list_temp): list_temp[x] = list_temp[x].replace(scrapertools.find_single_match(url, patron_host), response.host.rstrip('/')) url = response.url_new #Verificamos si se ha cargado una página, y si además tiene la estructura correcta if not response.sucess: # Si ERROR o lista de errores ... return itemlist # ... Salimos patron_temp = '<div\s*class="card-header">\s*<button\s*type="button"\s*data-toggle=' patron_temp += '"collapse"\s*data-target="#collapse-\d+">\s*<span>Temporada\s*(\d+)' patron_temp += '<\/span>(.*?)<\/table>\s*<\/div>\s*<\/div>\s*<\/div>' matches_temp = re.compile(patron_temp, re.DOTALL).findall(data) #logger.debug("PATRON_temp: " + patron_temp) #logger.debug(matches_temp) #logger.debug(data) patron = '<tr>(?:\s*<td>(\d+)<\/td>)?(?:\s*<td>([^<]*)<\/td>)?\s*<td>([^<]+)<\/td>' patron += '\s*<td>([^<]*)<\/td>\s*<td\s*class=[^<]+<\/td>(?:\s*<td>([^<]+)<\/td>)?' patron += '(?:\s*<td\s*class=[^<]+<\/td>)?\s*<td\s*class=[^<]+<\/td>\s*<td>\s*' patron += '<a\s*class="[^>]+href="([^"]+)"' for season_num, episodes in matches_temp: matches = re.compile(patron, re.DOTALL).findall(episodes) #logger.debug("PATRON: " + patron) #logger.debug(matches) #logger.debug(episodes) # Recorremos todos los episodios generando un Item local por cada uno en Itemlist for episode_num, scrapedserver, scrapedquality, scrapedlanguage, scrapedsize, scrapedurl in matches: server = scrapedserver if not server: server = 'torrent' item_local = item.clone() item_local.action = "findvideos" item_local.contentType = "episode" if item_local.library_playcounts: del item_local.library_playcounts if item_local.library_urls: del item_local.library_urls if item_local.path: del item_local.path if item_local.update_last: del item_local.update_last if item_local.update_next: del item_local.update_next if item_local.channel_host: del item_local.channel_host if item_local.active: del item_local.active if item_local.contentTitle: del item_local.infoLabels['title'] if item_local.season_colapse: del item_local.season_colapse item_local.url = url # Usamos las url de la temporada, no hay de episodio url_base64 = generictools.convert_url_base64(scrapedurl, host_torrent) item_local.matches = [] item_local.matches.append((episode_num, server, scrapedquality, scrapedlanguage, scrapedsize, url_base64)) # Salvado Matches de cada episodio item_local.context = "['buscar_trailer']" if not item_local.infoLabels['poster_path']: item_local.thumbnail = item_local.infoLabels['thumbnail'] # Extraemos números de temporada y episodio try: item_local.contentSeason = int(season_num) except: item_local.contentSeason = 1 try: item_local.contentEpisodeNumber = int(episode_num) except: item_local.contentEpisodeNumber = 0 item_local.title = '%sx%s - ' % (str(item_local.contentSeason), str(item_local.contentEpisodeNumber).zfill(2)) # Procesamos Calidad if scrapedquality: item_local.quality = scrapertools.remove_htmltags(scrapedquality) # iniciamos calidad if '[720p]' in scrapedquality.lower() or '720p' in scrapedquality.lower(): item_local.quality = '720p' if '[1080p]' in scrapedquality.lower() or '1080p' in scrapedquality.lower(): item_local.quality = '1080p' if '4k' in scrapedquality.lower(): item_local.quality = '4K' if '3d' in scrapedquality.lower() and not '3d' in item_local.quality.lower(): item_local.quality += ', 3D' if not item_local.quality: item_local.quality = '720p' # Comprobamos si hay más de un enlace por episodio, entonces los agrupamos if len(itemlist) > 0 and item_local.contentSeason == itemlist[-1].contentSeason \ and item_local.contentEpisodeNumber == itemlist[-1].contentEpisodeNumber \ and itemlist[-1].contentEpisodeNumber != 0: # solo guardamos un episodio ... if itemlist[-1].quality: if item_local.quality not in itemlist[-1].quality: itemlist[-1].quality += ", " + item_local.quality # ... pero acumulamos las calidades else: itemlist[-1].quality = item_local.quality itemlist[-1].matches.append(item_local.matches[0]) # Salvado Matches en el episodio anterior continue # ignoramos el episodio duplicado if season_display > 0: # Son de la temporada estos episodios? if item_local.contentSeason > season_display: break elif item_local.contentSeason < season_display: continue if modo_ultima_temp_alt and item.library_playcounts: # Si solo se actualiza la última temporada de Videoteca if item_local.contentSeason < max_temp: continue # Sale del bucle actual del FOR itemlist.append(item_local.clone()) #logger.debug(item_local) if len(itemlist) > 1: itemlist = sorted(itemlist, key=lambda it: (int(it.contentSeason), int(it.contentEpisodeNumber))) #clasificamos if item.season_colapse and not item.add_videolibrary: #Si viene de listado, mostramos solo Temporadas item, itemlist = generictools.post_tmdb_seasons(item, itemlist) if not item.season_colapse: #Si no es pantalla de Temporadas, pintamos todo # Pasada por TMDB y clasificación de lista por temporada y episodio tmdb.set_infoLabels(itemlist, True, idioma_busqueda=idioma) #Llamamos al método para el maquillaje de los títulos obtenidos desde TMDB item, itemlist = generictools.post_tmdb_episodios(item, itemlist) #logger.debug(item) return itemlist def actualizar_titulos(item): logger.info() #Llamamos al método que actualiza el título con tmdb.find_and_set_infoLabels item = generictools.update_title(item) #Volvemos a la siguiente acción en el canal return item def search(item, texto): logger.info() texto = texto.replace(" ", "+") try: item.url = host + 'buscar/page/1/?buscar=' + texto item.extra = 'search' if texto: return listado(item) else: return [] except: for line in sys.exc_info(): logger.error("{0}".format(line)) logger.error(traceback.format_exc(1)) return [] def newest(categoria): logger.info() itemlist = [] item = Item() item.title = "newest" item.category_new = "newest" item.channel = channel try: if categoria in ['peliculas', 'latino', 'torrent']: item.url = host + "peliculas/page/1/" if channel == 'yestorrent': item.url = host + "Descargar-peliculas-completas/page/1/" item.extra = "peliculas" item.extra2 = "novedades" item.action = "listado" itemlist.extend(listado(item)) if len(itemlist) > 0 and (">> Página siguiente" in itemlist[-1].title or "Pagina siguiente >>" in itemlist[-1].title): itemlist.pop() if categoria in ['series', 'latino', 'torrent']: item.category_new= 'newest' item.url = host + "series/page/1/" item.extra = "series" item.extra2 = "novedades" item.action = "listado" itemlist.extend(listado(item)) if len(itemlist) > 0 and (">> Página siguiente" in itemlist[-1].title or "Pagina siguiente >>" in itemlist[-1].title): itemlist.pop() # Se captura la excepción, para no interrumpir al canal novedades si un canal falla except: for line in sys.exc_info(): logger.error("{0}".format(line)) logger.error(traceback.format_exc(1)) return [] return itemlist<|fim▁end|>
<|file_name|>__openerp__.py<|end_file_name|><|fim▁begin|># -*- coding: utf-8 -*- # © <YEAR(S)> <AUTHOR(S)> # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl.html). { "name": "Account Analytic Warehouse", "summary": "Add analytic in stock_warehouse", "version": "9.0.1.0.0", "category": "Accounting", "website": "https://odoo-community.org/", "author": "<Deysy Mascorro>, Odoo Community Association (OCA)", "license": "AGPL-3", "application": False, "installable": True, "external_dependencies": { "python": [],<|fim▁hole|> "account", "stock" ], "data": [ "views/stock_warehouse_view.xml", "views/stock_location_view.xml", ], "demo": [ ], "qweb": [ ] }<|fim▁end|>
"bin": [], }, "depends": [ "base",
<|file_name|>hammer_v0.py<|end_file_name|><|fim▁begin|># coding=utf-8 # Copyright 2022 The Google Research Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Adroit hammer environment variant with RGB and proprioceptive observations. """ from d4rl.hand_manipulation_suite import HammerEnvV0 import numpy as np class VisualHammerEnvV0(HammerEnvV0): """Hammer environment with visual and proprioceptive observations.""" def __init__(self, camera_id, im_size, **kwargs): self._camera_id = camera_id self.im_size = im_size super().__init__(**kwargs) def get_obs(self): rgb = self.physics.render(<|fim▁hole|> sensordata = np.clip( self.physics.data.sensordata.ravel().copy()[:41], -5.0, 5.0) original_obs = super().get_obs() return {'rgb': rgb, 'qpos': qp[:-6], 'qvel': qv[:-6], 'palm_pos': palm_pos, 'tactile': sensordata, 'original_obs': original_obs}<|fim▁end|>
self.im_size, self.im_size, camera_id=self._camera_id) qp = self.physics.data.qpos.ravel().copy() qv = np.clip(self.physics.data.qvel.ravel(), -1.0, 1.0).copy() palm_pos = self.physics.data.site_xpos[self.S_grasp_sid].ravel().copy()
<|file_name|>outputstreamadaptor.cpp<|end_file_name|><|fim▁begin|>#include "bond/api/libio.h" #include "bond/io/outputstream.h" #include "bond/io/outputstreamadaptor.h" #include <cstdio> namespace Bond { void OutputStreamAdaptor::Print(const char *str) { const char *format = ((mFlags & IO::Left) != 0) ? "%-*s" : "%*s"; mStream->Print(format, mWidth, str); mWidth = 0; } void OutputStreamAdaptor::Print(bool value) { if ((mFlags & IO::BoolAlpha) != 0) { Print(value ? "true" : "false"); } else { Print(value ? int32_t(1) : int32_t(0)); } } void OutputStreamAdaptor::Print(char value) { const char *format = ((mFlags & IO::Left) != 0) ? "%-*c" : "%*c"; mStream->Print(format, mWidth, value); mWidth = 0; } void OutputStreamAdaptor::Print(int32_t value) { char format[16]; FormatInteger(format, BOND_PRId32, BOND_PRIx32, BOND_PRIo32); mStream->Print(format, mWidth, value); mWidth = 0; } void OutputStreamAdaptor::Print(uint32_t value) { char format[16]; FormatInteger(format, BOND_PRIu32, BOND_PRIx32, BOND_PRIo32); mStream->Print(format, mWidth, value); mWidth = 0; } void OutputStreamAdaptor::Print(int64_t value) { char format[16]; FormatInteger(format, BOND_PRId64, BOND_PRIx64, BOND_PRIo64); mStream->Print(format, mWidth, value); mWidth = 0; } void OutputStreamAdaptor::Print(uint64_t value) { char format[16]; FormatInteger(format, BOND_PRIu64, BOND_PRIx64, BOND_PRIo64); mStream->Print(format, mWidth, value); mWidth = 0; } void OutputStreamAdaptor::Print(double value) { char format[16]; FormatFloat(format); mStream->Print(format, mWidth, mPrecision, value); mWidth = 0; } void OutputStreamAdaptor::FormatInteger(char *format, const char *dec, const char *hex, const char *oct) const { *format++ = '%'; if ((mFlags & IO::Left) != 0) { *format++ = '-'; } if (((mFlags & IO::ShowBase) != 0) && ((mFlags & (IO::Hex | IO::Oct)) != 0)) { *format++ = '#'; } if ((mFlags & IO::Zero) != 0) {<|fim▁hole|> *format++ = '*'; const char *specifier = ((mFlags & IO::Hex) != 0) ? hex : ((mFlags & IO::Oct) != 0) ? oct : dec; while (*specifier) { *format++ = *specifier++; } *format++ = '\0'; } void OutputStreamAdaptor::FormatFloat(char *format) const { *format++ = '%'; if ((mFlags & IO::Left) != 0) { *format++ = '-'; } if ((mFlags & IO::ShowPoint) != 0) { *format++ = '#'; } if ((mFlags & IO::Zero) != 0) { *format++ = '0'; } *format++ = '*'; *format++ = '.'; *format++ = '*'; if ((mFlags & IO::Fixed) != 0) { *format++ = 'f'; } else if ((mFlags & IO::Scientific) != 0) { *format++ = 'e'; } else { *format++ = 'g'; } *format++ = '\0'; } }<|fim▁end|>
*format++ = '0'; }
<|file_name|>mod.rs<|end_file_name|><|fim▁begin|>// Copyright 2018 MaidSafe.net limited. // // This SAFE Network Software is licensed to you under The General Public License (GPL), version 3. // Unless required by applicable law or agreed to in writing, the SAFE Network Software distributed // under the GPL Licence is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY // KIND, either express or implied. Please review the Licences for the specific language governing // permissions and limitations relating to use of the SAFE Network Software. /// `FileHelper` provides functions for CRUD on file. pub mod file_helper; mod data_map; mod dir; mod errors; mod file; mod reader; #[cfg(test)] mod tests; mod writer; pub use self::dir::create_dir; pub use self::errors::NfsError; pub use self::file::File; pub use self::reader::Reader; pub use self::writer::{Mode, Writer};<|fim▁hole|>pub type NfsFuture<T> = Future<Item = T, Error = NfsError>;<|fim▁end|>
use futures::Future; /// Helper type for futures that can result in `NfsError`.
<|file_name|>p3.py<|end_file_name|><|fim▁begin|># $Id: p3.py,v 1.2 2003/11/18 19:04:03 phr Exp phr $ # Simple p3 encryption "algorithm": it's just SHA used as a stream # cipher in output feedback mode. # Author: Paul Rubin, Fort GNOX Cryptography, <phr-crypto at nightsong.com>. # Algorithmic advice from David Wagner, Richard Parker, Bryan # Olson, and Paul Crowley on sci.crypt is gratefully acknowledged. # Copyright 2002,2003 by Paul Rubin # Copying license: same as Python 2.3 license # Please include this revision number in any bug reports: $Revision: 1.2 $. from string import join from array import array try: import hashlib as sha except: import sha from time import time class CryptError(Exception): pass def _hash(str): return sha.new(str).digest() _ivlen = 16 _maclen = 8 _state = _hash(`time()`) try: import os _pid = `os.getpid()` except ImportError, AttributeError: _pid = '' def _expand_key(key, clen): blocks = (clen+19)/20 xkey=[] seed=key for i in xrange(blocks): seed=sha.new(key+seed).digest() xkey.append(seed) j = join(xkey,'') return array ('L', j) def p3_encrypt(plain,key): global _state H = _hash # change _state BEFORE using it to compute nonce, in case there's # a thread switch between computing the nonce and folding it into # the state. This way if two threads compute a nonce from the # same data, they won't both get the same nonce. (There's still # a small danger of a duplicate nonce--see below). _state = 'X'+_state # Attempt to make nlist unique for each call, so we can get a # unique nonce. It might be good to include a process ID or # something, but I don't know if that's portable between OS's. # Since is based partly on both the key and plaintext, in the # worst case (encrypting the same plaintext with the same key in # two separate Python instances at the same time), you might get # identical ciphertexts for the identical plaintexts, which would # be a security failure in some applications. Be careful. nlist = [`time()`, _pid, _state, `len(plain)`,plain, key] nonce = H(join(nlist,','))[:_ivlen] _state = H('update2'+_state+nonce) k_enc, k_auth = H('enc'+key+nonce), H('auth'+key+nonce) n=len(plain) # cipher size not counting IV stream = array('L', plain+'0000'[n&3:]) # pad to fill 32-bit words xkey = _expand_key(k_enc, n+4) for i in xrange(len(stream)): stream[i] = stream[i] ^ xkey[i] ct = nonce + stream.tostring()[:n] auth = _hmac(ct, k_auth) return ct + auth[:_maclen] def p3_decrypt(cipher,key): H = _hash n=len(cipher)-_ivlen-_maclen # length of ciphertext if n < 0: raise CryptError, "invalid ciphertext" nonce,stream,auth = \ cipher[:_ivlen], cipher[_ivlen:-_maclen]+'0000'[n&3:],cipher[-_maclen:] k_enc, k_auth = H('enc'+key+nonce), H('auth'+key+nonce) vauth = _hmac (cipher[:-_maclen], k_auth)[:_maclen] if auth != vauth: raise CryptError, "invalid key or ciphertext" stream = array('L', stream) xkey = _expand_key (k_enc, n+4) for i in xrange (len(stream)): stream[i] = stream[i] ^ xkey[i] plain = stream.tostring()[:n] return plain # RFC 2104 HMAC message authentication code # This implementation is faster than Python 2.2's hmac.py, and also works in # old Python versions (at least as old as 1.5.2). from string import translate def _hmac_setup(): global _ipad, _opad, _itrans, _otrans _itrans = array('B',[0]*256) _otrans = array('B',[0]*256) for i in xrange(256): _itrans[i] = i ^ 0x36 _otrans[i] = i ^ 0x5c _itrans = _itrans.tostring() _otrans = _otrans.tostring() _ipad = '\x36'*64 _opad = '\x5c'*64 def _hmac(msg, key): if len(key)>64: key=sha.new(key).digest() ki = (translate(key,_itrans)+_ipad)[:64] # inner<|fim▁hole|> ko = (translate(key,_otrans)+_opad)[:64] # outer return sha.new(ko+sha.new(ki+msg).digest()).digest() # # benchmark and unit test # def _time_p3(n=1000,len=20): plain="a"*len t=time() for i in xrange(n): p3_encrypt(plain,"abcdefgh") dt=time()-t print "plain p3:", n,len,dt,"sec =",n*len/dt,"bytes/sec" def _speed(): _time_p3(len=5) _time_p3() _time_p3(len=200) _time_p3(len=2000,n=100) def _test(): e=p3_encrypt d=p3_decrypt plain="test plaintext" key = "test key" c1 = e(plain,key) c2 = e(plain,key) assert c1!=c2 assert d(c2,key)==plain assert d(c1,key)==plain c3 = c2[:20]+chr(1+ord(c2[20]))+c2[21:] # change one ciphertext character try: print d(c3,key) # should throw exception print "auth verification failure" except CryptError: pass try: print d(c2,'wrong key') # should throw exception print "test failure" except CryptError: pass _hmac_setup() #_test() # _speed() # uncomment to run speed test<|fim▁end|>
<|file_name|>dnsblocker.go<|end_file_name|><|fim▁begin|>package main import ( "bufio" "flag" "fmt" "io" "os" "strings" "github.com/coocood/freecache" "github.com/miekg/dns" "github.com/op/go-logging" ) var logger = logging.MustGetLogger("dnsblocker") var logFormatter = logging.MustStringFormatter( "%{color}%{time} [%{level:.5s}] %{message}%{color:reset}", ) var dnsServer string var listen string var hostsPath string var logLevelStr string var blocked = []string{} var cacheSize int var cacheDuration int var cache *freecache.Cache func init() { flag.StringVar(&dnsServer, "dns-server", "192.168.1.1:domain", "DNS server for proxy not blocked queries") flag.StringVar(&listen, "listen", ":domain", "Listen is pair 'ip:port'") flag.StringVar(&hostsPath, "hosts-path", "hosts", "Path to hosts file") flag.StringVar(&logLevelStr, "log-level", "INFO", "Set minimum log level") flag.IntVar(&cacheSize, "cache-size", 1024*1024, "Set cache size into bytes") flag.IntVar(&cacheDuration, "cache-duration", 5*60, "Set cache duration into seconds") flag.Parse() configureLogger() configureCache() logger.Infof("Start listen on: %s", listen) logger.Infof("DNS source: %s", dnsServer) logger.Infof("Hosts file: %s", hostsPath) logger.Infof("Log level: %s", logLevelStr) logger.Infof("Cache size: %d", cacheSize) logger.Infof("Cache duration: %d", cacheDuration) loadBlocked() } func configureLogger() { logging.SetBackend( logging.MultiLogger( logging.NewBackendFormatter( logging.NewLogBackend(os.Stdout, "", 0), logFormatter, ), ), ) normalizedLogLevel := strings.ToUpper(logLevelStr) if logLevel, levelErr := logging.LogLevel(normalizedLogLevel); nil == levelErr { logging.SetLevel(logLevel, "") } else { logger.Fatal(levelErr) } } func loadBlocked() { file, openErr := os.Open(hostsPath) if nil != openErr { panic(openErr) } defer file.Close() buffer := bufio.NewReader(file) for { line, readErr := buffer.ReadString('\n') if readErr == io.EOF { break } else if nil != readErr { panic(readErr) } host := dns.Fqdn(line[0 : len(line)-1]) blocked = append(blocked, host) } logger.Infof("Loaded %d domains", len(blocked)) } func configureCache() { cache = freecache.NewCache(cacheSize) } func isBlocked(requestedName string) bool { for _, name := range blocked { if dns.IsSubDomain(name, requestedName) { return true } } return false } func messageCacheKey(message *dns.Msg) string { questions := make([]string, 0, len(message.Question)) for _, question := range message.Question { questionRow := fmt.Sprintf("%s %s %s", question.Name, dns.ClassToString[question.Qclass], dns.TypeToString[question.Qtype]) questions = append(questions, questionRow) } return strings.Join(questions, ",") } func fetchCache(cacheKey string) (*dns.Msg, error) { messageData, cacheErr := cache.Get([]byte(cacheKey)) if freecache.ErrNotFound == cacheErr { logger.Infof("Message for key %s not found in cache", cacheKey) return nil, cacheErr } if nil != cacheErr { logger.Errorf("Cache error for key %s: %s", cacheKey, cacheErr) return nil, cacheErr } logger.Infof("Message for key %s found in cache", cacheKey) dnsMessage := new(dns.Msg) if unpackErr := dnsMessage.Unpack(messageData); nil != unpackErr { logger.Errorf("Error unpack mesasge with key %s: %s", cacheKey, unpackErr)<|fim▁hole|> return nil, unpackErr } logger.Debug("Success unpack message") return dnsMessage, nil } func writeCache(cacheKey string, dnsMessage *dns.Msg) error { messageData, packErr := dnsMessage.Pack() if nil != packErr { logger.Errorf("Error pack message with key %s: %s", cacheKey, packErr) return packErr } logger.Debugf("Pack message with key %s is success", cacheKey) if cacheErr := cache.Set([]byte(cacheKey), messageData, cacheDuration); nil != cacheErr { logger.Errorf("Error write cache for key %s: %s", cacheKey, cacheErr) return cacheErr } logger.Debugf("Write cache for key %s is success", cacheKey) return nil } func dnsExchangeWithCache(requestMessage *dns.Msg) (*dns.Msg, error) { cacheKey := messageCacheKey(requestMessage) if cachedResponseMessage, fetchErr := fetchCache(cacheKey); nil == fetchErr { cachedResponseMessage.SetReply(requestMessage) return cachedResponseMessage, nil } else if freecache.ErrNotFound != fetchErr { return nil, fetchErr } responseMessage, exchangeErr := dns.Exchange(requestMessage, dnsServer) if nil != exchangeErr { logger.Errorf("Exchange error for key %s: %s", cacheKey, exchangeErr) return nil, exchangeErr } logger.Debugf("Exchange for key %s is success", cacheKey) if writeErr := writeCache(cacheKey, responseMessage); nil != writeErr { return nil, writeErr } return responseMessage, nil } func proxyRequest(writer dns.ResponseWriter, requestMessage *dns.Msg) { responseMessage, fetchErr := dnsExchangeWithCache(requestMessage) if nil == fetchErr { logger.Debug("Response message: %+v", responseMessage) writeResponse(writer, responseMessage) } else { logger.Errorf("Fetch error: %s", fetchErr) errorMessage := new(dns.Msg) errorMessage.SetRcode(requestMessage, dns.RcodeServerFailure) logger.Errorf("Error message: %+v", errorMessage) writeResponse(writer, errorMessage) } } func errorResponse(writer dns.ResponseWriter, requestMessage *dns.Msg, responseCode int) { responseMessage := new(dns.Msg) responseMessage.SetRcode(requestMessage, responseCode) logger.Debugf("Block response: %+v", responseMessage) writeResponse(writer, responseMessage) } func handler(writer dns.ResponseWriter, requestMessage *dns.Msg) { logger.Debugf("Query message: %+v", requestMessage) logger.Infof("Accept request from %v", writer.RemoteAddr()) if 1 != len(requestMessage.Question) { logger.Warning("Can not process message with multiple questions") errorResponse(writer, requestMessage, dns.RcodeFormatError) return } question := requestMessage.Question[0] logger.Infof("Request name: %s", question.Name) if isBlocked(question.Name) { logger.Infof("Block name: %s", question.Name) errorResponse(writer, requestMessage, dns.RcodeRefused) } else { logger.Infof("Request name %s is proxed", question.Name) proxyRequest(writer, requestMessage) } } func writeResponse(writer dns.ResponseWriter, message *dns.Msg) { if writeErr := writer.WriteMsg(message); nil == writeErr { logger.Debug("Writer success") } else { logger.Errorf("Writer error: %s", writeErr) } } func main() { server := &dns.Server{ Addr: listen, Net: "udp", Handler: dns.HandlerFunc(handler), } if serverErr := server.ListenAndServe(); nil != serverErr { logger.Fatal(serverErr) } }<|fim▁end|>
<|file_name|>environment.js<|end_file_name|><|fim▁begin|>module.exports = function() { return { modulePrefix: 'some-cool-app', EmberENV: { asdflkmawejf: ';jlnu3yr23'<|fim▁hole|> autoboot: false } }; };<|fim▁end|>
}, APP: {
<|file_name|>EncodeBufr302053.cc<|end_file_name|><|fim▁begin|>/* Kvalobs - Free Quality Control Software for Meteorological Observations $Id: Data.h,v 1.2.6.2 2007/09/27 09:02:22 paule Exp $ Copyright (C) 2007 met.no Contact information: Norwegian Meteorological Institute Box 43 Blindern 0313 OSLO NORWAY email: [email protected] This file is part of KVALOBS KVALOBS is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. KVALOBS is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with KVALOBS; if not, write to the Free Software Foundation Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA */ #include <boost/assign.hpp> #include "EncodeBufr302053.h" EncodeBufr302053:: EncodeBufr302053() { } std::string EncodeBufr302053::<|fim▁hole|>{ return "302053"; } std::list<int> EncodeBufr302053:: encodeIds()const { std::list<int> ids; boost::assign::push_back( ids )(302053); return ids; } void EncodeBufr302053:: encode( ) { bufr->addValue( 7032, stationInfo->heightVisability(), "h_V, height of visibility sensor above marine deck.", false ); bufr->addValue( 7033, FLT_MAX, "h_V, height of visibility sensor above water surface.", false ); bufr->addValue( 20001, data->VV, "VV, horisental visibility" ); }<|fim▁end|>
logIdentifier() const
<|file_name|>snake_game.py<|end_file_name|><|fim▁begin|>#!/usr/bin/python # Copyright (C) 2013 rapidhere # # Author: rapidhere <[email protected]> # Maintainer: rapidhere <[email protected]> # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. import skapp from optparse import OptionParser import sys parser = OptionParser( usage = "%prog [options]", description = """A simple snake game.Suggest that resize your terminal window at a property size befor playing!""", epilog = "[email protected]", version = "0.1" ) parser.add_option( "","--key-help", action = "store_true",default = False, help = "show game keys" ) opts,args = parser.parse_args() parser.destroy() if opts.key_help: print "'w' or 'W' or UP-Arrow up" print "'a' or 'A' or LF-Arrow left" print "'s' or 'S' or DW-Arrow down"<|fim▁hole|> app = skapp.SKApp() app.run()<|fim▁end|>
print "'d' or 'D' or RG-Arrpw right" print "'q' or 'Q' quit" sys.exit(0) else:
<|file_name|>views.py<|end_file_name|><|fim▁begin|>""" Django Views for service status app """ <|fim▁hole|>import json import time from celery.exceptions import TimeoutError from django.http import HttpResponse from djcelery import celery from openedx.core.djangoapps.service_status.tasks import delayed_ping def index(_): """ An empty view """ return HttpResponse() def celery_status(_): """ A view that returns Celery stats """ stats = celery.control.inspect().stats() or {} return HttpResponse(json.dumps(stats, indent=4), content_type="application/json") def celery_ping(_): """ A Simple view that checks if Celery can process a simple task """ start = time.time() result = delayed_ping.apply_async(('ping', 0.1)) task_id = result.id # Wait until we get the result try: value = result.get(timeout=4.0) success = True except TimeoutError: value = None success = False output = { 'success': success, 'task_id': task_id, 'value': value, 'time': time.time() - start, } return HttpResponse(json.dumps(output, indent=4), content_type="application/json")<|fim▁end|>
from __future__ import absolute_import
<|file_name|>script_runtime.rs<|end_file_name|><|fim▁begin|>/* This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this * file, You can obtain one at https://mozilla.org/MPL/2.0/. */ //! The script runtime contains common traits and structs commonly used by the //! script thread, the dom, and the worker threads. #![allow(dead_code)] use crate::body::BodyMixin; use crate::dom::bindings::codegen::Bindings::PromiseBinding::PromiseJobCallback; use crate::dom::bindings::codegen::Bindings::ResponseBinding::ResponseBinding::ResponseMethods; use crate::dom::bindings::codegen::Bindings::ResponseBinding::ResponseType as DOMResponseType; use crate::dom::bindings::conversions::get_dom_class; use crate::dom::bindings::conversions::private_from_object; use crate::dom::bindings::conversions::root_from_handleobject; use crate::dom::bindings::error::{throw_dom_exception, Error}; use crate::dom::bindings::inheritance::Castable; use crate::dom::bindings::refcounted::{trace_refcounted_objects, LiveDOMReferences}; use crate::dom::bindings::refcounted::{Trusted, TrustedPromise}; use crate::dom::bindings::reflector::DomObject; use crate::dom::bindings::root::trace_roots; use crate::dom::bindings::settings_stack; use crate::dom::bindings::trace::{trace_traceables, JSTraceable}; use crate::dom::bindings::utils::DOM_CALLBACKS; use crate::dom::event::{Event, EventBubbles, EventCancelable, EventStatus}; use crate::dom::eventtarget::EventTarget; use crate::dom::globalscope::GlobalScope; use crate::dom::promise::Promise; use crate::dom::promiserejectionevent::PromiseRejectionEvent; use crate::dom::response::Response; use crate::microtask::{EnqueuedPromiseCallback, Microtask, MicrotaskQueue}; use crate::realms::{AlreadyInRealm, InRealm}; use crate::script_module::EnsureModuleHooksInitialized; use crate::script_thread::trace_thread; use crate::task::TaskBox; use crate::task_source::networking::NetworkingTaskSource; use crate::task_source::{TaskSource, TaskSourceName}; use js::glue::{CollectServoSizes, CreateJobQueue, DeleteJobQueue, DispatchableRun}; use js::glue::{JobQueueTraps, RUST_js_GetErrorMessage, SetBuildId, StreamConsumerConsumeChunk}; use js::glue::{ StreamConsumerNoteResponseURLs, StreamConsumerStreamEnd, StreamConsumerStreamError, }; use js::jsapi::ContextOptionsRef; use js::jsapi::GetPromiseUserInputEventHandlingState; use js::jsapi::InitConsumeStreamCallback; use js::jsapi::InitDispatchToEventLoop; use js::jsapi::MimeType; use js::jsapi::PromiseUserInputEventHandlingState; use js::jsapi::StreamConsumer as JSStreamConsumer; use js::jsapi::{BuildIdCharVector, DisableIncrementalGC, GCDescription, GCProgress}; use js::jsapi::{Dispatchable as JSRunnable, Dispatchable_MaybeShuttingDown}; use js::jsapi::{HandleObject, Heap, JobQueue}; use js::jsapi::{JSContext as RawJSContext, JSTracer, SetDOMCallbacks, SetGCSliceCallback}; use js::jsapi::{ JSGCInvocationKind, JSGCStatus, JS_AddExtraGCRootsTracer, JS_RequestInterruptCallback, JS_SetGCCallback, }; use js::jsapi::{JSGCMode, JSGCParamKey, JS_SetGCParameter, JS_SetGlobalJitCompilerOption}; use js::jsapi::{ JSJitCompilerOption, JS_SetOffthreadIonCompilationEnabled, JS_SetParallelParsingEnabled, }; use js::jsapi::{JSObject, PromiseRejectionHandlingState, SetPreserveWrapperCallback}; use js::jsapi::{SetJobQueue, SetProcessBuildIdOp, SetPromiseRejectionTrackerCallback}; use js::jsval::UndefinedValue; use js::panic::wrap_panic; use js::rust::wrappers::{GetPromiseIsHandled, JS_GetPromiseResult}; use js::rust::Handle; use js::rust::HandleObject as RustHandleObject; use js::rust::IntoHandle; use js::rust::ParentRuntime; use js::rust::Runtime as RustRuntime; use js::rust::{JSEngine, JSEngineHandle}; use malloc_size_of::MallocSizeOfOps; use msg::constellation_msg::PipelineId; use profile_traits::mem::{Report, ReportKind, ReportsChan}; use servo_config::opts; use servo_config::pref; use std::cell::Cell; use std::ffi::CString; use std::fmt; use std::io::{stdout, Write}; use std::ops::Deref; use std::os; use std::os::raw::c_void; use std::ptr; use std::rc::Rc; use std::sync::Mutex; use std::thread; use std::time::Duration; use style::thread_state::{self, ThreadState}; use time::{now, Tm}; static JOB_QUEUE_TRAPS: JobQueueTraps = JobQueueTraps { getIncumbentGlobal: Some(get_incumbent_global), enqueuePromiseJob: Some(enqueue_promise_job), empty: Some(empty), }; /// Common messages used to control the event loops in both the script and the worker pub enum CommonScriptMsg { /// Requests that the script thread measure its memory usage. The results are sent back via the /// supplied channel. CollectReports(ReportsChan), /// Generic message that encapsulates event handling. Task( ScriptThreadEventCategory, Box<dyn TaskBox>, Option<PipelineId>, TaskSourceName, ), } impl fmt::Debug for CommonScriptMsg { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { match *self { CommonScriptMsg::CollectReports(_) => write!(f, "CollectReports(...)"), CommonScriptMsg::Task(ref category, ref task, _, _) => { f.debug_tuple("Task").field(category).field(task).finish() }, } } } /// A cloneable interface for communicating with an event loop. pub trait ScriptChan: JSTraceable { /// Send a message to the associated event loop. fn send(&self, msg: CommonScriptMsg) -> Result<(), ()>; /// Clone this handle. fn clone(&self) -> Box<dyn ScriptChan + Send>; } #[derive(Clone, Copy, Debug, Eq, Hash, JSTraceable, PartialEq)] pub enum ScriptThreadEventCategory { AttachLayout, ConstellationMsg, DevtoolsMsg, DocumentEvent, DomEvent, FileRead, FormPlannedNavigation, HistoryEvent, ImageCacheMsg, InputEvent, NetworkEvent, PortMessage, Resize, ScriptEvent, SetScrollState, SetViewport, StylesheetLoad, TimerEvent, UpdateReplacedElement, WebSocketEvent, WorkerEvent, WorkletEvent, ServiceWorkerEvent, EnterFullscreen, ExitFullscreen, PerformanceTimelineTask, WebGPUMsg, } /// An interface for receiving ScriptMsg values in an event loop. Used for synchronous DOM /// APIs that need to abstract over multiple kinds of event loops (worker/main thread) with /// different Receiver interfaces. pub trait ScriptPort { fn recv(&self) -> Result<CommonScriptMsg, ()>; } #[allow(unsafe_code)] unsafe extern "C" fn get_incumbent_global(_: *const c_void, _: *mut RawJSContext) -> *mut JSObject { let mut result = ptr::null_mut(); wrap_panic(&mut || { let incumbent_global = GlobalScope::incumbent(); assert!(incumbent_global.is_some()); result = incumbent_global .map(|g| g.reflector().get_jsobject().get()) .unwrap_or(ptr::null_mut()) }); result } #[allow(unsafe_code)] unsafe extern "C" fn empty(extra: *const c_void) -> bool { let mut result = false; wrap_panic(&mut || { let microtask_queue = &*(extra as *const MicrotaskQueue); result = microtask_queue.empty() }); result } /// SM callback for promise job resolution. Adds a promise callback to the current /// global's microtask queue. #[allow(unsafe_code)] unsafe extern "C" fn enqueue_promise_job( extra: *const c_void, cx: *mut RawJSContext, promise: HandleObject, job: HandleObject, _allocation_site: HandleObject, incumbent_global: HandleObject, ) -> bool { let cx = JSContext::from_ptr(cx); let mut result = false; wrap_panic(&mut || { let microtask_queue = &*(extra as *const MicrotaskQueue); let global = if !incumbent_global.is_null() {<|fim▁hole|> let realm = AlreadyInRealm::assert_for_cx(cx); GlobalScope::from_context(*cx, InRealm::in_realm(&realm)) }; let pipeline = global.pipeline_id(); let interaction = if promise.get().is_null() { PromiseUserInputEventHandlingState::DontCare } else { GetPromiseUserInputEventHandlingState(promise) }; let is_user_interacting = interaction == PromiseUserInputEventHandlingState::HadUserInteractionAtCreation; microtask_queue.enqueue( Microtask::Promise(EnqueuedPromiseCallback { callback: PromiseJobCallback::new(cx, job.get()), pipeline, is_user_interacting, }), cx, ); result = true }); result } #[allow(unsafe_code, unrooted_must_root)] /// https://html.spec.whatwg.org/multipage/#the-hostpromiserejectiontracker-implementation unsafe extern "C" fn promise_rejection_tracker( cx: *mut RawJSContext, _muted_errors: bool, promise: HandleObject, state: PromiseRejectionHandlingState, _data: *mut c_void, ) { // TODO: Step 2 - If script's muted errors is true, terminate these steps. // Step 3. let cx = JSContext::from_ptr(cx); let in_realm_proof = AlreadyInRealm::assert_for_cx(cx); let global = GlobalScope::from_context(*cx, InRealm::Already(&in_realm_proof)); wrap_panic(&mut || { match state { // Step 4. PromiseRejectionHandlingState::Unhandled => { global.add_uncaught_rejection(promise); }, // Step 5. PromiseRejectionHandlingState::Handled => { // Step 5-1. if global .get_uncaught_rejections() .borrow() .contains(&Heap::boxed(promise.get())) { global.remove_uncaught_rejection(promise); return; } // Step 5-2. if !global .get_consumed_rejections() .borrow() .contains(&Heap::boxed(promise.get())) { return; } // Step 5-3. global.remove_consumed_rejection(promise); let target = Trusted::new(global.upcast::<EventTarget>()); let promise = Promise::new_with_js_promise(Handle::from_raw(promise), cx); let trusted_promise = TrustedPromise::new(promise.clone()); // Step 5-4. global.dom_manipulation_task_source().queue( task!(rejection_handled_event: move || { let target = target.root(); let cx = target.global().get_cx(); let root_promise = trusted_promise.root(); rooted!(in(*cx) let mut reason = UndefinedValue()); JS_GetPromiseResult(root_promise.reflector().get_jsobject(), reason.handle_mut()); let event = PromiseRejectionEvent::new( &target.global(), atom!("rejectionhandled"), EventBubbles::DoesNotBubble, EventCancelable::Cancelable, root_promise, reason.handle() ); event.upcast::<Event>().fire(&target); }), global.upcast(), ).unwrap(); }, }; }) } #[allow(unsafe_code, unrooted_must_root)] /// https://html.spec.whatwg.org/multipage/#notify-about-rejected-promises pub fn notify_about_rejected_promises(global: &GlobalScope) { let cx = global.get_cx(); unsafe { // Step 2. if global.get_uncaught_rejections().borrow().len() > 0 { // Step 1. let uncaught_rejections: Vec<TrustedPromise> = global .get_uncaught_rejections() .borrow() .iter() .map(|promise| { let promise = Promise::new_with_js_promise(Handle::from_raw(promise.handle()), cx); TrustedPromise::new(promise) }) .collect(); // Step 3. global.get_uncaught_rejections().borrow_mut().clear(); let target = Trusted::new(global.upcast::<EventTarget>()); // Step 4. global.dom_manipulation_task_source().queue( task!(unhandled_rejection_event: move || { let target = target.root(); let cx = target.global().get_cx(); for promise in uncaught_rejections { let promise = promise.root(); // Step 4-1. let promise_is_handled = GetPromiseIsHandled(promise.reflector().get_jsobject()); if promise_is_handled { continue; } // Step 4-2. rooted!(in(*cx) let mut reason = UndefinedValue()); JS_GetPromiseResult(promise.reflector().get_jsobject(), reason.handle_mut()); let event = PromiseRejectionEvent::new( &target.global(), atom!("unhandledrejection"), EventBubbles::DoesNotBubble, EventCancelable::Cancelable, promise.clone(), reason.handle() ); let event_status = event.upcast::<Event>().fire(&target); // Step 4-3. if event_status == EventStatus::Canceled { // TODO: The promise rejection is not handled; we need to add it back to the list. } // Step 4-4. if !promise_is_handled { target.global().add_consumed_rejection(promise.reflector().get_jsobject().into_handle()); } } }), global.upcast(), ).unwrap(); } } } #[derive(JSTraceable)] pub struct Runtime { rt: RustRuntime, pub microtask_queue: Rc<MicrotaskQueue>, job_queue: *mut JobQueue, } impl Drop for Runtime { #[allow(unsafe_code)] fn drop(&mut self) { unsafe { DeleteJobQueue(self.job_queue); } THREAD_ACTIVE.with(|t| { LiveDOMReferences::destruct(); t.set(false); }); } } impl Deref for Runtime { type Target = RustRuntime; fn deref(&self) -> &RustRuntime { &self.rt } } pub struct JSEngineSetup(JSEngine); impl JSEngineSetup { pub fn new() -> Self { let engine = JSEngine::init().unwrap(); *JS_ENGINE.lock().unwrap() = Some(engine.handle()); Self(engine) } } impl Drop for JSEngineSetup { fn drop(&mut self) { *JS_ENGINE.lock().unwrap() = None; while !self.0.can_shutdown() { thread::sleep(Duration::from_millis(50)); } } } lazy_static! { static ref JS_ENGINE: Mutex<Option<JSEngineHandle>> = Mutex::new(None); } #[allow(unsafe_code)] pub unsafe fn new_child_runtime( parent: ParentRuntime, networking_task_source: Option<NetworkingTaskSource>, ) -> Runtime { new_rt_and_cx_with_parent(Some(parent), networking_task_source) } #[allow(unsafe_code)] pub fn new_rt_and_cx(networking_task_source: Option<NetworkingTaskSource>) -> Runtime { unsafe { new_rt_and_cx_with_parent(None, networking_task_source) } } #[allow(unsafe_code)] unsafe fn new_rt_and_cx_with_parent( parent: Option<ParentRuntime>, networking_task_source: Option<NetworkingTaskSource>, ) -> Runtime { LiveDOMReferences::initialize(); let (cx, runtime) = if let Some(parent) = parent { let runtime = RustRuntime::create_with_parent(parent); let cx = runtime.cx(); (cx, runtime) } else { let runtime = RustRuntime::new(JS_ENGINE.lock().unwrap().as_ref().unwrap().clone()); (runtime.cx(), runtime) }; JS_AddExtraGCRootsTracer(cx, Some(trace_rust_roots), ptr::null_mut()); // Needed for debug assertions about whether GC is running. if cfg!(debug_assertions) { JS_SetGCCallback(cx, Some(debug_gc_callback), ptr::null_mut()); } if opts::get().gc_profile { SetGCSliceCallback(cx, Some(gc_slice_callback)); } unsafe extern "C" fn empty_wrapper_callback(_: *mut RawJSContext, _: HandleObject) -> bool { true } SetDOMCallbacks(cx, &DOM_CALLBACKS); SetPreserveWrapperCallback(cx, Some(empty_wrapper_callback)); // Pre barriers aren't working correctly at the moment DisableIncrementalGC(cx); unsafe extern "C" fn dispatch_to_event_loop( closure: *mut c_void, dispatchable: *mut JSRunnable, ) -> bool { let networking_task_src: &NetworkingTaskSource = &*(closure as *mut NetworkingTaskSource); let runnable = Runnable(dispatchable); let task = task!(dispatch_to_event_loop_message: move || { runnable.run(RustRuntime::get(), Dispatchable_MaybeShuttingDown::NotShuttingDown); }); networking_task_src.queue_unconditionally(task).is_ok() } if let Some(source) = networking_task_source { let networking_task_src = Box::new(source); InitDispatchToEventLoop( cx, Some(dispatch_to_event_loop), Box::into_raw(networking_task_src) as *mut c_void, ); } InitConsumeStreamCallback(cx, Some(consume_stream), Some(report_stream_error)); let microtask_queue = Rc::new(MicrotaskQueue::default()); let job_queue = CreateJobQueue( &JOB_QUEUE_TRAPS, &*microtask_queue as *const _ as *const c_void, ); SetJobQueue(cx, job_queue); SetPromiseRejectionTrackerCallback(cx, Some(promise_rejection_tracker), ptr::null_mut()); EnsureModuleHooksInitialized(runtime.rt()); set_gc_zeal_options(cx); // Enable or disable the JITs. let cx_opts = &mut *ContextOptionsRef(cx); JS_SetGlobalJitCompilerOption( cx, JSJitCompilerOption::JSJITCOMPILER_BASELINE_ENABLE, pref!(js.baseline.enabled) as u32, ); JS_SetGlobalJitCompilerOption( cx, JSJitCompilerOption::JSJITCOMPILER_ION_ENABLE, pref!(js.ion.enabled) as u32, ); cx_opts.set_asmJS_(pref!(js.asmjs.enabled)); let wasm_enabled = pref!(js.wasm.enabled); cx_opts.set_wasm_(wasm_enabled); if wasm_enabled { // If WASM is enabled without setting the buildIdOp, // initializing a module will report an out of memory error. // https://dxr.mozilla.org/mozilla-central/source/js/src/wasm/WasmTypes.cpp#458 SetProcessBuildIdOp(Some(servo_build_id)); } cx_opts.set_wasmBaseline_(pref!(js.wasm.baseline.enabled)); cx_opts.set_wasmIon_(pref!(js.wasm.ion.enabled)); cx_opts.set_extraWarnings_(pref!(js.strict.enabled)); // TODO: handle js.strict.debug.enabled // TODO: handle js.throw_on_asmjs_validation_failure (needs new Spidermonkey) JS_SetGlobalJitCompilerOption( cx, JSJitCompilerOption::JSJITCOMPILER_NATIVE_REGEXP_ENABLE, pref!(js.native_regex.enabled) as u32, ); JS_SetParallelParsingEnabled(cx, pref!(js.parallel_parsing.enabled)); JS_SetOffthreadIonCompilationEnabled(cx, pref!(js.offthread_compilation.enabled)); JS_SetGlobalJitCompilerOption( cx, JSJitCompilerOption::JSJITCOMPILER_BASELINE_WARMUP_TRIGGER, if pref!(js.baseline.unsafe_eager_compilation.enabled) { 0 } else { u32::max_value() }, ); JS_SetGlobalJitCompilerOption( cx, JSJitCompilerOption::JSJITCOMPILER_ION_NORMAL_WARMUP_TRIGGER, if pref!(js.ion.unsafe_eager_compilation.enabled) { 0 } else { u32::max_value() }, ); // TODO: handle js.discard_system_source.enabled // TODO: handle js.asyncstack.enabled (needs new Spidermonkey) // TODO: handle js.throw_on_debugee_would_run (needs new Spidermonkey) // TODO: handle js.dump_stack_on_debugee_would_run (needs new Spidermonkey) cx_opts.set_werror_(pref!(js.werror.enabled)); // TODO: handle js.shared_memory.enabled JS_SetGCParameter( cx, JSGCParamKey::JSGC_MAX_BYTES, in_range(pref!(js.mem.max), 1, 0x100) .map(|val| (val * 1024 * 1024) as u32) .unwrap_or(u32::max_value()), ); // NOTE: This is disabled above, so enabling it here will do nothing for now. let js_gc_mode = if pref!(js.mem.gc.incremental.enabled) { JSGCMode::JSGC_MODE_INCREMENTAL } else if pref!(js.mem.gc.per_zone.enabled) { JSGCMode::JSGC_MODE_ZONE } else { JSGCMode::JSGC_MODE_GLOBAL }; JS_SetGCParameter(cx, JSGCParamKey::JSGC_MODE, js_gc_mode as u32); if let Some(val) = in_range(pref!(js.mem.gc.incremental.slice_ms), 0, 100_000) { JS_SetGCParameter(cx, JSGCParamKey::JSGC_SLICE_TIME_BUDGET_MS, val as u32); } JS_SetGCParameter( cx, JSGCParamKey::JSGC_COMPACTING_ENABLED, pref!(js.mem.gc.compacting.enabled) as u32, ); if let Some(val) = in_range(pref!(js.mem.gc.high_frequency_time_limit_ms), 0, 10_000) { JS_SetGCParameter(cx, JSGCParamKey::JSGC_HIGH_FREQUENCY_TIME_LIMIT, val as u32); } JS_SetGCParameter( cx, JSGCParamKey::JSGC_DYNAMIC_MARK_SLICE, pref!(js.mem.gc.dynamic_mark_slice.enabled) as u32, ); JS_SetGCParameter( cx, JSGCParamKey::JSGC_DYNAMIC_HEAP_GROWTH, pref!(js.mem.gc.dynamic_heap_growth.enabled) as u32, ); if let Some(val) = in_range(pref!(js.mem.gc.low_frequency_heap_growth), 0, 10_000) { JS_SetGCParameter(cx, JSGCParamKey::JSGC_LOW_FREQUENCY_HEAP_GROWTH, val as u32); } if let Some(val) = in_range(pref!(js.mem.gc.high_frequency_heap_growth_min), 0, 10_000) { JS_SetGCParameter( cx, JSGCParamKey::JSGC_HIGH_FREQUENCY_HEAP_GROWTH_MIN, val as u32, ); } if let Some(val) = in_range(pref!(js.mem.gc.high_frequency_heap_growth_max), 0, 10_000) { JS_SetGCParameter( cx, JSGCParamKey::JSGC_HIGH_FREQUENCY_HEAP_GROWTH_MAX, val as u32, ); } if let Some(val) = in_range(pref!(js.mem.gc.high_frequency_low_limit_mb), 0, 10_000) { JS_SetGCParameter(cx, JSGCParamKey::JSGC_HIGH_FREQUENCY_LOW_LIMIT, val as u32); } if let Some(val) = in_range(pref!(js.mem.gc.high_frequency_high_limit_mb), 0, 10_000) { JS_SetGCParameter(cx, JSGCParamKey::JSGC_HIGH_FREQUENCY_HIGH_LIMIT, val as u32); } if let Some(val) = in_range(pref!(js.mem.gc.allocation_threshold_factor), 0, 10_000) { JS_SetGCParameter(cx, JSGCParamKey::JSGC_NON_INCREMENTAL_FACTOR, val as u32); } if let Some(val) = in_range( pref!(js.mem.gc.allocation_threshold_avoid_interrupt_factor), 0, 10_000, ) { JS_SetGCParameter(cx, JSGCParamKey::JSGC_AVOID_INTERRUPT_FACTOR, val as u32); } if let Some(val) = in_range(pref!(js.mem.gc.empty_chunk_count_min), 0, 10_000) { JS_SetGCParameter(cx, JSGCParamKey::JSGC_MIN_EMPTY_CHUNK_COUNT, val as u32); } if let Some(val) = in_range(pref!(js.mem.gc.empty_chunk_count_max), 0, 10_000) { JS_SetGCParameter(cx, JSGCParamKey::JSGC_MAX_EMPTY_CHUNK_COUNT, val as u32); } Runtime { rt: runtime, microtask_queue, job_queue, } } fn in_range<T: PartialOrd + Copy>(val: T, min: T, max: T) -> Option<T> { if val < min || val >= max { None } else { Some(val) } } #[allow(unsafe_code)] unsafe extern "C" fn get_size(obj: *mut JSObject) -> usize { match get_dom_class(obj) { Ok(v) => { let dom_object = private_from_object(obj) as *const c_void; if dom_object.is_null() { return 0; } let mut ops = MallocSizeOfOps::new(servo_allocator::usable_size, None, None); (v.malloc_size_of)(&mut ops, dom_object) }, Err(_e) => { return 0; }, } } #[allow(unsafe_code)] pub fn get_reports(cx: *mut RawJSContext, path_seg: String) -> Vec<Report> { let mut reports = vec![]; unsafe { let mut stats = ::std::mem::zeroed(); if CollectServoSizes(cx, &mut stats, Some(get_size)) { let mut report = |mut path_suffix, kind, size| { let mut path = path![path_seg, "js"]; path.append(&mut path_suffix); reports.push(Report { path: path, kind: kind, size: size as usize, }) }; // A note about possibly confusing terminology: the JS GC "heap" is allocated via // mmap/VirtualAlloc, which means it's not on the malloc "heap", so we use // `ExplicitNonHeapSize` as its kind. report( path!["gc-heap", "used"], ReportKind::ExplicitNonHeapSize, stats.gcHeapUsed, ); report( path!["gc-heap", "unused"], ReportKind::ExplicitNonHeapSize, stats.gcHeapUnused, ); report( path!["gc-heap", "admin"], ReportKind::ExplicitNonHeapSize, stats.gcHeapAdmin, ); report( path!["gc-heap", "decommitted"], ReportKind::ExplicitNonHeapSize, stats.gcHeapDecommitted, ); // SpiderMonkey uses the system heap, not jemalloc. report( path!["malloc-heap"], ReportKind::ExplicitSystemHeapSize, stats.mallocHeap, ); report( path!["non-heap"], ReportKind::ExplicitNonHeapSize, stats.nonHeap, ); } } reports } thread_local!(static GC_CYCLE_START: Cell<Option<Tm>> = Cell::new(None)); thread_local!(static GC_SLICE_START: Cell<Option<Tm>> = Cell::new(None)); #[allow(unsafe_code)] unsafe extern "C" fn gc_slice_callback( _cx: *mut RawJSContext, progress: GCProgress, desc: *const GCDescription, ) { match progress { GCProgress::GC_CYCLE_BEGIN => GC_CYCLE_START.with(|start| { start.set(Some(now())); println!("GC cycle began"); }), GCProgress::GC_SLICE_BEGIN => GC_SLICE_START.with(|start| { start.set(Some(now())); println!("GC slice began"); }), GCProgress::GC_SLICE_END => GC_SLICE_START.with(|start| { let dur = now() - start.get().unwrap(); start.set(None); println!("GC slice ended: duration={}", dur); }), GCProgress::GC_CYCLE_END => GC_CYCLE_START.with(|start| { let dur = now() - start.get().unwrap(); start.set(None); println!("GC cycle ended: duration={}", dur); }), }; if !desc.is_null() { let desc: &GCDescription = &*desc; let invocation_kind = match desc.invocationKind_ { JSGCInvocationKind::GC_NORMAL => "GC_NORMAL", JSGCInvocationKind::GC_SHRINK => "GC_SHRINK", }; println!( " isZone={}, invocation_kind={}", desc.isZone_, invocation_kind ); } let _ = stdout().flush(); } #[allow(unsafe_code)] unsafe extern "C" fn debug_gc_callback( _cx: *mut RawJSContext, status: JSGCStatus, _data: *mut os::raw::c_void, ) { match status { JSGCStatus::JSGC_BEGIN => thread_state::enter(ThreadState::IN_GC), JSGCStatus::JSGC_END => thread_state::exit(ThreadState::IN_GC), } } thread_local!( static THREAD_ACTIVE: Cell<bool> = Cell::new(true); ); #[allow(unsafe_code)] unsafe extern "C" fn trace_rust_roots(tr: *mut JSTracer, _data: *mut os::raw::c_void) { if !THREAD_ACTIVE.with(|t| t.get()) { return; } debug!("starting custom root handler"); trace_thread(tr); trace_traceables(tr); trace_roots(tr); trace_refcounted_objects(tr); settings_stack::trace(tr); debug!("done custom root handler"); } #[allow(unsafe_code)] unsafe extern "C" fn servo_build_id(build_id: *mut BuildIdCharVector) -> bool { let servo_id = b"Servo\0"; SetBuildId(build_id, &servo_id[0], servo_id.len()) } #[allow(unsafe_code)] #[cfg(feature = "debugmozjs")] unsafe fn set_gc_zeal_options(cx: *mut RawJSContext) { use js::jsapi::{JS_SetGCZeal, JS_DEFAULT_ZEAL_FREQ}; let level = match pref!(js.mem.gc.zeal.level) { level @ 0..=14 => level as u8, _ => return, }; let frequency = match pref!(js.mem.gc.zeal.frequency) { frequency if frequency >= 0 => frequency as u32, _ => JS_DEFAULT_ZEAL_FREQ, }; JS_SetGCZeal(cx, level, frequency); } #[allow(unsafe_code)] #[cfg(not(feature = "debugmozjs"))] unsafe fn set_gc_zeal_options(_: *mut RawJSContext) {} #[repr(transparent)] /// A wrapper around a JSContext that is Send, /// enabling an interrupt to be requested /// from a thread other than the one running JS using that context. pub struct ContextForRequestInterrupt(*mut RawJSContext); impl ContextForRequestInterrupt { pub fn new(context: *mut RawJSContext) -> ContextForRequestInterrupt { ContextForRequestInterrupt(context) } #[allow(unsafe_code)] /// Can be called from any thread, to request the callback set by /// JS_AddInterruptCallback to be called /// on the thread where that context is running. pub fn request_interrupt(&self) { unsafe { JS_RequestInterruptCallback(self.0); } } } #[allow(unsafe_code)] /// It is safe to call `JS_RequestInterruptCallback(cx)` from any thread. /// See the docs for the corresponding `requestInterrupt` method, /// at `mozjs/js/src/vm/JSContext.h`. unsafe impl Send for ContextForRequestInterrupt {} #[derive(Clone, Copy)] #[repr(transparent)] pub struct JSContext(*mut RawJSContext); #[allow(unsafe_code)] impl JSContext { pub unsafe fn from_ptr(raw_js_context: *mut RawJSContext) -> Self { JSContext(raw_js_context) } } #[allow(unsafe_code)] impl Deref for JSContext { type Target = *mut RawJSContext; fn deref(&self) -> &Self::Target { &self.0 } } pub struct StreamConsumer(*mut JSStreamConsumer); #[allow(unsafe_code)] impl StreamConsumer { pub fn consume_chunk(&self, stream: &[u8]) -> bool { unsafe { let stream_ptr = stream.as_ptr(); return StreamConsumerConsumeChunk(self.0, stream_ptr, stream.len()); } } pub fn stream_end(&self) { unsafe { StreamConsumerStreamEnd(self.0); } } pub fn stream_error(&self, error_code: usize) { unsafe { StreamConsumerStreamError(self.0, error_code); } } pub fn note_response_urls( &self, maybe_url: Option<String>, maybe_source_map_url: Option<String>, ) { unsafe { let maybe_url = maybe_url.map(|url| CString::new(url).unwrap()); let maybe_source_map_url = maybe_source_map_url.map(|url| CString::new(url).unwrap()); let maybe_url_param = match maybe_url.as_ref() { Some(url) => url.as_ptr(), None => ptr::null(), }; let maybe_source_map_url_param = match maybe_source_map_url.as_ref() { Some(url) => url.as_ptr(), None => ptr::null(), }; StreamConsumerNoteResponseURLs(self.0, maybe_url_param, maybe_source_map_url_param); } } } /// Implements the steps to compile webassembly response mentioned here /// <https://webassembly.github.io/spec/web-api/#compile-a-potential-webassembly-response> #[allow(unsafe_code)] unsafe extern "C" fn consume_stream( _cx: *mut RawJSContext, obj: HandleObject, _mime_type: MimeType, _consumer: *mut JSStreamConsumer, ) -> bool { let cx = JSContext::from_ptr(_cx); let in_realm_proof = AlreadyInRealm::assert_for_cx(cx); let global = GlobalScope::from_context(*cx, InRealm::Already(&in_realm_proof)); //Step 2.1 Upon fulfillment of source, store the Response with value unwrappedSource. if let Ok(unwrapped_source) = root_from_handleobject::<Response>(RustHandleObject::from_raw(obj), *cx) { //Step 2.2 Let mimeType be the result of extracting a MIME type from response’s header list. let mimetype = unwrapped_source.Headers().extract_mime_type(); //Step 2.3 If mimeType is not `application/wasm`, return with a TypeError and abort these substeps. if !&mimetype[..].eq_ignore_ascii_case(b"application/wasm") { throw_dom_exception( cx, &global, Error::Type("Response has unsupported MIME type".to_string()), ); return false; } //Step 2.4 If response is not CORS-same-origin, return with a TypeError and abort these substeps. match unwrapped_source.Type() { DOMResponseType::Basic | DOMResponseType::Cors | DOMResponseType::Default => {}, _ => { throw_dom_exception( cx, &global, Error::Type("Response.type must be 'basic', 'cors' or 'default'".to_string()), ); return false; }, } //Step 2.5 If response’s status is not an ok status, return with a TypeError and abort these substeps. if !unwrapped_source.Ok() { throw_dom_exception( cx, &global, Error::Type("Response does not have ok status".to_string()), ); return false; } // Step 2.6.1 If response body is locked, return with a TypeError and abort these substeps. if unwrapped_source.is_locked() { throw_dom_exception( cx, &global, Error::Type("There was an error consuming the Response".to_string()), ); return false; } // Step 2.6.2 If response body is alreaady consumed, return with a TypeError and abort these substeps. if unwrapped_source.is_disturbed() { throw_dom_exception( cx, &global, Error::Type("Response already consumed".to_string()), ); return false; } unwrapped_source.set_stream_consumer(Some(StreamConsumer(_consumer))); } else { //Step 3 Upon rejection of source, return with reason. throw_dom_exception( cx, &global, Error::Type("expected Response or Promise resolving to Response".to_string()), ); return false; } return true; } #[allow(unsafe_code)] unsafe extern "C" fn report_stream_error(_cx: *mut RawJSContext, error_code: usize) { error!( "Error initializing StreamConsumer: {:?}", RUST_js_GetErrorMessage(ptr::null_mut(), error_code as u32) ); } pub struct Runnable(*mut JSRunnable); #[allow(unsafe_code)] unsafe impl Sync for Runnable {} #[allow(unsafe_code)] unsafe impl Send for Runnable {} #[allow(unsafe_code)] impl Runnable { fn run(&self, cx: *mut RawJSContext, maybe_shutting_down: Dispatchable_MaybeShuttingDown) { unsafe { DispatchableRun(cx, self.0, maybe_shutting_down); } } }<|fim▁end|>
GlobalScope::from_object(incumbent_global.get()) } else {
<|file_name|>CidrAddressBlockTranslatorFactory.java<|end_file_name|><|fim▁begin|>// Copyright 2017 The Nomulus Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and<|fim▁hole|>// limitations under the License. package google.registry.model.translators; import google.registry.util.CidrAddressBlock; /** Stores {@link CidrAddressBlock} as a canonicalized string. */ public class CidrAddressBlockTranslatorFactory extends AbstractSimpleTranslatorFactory<CidrAddressBlock, String> { public CidrAddressBlockTranslatorFactory() { super(CidrAddressBlock.class); } @Override SimpleTranslator<CidrAddressBlock, String> createTranslator() { return new SimpleTranslator<CidrAddressBlock, String>(){ @Override public CidrAddressBlock loadValue(String datastoreValue) { return CidrAddressBlock.create(datastoreValue); } @Override public String saveValue(CidrAddressBlock pojoValue) { return pojoValue.toString(); }}; } }<|fim▁end|>
<|file_name|>validators.py<|end_file_name|><|fim▁begin|>""" """ # Created on 2016.08.09 # # Author: Giovanni Cannata # # Copyright 2016, 2017 Giovanni Cannata # # This file is part of ldap3. # # ldap3 is free software: you can redistribute it and/or modify # it under the terms of the GNU Lesser General Public License as published # by the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # ldap3 is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public License # along with ldap3 in the COPYING and COPYING.LESSER files. # If not, see <http://www.gnu.org/licenses/>. from datetime import datetime from ... import SEQUENCE_TYPES, STRING_TYPES from .formatters import format_time from ...utils.conv import to_raw # Validators return True if value is valid, False if value is not valid, # or a value different from True and False that is a valid value to substitute to the input value def check_type(input_value, value_type): if isinstance(input_value, value_type): return True if isinstance(input_value, SEQUENCE_TYPES): for value in input_value: if not isinstance(value, value_type): return False return True return False def always_valid(input_value): return True def validate_generic_single_value(input_value): if not isinstance(input_value, SEQUENCE_TYPES): return True try: # object couldn't have a __len__ method if len(input_value) == 1: return True except Exception: pass return False def validate_integer(input_value): if check_type(input_value, (float, bool)): return False if str is bytes: # Python 2, check for long too if check_type(input_value, (int, long)): return True else: # Python 3, int only if check_type(input_value, int): return True sequence = True # indicates if a sequence must be returned if not isinstance(input_value, SEQUENCE_TYPES): sequence = False input_value = [input_value] else: sequence = True # indicates if a sequence must be returned valid_values = [] # builds a list of valid int values for element in input_value: try: # try to convert any type to int, an invalid conversion raise TypeError of ValueError, if both are valid and equal then then int() value is used float_value = float(element) int_value = int(element) if float_value == int_value: valid_values.append(int(element)) else: return False except (ValueError, TypeError): return False if sequence: return valid_values else: return valid_values[0] def validate_bytes(input_value): return check_type(input_value, bytes) def validate_boolean(input_value): # it could be a real bool or the string TRUE or FALSE, # only a single valued is allowed if validate_generic_single_value(input_value): # valid only if a single value or a sequence with a single element if isinstance(input_value, SEQUENCE_TYPES): input_value = input_value[0] if isinstance(input_value, bool): if input_value: return 'TRUE' else: return 'FALSE' if isinstance(input_value, STRING_TYPES):<|fim▁hole|> if input_value.lower() == 'true': return 'TRUE' elif input_value.lower() == 'false': return 'FALSE' return False def validate_time(input_value): # if datetime object doesn't have a timezone it's considered local time and is adjusted to UTC if not isinstance(input_value, SEQUENCE_TYPES): sequence = False input_value = [input_value] else: sequence = True # indicates if a sequence must be returned valid_values = [] changed = False for element in input_value: if isinstance(element, STRING_TYPES): # tries to check if it is already be a Generalized Time if isinstance(format_time(to_raw(element)), datetime): # valid Generalized Time string valid_values.append(element) else: return False elif isinstance(element, datetime): changed = True if element.tzinfo: # a datetime with a timezone valid_values.append(element.strftime('%Y%m%d%H%M%S%z')) else: # datetime without timezone, assumed local and adjusted to UTC offset = datetime.now() - datetime.utcnow() valid_values.append((element - offset).strftime('%Y%m%d%H%M%SZ')) else: return False if changed: if sequence: return valid_values else: return valid_values[0] else: return True<|fim▁end|>
<|file_name|>campaign.cpp<|end_file_name|><|fim▁begin|>/****************************************************************************** * irrlamb - https://github.com/jazztickets/irrlamb * Copyright (C) 2019 Alan Witkowski * * This program is free software: you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program. If not, see <http://www.gnu.org/licenses/>. *******************************************************************************/ #include <campaign.h> #include <globals.h> #include <log.h> #include <framework.h> #include <level.h> #include <save.h> #include <tinyxml2/tinyxml2.h> _Campaign Campaign; using namespace tinyxml2; // Loads the campaign data int _Campaign::Init() { Campaigns.clear(); Log.Write("Loading campaign file main.xml"); // Open the XML file std::string LevelFile = std::string("levels/main.xml"); XMLDocument Document; if(Document.LoadFile(LevelFile.c_str()) != XML_SUCCESS) { Log.Write("Error loading level file with error id = %d", Document.ErrorID()); Log.Write("Error string: %s", Document.ErrorStr()); Close(); return 0; } // Check for level tag XMLElement *CampaignsElement = Document.FirstChildElement("campaigns"); if(!CampaignsElement) { Log.Write("Could not find campaigns tag"); return 0; } // Load campaigns XMLElement *CampaignElement = CampaignsElement->FirstChildElement("campaign"); for(; CampaignElement != 0; CampaignElement = CampaignElement->NextSiblingElement("campaign")) { _CampaignInfo Campaign;<|fim▁hole|> Campaign.Show = true; Campaign.Column = 1; Campaign.Row = 0; Campaign.Name = CampaignElement->Attribute("name"); CampaignElement->QueryBoolAttribute("show", &Campaign.Show); CampaignElement->QueryIntAttribute("column", &Campaign.Column); CampaignElement->QueryIntAttribute("row", &Campaign.Row); // Get levels XMLElement *LevelElement = CampaignElement->FirstChildElement("level"); for(; LevelElement != 0; LevelElement = LevelElement->NextSiblingElement("level")) { _LevelInfo Level; Level.File = LevelElement->GetText(); Level.DataPath = Framework.GetWorkingPath() + "levels/" + Level.File + "/"; Level.Unlocked = 0; LevelElement->QueryIntAttribute("unlocked", &Level.Unlocked); ::Level.Init(Level.File, true); Level.NiceName = ::Level.LevelNiceName; Campaign.Levels.push_back(Level); } Campaigns.push_back(Campaign); } return 1; } // Closes the campaign system int _Campaign::Close() { Campaigns.clear(); return 1; } // Get number of completed levels for a campaign int _Campaign::GetCompletedLevels(int CampaignIndex) { int Completed = 0; for(auto &Level : Campaigns[CampaignIndex].Levels) { if(Save.LevelStats[Level.File].WinCount) Completed++; } return Completed; } // Get next level or 1st level in next campaign, return false if no next level bool _Campaign::GetNextLevel(uint32_t &Campaign, uint32_t &Level, bool Update) { uint32_t NewCampaign = Campaign; uint32_t NewLevel = Level; if(NewCampaign >= Campaigns.size()) return false; if(NewLevel+1 >= Campaigns[NewCampaign].Levels.size()) { return false; } else { NewLevel++; } // Return values if(Update) { Campaign = NewCampaign; Level = NewLevel; } return true; }<|fim▁end|>
<|file_name|>ex2.py<|end_file_name|><|fim▁begin|># A comment, this is so you can read your program later. # Anything after the # is ignored by python. print "I could have code like this." # and the comment after is ignored # You can also use a comment to "disable" or comment out a piece of code: # print "This won't run." print "This will run." # Adding another few lines just for fun. print 'Q: Why does the "#" in "print "Hi # there." not get ignored?'<|fim▁hole|># print 'The # in that code is inside a string, so it will put into the string until the ending " character is hit. These pound characters are just considered characters and are not considered comments.' # Another way to put it: (aren't instead of "are not") print "The # in that code is inside a string, so it will put into the string until the ending \" character is hit. These pound characters are just considered characters and aren't considered comments." # The backslash will escape the special character, as seen on the code above. Isaac Albeniz - Asturias :-)<|fim▁end|>
<|file_name|>render_target.rs<|end_file_name|><|fim▁begin|>//! Defines the render target for tray, where our image will be written too //! during rendering use std::vec::Vec; use std::{iter, cmp, f32}; use std::sync::Mutex; use film::Colorf; use film::filter::Filter; use sampler::Region; const FILTER_TABLE_SIZE: usize = 16; /// A struct containing results of an image sample where a ray was fired through /// continuous pixel coordinates [x, y] and color `color` was computed pub struct ImageSample { pub x: f32, pub y: f32, pub color: Colorf, } impl ImageSample { pub fn new(x: f32, y: f32, color: Colorf) -> ImageSample { ImageSample { x: x, y: y, color: color } } } /// `RenderTarget` is a RGBF render target to write our image too while rendering pub struct RenderTarget { width: usize, height: usize, pixels_locked: Vec<Mutex<Vec<Colorf>>>, lock_size: (i32, i32), filter: Box<Filter + Send + Sync>, filter_table: Vec<f32>, filter_pixel_width: (i32, i32), } impl RenderTarget { /// Create a render target with `width * height` pixels pub fn new(image_dim: (usize, usize), lock_size: (usize, usize), filter: Box<Filter + Send + Sync>) -> RenderTarget { if image_dim.0 % lock_size.0 != 0 || image_dim.1 % lock_size.1 != 0 { panic!("Image with dimension {:?} not evenly divided by blocks of {:?}", image_dim, lock_size); } let width = image_dim.0; let height = image_dim.1; let filter_pixel_width = (f32::floor(filter.width() / 0.5) as i32, f32::floor(filter.height() / 0.5) as i32); let mut filter_table: Vec<f32> = iter::repeat(0.0).take(FILTER_TABLE_SIZE * FILTER_TABLE_SIZE) .collect(); for y in 0..FILTER_TABLE_SIZE { let fy = (y as f32 + 0.5) * filter.height() / FILTER_TABLE_SIZE as f32; for x in 0..FILTER_TABLE_SIZE { let fx = (x as f32 + 0.5) * filter.width() / FILTER_TABLE_SIZE as f32; filter_table[y * FILTER_TABLE_SIZE + x] = filter.weight(fx, fy); } } let x_blocks = width / lock_size.0; let y_blocks = height / lock_size.1; let mut pixels_locked = Vec::with_capacity(x_blocks * y_blocks); for _ in 0..x_blocks * y_blocks { pixels_locked.push(Mutex::new(iter::repeat(Colorf::broadcast(0.0)) .take(lock_size.0 * lock_size.1).collect())); } RenderTarget { width: width, height: height, pixels_locked: pixels_locked, lock_size: (lock_size.0 as i32, lock_size.1 as i32), filter: filter, filter_table: filter_table, filter_pixel_width: filter_pixel_width, } } /// Write all the image samples to the render target pub fn write(&self, samples: &[ImageSample], region: &Region) { // Determine which blocks we touch with our set of samples let x_range = (cmp::max(region.start.0 as i32 - self.filter_pixel_width.0, 0), cmp::min(region.end.0 as i32 + self.filter_pixel_width.0, self.width as i32 - 1)); let y_range = (cmp::max(region.start.1 as i32 - self.filter_pixel_width.1, 0), cmp::min(region.end.1 as i32 + self.filter_pixel_width.1, self.height as i32 - 1)); if x_range.1 - x_range.0 < 0 || y_range.1 - y_range.0 < 0 { return; } let block_x_range = (x_range.0 / self.lock_size.0, x_range.1 / self.lock_size.0); let block_y_range = (y_range.0 / self.lock_size.1, y_range.1 / self.lock_size.1); // Temporary storage for filtered samples so we can compute the filtered results for // the block we're writing too without having to get the lock let mut filtered_samples: Vec<_> = iter::repeat(Colorf::broadcast(0.0)) .take((self.lock_size.0 * self.lock_size.1) as usize).collect(); let blocks_per_row = self.width as i32 / self.lock_size.0; for y in block_y_range.0..block_y_range.1 + 1 { for x in block_x_range.0..block_x_range.1 + 1 { let block_x_start = x * self.lock_size.0; let block_y_start = y * self.lock_size.1; let x_write_range = (cmp::max(x_range.0, block_x_start), cmp::min(x_range.1 + 1, block_x_start + self.lock_size.0)); let y_write_range = (cmp::max(y_range.0, block_y_start), cmp::min(y_range.1 + 1, block_y_start + self.lock_size.1)); let block_samples = samples.iter().filter(|s| { s.x >= (x_write_range.0 - self.filter_pixel_width.0) as f32 && s.x < (x_write_range.1 + self.filter_pixel_width.0) as f32 && s.y >= (y_write_range.0 - self.filter_pixel_width.1) as f32 && s.y < (y_write_range.1 + self.filter_pixel_width.1) as f32 }); for c in &mut filtered_samples { *c = Colorf::broadcast(0.0); } // Compute the filtered samples for the block for c in block_samples { let img_x = c.x - 0.5; let img_y = c.y - 0.5; for iy in y_write_range.0..y_write_range.1 { let fy = f32::abs(iy as f32 - img_y) * self.filter.inv_height(); // While we know this sample effects some pixels in this block it may not // necessarily effect this specific pixel, so double check that it's in // the filter's dimensions. if fy > self.filter.height() { continue; } let fy_idx = cmp::min((fy * FILTER_TABLE_SIZE as f32) as usize, FILTER_TABLE_SIZE - 1); for ix in x_write_range.0..x_write_range.1 { let fx = f32::abs(ix as f32 - img_x) * self.filter.inv_width(); // Check that we're also in the width of the filter if fx > self.filter.width() { continue; } let fx_idx = cmp::min((fx * FILTER_TABLE_SIZE as f32) as usize, FILTER_TABLE_SIZE - 1); let weight = self.filter_table[fy_idx * FILTER_TABLE_SIZE + fx_idx]; let px = ((iy - block_y_start) * self.lock_size.0 + ix - block_x_start) as usize; // TODO: Can't currently overload the += operator // Coming soon though, see RFC #953 https://github.com/rust-lang/rfcs/pull/953 filtered_samples[px].r += weight * c.color.r; filtered_samples[px].g += weight * c.color.g; filtered_samples[px].b += weight * c.color.b; filtered_samples[px].a += weight; } } } // Acquire lock for the block and write the filtered samples let block_idx = (y * blocks_per_row + x) as usize; let mut pixels = self.pixels_locked[block_idx].lock().unwrap(); for iy in y_write_range.0..y_write_range.1 { for ix in x_write_range.0..x_write_range.1 { let px = ((iy - block_y_start) * self.lock_size.0 + ix - block_x_start) as usize; let c = &filtered_samples[px]; pixels[px].r += c.r; pixels[px].g += c.g; pixels[px].b += c.b; pixels[px].a += c.a; } } } } } /// Clear the render target to black pub fn clear(&mut self) { let x_blocks = self.width / self.lock_size.0 as usize; let y_blocks = self.height / self.lock_size.1 as usize; for by in 0..y_blocks { for bx in 0..x_blocks { let block_idx = (by * x_blocks + bx) as usize; let mut pixels = self.pixels_locked[block_idx].lock().unwrap(); for p in pixels.iter_mut() { *p = Colorf::broadcast(0.0); } } } } /// Get the dimensions of the render target pub fn dimensions(&self) -> (usize, usize) { (self.width, self.height) } /// Convert the floating point color buffer to 24bpp sRGB for output to an image pub fn get_render(&self) -> Vec<u8> { let mut render: Vec<u8> = iter::repeat(0u8).take(self.width * self.height * 3).collect(); let x_blocks = self.width / self.lock_size.0 as usize; let y_blocks = self.height / self.lock_size.1 as usize; for by in 0..y_blocks { for bx in 0..x_blocks { let block_x_start = bx * self.lock_size.0 as usize; let block_y_start = by * self.lock_size.1 as usize; let block_idx = (by * x_blocks + bx) as usize; let pixels = self.pixels_locked[block_idx].lock().unwrap(); for y in 0..self.lock_size.1 as usize { for x in 0..self.lock_size.0 as usize { let c = &pixels[y * self.lock_size.0 as usize + x]; if c.a > 0.0 { let cn = (*c / c.a).clamp().to_srgb(); let px = (y + block_y_start) * self.width * 3 + (x + block_x_start) * 3; for i in 0..3 { render[px + i] = (cn[i] * 255.0) as u8; } } } } } } render } /// Get the blocks that have had pixels written too them. Returns the size of each block,<|fim▁hole|> /// The block's pixels are stored in the same order their position appears in the block /// positions vec and contain `dim.0 * dim.1 * 4` f32's per block. pub fn get_rendered_blocks(&self) -> ((usize, usize), Vec<(usize, usize)>, Vec<f32>) { let block_size = (self.lock_size.0 as usize, self.lock_size.1 as usize); let mut blocks = Vec::new(); let mut render = Vec::new(); let x_blocks = self.width / block_size.0; let y_blocks = self.height / block_size.1; for by in 0..y_blocks { for bx in 0..x_blocks { let block_x_start = bx * block_size.0; let block_y_start = by * block_size.1; let block_idx = by * x_blocks + bx; let pixels = self.pixels_locked[block_idx].lock().unwrap(); if pixels.iter().fold(true, |acc, px| acc && px.a != 0.0) { blocks.push((block_x_start, block_y_start)); for y in 0..block_size.1 { for x in 0..block_size.0 { let c = &pixels[y * block_size.0 + x]; for i in 0..4 { render.push(c[i]); } } } } } } (block_size, blocks, render) } /// Get the raw floating point framebuffer pub fn get_renderf32(&self) -> Vec<f32> { let mut render: Vec<f32> = iter::repeat(0.0).take(self.width * self.height * 4).collect(); let x_blocks = self.width / self.lock_size.0 as usize; let y_blocks = self.height / self.lock_size.1 as usize; for by in 0..y_blocks { for bx in 0..x_blocks { let block_x_start = bx * self.lock_size.0 as usize; let block_y_start = by * self.lock_size.1 as usize; let block_idx = (by * x_blocks + bx) as usize; let pixels = self.pixels_locked[block_idx].lock().unwrap(); for y in 0..self.lock_size.1 as usize { for x in 0..self.lock_size.0 as usize { let c = &pixels[y * self.lock_size.0 as usize + x]; let px = (y + block_y_start) * self.width * 4 + (x + block_x_start) * 4; for i in 0..4 { render[px + i] = c[i]; } } } } } render } }<|fim▁end|>
/// a list of block positions in pixels and then pixels for the blocks (in a single f32 vec).