supercat666's picture
add igv
78c921d
raw
history blame
13.8 kB
// Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing,
// software distributed under the License is distributed on an
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied. See the License for the
// specific language governing permissions and limitations
// under the License.
import {
toUint8Array,
joinUint8Arrays,
ArrayBufferViewInput,
toUint8ArrayIterator,
toUint8ArrayAsyncIterator
} from '../util/buffer.js';
import { ReadableDOMStreamOptions } from './interfaces.js';
type Uint8ArrayGenerator = Generator<Uint8Array, null, { cmd: 'peek' | 'read'; size: number }>;
type AsyncUint8ArrayGenerator = AsyncGenerator<Uint8Array, null, { cmd: 'peek' | 'read'; size: number }>;
/** @ignore */
export default {
fromIterable<T extends ArrayBufferViewInput>(source: Iterable<T> | T): Uint8ArrayGenerator {
return pump(fromIterable<T>(source));
},
fromAsyncIterable<T extends ArrayBufferViewInput>(source: AsyncIterable<T> | PromiseLike<T>): AsyncUint8ArrayGenerator {
return pump(fromAsyncIterable<T>(source));
},
fromDOMStream<T extends ArrayBufferViewInput>(source: ReadableStream<T>): AsyncUint8ArrayGenerator {
return pump(fromDOMStream<T>(source));
},
fromNodeStream(stream: NodeJS.ReadableStream): AsyncUint8ArrayGenerator {
return pump(fromNodeStream(stream));
},
// @ts-ignore
toDOMStream<T>(source: Iterable<T> | AsyncIterable<T>, options?: ReadableDOMStreamOptions): ReadableStream<T> {
throw new Error(`"toDOMStream" not available in this environment`);
},
// @ts-ignore
toNodeStream<T>(source: Iterable<T> | AsyncIterable<T>, options?: import('stream').ReadableOptions): import('stream').Readable {
throw new Error(`"toNodeStream" not available in this environment`);
},
};
/** @ignore */
const pump = <T extends Uint8ArrayGenerator | AsyncUint8ArrayGenerator>(iterator: T) => { iterator.next(); return iterator; };
/** @ignore */
function* fromIterable<T extends ArrayBufferViewInput>(source: Iterable<T> | T): Uint8ArrayGenerator {
let done: boolean | undefined, threw = false;
let buffers: Uint8Array[] = [], buffer: Uint8Array;
let cmd: 'peek' | 'read', size: number, bufferLength = 0;
function byteRange() {
if (cmd === 'peek') {
return joinUint8Arrays(buffers, size)[0];
}
[buffer, buffers, bufferLength] = joinUint8Arrays(buffers, size);
return buffer;
}
// Yield so the caller can inject the read command before creating the source Iterator
({ cmd, size } = yield <any>null);
// initialize the iterator
const it = toUint8ArrayIterator(source)[Symbol.iterator]();
try {
do {
// read the next value
({ done, value: buffer } = Number.isNaN(size - bufferLength) ?
it.next() : it.next(size - bufferLength));
// if chunk is not null or empty, push it onto the queue
if (!done && buffer.byteLength > 0) {
buffers.push(buffer);
bufferLength += buffer.byteLength;
}
// If we have enough bytes in our buffer, yield chunks until we don't
if (done || size <= bufferLength) {
do {
({ cmd, size } = yield byteRange());
} while (size < bufferLength);
}
} while (!done);
} catch (e) {
(threw = true) && (typeof it.throw === 'function') && (it.throw(e));
} finally {
(threw === false) && (typeof it.return === 'function') && (it.return(null!));
}
return null;
}
/** @ignore */
async function* fromAsyncIterable<T extends ArrayBufferViewInput>(source: AsyncIterable<T> | PromiseLike<T>): AsyncUint8ArrayGenerator {
let done: boolean | undefined, threw = false;
let buffers: Uint8Array[] = [], buffer: Uint8Array;
let cmd: 'peek' | 'read', size: number, bufferLength = 0;
function byteRange() {
if (cmd === 'peek') {
return joinUint8Arrays(buffers, size)[0];
}
[buffer, buffers, bufferLength] = joinUint8Arrays(buffers, size);
return buffer;
}
// Yield so the caller can inject the read command before creating the source AsyncIterator
({ cmd, size } = (yield <any>null)!);
// initialize the iterator
const it = toUint8ArrayAsyncIterator(source)[Symbol.asyncIterator]();
try {
do {
// read the next value
({ done, value: buffer } = Number.isNaN(size - bufferLength)
? await it.next()
: await it.next(size - bufferLength));
// if chunk is not null or empty, push it onto the queue
if (!done && buffer.byteLength > 0) {
buffers.push(buffer);
bufferLength += buffer.byteLength;
}
// If we have enough bytes in our buffer, yield chunks until we don't
if (done || size <= bufferLength) {
do {
({ cmd, size } = yield byteRange());
} while (size < bufferLength);
}
} while (!done);
} catch (e) {
(threw = true) && (typeof it.throw === 'function') && (await it.throw(e));
} finally {
(threw === false) && (typeof it.return === 'function') && (await it.return(new Uint8Array(0)));
}
return null;
}
// All this manual Uint8Array chunk management can be avoided if/when engines
// add support for ArrayBuffer.transfer() or ArrayBuffer.prototype.realloc():
// https://github.com/domenic/proposal-arraybuffer-transfer
/** @ignore */
async function* fromDOMStream<T extends ArrayBufferViewInput>(source: ReadableStream<T>): AsyncUint8ArrayGenerator {
let done = false, threw = false;
let buffers: Uint8Array[] = [], buffer: Uint8Array;
let cmd: 'peek' | 'read', size: number, bufferLength = 0;
function byteRange() {
if (cmd === 'peek') {
return joinUint8Arrays(buffers, size)[0];
}
[buffer, buffers, bufferLength] = joinUint8Arrays(buffers, size);
return buffer;
}
// Yield so the caller can inject the read command before we establish the ReadableStream lock
({ cmd, size } = yield <any>null);
// initialize the reader and lock the stream
const it = new AdaptiveByteReader(source);
try {
do {
// read the next value
({ done, value: buffer } = Number.isNaN(size - bufferLength)
? await it['read']()
: await it['read'](size - bufferLength));
// if chunk is not null or empty, push it onto the queue
if (!done && buffer.byteLength > 0) {
buffers.push(toUint8Array(buffer));
bufferLength += buffer.byteLength;
}
// If we have enough bytes in our buffer, yield chunks until we don't
if (done || size <= bufferLength) {
do {
({ cmd, size } = yield byteRange());
} while (size < bufferLength);
}
} while (!done);
} catch (e) {
(threw = true) && (await it['cancel'](e));
} finally {
(threw === false) ? (await it['cancel']())
: source['locked'] && it.releaseLock();
}
return null;
}
/** @ignore */
class AdaptiveByteReader<T extends ArrayBufferViewInput> {
private reader: ReadableStreamDefaultReader<T> | null = null;
constructor(private source: ReadableStream<T>) {
this.reader = this.source['getReader']();
// We have to catch and swallow errors here to avoid uncaught promise rejection exceptions
// that seem to be raised when we call `releaseLock()` on this reader. I'm still mystified
// about why these errors are raised, but I'm sure there's some important spec reason that
// I haven't considered. I hate to employ such an anti-pattern here, but it seems like the
// only solution in this case :/
this.reader['closed'].catch(() => { });
}
get closed(): Promise<void> {
return this.reader ? this.reader['closed'].catch(() => { }) : Promise.resolve();
}
releaseLock(): void {
if (this.reader) {
this.reader.releaseLock();
}
this.reader = null;
}
async cancel(reason?: any): Promise<void> {
const { reader, source } = this;
reader && (await reader['cancel'](reason).catch(() => { }));
source && (source['locked'] && this.releaseLock());
}
async read(size?: number): Promise<ReadableStreamReadValueResult<Uint8Array>> {
if (size === 0) {
return { done: this.reader == null, value: new Uint8Array(0) } as ReadableStreamReadValueResult<Uint8Array>;
}
const result = await this.reader!.read() as ReadableStreamReadValueResult<any>;
!result.done && (result.value = toUint8Array(result));
return result;
}
}
/** @ignore */
type EventName = 'end' | 'error' | 'readable';
/** @ignore */
type Event = [EventName, (_: any) => void, Promise<[EventName, Error | null]>];
/** @ignore */
const onEvent = <T extends string>(stream: NodeJS.ReadableStream, event: T) => {
const handler = (_: any) => resolve([event, _]);
let resolve: (value: [T, any] | PromiseLike<[T, any]>) => void;
return [event, handler, new Promise<[T, any]>(
(r) => (resolve = r) && stream['once'](event, handler)
)] as Event;
};
/** @ignore */
async function* fromNodeStream(stream: NodeJS.ReadableStream): AsyncUint8ArrayGenerator {
const events: Event[] = [];
let event: EventName = 'error';
let done = false, err: Error | null = null;
let cmd: 'peek' | 'read', size: number, bufferLength = 0;
let buffers: Uint8Array[] = [], buffer: Uint8Array | Buffer | string;
function byteRange() {
if (cmd === 'peek') {
return joinUint8Arrays(buffers, size)[0];
}
[buffer, buffers, bufferLength] = joinUint8Arrays(buffers, size);
return buffer;
}
// Yield so the caller can inject the read command before we
// add the listener for the source stream's 'readable' event.
({ cmd, size } = yield <any>null);
// ignore stdin if it's a TTY
if ((stream as any)['isTTY']) {
yield new Uint8Array(0);
return null;
}
try {
// initialize the stream event handlers
events[0] = onEvent(stream, 'end');
events[1] = onEvent(stream, 'error');
do {
events[2] = onEvent(stream, 'readable');
// wait on the first message event from the stream
[event, err] = await Promise.race(events.map((x) => x[2]));
// if the stream emitted an Error, rethrow it
if (event === 'error') { break; }
if (!(done = event === 'end')) {
// If the size is NaN, request to read everything in the stream's internal buffer
if (!Number.isFinite(size - bufferLength)) {
buffer = toUint8Array(stream['read']());
} else {
buffer = toUint8Array(stream['read'](size - bufferLength));
// If the byteLength is 0, then the requested amount is more than the stream has
// in its internal buffer. In this case the stream needs a "kick" to tell it to
// continue emitting readable events, so request to read everything the stream
// has in its internal buffer right now.
if ((buffer as Uint8Array).byteLength < (size - bufferLength)) {
buffer = toUint8Array(stream['read']());
}
}
// if chunk is not null or empty, push it onto the queue
if ((buffer as Uint8Array).byteLength > 0) {
buffers.push(buffer as Uint8Array);
bufferLength += (buffer as Uint8Array).byteLength;
}
}
// If we have enough bytes in our buffer, yield chunks until we don't
if (done || size <= bufferLength) {
do {
({ cmd, size } = yield byteRange());
} while (size < bufferLength);
}
} while (!done);
} finally {
await cleanup(events, event === 'error' ? err : null);
}
return null;
function cleanup<T extends Error | null | void>(events: Event[], err?: T) {
buffer = buffers = <any>null;
return new Promise<void>((resolve, reject) => {
for (const [evt, fn] of events) {
stream['off'](evt, fn);
}
try {
// Some stream implementations don't call the destroy callback,
// because it's really a node-internal API. Just calling `destroy`
// here should be enough to conform to the ReadableStream contract
const destroy = (stream as any)['destroy'];
destroy && destroy.call(stream, err);
err = undefined;
} catch (e) { err = e as T || err; } finally {
err != null ? reject(err) : resolve();
}
});
}
}