Compare commits

..

No commits in common. "main" and "dev" have entirely different histories.
main ... dev

9 changed files with 68 additions and 218 deletions

View File

@ -26,15 +26,4 @@
- [ ] FLAC (need tested) - [ ] FLAC (need tested)
- [ ] Wrap video element with customElements (Prototyping / Lit-html + Typescript) - [ ] Wrap video element with customElements (Prototyping / Lit-html + Typescript)
- [ ] Add WebCodecs polyfill with ffmpeg or libav (Todo / WASM) - [ ] Add WebCodecs polyfill with ffmpeg or libav (Todo / WASM)
- [x] Chrome/Edge/Android Webview: WebCodecs Native support - [ ] Danmuku integrated (Todo / Typescript)
- [ ] FIREFOX
- [x] VP8/VP9/AV1 native support
- [x] AVC/HEVC 8bit native support
- [ ] AVC/HEVC >= 10bit polyfill needed
- [ ] Firefox Android not support
- [ ] Safari
- [x] VP8/VP9/AV1 native support
- [x] AVC/HEVC 8bit native support
- [ ] AVC/HEVC >= 10bit polyfill needed for some devices
- [ ] Audio Decoder polyfill needed
- [ ] Danmuku integration (Todo / Typescript)

View File

@ -3,8 +3,7 @@
<head></head> <head></head>
<body> <body>
<!-- <my-element />--> <!-- <my-element />-->
<!-- <video-pipeline-demo src="/api/static/video/test-hevc.mkv" width="800" height="450"></video-pipeline-demo> --> <!-- <video-pipeline-demo src="/api/static/video/test.webm"></video-pipeline-demo>-->
<video-pipeline-demo src="/api/static/video/huge/test8.mkv" width="800" height="450"></video-pipeline-demo> <video-pipeline-demo src="/api/static/video/huge/[LoliHouse] Amagami-san Chi no Enmusubi - 23 [WebRip 1080p HEVC-10bit AAC SRTx2].mkv" width="800" height="450" />
<!-- <video-pipeline-demo src="/api/static/video/huge/[LoliHouse] Amagami-san Chi no Enmusubi - 23 [WebRip 1080p HEVC-10bit AAC SRTx2].mkv" width="800" height="450" /> -->
</body> </body>

View File

@ -13,10 +13,7 @@ import {
fromEvent, fromEvent,
share, share,
takeUntil, takeUntil,
firstValueFrom, firstValueFrom, tap, throwIfEmpty, ReplaySubject, finalize, of, interval,
tap,
throwIfEmpty,
ReplaySubject,
} from 'rxjs'; } from 'rxjs';
import { createMatroska } from '@konoplayer/matroska/model'; import { createMatroska } from '@konoplayer/matroska/model';
import { createRef, ref, type Ref } from 'lit/directives/ref.js'; import { createRef, ref, type Ref } from 'lit/directives/ref.js';
@ -81,47 +78,35 @@ export class VideoPipelineDemo extends LitElement {
videoTrackDecoder, videoTrackDecoder,
audioTrackDecoder, audioTrackDecoder,
}, },
totalSize, totalSize
} = await firstValueFrom( } = await firstValueFrom(
createMatroska({ createMatroska({
url: src, url: src,
}).pipe(throwIfEmpty(() => new Error('failed to extract matroska'))) }).pipe(
); throwIfEmpty(() => new Error("failed to extract matroska"))
)
)
console.debug(`[MATROSKA]: loaded metadata, total size ${totalSize} bytes`); console.debug(`[MATROSKA]: loaded metadata, total size ${totalSize} bytes`)
const currentCluster$ = this.seeked$.pipe( const currentCluster$ = this.seeked$.pipe(
switchMap((seekTime) => seek(seekTime)), switchMap((seekTime) => seek(seekTime)),
share({ share({ resetOnRefCountZero: false, resetOnError: false, resetOnComplete: false }),
resetOnRefCountZero: false,
resetOnError: false,
resetOnComplete: false,
})
); );
defaultVideoTrack$ defaultVideoTrack$
.pipe( .pipe(take(1), takeUntil(destroyRef$), tap((track) => console.debug('[MATROSKA]: video track loaded,', track)))
take(1),
takeUntil(destroyRef$),
tap((track) => console.debug('[MATROSKA]: video track loaded,', track))
)
.subscribe(this.videoTrack$.next.bind(this.videoTrack$)); .subscribe(this.videoTrack$.next.bind(this.videoTrack$));
defaultAudioTrack$ defaultAudioTrack$
.pipe( .pipe(take(1), takeUntil(destroyRef$), tap((track) => console.debug('[MATROSKA]: audio track loaded,', track)))
take(1),
takeUntil(destroyRef$),
tap((track) => console.debug('[MATROSKA]: audio track loaded,', track))
)
.subscribe(this.audioTrack$.next.bind(this.audioTrack$)); .subscribe(this.audioTrack$.next.bind(this.audioTrack$));
this.videoTrack$ this.videoTrack$
.pipe( .pipe(
takeUntil(this.destroyRef$), takeUntil(this.destroyRef$),
switchMap((track) => switchMap((track) =>
track?.configuration track?.configuration ? videoTrackDecoder(track, currentCluster$) : EMPTY
? videoTrackDecoder(track, currentCluster$)
: EMPTY
), ),
switchMap(({ frame$ }) => frame$) switchMap(({ frame$ }) => frame$)
) )
@ -135,9 +120,7 @@ export class VideoPipelineDemo extends LitElement {
.pipe( .pipe(
takeUntil(this.destroyRef$), takeUntil(this.destroyRef$),
switchMap((track) => switchMap((track) =>
track?.configuration track?.configuration ? audioTrackDecoder(track, currentCluster$) : EMPTY
? audioTrackDecoder(track, currentCluster$)
: EMPTY
), ),
switchMap(({ frame$ }) => frame$) switchMap(({ frame$ }) => frame$)
) )
@ -161,25 +144,22 @@ export class VideoPipelineDemo extends LitElement {
), ),
}).pipe( }).pipe(
takeUntil(this.destroyRef$), takeUntil(this.destroyRef$),
map( map(({ ended, paused, videoBuffered, audioBuffered }) => !paused && !ended && !!(videoBuffered || audioBuffered)),
({ ended, paused, videoBuffered, audioBuffered }) =>
!paused && !ended && !!(videoBuffered || audioBuffered)
),
tap((enabled) => { tap((enabled) => {
if (enabled) { if (enabled) {
playableStartTime = performance.now(); playableStartTime = performance.now()
} }
}), }),
share() share()
); )
let nextAudioStartTime = 0; let nextAudioStartTime = 0;
playable playable
.pipe( .pipe(
tap(() => { tap(() => {
nextAudioStartTime = 0; nextAudioStartTime = 0
}), }),
switchMap((enabled) => (enabled ? animationFrames() : EMPTY)) switchMap((enabled) => (enabled ? animationFrames() : EMPTY)),
) )
.subscribe(() => { .subscribe(() => {
const audioFrameBuffer = this.audioFrameBuffer$.getValue(); const audioFrameBuffer = this.audioFrameBuffer$.getValue();
@ -189,7 +169,7 @@ export class VideoPipelineDemo extends LitElement {
let audioChanged = false; let audioChanged = false;
while (audioFrameBuffer.size > 0) { while (audioFrameBuffer.size > 0) {
const firstAudio = audioFrameBuffer.peek(); const firstAudio = audioFrameBuffer.peek();
if (firstAudio && firstAudio.timestamp / 1000 <= accTime) { if (firstAudio && (firstAudio.timestamp / 1000) <= accTime) {
const audioFrame = audioFrameBuffer.dequeue()!; const audioFrame = audioFrameBuffer.dequeue()!;
audioChanged = true; audioChanged = true;
if (audioContext) { if (audioContext) {
@ -207,15 +187,10 @@ export class VideoPipelineDemo extends LitElement {
const fadeLength = Math.min(50, audioFrame.numberOfFrames); const fadeLength = Math.min(50, audioFrame.numberOfFrames);
for (let channel = 0; channel < numberOfChannels; channel++) { for (let channel = 0; channel < numberOfChannels; channel++) {
const channelData = new Float32Array(numberOfFrames); const channelData = new Float32Array(numberOfFrames);
audioFrame.copyTo(channelData, { audioFrame.copyTo(channelData, { planeIndex: channel, frameCount: numberOfFrames });
format: 'f32-planar',
planeIndex: channel,
frameCount: numberOfFrames,
});
for (let i = 0; i < fadeLength; i++) { for (let i = 0; i < fadeLength; i++) {
channelData[i] *= i / fadeLength; // fade-in channelData[i] *= i / fadeLength; // fade-in
channelData[audioFrame.numberOfFrames - 1 - i] *= channelData[audioFrame.numberOfFrames - 1 - i] *= i / fadeLength; // fade-out
i / fadeLength; // fade-out
} }
audioBuffer.copyToChannel(channelData, channel); audioBuffer.copyToChannel(channelData, channel);
} }
@ -247,7 +222,9 @@ export class VideoPipelineDemo extends LitElement {
}); });
playable playable
.pipe(switchMap((enabled) => (enabled ? animationFrames() : EMPTY))) .pipe(
switchMap((enabled) => (enabled ? animationFrames() : EMPTY)),
)
.subscribe(async () => { .subscribe(async () => {
const renderingContext = this.renderingContext; const renderingContext = this.renderingContext;
const videoFrameBuffer = this.videoFrameBuffer$.getValue(); const videoFrameBuffer = this.videoFrameBuffer$.getValue();
@ -256,7 +233,7 @@ export class VideoPipelineDemo extends LitElement {
const accTime = nowTime - playableStartTime; const accTime = nowTime - playableStartTime;
while (videoFrameBuffer.size > 0) { while (videoFrameBuffer.size > 0) {
const firstVideo = videoFrameBuffer.peek(); const firstVideo = videoFrameBuffer.peek();
if (firstVideo && firstVideo.timestamp / 1000 <= accTime) { if (firstVideo && (firstVideo.timestamp / 1000) <= accTime) {
const videoFrame = videoFrameBuffer.dequeue()!; const videoFrame = videoFrameBuffer.dequeue()!;
videoChanged = true; videoChanged = true;
if (renderingContext) { if (renderingContext) {
@ -275,30 +252,12 @@ export class VideoPipelineDemo extends LitElement {
fromEvent(document.body, 'click') fromEvent(document.body, 'click')
.pipe(takeUntil(this.destroyRef$)) .pipe(takeUntil(this.destroyRef$))
.subscribe(async () => { .subscribe(() => {
const permissionStatus = await navigator.permissions.query({
name: 'microphone',
});
if (permissionStatus.state === 'prompt') {
await navigator.mediaDevices.getUserMedia({
audio: true,
});
}
this.audioContext.resume(); this.audioContext.resume();
this.audioFrameBuffer$.next(this.audioFrameBuffer$.getValue()); this.audioFrameBuffer$.next(this.audioFrameBuffer$.getValue());
}); });
const permissionStatus = await navigator.permissions.query({ this.seeked$.next(0)
name: 'microphone',
});
if (permissionStatus.state === 'granted') {
await navigator.mediaDevices.getUserMedia({
audio: true,
});
this.audioContext.resume();
}
this.seeked$.next(0);
} }
async connectedCallback() { async connectedCallback() {

View File

@ -5,7 +5,7 @@
} }
``` ```
^https://konoplayer.com/api/static/*** resSpeed://10240 #^https://konoplayer.com/api/static/*** resSpeed://10240+
^https://konoplayer.com/api*** reqHeaders://{x-forwarded.json} http://127.0.0.1:5001/api$1 ^https://konoplayer.com/api*** reqHeaders://{x-forwarded.json} http://127.0.0.1:5001/api$1
^https://konoplayer.com/*** reqHeaders://{x-forwarded.json} http://127.0.0.1:5000/$1 excludeFilter://^https://konoplayer.com/api weinre://test ^https://konoplayer.com/*** reqHeaders://{x-forwarded.json} http://127.0.0.1:5000/$1 excludeFilter://^https://konoplayer.com/api weinre://test
^wss://konoplayer.com/*** reqHeaders://{x-forwarded.json} ws://127.0.0.1:5000/$1 excludeFilter://^wss://konoplayer.com/api ^wss://konoplayer.com/*** reqHeaders://{x-forwarded.json} ws://127.0.0.1:5000/$1 excludeFilter://^wss://konoplayer.com/api

View File

@ -23,9 +23,3 @@ export class ParseCodecErrors extends Error {
super('failed to parse codecs'); super('failed to parse codecs');
} }
} }
export class UnimplementedError extends Error {
constructor(detail: string) {
super(`unimplemented: ${detail}`);
}
}

View File

@ -6,16 +6,10 @@ import {
shareReplay, shareReplay,
map, map,
combineLatest, combineLatest,
of, of, type Observable, delayWhen, pipe, finalize, tap, throwIfEmpty,
type Observable,
delayWhen,
throwIfEmpty,
} from 'rxjs'; } from 'rxjs';
import { isTagIdPos } from '../util'; import { isTagIdPos } from '../util';
import { import {createRangedEbmlStream, type CreateRangedEbmlStreamOptions} from './resource';
createRangedEbmlStream,
type CreateRangedEbmlStreamOptions,
} from './resource';
import { type MatroskaSegmentModel, createMatroskaSegment } from './segment'; import { type MatroskaSegmentModel, createMatroskaSegment } from './segment';
export type CreateMatroskaOptions = Omit< export type CreateMatroskaOptions = Omit<
@ -30,9 +24,7 @@ export interface MatroskaModel {
segment: MatroskaSegmentModel; segment: MatroskaSegmentModel;
} }
export function createMatroska( export function createMatroska(options: CreateMatroskaOptions): Observable<MatroskaModel> {
options: CreateMatroskaOptions
): Observable<MatroskaModel> {
const metadataRequest$ = createRangedEbmlStream({ const metadataRequest$ = createRangedEbmlStream({
...options, ...options,
byteStart: 0, byteStart: 0,
@ -40,20 +32,21 @@ export function createMatroska(
return metadataRequest$.pipe( return metadataRequest$.pipe(
switchMap(({ totalSize, ebml$, response }) => { switchMap(({ totalSize, ebml$, response }) => {
/** /**
* while [matroska v4](https://www.matroska.org/technical/elements.html) doc tell that there is only one segment in a file * while [matroska v4](https://www.matroska.org/technical/elements.html) doc tell that there is only one segment in a file
* some mkv generated by strange tools will emit several * some mkv generated by strange tools will emit several
*/ */
const segment$ = ebml$.pipe( const segment$ = ebml$.pipe(
filter(isTagIdPos(EbmlTagIdEnum.Segment, EbmlTagPosition.Start)), filter(isTagIdPos(EbmlTagIdEnum.Segment, EbmlTagPosition.Start)),
map((startTag) => map((startTag) => createMatroskaSegment({
createMatroskaSegment({
startTag, startTag,
matroskaOptions: options, matroskaOptions: options,
ebml$, ebml$,
}) })),
delayWhen(
({ loadedMetadata$ }) => loadedMetadata$
), ),
delayWhen(({ loadedMetadata$ }) => loadedMetadata$),
take(1), take(1),
shareReplay(1) shareReplay(1)
); );
@ -62,7 +55,7 @@ export function createMatroska(
filter(isTagIdPos(EbmlTagIdEnum.EBML, EbmlTagPosition.End)), filter(isTagIdPos(EbmlTagIdEnum.EBML, EbmlTagPosition.End)),
take(1), take(1),
shareReplay(1), shareReplay(1),
throwIfEmpty(() => new Error('failed to find head tag')) throwIfEmpty(() => new Error("failed to find head tag"))
); );
return combineLatest({ return combineLatest({

View File

@ -24,7 +24,6 @@ import {
finalize, finalize,
delayWhen, delayWhen,
from, from,
combineLatest,
} from 'rxjs'; } from 'rxjs';
import type { CreateMatroskaOptions } from '.'; import type { CreateMatroskaOptions } from '.';
import { type ClusterType, TrackTypeRestrictionEnum } from '../schema'; import { type ClusterType, TrackTypeRestrictionEnum } from '../schema';
@ -114,7 +113,7 @@ export function createMatroskaSegment({
filter(({ canComplete }) => canComplete), filter(({ canComplete }) => canComplete),
map(({ segment }) => segment), map(({ segment }) => segment),
take(1), take(1),
shareReplay(1) shareReplay(1),
); );
const loadedRemoteCues$ = loadedMetadata$.pipe( const loadedRemoteCues$ = loadedMetadata$.pipe(
@ -305,33 +304,22 @@ export function createMatroskaSegment({
map(({ decoder, frame$ }) => { map(({ decoder, frame$ }) => {
const clusterSystem = segment.cluster; const clusterSystem = segment.cluster;
const infoSystem = segment.info; const infoSystem = segment.info;
const trackSystem = segment.track;
const timestampScale = Number(infoSystem.info.TimestampScale) / 1000; const timestampScale = Number(infoSystem.info.TimestampScale) / 1000;
const frameProcessing = trackSystem.buildFrameEncodingProcessor(
track.trackEntry
);
const decodeSubscription = cluster$.subscribe((cluster) => { const decodeSubscription = cluster$.subscribe((cluster) => {
for (const block of clusterSystem.enumerateBlocks( for (const block of clusterSystem.enumerateBlocks(
cluster, cluster,
track.trackEntry track.trackEntry
)) { )) {
const blockTime = const blockTime = (Number(cluster.Timestamp) + block.relTime) * timestampScale;
(Number(cluster.Timestamp) + block.relTime) * timestampScale;
const blockDuration = const blockDuration =
frames.length > 1 frames.length > 1 ? track.predictBlockDuration(blockTime) * timestampScale : 0;
? track.predictBlockDuration(blockTime) * timestampScale
: 0;
const perFrameDuration = const perFrameDuration =
frames.length > 1 && blockDuration frames.length > 1 && blockDuration
? blockDuration / block.frames.length ? blockDuration / block.frames.length
: 0; : 0;
for (let frame of block.frames) { for (const frame of block.frames) {
if (frameProcessing) {
frame = frameProcessing(frame);
}
const chunk = new EncodedVideoChunk({ const chunk = new EncodedVideoChunk({
type: block.keyframe ? 'key' : 'delta', type: block.keyframe ? 'key' : 'delta',
data: frame, data: frame,
@ -346,12 +334,13 @@ export function createMatroskaSegment({
return { return {
track, track,
decoder, decoder,
frame$: frame$.pipe( frame$: frame$
.pipe(
finalize(() => { finalize(() => {
decodeSubscription.unsubscribe(); decodeSubscription.unsubscribe();
}) })
), )
}; }
}) })
); );
}; };
@ -364,20 +353,14 @@ export function createMatroskaSegment({
map(({ decoder, frame$ }) => { map(({ decoder, frame$ }) => {
const clusterSystem = segment.cluster; const clusterSystem = segment.cluster;
const infoSystem = segment.info; const infoSystem = segment.info;
const trackSystem = segment.track;
const timestampScale = Number(infoSystem.info.TimestampScale) / 1000; const timestampScale = Number(infoSystem.info.TimestampScale) / 1000;
const frameProcessing = trackSystem.buildFrameEncodingProcessor(
track.trackEntry
);
const decodeSubscription = cluster$.subscribe((cluster) => { const decodeSubscription = cluster$.subscribe((cluster) => {
for (const block of clusterSystem.enumerateBlocks( for (const block of clusterSystem.enumerateBlocks(
cluster, cluster,
track.trackEntry track.trackEntry
)) { )) {
const blockTime = const blockTime = (Number(cluster.Timestamp) + block.relTime) * timestampScale;
(Number(cluster.Timestamp) + block.relTime) * timestampScale;
const blockDuration = const blockDuration =
frames.length > 1 ? track.predictBlockDuration(blockTime) : 0; frames.length > 1 ? track.predictBlockDuration(blockTime) : 0;
const perFrameDuration = const perFrameDuration =
@ -386,10 +369,7 @@ export function createMatroskaSegment({
: 0; : 0;
let i = 0; let i = 0;
for (let frame of block.frames) { for (const frame of block.frames) {
if (frameProcessing) {
frame = frameProcessing(frame);
}
const chunk = new EncodedAudioChunk({ const chunk = new EncodedAudioChunk({
type: block.keyframe ? 'key' : 'delta', type: block.keyframe ? 'key' : 'delta',
data: frame, data: frame,
@ -407,8 +387,7 @@ export function createMatroskaSegment({
decoder, decoder,
frame$: frame$.pipe(finalize(() => decodeSubscription.unsubscribe())), frame$: frame$.pipe(finalize(() => decodeSubscription.unsubscribe())),
}; };
}) }));
);
}; };
const defaultVideoTrack$ = loadedMetadata$.pipe( const defaultVideoTrack$ = loadedMetadata$.pipe(
@ -443,6 +422,6 @@ export function createMatroskaSegment({
videoTrackDecoder, videoTrackDecoder,
audioTrackDecoder, audioTrackDecoder,
defaultVideoTrack$, defaultVideoTrack$,
defaultAudioTrack$, defaultAudioTrack$
}; };
} }

View File

@ -7,7 +7,7 @@ import {
type TrackEntryType, type TrackEntryType,
} from '../schema'; } from '../schema';
import { type SegmentComponent } from './segment'; import { type SegmentComponent } from './segment';
import { SegmentComponentSystemTrait } from './segment-component'; import {SegmentComponentSystemTrait} from "./segment-component";
export abstract class BlockViewTrait { export abstract class BlockViewTrait {
abstract get keyframe(): boolean; abstract get keyframe(): boolean;
@ -82,21 +82,6 @@ export class ClusterSystem extends SegmentComponentSystemTrait<
cluster: ClusterType, cluster: ClusterType,
track: TrackEntryType track: TrackEntryType
): Generator<BlockViewTrait> { ): Generator<BlockViewTrait> {
if (cluster.BlockGroup && cluster.SimpleBlock) {
const blocks = [];
for (const block of cluster.BlockGroup) {
if (block.Block.track === track.TrackNumber) {
blocks.push(new BlockGroupView(block));
}
}
for (const block of cluster.SimpleBlock) {
if (block.track === track.TrackNumber) {
blocks.push(new SimpleBlockView(block));
}
}
blocks.sort((a, b) => a.relTime - b.relTime);
yield* blocks;
} else {
if (cluster.SimpleBlock) { if (cluster.SimpleBlock) {
for (const block of cluster.SimpleBlock) { for (const block of cluster.SimpleBlock) {
if (block.track === track.TrackNumber) { if (block.track === track.TrackNumber) {
@ -112,5 +97,4 @@ export class ClusterSystem extends SegmentComponentSystemTrait<
} }
} }
} }
}
} }

View File

@ -1,6 +1,5 @@
import { import {
ParseCodecErrors, ParseCodecErrors,
UnimplementedError,
UnsupportedCodecError, UnsupportedCodecError,
} from '@konoplayer/core/errors'; } from '@konoplayer/core/errors';
import { import {
@ -16,14 +15,13 @@ import {
type VideoDecoderConfigExt, type VideoDecoderConfigExt,
} from '../codecs'; } from '../codecs';
import { import {
ContentCompAlgoRestrictionEnum,
ContentEncodingTypeRestrictionEnum,
TrackEntrySchema, TrackEntrySchema,
type TrackEntryType, type TrackEntryType,
TrackTypeRestrictionEnum, TrackTypeRestrictionEnum,
} from '../schema'; } from '../schema';
import type { SegmentComponent } from './segment'; import type { SegmentComponent } from './segment';
import { SegmentComponentSystemTrait } from './segment-component'; import {SegmentComponentSystemTrait} from "./segment-component";
import {pick} from "lodash-es";
export interface GetTrackEntryOptions { export interface GetTrackEntryOptions {
priority?: (v: SegmentComponent<TrackEntryType>) => number; priority?: (v: SegmentComponent<TrackEntryType>) => number;
@ -228,49 +226,4 @@ export class TrackSystem extends SegmentComponentSystemTrait<
} }
return true; return true;
} }
buildFrameEncodingProcessor(
track: TrackEntryType
): undefined | ((source: Uint8Array) => Uint8Array) {
let encodings = track.ContentEncodings?.ContentEncoding;
if (!encodings?.length) {
return undefined;
}
encodings = encodings.toSorted(
(a, b) => Number(b.ContentEncodingOrder) - Number(a.ContentEncodingOrder)
);
const processors: Array<(source: Uint8Array) => Uint8Array> = [];
for (const encoing of encodings) {
if (
encoing.ContentEncodingType ===
ContentEncodingTypeRestrictionEnum.COMPRESSION
) {
const compression = encoing.ContentCompression;
const algo = compression?.ContentCompAlgo;
if (algo === ContentCompAlgoRestrictionEnum.HEADER_STRIPPING) {
const settings = compression?.ContentCompSettings;
if (settings?.length) {
processors.push((source: Uint8Array) => {
const dest = new Uint8Array(source.length + settings.length);
dest.set(source);
dest.set(settings, source.length);
return dest;
});
}
} else {
// TODO: dynamic import packages to support more compression algos
throw new UnimplementedError(
`compression algo ${ContentCompAlgoRestrictionEnum[algo as ContentCompAlgoRestrictionEnum]}`
);
}
}
}
return function processor(source: Uint8Array): Uint8Array<ArrayBufferLike> {
let dest = source;
for (const processor of processors) {
dest = processor(dest);
}
return dest;
};
}
} }