mirror of
https://github.com/koush/scrypted.git
synced 2026-03-01 16:52:56 +00:00
google-device-access: wip webrtc
This commit is contained in:
@@ -1,303 +1,409 @@
|
||||
import { RTCAVMessage, FFMpegInput, MediaManager, ScryptedMimeTypes } from "@scrypted/sdk/types";
|
||||
import child_process from 'child_process';
|
||||
import net from 'net';
|
||||
import { listenZero } from "./listen-cluster";
|
||||
import { listenZero, listenZeroSingleClient } from "./listen-cluster";
|
||||
import { ffmpegLogInitialOutput } from "./media-helpers";
|
||||
|
||||
let wrtc: any;
|
||||
function initalizeWebRtc() {
|
||||
if (wrtc)
|
||||
return;
|
||||
try {
|
||||
wrtc = require('wrtc');
|
||||
}
|
||||
catch (e) {
|
||||
console.warn('loading wrtc failed. trying @koush/wrtc fallback.');
|
||||
wrtc = require('@koush/wrtc');
|
||||
}
|
||||
if (wrtc)
|
||||
return;
|
||||
try {
|
||||
wrtc = require('wrtc');
|
||||
}
|
||||
catch (e) {
|
||||
console.warn('loading wrtc failed. trying @koush/wrtc fallback.');
|
||||
wrtc = require('@koush/wrtc');
|
||||
}
|
||||
|
||||
Object.assign(global, wrtc);
|
||||
Object.assign(global, wrtc);
|
||||
}
|
||||
|
||||
interface RTCSession {
|
||||
pc: RTCPeerConnection;
|
||||
pendingCandidates: RTCIceCandidate[];
|
||||
resolve?: (value: any) => void;
|
||||
pc: RTCPeerConnection;
|
||||
pendingCandidates: RTCIceCandidate[];
|
||||
resolve?: (value: any) => void;
|
||||
}
|
||||
|
||||
const rtcSessions: { [id: string]: RTCSession } = {};
|
||||
|
||||
export function addBuiltins(console: Console, mediaManager: MediaManager) {
|
||||
// older scrypted runtime won't have this property, and wrtc will be built in.
|
||||
if (!mediaManager.builtinConverters)
|
||||
return;
|
||||
// older scrypted runtime won't have this property, and wrtc will be built in.
|
||||
if (!mediaManager.builtinConverters)
|
||||
return;
|
||||
|
||||
mediaManager.builtinConverters.push({
|
||||
fromMimeType: ScryptedMimeTypes.RTCAVAnswer,
|
||||
toMimeType: ScryptedMimeTypes.RTCAVOffer,
|
||||
async convert(data: string | Buffer, fromMimeType: string): Promise<Buffer | string> {
|
||||
const rtcInput: RTCAVMessage = JSON.parse(data.toString());
|
||||
const { id } = rtcInput;
|
||||
const session = rtcSessions[id];
|
||||
const pc = rtcSessions[id].pc;
|
||||
let pendingCandidates: RTCIceCandidateInit[] = [];
|
||||
const rtcSessions: { [id: string]: RTCSession } = {};
|
||||
mediaManager.builtinConverters.push({
|
||||
fromMimeType: ScryptedMimeTypes.RTCAVAnswer,
|
||||
toMimeType: ScryptedMimeTypes.RTCAVOffer,
|
||||
async convert(data: string | Buffer, fromMimeType: string): Promise<Buffer | string> {
|
||||
const rtcInput: RTCAVMessage = JSON.parse(data.toString());
|
||||
const { id } = rtcInput;
|
||||
const session = rtcSessions[id];
|
||||
const pc = rtcSessions[id].pc;
|
||||
let pendingCandidates: RTCIceCandidateInit[] = [];
|
||||
|
||||
// safari sends the candidates before the RTC Answer? watch for that.
|
||||
if (!pc.remoteDescription) {
|
||||
if (!rtcInput.description) {
|
||||
// can't do anything with this yet, candidates out of order.
|
||||
pendingCandidates.push(...(rtcInput.candidates || []));
|
||||
}
|
||||
else {
|
||||
await pc.setRemoteDescription(rtcInput.description);
|
||||
if (!rtcInput.candidates)
|
||||
rtcInput.candidates = [];
|
||||
rtcInput.candidates.push(...pendingCandidates);
|
||||
pendingCandidates = [];
|
||||
}
|
||||
}
|
||||
|
||||
if (pc.remoteDescription && rtcInput.candidates?.length) {
|
||||
for (const candidate of rtcInput.candidates) {
|
||||
pc.addIceCandidate(candidate);
|
||||
}
|
||||
}
|
||||
else if (!session.pendingCandidates.length) {
|
||||
// wait for candidates to come in.
|
||||
await new Promise(resolve => session.resolve = resolve);
|
||||
}
|
||||
const ret: RTCAVMessage = {
|
||||
id,
|
||||
candidates: session.pendingCandidates,
|
||||
description: null,
|
||||
configuration: null,
|
||||
};
|
||||
session.pendingCandidates = [];
|
||||
return Buffer.from(JSON.stringify(ret));
|
||||
// safari sends the candidates before the RTC Answer? watch for that.
|
||||
if (!pc.remoteDescription) {
|
||||
if (!rtcInput.description) {
|
||||
// can't do anything with this yet, candidates out of order.
|
||||
pendingCandidates.push(...(rtcInput.candidates || []));
|
||||
}
|
||||
});
|
||||
|
||||
mediaManager.builtinConverters.push({
|
||||
fromMimeType: ScryptedMimeTypes.FFmpegInput,
|
||||
toMimeType: ScryptedMimeTypes.RTCAVOffer,
|
||||
async convert(ffInputBuffer: string | Buffer, fromMimeType: string): Promise<Buffer | string> {
|
||||
const ffInput: FFMpegInput = JSON.parse(ffInputBuffer.toString());
|
||||
|
||||
const configuration: RTCConfiguration = {
|
||||
iceServers: [
|
||||
{
|
||||
urls: ["turn:turn0.clockworkmod.com", "turn:n0.clockworkmod.com", "turn:n1.clockworkmod.com"],
|
||||
username: "foo",
|
||||
credential: "bar",
|
||||
},
|
||||
],
|
||||
};
|
||||
|
||||
initalizeWebRtc();
|
||||
const pc = new RTCPeerConnection(configuration);
|
||||
const id = Math.random().toString();
|
||||
const session: RTCSession = {
|
||||
pc,
|
||||
pendingCandidates: [],
|
||||
};
|
||||
rtcSessions[id] = session;
|
||||
|
||||
pc.onicecandidate = evt => {
|
||||
if (evt.candidate) {
|
||||
// console.log('local candidate', evt.candidate);
|
||||
session.pendingCandidates.push(evt.candidate);
|
||||
session.resolve?.(null);
|
||||
}
|
||||
}
|
||||
|
||||
const { RTCVideoSource, RTCAudioSource } = wrtc.nonstandard;
|
||||
|
||||
const videoSource = new RTCVideoSource();
|
||||
pc.addTrack(videoSource.createTrack());
|
||||
|
||||
|
||||
let audioPort: number;
|
||||
|
||||
// wrtc causes browser to hang if there's no audio track? so always make sure one exists.
|
||||
const noAudio = ffInput.mediaStreamOptions && ffInput.mediaStreamOptions.audio === null;
|
||||
|
||||
let audioServer: net.Server;
|
||||
if (!noAudio) {
|
||||
const audioSource = new RTCAudioSource();
|
||||
pc.addTrack(audioSource.createTrack());
|
||||
|
||||
audioServer = net.createServer(async (socket) => {
|
||||
audioServer.close()
|
||||
const { sample_rate, channels } = await sampleInfo;
|
||||
const bitsPerSample = 16;
|
||||
const channelCount = channels[1] === 'mono' ? 1 : 2;
|
||||
const sampleRate = parseInt(sample_rate[1]);
|
||||
|
||||
const toRead = sampleRate / 100 * channelCount * 2;
|
||||
socket.on('readable', () => {
|
||||
while (true) {
|
||||
const buffer: Buffer = socket.read(toRead);
|
||||
if (!buffer)
|
||||
return;
|
||||
|
||||
const ab = buffer.buffer.slice(buffer.byteOffset, buffer.byteOffset + toRead)
|
||||
const samples = new Int16Array(ab); // 10 ms of 16-bit mono audio
|
||||
|
||||
const data = {
|
||||
samples,
|
||||
sampleRate,
|
||||
bitsPerSample,
|
||||
channelCount,
|
||||
};
|
||||
try {
|
||||
audioSource.onData(data);
|
||||
}
|
||||
catch (e) {
|
||||
cp.kill();
|
||||
console.error(e);
|
||||
}
|
||||
}
|
||||
});
|
||||
});
|
||||
audioPort = await listenZero(audioServer);
|
||||
}
|
||||
|
||||
const videoServer = net.createServer(async (socket) => {
|
||||
videoServer.close()
|
||||
const res = await resolution;
|
||||
const width = parseInt(res[2]);
|
||||
const height = parseInt(res[3]);
|
||||
const toRead = parseInt(res[2]) * parseInt(res[3]) * 1.5;
|
||||
socket.on('readable', () => {
|
||||
while (true) {
|
||||
const buffer: Buffer = socket.read(toRead);
|
||||
if (!buffer)
|
||||
return;
|
||||
const data = new Uint8ClampedArray(buffer);
|
||||
const frame = { width, height, data };
|
||||
try {
|
||||
videoSource.onFrame(frame)
|
||||
}
|
||||
catch (e) {
|
||||
cp.kill();
|
||||
console.error(e);
|
||||
}
|
||||
}
|
||||
});
|
||||
});
|
||||
const videoPort = await listenZero(videoServer);
|
||||
|
||||
const args = [
|
||||
'-hide_banner',
|
||||
// don't think this is actually necessary but whatever.
|
||||
'-y',
|
||||
];
|
||||
|
||||
args.push(...ffInput.inputArguments);
|
||||
|
||||
if (!noAudio) {
|
||||
// create a dummy audio track if none actually exists.
|
||||
// this track will only be used if no audio track is available.
|
||||
// https://stackoverflow.com/questions/37862432/ffmpeg-output-silent-audio-track-if-source-has-no-audio-or-audio-is-shorter-th
|
||||
args.push('-f', 'lavfi', '-i', 'anullsrc=cl=1', '-shortest');
|
||||
|
||||
args.push('-vn');
|
||||
args.push('-acodec', 'pcm_s16le');
|
||||
args.push('-f', 's16le');
|
||||
args.push(`tcp://127.0.0.1:${audioPort}`);
|
||||
}
|
||||
|
||||
args.push('-an');
|
||||
// chromecast seems to crap out on higher than 15fps??? is there
|
||||
// some webrtc video negotiation that is failing here?
|
||||
args.push('-r', '15');
|
||||
args.push('-vcodec', 'rawvideo');
|
||||
args.push('-pix_fmt', 'yuv420p');
|
||||
args.push('-f', 'rawvideo');
|
||||
args.push(`tcp://127.0.0.1:${videoPort}`);
|
||||
|
||||
console.log(ffInput);
|
||||
console.log(args);
|
||||
|
||||
const cp = child_process.spawn(await mediaManager.getFFmpegPath(), args, {
|
||||
// DO NOT IGNORE STDIO, NEED THE DATA FOR RESOLUTION PARSING, ETC.
|
||||
});
|
||||
ffmpegLogInitialOutput(console, cp);
|
||||
cp.on('error', e => console.error('ffmpeg error', e));
|
||||
|
||||
cp.on('exit', () => {
|
||||
videoServer.close();
|
||||
audioServer?.close();
|
||||
pc.close();
|
||||
});
|
||||
|
||||
const resolution = new Promise<Array<string>>(resolve => {
|
||||
cp.stdout.on('data', data => {
|
||||
const stdout = data.toString();
|
||||
const res = /(([0-9]{2,5})x([0-9]{2,5}))/.exec(stdout);
|
||||
if (res)
|
||||
resolve(res);
|
||||
});
|
||||
cp.stderr.on('data', data => {
|
||||
const stdout = data.toString();
|
||||
const res = /(([0-9]{2,5})x([0-9]{2,5}))/.exec(stdout);
|
||||
if (res)
|
||||
resolve(res);
|
||||
});
|
||||
});
|
||||
|
||||
interface SampleInfo {
|
||||
sample_rate: string[];
|
||||
channels: string[];
|
||||
}
|
||||
|
||||
const sampleInfo = new Promise<SampleInfo>(resolve => {
|
||||
const parser = (data: Buffer) => {
|
||||
const stdout = data.toString();
|
||||
const sample_rate = /([0-9]+) Hz/i.exec(stdout)
|
||||
const channels = /Audio:.* (stereo|mono)/.exec(stdout)
|
||||
if (sample_rate && channels) {
|
||||
resolve({
|
||||
sample_rate, channels,
|
||||
});
|
||||
}
|
||||
};
|
||||
cp.stdout.on('data', parser);
|
||||
cp.stderr.on('data', parser);
|
||||
});
|
||||
|
||||
const checkConn = () => {
|
||||
if (pc.iceConnectionState === 'failed' || pc.connectionState === 'failed') {
|
||||
delete rtcSessions[id];
|
||||
cp.kill();
|
||||
}
|
||||
}
|
||||
|
||||
pc.onconnectionstatechange = checkConn;
|
||||
pc.oniceconnectionstatechange = checkConn;
|
||||
|
||||
setTimeout(() => {
|
||||
if (pc.connectionState !== 'connected') {
|
||||
pc.close();
|
||||
cp.kill();
|
||||
}
|
||||
}, 60000);
|
||||
|
||||
const offer = await pc.createOffer({
|
||||
offerToReceiveAudio: false,
|
||||
offerToReceiveVideo: false,
|
||||
});
|
||||
await pc.setLocalDescription(offer);
|
||||
|
||||
const ret: RTCAVMessage = {
|
||||
id,
|
||||
candidates: [],
|
||||
description: offer,
|
||||
configuration,
|
||||
}
|
||||
|
||||
return Buffer.from(JSON.stringify(ret));
|
||||
else {
|
||||
await pc.setRemoteDescription(rtcInput.description);
|
||||
if (!rtcInput.candidates)
|
||||
rtcInput.candidates = [];
|
||||
rtcInput.candidates.push(...pendingCandidates);
|
||||
pendingCandidates = [];
|
||||
}
|
||||
})
|
||||
}
|
||||
|
||||
if (pc.remoteDescription && rtcInput.candidates?.length) {
|
||||
for (const candidate of rtcInput.candidates) {
|
||||
pc.addIceCandidate(candidate);
|
||||
}
|
||||
}
|
||||
else if (!session.pendingCandidates.length) {
|
||||
// wait for candidates to come in.
|
||||
await new Promise(resolve => session.resolve = resolve);
|
||||
}
|
||||
const ret: RTCAVMessage = {
|
||||
id,
|
||||
candidates: session.pendingCandidates,
|
||||
description: null,
|
||||
configuration: null,
|
||||
};
|
||||
session.pendingCandidates = [];
|
||||
return Buffer.from(JSON.stringify(ret));
|
||||
}
|
||||
});
|
||||
|
||||
mediaManager.builtinConverters.push({
|
||||
fromMimeType: ScryptedMimeTypes.FFmpegInput,
|
||||
toMimeType: ScryptedMimeTypes.RTCAVOffer,
|
||||
async convert(ffInputBuffer: string | Buffer, fromMimeType: string): Promise<Buffer | string> {
|
||||
const ffInput: FFMpegInput = JSON.parse(ffInputBuffer.toString());
|
||||
|
||||
const configuration: RTCConfiguration = {
|
||||
iceServers: [
|
||||
{
|
||||
urls: ["turn:turn0.clockworkmod.com", "turn:n0.clockworkmod.com", "turn:n1.clockworkmod.com"],
|
||||
username: "foo",
|
||||
credential: "bar",
|
||||
},
|
||||
],
|
||||
};
|
||||
|
||||
initalizeWebRtc();
|
||||
const pc = new RTCPeerConnection(configuration);
|
||||
const id = Math.random().toString();
|
||||
const session: RTCSession = {
|
||||
pc,
|
||||
pendingCandidates: [],
|
||||
};
|
||||
rtcSessions[id] = session;
|
||||
|
||||
pc.onicecandidate = evt => {
|
||||
if (evt.candidate) {
|
||||
// console.log('local candidate', evt.candidate);
|
||||
session.pendingCandidates.push(evt.candidate);
|
||||
session.resolve?.(null);
|
||||
}
|
||||
}
|
||||
|
||||
const { RTCVideoSource, RTCAudioSource } = wrtc.nonstandard;
|
||||
|
||||
const videoSource = new RTCVideoSource();
|
||||
pc.addTrack(videoSource.createTrack());
|
||||
|
||||
|
||||
let audioPort: number;
|
||||
|
||||
// wrtc causes browser to hang if there's no audio track? so always make sure one exists.
|
||||
const noAudio = ffInput.mediaStreamOptions && ffInput.mediaStreamOptions.audio === null;
|
||||
|
||||
let audioServer: net.Server;
|
||||
if (!noAudio) {
|
||||
const audioSource = new RTCAudioSource();
|
||||
pc.addTrack(audioSource.createTrack());
|
||||
|
||||
audioServer = net.createServer(async (socket) => {
|
||||
audioServer.close()
|
||||
const { sample_rate, channels } = await sampleInfo;
|
||||
const bitsPerSample = 16;
|
||||
const channelCount = channels[1] === 'mono' ? 1 : 2;
|
||||
const sampleRate = parseInt(sample_rate[1]);
|
||||
|
||||
const toRead = sampleRate / 100 * channelCount * 2;
|
||||
socket.on('readable', () => {
|
||||
while (true) {
|
||||
const buffer: Buffer = socket.read(toRead);
|
||||
if (!buffer)
|
||||
return;
|
||||
|
||||
const ab = buffer.buffer.slice(buffer.byteOffset, buffer.byteOffset + toRead)
|
||||
const samples = new Int16Array(ab); // 10 ms of 16-bit mono audio
|
||||
|
||||
const data = {
|
||||
samples,
|
||||
sampleRate,
|
||||
bitsPerSample,
|
||||
channelCount,
|
||||
};
|
||||
try {
|
||||
audioSource.onData(data);
|
||||
}
|
||||
catch (e) {
|
||||
cp.kill();
|
||||
console.error(e);
|
||||
}
|
||||
}
|
||||
});
|
||||
});
|
||||
audioPort = await listenZero(audioServer);
|
||||
}
|
||||
|
||||
const videoServer = net.createServer(async (socket) => {
|
||||
videoServer.close()
|
||||
const res = await resolution;
|
||||
const width = parseInt(res[2]);
|
||||
const height = parseInt(res[3]);
|
||||
const toRead = parseInt(res[2]) * parseInt(res[3]) * 1.5;
|
||||
socket.on('readable', () => {
|
||||
while (true) {
|
||||
const buffer: Buffer = socket.read(toRead);
|
||||
if (!buffer)
|
||||
return;
|
||||
const data = new Uint8ClampedArray(buffer);
|
||||
const frame = { width, height, data };
|
||||
try {
|
||||
videoSource.onFrame(frame)
|
||||
}
|
||||
catch (e) {
|
||||
cp.kill();
|
||||
console.error(e);
|
||||
}
|
||||
}
|
||||
});
|
||||
});
|
||||
const videoPort = await listenZero(videoServer);
|
||||
|
||||
const args = [
|
||||
'-hide_banner',
|
||||
// don't think this is actually necessary but whatever.
|
||||
'-y',
|
||||
];
|
||||
|
||||
args.push(...ffInput.inputArguments);
|
||||
|
||||
if (!noAudio) {
|
||||
// create a dummy audio track if none actually exists.
|
||||
// this track will only be used if no audio track is available.
|
||||
// https://stackoverflow.com/questions/37862432/ffmpeg-output-silent-audio-track-if-source-has-no-audio-or-audio-is-shorter-th
|
||||
args.push('-f', 'lavfi', '-i', 'anullsrc=cl=1', '-shortest');
|
||||
|
||||
args.push('-vn');
|
||||
args.push('-acodec', 'pcm_s16le');
|
||||
args.push('-f', 's16le');
|
||||
args.push(`tcp://127.0.0.1:${audioPort}`);
|
||||
}
|
||||
|
||||
args.push('-an');
|
||||
// chromecast seems to crap out on higher than 15fps??? is there
|
||||
// some webrtc video negotiation that is failing here?
|
||||
args.push('-r', '15');
|
||||
args.push('-vcodec', 'rawvideo');
|
||||
args.push('-pix_fmt', 'yuv420p');
|
||||
args.push('-f', 'rawvideo');
|
||||
args.push(`tcp://127.0.0.1:${videoPort}`);
|
||||
|
||||
console.log(ffInput);
|
||||
console.log(args);
|
||||
|
||||
const cp = child_process.spawn(await mediaManager.getFFmpegPath(), args, {
|
||||
// DO NOT IGNORE STDIO, NEED THE DATA FOR RESOLUTION PARSING, ETC.
|
||||
});
|
||||
ffmpegLogInitialOutput(console, cp);
|
||||
cp.on('error', e => console.error('ffmpeg error', e));
|
||||
|
||||
cp.on('exit', () => {
|
||||
videoServer.close();
|
||||
audioServer?.close();
|
||||
pc.close();
|
||||
});
|
||||
|
||||
const resolution = new Promise<Array<string>>(resolve => {
|
||||
cp.stdout.on('data', data => {
|
||||
const stdout = data.toString();
|
||||
const res = /(([0-9]{2,5})x([0-9]{2,5}))/.exec(stdout);
|
||||
if (res)
|
||||
resolve(res);
|
||||
});
|
||||
cp.stderr.on('data', data => {
|
||||
const stdout = data.toString();
|
||||
const res = /(([0-9]{2,5})x([0-9]{2,5}))/.exec(stdout);
|
||||
if (res)
|
||||
resolve(res);
|
||||
});
|
||||
});
|
||||
|
||||
interface SampleInfo {
|
||||
sample_rate: string[];
|
||||
channels: string[];
|
||||
}
|
||||
|
||||
const sampleInfo = new Promise<SampleInfo>(resolve => {
|
||||
const parser = (data: Buffer) => {
|
||||
const stdout = data.toString();
|
||||
const sample_rate = /([0-9]+) Hz/i.exec(stdout)
|
||||
const channels = /Audio:.* (stereo|mono)/.exec(stdout)
|
||||
if (sample_rate && channels) {
|
||||
resolve({
|
||||
sample_rate, channels,
|
||||
});
|
||||
}
|
||||
};
|
||||
cp.stdout.on('data', parser);
|
||||
cp.stderr.on('data', parser);
|
||||
});
|
||||
|
||||
const checkConn = () => {
|
||||
if (pc.iceConnectionState === 'failed' || pc.connectionState === 'failed') {
|
||||
delete rtcSessions[id];
|
||||
cp.kill();
|
||||
}
|
||||
}
|
||||
|
||||
pc.onconnectionstatechange = checkConn;
|
||||
pc.oniceconnectionstatechange = checkConn;
|
||||
|
||||
setTimeout(() => {
|
||||
if (pc.connectionState !== 'connected') {
|
||||
pc.close();
|
||||
cp.kill();
|
||||
}
|
||||
}, 60000);
|
||||
|
||||
const offer = await pc.createOffer({
|
||||
offerToReceiveAudio: false,
|
||||
offerToReceiveVideo: false,
|
||||
});
|
||||
await pc.setLocalDescription(offer);
|
||||
|
||||
const ret: RTCAVMessage = {
|
||||
id,
|
||||
candidates: [],
|
||||
description: offer,
|
||||
configuration,
|
||||
}
|
||||
|
||||
return Buffer.from(JSON.stringify(ret));
|
||||
}
|
||||
})
|
||||
}
|
||||
|
||||
export async function createRTCPeerConnectionSink(console: Console, mediaManager: MediaManager, sendOffer: (offer: RTCAVMessage) => Promise<RTCAVMessage>): Promise<{
|
||||
ffmpegInput: FFMpegInput,
|
||||
peerConnection: RTCPeerConnection,
|
||||
}> {
|
||||
initalizeWebRtc();
|
||||
const { RTCVideoSink, RTCAudioSink } = wrtc.nonstandard;
|
||||
|
||||
const pc = new RTCPeerConnection();
|
||||
pc.createDataChannel("dataSendChannel");
|
||||
pc.addTransceiver("audio", {
|
||||
direction: 'recvonly'
|
||||
});
|
||||
pc.addTransceiver("video", {
|
||||
direction: 'recvonly',
|
||||
});
|
||||
const offer = await pc.createOffer();
|
||||
await pc.setLocalDescription(offer);
|
||||
|
||||
pc.onicecandidate = async (evt) => {
|
||||
if (evt.candidate)
|
||||
return;
|
||||
const offer = await pc.createOffer({
|
||||
offerToReceiveAudio: true,
|
||||
offerToReceiveVideo: true,
|
||||
});
|
||||
// await pc.setLocalDescription(offer);
|
||||
|
||||
const offerWithCandidates: RTCAVMessage = {
|
||||
id: undefined,
|
||||
candidates: [],
|
||||
description: {
|
||||
sdp: offer.sdp,
|
||||
type: 'offer',
|
||||
},
|
||||
configuration: {},
|
||||
};
|
||||
const answer = await sendOffer(offerWithCandidates);
|
||||
await pc.setRemoteDescription(answer.description);
|
||||
};
|
||||
|
||||
await pc.setLocalDescription(offer);
|
||||
|
||||
pc.onconnectionstatechange = () => console.log(pc.connectionState);
|
||||
pc.onsignalingstatechange = () => console.log(pc.connectionState);
|
||||
|
||||
return new Promise(resolve => {
|
||||
let audioSink: any;
|
||||
let videoSink: any;
|
||||
|
||||
pc.ontrack = async (ev) => {
|
||||
const { track } = ev;
|
||||
const {url: videoUrl, clientPromise: videoClientPromise} = await listenZeroSingleClient();
|
||||
// const {url: audioUrl, clientPromise: audioClientPromise} = await listenZeroSingleClient();
|
||||
|
||||
const cleanup = () => {
|
||||
console.log('peer connection cleaned up');
|
||||
// audioSink?.stop();
|
||||
// videoSink?.stop();
|
||||
// pc.close()
|
||||
};
|
||||
videoClientPromise.catch(cleanup);
|
||||
// audioClientPromise.catch(cleanup);
|
||||
videoClientPromise.then(videoClient => {
|
||||
videoSink.onframe = (frame: any) => {
|
||||
videoClient.write(frame.frame.data);
|
||||
}
|
||||
videoClient.once('close', cleanup);
|
||||
});
|
||||
// audioClientPromise.then(audioClient => {
|
||||
// audioSink.ondata = (data: any) => {
|
||||
// audioClient.write(data.frame.data);
|
||||
// }
|
||||
// audioClient.once('close', cleanup);
|
||||
// });
|
||||
|
||||
if (track.kind === 'audio') {
|
||||
audioSink = new RTCAudioSink(track);
|
||||
audioSink.ondata = () => {};
|
||||
}
|
||||
if (track.kind === 'video') {
|
||||
videoSink = new RTCVideoSink(track);
|
||||
videoSink.onframe = (frame: any) => {
|
||||
if (!audioSink)
|
||||
return;
|
||||
|
||||
videoSink.onframe = () => {};
|
||||
resolve({
|
||||
peerConnection: pc,
|
||||
ffmpegInput: {
|
||||
url: undefined,
|
||||
inputArguments: [
|
||||
// '-acodec', 'pcm_s16be',
|
||||
// '-f', 's16be',
|
||||
// '-i', audioUrl,
|
||||
'-vcodec', 'rawvideo',
|
||||
'-s', `${frame.frame.width}x${frame.frame.height}`,
|
||||
'-pix_fmt', 'yuv420p',
|
||||
'-f', 'rawvideo',
|
||||
'-i', videoUrl,
|
||||
]
|
||||
},
|
||||
});
|
||||
}
|
||||
}
|
||||
};
|
||||
})
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user