mirror of
https://github.com/koush/scrypted.git
synced 2026-04-12 03:20:21 +01:00
1005 lines
34 KiB
TypeScript
1005 lines
34 KiB
TypeScript
import { Deferred } from '@scrypted/common/src/deferred';
|
|
import { sleep } from '@scrypted/common/src/sleep';
|
|
import sdk, { Camera, DeviceProvider, DeviceState, EventListenerRegister, MediaObject, MediaStreamDestination, MixinDeviceBase, MixinProvider, MotionSensor, ObjectDetection, ObjectDetectionGeneratorResult, ObjectDetectionModel, ObjectDetectionTypes, ObjectDetector, ObjectsDetected, ScryptedDevice, ScryptedDeviceType, ScryptedInterface, ScryptedMimeTypes, ScryptedNativeId, Setting, Settings, SettingValue, VideoCamera, VideoFrame, VideoFrameGenerator } from '@scrypted/sdk';
|
|
import { StorageSettings } from '@scrypted/sdk/storage-settings';
|
|
import crypto from 'crypto';
|
|
import { AutoenableMixinProvider } from "../../../common/src/autoenable-mixin-provider";
|
|
import { SettingsMixinDeviceBase } from "../../../common/src/settings-mixin";
|
|
import { FFmpegVideoFrameGenerator, sharpLib } from './ffmpeg-videoframes';
|
|
import { serverSupportsMixinEventMasking } from './server-version';
|
|
import { getAllDevices, safeParseJson } from './util';
|
|
|
|
const polygonOverlap = require('polygon-overlap');
|
|
const insidePolygon = require('point-inside-polygon');
|
|
|
|
const { systemManager } = sdk;
|
|
|
|
const defaultDetectionDuration = 20;
|
|
const defaultDetectionInterval = 60;
|
|
const defaultDetectionTimeout = 60;
|
|
const defaultMotionDuration = 10;
|
|
|
|
const BUILTIN_MOTION_SENSOR_ASSIST = 'Assist';
|
|
const BUILTIN_MOTION_SENSOR_REPLACE = 'Replace';
|
|
|
|
const objectDetectionPrefix = `${ScryptedInterface.ObjectDetection}:`;
|
|
|
|
type ClipPath = [number, number][];
|
|
type Zones = { [zone: string]: ClipPath };
|
|
interface ZoneInfo {
|
|
exclusion?: boolean;
|
|
type?: 'Intersect' | 'Contain';
|
|
classes?: string[];
|
|
scoreThreshold?: number;
|
|
secondScoreThreshold?: number;
|
|
}
|
|
type ZoneInfos = { [zone: string]: ZoneInfo };
|
|
|
|
class ObjectDetectionMixin extends SettingsMixinDeviceBase<VideoCamera & Camera & MotionSensor & ObjectDetector> implements ObjectDetector, Settings {
|
|
motionListener: EventListenerRegister;
|
|
motionMixinListener: EventListenerRegister;
|
|
detections = new Map<string, MediaObject>();
|
|
cameraDevice: ScryptedDevice & Camera & VideoCamera & MotionSensor & ObjectDetector;
|
|
storageSettings = new StorageSettings(this, {
|
|
newPipeline: {
|
|
title: 'Video Pipeline',
|
|
description: 'Configure how frames are provided to the video analysis pipeline.',
|
|
onGet: async () => {
|
|
const choices = [
|
|
'Default',
|
|
...getAllDevices().filter(d => d.interfaces.includes(ScryptedInterface.VideoFrameGenerator)).map(d => d.name),
|
|
];
|
|
if (!this.hasMotionType)
|
|
choices.push('Snapshot');
|
|
return {
|
|
choices,
|
|
}
|
|
},
|
|
onPut: () => {
|
|
this.endObjectDetection();
|
|
this.maybeStartMotionDetection();
|
|
},
|
|
defaultValue: 'Default',
|
|
},
|
|
motionSensorSupplementation: {
|
|
title: 'Built-In Motion Sensor',
|
|
description: `This camera has a built in motion sensor. Using ${this.objectDetection.name} may be unnecessary and will use additional CPU. Replace will ignore the built in motion sensor. Filter will verify the motion sent by built in motion sensor. The Default is ${BUILTIN_MOTION_SENSOR_REPLACE}.`,
|
|
choices: [
|
|
'Default',
|
|
BUILTIN_MOTION_SENSOR_ASSIST,
|
|
BUILTIN_MOTION_SENSOR_REPLACE,
|
|
],
|
|
defaultValue: "Default",
|
|
onPut: () => {
|
|
this.endObjectDetection();
|
|
this.maybeStartMotionDetection();
|
|
}
|
|
},
|
|
detectionDuration: {
|
|
title: 'Detection Duration',
|
|
subgroup: 'Advanced',
|
|
description: 'The duration in seconds to analyze video when motion occurs.',
|
|
type: 'number',
|
|
defaultValue: defaultDetectionDuration,
|
|
},
|
|
detectionTimeout: {
|
|
title: 'Detection Timeout',
|
|
subgroup: 'Advanced',
|
|
description: 'Timeout in seconds before removing an object that is no longer detected.',
|
|
type: 'number',
|
|
defaultValue: defaultDetectionTimeout,
|
|
},
|
|
motionDuration: {
|
|
title: 'Motion Duration',
|
|
description: 'The duration in seconds to wait to reset the motion sensor.',
|
|
type: 'number',
|
|
defaultValue: defaultMotionDuration,
|
|
},
|
|
motionAsObjects: {
|
|
title: 'Motion Detection Objects',
|
|
description: 'Report motion detections as objects (useful for debugging).',
|
|
type: 'boolean',
|
|
},
|
|
detectionInterval: {
|
|
type: 'number',
|
|
defaultValue: defaultDetectionInterval,
|
|
hide: true,
|
|
},
|
|
});
|
|
motionTimeout: NodeJS.Timeout;
|
|
zones = this.getZones();
|
|
zoneInfos = this.getZoneInfos();
|
|
detectionIntervalTimeout: NodeJS.Timeout;
|
|
analyzeStop = 0;
|
|
detectorSignal = new Deferred<void>().resolve();
|
|
get detectorRunning() {
|
|
return !this.detectorSignal.finished;
|
|
}
|
|
|
|
constructor(public plugin: ObjectDetectionPlugin, mixinDevice: VideoCamera & Camera & MotionSensor & ObjectDetector & Settings, mixinDeviceInterfaces: ScryptedInterface[], mixinDeviceState: { [key: string]: any }, providerNativeId: string, public objectDetection: ObjectDetection & ScryptedDevice, public model: ObjectDetectionModel, group: string, public hasMotionType: boolean, public settings: Setting[]) {
|
|
super({
|
|
mixinDevice, mixinDeviceState,
|
|
mixinProviderNativeId: providerNativeId,
|
|
mixinDeviceInterfaces,
|
|
group,
|
|
groupKey: "objectdetectionplugin:" + objectDetection.id,
|
|
mixinStorageSuffix: objectDetection.id,
|
|
});
|
|
|
|
this.cameraDevice = systemManager.getDeviceById<Camera & VideoCamera & MotionSensor & ObjectDetector>(this.id);
|
|
|
|
this.bindObjectDetection();
|
|
this.register();
|
|
this.resetDetectionTimeout();
|
|
}
|
|
|
|
clearDetectionTimeout() {
|
|
clearTimeout(this.detectionIntervalTimeout);
|
|
this.detectionIntervalTimeout = undefined;
|
|
}
|
|
|
|
resetDetectionTimeout() {
|
|
this.clearDetectionTimeout();
|
|
this.detectionIntervalTimeout = setInterval(async () => {
|
|
if (this.hasMotionType) {
|
|
// force a motion detection restart if it quit
|
|
if (this.motionSensorSupplementation === BUILTIN_MOTION_SENSOR_REPLACE)
|
|
this.startPipelineAnalysis();
|
|
return;
|
|
}
|
|
}, this.storageSettings.values.detectionInterval * 1000);
|
|
}
|
|
|
|
clearMotionTimeout() {
|
|
clearTimeout(this.motionTimeout);
|
|
this.motionTimeout = undefined;
|
|
}
|
|
|
|
resetMotionTimeout() {
|
|
this.clearMotionTimeout();
|
|
this.motionTimeout = setTimeout(() => {
|
|
this.motionDetected = false;
|
|
// if (this.motionSensorSupplementation === BUILTIN_MOTION_SENSOR_ASSIST) {
|
|
// this.console.log(`${this.objectDetection.name} timed out confirming motion, stopping video detection.`)
|
|
// this.endObjectDetection();
|
|
// }
|
|
}, this.storageSettings.values.motionDuration * 1000);
|
|
}
|
|
|
|
getCurrentSettings() {
|
|
if (!this.settings)
|
|
return;
|
|
|
|
const ret: { [key: string]: any } = {};
|
|
for (const setting of this.settings) {
|
|
ret[setting.key] = (setting.multiple ? safeParseJson(this.storage.getItem(setting.key)) : this.storage.getItem(setting.key))
|
|
|| setting.value;
|
|
}
|
|
|
|
if (this.hasMotionType)
|
|
ret['motionAsObjects'] = this.storageSettings.values.motionAsObjects;
|
|
|
|
return ret;
|
|
}
|
|
|
|
async maybeStartMotionDetection() {
|
|
if (!this.hasMotionType)
|
|
return;
|
|
if (this.motionSensorSupplementation !== BUILTIN_MOTION_SENSOR_REPLACE)
|
|
return;
|
|
this.startPipelineAnalysis();
|
|
}
|
|
|
|
endObjectDetection() {
|
|
this.detectorSignal.resolve();
|
|
}
|
|
|
|
bindObjectDetection() {
|
|
if (this.hasMotionType)
|
|
this.motionDetected = false;
|
|
|
|
this.endObjectDetection();
|
|
|
|
this.maybeStartMotionDetection();
|
|
}
|
|
|
|
async register() {
|
|
const model = await this.objectDetection.getDetectionModel();
|
|
|
|
if (!this.hasMotionType) {
|
|
this.motionListener = this.cameraDevice.listen(ScryptedInterface.MotionSensor, async () => {
|
|
if (!this.cameraDevice.motionDetected) {
|
|
if (this.detectorRunning) {
|
|
// allow anaysis due to user request.
|
|
if (this.analyzeStop > Date.now())
|
|
return;
|
|
|
|
this.console.log('motion stopped, cancelling ongoing detection')
|
|
this.endObjectDetection();
|
|
}
|
|
return;
|
|
}
|
|
|
|
this.startPipelineAnalysis();
|
|
});
|
|
|
|
return;
|
|
}
|
|
|
|
if (this.hasMotionType) {
|
|
this.motionMixinListener = this.cameraDevice.listen({
|
|
event: ScryptedInterface.MotionSensor,
|
|
mixinId: this.id,
|
|
}, async (source, details, data) => {
|
|
if (this.motionSensorSupplementation !== BUILTIN_MOTION_SENSOR_ASSIST)
|
|
return;
|
|
if (data) {
|
|
if (this.motionDetected)
|
|
return;
|
|
if (!this.detectorRunning)
|
|
this.console.log('built in motion sensor started motion, starting video detection.');
|
|
this.startPipelineAnalysis();
|
|
return;
|
|
}
|
|
|
|
this.clearMotionTimeout();
|
|
if (this.detectorRunning) {
|
|
this.console.log('built in motion sensor ended motion, stopping video detection.')
|
|
this.endObjectDetection();
|
|
}
|
|
if (this.motionDetected)
|
|
this.motionDetected = false;
|
|
});
|
|
}
|
|
}
|
|
|
|
startPipelineAnalysis() {
|
|
if (!this.detectorSignal.finished)
|
|
return;
|
|
|
|
const signal = this.detectorSignal = new Deferred();
|
|
if (!this.hasMotionType)
|
|
this.plugin.objectDetectionStarted(this.console);
|
|
|
|
const options = {
|
|
onObjectDetection: () => this.plugin.trackDetection(),
|
|
};
|
|
|
|
const start = Date.now();
|
|
this.runPipelineAnalysis(signal, options)
|
|
.catch(e => {
|
|
this.console.error('Video Analysis ended with error', e);
|
|
}).finally(() => {
|
|
if (!this.hasMotionType)
|
|
this.plugin.objectDetectionEnded(this.console);
|
|
else
|
|
this.console.log('Video Analysis motion detection ended.');
|
|
signal.resolve();
|
|
});
|
|
}
|
|
|
|
async runPipelineAnalysis(signal: Deferred<void>, options: {
|
|
onObjectDetection: () => void,
|
|
}) {
|
|
const start = Date.now();
|
|
this.analyzeStop = start + this.getDetectionDuration();
|
|
|
|
let lastStatusTime = Date.now();
|
|
let lastStatus = 'starting';
|
|
const updatePipelineStatus = (status: string) => {
|
|
lastStatus = status;
|
|
lastStatusTime = Date.now();
|
|
}
|
|
|
|
let frameGenerator: AsyncGenerator<VideoFrame & MediaObject, void>;
|
|
let detectionGenerator: AsyncGenerator<ObjectDetectionGeneratorResult, void>;
|
|
const interval = setInterval(() => {
|
|
if (Date.now() - lastStatusTime > 30000) {
|
|
signal.resolve();
|
|
this.console.error('VideoAnalysis is hung and will terminate:', lastStatus);
|
|
}
|
|
}, 30000);
|
|
signal.promise.finally(() => clearInterval(interval));
|
|
|
|
let newPipeline: string = this.newPipeline;
|
|
if (!this.hasMotionType && (!newPipeline || newPipeline === 'Default')) {
|
|
if (this.plugin.shouldUseSnapshotPipeline()) {
|
|
newPipeline = 'Snapshot';
|
|
this.console.warn(`Due to limited performance, Snapshot mode is being used with ${this.plugin.statsSnapshotConcurrent} actively detecting cameras.`);
|
|
}
|
|
}
|
|
|
|
if (newPipeline === 'Snapshot' && !this.hasMotionType) {
|
|
this.console.log('decoder:', 'Snapshot +', this.objectDetection.name);
|
|
const self = this;
|
|
frameGenerator = (async function* gen() {
|
|
try {
|
|
while (!signal.finished) {
|
|
const now = Date.now();
|
|
const sleeper = async () => {
|
|
const diff = now + 1100 - Date.now();
|
|
if (diff > 0)
|
|
await sleep(diff);
|
|
};
|
|
let image: MediaObject & VideoFrame;
|
|
try {
|
|
updatePipelineStatus('takePicture');
|
|
const mo = await self.cameraDevice.takePicture({
|
|
reason: 'event',
|
|
});
|
|
updatePipelineStatus('converting image');
|
|
image = await sdk.mediaManager.convertMediaObject(mo, ScryptedMimeTypes.Image);
|
|
}
|
|
catch (e) {
|
|
self.console.error('Video analysis snapshot failed. Will retry in a moment.');
|
|
await sleeper();
|
|
continue;
|
|
}
|
|
|
|
// self.console.log('yield')
|
|
updatePipelineStatus('processing image');
|
|
yield image;
|
|
// self.console.log('done yield')
|
|
await sleeper();
|
|
}
|
|
}
|
|
finally {
|
|
self.console.log('Snapshot generation finished.');
|
|
}
|
|
})();
|
|
}
|
|
else {
|
|
const destination: MediaStreamDestination = this.hasMotionType ? 'low-resolution' : 'local-recorder';
|
|
const videoFrameGenerator = systemManager.getDeviceById<VideoFrameGenerator>(newPipeline);
|
|
if (!videoFrameGenerator)
|
|
throw new Error('invalid VideoFrameGenerator');
|
|
this.console.log(videoFrameGenerator.name, '+', this.objectDetection.name);
|
|
updatePipelineStatus('getVideoStream');
|
|
const stream = await this.cameraDevice.getVideoStream({
|
|
destination,
|
|
// ask rebroadcast to mute audio, not needed.
|
|
audio: null,
|
|
});
|
|
|
|
frameGenerator = await videoFrameGenerator.generateVideoFrames(stream, {
|
|
resize: this.model?.inputSize ? {
|
|
width: this.model.inputSize[0],
|
|
height: this.model.inputSize[1],
|
|
} : undefined,
|
|
format: this.model?.inputFormat,
|
|
});
|
|
}
|
|
|
|
signal.promise.finally(() => frameGenerator.return());
|
|
|
|
const currentDetections = new Set<string>();
|
|
let lastReport = 0;
|
|
detectionGenerator = await sdk.connectRPCObject(await this.objectDetection.generateObjectDetections(frameGenerator, {
|
|
settings: this.getCurrentSettings(),
|
|
sourceId: this.id,
|
|
}));
|
|
signal.promise.finally(() => detectionGenerator.return());
|
|
|
|
updatePipelineStatus('waiting result');
|
|
|
|
for await (const detected of detectionGenerator) {
|
|
if (signal.finished) {
|
|
break;
|
|
}
|
|
if (!this.hasMotionType && Date.now() > this.analyzeStop) {
|
|
break;
|
|
}
|
|
|
|
// apply the zones to the detections and get a shallow copy list of detections after
|
|
// exclusion zones have applied
|
|
const zonedDetections = this.applyZones(detected.detected);
|
|
detected.detected.detections = zonedDetections;
|
|
|
|
// this.console.warn('dps', detections / (Date.now() - start) * 1000);
|
|
|
|
if (!this.hasMotionType) {
|
|
options.onObjectDetection();
|
|
|
|
for (const d of detected.detected.detections) {
|
|
currentDetections.add(d.className);
|
|
}
|
|
|
|
const now = Date.now();
|
|
if (now > lastReport + 10000) {
|
|
const found = [...currentDetections.values()];
|
|
if (!found.length)
|
|
found.push('[no detections]');
|
|
this.console.log(`[${Math.round((now - start) / 100) / 10}s] Detected:`, ...found);
|
|
currentDetections.clear();
|
|
lastReport = now;
|
|
}
|
|
}
|
|
|
|
if (detected.detected.detectionId) {
|
|
updatePipelineStatus('creating jpeg');
|
|
const jpeg = await detected.videoFrame.toBuffer({
|
|
format: 'jpg',
|
|
});
|
|
updatePipelineStatus('converting jpeg');
|
|
const mo = await sdk.mediaManager.createMediaObject(jpeg, 'image/jpeg');
|
|
this.setDetection(detected.detected, mo);
|
|
// this.console.log('image saved', detected.detected.detections);
|
|
}
|
|
this.reportObjectDetections(detected.detected);
|
|
if (this.hasMotionType) {
|
|
await sleep(250);
|
|
}
|
|
updatePipelineStatus('waiting result');
|
|
// this.handleDetectionEvent(detected.detected);
|
|
}
|
|
|
|
}
|
|
|
|
normalizeBox(boundingBox: [number, number, number, number], inputDimensions: [number, number]) {
|
|
let [x, y, width, height] = boundingBox;
|
|
let x2 = x + width;
|
|
let y2 = y + height;
|
|
// the zones are point paths in percentage format
|
|
x = x * 100 / inputDimensions[0];
|
|
y = y * 100 / inputDimensions[1];
|
|
x2 = x2 * 100 / inputDimensions[0];
|
|
y2 = y2 * 100 / inputDimensions[1];
|
|
const box = [[x, y], [x2, y], [x2, y2], [x, y2]];
|
|
return box;
|
|
}
|
|
|
|
getDetectionDuration() {
|
|
// when motion type, the detection interval is a keepalive reset.
|
|
// the duration needs to simply be an arbitrarily longer time.
|
|
return this.hasMotionType ? this.storageSettings.values.detectionInterval * 1000 * 5 : this.storageSettings.values.detectionDuration * 1000;
|
|
}
|
|
|
|
applyZones(detection: ObjectsDetected) {
|
|
// determine zones of the objects, if configured.
|
|
if (!detection.detections)
|
|
return [];
|
|
let copy = detection.detections.slice();
|
|
for (const o of detection.detections) {
|
|
if (!o.boundingBox)
|
|
continue;
|
|
|
|
o.zones = []
|
|
const box = this.normalizeBox(o.boundingBox, detection.inputDimensions);
|
|
|
|
let included: boolean;
|
|
for (const [zone, zoneValue] of Object.entries(this.zones)) {
|
|
if (zoneValue.length < 3) {
|
|
// this.console.warn(zone, 'Zone is unconfigured, skipping.');
|
|
continue;
|
|
}
|
|
|
|
const zoneInfo = this.zoneInfos[zone];
|
|
// track if there are any inclusion zones
|
|
if (!zoneInfo?.exclusion && !included)
|
|
included = false;
|
|
|
|
let match = false;
|
|
if (zoneInfo?.type === 'Contain') {
|
|
match = insidePolygon(box[0], zoneValue) &&
|
|
insidePolygon(box[1], zoneValue) &&
|
|
insidePolygon(box[2], zoneValue) &&
|
|
insidePolygon(box[3], zoneValue);
|
|
}
|
|
else {
|
|
match = polygonOverlap(box, zoneValue);
|
|
}
|
|
|
|
if (match && zoneInfo?.classes?.length) {
|
|
match = zoneInfo.classes.includes(o.className);
|
|
}
|
|
if (match) {
|
|
o.zones.push(zone);
|
|
|
|
if (zoneInfo?.exclusion && match) {
|
|
copy = copy.filter(c => c !== o);
|
|
break;
|
|
}
|
|
|
|
included = true;
|
|
}
|
|
}
|
|
|
|
// if this is a motion sensor and there are no inclusion zones set up,
|
|
// use a default inclusion zone that crops the top and bottom to
|
|
// prevents errant motion from the on screen time changing every second.
|
|
if (this.hasMotionType && included === undefined) {
|
|
const defaultInclusionZone = [[0, 10], [100, 10], [100, 90], [0, 90]];
|
|
included = polygonOverlap(box, defaultInclusionZone);
|
|
}
|
|
|
|
// if there are inclusion zones and this object
|
|
// was not in any of them, filter it out.
|
|
if (included === false)
|
|
copy = copy.filter(c => c !== o);
|
|
}
|
|
|
|
return copy;
|
|
}
|
|
|
|
reportObjectDetections(detection: ObjectsDetected) {
|
|
if (this.hasMotionType) {
|
|
const found = detection.detections?.find(d => d.className === 'motion');
|
|
if (found) {
|
|
if (!this.motionDetected)
|
|
this.motionDetected = true;
|
|
this.resetMotionTimeout();
|
|
|
|
// if (this.motionSensorSupplementation === BUILTIN_MOTION_SENSOR_ASSIST) {
|
|
// if (!this.motionDetected) {
|
|
// this.motionDetected = true;
|
|
// this.console.log(`${this.objectDetection.name} confirmed motion, stopping video detection.`)
|
|
// this.endObjectDetection();
|
|
// this.clearMotionTimeout();
|
|
// }
|
|
// }
|
|
// else {
|
|
// if (!this.motionDetected)
|
|
// this.motionDetected = true;
|
|
// this.resetMotionTimeout();
|
|
// }
|
|
|
|
const areas = detection.detections.filter(d => d.className === 'motion' && d.score !== 1).map(d => d.score)
|
|
if (areas.length)
|
|
this.console.log('detection areas', areas);
|
|
}
|
|
}
|
|
|
|
if (!this.hasMotionType || this.storageSettings.values.motionAsObjects)
|
|
this.onDeviceEvent(ScryptedInterface.ObjectDetector, detection);
|
|
}
|
|
|
|
setDetection(detection: ObjectsDetected, detectionInput: MediaObject) {
|
|
if (!detection.detectionId)
|
|
detection.detectionId = crypto.randomBytes(4).toString('hex');
|
|
|
|
this.console.log('retaining detection image', ...detection.detections);
|
|
|
|
const { detectionId } = detection;
|
|
this.detections.set(detectionId, detectionInput);
|
|
setTimeout(() => {
|
|
this.detections.delete(detectionId);
|
|
}, 2000);
|
|
}
|
|
|
|
async getNativeObjectTypes(): Promise<ObjectDetectionTypes> {
|
|
if (this.mixinDeviceInterfaces.includes(ScryptedInterface.ObjectDetector))
|
|
return this.mixinDevice.getObjectTypes();
|
|
return {};
|
|
}
|
|
|
|
async getObjectTypes(): Promise<ObjectDetectionTypes> {
|
|
const ret = await this.getNativeObjectTypes();
|
|
if (!ret.classes)
|
|
ret.classes = [];
|
|
ret.classes.push(...(await this.objectDetection.getDetectionModel()).classes);
|
|
return ret;
|
|
}
|
|
|
|
async getDetectionInput(detectionId: any): Promise<MediaObject> {
|
|
const detection = this.detections.get(detectionId);
|
|
if (detection)
|
|
return detection;
|
|
if (this.mixinDeviceInterfaces.includes(ScryptedInterface.ObjectDetector))
|
|
return this.mixinDevice.getDetectionInput(detectionId);
|
|
throw new Error('Detection not found. It may have expired.');
|
|
}
|
|
|
|
get motionSensorSupplementation() {
|
|
if (!serverSupportsMixinEventMasking() || !this.interfaces.includes(ScryptedInterface.MotionSensor))
|
|
return BUILTIN_MOTION_SENSOR_REPLACE;
|
|
|
|
const supp = this.storage.getItem('motionSensorSupplementation');
|
|
switch (supp) {
|
|
case BUILTIN_MOTION_SENSOR_REPLACE:
|
|
return BUILTIN_MOTION_SENSOR_REPLACE;
|
|
case BUILTIN_MOTION_SENSOR_ASSIST:
|
|
return BUILTIN_MOTION_SENSOR_ASSIST;
|
|
}
|
|
|
|
// if (this.mixinDeviceInterfaces?.includes(ScryptedInterface.MotionSensor))
|
|
// return BUILTIN_MOTION_SENSOR_ASSIST;
|
|
return BUILTIN_MOTION_SENSOR_REPLACE;
|
|
}
|
|
|
|
get newPipeline() {
|
|
const newPipeline = this.storageSettings.values.newPipeline;
|
|
if (!newPipeline)
|
|
return newPipeline;
|
|
if (newPipeline === 'Snapshot')
|
|
return newPipeline;
|
|
const pipelines = getAllDevices().filter(d => d.interfaces.includes(ScryptedInterface.VideoFrameGenerator));
|
|
const webcodec = pipelines.find(p => p.nativeId === 'webcodec');
|
|
const gstreamer = pipelines.find(p => p.nativeId === 'gstreamer');
|
|
const libav = pipelines.find(p => p.nativeId === 'libav');
|
|
const ffmpeg = pipelines.find(p => p.nativeId === 'ffmpeg');
|
|
const use = pipelines.find(p => p.name === newPipeline) || webcodec || gstreamer || libav || ffmpeg;
|
|
return use.id;
|
|
}
|
|
|
|
async getMixinSettings(): Promise<Setting[]> {
|
|
const settings: Setting[] = [];
|
|
|
|
try {
|
|
this.settings = (await this.objectDetection.getDetectionModel(this.getCurrentSettings())).settings;
|
|
}
|
|
catch (e) {
|
|
}
|
|
|
|
if (this.settings) {
|
|
settings.push(...this.settings.map(setting =>
|
|
Object.assign({}, setting, {
|
|
placeholder: setting.placeholder?.toString(),
|
|
value: (setting.multiple ? safeParseJson(this.storage.getItem(setting.key)) : this.storage.getItem(setting.key))
|
|
|| setting.value,
|
|
} as Setting))
|
|
);
|
|
}
|
|
|
|
this.storageSettings.settings.motionSensorSupplementation.hide = !this.hasMotionType || !this.mixinDeviceInterfaces.includes(ScryptedInterface.MotionSensor);
|
|
this.storageSettings.settings.detectionDuration.hide = this.hasMotionType;
|
|
this.storageSettings.settings.detectionTimeout.hide = this.hasMotionType;
|
|
this.storageSettings.settings.motionDuration.hide = !this.hasMotionType;
|
|
this.storageSettings.settings.motionAsObjects.hide = !this.hasMotionType;
|
|
|
|
settings.push(...await this.storageSettings.getSettings());
|
|
|
|
settings.push({
|
|
key: 'zones',
|
|
title: 'Zones',
|
|
type: 'string',
|
|
description: 'Enter the name of a new zone or delete an existing zone.',
|
|
multiple: true,
|
|
value: Object.keys(this.zones),
|
|
choices: Object.keys(this.zones),
|
|
combobox: true,
|
|
});
|
|
|
|
for (const [name, value] of Object.entries(this.zones)) {
|
|
const zi = this.zoneInfos[name];
|
|
|
|
const subgroup = `Zone: ${name}`;
|
|
settings.push({
|
|
subgroup,
|
|
key: `zone-${name}`,
|
|
title: `Open Zone Editor`,
|
|
type: 'clippath',
|
|
value: JSON.stringify(value),
|
|
});
|
|
|
|
settings.push({
|
|
subgroup,
|
|
key: `zoneinfo-exclusion-${name}`,
|
|
title: `Exclusion Zone`,
|
|
description: 'Detections in this zone will be excluded.',
|
|
type: 'boolean',
|
|
value: zi?.exclusion,
|
|
});
|
|
|
|
settings.push({
|
|
subgroup,
|
|
key: `zoneinfo-type-${name}`,
|
|
title: `Zone Type`,
|
|
description: 'An Intersect zone will match objects that are partially or fully inside the zone. A Contain zone will only match objects that are fully inside the zone.',
|
|
choices: [
|
|
'Intersect',
|
|
'Contain',
|
|
],
|
|
value: zi?.type || 'Intersect',
|
|
});
|
|
}
|
|
|
|
if (!this.hasMotionType) {
|
|
settings.push(
|
|
{
|
|
title: 'Analyze',
|
|
description: 'Analyzes the video stream for 1 minute. Results will be shown in the Console.',
|
|
key: 'analyzeButton',
|
|
type: 'button',
|
|
},
|
|
);
|
|
}
|
|
|
|
return settings;
|
|
}
|
|
|
|
getZones(): Zones {
|
|
try {
|
|
return JSON.parse(this.storage.getItem('zones'));
|
|
}
|
|
catch (e) {
|
|
return {};
|
|
}
|
|
}
|
|
|
|
getZoneInfos(): ZoneInfos {
|
|
try {
|
|
return JSON.parse(this.storage.getItem('zoneInfos'));
|
|
}
|
|
catch (e) {
|
|
return {};
|
|
}
|
|
}
|
|
|
|
async putMixinSetting(key: string, value: string | number | boolean | string[] | number[]): Promise<void> {
|
|
let vs = value?.toString();
|
|
|
|
if (key === 'zones') {
|
|
const newZones: Zones = {};
|
|
const newZoneInfos: ZoneInfos = {};
|
|
for (const name of value as string[]) {
|
|
newZones[name] = this.zones[name] || [];
|
|
newZoneInfos[name] = this.zoneInfos[name];
|
|
}
|
|
this.zones = newZones;
|
|
this.zoneInfos = newZoneInfos;
|
|
this.storage.setItem('zones', JSON.stringify(newZones));
|
|
this.storage.setItem('zoneInfos', JSON.stringify(newZoneInfos));
|
|
return;
|
|
}
|
|
if (key.startsWith('zone-')) {
|
|
const zoneName = key.substring('zone-'.length);
|
|
if (this.zones[zoneName]) {
|
|
this.zones[zoneName] = JSON.parse(vs);
|
|
this.storage.setItem('zones', JSON.stringify(this.zones));
|
|
}
|
|
return;
|
|
}
|
|
if (key.startsWith('zoneinfo-')) {
|
|
const [zkey, zoneName] = key.substring('zoneinfo-'.length).split('-');
|
|
this.zoneInfos[zoneName] ||= {};
|
|
this.zoneInfos[zoneName][zkey] = value;
|
|
this.storage.setItem('zoneInfos', JSON.stringify(this.zoneInfos));
|
|
return;
|
|
}
|
|
|
|
if (this.storageSettings.settings[key]) {
|
|
return this.storageSettings.putSetting(key, value);
|
|
}
|
|
|
|
if (value && this.settings?.find(s => s.key === key)?.multiple) {
|
|
vs = JSON.stringify(value);
|
|
}
|
|
|
|
if (key === 'analyzeButton') {
|
|
// await this.snapshotDetection();
|
|
this.startPipelineAnalysis();
|
|
this.analyzeStop = Date.now() + 60000;
|
|
}
|
|
else {
|
|
const settings = this.getCurrentSettings();
|
|
if (settings && settings[key]) {
|
|
this.storage.setItem(key, vs);
|
|
settings[key] = value;
|
|
}
|
|
this.bindObjectDetection();
|
|
}
|
|
}
|
|
|
|
async release() {
|
|
super.release();
|
|
this.clearDetectionTimeout();
|
|
this.clearMotionTimeout();
|
|
this.motionListener?.removeListener();
|
|
this.motionMixinListener?.removeListener();
|
|
this.endObjectDetection();
|
|
}
|
|
}
|
|
|
|
class ObjectDetectorMixin extends MixinDeviceBase<ObjectDetection> implements MixinProvider {
|
|
currentMixins = new Set<ObjectDetectionMixin>();
|
|
|
|
constructor(mixinDevice: ObjectDetection, mixinDeviceInterfaces: ScryptedInterface[], mixinDeviceState: DeviceState, public plugin: ObjectDetectionPlugin, public model: ObjectDetectionModel) {
|
|
super({ mixinDevice, mixinDeviceInterfaces, mixinDeviceState, mixinProviderNativeId: plugin.nativeId });
|
|
|
|
// trigger mixin creation. todo: fix this to not be stupid hack.
|
|
for (const id of Object.keys(systemManager.getSystemState())) {
|
|
const device = systemManager.getDeviceById<VideoCamera & Settings>(id);
|
|
if (!device.mixins?.includes(this.id))
|
|
continue;
|
|
device.probe();
|
|
}
|
|
}
|
|
|
|
async canMixin(type: ScryptedDeviceType, interfaces: string[]): Promise<string[]> {
|
|
const hasMotionType = this.model.classes.includes('motion');
|
|
const prefix = `${objectDetectionPrefix}${hasMotionType}`;
|
|
const thisPrefix = `${prefix}:${this.id}`;
|
|
|
|
const found = interfaces.find(iface => iface.startsWith(prefix) && iface !== thisPrefix);
|
|
if (found)
|
|
return;
|
|
// this.console.log('found', found);
|
|
|
|
if ((type === ScryptedDeviceType.Camera || type === ScryptedDeviceType.Doorbell) && (interfaces.includes(ScryptedInterface.VideoCamera) || interfaces.includes(ScryptedInterface.Camera))) {
|
|
const ret: string[] = [
|
|
ScryptedInterface.ObjectDetector,
|
|
ScryptedInterface.Settings,
|
|
thisPrefix,
|
|
];
|
|
const model = await this.mixinDevice.getDetectionModel();
|
|
|
|
if (model.classes?.includes('motion')) {
|
|
ret.push(
|
|
ScryptedInterface.MotionSensor,
|
|
);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
async getMixin(mixinDevice: any, mixinDeviceInterfaces: ScryptedInterface[], mixinDeviceState: { [key: string]: any }) {
|
|
let objectDetection = systemManager.getDeviceById<ObjectDetection>(this.id);
|
|
const hasMotionType = this.model.classes.includes('motion');
|
|
const group = hasMotionType ? 'Motion Detection' : 'Object Detection';
|
|
// const group = objectDetection.name.replace('Plugin', '').trim();
|
|
|
|
const settings = this.model.settings;
|
|
|
|
const ret = new ObjectDetectionMixin(this.plugin, mixinDevice, mixinDeviceInterfaces, mixinDeviceState, this.mixinProviderNativeId, objectDetection, this.model, group, hasMotionType, settings);
|
|
this.currentMixins.add(ret);
|
|
return ret;
|
|
}
|
|
|
|
async releaseMixin(id: string, mixinDevice: any) {
|
|
this.currentMixins.delete(mixinDevice);
|
|
return mixinDevice.release();
|
|
}
|
|
}
|
|
|
|
interface ObjectDetectionStatistics {
|
|
dps: number;
|
|
sampleTime: number;
|
|
}
|
|
|
|
class ObjectDetectionPlugin extends AutoenableMixinProvider implements Settings, DeviceProvider {
|
|
currentMixins = new Set<ObjectDetectorMixin>();
|
|
objectDetectionStatistics = new Map<number, ObjectDetectionStatistics>();
|
|
statsSnapshotTime: number;
|
|
statsSnapshotDetections: number;
|
|
statsSnapshotConcurrent = 0;
|
|
storageSettings = new StorageSettings(this, {
|
|
activeMotionDetections: {
|
|
title: 'Active Motion Detection Sessions',
|
|
readonly: true,
|
|
mapGet: () => {
|
|
return [...this.currentMixins.values()]
|
|
.reduce((c1, v1) => c1 + [...v1.currentMixins.values()]
|
|
.reduce((c2, v2) => c2 + (v2.hasMotionType && v2.detectorRunning ? 1 : 0), 0), 0);
|
|
}
|
|
},
|
|
activeObjectDetections: {
|
|
title: 'Active Object Detection Sessions',
|
|
readonly: true,
|
|
mapGet: () => {
|
|
// could use the stats variable...
|
|
return [...this.currentMixins.values()]
|
|
.reduce((c1, v1) => c1 + [...v1.currentMixins.values()]
|
|
.reduce((c2, v2) => c2 + (!v2.hasMotionType && v2.detectorRunning ? 1 : 0), 0), 0);
|
|
},
|
|
}
|
|
});
|
|
|
|
shouldUseSnapshotPipeline() {
|
|
this.pruneOldStatistics();
|
|
|
|
for (const [k, v] of this.objectDetectionStatistics.entries()) {
|
|
// check the stats history to see if any sessions
|
|
// with same or lower number of cameras were on the struggle bus.
|
|
if (v.dps < 2 && k <= this.statsSnapshotConcurrent)
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
pruneOldStatistics() {
|
|
const now = Date.now();
|
|
for (const [k, v] of this.objectDetectionStatistics.entries()) {
|
|
// purge the stats every hour
|
|
if (Date.now() - v.sampleTime > 60 * 60 * 1000)
|
|
this.objectDetectionStatistics.delete(k);
|
|
}
|
|
}
|
|
|
|
trackDetection() {
|
|
this.statsSnapshotDetections++;
|
|
}
|
|
|
|
objectDetectionStarted(console?: Console) {
|
|
this.resetStats(console);
|
|
|
|
this.statsSnapshotConcurrent++;
|
|
}
|
|
objectDetectionEnded(console?: Console) {
|
|
this.resetStats(console);
|
|
|
|
this.statsSnapshotConcurrent--;
|
|
}
|
|
|
|
resetStats(console?: Console) {
|
|
const now = Date.now();
|
|
const concurrentSessions = this.statsSnapshotConcurrent;
|
|
if (concurrentSessions) {
|
|
const duration = now - this.statsSnapshotTime;
|
|
const stats: ObjectDetectionStatistics = {
|
|
sampleTime: now,
|
|
dps: this.statsSnapshotDetections / (duration / 1000),
|
|
};
|
|
|
|
// ignore short/busted sessions.
|
|
if (duration > 10000 && this.statsSnapshotDetections) {
|
|
this.objectDetectionStatistics.set(concurrentSessions, stats);
|
|
this.pruneOldStatistics();
|
|
}
|
|
|
|
const str = `video analysis, ${concurrentSessions} camera(s), dps: ${Math.round(stats.dps * 10) / 10} (${this.statsSnapshotDetections}/${Math.round(duration / 1000)})`;
|
|
this.console.log(str);
|
|
console?.log(str);
|
|
}
|
|
|
|
this.statsSnapshotDetections = 0;
|
|
this.statsSnapshotTime = now;
|
|
}
|
|
|
|
constructor(nativeId?: ScryptedNativeId) {
|
|
super(nativeId);
|
|
|
|
process.nextTick(() => {
|
|
sdk.deviceManager.onDevicesChanged({
|
|
devices: [
|
|
{
|
|
name: 'FFmpeg Frame Generator',
|
|
type: ScryptedDeviceType.Builtin,
|
|
interfaces: sharpLib ? [
|
|
ScryptedInterface.VideoFrameGenerator,
|
|
] : [],
|
|
nativeId: 'ffmpeg',
|
|
}
|
|
]
|
|
})
|
|
})
|
|
}
|
|
|
|
async getDevice(nativeId: string): Promise<any> {
|
|
if (nativeId === 'ffmpeg')
|
|
return new FFmpegVideoFrameGenerator('ffmpeg');
|
|
}
|
|
|
|
async releaseDevice(id: string, nativeId: string): Promise<void> {
|
|
}
|
|
|
|
getSettings(): Promise<Setting[]> {
|
|
return this.storageSettings.getSettings();
|
|
}
|
|
|
|
putSetting(key: string, value: SettingValue): Promise<void> {
|
|
return this.storageSettings.putSetting(key, value);
|
|
}
|
|
|
|
async canMixin(type: ScryptedDeviceType, interfaces: string[]): Promise<string[]> {
|
|
if (!interfaces.includes(ScryptedInterface.ObjectDetection))
|
|
return;
|
|
return [ScryptedInterface.MixinProvider];
|
|
}
|
|
|
|
async getMixin(mixinDevice: ObjectDetection, mixinDeviceInterfaces: ScryptedInterface[], mixinDeviceState: { [key: string]: any; }): Promise<any> {
|
|
const model = await mixinDevice.getDetectionModel();
|
|
const ret = new ObjectDetectorMixin(mixinDevice, mixinDeviceInterfaces, mixinDeviceState, this, model);
|
|
this.currentMixins.add(ret);
|
|
return ret;
|
|
}
|
|
|
|
async releaseMixin(id: string, mixinDevice: any): Promise<void> {
|
|
// what does this mean to make a mixin provider no longer available?
|
|
// just ignore it until reboot?
|
|
this.currentMixins.delete(mixinDevice);
|
|
}
|
|
}
|
|
|
|
export default ObjectDetectionPlugin;
|