coreml: fixup detection test

This commit is contained in:
Koushik Dutta
2024-04-12 22:04:55 -07:00
parent c0c938d9c4
commit 27a1c5269a
3 changed files with 43 additions and 183 deletions

View File

@@ -120,15 +120,15 @@ class CoreMLPlugin(PredictPlugin, scrypted_sdk.Settings, scrypted_sdk.DeviceProv
self.loop = asyncio.get_event_loop()
self.minThreshold = 0.2
asyncio.ensure_future(self.prepareVisionFramework(), loop=self.loop)
asyncio.ensure_future(self.prepareRecognitionModels(), loop=self.loop)
async def prepareVisionFramework(self):
async def prepareRecognitionModels(self):
try:
await scrypted_sdk.deviceManager.onDevicesChanged(
{
"devices": [
{
"nativeId": "vision",
"nativeId": "recognition",
"type": scrypted_sdk.ScryptedDeviceType.Builtin.value,
"interfaces": [
scrypted_sdk.ScryptedInterface.ObjectDetection.value,

View File

@@ -1,31 +1,16 @@
from __future__ import annotations
import asyncio
from asyncio import Future
import base64
import concurrent.futures
import os
from typing import Any, Tuple, List
import coremltools as ct
import numpy as np
# import Quartz
import scrypted_sdk
# from Foundation import NSData, NSMakeSize
from PIL import Image
from scrypted_sdk import (
Setting,
SettingValue,
ObjectDetectionSession,
ObjectsDetected,
ObjectDetectionResult,
)
import traceback
# import Vision
from predict import PredictPlugin
from common import yolo
from common.text import prepare_text_result, process_text_result
from predict.recognize import RecognizeDetection
def euclidean_distance(arr1, arr2):
return np.linalg.norm(arr1 - arr2)
@@ -41,30 +26,10 @@ def cosine_similarity(vector_a, vector_b):
predictExecutor = concurrent.futures.ThreadPoolExecutor(8, "Vision-Predict")
class CoreMLRecognition(PredictPlugin):
class CoreMLRecognition(RecognizeDetection):
def __init__(self, nativeId: str | None = None):
super().__init__(nativeId=nativeId)
self.inputheight = 320
self.inputwidth = 320
self.labels = {
0: "face",
1: "plate",
2: "text",
}
self.loop = asyncio.get_event_loop()
self.minThreshold = 0.7
self.detectModel = self.downloadModel("scrypted_yolov9c_flt_320")
self.detectInput = self.detectModel.get_spec().description.input[0].name
self.textModel = self.downloadModel("vgg_english_g2")
self.textInput = self.textModel.get_spec().description.input[0].name
self.faceModel = self.downloadModel("inception_resnet_v1")
self.faceInput = self.faceModel.get_spec().description.input[0].name
def downloadModel(self, model: str):
model_version = "v7"
mlmodel = "model"
@@ -82,25 +47,24 @@ class CoreMLRecognition(PredictPlugin):
)
modelFile = os.path.dirname(p)
return ct.models.MLModel(modelFile)
model = ct.models.MLModel(modelFile)
inputName = model.get_spec().description.input[0].name
return model, inputName
def predictDetectModel(self, input):
model, inputName = self.detectModel
out_dict = model.predict({inputName: input})
results = list(out_dict.values())[0][0]
return results
async def getSettings(self) -> list[Setting]:
pass
async def putSetting(self, key: str, value: SettingValue):
self.storage.setItem(key, value)
await self.onDeviceEvent(scrypted_sdk.ScryptedInterface.Settings.value, None)
await scrypted_sdk.deviceManager.requestRestart()
# width, height, channels
def get_input_details(self) -> Tuple[int, int, int]:
return (self.inputwidth, self.inputheight, 3)
def get_input_size(self) -> Tuple[float, float]:
return (self.inputwidth, self.inputheight)
def get_input_format(self) -> str:
return "rgb"
def predictFaceModel(self, input):
model, inputName = self.faceModel
out_dict = model.predict({inputName: input})
return out_dict["var_2167"][0]
def predictTextModel(self, input):
model, inputName = self.textModel
return model.predict({inputName: input})
# def predictVision(self, input: Image.Image) -> asyncio.Future[list[Prediction]]:
# buffer = input.tobytes()
@@ -164,124 +128,3 @@ class CoreMLRecognition(PredictPlugin):
# objs = await future
# ret = self.create_detection_result(objs, src_size, cvss)
# return ret
async def detect_once(self, input: Image.Image, settings: Any, src_size, cvss):
out_dict = await asyncio.get_event_loop().run_in_executor(
predictExecutor, lambda: self.detectModel.predict({self.detectInput: input})
)
results = list(out_dict.values())[0][0]
objs = yolo.parse_yolov9(results)
ret = self.create_detection_result(objs, src_size, cvss)
return ret
async def setEmbedding(self, d: ObjectDetectionResult, image: scrypted_sdk.Image):
try:
l, t, w, h = d["boundingBox"]
face = await image.toBuffer(
{
"crop": {
"left": l,
"top": t,
"width": w,
"height": h,
},
"resize": {
"width": 160,
"height": 160,
},
"format": "rgb",
}
)
faceImage = Image.frombuffer("RGB", (160, 160), face)
image_tensor = np.array(faceImage).astype(np.float32).transpose([2, 0, 1])
processed_tensor = (image_tensor - 127.5) / 128.0
processed_tensor = np.expand_dims(processed_tensor, axis=0)
out_dict = await asyncio.get_event_loop().run_in_executor(
predictExecutor,
lambda: self.faceModel.predict({self.faceInput: processed_tensor}),
)
output = out_dict["var_2167"][0]
b = output.tobytes()
embedding = str(base64.encodebytes(b))
d["embedding"] = embedding
except Exception as e:
traceback.print_exc()
pass
async def setLabel(self, d: ObjectDetectionResult, image: scrypted_sdk.Image):
try:
image_tensor = await prepare_text_result(d, image)
out_dict = self.textModel.predict({self.textInput: image_tensor})
preds = out_dict["linear_2"]
d['label'] = process_text_result(preds)
except Exception as e:
traceback.print_exc()
pass
async def run_detection_image(
self, image: scrypted_sdk.Image, detection_session: ObjectDetectionSession
) -> ObjectsDetected:
ret = await super().run_detection_image(image, detection_session)
detections = ret["detections"]
# non max suppression on detections
for i in range(len(detections)):
d1 = detections[i]
if d1["score"] < self.minThreshold:
continue
for j in range(i + 1, len(detections)):
d2 = detections[j]
if d2["score"] < self.minThreshold:
continue
if d1["className"] != d2["className"]:
continue
l1, t1, w1, h1 = d1["boundingBox"]
l2, t2, w2, h2 = d2["boundingBox"]
r1 = l1 + w1
b1 = t1 + h1
r2 = l2 + w2
b2 = t2 + h2
left = max(l1, l2)
top = max(t1, t2)
right = min(r1, r2)
bottom = min(b1, b2)
if left < right and top < bottom:
area1 = (r1 - l1) * (b1 - t1)
area2 = (r2 - l2) * (b2 - t2)
intersect = (right - left) * (bottom - top)
iou = intersect / (area1 + area2 - intersect)
if iou > 0.5:
if d1["score"] > d2["score"]:
d2["score"] = 0
else:
d1["score"] = 0
# remove anything with score 0
ret["detections"] = [d for d in detections if d["score"] >= self.minThreshold]
futures: List[Future] = []
for d in ret["detections"]:
if d["className"] == "face":
futures.append(asyncio.ensure_future(self.setEmbedding(d, image)))
elif d["className"] == "plate":
futures.append(asyncio.ensure_future(self.setLabel(d, image)))
if len(futures):
await asyncio.wait(futures)
return ret

View File

@@ -112,14 +112,13 @@ class RecognizeDetection(PredictPlugin):
processed_tensor = (image_tensor - 127.5) / 128.0
processed_tensor = np.expand_dims(processed_tensor, axis=0)
out_dict = await asyncio.get_event_loop().run_in_executor(
output = await asyncio.get_event_loop().run_in_executor(
predictExecutor,
lambda: self.predictFaceModel(processed_tensor)
)
output = out_dict["var_2167"][0]
b = output.tobytes()
embedding = str(base64.encodebytes(b))
embedding = base64.b64encode(b).decode("utf-8")
d["embedding"] = embedding
except Exception as e:
@@ -209,5 +208,23 @@ class RecognizeDetection(PredictPlugin):
if len(futures):
await asyncio.wait(futures)
last = None
for d in ret['detections']:
if d["className"] != "face":
continue
check = d.get("embedding")
if check is None:
continue
# decode base64 string check
embedding = base64.b64decode(check)
embedding = np.frombuffer(embedding, dtype=np.float32)
if last is None:
last = embedding
continue
# convert to numpy float32 arrays
similarity = cosine_similarity(last, embedding)
print('similarity', similarity)
last = embedding
return ret