-
Scout Monitoring
Free Django app performance insights with Scout Monitoring. Get Scout setup in minutes, and let us sweat the small stuff. A couple lines in settings.py is all you need to start monitoring your apps. Sign up for our free tier today.
-
tflite2tensorflow
Discontinued Generate saved_model, tfjs, tf-trt, EdgeTPU, CoreML, quantized tflite, ONNX, OpenVINO, Myriad Inference Engine blob and .pb from .tflite. Support for building environments with Docker. It is possible to directly access the host PC GUI and the camera to verify the operation. NVIDIA GPU (dGPU) support. Intel iHD GPU (iGPU) support. Supports inverse quantization of INT8 quantization model.
ToucanTrack can be found at the following repository: https://github.com/noahcoolboy/toucan-track/tree/main
They come in the form of tflite models, so I had to convert them to onnx. I used tf2onnx for converting the pose landmark model and tflite2tensorflow for converting the pose detection model. For improving performance, I had created a small script which modified the landmark models for supporting batch inference. This script is not included in the repository, but do tell me if you need it!
They come in the form of tflite models, so I had to convert them to onnx. I used tf2onnx for converting the pose landmark model and tflite2tensorflow for converting the pose detection model. For improving performance, I had created a small script which modified the landmark models for supporting batch inference. This script is not included in the repository, but do tell me if you need it!
If you're looking for the differences in terms of how inference is done, I recommend you take a look at MediaPipe's source code. MediaPipe doesn't use raw code, but uses a "graph" instead (eg. pose_landmark_cpu.pbtxt), which can be visualised using MediaPipe Viz. I also used axinc-ai/ailia-models as the base (preprocessing, inference, postprocessing, etc...) which I further built upon (keypoint refinement, roi from keypoints, filtering / smoothing, etc...)
If you're looking for the differences in terms of how inference is done, I recommend you take a look at MediaPipe's source code. MediaPipe doesn't use raw code, but uses a "graph" instead (eg. pose_landmark_cpu.pbtxt), which can be visualised using MediaPipe Viz. I also used axinc-ai/ailia-models as the base (preprocessing, inference, postprocessing, etc...) which I further built upon (keypoint refinement, roi from keypoints, filtering / smoothing, etc...)
This looks amazing! I really would like to try it out but I don't feel like downloading CL-Eye Platform SDK from a dropbox. Especially since this is usually a paid product. Are there any possible alternatives for this driver? Something like this one? Thank you!