mirror of
https://github.com/azaion/autopilot.git
synced 2026-04-22 22:06:34 +00:00
- addedd NCNN model support to rtsp_ai_player
- added printing of inference FPS - simple AI test bench which can be used to compare models
This commit is contained in:
@@ -1,5 +1,6 @@
|
||||
#include <QDebug>
|
||||
#include <opencv2/highgui.hpp>
|
||||
|
||||
#include "aiengine.h"
|
||||
#include "aiengineinference.h"
|
||||
#include "aiengineimagesaver.h"
|
||||
@@ -8,10 +9,14 @@
|
||||
#include "src-opi5/aiengineinferenceopi5.h"
|
||||
#elif defined(OPENCV_BUILD)
|
||||
#include "src-opencv-onnx/aiengineinferenceopencvonnx.h"
|
||||
#elif defined(NCNN_BUILD)
|
||||
#include "src-ncnn/aiengineinferencencnn.h"
|
||||
#else
|
||||
#include "src-onnx-runtime/aiengineinferenceonnxruntime.h"
|
||||
#endif
|
||||
|
||||
|
||||
|
||||
AiEngine::AiEngine(QString modelPath, QObject *parent)
|
||||
: QObject{parent}
|
||||
{
|
||||
@@ -27,6 +32,8 @@ AiEngine::AiEngine(QString modelPath, QObject *parent)
|
||||
mInference3->initialize(2);
|
||||
#elif defined(OPENCV_BUILD)
|
||||
mInference = new AiEngineInferenceOpencvOnnx(modelPath);
|
||||
#elif defined(NCNN_BUILD)
|
||||
mInference = new AiEngineInferencevNcnn(modelPath);
|
||||
#else
|
||||
mInference = new AiEngineInferencevOnnxRuntime(modelPath);
|
||||
#endif
|
||||
@@ -76,6 +83,7 @@ void AiEngine::inferenceResultsReceivedSlot(AiEngineInferenceResult result)
|
||||
{
|
||||
mFrameCounter++;
|
||||
qDebug() << "FPS = " << (mFrameCounter / (mElapsedTimer.elapsed()/1000.0f));
|
||||
//qDebug() << "DEBUG. inference frame counter:" << mFrameCounter;
|
||||
|
||||
//qDebug() << "AiEngine got inference results in thread: " << QThread::currentThreadId();
|
||||
if (mGimbalClient != nullptr) {
|
||||
@@ -97,19 +105,24 @@ void AiEngine::frameReceivedSlot(cv::Mat frame)
|
||||
{
|
||||
//qDebug() << "AiEngine got frame from RTSP listener in thread: " << QThread::currentThreadId();
|
||||
//cv::imshow("Received Frame", frame);
|
||||
static int framecounter = 0;
|
||||
//qDebug() << "DEBUG. RTSP frame counter:" << framecounter;
|
||||
|
||||
if (mInference->isActive() == false) {
|
||||
//qDebug() << "AiEngine. Inference thread is free. Sending frame to it.";
|
||||
emit inferenceFrame(frame);
|
||||
framecounter++;
|
||||
}
|
||||
#ifdef OPI5_BUILD
|
||||
else if (mInference2->isActive() == false) {
|
||||
//qDebug() << "AiEngine. Inference thread is free. Sending frame to it.";
|
||||
emit inferenceFrame2(frame);
|
||||
framecounter++;
|
||||
}
|
||||
else if (mInference3->isActive() == false) {
|
||||
//qDebug() << "AiEngine. Inference thread is free. Sending frame to it.";
|
||||
emit inferenceFrame3(frame);
|
||||
framecounter++;
|
||||
}
|
||||
#endif
|
||||
}
|
||||
|
||||
@@ -3,10 +3,8 @@
|
||||
#include <QString>
|
||||
|
||||
#ifdef OPI5_BUILD
|
||||
QString rtspVideoUrl = "rtsp://192.168.0.1:8554/live.stream";
|
||||
QString rtspVideoUrl = "rtsp://192.168.168.91:8554/live.stream";
|
||||
#else
|
||||
// Video file from the local MTX RTSP server or gimbal camera.
|
||||
QString rtspVideoUrl = "rtsp://localhost:8554/live.stream";
|
||||
//QString rtspVideoUrl = "rtsp://192.168.0.25:8554/main.264";
|
||||
#endif
|
||||
|
||||
|
||||
@@ -1,5 +1,7 @@
|
||||
#include <QDebug>
|
||||
#include <QtConcurrent/QtConcurrent>
|
||||
#include <opencv2/core.hpp>
|
||||
#include <opencv2/imgcodecs.hpp>
|
||||
#include "aienginertsplistener.h"
|
||||
#include "aiengineconfig.h"
|
||||
|
||||
@@ -39,6 +41,45 @@ void AiEngineRtspListener::stopListening()
|
||||
|
||||
void AiEngineRtspListener::listenLoop(void)
|
||||
{
|
||||
#ifdef AI_BENCH
|
||||
QThread::msleep(2000);
|
||||
|
||||
QString directoryPath = "/home/pama/tmp/photos";
|
||||
QDir directory(directoryPath);
|
||||
|
||||
// Ensure the directory exists
|
||||
if (!directory.exists()) {
|
||||
qDebug() << "Directory does not exist!";
|
||||
exit(1);
|
||||
}
|
||||
|
||||
QStringList filters;
|
||||
filters << "*.jpg" << "*.jpeg" << "*.png" << "*.bmp";
|
||||
directory.setNameFilters(filters);
|
||||
|
||||
QFileInfoList fileInfoList = directory.entryInfoList(QDir::Files, QDir::Name);
|
||||
std::sort(fileInfoList.begin(), fileInfoList.end(), [](const QFileInfo &a, const QFileInfo &b) {
|
||||
return a.fileName() < b.fileName();
|
||||
});
|
||||
|
||||
qDebug() << "Images in folder:" << fileInfoList.size();
|
||||
|
||||
for (const QFileInfo &fileInfo : fileInfoList) {
|
||||
QString filePath = fileInfo.absoluteFilePath();
|
||||
cv::Mat image = cv::imread(filePath.toStdString());
|
||||
if (image.empty()) {
|
||||
qDebug() << "Failed to load image";
|
||||
exit(1);
|
||||
}
|
||||
|
||||
emit frameReceived(image.clone());
|
||||
QThread::msleep(1500);
|
||||
}
|
||||
|
||||
QThread::msleep(2000);
|
||||
qDebug() << "Sleep 2000ms and exit";
|
||||
exit(0);
|
||||
#else
|
||||
qDebug() << "AiEngineRtspListener loop running in thread: " << QThread::currentThreadId();
|
||||
|
||||
mCap.open(rtspVideoUrl.toStdString());
|
||||
@@ -51,4 +92,5 @@ void AiEngineRtspListener::listenLoop(void)
|
||||
emit frameReceived(frame.clone());
|
||||
}
|
||||
}
|
||||
#endif
|
||||
}
|
||||
|
||||
@@ -1,12 +1,16 @@
|
||||
QT += core network serialport
|
||||
QT -= gui
|
||||
CONFIG += c++11 concurrent console
|
||||
CONFIG += concurrent console c++17
|
||||
MOC_DIR = moc
|
||||
OBJECTS_DIR = obj
|
||||
|
||||
SOURCES = $$PWD/*.cpp
|
||||
HEADERS = $$PWD/*.h
|
||||
|
||||
ai_bench {
|
||||
QMAKE_CXXFLAGS += -DAI_BENCH
|
||||
}
|
||||
|
||||
gimbal {
|
||||
message("Using real gimbal camera.")
|
||||
QMAKE_CXXFLAGS += -DGIMBAL
|
||||
@@ -35,6 +39,17 @@ opi5 {
|
||||
SOURCES += $$PWD/src-opi5/*.c $$PWD/src-opi5/*.cpp $$PWD/src-opi5/*.cc
|
||||
HEADERS += $$PWD/src-opi5/*.h
|
||||
}
|
||||
else:ncnn {
|
||||
message("NCNN build")
|
||||
CONFIG += link_pkgconfig
|
||||
PKGCONFIG += opencv4
|
||||
QMAKE_CXXFLAGS += -DNCNN_BUILD -fopenmp
|
||||
QMAKE_LFLAGS += -fopenmp
|
||||
INCLUDEPATH += /opt/ncnn/include
|
||||
LIBS += /opt/ncnn/lib/libncnn.a -lgomp
|
||||
SOURCES += $$PWD/src-ncnn/*.cpp
|
||||
HEADERS += $$PWD/src-ncnn/*.h
|
||||
}
|
||||
else:opencv {
|
||||
message("OpenCV build")
|
||||
message("You must use YOLOv8 ONNX files. Azaion model does not work!")
|
||||
|
||||
@@ -0,0 +1,408 @@
|
||||
#include <QDebug>
|
||||
#include <QThread>
|
||||
#include <iostream>
|
||||
#include <vector>
|
||||
#include "aiengineinferencencnn.h"
|
||||
|
||||
|
||||
const int target_size = 640;
|
||||
const float prob_threshold = 0.25f;
|
||||
const float nms_threshold = 0.45f;
|
||||
const int num_labels = 10; // 80 object labels in COCO, 10 in Azaion
|
||||
const int MAX_STRIDE = 32;
|
||||
|
||||
|
||||
char* getCharPointerCopy(const QString& modelPath) {
|
||||
QByteArray byteArray = modelPath.toUtf8();
|
||||
char* cString = new char[byteArray.size() + 1]; // Allocate memory
|
||||
std::strcpy(cString, byteArray.constData()); // Copy the data
|
||||
return cString; // Remember to delete[] this when done!
|
||||
}
|
||||
|
||||
|
||||
AiEngineInferencevNcnn::AiEngineInferencevNcnn(QString modelPath, QObject *parent) :
|
||||
AiEngineInference{modelPath, parent}
|
||||
{
|
||||
qDebug() << "TUOMAS AiEngineInferencevNcnn() mModelPath=" << mModelPath;
|
||||
|
||||
yolov8.opt.num_threads = 4;
|
||||
yolov8.opt.use_vulkan_compute = false;
|
||||
|
||||
QString paramPath = modelPath.chopped(3).append("param");
|
||||
char *model = getCharPointerCopy(modelPath);
|
||||
char *param = getCharPointerCopy(paramPath);
|
||||
|
||||
qDebug() << "model:" << model;
|
||||
qDebug() << "param:" << param;
|
||||
|
||||
yolov8.load_param(param);
|
||||
yolov8.load_model(model);
|
||||
}
|
||||
|
||||
|
||||
static inline float intersection_area(const Object& a, const Object& b)
|
||||
{
|
||||
cv::Rect_<float> inter = a.rect & b.rect;
|
||||
return inter.area();
|
||||
}
|
||||
|
||||
|
||||
static void qsort_descent_inplace(std::vector<Object>& objects, int left, int right)
|
||||
{
|
||||
int i = left;
|
||||
int j = right;
|
||||
float p = objects[(left + right) / 2].prob;
|
||||
|
||||
while (i <= j)
|
||||
{
|
||||
while (objects[i].prob > p)
|
||||
i++;
|
||||
|
||||
while (objects[j].prob < p)
|
||||
j--;
|
||||
|
||||
if (i <= j)
|
||||
{
|
||||
// swap
|
||||
std::swap(objects[i], objects[j]);
|
||||
|
||||
i++;
|
||||
j--;
|
||||
}
|
||||
}
|
||||
|
||||
#pragma omp parallel sections
|
||||
{
|
||||
#pragma omp section
|
||||
{
|
||||
if (left < j) qsort_descent_inplace(objects, left, j);
|
||||
}
|
||||
#pragma omp section
|
||||
{
|
||||
if (i < right) qsort_descent_inplace(objects, i, right);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
static void qsort_descent_inplace(std::vector<Object>& objects)
|
||||
{
|
||||
if (objects.empty())
|
||||
return;
|
||||
|
||||
qsort_descent_inplace(objects, 0, objects.size() - 1);
|
||||
}
|
||||
|
||||
|
||||
static void nms_sorted_bboxes(const std::vector<Object>& faceobjects, std::vector<int>& picked, float nms_threshold, bool agnostic = false)
|
||||
{
|
||||
picked.clear();
|
||||
|
||||
const int n = faceobjects.size();
|
||||
|
||||
std::vector<float> areas(n);
|
||||
for (int i = 0; i < n; i++)
|
||||
{
|
||||
areas[i] = faceobjects[i].rect.area();
|
||||
}
|
||||
|
||||
for (int i = 0; i < n; i++)
|
||||
{
|
||||
const Object& a = faceobjects[i];
|
||||
|
||||
int keep = 1;
|
||||
for (int j = 0; j < (int)picked.size(); j++)
|
||||
{
|
||||
const Object& b = faceobjects[picked[j]];
|
||||
|
||||
if (!agnostic && a.label != b.label)
|
||||
continue;
|
||||
|
||||
// intersection over union
|
||||
float inter_area = intersection_area(a, b);
|
||||
float union_area = areas[i] + areas[picked[j]] - inter_area;
|
||||
// float IoU = inter_area / union_area
|
||||
if (inter_area / union_area > nms_threshold)
|
||||
keep = 0;
|
||||
}
|
||||
|
||||
if (keep)
|
||||
picked.push_back(i);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
static inline float sigmoid(float x)
|
||||
{
|
||||
return static_cast<float>(1.f / (1.f + exp(-x)));
|
||||
}
|
||||
|
||||
|
||||
static inline float clampf(float d, float min, float max)
|
||||
{
|
||||
const float t = d < min ? min : d;
|
||||
return t > max ? max : t;
|
||||
}
|
||||
|
||||
|
||||
static void parse_yolov8_detections(
|
||||
float* inputs, float confidence_threshold,
|
||||
int num_channels, int num_anchors, int num_labels,
|
||||
int infer_img_width, int infer_img_height,
|
||||
std::vector<Object>& objects)
|
||||
{
|
||||
std::vector<Object> detections;
|
||||
cv::Mat output = cv::Mat((int)num_channels, (int)num_anchors, CV_32F, inputs).t();
|
||||
|
||||
for (int i = 0; i < num_anchors; i++)
|
||||
{
|
||||
const float* row_ptr = output.row(i).ptr<float>();
|
||||
const float* bboxes_ptr = row_ptr;
|
||||
const float* scores_ptr = row_ptr + 4;
|
||||
const float* max_s_ptr = std::max_element(scores_ptr, scores_ptr + num_labels);
|
||||
float score = *max_s_ptr;
|
||||
if (score > confidence_threshold)
|
||||
{
|
||||
float x = *bboxes_ptr++;
|
||||
float y = *bboxes_ptr++;
|
||||
float w = *bboxes_ptr++;
|
||||
float h = *bboxes_ptr;
|
||||
|
||||
float x0 = clampf((x - 0.5f * w), 0.f, (float)infer_img_width);
|
||||
float y0 = clampf((y - 0.5f * h), 0.f, (float)infer_img_height);
|
||||
float x1 = clampf((x + 0.5f * w), 0.f, (float)infer_img_width);
|
||||
float y1 = clampf((y + 0.5f * h), 0.f, (float)infer_img_height);
|
||||
|
||||
cv::Rect_<float> bbox;
|
||||
bbox.x = x0;
|
||||
bbox.y = y0;
|
||||
bbox.width = x1 - x0;
|
||||
bbox.height = y1 - y0;
|
||||
Object object;
|
||||
object.label = max_s_ptr - scores_ptr;
|
||||
object.prob = score;
|
||||
object.rect = bbox;
|
||||
detections.push_back(object);
|
||||
}
|
||||
}
|
||||
objects = detections;
|
||||
}
|
||||
|
||||
|
||||
int AiEngineInferencevNcnn::detect_yolov8(const cv::Mat& bgr, std::vector<Object>& objects)
|
||||
{
|
||||
/*
|
||||
int img_w = bgr.cols;
|
||||
int img_h = bgr.rows;
|
||||
|
||||
// letterbox pad to multiple of MAX_STRIDE
|
||||
int w = img_w;
|
||||
int h = img_h;
|
||||
float scale = 1.f;
|
||||
if (w > h)
|
||||
{
|
||||
scale = (float)target_size / w;
|
||||
w = target_size;
|
||||
h = h * scale;
|
||||
}
|
||||
else
|
||||
{
|
||||
scale = (float)target_size / h;
|
||||
h = target_size;
|
||||
w = w * scale;
|
||||
}
|
||||
*/
|
||||
int w = 640;
|
||||
int h = 640;
|
||||
int img_w = 640;
|
||||
int img_h = 640;
|
||||
float scale = 1.f;
|
||||
|
||||
|
||||
ncnn::Mat in = ncnn::Mat::from_pixels_resize(bgr.data, ncnn::Mat::PIXEL_BGR2RGB, img_w, img_h, w, h);
|
||||
|
||||
int wpad = (target_size + MAX_STRIDE - 1) / MAX_STRIDE * MAX_STRIDE - w;
|
||||
int hpad = (target_size + MAX_STRIDE - 1) / MAX_STRIDE * MAX_STRIDE - h;
|
||||
ncnn::Mat in_pad;
|
||||
ncnn::copy_make_border(in, in_pad, hpad / 2, hpad - hpad / 2, wpad / 2, wpad - wpad / 2, ncnn::BORDER_CONSTANT, 114.f);
|
||||
|
||||
const float norm_vals[3] = {1 / 255.f, 1 / 255.f, 1 / 255.f};
|
||||
in_pad.substract_mean_normalize(0, norm_vals);
|
||||
|
||||
auto start = std::chrono::high_resolution_clock::now();
|
||||
|
||||
ncnn::Extractor ex = yolov8.create_extractor();
|
||||
|
||||
ex.input("in0", in_pad);
|
||||
|
||||
std::vector<Object> proposals;
|
||||
|
||||
// stride 32
|
||||
{
|
||||
ncnn::Mat out;
|
||||
ex.extract("out0", out);
|
||||
|
||||
std::vector<Object> objects32;
|
||||
parse_yolov8_detections(
|
||||
(float*)out.data, prob_threshold,
|
||||
out.h, out.w, num_labels,
|
||||
in_pad.w, in_pad.h,
|
||||
objects32);
|
||||
proposals.insert(proposals.end(), objects32.begin(), objects32.end());
|
||||
}
|
||||
|
||||
// sort all proposals by score from highest to lowest
|
||||
qsort_descent_inplace(proposals);
|
||||
|
||||
// apply nms with nms_threshold
|
||||
std::vector<int> picked;
|
||||
nms_sorted_bboxes(proposals, picked, nms_threshold);
|
||||
|
||||
int count = picked.size();
|
||||
|
||||
objects.resize(count);
|
||||
for (int i = 0; i < count; i++)
|
||||
{
|
||||
objects[i] = proposals[picked[i]];
|
||||
|
||||
// adjust offset to original unpadded
|
||||
float x0 = (objects[i].rect.x - (wpad / 2)) / scale;
|
||||
float y0 = (objects[i].rect.y - (hpad / 2)) / scale;
|
||||
float x1 = (objects[i].rect.x + objects[i].rect.width - (wpad / 2)) / scale;
|
||||
float y1 = (objects[i].rect.y + objects[i].rect.height - (hpad / 2)) / scale;
|
||||
|
||||
// clip
|
||||
x0 = std::max(std::min(x0, (float)(img_w - 1)), 0.f);
|
||||
y0 = std::max(std::min(y0, (float)(img_h - 1)), 0.f);
|
||||
x1 = std::max(std::min(x1, (float)(img_w - 1)), 0.f);
|
||||
y1 = std::max(std::min(y1, (float)(img_h - 1)), 0.f);
|
||||
|
||||
objects[i].rect.x = x0;
|
||||
objects[i].rect.y = y0;
|
||||
objects[i].rect.width = x1 - x0;
|
||||
objects[i].rect.height = y1 - y0;
|
||||
}
|
||||
|
||||
auto end = std::chrono::high_resolution_clock::now();
|
||||
auto duration = std::chrono::duration_cast<std::chrono::milliseconds>(end - start);
|
||||
std::cout << "Time taken: " << duration.count() << " milliseconds" << std::endl;
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
static cv::Mat draw_objects(const cv::Mat& bgr, const std::vector<Object>& objects)
|
||||
{
|
||||
static const char* class_names[] = {
|
||||
"Armoured vehicles",
|
||||
"Truck",
|
||||
"Passenger car",
|
||||
"Artillery",
|
||||
"Shadow of the vehicle",
|
||||
"Trenches",
|
||||
"Military",
|
||||
"Ramps",
|
||||
"Tank with additional protection",
|
||||
"Smoke"
|
||||
};
|
||||
|
||||
static const unsigned char colors[19][3] = {
|
||||
{54, 67, 244},
|
||||
{99, 30, 233},
|
||||
{176, 39, 156},
|
||||
{183, 58, 103},
|
||||
{181, 81, 63},
|
||||
{243, 150, 33},
|
||||
{244, 169, 3},
|
||||
{212, 188, 0},
|
||||
{136, 150, 0},
|
||||
{80, 175, 76},
|
||||
{74, 195, 139},
|
||||
{57, 220, 205},
|
||||
{59, 235, 255},
|
||||
{7, 193, 255},
|
||||
{0, 152, 255},
|
||||
{34, 87, 255},
|
||||
{72, 85, 121},
|
||||
{158, 158, 158},
|
||||
{139, 125, 96}
|
||||
};
|
||||
|
||||
int color_index = 0;
|
||||
|
||||
cv::Mat image = bgr.clone();
|
||||
|
||||
for (size_t i = 0; i < objects.size(); i++)
|
||||
{
|
||||
const Object& obj = objects[i];
|
||||
|
||||
const unsigned char* color = colors[color_index % 19];
|
||||
color_index++;
|
||||
|
||||
cv::Scalar cc(color[0], color[1], color[2]);
|
||||
|
||||
fprintf(stderr, "%d = %.5f at %.2f %.2f %.2f x %.2f\n", obj.label, obj.prob,
|
||||
obj.rect.x, obj.rect.y, obj.rect.width, obj.rect.height);
|
||||
|
||||
cv::rectangle(image, obj.rect, cc, 2);
|
||||
|
||||
char text[256];
|
||||
sprintf(text, "%s %.1f%%", class_names[obj.label], obj.prob * 100);
|
||||
|
||||
int baseLine = 0;
|
||||
cv::Size label_size = cv::getTextSize(text, cv::FONT_HERSHEY_SIMPLEX, 0.5, 1, &baseLine);
|
||||
|
||||
int x = obj.rect.x;
|
||||
int y = obj.rect.y - label_size.height - baseLine;
|
||||
if (y < 0)
|
||||
y = 0;
|
||||
if (x + label_size.width > image.cols)
|
||||
x = image.cols - label_size.width;
|
||||
|
||||
cv::rectangle(image, cv::Rect(cv::Point(x, y), cv::Size(label_size.width, label_size.height + baseLine)),
|
||||
cc, -1);
|
||||
|
||||
cv::putText(image, text, cv::Point(x, y + label_size.height),
|
||||
cv::FONT_HERSHEY_SIMPLEX, 0.5, cv::Scalar(255, 255, 255));
|
||||
}
|
||||
|
||||
return image;
|
||||
}
|
||||
|
||||
|
||||
void AiEngineInferencevNcnn::performInferenceSlot(cv::Mat frame)
|
||||
{
|
||||
mActive = true;
|
||||
|
||||
cv::Mat scaledImage = resizeAndPad(frame);
|
||||
|
||||
std::vector<Object> objects;
|
||||
detect_yolov8(scaledImage, objects);
|
||||
|
||||
if (objects.empty() == false) {
|
||||
AiEngineInferenceResult result;
|
||||
result.frame = draw_objects(scaledImage, objects);
|
||||
|
||||
for (uint i = 0; i < objects.size(); i++) {
|
||||
const Object &detection = objects[i];
|
||||
AiEngineObject object;
|
||||
object.classId = detection.label;
|
||||
object.classStr = mClassNames[detection.label];
|
||||
object.propability = detection.prob;
|
||||
object.rectangle.top = detection.rect.y;
|
||||
object.rectangle.left = detection.rect.x;
|
||||
object.rectangle.bottom = detection.rect.y + detection.rect.height;
|
||||
object.rectangle.right = detection.rect.x + detection.rect.width;
|
||||
result.objects.append(object);
|
||||
}
|
||||
|
||||
emit resultsReady(result);
|
||||
}
|
||||
|
||||
mActive = false;
|
||||
}
|
||||
|
||||
|
||||
void AiEngineInferencevNcnn::initialize(int number)
|
||||
{
|
||||
(void)number;
|
||||
}
|
||||
@@ -0,0 +1,31 @@
|
||||
#pragma once
|
||||
|
||||
#include <QObject>
|
||||
#include <opencv2/core.hpp>
|
||||
#include <ncnn/net.h>
|
||||
#include "aiengineinference.h"
|
||||
|
||||
|
||||
struct Object
|
||||
{
|
||||
cv::Rect_<float> rect;
|
||||
int label;
|
||||
float prob;
|
||||
};
|
||||
|
||||
|
||||
class AiEngineInferencevNcnn : public AiEngineInference
|
||||
{
|
||||
Q_OBJECT
|
||||
public:
|
||||
explicit AiEngineInferencevNcnn(QString modelPath, QObject *parent = nullptr);
|
||||
void initialize(int number);
|
||||
|
||||
public slots:
|
||||
void performInferenceSlot(cv::Mat frame) override;
|
||||
|
||||
private:
|
||||
int detect_yolov8(const cv::Mat& bgr, std::vector<Object>& objects);
|
||||
ncnn::Net yolov8;
|
||||
};
|
||||
|
||||
@@ -4,9 +4,9 @@
|
||||
#include "aiengineinferenceonnxruntime.h"
|
||||
|
||||
|
||||
static const float confThreshold = 0.2f;
|
||||
static const float iouThreshold = 0.4f;
|
||||
static const float maskThreshold = 0.5f;
|
||||
static const float confThreshold = 0.25f;
|
||||
static const float iouThreshold = 0.45f;
|
||||
static const float maskThreshold = 0.45f;
|
||||
|
||||
|
||||
AiEngineInferencevOnnxRuntime::AiEngineInferencevOnnxRuntime(QString modelPath, QObject *parent) :
|
||||
|
||||
@@ -70,6 +70,7 @@ void AiEngineInferenceOpi5::freeImageBuffer(image_buffer_t& imgBuffer)
|
||||
|
||||
cv::Mat AiEngineInferenceOpi5::resizeToHalfAndAssigntoTopLeft640x640(const cv::Mat& inputFrame)
|
||||
{
|
||||
/*
|
||||
// Resize input frame to half size
|
||||
cv::Mat resizedFrame;
|
||||
cv::resize(inputFrame, resizedFrame, cv::Size(), 0.5, 0.5);
|
||||
@@ -81,6 +82,25 @@ cv::Mat AiEngineInferenceOpi5::resizeToHalfAndAssigntoTopLeft640x640(const cv::M
|
||||
cv::Rect roi(0, 0, resizedFrame.cols, resizedFrame.rows);
|
||||
resizedFrame.copyTo(outputFrame(roi));
|
||||
|
||||
return outputFrame;
|
||||
*/
|
||||
|
||||
const int targetWidth = 640;
|
||||
const int targetHeight = 640;
|
||||
float aspectRatio = static_cast<float>(inputFrame.cols) / static_cast<float>(inputFrame.rows);
|
||||
int newWidth = targetWidth;
|
||||
int newHeight = static_cast<int>(targetWidth / aspectRatio);
|
||||
if (newHeight > targetHeight) {
|
||||
newHeight = targetHeight;
|
||||
newWidth = static_cast<int>(targetHeight * aspectRatio);
|
||||
}
|
||||
|
||||
cv::Mat resizedFrame;
|
||||
cv::resize(inputFrame, resizedFrame, cv::Size(newWidth, newHeight));
|
||||
cv::Mat outputFrame = cv::Mat::zeros(targetHeight, targetWidth, inputFrame.type());
|
||||
cv::Rect roi(cv::Point(0, 0), resizedFrame.size());
|
||||
resizedFrame.copyTo(outputFrame(roi));
|
||||
|
||||
return outputFrame;
|
||||
}
|
||||
|
||||
@@ -91,7 +111,9 @@ void AiEngineInferenceOpi5::drawObjects(cv::Mat& image, const object_detect_resu
|
||||
const object_detect_result& result = result_list.results[i];
|
||||
|
||||
if (result.cls_id >= mClassNames.size()) {
|
||||
continue;
|
||||
//result.cls_id = result.cls_id % mClassNames.size();
|
||||
qDebug() << "Class id >= mClassNames.size() Reducing it.";
|
||||
//continue;
|
||||
}
|
||||
|
||||
fprintf(stderr, "TUOMAS [%d] prop = %f\n", i, result.prop);
|
||||
@@ -106,7 +128,7 @@ void AiEngineInferenceOpi5::drawObjects(cv::Mat& image, const object_detect_resu
|
||||
// Text
|
||||
char c_text[256];
|
||||
//sprintf(c_text, "%s %d%%", coco_cls_to_name(result.cls_id), (int)(round(result.prop * 100)));
|
||||
sprintf(c_text, "%s %d%%", mClassNames[result.cls_id].toStdString().c_str(), (int)(round(result.prop * 100)));
|
||||
sprintf(c_text, "%s %d%%", mClassNames[result.cls_id % mClassNames.size()].toStdString().c_str(), (int)(round(result.prop * 100)));
|
||||
cv::Point textOrg(left, top - 5);
|
||||
cv::putText(image, std::string(c_text), textOrg, cv::FONT_HERSHEY_COMPLEX, result.prop, cv::Scalar(0, 0, 255), 1, cv::LINE_AA);
|
||||
}
|
||||
@@ -131,23 +153,25 @@ void AiEngineInferenceOpi5::performInferenceSlot(cv::Mat frame)
|
||||
return;
|
||||
}
|
||||
|
||||
AiEngineInferenceResult result;
|
||||
for (int i = 0; i < od_results.count; i++) {
|
||||
object_detect_result *det_result = &(od_results.results[i]);
|
||||
if (od_results.count > 0) {
|
||||
AiEngineInferenceResult result;
|
||||
for (int i = 0; i < od_results.count; i++) {
|
||||
object_detect_result *det_result = &(od_results.results[i]);
|
||||
qDebug() << "TUOMAS box:" << det_result->box.top << det_result->box.left << det_result->box.bottom << det_result->box.right;
|
||||
AiEngineObject object;
|
||||
object.classId = det_result->cls_id;
|
||||
object.propability = det_result->prop;
|
||||
object.rectangle.top = det_result->box.top;
|
||||
object.rectangle.left = det_result->box.left;
|
||||
object.rectangle.bottom = det_result->box.bottom;
|
||||
object.rectangle.right = det_result->box.right;
|
||||
result.objects.append(object);
|
||||
}
|
||||
|
||||
AiEngineObject object;
|
||||
object.classId = det_result->cls_id;
|
||||
object.propability = det_result->prop;
|
||||
object.rectangle.top = det_result->box.top;
|
||||
object.rectangle.left = det_result->box.left;
|
||||
object.rectangle.bottom = det_result->box.bottom;
|
||||
object.rectangle.right = det_result->box.right;
|
||||
result.objects.append(object);
|
||||
drawObjects(scaledFrame, od_results);
|
||||
result.frame = scaledFrame.clone();
|
||||
emit resultsReady(result);
|
||||
}
|
||||
|
||||
drawObjects(scaledFrame, od_results);
|
||||
result.frame = scaledFrame.clone();
|
||||
emit resultsReady(result);
|
||||
|
||||
mActive = false;
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user