将yolo-fastest自训练模型转成rknn,并在rv1126下实现推理

2023-11-07 18:40

本文主要是介绍将yolo-fastest自训练模型转成rknn,并在rv1126下实现推理,希望对大家解决编程问题提供一定的参考价值,需要的开发者们随着小编来一起学习吧!

关于如何用训练自己的yolo-fastest模型,上一篇博文已经说明,现记录先近期的实验。

环境:
系统:ubuntu1804
软件:rknn-toolkit 1.6.0(根据Rockchip_Quick_Start_RKNN_Toolkit_Vx.x.x_CN.pdf文档,部署好其他软件环境,opencv numpy tensorflow…等一系列依赖)
硬件:rv1126开发板(rp pro-rv1126 2+8)

一、模型转换
1、将准备好相应的文件
(1)yolo-fastest.cfg,该文件是自己训练时候修改过的配置文件
(2)yolo-fastest_best.weights, 自训练的权重文件
(3)ai_0006.jpg,需要推理的图片
(4)dataset.txt,该文件的内容是推理图片的路径,如下

./ai_0006.jpg

(5)trans-yolofastest.py,内容如下

from PIL import Image
import numpy as np
from matplotlib import pyplot as plt
import re
import math
import random
from rknn.api import RKNNif __name__ == '__main__':rknn=RKNN()print('load model...')ret = rknn.load_darknet(model='./yolo-fastest.cfg', weight='./yolo-fastest_best.weights')if ret != 0:print('load err...')exit(ret)print('done')rknn.config(reorder_channel='0,1,2', mean_values=[[0,0,0]],std_values=[[255,255,255]],target_platform=['rv1126'])print('building...')ret = rknn.build(do_quantization=True, dataset='./dataset.txt')if ret != 0:print('build fail!')exit(ret)print('done')ret = rknn.export_rknn('./yolo-fastest.rknn')if ret != 0:print('export fail!')exit(ret)exit(0)

2、对模型进行转换
(1)确保已经在文章开头描述的软件环境中
(2)运行模型转换代码

yolo-fast-zyj$ python3 trans-yolofastest.py

运行结果如下图,并查看路径下是否已经生成rknn模型
在这里插入图片描述
二、模型推理
1、准备好推理代码文件run_yolo-fastest_rknn.py,需要修改几个关键的地方
(1)GRID0、GRID1根据yolo算法的grid cell来修改;就是输出单元大小,例如yolov3是13x13,26x26,52x52。
(2)LISTSIZE=NUL_CLS+5,就是识别种类加5,比如yolov4识别80种类,则LISTSIZE=80+5,我这里只识别两个种类,所以填的LISTSIZE=7
(3)CLASSES为识别种类,也就是标注的时候填的类别名称
(4)masks和anchors根据yolo-fastest.cfg文件来填写
(5)配置目标NPU和ID,rknn.init_runtime(target=‘rv1126’,device_id=‘6de927292515e514’)
(6)图像处理outputs在处理时要注意输出的维度,有时候reshape会报异常是因为你前面的GRID0~2配置不对。
具体修改后的代码如下

from PIL import Image
import numpy as np
from matplotlib import pyplot as pltimport re
import math
import random
import cv2from rknn.api import RKNNGRID0 = 10
GRID1 = 20
GRID2 = 52
LISTSIZE = 7
SPAN = 3
NUM_CLS = 2
MAX_BOXES = 500
OBJ_THRESH = 0.5
NMS_THRESH = 0.6'''
CLASSES = ("person", "bicycle", "car","motorbike ","aeroplane ","bus ","train","truck ","boat","traffic light","fire hydrant","stop sign ","parking meter","bench","bird","cat","dog ","horse ","sheep","cow","elephant","bear","zebra ","giraffe","backpack","umbrella","handbag","tie","suitcase","frisbee","skis","snowboard","sports ball","kite","baseball bat","baseball glove","skateboard","surfboard","tennis racket","bottle","wine glass","cup","fork","knife ","spoon","bowl","banana","apple","sandwich","orange","broccoli","carrot","hot dog","pizza ","donut","cake","chair","sofa","pottedplant","bed","diningtable","toilet ","tvmonitor","laptop	","mouse	","remote ","keyboard ","cell phone","microwave ","oven ","toaster","sink","refrigerator ","book","clock","vase","scissors ","teddy bear ","hair drier", "toothbrush ")CLASSES = ("aeroplane","bicycle","bird","boat","bottle","bus","car","cat","chair","cow","diningtable","dog","horse","motorbike","person","pottedplant",
"sheep","sofa","train","tvmonitor")
'''
CLASSES = ("zyj","muzhuang")def sigmoid(x):return 1 / (1 + np.exp(-x))def process(input, mask, anchors):anchors = [anchors[i] for i in mask]grid_h, grid_w = map(int, input.shape[0:2])box_confidence = sigmoid(input[..., 4])box_confidence = np.expand_dims(box_confidence, axis=-1)box_class_probs = sigmoid(input[..., 5:])box_xy = sigmoid(input[..., :2])box_wh = np.exp(input[..., 2:4])box_wh = box_wh * anchorscol = np.tile(np.arange(0, grid_w), grid_w).reshape(-1, grid_w)row = np.tile(np.arange(0, grid_h).reshape(-1, 1), grid_h)col = col.reshape(grid_h, grid_w, 1, 1).repeat(3, axis=-2)row = row.reshape(grid_h, grid_w, 1, 1).repeat(3, axis=-2)grid = np.concatenate((col, row), axis=-1)box_xy += gridbox_xy /= (grid_w, grid_h)box_wh /= (416, 416)box_xy -= (box_wh / 2.)box = np.concatenate((box_xy, box_wh), axis=-1)return box, box_confidence, box_class_probsdef filter_boxes(boxes, box_confidences, box_class_probs):"""Filter boxes with object threshold.# Argumentsboxes: ndarray, boxes of objects.box_confidences: ndarray, confidences of objects.box_class_probs: ndarray, class_probs of objects.# Returnsboxes: ndarray, filtered boxes.classes: ndarray, classes for boxes.scores: ndarray, scores for boxes."""box_scores = box_confidences * box_class_probsbox_classes = np.argmax(box_scores, axis=-1)box_class_scores = np.max(box_scores, axis=-1)pos = np.where(box_class_scores >= OBJ_THRESH)boxes = boxes[pos]classes = box_classes[pos]scores = box_class_scores[pos]return boxes, classes, scoresdef nms_boxes(boxes, scores):"""Suppress non-maximal boxes.# Argumentsboxes: ndarray, boxes of objects.scores: ndarray, scores of objects.# Returnskeep: ndarray, index of effective boxes."""x = boxes[:, 0]y = boxes[:, 1]w = boxes[:, 2]h = boxes[:, 3]areas = w * horder = scores.argsort()[::-1]keep = []while order.size > 0:i = order[0]keep.append(i)xx1 = np.maximum(x[i], x[order[1:]])yy1 = np.maximum(y[i], y[order[1:]])xx2 = np.minimum(x[i] + w[i], x[order[1:]] + w[order[1:]])yy2 = np.minimum(y[i] + h[i], y[order[1:]] + h[order[1:]])w1 = np.maximum(0.0, xx2 - xx1 + 0.00001)h1 = np.maximum(0.0, yy2 - yy1 + 0.00001)inter = w1 * h1ovr = inter / (areas[i] + areas[order[1:]] - inter)inds = np.where(ovr <= NMS_THRESH)[0]order = order[inds + 1]keep = np.array(keep)return keepdef yolov4_post_process(input_data):# yolov3# masks = [[6, 7, 8], [3, 4, 5], [0, 1, 2]]# anchors = [[10, 13], [16, 30], [33, 23], [30, 61], [62, 45],#          [59, 119], [116, 90], [156, 198], [373, 326]]# yolov3-tiny# masks = [[3, 4, 5], [0, 1, 2]]# anchors = [[10, 14], [23, 27], [37, 58], [81, 82], [135, 169], [344, 319]]#yolov4#masks = [[6, 7, 8], [3, 4, 5], [0, 1, 2]]#anchors = [[12, 16], [19, 36], [40, 28], [36, 75], [76, 55], [72, 146], [142, 110], [192, 243], [459, 401]]#yolov4-tiny#masks = [[1, 2, 3], [3, 4, 5]]#anchors = [[10, 14], [23, 27], [37, 58], [81, 82], [135, 169], [344, 319]]#yolo-fastestmasks = [[0, 1, 2], [3, 4, 5]]anchors = [[26, 48], [67, 84], [72, 175], [189, 126], [137, 236], [265, 259]]boxes, classes, scores = [], [], []for input,mask in zip(input_data, masks):b, c, s = process(input, mask, anchors)b, c, s = filter_boxes(b, c, s)boxes.append(b)classes.append(c)scores.append(s)boxes = np.concatenate(boxes)classes = np.concatenate(classes)scores = np.concatenate(scores)nboxes, nclasses, nscores = [], [], []for c in set(classes):inds = np.where(classes == c)b = boxes[inds]c = classes[inds]s = scores[inds]keep = nms_boxes(b, s)nboxes.append(b[keep])nclasses.append(c[keep])nscores.append(s[keep])if not nclasses and not nscores:return None, None, Noneboxes = np.concatenate(nboxes)classes = np.concatenate(nclasses)scores = np.concatenate(nscores)return boxes, classes, scoresdef draw(image, boxes, scores, classes):"""Draw the boxes on the image.# Argument:image: original image.boxes: ndarray, boxes of objects.classes: ndarray, classes of objects.scores: ndarray, scores of objects.all_classes: all classes name."""for box, score, cl in zip(boxes, scores, classes):x, y, w, h = boxprint('class: {}, score: {}'.format(CLASSES[cl], score))print('box coordinate left,top,right,down: [{}, {}, {}, {}]'.format(x, y, x+w, y+h))x *= image.shape[1]y *= image.shape[0]w *= image.shape[1]h *= image.shape[0]top = max(0, np.floor(x + 0.5).astype(int))left = max(0, np.floor(y + 0.5).astype(int))right = min(image.shape[1], np.floor(x + w + 0.5).astype(int))bottom = min(image.shape[0], np.floor(y + h + 0.5).astype(int))# print('class: {}, score: {}'.format(CLASSES[cl], score))# print('box coordinate left,top,right,down: [{}, {}, {}, {}]'.format(top, left, right, bottom))cv2.rectangle(image, (top, left), (right, bottom), (255, 0, 0), 2)cv2.putText(image, '{0} {1:.2f}'.format(CLASSES[cl], score),(top, left - 6),cv2.FONT_HERSHEY_SIMPLEX,0.6, (0, 0, 255), 2)if __name__ == '__main__':# Create RKNN objectrknn = RKNN()# Load tensorflow modelprint('--> Loading model')ret = rknn.load_rknn('./yolo-fastest.rknn')if ret != 0:print('load rknn model failed')exit(ret)print('done')# Set inputsim_file = 'ai_0006.jpg'img = cv2.imread(im_file)orig_img = cv2.resize(img, (320,320))img = cv2.cvtColor(orig_img, cv2.COLOR_BGR2RGB)# init runtime environmentprint('--> Init runtime environment')ret = rknn.init_runtime(target='rv1126',device_id='6de927292515e514')if ret != 0:print('Init runtime environment failed')exit(ret)print('done')# Inferenceprint('--> Running model')outputs = rknn.inference(inputs=[img])rknn.release()#input0_data = np.reshape(outputs[2], (SPAN, LISTSIZE, GRID0, GRID0))input1_data = np.reshape(outputs[1], (SPAN, LISTSIZE, GRID1, GRID1))input2_data = np.reshape(outputs[0], (SPAN, LISTSIZE, GRID0, GRID0))input_data = []#input_data.append(np.transpose(input0_data, (2, 3, 0, 1)))input_data.append(np.transpose(input1_data, (2, 3, 0, 1)))input_data.append(np.transpose(input2_data, (2, 3, 0, 1)))boxes, classes, scores = yolov4_post_process(input_data)if boxes is not None:draw(orig_img, boxes, scores, classes)cv2.imshow("results",orig_img)cv2.waitKeyEx(0)print('done')exit(0)

2、运行代码及结果如下:

yolo-fast-zyj$ pyhton3 run_yolo-fastest_rknn.py

在这里插入图片描述
三、总结
倒腾了好几次,最后终于搞好了。最主要还是要到官方去找下资料,细心点就OK了。

这篇关于将yolo-fastest自训练模型转成rknn,并在rv1126下实现推理的文章就介绍到这儿,希望我们推荐的文章对编程师们有所帮助!



http://www.chinasem.cn/article/365557

相关文章

SpringBoot3实现Gzip压缩优化的技术指南

《SpringBoot3实现Gzip压缩优化的技术指南》随着Web应用的用户量和数据量增加,网络带宽和页面加载速度逐渐成为瓶颈,为了减少数据传输量,提高用户体验,我们可以使用Gzip压缩HTTP响应,... 目录1、简述2、配置2.1 添加依赖2.2 配置 Gzip 压缩3、服务端应用4、前端应用4.1 N

SpringBoot实现数据库读写分离的3种方法小结

《SpringBoot实现数据库读写分离的3种方法小结》为了提高系统的读写性能和可用性,读写分离是一种经典的数据库架构模式,在SpringBoot应用中,有多种方式可以实现数据库读写分离,本文将介绍三... 目录一、数据库读写分离概述二、方案一:基于AbstractRoutingDataSource实现动态

Python FastAPI+Celery+RabbitMQ实现分布式图片水印处理系统

《PythonFastAPI+Celery+RabbitMQ实现分布式图片水印处理系统》这篇文章主要为大家详细介绍了PythonFastAPI如何结合Celery以及RabbitMQ实现简单的分布式... 实现思路FastAPI 服务器Celery 任务队列RabbitMQ 作为消息代理定时任务处理完整

Java枚举类实现Key-Value映射的多种实现方式

《Java枚举类实现Key-Value映射的多种实现方式》在Java开发中,枚举(Enum)是一种特殊的类,本文将详细介绍Java枚举类实现key-value映射的多种方式,有需要的小伙伴可以根据需要... 目录前言一、基础实现方式1.1 为枚举添加属性和构造方法二、http://www.cppcns.co

使用Python实现快速搭建本地HTTP服务器

《使用Python实现快速搭建本地HTTP服务器》:本文主要介绍如何使用Python快速搭建本地HTTP服务器,轻松实现一键HTTP文件共享,同时结合二维码技术,让访问更简单,感兴趣的小伙伴可以了... 目录1. 概述2. 快速搭建 HTTP 文件共享服务2.1 核心思路2.2 代码实现2.3 代码解读3.

MySQL双主搭建+keepalived高可用的实现

《MySQL双主搭建+keepalived高可用的实现》本文主要介绍了MySQL双主搭建+keepalived高可用的实现,文中通过示例代码介绍的非常详细,对大家的学习或者工作具有一定的参考学习价值,... 目录一、测试环境准备二、主从搭建1.创建复制用户2.创建复制关系3.开启复制,确认复制是否成功4.同

Java实现文件图片的预览和下载功能

《Java实现文件图片的预览和下载功能》这篇文章主要为大家详细介绍了如何使用Java实现文件图片的预览和下载功能,文中的示例代码讲解详细,感兴趣的小伙伴可以跟随小编一起学习一下... Java实现文件(图片)的预览和下载 @ApiOperation("访问文件") @GetMapping("

使用Sentinel自定义返回和实现区分来源方式

《使用Sentinel自定义返回和实现区分来源方式》:本文主要介绍使用Sentinel自定义返回和实现区分来源方式,具有很好的参考价值,希望对大家有所帮助,如有错误或未考虑完全的地方,望不吝赐教... 目录Sentinel自定义返回和实现区分来源1. 自定义错误返回2. 实现区分来源总结Sentinel自定

Java实现时间与字符串互相转换详解

《Java实现时间与字符串互相转换详解》这篇文章主要为大家详细介绍了Java中实现时间与字符串互相转换的相关方法,文中的示例代码讲解详细,感兴趣的小伙伴可以跟随小编一起学习一下... 目录一、日期格式化为字符串(一)使用预定义格式(二)自定义格式二、字符串解析为日期(一)解析ISO格式字符串(二)解析自定义

opencv图像处理之指纹验证的实现

《opencv图像处理之指纹验证的实现》本文主要介绍了opencv图像处理之指纹验证的实现,文中通过示例代码介绍的非常详细,对大家的学习或者工作具有一定的参考学习价值,需要的朋友们下面随着小编来一起学... 目录一、简介二、具体案例实现1. 图像显示函数2. 指纹验证函数3. 主函数4、运行结果三、总结一、