Initial commit

This commit is contained in:
琉璃月光
2025-11-18 17:16:08 +08:00
commit caeb0457f4
26 changed files with 1141 additions and 0 deletions

235
obb_angle/obb_angle.py Normal file
View File

@ -0,0 +1,235 @@
import cv2
import numpy as np
import math
from shapely.geometry import Polygon
from rknnlite.api import RKNNLite
import os
# ------------------- 全局配置变量 -------------------
# 模型相关
CLASSES = ['clamp']
nmsThresh = 0.4
objectThresh = 0.35
# 可视化与保存控制(全局变量,可外部修改)
DRAW_RESULT = True # 是否在输出图像上绘制旋转框
SAVE_PATH = None # 保存路径,如 "./result.jpg";设为 None 则不保存
# RKNN 单例
_rknn_instance = None
# ------------------- RKNN 管理函数 -------------------
def init_rknn(model_path):
"""只加载一次 RKNN 模型"""
global _rknn_instance
if _rknn_instance is None:
_rknn_instance = RKNNLite(verbose=False)
ret = _rknn_instance.load_rknn(model_path)
if ret != 0:
print(f"[ERROR] Failed to load RKNN model: {ret}")
return None
ret = _rknn_instance.init_runtime(core_mask=RKNNLite.NPU_CORE_0)
if ret != 0:
print(f"[ERROR] Failed to init runtime: {ret}")
return None
return _rknn_instance
def release_rknn():
"""释放 RKNN 对象"""
global _rknn_instance
if _rknn_instance:
_rknn_instance.release()
_rknn_instance = None
# ------------------- 工具函数 -------------------
def letterbox_resize(image, size, bg_color=114):
target_width, target_height = size
image_height, image_width, _ = image.shape
scale = min(target_width / image_width, target_height / image_height)
new_width, new_height = int(image_width * scale), int(image_height * scale)
image_resized = cv2.resize(image, (new_width, new_height), interpolation=cv2.INTER_AREA)
canvas = np.ones((target_height, target_width, 3), dtype=np.uint8) * bg_color
offset_x, offset_y = (target_width - new_width) // 2, (target_height - new_height) // 2
canvas[offset_y:offset_y + new_height, offset_x:offset_x + new_width] = image_resized
return canvas, scale, offset_x, offset_y
class DetectBox:
def __init__(self, classId, score, xmin, ymin, xmax, ymax, angle):
self.classId = classId
self.score = score
self.xmin = xmin
self.ymin = ymin
self.xmax = xmax
self.ymax = ymax
self.angle = angle
def rotate_rectangle(x1, y1, x2, y2, a):
cx, cy = (x1 + x2) / 2, (y1 + y2) / 2
cos_a, sin_a = math.cos(a), math.sin(a)
pts = [(x1, y1), (x1, y2), (x2, y2), (x2, y1)]
return [[int(cx + (xx - cx) * cos_a - (yy - cy) * sin_a),
int(cy + (xx - cx) * sin_a + (yy - cy) * cos_a)] for xx, yy in pts]
def intersection(g, p):
g = Polygon(np.array(g).reshape(-1, 2))
p = Polygon(np.array(p).reshape(-1, 2))
if not g.is_valid or not p.is_valid:
return 0
inter = g.intersection(p).area
union = g.area + p.area - inter
return 0 if union == 0 else inter / union
def NMS(detectResult):
predBoxs = []
sort_detectboxs = sorted(detectResult, key=lambda x: x.score, reverse=True)
for i in range(len(sort_detectboxs)):
if sort_detectboxs[i].classId == -1:
continue
p1 = rotate_rectangle(sort_detectboxs[i].xmin, sort_detectboxs[i].ymin,
sort_detectboxs[i].xmax, sort_detectboxs[i].ymax,
sort_detectboxs[i].angle)
predBoxs.append(sort_detectboxs[i])
for j in range(i + 1, len(sort_detectboxs)):
if sort_detectboxs[j].classId == sort_detectboxs[i].classId:
p2 = rotate_rectangle(sort_detectboxs[j].xmin, sort_detectboxs[j].ymin,
sort_detectboxs[j].xmax, sort_detectboxs[j].ymax,
sort_detectboxs[j].angle)
if intersection(p1, p2) > nmsThresh:
sort_detectboxs[j].classId = -1
return predBoxs
def sigmoid(x):
x = np.clip(x, -709, 709) # 防止 exp 溢出
return np.where(x >= 0, 1 / (1 + np.exp(-x)), np.exp(x) / (1 + np.exp(x)))
def softmax(x, axis=-1):
exp_x = np.exp(x - np.max(x, axis=axis, keepdims=True))
return exp_x / np.sum(exp_x, axis=axis, keepdims=True)
def process(out, model_w, model_h, stride, angle_feature, index, scale_w=1, scale_h=1):
class_num = len(CLASSES)
angle_feature = angle_feature.reshape(-1)
xywh = out[:, :64, :]
conf = sigmoid(out[:, 64:, :])
conf = conf.reshape(-1)
boxes = []
for ik in range(model_h * model_w * class_num):
if conf[ik] > objectThresh:
w = ik % model_w
h = (ik % (model_w * model_h)) // model_w
c = ik // (model_w * model_h)
xywh_ = xywh[0, :, (h * model_w) + w].reshape(1, 4, 16, 1)
data = np.arange(16).reshape(1, 1, 16, 1)
xywh_ = softmax(xywh_, 2)
xywh_ = np.sum(xywh_ * data, axis=2).reshape(-1)
xywh_add = xywh_[:2] + xywh_[2:]
xywh_sub = (xywh_[2:] - xywh_[:2]) / 2
angle = (angle_feature[index + (h * model_w) + w] - 0.25) * math.pi
cos_a, sin_a = math.cos(angle), math.sin(angle)
xy = xywh_sub[0] * cos_a - xywh_sub[1] * sin_a, xywh_sub[0] * sin_a + xywh_sub[1] * cos_a
xywh1 = np.array([xy[0] + w + 0.5, xy[1] + h + 0.5, xywh_add[0], xywh_add[1]])
xywh1 *= stride
xmin = (xywh1[0] - xywh1[2] / 2) * scale_w
ymin = (xywh1[1] - xywh1[3] / 2) * scale_h
xmax = (xywh1[0] + xywh1[2] / 2) * scale_w
ymax = (xywh1[1] + xywh1[3] / 2) * scale_h
boxes.append(DetectBox(c, conf[ik], xmin, ymin, xmax, ymax, angle))
return boxes
# ------------------- 主推理函数 -------------------
def detect_two_box_angle(model_path, rgb_frame):
"""
输入模型路径和 RGB 图像numpy array输出夹角和结果图像。
可视化和保存由全局变量 DRAW_RESULT 和 SAVE_PATH 控制。
"""
global _rknn_instance, DRAW_RESULT, SAVE_PATH
if not isinstance(rgb_frame, np.ndarray) or rgb_frame is None:
print(f"[ERROR] detect_two_box_angle 接收到错误类型: {type(rgb_frame)}")
return None, np.zeros((640, 640, 3), np.uint8)
# 注意:输入是 BGR因为 cv2.imread 返回 BGR但内部会转为 RGB 给模型
img = rgb_frame.copy()
img_resized, scale, offset_x, offset_y = letterbox_resize(img, (640, 640))
infer_img = np.expand_dims(cv2.cvtColor(img_resized, cv2.COLOR_BGR2RGB), 0)
try:
rknn = init_rknn(model_path)
if rknn is None:
return None, img
results = rknn.inference([infer_img])
except Exception as e:
print(f"[ERROR] RKNN 推理失败: {e}")
return None, img
outputs = []
for x in results[:-1]:
index, stride = 0, 0
if x.shape[2] == 20:
stride, index = 32, 20*4*20*4 + 20*2*20*2
elif x.shape[2] == 40:
stride, index = 16, 20*4*20*4
elif x.shape[2] == 80:
stride, index = 8, 0
feature = x.reshape(1, 65, -1)
outputs += process(feature, x.shape[3], x.shape[2], stride, results[-1], index)
predbox = NMS(outputs)
print(f"[DEBUG] 检测到 {len(predbox)} 个框")
if len(predbox) < 2:
print("检测少于两个目标,无法计算夹角。")
return None, img
predbox = sorted(predbox, key=lambda x: x.score, reverse=True)
box1, box2 = predbox[:2]
output_img = img.copy() if DRAW_RESULT else img # 若不绘制,则直接用原图
if DRAW_RESULT:
for box in [box1, box2]:
xmin = int((box.xmin - offset_x) / scale)
ymin = int((box.ymin - offset_y) / scale)
xmax = int((box.xmax - offset_x) / scale)
ymax = int((box.ymax - offset_y) / scale)
points = rotate_rectangle(xmin, ymin, xmax, ymax, box.angle)
cv2.polylines(output_img, [np.array(points, np.int32)], True, (0, 255, 0), 2)
def main_direction(box):
w, h = (box.xmax - box.xmin)/scale, (box.ymax - box.ymin)/scale
direction = box.angle if w >= h else box.angle + np.pi/2
return direction % np.pi
dir1 = main_direction(box1)
dir2 = main_direction(box2)
diff = abs(dir1 - dir2)
diff = min(diff, np.pi - diff)
angle_deg = np.degrees(diff)
# 保存结果(如果需要)
if SAVE_PATH:
save_dir = os.path.dirname(SAVE_PATH)
if save_dir: # 非空目录才创建
os.makedirs(save_dir, exist_ok=True)
cv2.imwrite(SAVE_PATH, output_img)
return angle_deg, output_img
# ------------------- 示例调用 -------------------
if __name__ == "__main__":
MODEL_PATH = "./obb.rknn"
IMAGE_PATH = "./11.jpg"
# === 全局控制开关 ===
DRAW_RESULT = True # 是否绘制框
SAVE_PATH = "./result11.jpg" # 保存路径,设为 None 则不保存
frame = cv2.imread(IMAGE_PATH)
if frame is None:
print(f"[ERROR] 无法读取图像: {IMAGE_PATH}")
else:
angle_deg, output_image = detect_two_box_angle(MODEL_PATH, frame)
if angle_deg is not None:
print(f"检测到的角度差: {angle_deg:.2f}°")
else:
print("未能成功检测到目标或计算角度差")