-
Notifications
You must be signed in to change notification settings - Fork 0
/
yolo2can5.py
191 lines (152 loc) · 6.73 KB
/
yolo2can5.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
import cv2
import can
import torch
from pathlib import Path
from PIL import Image, ImageDraw
import pyrealsense2 as rs
import numpy as np
# 初始化CAN总线
# bus = can.interface.Bus(bustype='pcan', channel='PCAN_USBBUS1', bitrate='500000')
# 模型文件路径
model_path = './weights/best.pt'
classes=['surface']
# 加载YOLOv5模型
model = torch.hub.load('.', 'custom', path=model_path, source='local')
device = torch.device('cuda') if torch.cuda.is_available() else torch.device('cpu')
model = model.to(device).eval()
# 设置间隔拍照的时间间隔(单位:毫秒)
interval = 3000
# 设置置信度阈值
conf_threshold = 0.90
# # 进行目标检测
# def detect_objects(frame):
# results = model(frame)
# # 解析模型结果
# detections = results.xyxy[0].cpu().numpy()
# # 将检测结果解析为物体的像素坐标信息
# # detected_objects = []
# for detection in detections:
# # 提取每个物体的左上角和右下角坐标
# x1, y1, x2, y2, confidence, class_idx = detection
# # 保留置信度的检测结果
# if confidence >= conf_threshold:
# # 获取物体的中心坐标
# center_x = int((x1 + x2) / 2)
# center_y = int((y1 + y2) / 2)
# # 添加物体的像素坐标信息到列表中
# # detected_objects.append((center_x, center_y))
# # 在图像上绘制待抓取的目标物体
# image = frame.copy()
# # 绘制边界框
# cv2.rectangle(image, (int(x1), int(y1)), (int(x2), int(y2)), (0, 255, 0), 2)
# # 标注物体的位置
# class_name = classes[int(class_idx)]
# text = f'{class_name}: {confidence:.2f}, Center: ({center_x:.2f}, {center_y:.2f})'
# cv2.putText(image, text, (int(x1), int(y1)-10), cv2.FONT_HERSHEY_SIMPLEX, 0.5, (0, 255, 0), 1)
# # 显示图像
# cv2.imshow('Grabbed object', image)
# cv2.waitKey(interval)
# cv2.destroyWindow('Grabbed object')
# # 将像素坐标信息转换为可以发送的格式,并通过PCAN发送出去
# print ('center_x , center_y ', center_x, center_y)
# can_data = [center_x & 0xFF, (center_x >> 8) & 0xFF, center_y & 0xFF, (center_y >> 8) & 0xFF, 0, 0, 0, 0]
# print ('sending data via PCAN: ', can_data)
# msg = can.Message(arbitration_id=0x123, data=can_data)
# # 发送到数据总线
# try:
# bus.send(msg)
# print ('message sent on {}'.format(bus.channel_info))
# except can.CanError:
# print ('Failed to send data via PCAN')
# else:
# print ('no detected objects')
# pass
# # 打开摄像头
# cap = cv2.VideoCapture(1)
# # 检查摄像头是否成功打开
# if not cap.isOpened():
# print ('Error: Failed to open camera.')
# exit()
# # 循环执行
# while True:
# ret, frame = cap.read()
# # 检查是否成功读取帧
# if not ret:
# print ('Error: Failed to read frame.')
# # 进行目标检测
# detect_objects(frame)
# # cv2.waitKey(interval)
# # cv2.destroyAllWindows()
# if cv2.waitKey(1) & 0xFF == ord('q'):
# break
# cap.release()
# cv2.destroyAllWindows()
pipeline = rs.pipeline()
config = rs.config()
config.enable_stream(rs.stream.depth, 640, 480, rs.format.z16, 30)
config.enable_stream(rs.stream.color, 640, 480, rs.format.rgb8, 30)
cfg = pipeline.start(config)
align = rs.align(rs.stream.color)
def get_aligned_images():
frames = pipeline.wait_for_frames()
aligned_frames = align.process(frames)
aligned_depth_frame = aligned_frames.get_depth_frame()
aligned_color_frame = aligned_frames.get_color_frame()
# 获取相机参数
depth_intrin = aligned_depth_frame.profile.as_video_stream_profile().intrinsics
color_intrin = aligned_color_frame.profile.as_video_stream_profile().intrinsics
img_color = np.asanyarray(aligned_color_frame.get_data())
img_depth = np.asanyarray(aligned_depth_frame.get_data())
depth_colormap = cv2.applyColorMap(cv2.convertScaleAbs(img_depth, alpha=0.008), cv2.COLORMAP_JET)
return color_intrin, depth_intrin, img_color, img_depth, aligned_depth_frame
def detect_objects(frame):
results = model(frame)
# 解析模型结果
detections = results.xyxy[0].cpu().numpy()
# 将检测结果解析为物体的像素坐标信息
detected_objects = []
for detection in detections:
# 提取每个物体的左上角和右下角坐标
x1, y1, x2, y2, confidence, class_idx = detection
# 保留置信度的检测结果
if confidence >= conf_threshold:
# 获取物体的中心坐标
center_x = int((x1 + x2) / 2)
center_y = int((y1 + y2) / 2)
# 添加物体的像素坐标信息到列表中
detected_objects.append((center_x, center_y))
return detected_objects
try:
while True:
color_intrin, depthe_intrin, img_color, img_depth, aligne_depth_frame = get_aligned_images()
if not img_color.any() or not img_depth.any():
continue
depth_colormap = cv2.applyColorMap(cv2.convertScaleAbs(img_depth, alpha=0.03), cv2.COLORMAP_JET)
images = np.hstack((img_color, depth_colormap))
center_xy = detect_objects(img_color)
for i in range(len(center_xy)):
ux = center_xy[i][0]
uy = center_xy[i][1]
dis = aligne_depth_frame.get_distance(ux, uy)
print ('distance: ', dis)
# 计算相机坐标系
camera_xyz = rs.rs2_deproject_pixel_to_point(depthe_intrin, (ux, uy), dis)
# 转换成三位小数
camera_xyz = np.round(np.array(camera_xyz), 3)
camera_xyz = np.array(list(camera_xyz)) * 1000
camera_xyz = list(camera_xyz)
print ('camera_xyz: ', camera_xyz)
cv2.circle(img_color, (ux, uy), 4, (255, 255, 255), 5)
cv2.putText(img_color, str(camera_xyz), (ux+20, uy+10), 0, 1,
[225, 255, 255], thickness=1, lineType=cv2.LINE_AA)
cv2.namedWindow('detection', flags=cv2.WINDOW_NORMAL | cv2.WINDOW_KEEPRATIO | cv2.WINDOW_GUI_EXPANDED)
cv2.resizeWindow('detection', 640, 480)
cv2.imshow('detection', img_color)
cv2.waitKey(2000)
key = cv2.waitKey(1)
if key & 0xFF == ord('q') or key == 27:
cv2.destroyAllWindows()
pipeline.stop()
break
finally:
pipeline.stop()