updates
This commit is contained in:
parent
4445715f4c
commit
d1b6929043
98
detect.py
98
detect.py
|
@ -11,6 +11,7 @@ def detect(save_txt=False, save_img=False, stream_img=False):
|
||||||
img_size = (320, 192) if ONNX_EXPORT else opt.img_size # (320, 192) or (416, 256) or (608, 352) for (height, width)
|
img_size = (320, 192) if ONNX_EXPORT else opt.img_size # (320, 192) or (416, 256) or (608, 352) for (height, width)
|
||||||
out, source, weights, half = opt.output, opt.source, opt.weights, opt.half
|
out, source, weights, half = opt.output, opt.source, opt.weights, opt.half
|
||||||
webcam = source == '0' or source.startswith('rtsp') or source.startswith('http')
|
webcam = source == '0' or source.startswith('rtsp') or source.startswith('http')
|
||||||
|
streams = source == 'streams.txt'
|
||||||
|
|
||||||
# Initialize
|
# Initialize
|
||||||
device = torch_utils.select_device(force_cpu=ONNX_EXPORT)
|
device = torch_utils.select_device(force_cpu=ONNX_EXPORT)
|
||||||
|
@ -47,7 +48,9 @@ def detect(save_txt=False, save_img=False, stream_img=False):
|
||||||
|
|
||||||
# Set Dataloader
|
# Set Dataloader
|
||||||
vid_path, vid_writer = None, None
|
vid_path, vid_writer = None, None
|
||||||
if webcam:
|
if streams:
|
||||||
|
dataset = LoadStreams(source, img_size=img_size, half=half)
|
||||||
|
elif webcam:
|
||||||
stream_img = True
|
stream_img = True
|
||||||
dataset = LoadWebcam(source, img_size=img_size, half=half)
|
dataset = LoadWebcam(source, img_size=img_size, half=half)
|
||||||
else:
|
else:
|
||||||
|
@ -60,56 +63,63 @@ def detect(save_txt=False, save_img=False, stream_img=False):
|
||||||
|
|
||||||
# Run inference
|
# Run inference
|
||||||
t0 = time.time()
|
t0 = time.time()
|
||||||
for path, img, im0, vid_cap in dataset:
|
for path, img, im0s, vid_cap in dataset:
|
||||||
t = time.time()
|
t = time.time()
|
||||||
save_path = str(Path(out) / Path(path).name)
|
|
||||||
|
|
||||||
# Get detections
|
# Get detections
|
||||||
img = torch.from_numpy(img).unsqueeze(0).to(device)
|
img = torch.from_numpy(img).to(device)
|
||||||
|
if img.ndimension() == 3:
|
||||||
|
img = img.unsqueeze(0)
|
||||||
pred, _ = model(img)
|
pred, _ = model(img)
|
||||||
det = non_max_suppression(pred.float(), opt.conf_thres, opt.nms_thres)[0]
|
|
||||||
|
|
||||||
s = '%gx%g ' % img.shape[2:] # print string
|
for i, det in enumerate(non_max_suppression(pred, opt.conf_thres, opt.nms_thres)): # detections per image
|
||||||
if det is not None and len(det):
|
if streams: # batch_size > 1
|
||||||
# Rescale boxes from img_size to im0 size
|
p, s, im0 = path[i], '%g: ' % i, im0s[i]
|
||||||
det[:, :4] = scale_coords(img.shape[2:], det[:, :4], im0.shape).round()
|
|
||||||
|
|
||||||
# Print results
|
|
||||||
for c in det[:, -1].unique():
|
|
||||||
n = (det[:, -1] == c).sum() # detections per class
|
|
||||||
s += '%g %ss, ' % (n, classes[int(c)]) # add to string
|
|
||||||
|
|
||||||
# Write results
|
|
||||||
for *xyxy, conf, _, cls in det:
|
|
||||||
if save_txt: # Write to file
|
|
||||||
with open(save_path + '.txt', 'a') as file:
|
|
||||||
file.write(('%g ' * 6 + '\n') % (*xyxy, cls, conf))
|
|
||||||
|
|
||||||
if save_img or stream_img: # Add bbox to image
|
|
||||||
label = '%s %.2f' % (classes[int(cls)], conf)
|
|
||||||
plot_one_box(xyxy, im0, label=label, color=colors[int(cls)])
|
|
||||||
|
|
||||||
print('%sDone. (%.3fs)' % (s, time.time() - t))
|
|
||||||
|
|
||||||
# Stream results
|
|
||||||
if stream_img:
|
|
||||||
cv2.imshow(weights, im0)
|
|
||||||
|
|
||||||
# Save results (image with detections)
|
|
||||||
if save_img:
|
|
||||||
if dataset.mode == 'images':
|
|
||||||
cv2.imwrite(save_path, im0)
|
|
||||||
else:
|
else:
|
||||||
if vid_path != save_path: # new video
|
p, s, im0 = path, '', im0s
|
||||||
vid_path = save_path
|
|
||||||
if isinstance(vid_writer, cv2.VideoWriter):
|
|
||||||
vid_writer.release() # release previous video writer
|
|
||||||
|
|
||||||
fps = vid_cap.get(cv2.CAP_PROP_FPS)
|
save_path = str(Path(out) / Path(p).name)
|
||||||
width = int(vid_cap.get(cv2.CAP_PROP_FRAME_WIDTH))
|
s += '%gx%g ' % img.shape[2:] # print string
|
||||||
height = int(vid_cap.get(cv2.CAP_PROP_FRAME_HEIGHT))
|
if det is not None and len(det):
|
||||||
vid_writer = cv2.VideoWriter(save_path, cv2.VideoWriter_fourcc(*opt.fourcc), fps, (width, height))
|
# Rescale boxes from img_size to im0 size
|
||||||
vid_writer.write(im0)
|
det[:, :4] = scale_coords(img.shape[2:], det[:, :4], im0.shape).round()
|
||||||
|
|
||||||
|
# Print results
|
||||||
|
for c in det[:, -1].unique():
|
||||||
|
n = (det[:, -1] == c).sum() # detections per class
|
||||||
|
s += '%g %ss, ' % (n, classes[int(c)]) # add to string
|
||||||
|
|
||||||
|
# Write results
|
||||||
|
for *xyxy, conf, _, cls in det:
|
||||||
|
if save_txt: # Write to file
|
||||||
|
with open(save_path + '.txt', 'a') as file:
|
||||||
|
file.write(('%g ' * 6 + '\n') % (*xyxy, cls, conf))
|
||||||
|
|
||||||
|
if save_img or stream_img: # Add bbox to image
|
||||||
|
label = '%s %.2f' % (classes[int(cls)], conf)
|
||||||
|
plot_one_box(xyxy, im0, label=label, color=colors[int(cls)])
|
||||||
|
|
||||||
|
print('%sDone. (%.3fs)' % (s, time.time() - t))
|
||||||
|
|
||||||
|
# Stream results
|
||||||
|
if stream_img:
|
||||||
|
cv2.imshow(p, im0)
|
||||||
|
|
||||||
|
# Save results (image with detections)
|
||||||
|
if save_img:
|
||||||
|
if dataset.mode == 'images':
|
||||||
|
cv2.imwrite(save_path, im0)
|
||||||
|
else:
|
||||||
|
if vid_path != save_path: # new video
|
||||||
|
vid_path = save_path
|
||||||
|
if isinstance(vid_writer, cv2.VideoWriter):
|
||||||
|
vid_writer.release() # release previous video writer
|
||||||
|
|
||||||
|
fps = vid_cap.get(cv2.CAP_PROP_FPS)
|
||||||
|
w = int(vid_cap.get(cv2.CAP_PROP_FRAME_WIDTH))
|
||||||
|
h = int(vid_cap.get(cv2.CAP_PROP_FRAME_HEIGHT))
|
||||||
|
vid_writer = cv2.VideoWriter(save_path, cv2.VideoWriter_fourcc(*opt.fourcc), fps, (w, h))
|
||||||
|
vid_writer.write(im0)
|
||||||
|
|
||||||
if save_txt or save_img:
|
if save_txt or save_img:
|
||||||
print('Results saved to %s' % os.getcwd() + os.sep + out)
|
print('Results saved to %s' % os.getcwd() + os.sep + out)
|
||||||
|
|
|
@ -3,7 +3,9 @@ import math
|
||||||
import os
|
import os
|
||||||
import random
|
import random
|
||||||
import shutil
|
import shutil
|
||||||
|
import time
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
from threading import Thread
|
||||||
|
|
||||||
import cv2
|
import cv2
|
||||||
import numpy as np
|
import numpy as np
|
||||||
|
@ -183,6 +185,64 @@ class LoadWebcam: # for inference
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|
||||||
|
class LoadStreams: # multiple IP or RTSP cameras
|
||||||
|
def __init__(self, path='streams.txt', img_size=416, half=False):
|
||||||
|
self.img_size = img_size
|
||||||
|
self.half = half # half precision fp16 images
|
||||||
|
with open(path, 'r') as f:
|
||||||
|
sources = f.read().splitlines()
|
||||||
|
|
||||||
|
n = len(sources)
|
||||||
|
self.imgs = [None] * n
|
||||||
|
self.sources = sources
|
||||||
|
for i, s in enumerate(sources):
|
||||||
|
# Start the thread to read frames from the video stream
|
||||||
|
cap = cv2.VideoCapture(0 if s == '0' else s)
|
||||||
|
fps = cap.get(cv2.CAP_PROP_FPS) % 100
|
||||||
|
width = int(cap.get(cv2.CAP_PROP_FRAME_WIDTH))
|
||||||
|
height = int(cap.get(cv2.CAP_PROP_FRAME_HEIGHT))
|
||||||
|
|
||||||
|
print('%g/%g: %gx%g at %.2f FPS %s...' % (i + 1, n, width, height, fps, s))
|
||||||
|
thread = Thread(target=self.update, args=([i, cap]))
|
||||||
|
thread.daemon = True
|
||||||
|
thread.start()
|
||||||
|
print('') # newline
|
||||||
|
time.sleep(0.5)
|
||||||
|
|
||||||
|
def update(self, index, cap):
|
||||||
|
# Read next stream frame in a daemon thread
|
||||||
|
while cap.isOpened():
|
||||||
|
_, self.imgs[index] = cap.read()
|
||||||
|
time.sleep(0.030) # 33.3 FPS to keep buffer empty
|
||||||
|
|
||||||
|
def __iter__(self):
|
||||||
|
self.count = -1
|
||||||
|
return self
|
||||||
|
|
||||||
|
def __next__(self):
|
||||||
|
self.count += 1
|
||||||
|
img0 = self.imgs.copy()
|
||||||
|
if cv2.waitKey(1) == ord('q'): # 'q' to quit
|
||||||
|
cv2.destroyAllWindows()
|
||||||
|
raise StopIteration
|
||||||
|
|
||||||
|
# Letterbox
|
||||||
|
img = [letterbox(x, new_shape=self.img_size, mode='square')[0] for x in img0]
|
||||||
|
|
||||||
|
# Stack
|
||||||
|
img = np.stack(img, 0)
|
||||||
|
|
||||||
|
# Normalize RGB
|
||||||
|
img = img[:, :, :, ::-1].transpose(0, 3, 1, 2) # BGR to RGB
|
||||||
|
img = np.ascontiguousarray(img, dtype=np.float16 if self.half else np.float32) # uint8 to fp16/fp32
|
||||||
|
img /= 255.0 # 0 - 255 to 0.0 - 1.0
|
||||||
|
|
||||||
|
return self.sources, img, img0, None
|
||||||
|
|
||||||
|
def __len__(self):
|
||||||
|
return 0 # 1E12 frames = 32 streams at 30 FPS for 30 years
|
||||||
|
|
||||||
|
|
||||||
class LoadImagesAndLabels(Dataset): # for training/testing
|
class LoadImagesAndLabels(Dataset): # for training/testing
|
||||||
def __init__(self, path, img_size=416, batch_size=16, augment=False, hyp=None, rect=True, image_weights=False,
|
def __init__(self, path, img_size=416, batch_size=16, augment=False, hyp=None, rect=True, image_weights=False,
|
||||||
cache_images=False):
|
cache_images=False):
|
||||||
|
|
Loading…
Reference in New Issue