Appendix A
Appendix A
import cv2 as cv
from playsound import playsound
import pyttsx3
import time
text_speak = pyttsx3.init()
# setting parameters
CONFIDENCE_THRESHOLD = 0.5
NMS_THRESHOLD = 0.5
yoloNet.setPreferableBackend(cv.dnn.DNN_BACKEND_CUDA)
yoloNet.setPreferableTarget(cv.dnn.DNN_TARGET_CUDA_FP16)
model = cv.dnn_DetectionModel(yoloNet)
model.setInputParams(size=(416, 416), scale=1/255, swapRB=True)
# setting camera
def ObjectDetector(image):
classes, scores, boxes = model.detect(
image, CONFIDENCE_THRESHOLD, NMS_THRESHOLD)
xii
color = COLORS[int(classid) % len(COLORS)]
label1 = "%s "% ((class_names)[classid[0]])
text_speak.say(label1)
text_speak.runAndWait()
time.sleep(0.5)
# label = "%f" % (score)
lab3 = "%s : %f "% ((class_names)[classid[0]],score)
cv.rectangle(image, box, color, 2)
cv.putText(frame,lab3, (box[0], box[1]-10), fonts, 0.5, color, 2)
camera = cv.VideoCapture(0)
counter = 0
capture = False
number = 0
while True:
ret, frame = camera.read()
orignal = frame.copy()
ObjectDetector(frame)
cv.imshow('oringal', orignal)
else:
counter = 0
cv.imshow('frame', frame)
key = cv.waitKey(1)
if key == ord('c'):
capture = True
number += 1
cv.imwrite(f'ReferenceImages/image{number}.png', orignal)
if key == ord('q'):
break
cv.destroyAllWindows()
xiii