我正在使用person-detection-action-recognition-0005预训练模型从openvino,以检测人和他们的行动。
通过上面的链接,我编写了一个python脚本来进行检测。
这是剧本。
import cv2
def main():
print(cv2.__file__)
frame = cv2.imread('/home/naveen/Downloads/person.jpg')
actionNet = cv2.dnn.readNet('person-detection-action-recognition-0005.bin',
'person-detection-action-recognition-0005.xml')
actionBlob = cv2.dnn.blobFromImage(frame, size=(680, 400))
actionNet.setInput(actionBlob)
# detection output
actionOut = actionNet.forward(['mbox_loc1/out/conv/flat',
'mbox_main_conf/out/conv/flat/softmax/flat',
'out/anchor1','out/anchor2',
'out/anchor3','out/anchor4'])
# this is the part where I dont know how to get person bbox
# and action label for those person fro actionOut
for detection in actionOut[2].reshape(-1, 3):
print('sitting ' +str( detection[0]))
print('standing ' +str(detection[1]))
print('raising hand ' +str(detection[2]))
现在,我不知道如何从输出变量(ActionOut)获取bbox和操作标签。我找不到任何文件或博客解释这一点。
有人有什么想法或建议,怎么做呢?
https://stackoverflow.com/questions/61272635
复制相似问题