Running DRP-AI Sample Application with USB Camera

1. Overview

This application note describes how to run the DRP-AI sample for the YOLOv2 model with the Renesas RZ/V2H EVK (RZ/V2H SoC-based embedded computer), using a USB Full HD resolution (1920x1080) camera.

This functionality is available with the Emcraft software distribution for the RZ/V2H, Release 1.0.5-42 or higher.

1.1. DRP-AI

DRP-AI is an AI accelerator developed by Renesas, designed to provide high-performance and low-power AI processing for embedded applications. It addresses the challenges of implementing AI on embedded devices, which may have limited resources or power constraints. The DRP-AI accelerator integrates pre- and post-processing with AI inference, resulting in superior AI processing performance compared to many other accelerators. The DRP-AI Translator tool maximises DRP-AI performance by enabling it to handle all operations of an AI model. The DRP-AI TVM, based on the Apache TVM ML compiler framework, supports multiple AI frameworks and allows operation in conjunction with the CPU, expanding the range of AI models that can be run. The DRP-AI Support Package provides the necessary driver and guide for operating the DRP-AI. The DRP-AI accelerator is embedded in the Renesas RZ/V2H series MPUs, delivering high-speed AI processing with high power efficiency.

Pleaser refer to the DRP-AI page on the Renesas web site for more information.

1.2. YOLO v2

YOLOv2 (You Only Look Once version 2) is a real-time object detection algorithm that aims to efficiently detect and classify objects in images or video frames. It is an improvement over the original YOLO algorithm, offering better accuracy and performance.

Please refer to original the YOLO v2 web site for more information.

Here is an example of YOLOv2 detecting objects:

wayland-screenshot-2023-09-19_17-41-26.png

In this example, YOLOv2 identifies and localises a bicycle, car, and dog within the input image. The confidence scores indicate the algorithm's level of confidence in the predictions, while the bounding boxes specify the location and size of each detected object.

1.3. Supported Classes

Here is a list of objects that the YOLOv2 sample is capable of detecting:

  • aeroplane

  • bicycle

  • bird

  • boat

  • bottle

  • bus

  • car

  • cat

  • chair

  • cow

  • diningtable

  • dog

  • horse

  • motorbike

  • person

  • pottedplant

  • sheep

  • sofa

  • train

  • tvmonitor

Detection example provided with YOLO v2 sample:

application_result_on_hdmi-20230926-120711.png

 

1.4. Logitech BRIO 4K USB Camera

While the demo supports any USB Full HD resolution (1920x1080) , Emcraft has explicitly validated with the Logitech BRIO 4K USB Camera is an Ultra HD camera manufactured by Logitech. It is designed to capture high-resolution video and provide exceptional image quality.

The Logitech BRIO 4K USB Camera can be purchased from various online retailers, including Amazon.

2. Running the DRP-AI Demo

2.1. Connecting Logitech BRIO 4K USB Camera

To connect the Logitech BRIO 4K USB Camera to the Renesas RZ/V2H kit, plug the camera into one of the USB 3.0 ports of the RZ/V2H kit.

2.2. Accessing Camera from Linux

The Logitech BRIO 4K USB Camera conforms to the USB Video Class (UVC) specification. Devices of this class are handled by the uvcvideo kernel driver, which implements the standard V4L2 API. Thus, a broad range of V4L2 compatible tools can be used for capturing video from the camera.

Depending on additional hardware connected to the Renesas board, the Logitech camera may not be the only V4L2 device present in the system. In order to access the camera the user needs to determine the correct video device file node. The below command lists all device node files registered for the Logitech camera:

user@rzv2hevkalpha:~$ v4l2-ctl --list-devices| grep -A3 Logi Logi 4K Stream Edition (usb-15860000.usb-1): /dev/video2 /dev/video3 /dev/video4

2.3. Installing Dependencies

To install the required build dependencies, run the following command:

user@rzv2hevkalpha:~$ sudo apt install libgtkd-3-dev

2.4. Modifying Source Code

Unpack the source code from the rzv2h_drpai-sample-application_ver0.90.tar.gz archive:

user@rzv2hevkalpha:~$ tar zxvf rzv2h_drpai-sample-application_ver0.90.tar.gz user@rzv2hevkalpha:~$ cd rzv2h_drpai-sample-application_ver0.90/app_yolov2_cam/src

and open the define.h file in a text editor:

Change INPUT_CAM_TYPE from 1 to 0 for the Input Camera value:

To exclude the installation of unnecessary libraries, open the CmakeList.txt file in a text editor:

and remove the following modules mmngr, gstreamer-1.0 gstbase-1.0 gstapp-1.0 mmngrbuf. These modules are not required for building this sample:

After editing the CMakeList.txt, target_link_libraries will look like this:

2.5. Building Source Code

To build the source code, create a build directory in src folder and run cmake in it:

2.6. Running the Sample

To run the sample, follow these steps:

  1. Go to rzv2h_drpai-sample-application_ver0.90/app_yolov2_cam/exe directory.

  2. Locate the app_yolov2_cam binary file.

  3. Replace the existing app_yolov2_cambinary file with the one that you have built in the previous step:

Switch user to root to access the Weston graphics:

Set the appropriate environment variables so the application can connect to Weston:

Run the sample, setting the DRP-AI frequencies as follows: DRP0 to 315MHz and AI-MAC to 315MHz:

 

By default, the application utilizes DRP-AI with the maximum operating frequency for DRP0 (420MHz) and AI-MAC (1GHz) when no arguments are provided.

For detailed instructions on setting the DRP-AI frequencies, please refer to the "Setting the DRP-AI (DRP0 and AI-MAC) frequency" section in the README.md file found in app_yolov2_cam directory of rzv2h_drpai-sample-application_ver0.90.tar.gz archive.