Web Analytics

Getting Started

1. Environment Setup

There are several ways to get the Kneron host example:

Note:
Kneron Host_lib also supproted in Windows 10, Mingw64 environment
See README_*.md in host_lib for details

2. File Structure

host_lib supports Kneron Chips, including both KL520 and KL720.
In this document file, KL520 examples in Linux are shown.

host_lib-|
         |->app_binaries-|        (*KL520 app binaries for example)
                         |-> *KL520/tiny_yolo_v3      (FW for object detection)
                         |-> *KL520/ssd_fd_lm         (FW for mask detection)
                         |-> *KL520/ota/ready_to_load (default path for ota)
                         |-> *KL520/dfw/minions.bin   (bin file for UART boot)
                         |-> ...
         |-> common               (common shared file between host and KL520)
         |-> dll                  (dll files for Windows MINGW64)
         |-> docs                 (images in README)
         |-> example/*KL520       (host side C++ examples)
         |-> python               (host side Python examples)
         |-> src                  (source files for host lib, which communicate with KL520)
         |-> input_images         (test image binary)
         |-> input_models/*KL520  (models for examples)
         |-> CMakeList.txt        (top level cmake list)
         |-> README_CPP.md        (simple instruction to build for C++ examples)
         |-> README_Python.md     (simple instruction for Python examples)
         |-> version.txt          (version information)


*KL520 stands for KL520 examples/models
 KL720 stands for KL720 examples/models

3. Compile and Build

This project use cmake to build the code. There is top level CMakeLists.txt, and src folder and example/KL520 folder have their own CMakeLists.txt.

Use Cmake to build:

# in linux
mkdir build
cd build
cmake ../
# for cmake, or append '-DBUILD_OPENCV_EX=on ..' to build with opencv applicaitons
make -j

And then you will see the bin directory, and all the example executables are under here.

4. Run the Default Application Tiny Yolo v3

Tiny Yolo v3 is an object detection algorithm, and it can classify 80 different classes.

Kneron device come with build-in binaries that can run 608x608 RGB565 image in Tiny Yolo v3 object detection.

4.1 C++ Example

Let’s take a look at the C++ example program, kl520_isi_async_parallel_yolo.cpp. In this test, we send two different images cars_street_at_night_608x608_rgb565.bin and car_park_barrier_608x608_rgb565.bin (two RGB565 binaries) into KL520, and get the detection results back.

#define ISI_IMAGE_FILE      (HOST_LIB_DIR "/input_images/cars_street_at_night_608x608_rgb565.bin")
#define ISI_IMAGE_FILE_T    (HOST_LIB_DIR "/input_images/car_park_barrier_608x608_rgb565.bin")
#define IMG_SOURCE_W        608
#define IMG_SOURCE_H        608
#define ISI_IMG_SIZE        (IMG_SOURCE_W * IMG_SOURCE_H * 2)
#define ISI_APP_ID          APP_TINY_YOLO3
// isi initial result memory size in bytes
#define ISI_RESULT_SIZE     2048

After we setup the images and image buffer size, the program uses kdp_start_isi_mode to notify the communication and inference setup.

uint16_t width  = IMG_SOURCE_W;
uint16_t height = IMG_SOURCE_H;
// image format flags
// IMAGE_FORMAT_SUB128: subtract 128 for R/G/B value in Kneron device
// NPU_FORMAT_RGB565: image input is RGB565
// IMAGE_FORMAT_PARALLEL_PROC: pipeline between ncpu and npu
uint32_t format = IMAGE_FORMAT_SUB128 | NPU_FORMAT_RGB565 | IMAGE_FORMAT_PARALLEL_PROC;

// Flash the firmware code with companion mode for tiny_yolo_v3 !!!
printf("starting ISI mode ...\n");
// call kdp_start_isi_mode to start isi mode        
int ret = kdp_start_isi_mode(
    dev_idx, ISI_APP_ID, ISI_RESULT_SIZE, width, height, format, &error_code, &image_buf_size);
if (ret != 0) {
    printf("could not set to ISI mode: %d ..\n", ret);
    return -1;
}

if (image_buf_size < 3) {
    printf("ISI mode window %d too small...\n", image_buf_size);
    return -1;
}

printf("ISI mode succeeded (window = %d)...\n", image_buf_size);
usleep(SLEEP_TIME);

Setup image buffer to load the image binaries.

int n_len = read_file_to_buf(img_buf1, ISI_IMAGE_FILE, ISI_IMG_SIZE);
if (n_len <= 0) {
    printf("reading image file 1 failed:%d...\n", n_len);
    return -1;
}

n_len = read_file_to_buf(img_buf2, ISI_IMAGE_FILE_T, ISI_IMG_SIZE);
if (n_len <= 0) {
    printf("reading image file 2 failed:%d...\n", n_len);
    return -1;
}

Since the KL520 application design uses image buffer to pipeline the image transfer and npu processing, we can fill up the image buffer first. Therefore, we can use kdp_isi_inference to send images to KL520, without getting the result back yet.

printf("starting ISI inference ...\n");
uint32_t img_id_tx = 1234;
uint32_t img_id_rx = img_id_tx;
uint32_t img_left = 12;
uint32_t result_size = 0;
uint32_t buf_len = ISI_IMG_SIZE;
char inf_res[ISI_RESULT_SIZE];

// start time for the first frame
double start_time = what_time_is_it_now();

// Send 2 images first
// do inference for each input
ret = do_inference(dev_idx, img_buf1, buf_len, img_id_tx, &error_code, &img_left);
if (ret)
    return ret;
img_id_tx++;
// do inference for each input
ret = do_inference(dev_idx, img_buf2, buf_len, img_id_tx, &error_code, &img_left);
if (ret)
    return ret;
img_id_tx++;

// Send the rest and get result in loop, with 2 images alternatively
uint32_t loop = 0;
if (test_loop > 3)
    loop = test_loop - 2;

while (loop && !check_ctl_break()) {
    // do inference for each input
    ret = do_inference(dev_idx, img_buf1, buf_len, img_id_tx, &error_code, &img_left);
    if (ret)
        return ret;
    img_id_tx++;

Then the program must wait and get back the result then send the next image. It uses kdp_isi_retrieve_res to get the result. After that, it uses kdp_isi_inference to send another images to KL520 to process. The loop is set to 300 to estimate the average FPS of 300 images.

    ret = do_get_result(dev_idx, img_id_rx, &error_code, &result_size, inf_res);
    if (ret)
        return ret;
    img_id_rx++;

    loop--;
    // Odd loop case
    if (loop == 0)
        break;
    // do inference for each input
    ret = do_inference(dev_idx, img_buf2, buf_len, img_id_tx, &error_code, &img_left);
    if (ret)
        return ret;
    img_id_tx++;

    // retrieve the detection results for each input
    ret = do_get_result(dev_idx, img_id_rx, &error_code, &result_size, inf_res);
    if (ret)
        return ret;
    img_id_rx++;

    loop--;
}

Then, the program needs to retrieve all the remaining image results, using kdp_isi_retrieve_res.

// Get last 2 results
// retrieve the detection results for each input
ret = do_get_result(dev_idx, img_id_rx, &error_code, &result_size, inf_res);
if (ret)
    return ret;
img_id_rx++;
// retrieve the detection results for each input
ret = do_get_result(dev_idx, img_id_rx, &error_code, &result_size, inf_res);
if (ret)
    return ret;
img_id_rx++;
// calculate the FPS
if (1) {
    double end_time = what_time_is_it_now();
    double elapsed_time, avg_elapsed_time, avg_fps;

    elapsed_time = (end_time - start_time) * 1000;
    avg_elapsed_time = elapsed_time / test_loop;
    avg_fps = 1000.0f / avg_elapsed_time;

    printf("\n=> Avg %.2f FPS (%.2f ms = %.2f/%d)\n\n",
        avg_fps, avg_elapsed_time, elapsed_time, test_loop);
}

Lastly, the program calls kdp_end_isi to end isi mode.

// call kdp_end_isi to end isi mode after finishing the inference for all frames
kdp_end_isi(dev_idx);

Run the executable binaries, and we can see initialization messages, and the buffer depth is 3. Image index is keep increasing and the expected 5 and 3 objects are toggling because we ping pong transfer two images, so the results are ping pong as 5 objects and 3 objects as well

Kneron@Ubuntu:~/host_lib/build/bin$ ./kl520_isi_async_parallel_yolo
init kdp host lib log....
adding devices....
start kdp host lib....
doing test :0....
starting ISI mode ...
ISI mode succeeded (window = 3)...
starting ISI inference ...
image 1234 -> 5 object(s)
image 1235 -> 3 object(s)
image 1236 -> 5 object(s)
image 1237 -> 3 object(s)

Lastly, we can see the average FPS for running 300 images is 11.7, each image take average 85ms to finish.

image 1532 -> 5 object(s)
image 1533 -> 3 object(s)

=> Avg 11.70 FPS (85.46 ms = 25637.14/300)

de init kdp host lib....

4.2. Python Example

Note:
visit Kneron Github for more python materials

Let’s take a look at the Python example program (path=python/examples_kl520), cam_isi_async_parallel_yolo.py. In this test, we send frames (RGB565) of camera into KL520, and get the detection results back.

Setup video capture device for the image width and height

"""User test cam yolo."""
image_source_h = 480
image_source_w = 640
app_id = constants.APP_TINY_YOLO3
image_size = image_source_w * image_source_h * 2
frames = []

# Setup video capture device.
capture = kdp_wrapper.setup_capture(0, image_source_w, image_source_h)
if capture is None:
    return -1

After we setup the video capture device, the program uses start_isi_parallel to notify the communication and inference setup.

# Start ISI mode.
if kdp_wrapper.start_isi_parallel_ext(dev_idx, app_id, image_source_w, image_source_h):
    return -1

Since the KL520 application design use image buffer to pipeline the image transfer and npu processing, we can fill up the image buffer first. Therefore, we can use fill_buffer to send frames to KL520, without getting the result back yet.

start_time = time.time()
# Fill up the image buffers.
ret, img_id_tx, img_left, buffer_depth = kdp_wrapper.fill_buffer(
    dev_idx, capture, image_size, frames)
if ret:
    return -1

Then the program must wait and get back the result then send the next image. It uses pipeline_inference to get the result and to send another images to KL520 to process. The loop is set to 1000 to estimate the average FPS of 1000 images.

# Send the rest and get result in loop, with 2 images alternatively
print("Companion image buffer depth = ", buffer_depth)
kdp_wrapper.pipeline_inference(
    dev_idx, app_id, test_loop - buffer_depth, image_size,
    capture, img_id_tx, img_left, buffer_depth, frames, handle_result)

end_time = time.time()
diff = end_time - start_time 
estimate_runtime = float(diff/test_loop)
fps = float(1/estimate_runtime)    
print("Parallel inference average estimate runtime is ", estimate_runtime)
print("Average FPS is ", fps)

Lastly, the program calls kdp_end_isi to end isi mode.

kdp_wrapper.kdp_exit_isi(dev_idx)

Run the Python example, and we can see initialization messages, and the buffer depth is 3. Image index keeps increasing and the number of detected objects are outputted

$ python3 main.py -t cam_isi_aync_parallel_yolo
adding devices...
start kdp host lib....
Task:    cam_isi_aysnc_parallel_yolo
starting ISI mode...
ISI mode succeeded (window = 3)...
starting ISI inference ...
Companion image buffer depth =  3
image 1234 -> 1 object(s)
image 1235 -> 1 object(s)
image 1236 -> 1 object(s)
image 1237 -> 2 object(s)

Lastly, we can see the average FPS for running 1000 images is 12.2, each image take average 82ms to finish.

image 2233 -> 2 object(s)

Parallel inference average estimate runtime is  0.08506509852409362
Average FPS is  11.755702601305547
de init kdp host lib....

5. Run OTA to Swap Another Pre-build Application Binary Mask Face Detection

Besides Tiny Yolo v3, Kneron also provides the following for OTA usage example

Kneron@ubuntu:~/host_lib/app_binaries/KL520$ ls
dfw/  ota/  readme.txt  ssd_fd_lm/  tiny_yolo_v3/

In order to swap the KL520 application, Kneron provides update application feature to update the firmware via USB.

Here is an introduction how to do it.

First, user can copy the target application into app_binaries/KL520/ota/ready_to_load. Here we will load the mask fd application into KL520. As you can see, there are 3 files,

Kneron@ubuntu:~/host_lib/app_binaries/KL520/ota/ready_to_load$ cp ../../ssd_fd_lm/*.bin .
Kneron@ubuntu:~/host_lib/app_binaries/KL520/ota/ready_to_load$ cp ../../../../input_models/KL520/ssd_fd_lm/models_520.nef .
Kneron@ubuntu:~/host_lib/app_binaries/KL520/ota/ready_to_load$ ls
fw_ncpu.bin  fw_scpu.bin  models_520.nef

Then we can go to build/bin, and run ./kl520_update_app_nef_model. This will load all 3 binaries into KL520, and program them into the flash so that even without power, the KL520 can still maintain the applications. Since the model are relatively large and flash programming is slow, users need to wait for couple mins to update the application. It takes about 3-5 minutes here.

Kneron@ubuntu:~/host_lib/build/bin$ ./update_app_nef_model
init kdp host lib log....
adding devices....
start kdp host lib....
doing test :0....
starting update fw ...
update SCPU firmware successed...
update NCPU firmware successed...
starting update model :1...
update model succeeded...
starting report sys status ...
report sys status succeeded...

FW firmware_id 01040000 build_id 00000000
de init kdp host lib....

As you can see the log, the SCPU and NCPU firmware are updated successfully, and model update successfully. Last, it will print out the SCPU firmware id and NCPU firmware id to ensure it is updated with the correct version of code.

After KL520 is updated with the mask face detection, we can start to run this corresponding example. We are going to try the interactive with camera example, therefore, we will make sure we build the apps require opencv. Make sure to do cmake -DBUILD_OPENCV_EX=on and make -j again to make the cam application.

Kneron@ubuntu:~/host_lib/build/bin$ ./k520_cam_isi_async_ssd_fd

A Camera output window will pop out and use light blue box to get the human face.

And when you put on the mask, there will be a green bounding box around the human face. After 1000 inference, the program will stop.

6. Build New Model Binary Based on MobileNet V2 Image Classification

Now, let's try to use Kneron tool chain to build a model binary for a public model.

In this example, we pick the MobileNet V2 from Keras application. This is a model that doing Image classification, which is 1000 classes from ImageNet. we can get a public MobileNetV2 keras model by executing Python code on Keras 2.2.4.

from keras.applications.mobilenet_v2 import MobileNetV2

model = MobileNetV2(include_top=True, weights='imagenet')
model.save('MobileNetV2.h5')

6.1. Model Conversion

After getting the keras model, copy the model and a folder of jpg/png images which fall into the categories of ImageNet to /data1 of vm. The recommended size of images is more than 100, which the tools will use it for quantization.

Kneron@ubuntu:~/data1$ ls
images/ MobileNetV2.h5

Run toolchain in vm and map ~/data1 folder of vm into /data1 of toolchain

$ sudo docker run -it --rm -v ~/data1:/data1 kneron/toolchain:linux_command_toolchain

After that, we will enter toolchain docker container, and it is at workspace/. Check if the /data1 is mapped with the external folder successfully or not.

root@6093d5502017:/workspace# ls
README.md   libs        onnx-keras      requirements.txt        tmp
examples    onnx-caffe  onnx-tensorflow scripts
root@6093d5502017:/workspace# cd /data1
root@6093d5502017:/data1# ls
images/ MobileNetV2.h5

Then we can run toolchain in vm to convert the keras model into onnx model. The command is:

$ python /workspace/onnx-keras/generate_onnx.py -o /data1/MobileNetV2.h5.onnx /data1/keras/MobileNetV2.h5 -O --duplicate-shared-weights

Then we can see that a MobileNetV2 ONNX model is created under /data1

root@6093d5502017:/data1# ls
images/     MobileNetV2.h5   MobileNetV2.h5.onnx

6.2. Model editing (remove softmax at the end)

When we check the MobileNetV2 ONNX model with Netron, we can see that the network's final output layer is a softmax layer, which cannot be handled by KL520 NPU. It is very common to see the softmax layer at the end of classification network, but it is not computation extensive layer, and we can move this softmax layer into network's post process.

Toolchain provides the Python script (onnx2onnx.py) to optimize the onnx model, and the script (editor.py) to cut layers starting from a specific layer. To remove the softmax layer, we can just simply run the onnx2onnx.py as follow:

$ python /workspace/scripts/onnx2onnx2.py /data1/MobileNetV2.h5.onnx -o /data1/MobileNetV2_opt.h5.onnx

After running onnx2onnx.py script, the optimized model MobileNetV2_opt.h5.onnx is saved in /data1. The final layer of the optimized onnx model is Gemm layer now.

root@6093d5502017:/data1# ls
images/     MobileNetV2.h5   MobileNetV2.h5.onnx   MobileNetV2_opt.h5.onnx  

6.3. Model Compile Flow (compile to .nef file)

Copy the /workspace/examples/batch_compile_input_params.json into /data1 and modify it before batch-compiling MobileNetV2.

root@6093d5502017:/data1# cp /workspace/examples/batch_compile_input_params.json .

root@6093d5502017:/data1# ls
images/                  MobileNetV2.h5                   MobileNetV2.h5.onnx   
MobileNetV2_opt.h5.onnx  batch_compile_input_params.json

The batch_compile_input_params.json is modified as:

specify the input_image_folder to be data1/images, this give the image folder path for the tool to do fixed point analysis.

specify the model input width and height to be "224" and "224", user needs to modify this to fit thier models input size.

tensorflow for public MobileNetV2 as img_preprocess_method. this implies using tensorflow default preprocess, which is X/127. 5 -1

specify the model "input_onnx_file" to be /data1/MobileNetV2_opt.h5.onnx, which is the model we just edited.

Modify the model_id to be 1000 for MobileNetV2.

{
    "input_image_folder": ["/data1/images"],
    "img_channel": ["RGB"],
    "model_input_width": [224],
    "model_input_height": [224],
    "img_preprocess_method": ["tensorflow"],
    "input_onnx_file": ["/data1/MobileNetV2_opt.h5.onnx"],
    "keep_aspect_ratio": ["False"],
    "command_addr": "0x30000000",
    "weight_addr": "0x40000000",
    "sram_addr": "0x50000000",
    "dram_addr": "0x60000000",
    "whether_encryption": "No",
    "encryption_key": "0x12345678",
    "model_id_list": [1000],
    "model_version_list": [1],
    "add_norm_list": ["False"],
    "dedicated_output_buffer": "True"
}

Execute the command to batch-compile MobileNetV2 model.

$ cd /workspace/scripts && ./fpAnalyserBatchCompiler.sh.x       

After batch-compilation, a new batch_compile folder with .nef file is present in /data1. This file will be used for running the model in KL520 later.

6.4. Estimated NPU Run Time for Model

We can use Toolchain to get the evaluation result of NPU performance for Kneron device. The result does not include the time to do pre/post process and cpu node process. Firstly copy the input_params.json under /workspace/examples to /data1. Same as doing batch compile

Modify the input_params.json as: tensorflow for public MobileNetV2 as img_preprocess_method, False for add_norm.

{
    "input_image_folder": ["/data1/images"],
    "img_channel": ["RGB"],
    "model_input_width": [224],
    "model_input_height": [224],
    "img_preprocess_method": ["tensorflow"],
    "input_onnx_file": ["/data1/MobileNetV2_opt.h5.onnx"],
    "keep_aspect_ratio": ["False"],
    "command_addr": "0x30000000",
    "weight_addr": "0x40000000",
    "sram_addr": "0x50000000",
    "dram_addr": "0x60000000",
    "whether_encryption": "No",
    "encryption_key": "0x12345678",
    "simulator_img_file": "/data1/images/n01503061_3560_bird.jpg",
    "emulator_img_folder": "/data1/images",
    "cmd_bin": "/data1/compiler/command.bin",
    "weight_bin": "/data1/compiler/weight.bin",
    "setup.bin": "/data1/compiler/setup.bin",
    "whether_npu_preprocess": false,
    "raw_img_fmt": "IMG",
    "radix": 8,
    "pad_mode": 1,
    "roate":0,
    "pCrop": {
        "crop_x": 0,
        "crop_y": 0,
        "crop_w": 0,
        "crop_h": 0,
        "bCropFirstly": false
    },
    "imgSize":{
        "width": 640,
        "height": 480
    },
    "add_norm": "False"
}

Execute the command to run evaluation for MobileNetV2 model.

$ cd /workspace/scripts && ./fpAnalyserCompilerIpevaluator.sh.x

After the evaluation process is finished, we can get the evaluation_result.txt under /data1/compiler.

root@92ace5540942:/data1# cd complier/
root@92ace5540942:/data1/complier# ls
command.bin  evaluation_result.txt  setup.bin  weight.bin

The evaluation result does not count in the time to do pre/post process and cpu node process. For this MobileNetV2 model, it will take around 15.7ms (total time)

[Evaluation Result]
estimate FPS float = 63.6315
total time = 15.7105 ms
total theoretical covolution time = 3.81611 ms
average DRAM bandwidth = 0.386193 GB/s
MAC efficiency to total time = 24.2901 %
MAC idle time = 2.3849 ms
MAC running time = 13.3256 ms

7. Run Model Binaries in DME Mode

Please refer to the example code in example/KL520/kl520_dme_async_mobilenet_classification

7.1. DME mode Introduction

In DME mode, the test images, model binaries, and configuration are dynamically sent from host to Kneron device via USB, and the detection results are dynamically retrieved back from Kneron device to host via USB.

In host side, 6 APIs are used for DME.

API Description Note
kdp_start_dme_ext Send model, models_520.nef, to Kneron device Call once
kdp_dme_configure Send DME configuration to Kneron device Call once
kdp_dme_inference Send image data to Kneron device and start inference Call multiple times to send image for inference
kdp_dme_get_status Poll the completed status from Kneron device Call multiple times after kdp_dme_inference (only in DME async mode)
kdp_dme_retrieve_res Retrieve the inference result (fix-point data) back to host Call multiple times after kdp_dme_get_status
kdp_end_dme End DME mode by sending command to Kneron device Call once

7.2. DME Mode Pre/Post Process

Kneron provides default pre/post process function in firmware code if it is run in DME mode.

Pre Process

The default preprocess function finish the tasks:

Reformat: Transfer original format, such as RGB565 or YUV422, to RGBA8888

Resize: Resize image size to model input size

Subtract: Subtract 128 for all data if configured

Right-shift for 1 bit if configured

Because this MobileNetV2 model needs to apply tensorflow preprocess, which is X/127.5 -1, it is similar to apply DME default pre process. In section 7.3, we will see how to config the pre process for this MobileNetV2 model.

Post Process

The default postprocess function in DME mode would send back the original output from KL520 NPU, and user needs to implement post process functions in order to get the correct result.

finish the tasks:

  1. KL520 sends the fix-point data of all output nodes to one with the sequence of total_out_number + (c/h/w/radix/scale) + (c/h/w/radix/scale) + ... + fp_data + fp_data + ...

    With this data format, host needs to converts it back to a shared structure, struct kdp_image_s, that can easily identify each output nodes parameters, such as channel, height, width, radix, and scales.

    // Prepare for postprocessing
    int output_num = inf_res[0];
    struct output_node_params *p_node_info;
    int r_len, offset;
    struct imagenet_result_s *det_res = (struct imagenet_result_s *)calloc(IMAGENET_TOP_MAX, sizeof(imagenet_result_s));
    
    struct kdp_image_s *image_p = (struct kdp_image_s *)calloc(1, sizeof(struct kdp_image_s));
    offset = sizeof(int) + output_num * sizeof(output_node_params);
    
    // Struct to pass the parameters
    RAW_INPUT_COL(image_p) = post_par.raw_input_col;
    RAW_INPUT_ROW(image_p) = post_par.raw_input_row;
    DIM_INPUT_COL(image_p) = post_par.model_input_row;
    DIM_INPUT_ROW(image_p) = post_par.model_input_row;
    RAW_FORMAT(image_p) = post_par.image_format;
    POSTPROC_RESULT_MEM_ADDR(image_p) = (uint32_t *)det_res;
    POSTPROC_OUTPUT_NUM(image_p) = output_num;
    
    for (int i = 0; i < output_num; i++) {
        if (check_ctl_break())
            return;
        p_node_info = (struct output_node_params *)(inf_res + sizeof(int) + i * sizeof(output_node_params));
        r_len = p_node_info->channel * p_node_info->height * round_up(p_node_info->width);
    
        POSTPROC_OUT_NODE_ADDR(image_p, i) = inf_res + offset;
        POSTPROC_OUT_NODE_ROW(image_p, i) = p_node_info->height;
        POSTPROC_OUT_NODE_CH(image_p, i) = p_node_info->channel;
        POSTPROC_OUT_NODE_COL(image_p, i) = p_node_info->width;
        POSTPROC_OUT_NODE_RADIX(image_p, i) = p_node_info->radix;
        POSTPROC_OUT_NODE_SCALE(image_p, i) = p_node_info->scale;
    
        offset = offset + r_len;
    }
  2. Host will call actual model post process function post_imgnet_classificaiton() in example/post_processing_ex.c

    int post_imgnet_classification(int model_id, struct kdp_image_s *image_p)
    {
       struct imagenet_post_globals_s *gp = &u_globals.imgnet;
       uint8_t *result_p;
       int i, len, data_size, div;
       float scale;
    
       data_size = (POSTPROC_OUTPUT_FORMAT(image_p) & BIT(0)) + 1;     /* 1 or 2 in bytes */
    
       int8_t *src_p = (int8_t *)POSTPROC_OUT_NODE_ADDR(image_p, 0);
       int grid_w = POSTPROC_OUT_NODE_COL(image_p, 0);
       len = grid_w * data_size;
       int grid_w_bytes_aligned = round_up(len);
       int w_bytes_to_skip = grid_w_bytes_aligned - len;
       len = grid_w_bytes_aligned;
    
       int ch = POSTPROC_OUT_NODE_CH(image_p, 0);
    
       /* Convert to float */
       scale = POSTPROC_OUT_NODE_SCALE(image_p, 0);
       div = 1 << POSTPROC_OUT_NODE_RADIX(image_p, 0);
       for (i = 0; i < ch; i++) {
           gp->temp[i].index = i;
           gp->temp[i].score = (float)*src_p;
           gp->temp[i].score = do_div_scale(gp->temp[i].score, div, scale);
           src_p += data_size + w_bytes_to_skip;
       }
    
       softmax(gp->temp, ch);
       qsort(gp->temp, ch, sizeof(struct imagenet_result_s), inet_comparator);
    
       result_p = (uint8_t *)(POSTPROC_RESULT_MEM_ADDR(image_p));
       len = sizeof(struct imagenet_result_s) * IMAGENET_TOP_MAX;
       memcpy(result_p, gp->temp, len);
       return len;
    }

    In KL520, the results are in 16-byte aligned format for each row, and these values are in fixed point. As a result, host need to convert these values back to floating by using the radix and scale value for each output node. do_div_scale() is for this purpose.

    static float do_div_scale(float v, int div, float scale)
    {
        return ((v / div) / scale);
    }

    After converting all the output values back to float, host post process program need to add back the softmax layer, which was cut in model editor (onnx2onnx.py). Lastly, host use qsort to find the top N probability from the 1000 classes.

    softmax(gp->temp, ch);
    qsort(gp->temp, ch, sizeof(struct imagenet_result_s), inet_comparator);

7.3. How to config DME based on the input images

After getting the parameters of input images and models, we can set DME configuration and postprocessing configuration as shown in the example code of kl520_ dme_async_mobilenet_classification.cpp.

The following settings mean that the host would send 640x480 RGB565 image to KL520, and KL520 will resize it to 224x224 for model input. Then KL520 will send back raw NPU output back to Host, which suggests host side need to perform post process in order to get the results. Please note that when we compile the model in section 6, we assign model id 1000 to this MobileNetV2 model, and we need to pass this model id in the dme config as well.

// parameters for postprocessing
post_par.raw_input_col   = 640;
post_par.raw_input_row   = 480;
post_par.model_input_row = 224;
post_par.model_input_col = 224;
post_par.image_format    = IMAGE_FORMAT_SUB128  | 
                           NPU_FORMAT_RGB565    | 
                           IMAGE_FORMAT_RAW_OUTPUT  | 
                           IMAGE_FORMAT_CHANGE_ASPECT_RATIO;

// dme configuration
dme_cfg.model_id     = IMAGENET_CLASSIFICATION_MOBILENET_V2_224_224_3;// model id when compiling in toolchain
dme_cfg.output_num   = 1;                             // number of output node for the model
dme_cfg.image_col    = post_par.raw_input_col;
dme_cfg.image_row    = post_par.raw_input_row;
dme_cfg.image_ch     = 3;
dme_cfg.image_format = post_par.image_format;

Then we can call API kdp_dme_configure() to config DME.

dat_size = sizeof(struct kdp_dme_cfg_s);
printf("starting DME configure ...\n");
int ret = kdp_dme_configure(dev_idx, (char *)&dme_cfg, dat_size, &model_id);
if (ret != 0) {
    printf("could not set to DME configure mode..\n");
    return -1;
}
printf("DME configure model [%d] succeeded...\n", model_id);

7.4. Run DME App

After building the example code using make -j, we can run the MobileNetV2 example in command line.

Kneron@ubuntu:~/host_lib/build/bin$ ./kl520_dme_async_mobilenet_classificaiton

The top 5 results for each image is printed out. After finishing the inference for 100 images, the average time of each frame and the fps is calculated and printed out.

/****************Top 5 Results***********************/
Index 281: score 0.595752
Index 285: score 0.162995
Index 282: score 0.146307
Index 904: score 0.007110
Index 173: score 0.001945
/****************Top 5 Results***********************/
Index 169: score 0.344763
Index 247: score 0.309466
Index 256: score 0.044287
Index 188: score 0.015038
Index 215: score 0.015038
[INFO] average time on 100 frames: 48.570648 ms/frame, fps: 20.588566

We can compare the classification index results with the index list on here: Imagenet Class Index

8. Create New SDK Application

8.1. KL520 Firmware Architecture

KL520 firmware is consisted of two bootloaders, IPL and SPL, and two RTOS (Real Time Operating System) images running on system cpu (SCPU) and NPU-assisting cpu (NCPU).

When IPL (Initial Program Loader) in ROM starts to run on SCPU after power-on or reset, it loads SPL (Secondary Program Loader) from flash (automatically or type 1 in UART menu), then SPL loads SCPU firmware image from flash, and finally SCPU firmware loads NCPU firmware image which runs on NCPU.

Both SCPU and NCPU firmware run RTOS with SCPU handling application, media input/output and peripheral drivers and NCPU handling CNN model pre/post processing. Two CPUs use interrupts and shared memory to achieve IPC (Inter Processor Communication).

The examples of SDK here are for SCPU RTOS firmware and NCPU RTOS firmware. Both uses ARM Keil RTX.

download the firmware source code: KL520 SDK section in Kneron Develop Center

8.2. Firmware components

8.3. Application Architecture

An application is consisted of one or multiple CNN models for specific purposes, and corresponding preprocessing and postprocessing. There are many kinds of applications depending on specific use cases. Some application could have their image processing streamlined for best performance and some may not. Some applications could have multiple models, and some may have single model.

A host mode application means that camera(s) and maybe display are located on the same board with KL520, and a companion mode application assumes camera image comes from outside like a different chip or PC.

Tiny Yolo is a single model application with streamlined processing. Both companion mode and host mode are supported. Figure below is a companion mode example.

8.4. Create New Application Project

Example: ./scpu/project/tiny_yolo_v3/companion/ Copy whole directory to a new one with files like this:

scpu/project/tiny_yolo_v3/companion/RTE/CMSIS/RTX_Config.c
scpu/project/tiny_yolo_v3/companion/RTE/CMSIS/RTX_Config.h
scpu/project/tiny_yolo_v3/companion/RTE/Device/ARMCM4_FP/startup_ARMCM4.s
scpu/project/tiny_yolo_v3/companion/RTE/Device/ARMCM4_FP/system_ARMCM4.c
scpu/project/tiny_yolo_v3/companion/companion.uvoptx
scpu/project/tiny_yolo_v3/companion/companion.uvprojx
scpu/project/tiny_yolo_v3/companion/main/main.c
scpu/project/tiny_yolo_v3/companion/mozart_96.sct
scpu/project/tiny_yolo_v3/companion/post_build.bat
scpu/project/tiny_yolo_v3/companion/vtor.ini

Host mode may include more files to support UART console, camera image inference:

scpu/project/tiny_yolo_v3/host/main/main.c
scpu/project/tiny_yolo_v3/host/main/kapp_tiny_yolo_console.c
scpu/project/tiny_yolo_v3/host/main/kapp_tiny_yolo_inf.c
scpu/project/tiny_yolo_v3/host/main/kapp_tiny_yolo_inf.h

Furthermore, make sure libraries are included such as these for companion application:

kapp.lib
kdp-system.lib
kcomm.lib
sdk.lib

Host mode application may include less libraries if no communication (kapp.lib or kcomm.lib) with another chip or PC is needed:

kdp-system.lib
sdk.lib

8.5. Create New NCPU Project

Example: ./ncpu/project/tiny_yolo_v3/ Copy whole directory toa new one with files like this:

ncpu/project/tiny_yolo_v3/RTE/CMSIS/RTX_Config.c
ncpu/project/tiny_yolo_v3/RTE/CMSIS/RTX_Config.h
ncpu/project/tiny_yolo_v3/RTE/Device/ARMCM4_FP/startup_ARMCM4.s
ncpu/project/tiny_yolo_v3/RTE/Device/ARMCM4_FP/system_ARMCM4.c
ncpu/project/tiny_yolo_v3/main/main.c
ncpu/project/tiny_yolo_v3/ncpu.uvoptx
ncpu/project/tiny_yolo_v3/ncpu.uvprojx
ncpu/project/tiny_yolo_v3/post_build.bat

Also, must include libraries:

kdpio-lib.lib
sdk-ncpu.lib

8.6. Create New Workspace to Include All Projects

Copy the workspace.uvmpw file to your directory, add/remove projects as needed. example_projects/tiny_yolo_v3_companion/workspace.uvmpw A companion application workspace usually contains these projects:

Project:sdk
Project:kcomm-lib
Project:companion
Project:sdk-ncpu
Project:ncpu

A host application workspace usually contains these projects:

Project:sdk
Project:companion
Project:sdk-ncpu
Project:ncpu

8.7. Create ISI Companion Application

Main tasks in main.c

Add operations for ISI command handler, e.g. in a shared directory/file (app/tiny_yolo_ops.c):

static struct kapp_ops kapp_tiny_yolo_ops = {
    .start          = tiny_yolo_start,
    .run_image      = tiny_yolo_run_image,
    .get_result_len = tiny_yolo_get_result_len,
};

/**
  .start: check application id at init time
  .run_image: pass image and parameters to middleware driver to
              run with the model(s) (model id TINY_YOLO_V3 here) 
              of the application
  .get_result_len: tell the length in bytes of a result buffer
**/

struct kapp_ops *tiny_yolo_get_ops(void)
{
    return &kapp_tiny_yolo_ops;
}

Register new ops with ISI command handler:

struct kapp_ops *ops;

ops = tiny_yolo_get_ops();
kcomm_enable_isi_cmds(ops);

Support multiple models:

When an application includes multiple models, each model needs a separate result memory, and all result memory buffers must be allocated in DDR using kmdw_ddr_reserve() because they are filled up by NCPU.

For companion mode this can be all done in .run_image callback function like age_gender ISI example where two models are run one after another. scpu/project/ssd_fd_lm/app/ssd_fd_lm_ops.c

Parallel image processing for NPU and NCPU:

When incoming images could be fed to NPU running model while previous image’s postposing to run on NCPU in parallel, a parallel bit can be set in image format to enable this feature.

define IMAGE_FORMAT_PARALLEL_PROC BIT27

8.8. Create Host Mode Application with MIPI Camera

Main tasks in main.c

Add your application needed console/camera/display and inference controls, such as those in:

scpu/project/tiny_yolo_v3/host/main/kapp_tiny_yolo_console.c
scpu/project/tiny_yolo_v3/host/main/kapp_tiny_yolo_inf.c
scpu/project/tiny_yolo_v3/host/main/kapp_tiny_yolo_inf.h
Host mode application can use the common operations as in companion mode with example in kapp_tiny_yolo_inf.c:
struct kapp_ops *ops;

ops = tiny_yolo_get_ops();
kcomm_enable_isi_cmds(ops);

8.9. Register New Pre/Post Processing and CPU functions

For application that using new model, users need to register the corresponding pre and post process functions. User can refer to tiny_yolo_v3_companion project's main function in ncpu.

First, user need to define an new model ID for the model. For example, TINY_YOLO_V3 is defined model ID for tiny yolo v3 model.

There is a default pre-processing function to handle scaling, cropping, rotation, 0-normalization with hardware acceleration.

If a special processing is needed for incoming raw image, this API can be called to register in void pre_processing_add(void) function.

kdpio_pre_processing_register(TINY_YOLO_V3, new_pre_yolo_v3);

Same procedure can be applied to post process as well. We need to add the following into the void post_processing_add(void) function.

kdpio_post_processing_register(TINY_YOLO_V3, new_post_yolo_v3);

Sometime, KL520 NPU cannot handle some layers in the model, and user need to implement a CPU function to complete the model. The user will require to register the cpu function so that the runtime library knows what to do during the cpu node. Users can do it in void cpu_processing_add(void) function, add the cpu funcitons:

kdpio_cpu_op_register(ZERO_UPSAMPLE, new_zero_upsample_op);

Please note that user needs to define an new cpu function ID for this cpu function.

9. More SDK example code for SOC peripheral drivers

KL520 also provides some simple examples to show how to use basic peripherals such as, I2C, PWM, DMA, GPIO... User can find them from sdkexamples folder.

There is also a PDF file to briefly describe the peripheral APIs. Please download it from the following link: KL520_Peripheral_Driver_APIs.pdf

10. Flash programming

10.1. Board Overview

10.2. Hardware Setting

10.2.1. Connecting UART0 (Program Flash via UART0 Interface)

UART0: Command Port

10.2.2. Connecting 5V power and trun on power switch

Power from Adapter or from Power Bank (USB PD)

10.2.3. Wake up chip from RTC power domain by pressing PTN button

Please do it every time after plugging in the power

10.2.4. Set bootstrap settings to manual boot mode (Program Flash via UART0 Interface)

  1. Must set bootstrap settings to manual boot mode
  2. Reset pin or power must be controllable to enter manual boot mode.

10.3. Program Flash via UART0 Interface

10.3.1. Flash programmer necessaries

  1. Open command terminal for flash programmer execution
    Tool path: kl520_sdk\utils\flash_programmer\flash_programmer.py
  2. Install Necessary python modules: kl520_sdk\utils\requirements.txt
  3. Limitations: Only the listed argument combinations below are allowed.

10.3.2. Edit python verification setting

  1. Check UART port number from device manager
  2. Edit setup.py, search “COM_ID” and modify the ID to match your UART port number

    EX: COM_ID = 3 # COM3

10.3.3 Firmware Binary Generation (FW + MODELS)

Generate flash final bin file from other seperate bin files.
The script combines .bin files in "flash_bin" in predefined order.
Morever, the addressing is in 4KB alignment

Command

$ python3 bin_gen.py

    options argument:

    -h, --help      Show this help message and exit
    -p, --CPU_ONLY  SPL/SCPU/NCPU only

Output
flash_image.bin

Note
The following bin files are must

flash_bin/
├── boot_spl.bin        // bool spl bin file
├── fw_ncpu.bin         // SCPU FW bin file (generated by Keil)
├── fw_scpu.bin         // NCPU FW bin file (generated by Keil)
├── models_520.nef      // model information(copied from [host_lib]/input_models/KL520/[app]/)

10.3.4 Flash Chip Programming (FW + DATA)

$ python flash_programmer.py -a flash_image.bin

Please press RESET BTN while you are seeing “Please press reset button!!”

Afterwards, just wait until all progresses are finished (erase, program, verify)

Note
flash_programmer.py -a means to do flash chip erase + programming + verification

10.3.5 Flash Verification (optional)

$ python flash_programmer.py -v flash_image.bin

10.3.6 Flash Erase (optional)

$ python flash_programmer.py -e

10.3.7 Flash Partial Programming (optional)

$ python flash_programmer.py -i 0x00002000 -p fw_scpu.bin

# "**-i**" means the flash index/address to program  
# "**-p**" means the FW binary to program   

10.4. Program Flash via JTAG/SWD Interface

Connect JTAG/SWD.

  1. Check your flash manufacturer: Winbond or Mxic or GigaDevice
  2. Select a specific device based on flash manufacturer

    EX:

    device KL520-WB //Winbond
    device KL520-MX //Mxic
    device KL520-GD //GigaDevice

    Copy the bin file to kl520_sdk\utils\JLink_programmer\bin folder
    EX: flash_image.bin, boot_spl.bin, fw_scpu.bin, fw_ncpu.bin, etc.

10.4.3. Double click "flash_prog.bat"

Afterwards, just wait until all progresses are finished (chip erase, program, verify)

10.4.4. Check programming result

Please ensure all the results are "O.K.", and enter "qc" to quit and close J-Link commander

To program specific bin file to specific flash address

  1. Copy the bin file to kl520_sdk\utils\JLink_programmer\bin\
  2. Select a specific device+’-P’ based on flash manufacturer

    EX:

    device KL520-WB-P   //Winbond
    device KL520-MX-P   //Mxic
    device KL520-GD-P   //GigaDevice
  3. Edit loadbin command: Load *.bin file into target memory

    Syntax:

    loadbin <filename>, <addr>  
    loadbin .\bin\boot_spl.bin,0x00000000
    loadbin .\bin\fw_scpu.bin,0x00002000
    loadbin .\bin\fw_ncpu.bin,0x00018000
  4. Double click “flash_prog_partial.bat” and wait until all progresses are finished

  5. Check programming result

    Please ensure the results is “O.K.”, and enter “qc” to quit and close J-Link commander

11. UART Device Firmware Write (DFW) Boot

11.1. Hardware Setting

Please refer to chapter 10.2

11.2. Design Principles

The purpose of this document is to descript how to develop/apply the DFW (Device Firmware Write) (via UART) Boot feature for KL520.

You can download firmware bin file to KL520 internal Memory space through UART and then directly boot KL520 up from internal memory, therefore external flash could be optional.

11.2.1 KL520 internal memory space address

  1. Minion FW starts from 0x10100000 and size must < 0x2000 bytes
  2. SCPU FW starts from 0x10102000
  3. NCPU FW starts from 0x28000000

11.2.2 UART_DFW_Boot Workflow

  1. Send Minion FW through UART0/XMODEM and also process the followings sequence through UART0 interface
  2. Minion is the bridge to handle host command, collect bin data and then save it into internal memory space

11.2.3 Host control principle

  1. Initial UART COM port with baud rate 115200 and send ‘2’ to tell KL520 to enter 2.UART(XMODEM) mode.
  2. Send Minion.bin through UART by XMODEM protocol. Reference: https://pythonhosted.org/xmodem/xmodem.html
  3. Switch to baud rate 921600.
  4. Send SCPU FW to address 0x10102000. Buf[n] = Msg_header + data_buf[4096]

    • Msg_header:

      typedef struct {
          uint16_t preamble;
          uint16_t crc16;
          uint32_t cmd;
          uint32_t addr;
          uint32_t len;
      } __attribute__((packed)) MsgHdr;
    • data_buf[4096]: transfer bin by unit 4096 bytes(max.) each time

    • Packet TX Preamble: PKTX_PAMB = 0xA583
    • crc16: calculate checksum from buf[4] to buf[n-1] except preamble and crc16
    • cmd for mem_write: 0x1004
    • cmd for mem_read: 0x1003 (read back for verification)
    • addr: 0x10102000
    • len: 4096 or remainder
  5. Send NCPU FW to address 0x28000000. Buf[n] = Msg_header + data_buf[4096]

    • Msg_header:

      typedef struct {
          uint16_t preamble;
          uint16_t crc16;
          uint32_t cmd;
          uint32_t addr;
          uint32_t len;
      } __attribute__((packed)) MsgHdr; 
    • data_buf[4096]: transfer bin by unit 4096 bytes(max.) each time

    • Packet TX Preamble: PKTX_PAMB = 0xA583
    • crc16: calculate checksum from buf[4] to buf[n-1] except preamble and crc16
    • cmd for mem_write: 0x1004
    • cmd for mem_read: 0x1003 (read back for verification)
    • addr: 0x28000000
    • len: 4096 or remainder
  6. Send CHIP_RUN command with address set as 0x10102000. Buf[n] = Msg_header

    • Msg_header:

      typedef struct {
          uint16_t preamble;
          uint16_t crc16;
          uint32_t cmd;
          uint32_t addr;
          uint32_t len;
      } __attribute__((packed)) MsgHdr;
    • Packet TX Preamble: PKTX_PAMB = 0xA583

    • crc16: calculate checksum from buf[4] to buf[n-1] except preamble and crc16
    • cmd for scup_run: 0x1005
    • addr: 0x10102000
    • len: 0

11.3. Build dfw_companion firmware

The dfw_companion project in mozart_sw\example_projects\dfw_companion is the example for users to develop the firmware without external SPI Flash drivers.

You can rebuild dfw_companion project then get the updated fw_scpu.bin/fw_ncpu.bin in mozart_sw\utils\bin_gen\flash_bin folder.

11.4. DFW via UART0 Interface with Python

11.4.1. DFW Boot necessaries

  1. Open command terminal for flash programmer execution Tool path: kl520_sdk\utils\dfw_boot\uart_dfw_boot.py
  2. Install Necessary python modules: kl520_sdk\utils\requirements.txt
  3. Limitations: Only the listed argument combinations below are allowed.

11.4.2. Edit python verification setting

Please refer to chapter 10.3.2

11.4.3 Chip Initialize and Send/Start Minion FW

$ python uart_dfw_boot.py -s

Please press RESET BTN while you are seeing “Please press reset button!!”

Afterwards, just wait until seeing “Xmodem sends Minion file DONE!!!”

11.4.4 DFW fw_scpu.bin to memory address 0x10102000

$ python uart_dfw_boot.py -i 0x10102000 -p fw_scpu.bin

11.4.5 DFW fw_ncpu.bin to memory address 0x28000000

$ python uart_dfw_boot.py -i 0x28000000 -p fw_ncpu.bin

11.4.6 Command to boot up KL520 from memory address 0x10102000

$ python uart_dfw_boot.py -i 0x10102000 -r

Note
To write specific bin file to specific memory address
-i" means the memory index/address to program
-p" means the FW binary to program

11.5. DFW via UART0 Interface with hostlib

The kl520_util_uart_dfw_boot example in host_lib\example\KL520\kl520_util_uart_dfw_boot is the reference code for users to develop UART_DFW_Boot feature on their own platform.

You can do integration test with kl520_util_uart_dfw_boot to write SCPU/NCPU firmware to KL520 internal memory space and boot from specific memory address directly.

11.5.1. Edit kl520_util_uart_dfw_boot.cpp to assign proper com port

Search “#define com_port” and modify the ID to match your UART port number listed on device management.

11.5.2. Build host_lib examples

Refer to host_lib\README_CPP.md for further information.

11.5.3 UART_DFW_Boot integration tests

  1. Copy fw_scpu.bin and fw_ncpu.bin to hostlib host_lib\app_binaries\KL520\dfw folder.
  2. Execute host_lib\build\bin\kl520_util_uart_dfw_boot, and please press reset button when you see "Please press RESET btn!!......"
  3. Please check the message showed after kl520_util_uart_dfw_boot executed, make sure minion bin is transmitted via UART/Xmodem successfully, send fw_scpu.bin/fw_ncpu.bin successfully, and reboot KL520 from specific memory address successfully.

    EX:

Appendix

Host Mode Example

This sample code is an application which KL520 chip plays as a host chip with connected display and cameras for Object detection. You will need to program the flash image to device first. The flash image includes SCPU/NCPU firmware and model. Then, you can Start/Stop object detection by simple command via UART.


Hardware Requirements

Kneron KL520 series AI SoC Development Kit


Firmware Preparation and Installation

WARNING:
DO NOT program host mode FW to Kneron Dongle


Run example

  1. Turn on KL520
  2. Press PTN button
  3. Select “1” to boot from SPI to boot from the programmed FW
  4. Once it’s started, RGB camera and LCD will be turned on, and image will be continuously captured by camera and displayed on LCD panel.
  5. Also, see command menu and type command (1) - (4) for functions in UART console window(ex. Putty)

Commands

  1. Type "1" in UART console window to Start Tiny Yolo

    When a person is detected, yellow box is drawn around the person.

    Other detected objects are drawn around by blue boxes.

    Console window shows detected objects with FPS info.

  2. Type "2" to Stop Object detection

  3. Type "3" to turn off Pipeline mode, or toggle the mode, at runtime.
    • This pipeline mode can be toggled on/off (1/0) to demonstrate performance improvement.
    • Note that there may be delay to see first good inference due to opening camera/sensor.

Supported/Unsupported Peripheral Table

Image Input

Peripherals Companion Host Mode
MIPI CSI RX x O
DVP x driver/example
UVC Host x specified cameras
USB(proprietary) V x
SPI Master, non-DMA x driver/example
SPI Slave, non-DMA x driver/example
SPI Master, DMA x x
SPI Slave, DMA x x
UART x x

Image/Result Output

Peripherals Companion Host Mode
MIPI DSI TX x x
MIPI CSI TX x x
DVP x O
UVC device x x
USB bulk x O
USB(proprietary) O x
SPI Master, non-DMA x driver/example
SPI Slave, non-DMA x driver/example
SPI Master, DMA x x
SPI Slave, DMA x x
UART x O
I2C x driver/example
I2S x x
INTEL 8080 x x