ModalAI Forum
    • Categories
    • Recent
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    YOLOv8 with NPU (VOXL2 Mini)

    Ask your questions right here!
    1
    1
    11
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • Y
      yashpyrrhus
      last edited by

      Dear Dev Team,

      My goal is to run the yolov8s-oiv model on the NPU. I've quantized the model to INT8 following this method:

      from ultralytics import YOLO
      model = YOLO('yolov8s-oiv7.pt')
      model.export(format='tflite', int8=True, data='dataset.yaml', imgsz=640, nms=False, single_cls=False)
      

      where dataset.yaml provides details to the calibration dataset.

      When I run the quantized model on the VOXL2 Mini (SDK 1.6.3), it silently falls back to the CPU. I have edited the voxl-tflite-server.conf file to set delegate to nnapi.

      I'd really appreciate any inputs on the best practices for quantizing YOLOv8 for the NPU, and references to relevant documentation. I also wished to run a sanity check to ensure the NPU works, can you please provide some quantized models that I can use to verify a functional NPU?

      Thanks,
      Yash

      1 Reply Last reply Reply Quote 0
      • First post
        Last post
      Powered by NodeBB | Contributors