Yes I've checked some examples from RKNN Toolkit 2 and they are working as expected.
Its a two steps process:
1. Use RKNN Toolkit2 to convert models (onnx, tensorflow, tflite, caffe, others) to RK3576 NPU
2. Use RNKK Toolkit Lite 2 to perform the inference on the board.
I can provide some instructions on how to do it, if anyone is interested. Everything is covered here.
At this moment I'm trying to find a fast style transfer model able to do inference (ideally something around 10fps, for 1920x1080 frames) on the H96 max M9s.