Prediction takes too long time on CPU

Describe your question/issue here! (delete this when you post)
I am running inference on a local server on a CPU in windows. The server is running inside a podman container. for 400 images it takes ~6min for the prediction. Is there a way I can reduce the duration. Is it possible to check chunks of images?

  • Project Type:
  • Operating System & Browser:
  • Project Universe Link or Workspace/Project ID:

Hi @Anunaya_Thakur ,

Can you please share more information:

  • what is the model you are running
  • what is the size of images you are processing

Could you also try to run outside of container, directly in WSL, and confirm timings?

Thanks!
Grzegorz