It sounds like you might be trying to train an object detection model or OCR at the moment. Object detection models infer bounding boxes around objects of interest and might not be suitable for your use case. OCR reads text off images and can’t identify the pointer in a meter.
If you want to derive precise angles (which you can use to convert to meter readings), I recommend using a keypoint detection model.
A contributor on our forum recently wrote a guide on how to estimate bottle orientation using keypoint detection. I think that’s a good place to start, just substitute bottles for the meter pointer/hand.
You can also search for keypoint detection specific models using Universe:
Oh, thanks for clarifying! In that case, I would recommend you use a object detection model to locate the meter itself, then use an OCR model (we’ve covered several here)
Okay so i only should train a model on your platform to identify the meter and then use e.g. Claude? The problem with GenAi is, we have no confidence value. I need the confidence.
We still need to update the post, but PaliGemma, a new multimodal model, does OCR the best and it can run locally on your device using Inference.
PaliGemma doesn’t provide confidence values either, but its worth noting that OCR-specific models and packages usually don’t provide confidence either.
We have an automated process and if the confidence is too low, an agent looks at the case. If its good enough, the values will be processed in the ERP system. So confidence is crucial and also a good quality.