I am using the FirebaseVision
Object detection to detect things from the CameraX
camera preview. It is detecting things find but I am trying to draw the bounding box of the items detected over the camera preview. In doing so the bounding box that firebase gives back is not for the image itself not the preview view to they appear in the wrong place.
The image size that I get back from firebase is 1200x1600
and the preview size is 2425x1440
How do I translate the bounding boxes returned from firebase to the correct screen coordinates?
What I ended up doing was I took the image size that the camera took, divided the width/height by the view width/height to get the scale size
if(isPortraitMode()){
_scaleY = overlayView.height.toFloat() / imageWidth.toFloat()
_scaleX = overlayView.width.toFloat() / imageHeight.toFloat()
}else{
_scaleY = overlayView.height.toFloat() / imageHeight.toFloat()
_scaleX = overlayView.width.toFloat() / imageWidth.toFloat()
}
Now that I have the scale I can then take the bounding box return by the firebase detector and translate the x and y coordinates by the scales
private fun translateX(x: Float): Float = x * _scaleX
private fun translateY(y: Float): Float = y * _scaleY
private fun translateRect(rect: Rect) = RectF(
translateX(rect.left.toFloat()),
translateY(rect.top.toFloat()),
translateX(rect.right.toFloat()),
translateY(rect.bottom.toFloat())
)
Which then gives you the scaled rect coordinates which you then draw on the screen