Not to be rude, but if you can't work that out you shouldn't be working on this. I'm sure the author would take a pull request, but but it really sounds like you are nitpicking.
Figures like precision and recall are often expressed either as 66% or 0.66. Confusion really isn't that big a problem.
If you would write mAP: 0.9 or mAP (%): 90 or similar my "professor" would be OK with that.
I guess general rule in science (or maybe engineering) is to define units. I have no doubt that e.g. when you are building rockets (or maybe model airplanes) you would know that e.g. velocity is in m/s. Until somebody thinks in some other way and things blow up.
I just wanted to make some fun comment about missing units and that is it. I have no doubt that anybody browsing that repo would not be confused about mAP being 95.6.
The algorithm is state of the art. There are several implementations of the algorithm, the original is with the mxnet framework, and my implementation with tensorflow framework has a slightly lower accuracy.
Generally it means something roughly like "the best known approach for this specific problem". Often it means "the best known approach for this specific dataset" (eg "SOTA on ImageNet").
> why is lower accuracy acceptable?
Lower accuracy is worse, but these numbers look close enough that it's probably acceptable for most people.
There are plenty of environments where TF is preferable to MXNet (eg, you have TPUs/want to use TFLite on mobile/want to slice the model weights up and use it for your own custom TF model).
It's probably lower accuracy because it wasn't trained as long. Those extra couple of points could take days (or more) of training.
The paper that presents this algorithm has the best known accuracy on the widerface dataset, which is why it is called state of the art. The authors of the paper published an implementation of this algorithm based on mxnet, but a lot people and companies use tensorflow instead of mxnet in their work, so just using the mxnet implemenation is not an option. Thats why I converted it to TF, with a slight decrease in accuracy on widerface. Then what is "acceptable" depends on your judgement, but the widerface dataset is extremely challenging, with many pictures having hundreds of small faces. In some of these pictures my implementation will miss a few faces, or find them but with lower probability. Overall for the vast majority of face detection applications, the two implementations will yield extremely similar results. I havent yet found a picture with a few dozen faces where the TF implementation performed not as well as the original one !
would you be interested in a smaller model file size or in a model that runs faster ?
For model file size I have a few ideas of things that could be tried easily.
For a faster model, this model is based on resnet50 architecture. The original authors also published a lighter mobileNet architecture. I havent converted it to tensorflow but I could try to take a look
https://medium.com/@jonathan_hui/map-mean-average-precision-...