I want to add the spp algorithm to my program so that the input and output images don't have to be fixed - tensorflow

https://github.com/nickliqian/cnn_captcha
Here is the git address, which is the code I am using.
Hello, I have a verification code recognition program. Now I can model the model and verify the recognition result, but I can only fix the image size of the input and output. Now I want to add the spp(Spatial Pyramid Pooling) algorithm into my program. I have tried many times and can't solve it normally. Can you help me? Thank you.
https://github.com/nickliqian/cnn_captcha
Let the input and output pictures not be in fixed size

Related

How to pick the right images for an object detection model for only 1 tag?

UseCase: I'm trying to extract certain parts of a screenshot which is taken from a game (with a tf object detection model) and extract the text within this part (custom model for the font used in the game).
I have trained a custom model based on SSD Mobilenet V2 and the object detection works quite okish, but sometimes the bounding box is off. I googled about selecting the right images and the right amount for training the custom model, but I couldn't find a good hint in the right direction.
I try to extract the following (surrounded by red):
The environmen can change:
Resolution of the game can be different (1920x1080, WHQD etc.)
Text in the box is not always the same
I have trained with 120 self made images (1920x1080) (90% for training 10% for test) (all of these images where a screenshot of the game) and as I mentioned the results are okish. Sometimes the detected area is off (cutting the content of the box or including a lot area of the box surroundings).
Maybe someone can help me/answering the following questions:
Could a bigger training dataset increase the accuracy?
Should I also take different resolutions into account when creating the training data?
Would it make sense to feed only the boxes without the rest of the game screenshot into the training? Or should I mix screenshots of the whole game and only box screenshots?
Thank you in advance ! :)

How to get labels for ILSVRC2012 Classification Task

The ILSVRC 2012 small classification dataset is not separated by folder and don't have a labels file. How get the labels for the training set?
I tried on nonpub downloads page but does not exist anymore, and i tried by the filenames but their don't have the synset id on it.
I've been having the same issue today following this tutorial on reproducing ImageNet Validation results. I think I've found an answer, even if partial
In the article they point out to this link to get the validation set for object detection. I downloaded it and had the same issue as yourself, it only contains images without labels. What I've found is that this same website had this other link for the bounding boxes. I've downloaded it and alongside with the bboxes it comes with the proper class for each image
Hope this helps!

ResourceExhaustedError When running network demo fourth try

I have 1600 videos and I want to make joint annotation label data about videos.
I've already made the open pose network and I put my videos as input of the network and saved the joint data as json file.
When I put my first video data as input, there are no errors. And when I put second, third video as input, there are no errors too.
But When I put the fourth video data as input, I got these error message.
enter image description here
enter image description here
these above images are the error message.(OOM)
The size of first, second, third, fourth video is the same.
When I change name first and fourth video name, I got the same error when putting fourth video.
I think this error is about the graph. but I couldn't know why exactly.
I think there are many genious on stackoverflow. So please answer my question... :)
I solve this problem by using cpu. not using gpu.
I use cpu only in tensorflow for solving this problem. and it works!

Tesseract cannot recognize my image correctly

I am developing an Android app now, it needs to recognize captcha from website.
I utilize the tess-two to recognize captcha and follow TrainingTesseract3 instructions to train my own traineddata (using jTessBoxEditor to correct characters), but it cannot recognize correctly and even cannot recognize it.
The below TIFF image is that I use to train my Tesseract, I collect many captchas and merge them into a image.
TIFF image
The image that I want to recognize
For example, the expected result of the above image should be k8666, but the actual result is only 66.
Does anyone give me a help? Thanks.
I tried your images using a .NET wrapper for tesseract-ocr Tesseract-ocr .Net Wrapper by Charliesw.
I got some better results like (K8EEE, K8656), i think you have to increase the text font and make it bold and i saved the image in tiff format with 96DPI resolution to get a better results than mine.

Printing labels on vector layers by geoserver print module (MapFish)

I use MapFish plugin for geoserver. Vector layers are printed fine but text labels on them do not appear in pdf. Although they are visible on the map before printing.
Help me please =)
Please, use latest mapfish-print. You need to download and compile it as described here https://github.com/mapfish/mapfish-print