reject binary on iTunes connect issue, can not see remove button - app-store-connect

As you can see you in the picture, I want to reject my binary when waiting for review, but i can not see the blue rectangle which is shown in the official guide in my view.
Any one can help? Thanks!

Related

sony-camera-api trackingFocus

I can turn tracking Focus on and use the actTrackingFocus. Once the actTrackingFocus is set how can I get the coordinates back from the camera so I can draw a box in the Liveview box showing what the camera is focused on?
That is not possible with the existing API unfortunately.
Appreciate that this is an old question, but if you are still trying and OK playing in python...
The tracking focus location is (apparently) reported via the frame info packets, and thus you have to enable them and then decode.
We are attempting to do this with pysony 1
Use 'python src/example/pygameLiveView -i' to see the reported locations. You might need to add your 'actTrackingFocus()' call to enable tracking focus, but they should be rendered (box with triangle corners) on screen.
Since none of the devs have a camera which support tracking focus, we'd love to hear whether it works on not. :-)

Focusing box on QRCode scanner on iOS 7

I'm investigating if it's possible to implement the same functionalities of ZBar library with iOS 7 api.
Everything was good so far thanks to this tutorial.
However, I now want to have a green box shown on the screen whenever the camera detects a QRCode. The green box is supposed to wrap around the QRCode.
From the delegate of AVCaptureMetadataOutput, I can grab AVMetadataObject but the bounds getting from this object is always very small which is not correct, given the fact that my QRCode is very big on the screen.
Anyone has any ideas on how to achieve the green focusing box?
P/S: I came across the documentation and couldn't understand this line "If the metadata originates from video, bounds
may be expressed as scalar values from 0. - 1.". This is for the bounds property of AVMetadataObject
You can look at this tutorial for qrCode scanning using iOS 7.
I had to do the same thing in my scanner app. Here is a link that I found very useful and pretty much answered all my questions.
He goes step by step from setting up the scanner to adding the bounding box.

Image processing - How detect array similar colors from a image on ios

I want to create a area of similar colors when user touches into a point on imageview. Who knows library or technique to resolve problems for ios.
Original image link:
http://cannshine.com/images/1.jpg
After touched by user, link:
http://cannshine.com/images/2.jpg
Please helping, thanks!
It looks to me like you want to create a magic wand tool, correct? I can't help you come up with a way to handle this and I don't know of any libraries that can select a group of like-colored pixels based on tolerance, but there is this article here I found. It covers implementing a magic wand tool using Objective-C. I found it in the answer to this similar question on SO.

Multiple notifications on a tile in metro apps?

Am working with tiles and badges in metro apps. Here am able to display one at a time either image or number on a tile , But unable to display both at a time. Can anyone help me how to display image and some-number on a Tile?
Thank you.
As MSDN states:
A notification badge conveys summary or status information concerning and specific to your app. They can be numeric (1-99) or one of a set of Windows-provided glyphs.
http://msdn.microsoft.com/en-us/library/windows/apps/hh779719.aspx
Note: When you hit limits like this you might consider building a custom image that simply has whatever you want. But then you will realize that rendering an image in WinRT is not possible yet. So you will want to offload to some web server... just helping you think this through.

iOS: compare a slice of an image to library of options

I'm basically trying to work out how to take a slice of an image, say a screenshot of an iPhone home screen, slice out the first icon and compare it to a set array of images in a library. Any help on where to start?
I'm no iPhone programmer, but I might be able to suggest a few things:
The SURF feature detection implemented in OpenCV should help you with this
There is a nice article on using OpenCV in Objective-C code.
A quick & dirty way might be to use the difference blend mode which should return the difference between the 1st image(top) and the 2nd image(bottom). If there is no difference the result will be completely black. So, the more black pixels in the difference result, potentially, the more similarities between the compared images.
I'm not an iOS developer, so I don't know if there is an image library that ships with sdk or if there's a free/opensource library for basic image processing. Still this should be trivial to implement:
e.g.
- (int)difference((int)topPixel,(int)bottomPixel)
{
return abs(topPixel-bottomPixel);
}
Note: Syntax might not be correct :)
HTH
This may not help you with taking a screenshot of the iOS home screen... But these articles show how to take snapshots from within a UIKit application:
https://developer.apple.com/library/prerelease/ios/#qa/qa1703/_index.html
https://developer.apple.com/library/prerelease/ios/#qa/qa1714/_index.html
Perhaps you would instruct the user to press home-power (buttons) to take a snapshot and store in the photo roll, then load that screenshot into an app to process the screenshot.
Hope this helps!