What does the word "tap" mean on mobile platform? I can see there is IsTapEnabled property on UI.Element(s), however I don't understand what kind of action is "tap". What is difference between "tap" and "pressed"?
Pressed event is published when you start touching a control (= your finger starts to touch the control on the screen).
Tap event is published after you release the touch from the control (= remove your finger from the screen). It's the equivalent of Clicked in desktop applications.
Related
i have implemented a context menu in my controller and when i tested it in simulator with a long press it is not being displayed. I am using xcode 7.1.1 and watchos for development. Please help me.
How did you test it in simulator? Simply pressing the mouse button for a long time is not the way. You have to go to the Menu item: Hardware->Force Touch Pressure, and change it to "Deep Press" From that moment on, your mouse taps will all be long presses. You have to change it back to "Shallow Press" in order for your normal mouse behavior to simulate regular tapping. Note there are also keyboard equivalents: shift-command 1 and shift-command 2.
Can a Mac app temporarily disable Notification Center, for example while it is doing a full-screen presentation? Apple says alerts are "automatically disabled while you’re presenting in Keynote. They also won’t appear if your display is mirrored on an external monitor." Do third-party apps also have the ability to temporary disable notifications?
I don't think so, but my assumptions are base on iOS, where you can't turn off notification center at all.
You can enable Focus (aka Do Not Disturb) which will hide current notifications and prevent new ones from popping up. See https://support.apple.com/guide/mac-help/turn-a-focus-on-or-off-mchl999b7c1a/mac for how to enable/disable it.
I do not know if applications themselves have control over enabling/disabling Focus.
I use the Touch Bar and add the Focus button to quickly enable it if I need to present or share screen. I do not see a keyboard shortcut you can set in Monterrey like you could in Big Sur.
I am writing a Mac app that uses a window which is glued to the foreground of the screen. I am listening for the NSWindowDidResignKeyNotification to detect when it is appropriate to hide the window.
For almost every purpose -- including when a different application enters focus -- this works perfectly.
On OSX Lion there is an idiosyncrasy. When my window is open and glued to the foreground, and then the user selects the Launchpad icon from the dock to expand the Launchpad, my window stays in the foreground, even over Launchpad. It appears that Launchpad is implemented as some sort of screen overlay rather than a window that would take focus.
Does anyone have any solution to detect when Launchpad enters focus so that I can dismiss my window?
Is there any way to remove the swipe event that pulls in the notification center from the top of the ipad when developing an ipad app? If there is an easy way in PhoneGap, that would be best, if not - I can manually implement any Objective-C solution.
The app is already in full screen, and there are some gestures that including swiping at the top. These iPads are distributed to the sales force of this company, and they want that feature (top swipe notification center) disabled.
Cheers!
When your app is in fullscreen mode, Notification Center switches to a two-swipe system: swipe once to bring up a little handle, and swipe down on the handle to bring up Notification Center. This should avoid most interference.
Other than that, there's no public way to disable this gesture.
Of course not. At least not without a jailbroken phone. Notification Center is built into the OS and cannot be disabled by third party apps. As #jtbandes pointed out, their is the "handle" option, but there is no way to completely disable the functionality.
When you have an app in fullscreen mode (in a fullscreen space) the menubar is hidden unless you "nudge" the top of the screen. However if you hit the spotlight hotkey the menubar politely shifts down into the screen with the spotlight search bar. Excellent.
I am migrating a statusbar/faceless application that I wish to have emulate this spotlight's menubar handling. Under snow leopard the app responds to a system hotkey by displaying a menu (it has a NSStatusBar item and I call popUpStatusItemMenu). Under Lion when the app is in fullscreen mode/space however, the menubar remains hidden for the fullscreen app (note however that the hot key IS bringing up the statusbar item and using the keyboard I can activate individual actions on the menu -- I just can't see the menu pop up).
To reiterate - anyone know how spotlight (under lion) displays the menubar in response to a hotkey? Or should I be asking when a better way to handle my UI would be under Lion?
thanks,
I feel reasonably certain that Spotlight is using private interfaces to change the presentation mode of the login session. You can change the presentation mode for an application using -[NSApplication setPresentationOptions:] (or the deprecated-but-still-available Carbon equivalent SetSystemUIMode()), but these explicitly only propagate to the login session from the “frontmost faceful application”.
I suggest filing a suggestion for an API to request that certain UI elements (menu bar, dock) are shown if they’re currently in autohide mode from the background.