Sometimes I want to "scan" documents with the back camera, and sometimes take selfies with the front camera. Is it possible to set the default camera before calling the Capture.capturePhoto() method?
I see that the default camera is always the one that was last used.
You should take a look at the new native camera API which lets you have more control over the camera.
This blog is about the library.
Related
I've got an application in React Native to be developed in Android and iOS, in which I'd like to take the camera stream, make some processing, and render it multiple times.
Imagine an app like Instagram, in which you can add filters to the camera live and it shows the applied filters with on live previews. It would be more or less the same.
I need to get the camera stream and be able to show it on screen multiple times.
I've tried react-native-camera, but it only lets me to get one instance of the camera to be shown.
With this same library, I've tried to take pictures in intervals of the camera and to render them via <Image> container but, of course, it's a performance kill and the app ends up crashing.
How can I achieve what I'm trying to do? Do you know of any module or any approach that allows me to do so?
Thank you.
We have an application for Mac OS X that needs to know when the user is watching a movie in full screen to change its behavior.
Is there any system programmatic "hooks" that allow native Objective-C application to know when fullscreen playback is started?
You can get a list of all windows by using the CGWindow API, like in the Son of Grab sample.
From there, you can look at the window levels to figure out which windows are full screen, but I am not aware of any way to look for video playback specifically, as different apps (VLC, QickTime Player) all use slightly different methods. Of course, you could hard code specific process names, and assume that they are doing video playback if they have a fullscreen window.
I'm following the code here to capture an image with the Google Glass camera.
https://developers.google.com/glass/develop/gdk/media-camera/camera#capturing_images_or_video
Intent intent = new Intent(MediaStore.ACTION_IMAGE_CAPTURE);
startActivityForResult(intent, 1);
Everything is working fine, except that the camera activity requires the user to "tap to accept". Is it possible to just take the photo after a second or so?
This is how the built-in camera app works, I just say "ok glass, take a picture" and it takes a picture without requiring any additional confirmation. The camera in my app is already being activated by the user with other voice/taps in the interface so this second confirmation tap is undesirable.
Instead of opening the Google Glass camera app, you could take the picture yourself: http://developer.android.com/training/camera/cameradirect.html
This is even mentioned in the GDK reference:
Building your own logic with the Android Camera API. Follow these guidelines if you are using this method:
Take a picture on a camera button click and a video on a long click, just like Glass does.
Indicate to the user whether a picture was taken or a video was recorded.
Keep the screen on during capture.
You can use Cuxtom Cam library to perform that task that you want.
I want to open the camera action like below pictures. just like camera open within some area only for focusing particular area of the image.
Please help me. Thanks in advance.
Normal Camera View
My Expectation
If you want to use the CameraCaptureTask, you will have to deal with its possibilities. It uses the built-in camera app of the phone, and available setting are the ones in that app.
If you prefer a customized photo taking app, you should use the PhotoCamera class:
PhotoCamera Class
You will find some guidance here:
How to create a base camera app for Windows Phone
We need to open the iPhone camera, to take images that will be saved to the camera roll.
I have read many examples here that all of them opens the UIImagePickerView.
Besides the fact that i cant understand why i have to open the picker view in order to open the camera , i just can't do that- i dont want the picker view, because i have my costumed photo album that we build, and we just need to have a little button in it, that opens the camera to take an image . without opening any other views above it .
Is that possible to use the camera without this pickerview that will cover my scene ?
or can i lead the user to the camera app and than take him back to my app ?
Thanks.
Instead of high level (i.e. Apple supplies the UI element) classes, you have to go to a more foundational (lower) level of API's, which would be AVCaptureDevice and AVCaptureDeviceInput.
And Apple has some nice source code available in their AVCam project.
If you want to display camera stream in you app without UIImagePickerController than you should you AVFoundation framework.
Here some some examples and tutorials:
take-photos-with AVFoundation
Custom camera
Displaying camera