I am trying to use setTouchAFPosition with Sony Camera Remote API version 2.40 and a a6500 Camera.
I am able to get the Liveview and actTakePicture to work but any time I call "setTouchAFPosition" I keep getting "Not Available Now".
Has anyone been able to use this function with the a6500 camera or similar?
It could have something to do with the a6500 not supporting "Tracking Focus". Try running the "getSupportedCameraFunction" and "getAvailableCameraFunction" commands to see if the "setTouchAFPosition" commands is returned in the list.
Related
Im facing with an issue on desktop Safari 13.1 version. If I open the console in the web inspector (with a regular macbook which has webcam and mic) and execute this command on any kind of website:
navigator.mediaDevices.enumerateDevices()
First time it will return in the Promise result with a videoinput and an audioinput.
Second time it will return only 2 audioinput. Videoinput is disappear.
Unfortunately I call this method several times while checking the available devices on my solution.
Any idea why does it happen and how could I get the accurate information about the devices even If I call it more than once?
See the results here
I've found the same issue, also on my iPad running iOS 13.
It seems you need to request camera access first in order to see the correct device list.
navigator.mediaDevices.getUserMedia({ video: true })
This will prompt you for access to the camera (you need to be on HTTPS or localhost).
Grant permission, then run this again and you should see the videoinput device(s) listed in the returned promise:
navigator.mediaDevices.enumerateDevices()
I guess this makes sense as a privacy feature that a website cannot check if a camera exists without first asking your permission.
I am developing a python remote control application using the open source pysony library. The program should be able to shoot images in a loop, download them and then delete the artifacts on the camera (We can't manually format the sd card, since the idea is that, of a remote control application).
I'd like to point out, that I have read this post about the correct way of remotely deleting files. I have read the Sony API documentation and have successfully managed to control everything I need to, but the deletion of images. The camera in question is a Sony a6300, updated to the latest firmware as is its API.
The problem in question is the fact that the camera returns a success response ({'result': [], 'id': 1}) after trying to delete a set of image URIs, but the images still remain on the camera. I am using the remote control app and am connected directly to the camera wifi (making this the standard 1:1 connection). When I issue the delete command, the screen of the camera shortly displays a "controlling with smartphone... you cannot directly operate this device" message.
I have searched all around the www an can't seem to find and aswer.
Thank you in advance!
I am trying to control my Camera DSC-RX1RM2 with Remote SDK.
With the PDF guide [Sony_CameraRemoteAPIbeta_API-Reference_v2.20.pdf],
I think I can use [Continuous shooting mode]API for my Camera,
But the result always return ["error": [12, "No Such Method"]].
I want to ask where is the problem?my camera or the SDK or my source?
Unfortunately, the DSC-RX1RM2 is not supported by the API. Stay tuned to the Sony Camera Remote API page for any updates on supported cameras - https://developer.sony.com/develop/cameras/.
The latest API does support the DSC-RX1RM2, just confirmed it.
Also check that your URLs are like:
http://ip:port/sony/camera
or
http://ip:port/sony/avContent
I didn't append camera or avContent at first and got similar No Such Method errors.
I'm making a follow focus app for Sony cameras. My app needs to be able to set focus marks and recall them. There are a lot of functions for doing AF on the Sony Camera API, but I'd like to be able to set focus manually to a given absolute value. Is there a way to accomplish this with the Sony Camera API?
If not, am I able to make a feature request for the next version?
Unfortunately there is no way to manually set the focus to a value in the API. But I will let the team know about the feature request.
Using the supplied Android demo from
https://developer.sony.com/downloads/all/sony-camera-remote-api-beta-sdk/
Connected up to the WIFI connection on a Sony QX1. The sample application finds the camera device and is able to connect to it.
The liveview is not displaying correctly. At most, one frame is shown and the code hits an exception in SimpleLiveViewSlicer.java
if (commonHeader[0] != (byte) 0xFF) {
throw new IOException("Unexpected data format. (Start byte)");
}
Shooting a photo does not seem to work. Zooming does work - lens is moving. Camera is working fine when using the PlayMemories app directly, so no hardware issue.
Hoping from advice from Sony on this one - standard hardware and demo application should work.
Can you provide some details of your setup?
What version of Android SDK are you compiling with?
What IDE and OS are you using?
Have you installed the latest firmware? (http://www.sony.co.uk/support/en/product/ILCE-QX1#SoftwareAndDownloads)
Edit:
We tested the sample code using a QX1 lens and the same setup as you and were able to run the sample code just fine.
One thing to check is whether the liveview is ready to transfer images. To confirm whether the camera is ready to transfer liveview images, the client can check “liveviewStatus” status of “getEvent” API (see API specification for details). Perhaps there is some timing issue due to connection speed that is causing the crash.