Transferring images from one iPad to another - objective-c

I have an app which displays an image when I say some word, and the word is from my dictionary which I had exclusively created for the app.
Now I want that image displayed on my friends iPad simultaneously, when it shows on my iPad. Is there any way for that?
Thanks and regards,
Christy

If both iPads are in the same local network, you can use cocoahttpserver to run a HTTP-Server on every device and announce the services via Bonjour. Once a new picture is available, that devices connects the server on the other devices and informs it, that new data is available. The second device than will connect the server running on the first device and download the picture.
the advantage of this approach: the protocols you are using are well-know and a lot of additionally tools are available.
But you can also have a similar, more downlevel appoach, where you design your own network protocol. In that case I'd use Asyncsocket

Use synchronize image with both ipad using internet or use APN to notify that image is ready.

Related

Is there a way to send multiple wifi camera a signal to take a picture at the same time?

I have 20 wifi cameras that are compatible with the camera remote api (all identical).
Using the API, is it possible to trigger all of them over wifi to take a picture at the same time?
I would basically like to trigger all 20 over wifi and then fetch the pictures from all of them. So any suggestions for how to do that would be appreciated. Mainly just triggering them together would be a great start.
Also, is it possible to send commands to the camera with the API when the camera is connected to a network via the play memories camera app or can commands only be sent via connections with the Smart Remote Control? Reason being, I can have multiple cameras connect to my network this way, but via the smart remote, I can only connect to a single pc to camera at a time as the camera acts like the hot spot instead of the other way around.
Thanks,
-Scott
That would be possible but since it's only possible to connect one camera to one phone at a time, you would need 20 phones as well and then you could send the command to take a picture to all the phones which would in turn trigger the camera connected to each one.
It is not possible to send commands to the camera other than by using the Camera Remote API.
It would appear that some cameras use WPS to make outgoing connections (ie. when using 5 AZ1 with LiveViewRemote)....
If you had such a camera maybe you could connect them all to the same WiFi router, and then use the API to initiate a connection to all at the same time and quickly tell each to take a picture in quick succession.
Here's a link which might help you get there:
https://github.com/Bloodevil/sony_camera_api/issues/8

Camera Remote API beta SDK: How to synchronize multiple DSC QX 100 cameras

Can we control multiple DSC-QX100 cameras using the Camera Remote API SDK from an iPad OS7?
The objective is to cause multiple cameras to "snap" picture at the exact same time. Perhaps each camera has an address (serial number)...can the software communicate with all cameras at the same time using multiple addresses? Need is limited to still photos and so-called fast, rapid photography. Video not necessary.
If so, how?
Unfortunately you can only control one QX100 lens at a time. This is because the lens connects over WiFi and you are limited to only a single WiFi connection at once on an iPad. It may be possible using a desktop PC with multiple wireless cards installed but that would be the only way.
As the Sony rep said, there's no way to do this with "officially supported" techniques.
The reason for this is that the camera acts as a WiFi Access Point (AP) - so while multiple devices can connect to it, most mobile devices can only connect to it and not anything else (since iOS and Android don't support connection to multiple APs simultaneously). This is also why you can't use other network interfaces when connected to the camera. (I don't know about iOS, but Android always prioritizes WiFi over cell network data, for example.)
Android devices have a feature called "WiFi Direct" that provides more flexibility in terms of peer-to-peer interconnection, but iOS does not support WFD. The QX100 DOES respond to WFD invites, and you can accept a pairing request with (if I remember correctly) a long-press of the shutter button. However, the official app only supports normal WiFi AP connections.
I have not yet attempted to see if using Sony's remote API in combination with the (unsupported but apparently present) WiFi Direct capability works.
More info on Wifi Direct and Android can be found at http://developer.android.com/guide/topics/connectivity/wifip2p.html
Marlin SONY, I disagree, wifi is Ethernet and by definition can handle multiple devices on the same network. If you run a phone or iPad as a hotspot and connect multiple devices, it works.
Multicam Switcher Basic is an example of a free app that supports cutting together multiple camera angles live. Unfortunately the app is still being developed so features like third party camera support isn't included, but it does show what is possible and awaits development.
I think this should be possible. Apps like CollabraCamâ„¢ (Multicam Social Video Production) or RecoLive MultiCam prove that is possible to use mulitple cams simultaneous.
I need someone too to develop an App to be able to use "two" sony dsc-qx for 3D shoots. Please, if You know how or who can do this contact me email#3-d.re

Web cam type video camera to iPad2 streaming over wired communication

I want to develop an App that requires wired communication between Web cam type video camera and iPad2. Basically I will directly connect Web cam and iPad2 using cable and when I start web cam, whatever images(picture/video) captured by web cam should be displayed on iPad2.
Based on my research on this I found that iPad2 cable is only made for iPod Program so the connector is not a traditional USB port I can't do direct communication between web cam and iPad2. Am I missing anything?
We are going to use Vivotek camera and they have mentioned here that we can use safari to receive the Motion-JPEG stream. I am wondering if that could also possible on iPad 2 and is it reliable?
Further I found Apple's MFi Program to develop electronic accessories that connect to iPod, iPhone, and iPad. Is there anyone out here used this already and know more about this if I can go for this?
Thanks.
You can receive a motion jpeg stream in mobile Safari or in a UIWebView in a custom app. I am not able to (yet) successfully receive a motion jpeg stream via an AVPlayer, AVPlayerItem or AVURLAsset.
Becoming a MFI authorized company is non-trivial (I tried once). They want larger established companies that have demonstrated they have the skills and manufacturing know-how/contacts to produce quality accessories.
Curious if you can step back from your initial requirement and see if you can figure out how to do it wirelessly for at least the last step to the iPad 2. Can go wired up to 2 feet away from the iPad and use a local private wifi network for that last 2 feet (say).
I recommend you add (use existing, or purchase) a wireless gateway. Connect the camera to the gateway, and then connect the iPAD to the wireless network, and then browse in safari to the camera and then you can view the image. There is no "hard wired" way to get this to work.
As for the "hard wired" portion of the question, I do not believe that is not possible without a lot of work and hardware. There is no "video in" on an ipad to make it a monitor for a camera.

Turn iPhone into a server programmatically?

I want to make my iPhone app display on a Mac's screen, kind of like AirPlay does with other machines. The only way I have heard to do this is, although I do not like it, turn the iPhone into a server. Unfortunately, I cannot figure out how to do so. I also wanted to set it up in such a way that my Mac automatically detects it. I have seen a similar setup in the game Chopper 2. My Mac app will have a simple timer that fires every few seconds to look for the iPhone, in the same way that Chopper 2's "Find iPhone" button does.
Is there a simple way to turn the iPhone into a server, or start a "session" like Game Center does?
One last thing: I know it is somehow possible, because another app I have actually gives my iPhone a web address at the click of a button. It is called the Dicenomicon, if you want proof.
First there is no easy way to redirect your display to Mac, even if you made it a server of some kind.
Second, to discover or publish customized services on WLAN, you may want to refer to the samples on Bonjour:
CocoaHTTPServer: a simple TCP/HTTP server.
WiTap: an app that discovers and connects to services of the same kind on WLAN by Bonjour.
I'm not really sure what you mean by "server", because there is no way to share the screen of an iPhone using the official SDK, although this is possible by jailbreaking.
It would be possible, however, to send data back and forth between the Mac and iPhone, and display the data on the iPhone, on a Mac. Using that data, you could try to recreate the interface on the Mac. All of this could be accomplished using sockets. A class that might help with that would be cocoaAsyncSocket, which makes network programming a lot easier.
The auto-discovery of iPhones on the local network is achievable with Bonjour. Without getting into too many details, NSNetService would allow you to publish a service for your app from an iPhone, and NSNetServiceBrowser would allow you to find that service on the local network from the Mac. From the NSNetServiceBrowser, you could establish a socket connection with the iPhone.
Good luck!
You might want to take a look at the GameKit APIs, I know they do something similar between two iOS devices.
http://developer.apple.com/library/ios/#documentation/NetworkingInternet/Conceptual/GameKit_Guide/Introduction/Introduction.html

Robust real-time communication between iOS App and Mac App

We're working on an exhibit (http://pulse.media.mit.edu) and I'm brand new to iOS, objective-c and Xcode. The exhibit deadline is in one week and I'm stuck.
The problem I'm having seemed simple enough.
Our exhibit has a projector and an iPad. The projector will be hooked up to a mac and be playing a video. The iPad will act as a controller for those videos. More simply:
I have 50 videos on a Mac. I need to develop an application on the Mac, that, when opened will loop one of the videos.
On an iPad, I need to develop an app that can change between the videos on the mac in real time. The iOS app is already designed, we're just struggling with some code.
On the mac, when the iPad tells the Mac to change video, we'd like it to switch between videos using Core Animation, like this (http://youtu.be/pyd8O-2mkgk?t=1m).
So my question: What is the most robust way to do this? It has to be able to run in a museum, for two months. Some things to consider:
We are 4,000 miles away and can't monitor it all the time. We'll
check nightly to see if it's still working, but it should run the day
without breaking.
If people unplug the iPad, it should still work.
It should be as robust as possible.
How can I best do this? Should I write to a database from the iPad onto a database running locally on the connected Mac and then monitor that database 10 times every second? Are sockets robust enough to use alone?
If you do suggest a way, can you please point me in the direction of some resources (frameworks, function names, etc) that can help me do this quickly?
Thank you for your time.
I would go for the server-in-the-middle option, because it will be easiest to debug, and requires nothing more than a working internet (wifi) connection on the client side. When having connection issues, all you need is someone who knows how to hook up an iPad or Mac to the internet. And you can see server-side which device has issues connecting.
Furthermore, using plain HTTP sounds like the best way to go for communication, and the backend can be written in any server side scripting language. Both clients should be polling the server every X seconds.
If you get this basic setup working well before your deadline, you could try to get the devices to connect to eachother directly (for a less sluggish user experience), and leave the server solution in there as a fallback method. The Mac app could function as a HTTP server, accepting the very same commands that the iPad normally sends to your server. The challenge will be to reliably know which IP to connect to. You could hardcode it or use Bonjour.