I would like to make a personal application to be installed on two iPhones. The first to be used as a webcam that transmits to the second via wifi.
Having no experience with xCode, I am looking for a code example to connect 2 devices via wifi and transmit a real-time video stream.
Unfortunately, the documentation and examples I found are deprecated or partial and inconsistent.
Where can I find some code examples to help me solve my problem, preferably in ObjectiveC (but also in Swift)?
Thank you
Related
I am trying to build a video streaming platform and I need to implement a limited devices login feature just like netflix. I have seen some people using node device detector to get the device type from the useragent, but I don't think this is a good solution since the user agent can be faked. Please any ideas on how to effectively implement this?
Found a solution to my own answer, I can use fingerprint.js to identify devices and store in mongodb. and I found an open source version for fingerprint.js that is broprint.js
I'm trying to create a virtual classroom. Since I'm not familiar with the web conferencing (or conferencing) terminology, I'm not sure if I'm understanding WebRTC's capabilities as I should.
I've looked in the examples for WebRTC, and all that I've found seem to be peer-to-peer connections. As I understand it, peer-to-peer connections are between two entities. However, virtual classrooms are different as far as I know; they require all parties to be connected to each other, so that when one user speaks/types, all users hear her.
Is such a thing possible with WebRTC? If so, what is it called and how can I read more about it?
Check out the open source Big Blue Button project (http://bigbluebutton.org/). They're currently Flash based but are actively moving towards webRTC. Rumor has it they'll be using Kurento as their MCU. They also have open source mobile (Android/iOS) application code.
According to http://www.html5rocks.com/en/tutorials/webrtc/infrastructure/, such a thing is possible:
Beyond one-to-one: multi-party WebRTC
You may also want to take a look at Justin Uberti's proposed IETF standard for
a REST API for access to TURN Services.
It's easy to imagine use cases for media streaming that go beyond a simple
one-to-one call: for example, video conferencing between a group of colleagues,
or a public event with one speaker and hundreds (or millions) of viewers.
A WebRTC app can use multiple RTCPeerConnections so to that every endpoint
connects to every other endpoint in a mesh configuration. This is the approach
taken by apps such as talky.io, and works remarkably well for a small handful
of peers. Beyond that, processing and bandwidth consumption becomes excessive,
especially for mobile clients.
Maybe you can try searching in the webrtc google group
hope this helps
I'm trying to implement a webRTC based video chat example. I've taken a look at Quickblox and their solution. In particular, I've given their online demo a test run: http://quickblox.github.io/sample-webrtc-web/
I've opened two tabs in Chrome logged in as different users. It does not seem to work. Does this work for anyone else? Besides that, it looks like a great solution to what I'm trying to achieve so I'm hoping it is just a wrong setting with my computer.
P.S. I can see the local video stream (i.e. my own face staring back at me in the mini view, but the place where I should be able to see the remote user is black).
I was wondering if it's possible to open an internet browser from code in your arduino IDE or any other sort of program connected to the arduino and in turn open a specific webpage.
My idea was to have a button on my arduino uno, once pressed, it would open a browser and a webpage in that browser on my connected computer.
I don't mind if the answer involves code outside of the arduino but it would be great if someone has an idea about how to do this sort of thing.
I've done a lot of research and I couldn't find anything relating to the topic which surprises me a lot as this seems like a simple task, anyway any responses would be greatly appreciated!
There are several ways to open a web page with the help of arduino.
But all of them have on thing in common and that is:
1.You receive the data in your PC
2.You tell your PC if the desired data is received open a webpage.
The best way I can think of telling you pc to open a web page is through a programming language. This could be C#,Java,Processing or any other language that has a library that support to monitor the serial port.
After that all you have to do is to say with the help of programming language if this received then open web page..
Here ara two useful links that might be of help:
http://forum.arduino.cc/index.php?topic=138974.0
http://whichlight.com/blog/arduino-serial-out-to-browser/comment-page-1/
The cheap and easy way would be to make your Arduino appear as a USB keyboard and simulate the keypresses required. Of course, this is prone to breakage if the user is typing, not very cross-platform, etc. but it is fairly simple and doesn't require a program on the machine. I received a small USB dongle attached to an ad once that did this. They found a key combination that worked on both Windows and Ubuntu.
With an ASP.Net application you can easily do that,here is a a tutorial on this issue.
And after getting a signal from arduino all you have to do is to write in ASP.NET to open the required link
Yes, its totally possible given that you know how to code in python.
You can have your own program running that will monitor request from arduino uno using serial port. When it receives a particular set of bytes (say "0xabcdef"), it will call the function to open a browser and a web page.
Here are some of the sites that will help you know how to interface arduino to python and opening web browser using python.
http://www.olgapanades.com/blog/controlling-arduino-with-python/
https://docs.python.org/2/library/webbrowser.html
I want to create audio, video and text messagtes chat. Is it possible using WebRTC? Or it only allow audio and video chats?
One side of my app will be implemented using browser. An other one - using C++ native API.
Does anyone have examples in native C++ API and/or javascript?
The WebRTC specification is still very much in flux, but there's a DataChannel API in the spec that is implemented in an early form in both Firefox and Chrome. DataChannels are intended to allow you to send arbitrary bytes between peers, and the spec provides for both reliable (TCP-like) and unreliable (UDP-like) channels.
I am not sure if WebRTC allows for text chatting. I was able to successfully create an Android Application that performed all of this, but only with the combination of Google's Libjingle and WebRTC libraries. Within the Libjingle library there are several example programs/pieces of code that demonstrate the library's functionality. The call example in Libjingle sounds very similar to what you are wanting to do, and is what I built my Android application out of. The only thing is I have not yet ported it to an web browser, so I am not sure if Libjingle will work with that.
I have begun looking into it, and I have found some people on the WebRTC discussion group that have developed a very nice Multi-user video chat application for a web browser that is built using WebRTC. It is capable of video (along with voice) communications as well as text chatting. I do not know if this matters, but it all occurs within a single interface (meaning that it does not seem to allow for separated/singular form communications -- text only, voice only, video only). I am sure that it would not be too difficult to separate them all out if you wanted/needed. They have posted all of their code onto GitHub and seem to be actively updating and improving it.