Screen share and video display simultaeously - webrtc

Currently, when I'm screen sharing I replace the webcam stream with the screen share stream so only one can be displayed at a time. So is there a way to access both I want to screen share as well. I'm using sipjs.

Related

Screen sharing location

I made videosurface that shows my cam video and a Videosurface that shows screen sharing.
When you start screen sharing, you will see a screen shared by two screens, the cam video videosurface and the videosurface screen sharing screen.
I want to see the screen shared by the other party, but other users do not see it on the videosurface created for screen sharing, but my cam video is visible.
Can you show the video on the videosurface I want?
Tankyou in advice.

Agora WebRTC Change video call view

I am working on a react-native project which uses Agora.io for video calling.
In a video call it shows my camera feed as fullscreen and the reciever's feed as thumbnail which is the opposite of the correct way.
I want to know, Is this the way agora works or is it possible to fix this..?
Because even in their website they have put the images in that way.
image on the home page
I appreciate any help regarding to fix this.
So it seems like you are presenting the local video stream to the larger view. You would need to switch this. Render the remote video stream on the larger view and the local video stream on the thumbnail view.

React Native: Multiple previews of camera stream

I've got an application in React Native to be developed in Android and iOS, in which I'd like to take the camera stream, make some processing, and render it multiple times.
Imagine an app like Instagram, in which you can add filters to the camera live and it shows the applied filters with on live previews. It would be more or less the same.
I need to get the camera stream and be able to show it on screen multiple times.
I've tried react-native-camera, but it only lets me to get one instance of the camera to be shown.
With this same library, I've tried to take pictures in intervals of the camera and to render them via <Image> container but, of course, it's a performance kill and the app ends up crashing.
How can I achieve what I'm trying to do? Do you know of any module or any approach that allows me to do so?
Thank you.

Is it possible to programmatically force an ios ap in split view to go full screen in ios9?

I am developing this app for ipad and ios9 and at some point the user will click on a button to watch a video. But I wouldn't want the user to just see it in the split view size he's currently on (like 1/3rd or 14th) instead I would like the app to close the other open app and for it to take over the entire screen.
Does anyone know if that's possible? Thanks.
By-default, the video always get played in full-screen mode. When video playing is finished, the player gets dismissed and you will see your screen (from where you played the video). In your case, in split view controller.
Do let me know if you need further details.

Turning Google Earth plugin animation into a video

I've created an animation which runs inside of the Google Earth plugin (browser) and I'd like to somehow encode this animation into a video format that I can upload to YouTube or a related video site. Are there any tools out there to help me do this?
**EDIT: more detail
This animation changes depending on user input. So it needs to be scalable. The user would click a button: download video after which a server would convert the animation.
You can use FRAPS to record a video of the animation running on your machine.