Agora WebRTC Change video call view - react-native

I am working on a react-native project which uses Agora.io for video calling.
In a video call it shows my camera feed as fullscreen and the reciever's feed as thumbnail which is the opposite of the correct way.
I want to know, Is this the way agora works or is it possible to fix this..?
Because even in their website they have put the images in that way.
image on the home page
I appreciate any help regarding to fix this.

So it seems like you are presenting the local video stream to the larger view. You would need to switch this. Render the remote video stream on the larger view and the local video stream on the thumbnail view.

Related

Facebook Cover Video in vue.js

Facebook's functionality which needs to be recreated is described really well in this video: https://youtu.be/CMKJoK3DI50?t=128
Our stack is Laravel + Vue.js 2 on a LAMP server.
the ability to upload the video, and after the video is uploaded, the ability to move the video behind the player so that it is "cropped" (see youtube video above, #2:20) - i am not sure if facebook actually crops their videos or simply leaves it whole and changes the placement of the video behind the viewable player so that a certain area of the video is shown, but the entire video is playing behind the "mask".
Any pointers on the best way to get this done in vue.js? And is the video actually cropped, or just "masked".
thanks, Sebastian

Preload first frame of video using react native video so user won't have to wait while playing

In my app i've list of videos like instagram and what i want to is to preload first frame of video so when user opens app they don't have to wait until it load.
I am using react-native-video. I can not cached video because its utilising too much memory and getting crash.
Any help/suggestion appreciated.
Try using ffmpeg-kit to generate the video frame and then render.

Snapchat style captions on recorded videos

I am using expo and i am trying to implement a feature similar to snapchat/instagrams draw on video and add text/caption to video before upload it. My problem is not the UI part but editing the original video and getting a uri for the new video
I know with images you can use libraries like expo-pixi then take a snapshot of a view but i am not how to go about this for recorded videos specifically
anyone would be kind enough to point me to the right direction?

Playing back a CMSampleBuffer

I'm trying to create an application that streams video and audio to an other computer.
On the "server" side, I'm able to capture (using AVCaptureSession) video and audio, to preview them and to send them over the network using the delegate and reconstruct everything back on the other side.
On "client" side I've then a CMSampleBuffer that contains audio and video and I don't find the way to play it back. I've checked AVPlayer and AVCaptureSession, but I don't understand the mechanism with a CMSampleBuffer for input.
Any ideas, links? Thank you!
you could try AVSampleBufferDisplayLayer.

Turning Google Earth plugin animation into a video

I've created an animation which runs inside of the Google Earth plugin (browser) and I'd like to somehow encode this animation into a video format that I can upload to YouTube or a related video site. Are there any tools out there to help me do this?
**EDIT: more detail
This animation changes depending on user input. So it needs to be scalable. The user would click a button: download video after which a server would convert the animation.
You can use FRAPS to record a video of the animation running on your machine.