I'm trying to create an application that streams video and audio to an other computer.
On the "server" side, I'm able to capture (using AVCaptureSession) video and audio, to preview them and to send them over the network using the delegate and reconstruct everything back on the other side.
On "client" side I've then a CMSampleBuffer that contains audio and video and I don't find the way to play it back. I've checked AVPlayer and AVCaptureSession, but I don't understand the mechanism with a CMSampleBuffer for input.
Any ideas, links? Thank you!
you could try AVSampleBufferDisplayLayer.
Related
iam learning app development so it is very important for me if you people guide me. I have main activity and and two fragments, 1 for listitem and another for now playing. When clicked list item it pass the url to 2nd(now playing) fragment and start playing the music, everything going normal but i want to do this in background/foreground service with notifications control and i want to use exoPlayer or which one is better to stream deferent types of audio. Please guys i need to achieve this in any how help me. Waiting for yours answer. Thank you
I am working on a react-native project which uses Agora.io for video calling.
In a video call it shows my camera feed as fullscreen and the reciever's feed as thumbnail which is the opposite of the correct way.
I want to know, Is this the way agora works or is it possible to fix this..?
Because even in their website they have put the images in that way.
image on the home page
I appreciate any help regarding to fix this.
So it seems like you are presenting the local video stream to the larger view. You would need to switch this. Render the remote video stream on the larger view and the local video stream on the thumbnail view.
My design calls for a video playing in the background of my login screen, exactly like 6snap has.
I would like to avoid the default behavior of stopping the user's music when the video starts to play. My video does not have sound.
I'm using:
<MediaElement Source="MyVideo.mp4" />
I tried setting IsMuted=true which didn't help. Does anyone have an idea how 6snap managed it?
Edit: currently trying the animated gif route. Using the ImageTools 3rd party library and having converted my MP4, it works fine. My 9 second 640x1136 3MB video became a 41MB GIF, so I have to reduce the quality drastically. Still trying to find a better way if possible.
You won't be able to do that with Background Audio and MediaElement, hence as MSDN says:
When a MediaElement control plays audio or video content, any background sounds or media already playing are halted. The app launches the playback experience when the user taps the control. Only one MediaElement control can operate at a time.
It's no matter you have no sound - when you start to play all background sounds/media are halted.
I'm not sure how the App you have mentioned achieved that, but maybe you can try with DirectX/XNA - thought I've not tried this and don't know if that would help.
I was able to display a video using NSURL from a webpage into a UIView, but i need to be able to find a way to stream a video from a web server into my iphone using avfoundation. i am a little confused as to what i have to do in order to do this. any pointers would be helpful. thank you.
If you just need simple movie playing then take a look at MPMoviePlayerController. For more advanced needs start at AVPlayer
I've created an animation which runs inside of the Google Earth plugin (browser) and I'd like to somehow encode this animation into a video format that I can upload to YouTube or a related video site. Are there any tools out there to help me do this?
**EDIT: more detail
This animation changes depending on user input. So it needs to be scalable. The user would click a button: download video after which a server would convert the animation.
You can use FRAPS to record a video of the animation running on your machine.