I am working on an action-camera like device. I need to implement the functionality to play recorded videos. Namely, the playback ui should be able to play/pause, seek, 2x speed playback and display time/duration of the AVI file. SDL 1 is used for the graphics and Buildroot for building.
I like libvlc, but for some reason it is really slow on the target system.
To write my own video playback ffplay as a foundation is the viable alternative. I find it to be low-level compared to libvlc though.
Could you please suggest me libraries like libvlc that are promising for video playback implementation in embedded Linux?
I'd look at the gstreamer framework, if it has support for hardware acceleration on the target.
Related
I successfully built and ran the test application from https://github.com/tensorflow/tensorflow/tree/master/tensorflow/contrib/hvx. I'd now like to benchmark HVX against the CPU implementation of https://github.com/tensorflow/tensorflow/tree/master/tensorflow/tools/benchmark, and if possible, the Android camera demo, to see how much it would help, but I wasn't able to find any documentation describing how to build said apps with HVX support (my builds run on the CPU). I'm testing on the Open-Q 820 development board with Android 7.0.
Is utilizing HVX acceleration outside the HVX test application, preferably with the benchmark and maybe the Android camera demos supported yet? If so, could someone please point me in the right direction? Thanks!
Currently, the Android demo app does not support the HVX runtime. But I'm sure that you can use the runtime with Android demo app by replacing .so file with HVX version. If you can wait for the official support, that would be happening soon, but no promise. Let me know if you have any questions :)
Has anyone played an ogg stream from iOS? What can I use without having to write my own decoder?
I'm using Apple's AVPlayer to play the stream.
This question has been asked as part of this thread:
iOS online radio streaming questions but it hasn't been answered.
There is no built-in support for playing ogg files in iOS, however, the source code for ogg compression/decompression is readily available (from xiph.com). It is fairly straightforward to compile this code for iOS and use it in an app.
I had a chance to compile the libraries and have made them available: "Precompiled Ogg Vorbis Libraries for iOS".
I have also made some sample source code available on GitHub that shows how to play decoded Ogg Vorbis audio on iOS: IDZAQAudioPlayer
.ogg file may also be encoded in Opus or FLAC, refer to Ogg wiki. IDZAQAudioPlayer may not cover all cases.
For Opus codec, you may refer Telegram for iPhone source code: https://telegram.org/apps
Use the VLCKit – ibvlc bindings for macOS, iOS, iPadOS and tvOS in Objective-C.
It's a library from the well-known media player "VLC Media Player," and it supports Cocoapods and can play/stream .ogg files on iOS seamlessly.
pod 'MobileVLCKit'
Check out this blog post for a more detailed tutorial.
Can we play mp4 songs using NAudio?
Please provide some code to play mp4 songs
I need a c# code that will help me play mp4 songs
You can play .mp4 files using NAudio, there is a demo in the official page.
Donwload the Demo-Apps and look for the "Media Foundation Demo". The C# code of the demo can be found at the Source Code of NAudio.
Link to Demo-Apps
Link to Source Code
The implementation is in the WasapiOutGuiThread class:
WasapiOutGuiThread : IWavePlayer
Update: NAudio can now play MP4 files using the MediaFoundationReader assuming you have the appropriate codecs installed. This works on Windows Vista and above, and if you are running Windows Server you may need to have the "Desktop Experience" installed.
Alternative approaches would be to see if the WPF Media Element can play it, or using the Windows Media Player ActiveX control (on an OS that can play MP4s).
I am using mono 2.10.4 and monobjc. I want to be able to play H264 video file and an mp3 file. Monotouch has examples of playing the same on their web site(http://wiki.ios.xamarin.com/HowTo/Video/HowTo%3a_Playback_a_Video.).
But can the same be accomplished using mono and monobjc ?. What would be the libraries that I would need ?.
There is a sample application named QTKitPlayer that demonstrate how to play a movie by using the QuickTime API; it is part of the Monobjc distribution.
What exactly is Adobe Air? I've seen a lot of people talking about it and I've even seen applications for it but I'm still not entirely sure what makes it unique or how it is different from other languages. Can someone please give me the concise version from a programmer's point of view?
Edit:
I wasn't familiar with Flex so I found this nice explanation: http://www.onflex.org/ted/2008/01/what-is-flex.php
In a nutshell.
Start with the assumption that you know what Flex and Actionscript are. Then take the fact that they both run exclusively in your browser and to all intents and purposes are for building web apps.
Now assume you want to develop the same app, with the same language and user interface resources, but run it as a desktop app on a workstion (PC, Mac, or Linux interchangeably).
AIR is what you add (as a link library) to Flex and Actionscript to accomplish that.
http://en.wikipedia.org/wiki/Adobe_Integrated_Runtime
Write cross-platform desktop apps in Flash, Flex, HTML, Ajax.
Adobe Air is a framework which allows to build desktop applications and it is based on HTML/JS and Flash.
Adobe Air its Flash Runtime that can run Flash inside it and provide access to your operation system.
Adobe Air can be used for gaming and software as usual Flash. Its stand alone flash player with extended and reach functionality. For example you can develop flash app that will interact with filesystem or hardware.
Also its support native extensions so you can extend Air using native C/Java libraries.
air can be produced as exe for windows, app for mac, ipa for ios, apk for android, linux with limitations and blackberry.
Adobe Air is cross platform language/tool for mobile, window and OSX application.