I'm trying to add CEA-608/708 into different video containers using ffmpeg. I have found out that ffmpeg will not support either of them,is there any other API to do this. I don't want web app or software i Want to do with API
Related
I've implemented smart home service for camera, including SYNC, QUERY, EXECUTE. Execute response is returning adequate hls url. However, casting that stream to Android tv via chromecast results in black screen.
Do I need some configuration to be able to stream hls?
Also, I have another question.
Is it possible to watch stream inside Google Home application on phone?
Chromecast supports hls protocol for video streaming now. A possible reason for the
black screen during casting the stream on Android TV using google chromecast can
be due to the usage of the custom driver instead of the generic driver. It would be
great to check if the streaming works fine with a generic driver. More information in
regards to this can be found here.
Currently, I'm working on POC which converts h264 stream to MP4 file. I'm using NReco FFMpegConverter for that purpose and also use the Azure Cloud Service to host my service.
This Azure Cloud Service capture concurrent streams and convert the respective MP4 file. From the testing of this app so far I realize better to use Azure Function or Azure WebJobs to use only for the conversion of H264 stream to MP4 file. Then cloud service has more room to handle concurrent streams.
That I want to know, Is it possible to use NReco FFMpegConverter in Azure Function or Azure WebJobs without an issue? are there any special configurations to consider because FFmpeg runs in a separate thread with NReco FFMpegConverter?
Google Speech API claims to support a number of codecs (https://cloud.google.com/speech/docs/basics). I'm interested in processing an archive of a session produced by TokBox WebRTC.
Is there a sample code that does something like this? Does the archive need to be converted to a compatible format?
The default audio codec for WebRTC is Opus, which is indeed supported by Google Speech API. The trick is getting the audio out of an OpenTok stream and forwarded along to the recognition service; unfortunately this is no small effort.
Although some work has been done on this in an experimental capacity, there is no official support at this time. Recommend reaching out to TokBox support directly to discuss the specifics of what you're trying to build (email support at tokbox.com).
Disclosure: I work at TokBox.
My company streams videos using IIS Media Services to Silverlight players, the streams are delivered as adaptive bitrates (Microsoft Smooth Streaming). Due to support for Silverlight plugin being dropped by all major browsers, we are planning to migrate our streaming platform to Azure.
I have checked the documentations, samples & read articles and couldn't find anything on how to use existing smooth streaming encoded video without having to re-encode. We have quite a large asset to migrate, around 400GB, re-encoding is not an option, also we plan to dynamically encrypt our content using AES. Does anyone know how to go about this?
You need to perform following steps
Create azure media services asset
Upload files for specified asset.
Then you need to run media encryptor encoder "Windows Azure Media Encryptor"
Configure delivery options
See https://github.com/Azure/azure-sdk-for-media-services/blob/dev/test/net/Scenario/JobTests.cs.
Method
private IAsset CreateSmoothAsset()
covering step 1&2.
There are various tests in this file to cover encrypting asset using "Windows Azure Media Encryptor" encoder(see usage of
GetMediaProcessor(_mediaContext, WindowsAzureMediaServicesTestConfiguration.MpEncryptorName);
)
To configure delivery of protected content see - https://azure.microsoft.com/en-us/documentation/articles/media-services-protect-with-aes128/ .
There is also media processor called "Windows Azure Media Packager" which will allow you to package your smooth asset for example to HLS.
You can onboard your existing Smooth streaming assets to Azure Media Services without re-encoding them and apply dynamic encryption of AES and dynamic packaging to different streaming formats such as HLS, MPEG-DASH and Smooth Streaming. However, there can be some limitations and constrains. If your content is already encrypted such as Smooth Streaming + PlayReady it is not supported to dynamic encrypt to AES. Your content needs to be in clear form if you want to use dynamic encryption. Also your Smooth Streaming assets needs to be Smooth Streaming spec complaint. There are tools which generates Smooth Streaming files which is not spec complaint and not supported by Azure Media Services.
You can use creating assets from existing storage blobs article to start
https://azure.microsoft.com/en-us/documentation/articles/media-services-copying-existing-blob/
I hope this answers your question.
Cenk
I am using wowza with amazon ec2 and I uploaded one smil file in amazon for testing. For videojs flash player, I am able to do multiple bitrate Http live streaming(HLS) using video-js-swf plugin and a smil file. For all videos I need to create smil file and have to upload to amazon, and it takes lots of time.
Is there any way to use multiple rendition files only(without creating smil file and upload to amazon ec2)?
I removed wowza server configuration and created script which generates m3u8 and ts files using ffmpeg from mp4 files(which is having different bitrate videos).
We can directly add this m3u8 file in the video src(in web app). This will play http live streaming in ios and android devices.
i'm not sure if it answers your question, however Wowza has the following feature:
http://www.wowza.com/forums/showthread.php?14664-New-API-for-resolving-MediaList-%28SMIL-file-requests%29-through-Java-API