get notified / check of end of transcoding for vimeo-api? - ruby

What's the best-practice method of checking for the release of the video after transcoding using the vimeo api?
Is there any way to set a callback url for notification?
I'll be using this ruby wrapper but can make a html request if need be....
Thanks.

There is not currently a transcoding callback (it is a planned feature), but the new API tells the transcoding status on all of your clip responses.
Look for "status", the valid fields are : available, uploading, transcoding, uploading_error, transcoding_error

Related

How do I test a webm videostream using jmeter?

AFAIK there are 2 video streaming plugins available for JMeter:
BlazeMeter - HLS Plugin and
UbikLoadPack Video Streaming plugin
UbikLoadPack as a prohibitive pricing tag and HLS Plugin doesn't test the format I need. Also I want a FOSS solution not a paid one.
Does anyone know of some other plugin or method I could test a webm video stream ?
Edit
#dmitri-t when I try to do this it just hangs. Here I found this script that shows how to test a video. But when I changed the parameters to my video and range to 0-100 it hanged.
Also the example is using HTTP and my video uses HTTPS.
Tried to include a timer. It hangs also.
Yet the video loads perfectly in Chrome with the same url I used in Jmeter.
I also tested the request with Postman. It ignores the range header. So what's probably happening in JMeter is that it's trying to load the whole continuous stream. How do I make it consider range header ?
I tested with Postman on an image in the same server to see if range header was being considered or if it was a server problem and range was respected correctly.
Content-Range header doesn't work also. Please check this related question I did relative to the range problem with streams: Request to a webm stream ignores range header
I don't think you need any form of plugin, you can simulate the browser playing the video using normal HTTP Request sampler sending simple HTTP GET request
Here is the evidence that "playing" an webm "stream" is nothing more than downloading it.
It would be a good idea to add Timers to simulate users watching the video till the end (or according to your test case)

Chromecast Receiver: Custom XHR Loader

I am writing a custom receiver for chromecast and was wondering if there is a way we can have our own, custom XHR loader functionality rather then the built in goog.net.XhrIo?
Basically I need to override the functionality of goog.net.XhrIo for all segments/fragments and media files.
I need this to send some beacons back to my servers for analytics.
Thanks!
If you are using the Media Player Library (MPL), then you can use skipRequest() and setResponse() to achieve what you want to do. Note that the Host class provides a number of overrides for updating Segments, Manifest, License and Captions request info.

FlowPlayer with Wowza

we're using flow player with Wowza. We've managed to get it so the stream starts on opening the page, however the stream is made up of several individual videos and between videos the user has to click play to resume the stream. Is there anyway to get flowplayer to automatically play the next video? Thanks
Dobro.
Assuming you are using the flash version of flowplayer (not the HTML5). Flowplayer offers a complete api, with events and properties to modify every aspect of the original configuration. For instance you could declare the 'onFinish' method in the clip, and then load another clip automatically
Check the flowplayer documentation:
http://flash.flowplayer.org/documentation/api/clip.html

Retrieving Timed Metadata on Chromecast

We are developing a custom receiver for HLS video playback on Chromecast.
Our stream has Timed ID3 Metadata embedded in the MPEG-2 Transport Stream (TS files). I need the exact position in the stream that these ID3 tags are located for our app to function properly.
In my Custom Receiver, I am registering for the Host.processMetadata event, and am receiving the metadata tags as the fragments are processed, but I am unable to determine the position of the tags in the stream.
I am looking for the best way to determine the position in the stream that the Timed Metadata is located. Is there an API Call I am missing?
Notes:
We are able to stream our HLS video, using the proper CORS headers.
We are getting the accurate position of the Timed Metadata when playing this stream on iOS and Android players.
We are working with an Android Sender.
Working with:
Cast Receiver 2.0.0
Media Player 1.0.0
Chromecast Firmware Version 26653
Thank you!
We are working on adding a new feature to MPL to address this exact same issue, to make the media time corresponding to the ID3 data available in processMetadata. I'll try to update this post when that is implemented and released.
Google has updated the Host API's ProcessMetadata method a week or two after I posted this question. The callback now includes the time in the stream that the metadata is located.
See the docs for more:
https://developers.google.com/cast/docs/reference/player/cast.player.api.Host#processMetadata

Chromecast testing for playable content receiver side

In the javascript api for HTML5 video there is a canPlayType method to check if something is playable or not. However I've noticed that on the Chromecast this function doesn't quite cover everything. On several occasions playing local content with the app I'm working on I'll get an error saying that it is not playable, but if I take off my error checking which uses canPlayType then I have no issues. I'm curious if anyone has nailed down a good way to figure out supported content other than simply checking strings against a list that we create using the supported media types from https://developers.google.com/cast/supported_media_types
The response to HTMLMediaElement.canPlayType()is supposed to be accurate and reliable. If you see any inaccuracy there on Chromecast devices, please file a bug with the appropriate information, thanks.

Resources