I have streams that contains multiple audio tracks and I'd like to provide a way to users so they can change the audio.
Does Chromecast web sender api has such functionality?
If not, how can this be implemented?
Yes, it is supported by APIs, read our documentation, specially the section "Using the Tracks APIs". Note that our Default an Styled receivers only support multiple Text Tracks but you can write your own receiver to support multiple Audio Tracks with the provided APIs.
Related
I've implemented smart home service for camera, including SYNC, QUERY, EXECUTE. Execute response is returning adequate hls url. However, casting that stream to Android tv via chromecast results in black screen.
Do I need some configuration to be able to stream hls?
Also, I have another question.
Is it possible to watch stream inside Google Home application on phone?
Chromecast supports hls protocol for video streaming now. A possible reason for the
black screen during casting the stream on Android TV using google chromecast can
be due to the usage of the custom driver instead of the generic driver. It would be
great to check if the streaming works fine with a generic driver. More information in
regards to this can be found here.
Google Speech API claims to support a number of codecs (https://cloud.google.com/speech/docs/basics). I'm interested in processing an archive of a session produced by TokBox WebRTC.
Is there a sample code that does something like this? Does the archive need to be converted to a compatible format?
The default audio codec for WebRTC is Opus, which is indeed supported by Google Speech API. The trick is getting the audio out of an OpenTok stream and forwarded along to the recognition service; unfortunately this is no small effort.
Although some work has been done on this in an experimental capacity, there is no official support at this time. Recommend reaching out to TokBox support directly to discuss the specifics of what you're trying to build (email support at tokbox.com).
Disclosure: I work at TokBox.
I am having a requirement to broadcast live audio on my website.
Scenarios are:
One user will talk/sing in my application (only audio) and
His followers will have to listen that live audio instantly in the same application(followers can listen only).
After broadcasting live audio the followers may give replay through chat.
Can we implement the above scenario using Tokbox?
Note: I am developing my web application in ASP.Net MVC5 + WEB API.
You should be able to do it with TokBox.
Use OT.initPublisher({ videoSource: null }) to publish audio only. See https://tokbox.com/developer/guides/audio-video/js/#voice And for singing you may want to tune the audio quality, see https://tokbox.com/developer/guides/audio-video/js/#audio-tuning
Subscribing to this stream will allow your users to listen to the live audio
You will be able access the audio by creating an audio-only archive. The TokBox API provides the audio-only archive for you but you will have to integrate a playback system into your chat system yourself. See https://tokbox.com/developer/guides/archiving/
I am trying to record audio on the wear and send it to the mobile and vice versa using the Channel API. However I can't find a working example of how to write this. Can anyone help? Thanks.
You can use WearCompanionLibrary that provides an API for that. The sample WclDemoSample uses that library and one of the features shown there is exactly that; it sends the audio stream (captured from microphone) from a wear device to the connected phone in real-time, using ChannelApi. If you don't want to use the library, you can take a look inside and see how it is done.
I would like to develop a chromecast receiver and sender app which would play the same video to multiple chromecasts and both those videos are in sync. Is there a low level access to video playback and can you use web sockets or other technology to synchronise playback between multiple devices?
You cannot control/communicate with two chromecasts using one sender. In other words, two chromecasts will have their own sessions and any synchronization between the two should be done out of band, using, say, a cloud backend or so, within each receiver.