gstreamer low latency audio capture- windows - windows

I am trying to use gstreamer for low-latency streaming purpose under windows, and what I am facing now is the audio capturing latency.
I tested audio capture latency by creating a loopback pipeline as follows:
gst-launch-1.0 directsoundsrc ! directsoundsink
However, the audio latency is higher than expected, more than ~200ms. I guess the directsoundsrc do some buffering for the audio samples, I tried to tune its parameters in a hope to get better latency (buffer-time/latency-time) but the results were still not applicable.
Does anyone knows better way to capture low-latency audio and feed it to gstreamer pipeline than the one I am using?

Related

ffmpeg streaming to youtube becomes slow and starts buffering

I'm using FFmpeg to broadcast to YouTube via rtmp.
I'm using a cloud computing instance.
Whether I put an HD video or low quality to broadcast, at the beginning it says that all is good and connection is Excellent, but after about 3 hours it becomes slow and starts to cut.
I don't understand why this is happening, and if you know guys any tips to fix that.
Thanks...

GStreamer H264 stream to HTML5 webpage

I have a Linux board with a camera connected to it that is recording whatever goes on to a mp4 file(s) on the SD card of the board. I'm using gstreamer which connects to the /dev/video1 source and uses H264 encoding. I run it with a command similar to this one:
gst-launch-1.0 v4l2src device=/dev/video1 ! video/x-h264,width=640,height=480,framerate=30/1 ! h264parse ! rtph264pay ! udpsink host={host} port={port}
The upper part works fine and records everything locally, but I'd also like to stream this video to a HTML5 webpage, which is meant to change camera options and have a live preview.
I tried using HTTP via tcpsink and HLS via hlssink, but both resulted in a 8-10 second delay, which is basically unusable. The only thing that has no delay is the UDP sink. As far as I know the only way to catch the UDP stream is by having a tool like FFMPEG in the middle, that can convert the UDP stream to MJPEG for instance and serve it to the webpage.
That would probably work, but the board doesn't have a very good CPU and is already at 50% utilization. Converting stream via FFMPEG would probably push it to 100%.
Is there any other way to stream to a webpage without delay?

Does the "live=1" on ffmpeg rtmp urls mean that the stream is live and you cannot rewind or pause it?

Or does it have some other meaning? I have searched all over the internet, and the documentation is very thin on it... If someone could point me to something that explains exactly what it is, I would appreciate it.
I am talking about this:
ffmpeg "rtmp://...... live=1" .....
tia.
Short answer is yes.
RTMP has live streaming support and vod support. 'live=1' means the rtmp is running a live streaming. In this case, the media server is receiving video feed from source in real-time. Therefore, rewind back to a previous time is not a supported action. Without 'live=1', RTMP is running on vod mode, which means the entire video pre-exist on media server, then the server is capable of rewinding back, or seek to a random position of the video.
Although technically, on client side (preferably with a software, not webpages), if you maintain buffer your self, you can rewind or pause one way or another. Since you are saving data as you are receiving from media server and everything is under your control, you will be capable of rewind and pause live streams. But you will have to implement the buffering and decoding mechanism yourself. ffmpeg command will not be able to help on this.

stream webcam using ffmpeg and live555

I am new to live555.
I want to stream my webcam from a windows 7 (64-bit) machine behind home LAN using ffmpeg as the encoder to a live555 server running on a Debian 64-bit linux machine in a data center over the WAN. I want to send a H.264 RTP/UDP stream from ffmpeg and the "testOnDemandRTSPServer" should send out RTSP streams to clients that connect to it.
I am using the following ffmpeg command which sends UDP data to port 1234, IP address AA.BB.CC.DD
.\ffmpeg.exe -f dshow -i video="Webcam C170":audio="Microphone (3- Webcam C170)" -an
-vcodec libx264 -f mpegts udp://AA.BB.CC.DD:1234
On the linux server I am running the testOnDemandRTSPServer on port 5555 which expects raw UDP data from from AA:BB:CC:DD:1234. I try to open the rtsp stream in VLC using rtsp://AA.BB.CC.DD:5555/mpeg2TransportStreamFromUDPSourceTest
But I get nothing in VLC. What am I doing wrong? How can I fix it?
From what I remember, it was non-trivial to write a DeviceSource class, the problem you're describing is definitely something that's discussed quite frequently on the live555 mailing list - you need to get yourself approved to the list a.s.a.p if you want to do anything related to rtsp development.
The problem you seem to be having is related to the fact that some video formats are written with streaming in mind, and the rtsp server can easily stream certain formats because they contain "sync bytes" and other 'markers' which it can use to determine where frame boundaries end. The simplest solution you could use is to get your hands on the SDK for the camera, and use that to request data from the camera. There are many different libraries and toolkits that let you access data from the camera - one of which would be the DirectX SDK. Once you have the camera data, you would need to encode it into a streamable format, you might be able to get the raw camera frames using DirectX, then convert that to mp4 / h264 frame data using ffmpeg (libavcodec, libavformat).
Once you have your encoded frame data, you feed that into your DeviceSource class, and it will take care of streaming the data for you. I wish I had code on hand, but I was bound by NDA to not remove code from the premises, although the general algorithm is documented on the live555 website, so I am able to explain it here.
I hope you have a bit more luck with this. If you get stuck, then remember to add code to your question. Right now the only thing that's stopping your original plan from working (stream file to VLC) is the file format you chose to stream.
One thing you can try is to increase the logging verbosity level of VLC to 2: VLC expects in-band parameter sets in which case it will print a debug message that it is waiting for parameter sets on the messages window. Just having the parameter sets in the SDP of the RTSP DESCRIBE is not sufficient. IIRC you can configure x264 to output parameter sets periodically or at least with every IDR frame.
Other things you can try:
You can test the stream with openRTSP before using VLC. If you use the openRTSP -d 5 -Q rtsp://xxx.xxx.xxx.xxx:5555/mpeg2TransportStreamFromUDPSourceTest options openRTSP will print quality statistics after streaming for 5 seconds. Then you will be able to verify that the testOnDemandRTSPServer is indeed relaying the stream, and that there is not a problem between the ffmpeg application and the testOnDemandRTSPServer.
Have you tried a different stream? Also, I had a similar problem due to issues with my firewall, you might want to make sure you can actually stream data through those ports.
If you are missing a Sync Byte, it's probably a stream issue - try using a different data source and see if that helps, try an .avi file or an .mp4 file, usually .mp4 files are easy to stream. If the streaming works with the .mp4 file, and not with your mpegts file, then it's a problem in your file - ffmpeg is trying to figure out where each "frame" or "frame set" of data ends so that it can try to stream discrete chunks.
It's been over 2 years since I last worked with this stuff, so let me know if you get anywhere.

Delay in displaying a video stream after streaming by libavformat

I am using libavformat library to stream a video at a network address like udp://127.0.0.1:1000. I use ffplay to display the received video stream at the network address. However, the video appears few second later (e.g. 5 6 seconds) at ffplay on the same machine. Do you know what is the reason?
More info:
I have written my own streaming application using libavformat. When I stream a 3sec 1080p video at 25fps, ffplay does not show anything. If I repeat streaming the same video once again, this time, ffplay starts displaying the previous streamed video as well as the current video. So, it looks like ffplay waits for a buffer to be filled up by some amount, and then displays the stream. But am I correct?
To do what you are describing you are (at least) encoding the video stream, sending it over a network socket and then decoding it again. If you are streaming from an already compressed source, there may even be an additional video decoding stage involved.
Depending on the video format, the compression and buffering settings, your network configuration and the hardware involved, a delay of several seconds is not out of the ordinary. People watching TV channels using their live streaming services often see such delays when comparing to the over-the-air signal, and TV stations are supposedly using professional equipment for the streaming process...
You might be able to get more specific help if you mention how you are using the libavformat library (especially if you have written your own streaming application), the codec settings and some basics about your video stream, such as its resolution and frame rate.

Resources