Flash - Streaming Live Video Over RTMFP Using A Format Other Than Spark?
Aug 16, 2010
I would like to serve high quality video (H.264) between a server and a single Flash Player client with low latency (RTMFP). The single client will be controlling a robotic device and seeing the results through the video stream, so low latency is important.
Option A: The Flash Player supports low latency streaming of video between two P2P nodes via the RTMFP protocol in Flash Player 10. It uses the Spark codec to encode the video and as far as I know it is not possible to encode using any other codec in the player.
Option B: Flash Media Live Encoder does support live encoding using other codecs (On2 and H.264) but cannot act as an RTMFP peer - this option would require Flash Media Server in the middle.It seems I cannot stream video to a single client over RTMFP with anything other than Spark.
There are 25 livestreams streaming to FMS4 server by fmle3.2,and client number is 20, (FMSCore.exe)CPU usage is about 20%~40%, how can i reduce cpu usage?
I need to create a simple video streaming without using Flash comunicator server or other streaming server.So I am thinking to use RTMFP but I haven' t found any good RTMFP tutorial for this issue.
Do you know tutorial for Video Streaming using RTMFP ?
In using Flash Live Media Encoder to aquire a webcam video and to send it to a Flash Media Development Server 3.5 for live video streaming to one client by the same local network.The client device doesn't support flash player but is able to read and play flv video format.So, if i get a flv video file from apache 2 it works.But I don't know how to get a live video streaming from the server. The client application needs a flv file to point to.Does FVS generate a flv file for live video streaming?
how to publish live video feed from webcam in h.264 format non VP6 format with FMS 3.5.2 without using Flash Media Live Encoder, and how to set all parametersto have a good quality and smooth video without interruption, i have a server with 50Mbit bandwidth output enough for a publisher and 10 clients meunderstand this thing's been months since I try but the quality ugly
I have an app that streams in video live, and archives it as it goes along.My question is this: What happens when I hit, or allow, 'pause' on a live stream, and then 'play'? Obviously the person on the other end streaming continues to do so, and my application continues to archive it, but what about the person receiving the stream?
Do they resume play from the archived copy, or do they seek ahead to the current point in the stream, missing everything they were paused for?
we are looking forward towards developing a very interesting community portal that would help the user to broadcast their live videos across the community.I've been checking over sites like ustream.tv, justin.tv and wondering what/how Technology they been using to do so.
I am doing a lot of research over the last few days checking over the medium to do this effectively and figure out some of the leading companies in the domain like Ooyala.com, brightcove.com providing servers/technology to broadcast videos seamlessly across the globe. I will be signing up with any of these providers soon.So my question is , how exactly would my website be catching with the live feed from the users cam, send the stream to ooyala/brightcove and further broadcast it to rest of the community users.
I can play the .flv movie after compilation through FlashDevelop but its not working if I move the whole directory into another PC or another Directory in the same PC.
I am looking to broadcast live video and need a camera that is better than a simple webcam. It looks like they used to make video cameras that would register as webcams, but stopped making them.If you wanted to broadcast an event live, what would be a good camera to use?
I have this project in mind which I want to create a video montage of different live streaming videos from a few different users with their webcam in different locations. There will be the website which the participants will be able to view the live video montage in real time. However, I am unsure of where to start with (live video & server).
I've tested an exciting tutorial about streaming live video with Flash Media Server 3.5..every thing went good and i could see my webcam broadcast from my machine through my web site but unfortunately.I can't see it from any other machine.I'm using Microsoft windows XP SP2and flash media live encoder 3but my web server run UNIX is this a problem?
We have video cameras from Point Grey Research that are typically used for instrumentation and measuring.To oversimplify, the Flycapture software library they provide essentially captures single frames as 2-dimensional arrays of 16-bit greyscale values. We want to measure and process this information in real time. We also want to take this video, write some simple overlays on it (rectangular outlines, histograms, etc.) and make it available for viewing on a local area network using an ordinary web browser.
That is, we need to programmatically generate a live FLV stream on the fly--where the source of the video is not actually a device, but our program. The resolution and bandwidth requirements are modest; 320x240 30 fps 8 bit-grey = less than three megabytes per second. Some very simple, fast, easy compression would do, and in fact compression is not really needed at all.
I'm puzzled as how to where to start. I've browsed dozens of free and commercial packages, and they all seem to assume that the video source is a device.
The FLV and RTMP specs are intimidating. I can't easily identify any trivial subset (e.g. no compression) that would be easy to implement myself, nor have I found free or inexpensive commercial libraries that would do it for me.
I'm using FMLE 3.1 to stream live video encoded with H.264 format with FMS 3 and Flash player 10 + AS 3 to connect to the stream. When I connect to the stream, just the audio is played. I'm able even to get the metadata information about the video, but I just receive the audio. I already tried some stuff like
1. "Flash 10 won't play live stream H.264 after iTunes install" [URL]. I tested it in a complete different environment than mine, but the same result.
2. I've tried some format to play method, but this is just to play files ns.play("mp4:saple.f4v"); ns.play("mp4:sample");
3. Also read "How do you watch and record a live h.264" [URL], but I don't get even to play the stream at first place. This is the code I'm using
import flash.media.Video; var video:Video = new Video(720, 480); var ncVideo = new NetConnection(); this.ncVideo.addEventListener(NetStatusEvent.NET_STATUS, onNetStatus); [Code] .....
Another detail is that when I record the video to a file for instance "sample.f4v", I put this video in the FMS but when I connect to this stream I receive the "FileStructureInvalid" error message. I went from changing the extension to .flv [URL] to the solution to flatten the files [URL] but this is not the case because I'm using FMS to stream the recorded video.
I am a Wedding/Events Photographer/VideographerOne of my client want his events be live streamed over internet, I am quiet new for this type of jobas to what Software / Hardware I have to purchase for Live Streaming of Video over Internet.. and how it will encode in realtime.?
I have Flash media server, flash media live encoder, and flash cs5.I'm gonna be using flash media live encoder to stream my desktop and webcam to my website. My current host (host gator) doesn't have rtmp enabled on their web hosting plan, only on their dedicated and vps hosting. My question is do I need to switch hosts to one that has rtmp enabled so I can stream to my website for people to view?Also, can I embed this stream into another part of my site?
I have a website (on my own Apache Server in my home office) that displays my video clips (.flv files) with no problem, I have no idea what ActionScript is or where it is used (but I am a basic user of JavaScript)And since this forum is for 'beginners', and since I don't even understand the questions being asked here , *where* do I start to acquire knowledge of ActionScript, or what books shoud I buy to start this learning (or where is the FAQ list please.I would only like to be able to show Live Streaming Video over my website,
I'm very new to FMS and I have been experimenting around with it as a part of my job. I already have the encoder, FMS, and my AIR application all talking, so that portion is going well.The problem that I want to tackle next though is to have the FMS server record the live stream to it's hard drive. All of the guides that I've found talk about how to make a DVR stream, but that's not what I want. I want to be able to have the live stream, and then separately I would like to have it recorded.
we're planning to set up a live streaming service, one operson live streaming a video to multiple users. so no more than one channels contemporary in uplaod, and at the beginning it won't be a open broadcast service (maybe later..) we initially don't know how many can be the users. I've red that there's no theoretically limit for the maximum number of users connected with a fms, is this also valid for live streaming, or has it some limitation due to be live? is it necessary to have our fms, or is it possible (and cheaper) use a server provider?
We have installed FMS on our godaddy server to work with and used live video streaming services. We are having problems with the video quality being choppy, slow and pauses continuously during live broadcast and motion. We can't figure what is wrong; video needs to be flawless and in high quality. The godaddy server is running:
Red Hat Fedora Core 7 Intel Core 2 Duo - 2.13 GHzRAM 2 GB RAM 250 GB Total Disk Space
I am facing below mentioned problem regarding video broadcasting with adobe flash media interactive server (Developers version). I am using DVRCAST for recording video from a webcam and broadcasting the same live. Please find video settings below. width:320 height:240 fps:15 quality:90 bandwidth:150 buffer time:0.01
Our problem is that videos are not playing smoothly in application. The motion is getting paused and interrupted in time of broadcast live. I downloaded and checked the recorded flv files from server and found the recorded file also has same problem. So this problem is appearing in time of streaming the video from client computer to fms server. This problem is not appearing when I am testing the application in my localhost.
My code for live streaming mentioned below: private function publishCamera():void { cam = Camera.getCamera(); cam.setMode(cameraSettings.width,cameraSettings.height,cameraSettings .fps); cam.setQuality(cameraSettings.bandwidth,0); cam.setQuality(cameraSettings.bandwidth,cameraSettings.quality); [Code] .....
We have an application that is using FMS to share slides (swf,jpg, etc) with another client. We added a video as an option. We would like to verify as often as possible the bandwidth quality between communications in order to suggest the user to close video if we notice bandwidth is not optimal.
We've been experimenting doing a bandwidth detection every 60 seconds, but I am wondering if there are specific guidelines to do this for live video, like how many times one should verify bandwidth? Is it ok to take one reading or should we take a few readings and then average out?
We are building a live RTMFP voice chat application with Cumulus. While the basic voice transfer works pretty easy using NetStreams, we have one big problem:There does not seem to be a way to manipulate the microphone data that the NetStream sends, and also not a way to manipulate the data the listening NetStream receives before it is played.However, this is exactly what we need. We do not want to transfer the normal microphone recorded audio, but first pitch it, then send it, then play it. Or first send it, then pitch it, then play it. But it seems that the whole audio recording, speex encoding, speex decoding and audio playback is completely encapsuled within the NetStream class.The only ways to achieve what we want (and all of them removing NetStream completely) seem to be:Send raw pitched audio data. That does work, but is of course a lot of data to send and will likely not work fast enough outside of our local LAN testing.
Pitch audio data, convert to ogg/mp3 using existing encoders for flash, send, decode ogg/mp3 and play. But this would mean encoding each and every sample packet that is received from the microphone, adding header stuff, etc. So this would likely not even yield that much of a benefit compared to raw audio data.2.1. This would actually be a good way if there was a Speex encoder/decoder for flash. But ironically, there is none other than the built in one (which is used for encoding/decoding audio in NetStreams) that cannot be explicitly used. ot offering it, Adobe...Send the data to the Cumulus server, pitch (and probably convert) there and send to the recipient. This would likely not even be that much faster than 1. and also throw away the exact benefit of RTMFP, P2P communication.
I have to install a multi-webcam setup that will be live streamed and accessible via a web browser.I'm looking into the quickest and most efficient solution, the less time and hassle it takes to implement, the better. That means buying hardware and/or software is not that much of an issue (as long as it's not 1000$+ flash server software).
The requirements / basic setup:The web server will be directly connected the cams (via ethernet / WiFi).The stream should be viewable by the max possible audience so no ActiveX / VLC-style object embedding. Flash would be ideal (even if it's a paid solution).Bandwidth is not an issue, max 2-3 clients at a time, most of the time only one.
Im Using Red5 to record live stream from the user. I am able to save the stream to disk and it generates an FLV file.
When I'm trying to read the FLV file with some FLV player - it seems like the file's meta-data is corrupted - which also makes the progress bar to act oddly. Also, I tried to play with the file using Xuggle. It seems impossible since the generated FLV do not have the codec-id in it.
Is it posible to convert a live stream into a flash format to save on bandwidth? If so, what player could be adapted to use it, and how could I convert it to begin with? If not, what is the best format to use for a live stream to get good quality without using too much bandwidth?
Is it possible to capture/encode a video file on a slow machine, then upload it live to a Red5 or Adobe Streaming server, and have that server stream to file live to flash clients? If so, what protocol is used to stream the data from the encoding machine to the Red5 / Adobe server? I see a lot of things about RTMP, but am I correct that this protocol is used between Streaming Server and Flash client, but not from encoding machine to Streaming Server?