I was told by my FMS host that even though FMEL can create multiple streams at different bit rates I could not send them. I need to use a different entry point for every stream which meant a different encoding machine for each rate.
The article Live dynamic streaming with Flash Media Server 3.5 mentions two different bit rates for each video size type.But why? Are these bitrates related to different compression ratios? And if so, why exactly two?
I want to create folders inside of VOD that I can use to store FLV in and them stream them to webpage.When I do this the webpage displays a white screen.It doesn't seem to find the FLV.
Just like the title says I am using Flash Media Live Encoder and I have three streams. The encoder is sending the streams just fine but I am not sure how to deliver it to the client. The only one that shows up is "livestream1". "Livestream2" and "Livestream3" won't even come up.
I have read in FMIS 4 "new features" that "absolute timecode" allows to switch audio tracks while playing a video (managing synch). Is there any exemple showing how to use this funtionnality (server config, flash player action script exemple)?
I need to merge multiple live audio streams into a single stream so that i can pass this stream as input to VOIP through a softphone.For this i tried the following approach:Created a new stream (str1) on FMS onAppStart and recorded the live streams (sent throgh microphone) in that new stream.
Below is the code : application.onAppStart = function() {
I'm using the Niagara 4100 streaming appliance to encode a live multiple bitrate streams for an event. The problem is the Niagara encoder doesn't allow me to attach an event name to the stream url like - 'streamevent%i?adbe-live-event=liveevent', instead it only allows 'streamevent%i'. So in the end I get three seperate events with a stream for each event.
Like so: Example: /Events/streamevent1/MNMMNNMxxxxx.stream /Events/streamevent2/MNMMNNMxxxxx.stream /Events/streamevent3/MNMMNNMxxxxx.stream AND /Streams/streamevent1/files /Streams/streamevent2/files /Streams/streamevent3/files
Is it possible to make it work this way without using the 'adbe-live-event' parameter using the .f4m manifest file?
I am using Flash Media server 4.5 and i read the tutorial if i want to stream the live feed, i may need to use the media live encoder. but what i found in media encoder is i have to manually setup everything and it only support camera devices. But in my case i have multiple video files keep received from another program and place it on file system (server),my goal is use the Flash Media server to perform a live boardcasting with these video file one by one. That means when client watching a live streaming, they will not notice the server is playing mov1, then mov2, then mov3, then mov4... and so on.
You can imagine i am trying to boardcast a live footage say for 60sec, but the video file will not recorded entirely after 60sec, instead for every 10sec i will save a new video file, so that when client watching the live by HLS [URL]when the time reach to 10sec, a mov1 video file available and FMS should boardcast this video on live123.when the time reach to 20sec, a mov2 video file available and FMS should Immediately follow the mov1 boardcast on live123.and so on...Also can FMS dynamically create a new streaming session (invoke by code), so that when client A uploading some video files to the server, the FMS open a new streaming session only stream cilent A video files?the configuration to boardcasting like screen size, bit rate, etc should be pre-defined on the server. [URL]
I have Flash Media Streaming Server 3.5 (not Interactive) running on RHEL5.5 x86_64 Linux.All is working well, however how do I prevent unauthorized access to connecting to the live stream and streaming content?How can I setup the server to require a user and password to stream live media to the server?I am new to this product and I have been reading some documentation but I have not found a clear cut answer on how to force a username and password to connect to the server to stream live content only.I am using the Adobe FMS Apache install, what files need changing?[code]I want to lock down a person from connecting to the server on the public internet and starting a live stream?Can this be done with a user name and password?
when i try to live stream with FMS! I can stream video with Flash media live encoder to the server but when i create the player to recieve the livestream from server,i can not recieve the live stream,can anyone give me a step by step tutorial of how to do it?
Only just getting started on this whole domain of learning, so go easy!If I set up a P2P video/audio chat (similar to the sample VideoPhone thing on the Cirrus site), can I get the stream from both parties to send to a server at the same time so that I can record it? If so, would I have to use a FMS to stream it to and perform the recording (and if so which version could I get away with)? Are there any (preferably free, or just tutorialised) solutions for the recording side of things?
Currently it seems like the only option for doing the P2P thing is to use Stratus/Cirrus unless I use FMS4 Enterprise.
how effective this kind of situation can be, in terms of quality of the stream and recording? Does any of this make sense?
I've had FMS running on my local machine for a while and have had a little experience writing FMS apps, but I've just tried recording audio for the first time using the standard vod application and I keep getting a "Write access denied for stream" error. My AS3 code is copied and pasted for various examples and am confident that it works.
I'm running Windows XP service pack 3 & FMIS 3.5.
I've had a look at the vod/media directory and under windows->properties the read-only attribute is ticked. Every time I un-tick this it reverts back to being ticked. I've googled this and MS say that most programs ignore the read-only attribute and that it only really applied to files. I've also tried the MS fix for setting the read-only attribute via cmd and still no joy (doesn't fix read-only attribute or FMS recording the audio after setting via cmd).
I've also tried our dev server install of FMS (running under linux) and am getting the same results.
Here's my AS3 code...
private function initApp(event:Event):void { removeEventListener(Event.ADDED_TO_STAGE,initApp);
i test the fms 4 update 1 rtmfp streams multicast after 10 minutes i get this message RTMFP Multicast stream has exceeded max duration allowed; closing stream. but i do not use IP multicast
I build a client side application where is only a FLVPlayback2.5 component and a short AS3 script.
[Code]....
My Encoder is setup with three streams: Vid: 500 kbps - Audio: 48 kbpsVid: 800 kbps - Audio: 48 kbpsVid: 1500 kbps - Audio: 48 kbps I start the encoder and everything looks fine in the log. In my browser (Safari or Firefox) I go to my html site and the stream starts after 6-8 sec. But anytime with the lowest bitrate 548 kbps and nothing look like the stream is switching to another bitrate. I tried it with the smil playlist and the result is the same. Only the lowest bitrate is plublished.
I have recently installed FMIS 3.5.3. In checking the access logs I find data in both logs that display the same stream stop and stream play time .I'm not sure why the time is the same (00:19:27 example below). Videos play fine when testing from work (T3 connection). However, occasionally a very slight hesitation when playing video from home (I have cable connection). [code]...
I'm having a problem with recording a live webcam stream. The last few seconds of the stream is getting cut off. The recording is stopped with the following piece of code:
I'm trying to stream a HDS live multi-bit stream, it seems to push to the FMS but my player doesn't display the stream.Are these settings and files correct? The documenation is confusing on what and which files need to be edited and/or created.
Encoder settings: Bit Rate: 150,500,700 FMS URL: rtmp://myserver/livepkgr Stream: liveevent%i?adbe-live-event?liveevent
FMS 4.5
I see the following directories being created when I start encoding and each directory has a single file with a .stream extension in them. Are these correct? fC:FMS-HOMEapplicationslivepkgrevents\_definst_liveevent1[code].....
I'm new to Adobe Flash and I'm just trying out the Flashe Media Server 4.5 to stream live video on a local machine. I'm using a FLME 3.2 to capture the video from my digital camera, which works fine, and I can connect to the server with no worries as well. But I have issues streaming the content in the Sample Video Player that comes with the server.
[Code]....
I've also tried playing some sample videos on it. it even refuses to play those. What could be the problem??
One server - is streaming server with Flash Media Interactive Server 3.5that host the application with asc files in FMS application directory.Second server - is the IIS web server that host thehtml,aspx,swf etc.. files.So basically i heve swf file on one server that have to connect to live streaming via rtmp on different server with different ip address.I did not find any clear explanation on this crossdomain rmtp issue.ys only HTTP can use crossdomain.xml policy filebut not rmtp.So what policy security procedure need to be done in order to enableswf file hosted on one server to connect and show rmtp live cam broadcasting from another server?
I am trying to create a flash player to stream an on online Internet radio station using FMS. I chose FMS after being told that I need RTMP server to extract the metadata. Our radio audio is being encoded using a DJ interface called SAM Broadcaster. But for some reason, it does not have an option to send over the stream directly to a Flash Media Server. Only options are either IceCAST or ShoutCAST. How should i setup the stream? Should it be Stream encoder>> IceCAST/SHOUTCast Server>>Flash Media Server>>Flash Player Client. Or should I setup Stream Encoder>>Flash Media Server>>Flash Player Client? Shouldn't Flash Media Server be an alternative to IceCast or Shoutcast? If that is the case, how do I send over the stream from SAM Broadcaster directly to Flash Media Server without restreaming through another streaming server?
'URL missing from Media tag' on FMS4.5 HTTP Live Stream playback.I have spent a whole day trying in vain to publish f4m live http streams from FMLE using FMS 4.5, and almost everything I try results in the error .The F4M document contains errors URL missing from Media tag...in both Flash Media Playback and the fms videoPlayer app.Other people seem to have seen this and solved it, but I've found nothing that helps. I have tried to publish a single stream, and with multiple bitrates. For the latter, I followed the Adobe video tutorial as follows:
Multiple bitrate live ================ I'm publishing 3 streams from FMLE, using FMS URL: rtmp://[serveraddress]/livepkgr[code]..........
What would be some top reasons an FLV will not stream with Flash Media Streaming Serve 3.5? I ran the FLVcheck.exe on the FLV and it passed; however, the FLV will not stream. If I could just get some good reasons that would cause an FLV from streaming I think that will give me a good starting point for troubleshooting.
Beside Flash Media Encoder, can I use a live stream from Windows Media Services?We have live streaming with Windows Media Server, but now want also offer to mobile devices, but we dont re-encode the entire streams (tvs) again, it would be great if we can use a stream from windows to tunnel it to flash media server.
Is there a way to stream data instead of media files? In particular I'm interested to stream swf files that will be dinamically loaded by another swf client. I see this interesting message talking about how to do this on Wowza but I prefer to find a solution also in FMS to use the SWF Verification feature.
I have FLV files streaming fine from the media folder in the applications folder.I want to split up the videos into folders to make them more manageable.If I create another folder in the applications folder at the same level as the media folder and put a video in there the video does not play.
I have an *.flv file on a FMS. When I play it on the client side the video plays just fine, but when I call Stream.play(filename, 0, -1, false) on the server side the video turns out really choppy.I both cases I use NetConnection to connect to an rtmp and NetStream to play the stream, but in one case I connect to a stream and request the server to play my file on that stream. Apparently that doesn't work with files? It works just fine for live streams.