Media Server :: Reconnecting Live HTTP Dynamic Stream Via FMLE 3.2?
Feb 22, 2011
As seen the tutorial [URL].. I have successfully implimented the steps. However the problem comes when the stream is disconnected and reconnected again due to network problems. Once the stream is reconnected the video stops coming, even though all 3 streams are succesfully publishing the stream to Flash Media Server but on the client side where I'm using OSMF media player the stream gets stops. Is there any setting that I need to impliment in the OSMF player to resolve this issue ?
'URL missing from Media tag' on FMS4.5 HTTP Live Stream playback.I have spent a whole day trying in vain to publish f4m live http streams from FMLE using FMS 4.5, and almost everything I try results in the error .The F4M document contains errors URL missing from Media tag...in both Flash Media Playback and the fms videoPlayer app.Other people seem to have seen this and solved it, but I've found nothing that helps. I have tried to publish a single stream, and with multiple bitrates. For the latter, I followed the Adobe video tutorial as follows:
Multiple bitrate live ================ I'm publishing 3 streams from FMLE, using FMS URL: rtmp://[serveraddress]/livepkgr[code]..........
I want to use customised encoder to publish live steams to a FMS4.5 server?Is it possible to connect encoder and FMS4.5 server on HTTP protocol?i read in other articles FMLE can connect to FMS on RTMP protocol.
I build a client side application where is only a FLVPlayback2.5 component and a short AS3 script.
[Code]....
My Encoder is setup with three streams: Vid: 500 kbps - Audio: 48 kbpsVid: 800 kbps - Audio: 48 kbpsVid: 1500 kbps - Audio: 48 kbps I start the encoder and everything looks fine in the log. In my browser (Safari or Firefox) I go to my html site and the stream starts after 6-8 sec. But anytime with the lowest bitrate 548 kbps and nothing look like the stream is switching to another bitrate. I tried it with the smil playlist and the result is the same. Only the lowest bitrate is plublished.
I read all information from above and related all posts.Now my question/problem is,Can we use DynamicStream to access and handle live stream user's web camera instead of normal NetStream?Our application is like video conferencing.In our application user's stream is not being stored on server. It is just being published as live.Now when I publish user's webcam stream as "record" (in ns.publish("stream_name","record"))I am able to get the stream.But when I publish the stream as "live" (in ns.publish("stream_name","live"))the stream is not being retrieved. Even sometimes I get an exception in "DynamicStream.as" provided by Adobe
I read that it is best to de-interlace video "at the hardware level" before sending it to FMLE 3.2 for encoding. I have looked and looked for a software or solution but can't find anything that I can work. Is it best to de-interlace before sending the video to FMLE? I am streaming live video that needs de-interlacing before sending to our site. I know there is a "de-interlace" option in FMLE 3.2 but read that it was better to de-interlace the signal before sending to FMLE. I found something called DScaler but don't know how to get the video from DScaler into FMLE 3.2.
I have FMIS 3.5.I've installed it with pretty much all the default values.I haven't changed any of the settings in either the LIVE or the VOD applications.Dynamic Bitrate Switching is working well for VOD, but not working at all for LIVE streams.Doing a regular bandwidth detection on both the LIVE and the VOD applications give similar, high bandwidth results. However, the LIVE application NetStreamInfo.maxBytesPerSecond is showing a very low bandwidth capability of around maxBytesPerSecond = 19016, where as for the VOD its achieving around 637110.I can play a single LIVE stream of a high quality smoothly without any error.
I don't know if this is relevent but I'm getting some error messages occassionally in the log of the live application saying : Dropping application (live/_definst_) message. Clients not allowed to broadcast message.These messages aren't consistent, and don't coincide with trying to use bitrate switching.I have tried downloading the Adobe sample StreamSwitching. fla and it won't play the LIVE streams at all.Using the opensource Longtail Video player it just always defaults to the lowest stream.Here is an example: http:[url]....
if I repeat this IIS topic, I couldn't find the answer to my problem any where on the internet. I installed FMS Dev 3.5 on Win 2003 Server with IIS 6.0 enabled. I don't have any issue with port 80 Listening, I used the IP address 192.168.0.21 for my web application (IIS) and 192.168.0.22 for FMS (I only have one network card and port 1935 is open under firewall). I can play the sample videos (RTMP, HTTP, and Dynamic Stream) using the Flash Media Start Screen (or from the location C: Program FilesAdobe Flash Media Server 3.5webrootindex.html) without any problem.
I then modified the IIS Default Website to look at the "webroot" folder (C: Program FilesAdobe Flash Media Server 3.5webroot). From IE, I can access the default web site by enter http://192.168.0.21/index.html. The website loads up correctly, and the RTMP video is playing perfectly. However, if I click Play Video (HTTP) or Dynamic Stream (tab), I receive "Connection Error. Please press Play to try again." I look at the log file (access.01.log) and see the error log "Session disconnect
So I have FMLE and FMS. to record and stream video at the same time. On server side I have "main" application.onPublish = function(client, potok){ potok.onStatus = function(info){ for(var i in info){ trace(i + "=" + info[i]); [Code] .....
I need to take the some stream which is coming from firewire on mac, and then transcode/encode it to perticular bitrates, then stream.I think from firewire to FMLE, its straight forward. How can i stream? How can i give my clients ( say from mobile, or from any player as a http url streaming) stream? Do i need to use Flash media server, if yes can anybody point out the steps to set-up the same.
I am using FMLE and fms 4 ent. When using fmle to broadcast webcam without audio. It goes really fast. I am on 1gbps connection with nothing else coming to the server at moment. I tested download speeds to and from the server and they are around 700mbs. I have tried all the different audio sample and bit rate settings. All have the same result. Once audio is turned on there becaomes a 15 second delay. What am I doing wrong.
i was modified the main.asc script in livepkgr, it used to call s.play(streamObj.name), which i assume this is the code it play the stream from FMLE to s.play("FileName") , which is a flv video file in my application. unfortunately when i start the FMLE stream to the FMS and watch the stream with the FMS Sample Video player, i was expecting to see the video content from the flv, but instead the camera footage in playing on the player.
i am very confusing now, i thought the s.play the only code to control what should be playing in application. or when FMLE is connected to server, the server must play the footage from FMLE by some hidden code, no matter what s.play wanna to play?
The auth_addin works as expected when using the Flash Media Live encoder requiring a username and password before publishing a stream.However, tools like Wirecast and even the sample broadcasting SWF that comes in the FMS Introduction page can publish a stream without providing any credentials!
I have Flash Media Streaming Server 3.5 (not Interactive) running on RHEL5.5 x86_64 Linux.All is working well, however how do I prevent unauthorized access to connecting to the live stream and streaming content?How can I setup the server to require a user and password to stream live media to the server?I am new to this product and I have been reading some documentation but I have not found a clear cut answer on how to force a username and password to connect to the server to stream live content only.I am using the Adobe FMS Apache install, what files need changing?[code]I want to lock down a person from connecting to the server on the public internet and starting a live stream?Can this be done with a user name and password?
I'm having a problem with recording a live webcam stream. The last few seconds of the stream is getting cut off. The recording is stopped with the following piece of code:
when i try to live stream with FMS! I can stream video with Flash media live encoder to the server but when i create the player to recieve the livestream from server,i can not recieve the live stream,can anyone give me a step by step tutorial of how to do it?
I'm trying to stream a HDS live multi-bit stream, it seems to push to the FMS but my player doesn't display the stream.Are these settings and files correct? The documenation is confusing on what and which files need to be edited and/or created.
Encoder settings: Bit Rate: 150,500,700 FMS URL: rtmp://myserver/livepkgr Stream: liveevent%i?adbe-live-event?liveevent
FMS 4.5
I see the following directories being created when I start encoding and each directory has a single file with a .stream extension in them. Are these correct? fC:FMS-HOMEapplicationslivepkgrevents\_definst_liveevent1[code].....
We tried to stream from FMS 3.5 flv files through HTTP,How can we prevent Progressive download when we stream flv via HTTP?We tried some configuration but it seems that files are always download to the local computer temp folder Can someone point to documentation that explain how to stream flv without Progressive download ?
I have FMS installed on windows server 2008. I have a large number of videos which are served to my clients using rtmp protocol.Now we have designed a new player with some controls which apparently do not work while an rtmp video is playing.
I've read suggest naming the stream in the FMLE to something like mp4:mystream.mp4 or mp4:mystream.f4v but then you can no longer view the stream, and according my the admin console the stream will not even publish using this syntax.
I can view the published stream if I name the FMLE stream to something like mystream.mp4 or just mystream, but then it doesn't record at all or records as a FLV file depending on how I code the main.asc file.
i am wondering is it possible to get embeded timecode from HTTP Stream?Flash Media Live Encoder contains a special built-in handler, onFI, that subscribing clients can use in their ActionScript code to access timecode information. The following client-side ActionScript code shows how to get timecode information using the onFI handler. The object ns is the NetStream object. You can get timecode and system date and time information, if timecode and system date and time were embedded in the stream, by accessing the tc, sd, and st properties of the info object that is passed as an argument to onFI():i am using this onFI method succesfully with netStream object when i am getting stream from RTMP.is it possible to get somehow same results with HTTP Streaming (f4m and osmf)
Beside Flash Media Encoder, can I use a live stream from Windows Media Services?We have live streaming with Windows Media Server, but now want also offer to mobile devices, but we dont re-encode the entire streams (tvs) again, it would be great if we can use a stream from windows to tunnel it to flash media server.
One server - is streaming server with Flash Media Interactive Server 3.5that host the application with asc files in FMS application directory.Second server - is the IIS web server that host thehtml,aspx,swf etc.. files.So basically i heve swf file on one server that have to connect to live streaming via rtmp on different server with different ip address.I did not find any clear explanation on this crossdomain rmtp issue.ys only HTTP can use crossdomain.xml policy filebut not rmtp.So what policy security procedure need to be done in order to enableswf file hosted on one server to connect and show rmtp live cam broadcasting from another server?
Somebody I know can't watch because they have a proxy server.The nc.connect() call to get the live stream fails because it goes direct from AS3 to FMS and doesn't go through the proxy server.
My rtmp client connects to the server ok and rtmp data flowing ok for about 4hours, in the wireshark trace I can see that the server just stops sending data roughly after 4hours. The server version is 3.5.1, and I also try it on a 4.0 server the data stops coming from the server as well after 5+hours. There are no errors or any indication of session disconneciton in the diagnostic logs. I have also looked thru the settings in applicaiton.xml and server.xml I just don't see any obviious reason for the issue. I do send rtmp Acknowledgement to server after the client receives "window size" bytes, and the client sends ping response to server.
We purchased Flash Media Streaming Server Interactive edition in order to have the security featuers of not allowing unauthorized connections streaming live content to the server. Is there a document with procedures that details how to make this happen, the allowedHTMLdomains.txt does not work when you place in IP ranges that you do not want connecting. I need some help just to figure out a way to keep someoen from connecting to the server and not streaming live content. Has anyone configured this in Linux because the documentation is mainly for Windows. Any help/advice would be great since this seems to be a common problem for anyone who runs Linux.
I am trying to config custom encryption key for different pieces of content at streaming level using jit.conf for On-demand Apple HTTP Live Streaming.I tried to set "HLSEncryptionScope content" in httpd.conf and placed jit.conf in ../webroot/vod but not success, the m3u8 generate not contains EXT-X-KEY.and, I tried and got nothing in access plugin log file while i http stream a m3u8 file, is it possible to log HTTP Live Streaming using plug-in?can plugin got any event triggered by http streaming?