Media Server :: Single RTMP Link For Multiple Servers?
Oct 3, 2011
can i use a single RTMP link included on FMLE with multiple servers, means, that RTMP link will do the redirection to the other RTMP links.with hight number of simultanous users, that feature will manage the use of servers
I have a case where I'm displaying 3 videos simultaneously using FMIS 3.5. When one is selected it will play the other two will switch and show a videoloop.
What is the proper way to accomlishing this. I actually have 6 videos (3 loops and 3 videos). Do I need a netconnection/netstream/video object for all of them or can they share something. What is the best practice?
I have some .mov files want to stream to Flash media server by ffmpeg. i have already tried to stream a single .mov by FFMPEG command in terminal and it works, the FMS can display the thing i streaming in live.Now i want to stream multiple files as one source, i tried to use above command one by one, but it seems Flash media server stop the streaming when file1 is finished, then start the stream with file2.It makes the stream player stopped when file1 is finish, and i have to refresh the web page in order to continue on file2.i am calling the FFMPEG command by a C program in linux, i wonder is there any method that i can prevent the FMS stopped when i switch the file source in FFMPEG? or is that possible to let FFMPEG constantly deliver the stream by multiple files source without stopped when a file finish
Is it possible to allocate multiple media locations to a single VOD application? I'm about to exceed the 2TB volume limitation of my SAN, yet the quantity of content that needs to be available for a given clients FMS application exceeds that.
If I'm streaming live video and chatting in an FMS app, should each component have its own instance of NetConnection or should they share one? Was thinking perhaps the quality of the live video stream would be hurt in a sharing situation but am not sure the architecture works like that.
From a development standpoint, I'd like to make one NetConnection instance in my main application file and connect with it. Then, when successfully connected, I'd like to pass that connection into any custom components that need it.
I have 3 seperate .swf on a single html page which are all communicating with FMS on separate connections. Do you think there will be any stability issues? or connection issues. All the .swf use the same main.asc. There are no apparant problems so far just need an expert opnion.....a side note is I've also noticed an error in the error logs that says its detected a backwards time stamp when recodring. Whats that all about? The application seems fine. Using FMS 3.0
I'm trying to make a software which sends video and audio data to a flash media server by using RTMP protocol. Currently, my program can communicate with a flash media server correctly. RTMP specifications does not describe about the raw data in video/audio messages, so I muxed raw H.264 and AAC data into video/audio messages and sent to the server. The server seems to accept them, but a video player cannot playback the stream sending from the server. The player just says "Loading..." For a test purpose, I sniffed the network packets between Wirecast and the flash media server and ripped off only video and audio data. Then, I muxed those data into video/audio message and sent to the flash media server. In this case, the video player connected to the server can playback the stream correctly.
I checked the stream sent from Wirecast, the stream seems not to be H.264 raw data because those data are not started from 0x17 instead of H.264 start code. With those situation, I am wondering what kind of container format I should use for H.264/AAC data to the flash media server.
I need to merge multiple live audio streams into a single stream so that i can pass this stream as input to VOIP through a softphone.For this i tried the following approach:Created a new stream (str1) on FMS onAppStart and recorded the live streams (sent throgh microphone) in that new stream.
Below is the code : application.onAppStart = function() {
Does anyone know if this is possible I have an FMS server in our production enviroment that contains multiple vHosts, at the moment I have one edge server setup for each host based on the comment at the bottom of this entry:[URL]
I am using Flash Media server 4.5 and i read the tutorial if i want to stream the live feed, i may need to use the media live encoder. but what i found in media encoder is i have to manually setup everything and it only support camera devices. But in my case i have multiple video files keep received from another program and place it on file system (server),my goal is use the Flash Media server to perform a live boardcasting with these video file one by one. That means when client watching a live streaming, they will not notice the server is playing mov1, then mov2, then mov3, then mov4... and so on.
You can imagine i am trying to boardcast a live footage say for 60sec, but the video file will not recorded entirely after 60sec, instead for every 10sec i will save a new video file, so that when client watching the live by HLS [URL]when the time reach to 10sec, a mov1 video file available and FMS should boardcast this video on live123.when the time reach to 20sec, a mov2 video file available and FMS should Immediately follow the mov1 boardcast on live123.and so on...Also can FMS dynamically create a new streaming session (invoke by code), so that when client A uploading some video files to the server, the FMS open a new streaming session only stream cilent A video files?the configuration to boardcasting like screen size, bit rate, etc should be pre-defined on the server. [URL]
I'm running Flash Media Streaming Server and have only been serving VOD up until now. I had my network administrator open up port 1935 to the outside world during the setup process and now I can't remember if that was actually required for streaming VOD to clients. Most documentation I've read says that this port should be open, but I seem to recall reading something at one point that suggested it wasn't necessary.
I've just started messing around with publishing live streams using Flash Media Live Encoder to the Flash Media Streaming Server. I have that working without issue but was surprised to find that no authentication is required before a client running the live encoder can publish a stream to the Flash Media Streaming Server. An authentication module is available however it only works with Flash Media Interactive Server and Flash Media Development Server.
If I leave port 1935 open to the outside world, there would be nothing to stop anybody anywhere from streaming video via my server. Anyone else running a default install of Flash Media Streaming Server and with port 1935 open to the outside should see that this is true of their setup as well. I'm wondering if I can safely close port 1935 without limiting the functionality of the server or if there's some way I can require authentication prior to publishing a live stream even though I'm not on the four-and-a-half-times-more-expensive edition of the product.
We had FMS2 installed before and the paths to all our videos are like rtmp://ServerName/sites/.... (the default path on FMS2) Now we upgrade to FMS4 and we would like to keep these paths the same because we have many HTMLs that reference these videos. However, the default path on FMS4 is rtmp://ServerName/vod/... Is there a way to change "vod" to "sites"?
I tried to change VOD_COMMON_DIR in fms.ini from /install_dir/webroot/vod to /install_dir/webroot/sites, and also changed the document root in httpd.conf, but rtmp://ServerName/sites/ is still not working.
I am trying to load test my 2 server load balanced FMS 2 environment which has been set up with the profile for 1000 connection 40MB. When I hit about 140 connections/server it fails with a "Bandwidth Penalty activated at time 1248945759.731000 with bandwidth 5945889 and limit 5000000" error logged in the event viewer immediately followed by many rejection errors: " immediately followed by many rejection errors: " Connection rejected by server. Reason : [ License.Limit.Exceeded ] : (_defaultRoot_, _defaultVHost_/DLvideo/_definst_) : Current server bandwidth usage exceeds license limit set. Rejecting connection."The load testing mechanism is a test harness that was created by a third party and seems to work well up until the above messages are logged. It basically is making 25 sustained streamed video connections per minute until it starts disgarding connections as noted above.
The servers are not even breathing hard (RAM is less that 50% consumed and CPU Utilization is < 5%) 1GB NIC on each server.What parameter(s) can be changed to push the servers harder.
I have a load balancing problem for example I have two FMS servers I want to make aconnection between them if the first server overload the second can handle the request
I run a FMS 3.5 Server and I want to add additional FMS Servers and load balance the traffic. The plan is to use a script for the developer version of FMS that forwards two incoming FME streams to the different servers, so each FMS Server has all streams. Is there a script for this? In addition there is a need for a script that somehow redirects the users to the server with the lowest users/bandwidth in use.
I'm running FMS 3.5 on RedHat, and have some server side actionscript that is attempting to perform an HTTP GET on a remote URL, e.g:
loadVars.send("http://someurl/");
This works on a Windows XP development server, but not on the RedHat deployment server. Are there any configuration options that I need to be aware of in order to allow my application to perform requests such as these (i.e. HTTP requests to remote servers)?
We are currently deciding what to use: Wowza(not so much, since they are undergoing the lawsuit) or Adobe Flash Media Server. The problem is that I couldn't find any information on how to configure the loadbalancer using FMS. I tested with Wowza and they use built-in loadbalancer but not so for FMS. Forgot to mention that we need it mostly for live events using RTMP protocol. My current configuration has 1 origin server + 2 edge servers(trial developers edition). I can manage to play livestream from the edge--origin servers using egde url. How can I configure the edge server so that when I connect to the origin server it redirects me to the least loaded edge server?
I tried to set up the connection limit for the origin server for 2 connections(one for the fmle, one for edge) and connect with 5-10 clients but my origin server didn't redirect the stream to the edge.
I need to do an in-place upgrade from 3.5.2 to 3.5.3 on my origin servers. Is there a recommended precedure to acheive this? I want to have the minimum downtime possible. CentOS, x64. I think something along the lines of :
I want to implement an internet chat application, I mean, I want a front end, which connects a web cam (mic and speaker). So, I can send videos and audios to a back end server (do flash streaming? but I am not sure what good product out there I can use), and also send both A/V to another computer on the internet. In other words, need a front end that sends local a/v from webcam (&mic) to backend and receives remote a/v from backend to webcam (and speaker) + backend streaming server.
i have a small LAN of about 8 computers all of which are running windows 7. I have installed FMS and XAMPP webserver on one of the machines. I want to stream live from one PC to all the other PCs on the LAN. I have a webpage with jwplayer embedded in it on my XAMPP webserver that is able to see the live stream when i start it locally. I mean the live stream works fine on the machine with the servers on it. But when i want to view the live stream from another machine in the LAN by accessing the webpage that has the jwpalyer from another machine, The jwplayer returns "server not found:rtmp://192.168.10.1/live" error. I was thinking that maybe a firewall is blocking the 1935 port but i have turned off the firewall of every PC on the LAN. I have unistalled any antivirus program on all the PCs. But i still get the same error when i try to access the live stream from another PC on the LAN.When i run netstat -a -n|find ":1935" i get 192.168.10.2:49184 192.168.10.1:1935 SYN_SENT and i think the request for the stream is sent but the conection is rejected.
This is the code for the webpage with jwplayer embedded in it. maybe it: <html>head> <title>JW FLV Media Player</title> <script type="text/JavaScript" src="swfobject.js"></script>
Somebody I know can't watch because they have a proxy server.The nc.connect() call to get the live stream fails because it goes direct from AS3 to FMS and doesn't go through the proxy server.
i am new on steaming & flash server; when we try to use RTMP over HTTP the outside client gets the internal IP address of the FMS server instead of the NAT one or public IP address, how can we solve this.
I was previously using FMS 3.5 to record from flash media live encoder using the DVR app. This worked well as I was able to seek into a live stream, all of it not just a window, and I was using RTMP for playback (not HTTP).
Now I would like to upgrade to 4.5, and keep using RTMP for DVR BUT add support for iOS. So my question is, is this possible? I've looked over the documentation many times and cannot figure it out. I see livepackager and that appears to support HLS but it also requires using adobe dynamic http streaming and it says nothing about true DVR (yes actually recording it).
So is the dvr app now obsolete, and if not how can I add iPhone support to it? I already have a custom player and really hope I don't need to rewrite it to use adobe http dynamic streaming just for this.
I am trying to encode and stream live video. I have downloaded both the Flash Media Encoder and Flash Media Server. To complete the process of streaming video, it appears that I need to obtain a "flash media server URL address" which is called "RTMP" on the encoder page. Where can I find this? Is this something that can be downloaded, or do I have to purchase this from a partner like Level 3 communications, AT&T, etc.?
i am having a serious issue with finding my RTMP address which may be something very simple to you all out there, i just installed Flash Media Server 3.5 at C:Program Files....i have purchased a video flash chat program and it needs to know the rtmp address which im not sure. I named the Flash Server EazyFlash.....not sure if i should install flash server somewhere else on server so website can see it??? The flash console and sample videos are working great but its all C: Drive??? would it be RTMP://EazyFlash? The software came with flash application which was a folder called videoflashchat which i copied to applications folder.
i re-stream one rtmp from server1 to 5 or more FMS servers, and make sure they are always connected if they lose connection for some reason it should keep trying to connect back.Please note i'm not a coder. I need simple example of how i can do this without hiring someone to do so.