Media Server :: Severside Script Not Work For HLS?
Feb 16, 2012
I am using some severside script ( create a main.asc under application directory) for streaming some pre recorded video files as a live to client. You can think of using a video file for live stream source rather than FMLE .
I follow this helpful instruction in this [URL]..It works perfectly fine if I watch it in rtmp.
But I wonder is this work for hls-live? Right now I am try to stream the content to an iPad.I tried to input the hls-live format URL to the browser URL bar, but seems the server cannot found such live stream event.Such as http://192.168.1.106/hls-live/sstream/a/version1/clientPlayStream
I checked the folder used for hls vod or live, it looks like the documents needed for hls is more than a simple rtmp application. Anybody knows the severside script application is work for HLS ?
i have FMS Installed And after its installed i also Install The Addon Auth to make a username and password so no one else can just start a Stream but after i install it and restart the server it wont work any the fms does work but the addon doesn't.
i have a orgin server and 2 edge servers and a load balancer
first off im confused in each edge server i set them to be remote, rtmp and changed the routeentry to <RouteEntry>*:*;96.44.***.***:1935</RouteEntry> on both servers
orgin server is just local
what my question is im streaming to orgin and its getting sent to 1 edge and other edge is doing nothing no traffic the stream i set it this way url:
'rtmp://173.254.***.***/live/live_test',
which is the 1 edge thats working if i setup another channel pointing to the other edge server ip adresss it doesnt work.. what am i doing wrong isnt all this supposed to be load balanced or am i missing steps...or did i do it wrong on the flash player config.
I have a problem with flash media streaming server 3.5.0. I try to make subfolder under application/vod/media and put .flv file there, but it just doesn`t work from there. Example url should be [URL]
I'm using FMS 4.0 64 bit The way I create new VHost on default adapter:
1. Create new directory named for example foo.host.com and copy Application.xml and Vhost.xml from _defaultVHost_ dir.
2. Execute startVHost command through http admin interface or Restart Vhost in admin console.
New Vhost still does not work untill I Restart the Server (which is not acceptable). I had no problems with startVHost or restartVHost on new vhosts on FMS 2 - I did not have to restart the whole server to make new vhost work.
I was wondering if any one could clarify something with IP multicast. From what I understand to get IP multicast to work the server needs to publish to the router box. So how can this be achieved if the FMS server is not in the same network as the router box? Would it require a second FMS server to be placed in the same network as the router box that connects to the main FMS server?
I try to do live stream by FMS 4.5.1 and FMLE 3.2 (H.264&MP3 encoding) It's all good no matter video or audio. But after about 30 mins, stream freeze. Refresh doesn't help, only get a black screen without any error message.
restart service or reboot my PC still the same. The only way to fix it is uninstall FMS and Install it again. HDS works again....but same problem happens after 30 mins or a couple of hours.
My problem and I SWFVerification enabled on the server and saw that FMLE does not connect more, I looked for some explanation and I saw that if that is enabled for SWFVerification you must also change UserAgentExceptions tag <Exception from = "" to =""/> at this point I added this change<Exception From="FME/0.0" to="FME/4.0"/>but I saw FMLE does not connect as little to solve this problem
I am in the midst of coding a multiplayer card game. Right now, to make things simple, I limit myself to code for 3 players for one room. The clients call functions in the server and vice versa. Functions such as player's turn, shuffling cards, etc. So what happens if there's more than 1 room? Won't the data from the different rooms get mixed up with one another in the server side, since everything will be happening in just 1 server and on the same function? I am abit confuse on this part. Can someone enlighten me on how different rooms work in a multiplayer game? How do I make sure the data don't get messed up?
I have a media player on my website. Visualizer of media player works when I test it on xampp localhost but it does not work when I upload it to my hosting. My website: [URL]
I install the latest version of apache before the FMS4. After the installation completed, I copied webroot, applications and other files to the web sites home directory. I also updated the FMS.ini file in the config folder. Because I am not familiar with FMS, I don't know whether the change is correct or not.
I have application that uses the send functionally in the NetStream on the server-side. When a connect to the app using rtmp I am able to see the send come thru but if the app uses rtmfp it does notI gone thru all the examples in setting up a mutlicast app and I know the app works because when I publish a video out everyone in the group sees the video. The only thing not working is send. I've also set dataReliable to true.
I'm wondering if anybody knows of a guide or a step by step process for configuring a VBrick HS.264 encoder to send video to my FMS for live broadcast. I'm experiencing issues configuring the Transmitters on the VBrick admin panel and unsure how to have FMS recognize the video stream to publish it. Has anybody used a VBrick encoder with FMS?
For the past while, my developer has been working with me on a small project to put together a live interactive whiteboard app, powered by Flash Media Interactive Server. Well, everything is finished & ready to go, but there seems to be a big problem:
The app was developed & tested on a windows machine, while my server and all my website files (to integrate with) is on a Linux system.
It's been a month since this problem arrose, and there had not been any further progress with this project for the time. My developer doesn't seem to know much about how the 'conversion process' works, if there is one.
Do you know if this is possible (to convert a windows-programmed app to work on a linux system)? If so, do you know roughly how it should be handled?
It's a bit vague - the whole thing with AMF. On one hand we have a client capable of serializing and sending everything from integers to XML documents and ByteArray streams, but on the other hand we have the FMS with a JavaScript host. Where is the bridge? :-) I mean how does one work with the objects, in particular, a ByteArray, received through the [generic] wire? It's easy with numbers and strings, but no word on more complex objects.
I installed FMS about a year ago, and realized that it broadcasts videos in "real time", just the point where the video is, without a "progressive display and download".Maybe there is a mode in FMS that makes it work in "progressive display and download" way in HTTTP connections, just like youtube or any other tube, so, it will be downloading the rest of a video, while playing its previous part.
We are working with FMS 4 server for a while for a 2 directions video application, and it works great with rtmp. We now want to use its rtmfp abilities after we used Cirrus for testing in the last few days and it also worked well. Locally - everything is working fine, but when we try the application on the remote server - we have some problems.
[Code]....
Again, it is working beautifully both locally and with Cirrus. RTMP also works well with the remote server.
I had FMS 3.5 developer package installed on my server installed about a month ago. I used to be able to go to the /webroot/ page to see the sample RTMP video play and it worked great. Today, I go to view that page and find that it no longer plays. I get "Connection Error. Please preee Play to try again."
Pressing the play button does nothing, so I click the RTMP thumbnail and it reloads ony to give me that same error again. I can click the HTTP thumbnail and see the video play just fine. But, I want RTMP and after over an hour on the phone with my managed hosting tech, we can't find anything wrong.
Here's the thing: I haven't done anything, I have not changed anything, I have not manipulated the server software whatsoever. It just stopped working and gives me "Connection Error."
Can Adobe Flash Media Streaming Server 3.5 run on AMD Athlon Dual Sock Quad Core?I just requested a Dell server to be added to our farm to run as a Media Server and to my surprise, while reading the requirements for FMS it states the following: 3.2GHz Intel® Pentium® 4 processor (dual Intel Xeon® or faster recommended)
I'm trying to troubleshoot a Flash Media Server working with a little video playback application I wrote a few years ago that has suddenly stopped working.I'm using CS3/Actionscript 3.My app uses the FLVPlayback Component, and was working well last time I checked. I recevied a report that the videos stopped working, and have been looking into it.I figured I'd add a bunch of event listeners to the FLVPlayback's ncMgr.netConnection so I could get debug info on things like io errors, net status, etc.The problem I'm running into is that the netConnection is null when I set it to anything on my Flash Media Server, and adding any event listeners to this netConnection throws errors.Here's what I've tried so far:
Playback of a local FLV file works fine.In the FLVPlayback documentation, I found an example and stole the URL of the stream they were using in the example, and that works fine, although it is an HTTP protocol stream rather than RTMP.Any attempt to access FLV files on my Media Server, which has worked fine in the past, basically cause the FLVPlayback object to sit and hang in "buffering" mode and never progresses beyond this point.The netConnection object in this case is null.Here's my code:[code]........
Again the purpose of this is to troubleshoot the video streaming from the Flash Media Server, and it seems like there is no netconnection to the server being created.Does this mean that the server is not working, or is there a problem with the way I'm trying to access the content on the server?This was all working fine before, and I have set up the server-side application .ASC files and such to allow things to work fine on the server end.
I'm trying to make a software which sends video and audio data to a flash media server by using RTMP protocol. Currently, my program can communicate with a flash media server correctly. RTMP specifications does not describe about the raw data in video/audio messages, so I muxed raw H.264 and AAC data into video/audio messages and sent to the server. The server seems to accept them, but a video player cannot playback the stream sending from the server. The player just says "Loading..." For a test purpose, I sniffed the network packets between Wirecast and the flash media server and ripped off only video and audio data. Then, I muxed those data into video/audio message and sent to the flash media server. In this case, the video player connected to the server can playback the stream correctly.
I checked the stream sent from Wirecast, the stream seems not to be H.264 raw data because those data are not started from 0x17 instead of H.264 start code. With those situation, I am wondering what kind of container format I should use for H.264/AAC data to the flash media server.
I have to a problem using the Flash Media Interactive Server Feature Explorer. I want use the sample: RecordStream. I can see the instance "RecordStream" in console FMS 3.5. and show me the video in app AIR, but does not save the .FLV in my server.