Media Server :: Allow Server To Publish Only Audio?
Jul 5, 2011
My clients use Flash Media Live Encoder to publish the audio+video to my server. In some case, I want to allow my clients to publish only Audio and no video but I cannot control it via the FMLE, they can even select the video to publish.
I wanted to know if there is any way in which on the server side, I can allow my client to publish only Audio and no Video?
"To serve streams over a cellular network, one of the streams must be audio-only. For more information, see HTTP Live Streaming Overview.To publish an audio-only stream, enter the following in the Flash Media Encoder Stream field:livestream%i?adbe-live-event=liveevent&adbe-audio-stream-name=livestream1_audio_only&adbe-audio-stream-src=livestream1If the encoder specifies individual query strings for each stream, use individual stream names instead of the variable %i:livestream1?adbe-live-event=liveevent&adbe-audio-stream-name=livestream1_audio_onlylivestream2?adbe-live-event=liveevent&adbe-audio-stream-name=livestream2_audio_onlyTo generate a set-level variant playlist when using an audio-only stream, specify the audio codec of the audio-nly stream. Specify the audio and the video codec of the streams that contain audio and video.For more information about using the Set-level F4M/M3U8 File Generator, see Publish and play live multi-bitrate streams over HTTP.
I'm trying to make a software which sends video and audio data to a flash media server by using RTMP protocol. Currently, my program can communicate with a flash media server correctly. RTMP specifications does not describe about the raw data in video/audio messages, so I muxed raw H.264 and AAC data into video/audio messages and sent to the server. The server seems to accept them, but a video player cannot playback the stream sending from the server. The player just says "Loading..." For a test purpose, I sniffed the network packets between Wirecast and the flash media server and ripped off only video and audio data. Then, I muxed those data into video/audio message and sent to the flash media server. In this case, the video player connected to the server can playback the stream correctly.
I checked the stream sent from Wirecast, the stream seems not to be H.264 raw data because those data are not started from 0x17 instead of H.264 start code. With those situation, I am wondering what kind of container format I should use for H.264/AAC data to the flash media server.
As titled, what is the way to record video/audio files using Flash Meida Server through rmtp, and allow users to access the recorded files through http?What I am trying to do, is to record a user's microphone's input and save it to the server.fterwards, I would like other users to be able to access the recorded files and mainuplating the audio data, by computeSpectrum(), to do some visualization of the audio. As I know computeSpectrum() cannot work on streaming files, so I think I need to access the recorded files using http instead of rmtp. Is that true?
im looking for a script that can record audio from a mic and save it to a fms server i came across a video email script but would like to have just a audio one that can do the samething as the video email with out the video part.
I ran a simple live video streaming application for the first time with actual users and ran into a couple of serious performance issues that had not turned up during testing. In this instance there was one video stream from a live web cam and used FMLE at 150 kbps using VP6 and MP3 @22k. There were 16 clients and everything worked pretty good for about 30 minutes. (although some clients said their audio and video were out of sync by up to 3 seconds)
Then individual clients would have either the video freeze or the video would continue and the audio would stop. These clints had to "disconnect" and then "connect" again to the application. This happened to all of the clients at one time or another for several minutes. I stopped and restarted the FMLE with progessively lower bandwidth settings down to 75 kbps but still clients were having the same issue.
I eventually stopped the FMLE and used the applications built in publisher at 45 kbps and that seemed to eliminate the freeze/dropping issue. But of course the video quality was very poor and some clients still reported that the audio was out of sync with the video. The server hosting the FMS application is a quad processor dell with lots of memory and network connectivity. The Flash Media Admin Console performance graph showed the total Bandwidth as 3 Mbps at maximum.
have an error in the publication of a Live (Webcam) to a CDN (Level3).when "ns.publish (STREAM_NAME 'live')" 'it is returned an error"NetStream.Record.NoAccess 'Thank you to those who use AS3 application to connect to their CDN (Akamai, Level3 ...) oftell me if they encounter a problem.I use the same code has always been to publish a Live (Webcam) to serversstreaming (FMS RED5, etc ...) and can be summarized as follows:nc = new NetConnection ()nc.connect (HOST_URL)../..ns = new NetStream (nc)ns.publish (MON_STREAM 'live') this works on all versions FMS 2.5, 3.0, 3.5, etc. ..Since a migration of my FMS version 2.5 to 3.5 this code does not work anymore!the worst is that Flash Media Encoder (FME 3.0) it is operating properly.So is there a feature to broadcast a live publishing point on a CDN (LEVEL3)?The CDN turning on the fact that if it works with FME is that it is the code that raisesproblem ... without more information.The FMS version uses "special" of FMS, but until now I had no worries!In general the user CDN use hardware encoders ...
i have problem with publish with fms.i have a song play in back i have microphone to sing and camera when publish in mp4 the voice is ok and camera good. But the song play in background dont have good quality.
1- can i separate voice and song in my microphone? if yes do you have an exemple? because the sound o f voice is ok but song in background not good
2- else can mix my original song mp4 to camera and microphone? and publish record all with fms?
i am building an live streaming website and i am use
1- FMS 2- Apache webserver
i have made the subcriber and i works well .but i need to build the publisher to allow users to broadcast thier stream from thier cameras.i have tested a publisher which was build using actionscript and it didn't work untill i have installed the fms on my localhost , but i need a publisher which any user can use in his web browser .
I want to build an application to stream h.264 video from the browser, how can I be able to do this? I can't seem to find the playerglobal.swc anywhere, I have Flash CS5.5
I need to protect my FMS, so that only me (my Username, or IP if its not possible) can start a live stream. I have been reading the pdf on plugin for developers, and i read that the authorization plugin can do this, but it is C++ and i dont know nothing about that, i can understand a little bit, but i do not know how to make the dll i want. Is there another way to do this, the server requests a user and password when i want to stream.
I am developing a flash media encoder. Encoder is sending 'publish' command in 'record' mode to FMS. But FMS is not giving any response (I am catching responses in wireshark) but creating the related folders in its application directory. So, I assumed that FMS has received the publish command and reay to receive the data. Then I sent some 20 media messages each of soze 500 bytes. FMS is sending 'onStatus' response for the publish after receiving the 8th media message and writing only contents of media messages from 9th to 20th.
i've created a sub-folder called dvr under fms 4.0 applications. i made a test of playing a video named sample.flv then at the same time i publish it to the same fms 4.0 server with a different video name: "mergedVideo".the asc at the fms server as following:
I have FMLE source streaming to FMS #1, and FMS #1 is multipoint publishing to FMS #2. It works great, but when I stop FMLE source, the multipoint publishing from FMS #1 to FMS #2 doesn't stop (unpublish), and continues to stream empty (blank black) stream between the FMS servers. So how do I ubpublish the multipoint publishing streams on the FMS #1? Here is my code in application live on FMS #1:
I've installed Flash Media Server and send stream to it use Flash Media Live Encoder via [URL]. And on my website, I've embed code to play this live stream via [URL]. All ok!
But, any user can install Flash Media Live Encoder and connect to my FMS, publish his/her stream (because url to publish and view is same) My question is: how to prevent end-user publish stream to my FMS, only allow end-user view my live stream?
I want to publish and record video stream in h.264 and audio stream in AAC format . Does flash player support to publish video stream in h.264 format ? And which version of flash player and FMS could implement this function ?
I have a requirement of publishing static content like video files mentioned in a playlist to be published to a live stream. It is like creating a TV station with pre-recorded content but with live stream. It will great if anyone let me know is it possible with FMS?
I know the FMS is good for streaming things video to clients and for clients to publish videos from their attached camera. Is it possible for a client to publish something other than a stream from a connected camera, for example, the contents of my swf stage? Is it possible to draw something like a text overlay onto a published stream?
I've searched quite a few threads:[code]But all of them are talking about playing multiple bit rate live video,my question is how I can publish multiple bit rate live videos in the first place?
That is , I don't need to terminate the current publishing video then republish again.I just want to change the publish mode,maybe from live to recorded mode.
I've noticed an issue while creating one way (A publish B play) stream through RTMFP. When I try to create connection between computers behind the same NAT (same IP) stream playing is not fired at all, but on the other hand the same code works on remote computers (different IPs).This issue arose after releasing final release of Flash Player, in a beta release everything works fine.We checked NAT settings and there are no differences for this cases.
I want to use customised encoder to publish live steams to a FMS4.5 server?Is it possible to connect encoder and FMS4.5 server on HTTP protocol?i read in other articles FMLE can connect to FMS on RTMP protocol.
I am developing a C++ program to connect FMS via RTMP, and then publish a live stream I can't use FMLE here because I need collect the H.264 NALU from somewhere and forward them to the FMS.I write a demo app, which parses H.264 encoded FLV file, and send the VideoTag in the FLV as RTMP Message content for publishing.Finally, I make it run and I can see some frames of video!But the problem is: The video does not play smoothly and it just updates some frames. Then I looked at the NetStream event and found that:Everytime when there is a NetStream.Buffer.Full event, the frame in the video is updated and display correctly But immediately there is another NetStream.Buffer.Empty event followed and the video is frozen.
<?xml version="1.0" encoding="utf-8"?> <mx:Application xmlns:mx="[URL]" layout="absolute" creationComplete="init();"> <mx:Script> <![CDATA[ import mx.controls.Alert; import mx.core.mx_internal; import flash.media.Camera; [Code] ..... So, I play the record video , but there is no image , just voice
I want to make sure that only authenticated users are publishing streams to my Adobe Flash Media Server.
Ideally I would like to limit the IPs which are allowed to publish streams to the server.
I read about the Adobe Authentication Plugin, but apparently it is broken since it works only for Adobe Flash Media Live Enconder. It doesn't secure the server for other encoders that publish streams.