Media Server :: Get NetStream.time On The Serverside?
Jan 26, 2010
I'm using the following code to attempt to get netStream.time on the serverside.....i've tried various ways googled it but have come up short...what code do i need to use so i can track a streams progress from the server side?
myStream = Stream.get("home1");
setInterval(time,5000,myStream.time); function time(myStream){ trace(myStream); }
I am developing a video player for FMS 3.0 and my client complains that when he leaves the player running overnight or after computer goes to sleep when he comes back the elapsed time displayed as NaN and when he tries to play, nothing happens. Why is this happening? Does it mean that the connection with the FMS dropped and I have to reconnect? How can test for the dropped connection?
I am streaming a playlist of multiple mp4 video files with flash 9 and FMS 3 and have been experiencing strange netStream.time values when the buffering is set for an extra long time on slow connections. When I have my buffer set to something like 60 seconds I'll start experiancing this problem:
the NetStream.time value will return 0 on progress events while buffering, but the moment the buffer is full, instead of reading 0,1,2 +++ it jumps to a really high value, in this case 53.5 seconds, and sticks there until the video (which seems to play correctly from the beginning) catches up to that time and then the netStream.time starts incrementing as it should.
Another relevant fact is that the time the netStream.time gets stuck to always corresponds to the end/begining point between two streams in the playlist. For whatever reason this usually seems to be the third video in the playlist. If the connection is relatively fast and the buffer is set to a smaller value,
I'm trying to get some xml on the serverside to parse. I can load the xml just fine so there is no problem there but i can't get the streamArray to populate nor can i get any of the if statments after loaded to execute. Here is the code feel free to use the xml document on my server.
var StreamArray = new Array();var location_xml = new XML(); location_xml.onLoad = function( success ) { if(location_xml.loaded == true) {trace(location_xml); if (location_xml.firstChild.hasChildNodes()) { for (var aNode = location_xml.firstChild.firstChild; aNode != null; aNode=aNode.nextSibling) {
I've implemented a serverside script that records the current stream to disk. This record method is called from another clientside flash application. On some occasions the recording suddenly stops after a few hours without giving any errors, and without there being any network dropouts (Flash Media Encoder doesn't report any dropouts). What can be the cause of this, and what can I do to make the recording more stable?
I am streaming pre recorded audio files (mp4) to an AIR client. I have tried two different solutions, streaming the file directly with the NetStream.play("mp4:xxxxx.m4a") and creating a server side playlist, adding the same sound clip and then streaming the playlist. The problem is, when streaming the playlist i get a few NetStream.Play.InsufficientBW, this does not happen when streaming the file direct. Both solutions uses bufferTime=1.0
I would like to use a server side playlist to implement a simple key solution so that the client dont know the full path to the file, but instead sends a key to a custom server side function that looks up the file path and creates a stream for the client.
I'm creating a video chat app using Flash Media Server 4 and Flex, using RTMFP for peer-to-peer.
No matter what I try, I cannot mute a user's NetStream. I tried receiveAudio(false) and that does not work.
I tried capturing the stream SoundTransform and setting the volume of that to 0, that does not work.
I tried setting the mx VideoDisplay volume to 0 and that does not work.
I can set the alpha of the VideoDisplay, so its really strange that I cannot set the volume.
BTW, I am not trying to mute the mic of the user who launched the app, that I can do. I'm trying to mute one of the other users stream in the Flex app, that way the app makes it possible to not hear people who are perhaps being offensive.
I have an FMIS 3.5 app where I am attempting to publish multiple streams to an edge server.I have found that if I loop through my list of streams and attempt to publish to the edge server (i.e. as fast as possible), the publishing will fail They succeed if I have some delay between the publish calls.
I'm working with a custom made Flash video player using FMS (Wowza, actually).The video is working, yet when I access the BytesLoaded and BytesTotal,I get 0 (zero) for each.This even persists as the video is loading and playing.It's as if it is not even being registered at all.Here's the code:
I'm recording audio using FMIS, the file is being saved out as a FLV file and I am able to play it back. Eventually that file may be deleted after some time and I would like to notify the user if the file is no longer available. I was expecting to see NetStream.Play.StreamNotFound after deleting the file on the server but I'm not getting that message. Instead I get:
--> NetStream.Play.Reset--> NetStream.Play.StartAnd nothing plays which is expected since the file is deleted, but I dont get the right message. I'm calling the file like so:ns.play ("flv:file_to_stream");
I'm streaming mp3 on flash media server 3.5 on linux version, and very often come accross with that problem, when the buffer.length become 0 on client side, the streaming will stop with NetStream.Play.Stop event, and will not be continued. I think this case it should wait always till the data arrives and never should stopping. Some more info : I tested on applications/live folder, and the mp3 files are created dynamically during the playing, and the creation speed is almost the same like the playing speed, so those are not fix files.
I have application that uses the send functionally in the NetStream on the server-side. When a connect to the app using rtmp I am able to see the send come thru but if the app uses rtmfp it does notI gone thru all the examples in setting up a mutlicast app and I know the app works because when I publish a video out everyone in the group sees the video. The only thing not working is send. I've also set dataReliable to true.
I'm trying to get the metaData as usual with NetStream object from FMS but in RTMFP protocol but doesn't work.However, with RTMP my code (which is coming from onMetaData example in livedocs)it works.
Is it need any special metchod/properties to get the metadata with rtmfp ?
I would like to have the possibility to detect when NetStream is not working (i.e. the name used in initialization can't be found on the server side). The StreamNotFound doesn't work at all. I've read that it's the fault of the Flash Media Server, which automaticaly creates new stream, if it cannot be found (of course it's an empty stream - what is wrong from my point of view). Is it true? If yes, can I disable it on the server side, so I could easily detect if the stream name is correct?
i've created a sub-folder called dvr under fms 4.0 applications. i made a test of playing a video named sample.flv then at the same time i publish it to the same fms 4.0 server with a different video name: "mergedVideo".the asc at the fms server as following:
I'm trying to send metaData from client to FMS server when recording Webcam video and have implemented sending of metaData as stated on Adobe's help page at: http:[url]....
However when I play the recorded video from FMS, my custom metaData properties trace "undefined".I can trace default metaData properties such as duration or videocodecid successfully,but not my custom properties such customProp,width or height. Here is part of my code that is related to the issue: private function
I've googled this and come up empty. All I am trying to do is create a netstream and assign a group to it like so:
var netstream; var groupSpec; groupSpec = new GroupSpecifier("foo"); trace("groupSpec: "+ groupSpec);
[Code]...
If I remove the groupSpec optional parameter I successfully connect the NetStream. According to the docs [on this page] I should be able to identify a GroupSpecifier in the NetStream constructor.
I notice that, when I pause the video, the event "NetStream.Buffer.Flush" is triggered. And according to the language reference: "Data has finished streaming, and the remaining buffer will be emptied.", I have to re-buffer it, right? However, also according to the reference, it shouldn't stop buffering:Starting with Flash Player 9.0.115.0, Flash Player no longer clears the buffer when NetStream.pause() is called. This behavior is called "smart pause". Before Flash Player 9.0.115.0, Flash Player waited for the buffer to fill up before resuming playback, which often caused a delay.I'm using Flash Professional to do the debugging, and the traced version number is: MAC 10,0,22,91, and the streaming server is FMS4
i have application that streams videos from fms .i try to play video of 240 seconds, but when playback reaches 180 seconds i got NetStream. play. stop . it happens regularly at the same time.why does it happen? is it encoding of the video??
We are working with FMS 4 server for a while for a 2 directions video application, and it works great with rtmp. We now want to use its rtmfp abilities after we used Cirrus for testing in the last few days and it also worked well. Locally - everything is working fine, but when we try the application on the remote server - we have some problems.
[Code]....
Again, it is working beautifully both locally and with Cirrus. RTMP also works well with the remote server.
I'm trying to add a custom callback handler to a NetStream client in a p2p application. The problem is, when I add such a handler, the NetStream Client doesn't function as it did before. It seems the NetStream.client object is changed. I know the default object is this. But changing the client to this doesn't solve the problem. The remoteControlStream, is the incoming stream. And the localControl stream is the stream being published. This is the localControlStream that's being send by the peer, and received as remoteControlStream:
private function initLocalControlStream():void{ localControlStream = new NetStream(nc, NetStream.DIRECT_CONNECTIONS); localControlStream.addEventListener(NetStatusEvent.NET_STATUS, localControlHandler); localControlStream.publish(myPeerID+"control"); var localControlStreamClient:Object = new Object(); [Code] .....
The onPeerConnect method of the localControlStream doesn't get called when I connect when the above handler is added. When I remove the that handler, the onPeerConnect method gets called. Obviously the problem is the NetStream.client.
I am developing a C++ program to connect FMS via RTMP, and then publish a live stream I can't use FMLE here because I need collect the H.264 NALU from somewhere and forward them to the FMS.I write a demo app, which parses H.264 encoded FLV file, and send the VideoTag in the FLV as RTMP Message content for publishing.Finally, I make it run and I can see some frames of video!But the problem is: The video does not play smoothly and it just updates some frames. Then I looked at the NetStream event and found that:Everytime when there is a NetStream.Buffer.Full event, the frame in the video is updated and display correctly But immediately there is another NetStream.Buffer.Empty event followed and the video is frozen.
<?xml version="1.0" encoding="utf-8"?> <mx:Application xmlns:mx="[URL]" layout="absolute" creationComplete="init();"> <mx:Script> <![CDATA[ import mx.controls.Alert; import mx.core.mx_internal; import flash.media.Camera; [Code] ..... So, I play the record video , but there is no image , just voice
NetStream.play(streamName, -1); This seems to be working wrong.if I have recorded an flv on server using FMS and FMLE with only audio with name "myaudio" and then after if I try to play a live stream using NetStream.play("myaudio", -1) then it plays the recorded stream. I believe that documentation says that it should start a live stream instead of playing recorded stream as the second argument is -1. Is this a bug in NetStream.play method?
I am developing a video system using the peer-assisted networking of FMS. but when I have published video successed, I can't receive the video at some endpoint.
I'm trying to create a simple application that streams the webcam to FMS 4.5 (running on a Windows 7 machine), and then displays the published stream locally. Here's my code:
All traces in the event handlers show the NetConnection connecting successfully, and the netstream publishing successfully. However, I get no video output from this. I've confirmed this using the sample video player in FMS 4.5 as well. There is also a weird traffic pattern reported in the FMS Admin console.There is some traffic (~4K) when the NetStream first publishes, but after that there is no activity.