I have an RSS feed that is being read into dynamic text fields. The content of the RSS feed needs to be updated like every minute or every half minute, as there are constantly new items in the feed. Is there a way in ActionScript to reload the feed in interval times, or maybe reload the whole script?
I am streaming to live videos with audio to our webpage, each with it's own accompanying audio. On my laptop, I could stream them without any trouble but it would overheat so I had two new computers built just for encoding. Now, on the new computers, the 2 videos stream fine but only on audio file works at a time. The other just gets hung up on one note of sound and repeats very annoyingly over and over.
We (the university I work for) want to add an IP camera to the top of a building (so it has to be IP, we can't put a machine up there, so USB is out of the question) to stream a live view of the quad. The problem is, I can't find a way to stream any IP Cameras through the flash live encoder. I tried a camera by Axis, but their capture driver only supported MJPEG which the live encoder does not.
In using Flash Live Media Encoder to aquire a webcam video and to send it to a Flash Media Development Server 3.5 for live video streaming to one client by the same local network.The client device doesn't support flash player but is able to read and play flv video format.So, if i get a flv video file from apache 2 it works.But I don't know how to get a live video streaming from the server. The client application needs a flv file to point to.Does FVS generate a flv file for live video streaming?
How would I go about this? any one know of any good tutorials of how to stream sounds from a recording? any special type of server i would HAVE to use?
I'm finally getting the hang of the Dynamic Streaming of live video content via FMS 3.5.2 and FMLE, and to a lesser extent, the DVR functionality that has newly been made available.
Inquiring minds want to know... is it possible to add DVR functionality to live Dynamically Streamed video content? In other words, I want to be able to provide DVR capability to our live videos that are being streamed at 3 different bitrates.
Is that currently possible? Jodi, can you ask David Hassoun, who seems to be the leading authority at the moment, if this is possible, and if so, if there's a tutorial we can access?
I was pretty confused while stetting up a FMS and reading the docs. Live Streaming is by default enabled for everyone without password. Then I discovered and installed the authentication module. But this does not work and is not well documented. Also the download page states that it will work with FMS but some lines below it is stated that: "the Flash Media Server Authentication Add-In is only available for the Flash Media Interactive Server and the Flash Media Developer Server. This Add-In does not work with the Flash Media Streaming Server." hmmm. Confusing. How can I prevent that everyone is using my server for live streaming?
I've tested an exciting tutorial about streaming live video with Flash Media Server 3.5..every thing went good and i could see my webcam broadcast from my machine through my web site but unfortunately.I can't see it from any other machine.I'm using Microsoft windows XP SP2and flash media live encoder 3but my web server run UNIX is this a problem?
I'm hoping there's a simple answer to this. I'm attempting to stream a live event using FMLE and FMS. The first thing that throws me off is that I cannot push the stream to the server using RTMPE, only RTMP[T]. I assume this is because the the encryption is handled by the server and not the encoder. However, for the sake of security I would like to disable RTMP connections from the flash player to the server. Yet it would seem that if I incorporate the server side scripting described here [URL] (namely the changes to my main.asc file), that my server would reject an RTMP connection from FMLE? Is that correct? If so, where do I go from here? Can I broadcast a live stream using RTMPE only? Do I need a different encoder or something, or am I just missing a very simple point?
The article Live dynamic streaming with Flash Media Server 3.5 mentions two different bit rates for each video size type.But why? Are these bitrates related to different compression ratios? And if so, why exactly two?
I herd about RED5, but unfortunately I can't find any examples for feeding it with external source. I know RTSP is in development, but I'm thinking about (named) pipes or something like this. There is is also project named xuggle that is as far as I understand ffmpeg wrapper for Java, but I herd they have also problems with live streaming.
we are looking forward towards developing a very interesting community portal that would help the user to broadcast their live videos across the community.I've been checking over sites like ustream.tv, justin.tv and wondering what/how Technology they been using to do so.
I am doing a lot of research over the last few days checking over the medium to do this effectively and figure out some of the leading companies in the domain like Ooyala.com, brightcove.com providing servers/technology to broadcast videos seamlessly across the globe. I will be signing up with any of these providers soon.So my question is , how exactly would my website be catching with the live feed from the users cam, send the stream to ooyala/brightcove and further broadcast it to rest of the community users.
I have to install a multi-webcam setup that will be live streamed and accessible via a web browser.I'm looking into the quickest and most efficient solution, the less time and hassle it takes to implement, the better. That means buying hardware and/or software is not that much of an issue (as long as it's not 1000$+ flash server software).
The requirements / basic setup:The web server will be directly connected the cams (via ethernet / WiFi).The stream should be viewable by the max possible audience so no ActiveX / VLC-style object embedding. Flash would be ideal (even if it's a paid solution).Bandwidth is not an issue, max 2-3 clients at a time, most of the time only one.
Im Using Red5 to record live stream from the user. I am able to save the stream to disk and it generates an FLV file.
When I'm trying to read the FLV file with some FLV player - it seems like the file's meta-data is corrupted - which also makes the progress bar to act oddly. Also, I tried to play with the file using Xuggle. It seems impossible since the generated FLV do not have the codec-id in it.
I now have managed to get the red5 server up and running and ive been using the samples to stream my webcam live from one computer from another and it works fine.Now to step 2. Making a custom client that autoconnects to my live feed.
We have a problem working with FMS 3.5 (installed on a IIS server). Configuration of the servers all based on IIS: webserver (authentification, xml data loaded by Flash app, data (on IIS) sended to and recieve from FMS). Seperate FMS 3.5 (latest update) server. Only running FMS (no other apps are running on this machine).
I have got success in making live streaming with telecast at different pc. The application which I have made has the features: I' am publishing live stream using the connection like: nc.connection("rtmp://10.8.4.56:1935/live"); publishing the stream like: ns.publish("mycamera","live"); now, on other pc I am running the file for telecast using the following code: making connection with server: nc.connection("rtmp://10.8.4.56:1935/live"); playing the live stream: ns.play("mycamera"); vid.attachNetStream(ns);
Now, that was for the live stream for publish and view. But I want little more than that.I want to record the live stream at the publish end simultaneously. I have tried the code at publish side like: nc.connection("rtmp://10.8.4.56:1935/live"); ns.publish("mycamera","live"); ns.publish("mycamera","record"); But it is giving me error: NetStream.Record.NoAccess when I am changing the connection point to tmp://10.8.4.56:1935/dvr", t is giving me the result but then effecting the live telecast.
I am in the market for a good (pro)camcorder to stream live university seminars. Currently I use a Canon Optura Xi with firewire output and use the miniDV tape for backup in case something interfers with the live stream. I want to get away from miniDV tape and find a video camera with SD/flash card recording. My dilemma, most new camcorders do not have firewire output for live recording. We do our live stream through the Adobe Media Encoder
We have video cameras from Point Grey Research that are typically used for instrumentation and measuring.To oversimplify, the Flycapture software library they provide essentially captures single frames as 2-dimensional arrays of 16-bit greyscale values. We want to measure and process this information in real time. We also want to take this video, write some simple overlays on it (rectangular outlines, histograms, etc.) and make it available for viewing on a local area network using an ordinary web browser.
That is, we need to programmatically generate a live FLV stream on the fly--where the source of the video is not actually a device, but our program. The resolution and bandwidth requirements are modest; 320x240 30 fps 8 bit-grey = less than three megabytes per second. Some very simple, fast, easy compression would do, and in fact compression is not really needed at all.
I'm puzzled as how to where to start. I've browsed dozens of free and commercial packages, and they all seem to assume that the video source is a device.
The FLV and RTMP specs are intimidating. I can't easily identify any trivial subset (e.g. no compression) that would be easy to implement myself, nor have I found free or inexpensive commercial libraries that would do it for me.
I'm using FMLE 3.1 to stream live video encoded with H.264 format with FMS 3 and Flash player 10 + AS 3 to connect to the stream. When I connect to the stream, just the audio is played. I'm able even to get the metadata information about the video, but I just receive the audio. I already tried some stuff like
1. "Flash 10 won't play live stream H.264 after iTunes install" [URL]. I tested it in a complete different environment than mine, but the same result.
2. I've tried some format to play method, but this is just to play files ns.play("mp4:saple.f4v"); ns.play("mp4:sample");
3. Also read "How do you watch and record a live h.264" [URL], but I don't get even to play the stream at first place. This is the code I'm using
import flash.media.Video; var video:Video = new Video(720, 480); var ncVideo = new NetConnection(); this.ncVideo.addEventListener(NetStatusEvent.NET_STATUS, onNetStatus); [Code] .....
Another detail is that when I record the video to a file for instance "sample.f4v", I put this video in the FMS but when I connect to this stream I receive the "FileStructureInvalid" error message. I went from changing the extension to .flv [URL] to the solution to flatten the files [URL] but this is not the case because I'm using FMS to stream the recorded video.
I am a Wedding/Events Photographer/VideographerOne of my client want his events be live streamed over internet, I am quiet new for this type of jobas to what Software / Hardware I have to purchase for Live Streaming of Video over Internet.. and how it will encode in realtime.?
We intend to develop a Flex client application the user to send audio from your microphone and mp3 at the same time to a server that Flash Media Server.It would be a mix of microphone and music in mp3 that is playing in the application at the same time transmitted to the listeners.What will be the version of Flash Media Server that enables this mixing of audio from microphone and music played on the client in flex?