Media Server :: Capturing Camera Feed Without FMS?
Nov 16, 2011
I'm not sure if this is the appropriate place to post this, but I was wandering its possible to build your own Player in Flash? The situation is: I have an external digital camera, and i need to capture its contents and display it in a player/window. This will all be on a local machine. I don't need to record anything, or stream it to other machines..... I just need the player to be able to display what ever the camera is displaying. Kind of like the Live Encoder (input and output views). With that I could hopefuly embed the player into other applications.
I have a video capture device on a machine. I have written a windows application to capture the video from this device. While capturing, if i try to broadcast using flash it does not connect. If i'm already broadcasting and then try to capture then my windows application does not get the capture pin of the device. So, flash seems to be taking the capture pin. Is there anyway to broadcast using the preview pin on the device so that i can capture using my application. Or is there way where i can split the capture pin and give one to my app and one to the broadcaster
I found the article titled 'Capturing live video' quite useful. This article is linked off the FMS start web page - the one that opens immediately after installing.[URL] The tutorial is better than most that I've found because it has a simple, working example. I was able to enter the code into a frame script in Flash CS5 and I used my own FMS address to connect both to a localhost and a remote server.
The sad part was that I found that I got the video and audio to round trip to and from the FMS when compiled (ctrl+enter) in the Flash CS5 program but when I try to publish this to a web page, I get audio but I can't get any video to display - neither eachoing straight from the camera or subscribing from the FMS. I've tried accessing the web page from my local file system and from a webserver. In both cases, I get audio (great news) but not video (bad news). I'm wondering why this might be. I checked warnings and errors but I'm not getting any. This solution is so simple and would pretty much solve my problems.
I'm new to Flash media server family. I need a server that can capture user recorded video/audio (through a flash recorder) and save it as a flv file on the server. which one of those server family is for me? I'm currently doing it using Red5, but would like to try out Flash media server as well.
There's just one thing I still don't get about the sample app: As a client, you always have to check the checkboxes (send/receive Audio/Video) in order to define who is sending/receiving what kind of stream. How is it possible to manage the feeds from the FMS server? So the clients can't set the checkboxes, but the server determines who is sending and who is receiving?
How to detect camera is in use or not? Adobe says, if Camera.getCamera(); returns null, it means camera not installed, or in use , i use below, then when i start swf, first one is ok, but then i start new instance of app, it throws and error " .... flash 10 r2has stopped" how can i solve that problem...
I have installed flash media server, flash cs5, media live encoder to be able to view my webcam feed via my website. Now in addition to these, i started streaming to localhost, and installed a dns updater which updates a name server according to my current ip. Then i wrote this dns address in the feed's source from flash 5 as rtmp://thednsname.somedns.net/live/livestream
so far so good, i also forwarded the ports 1935 incoming and outgoing from modem to my computer as well (because 1935 is said to be the default port number for media server). firewall -> closed (just in case); and my internet provider doesn't block ports...started the media server (with admin privilage in 7), started the live encoder, connect to localhost, start live feed.
NetStream.Play.StreamNotFound : Adobe Flash tried to play a live or recorded stream that does not exist. Source can't be found.
I'm trying to use Flash Media Server (on Amazon) to read in a video feed from a user's webcam, process it via my own executable, and return processed data (not video) to the user. For simplicity, let's say my executable finds the darkest pixel in the image and returns its coordinates, and I want the client to draw a cross at this location.
The client-side is working fine. It can send the video stream, and receive vertex data to draw. The server is half-working. It can accept connections, and send fake vertex data to the client. However I can't find any way to access the webcam stream.
Ideally, I would pass each frame from the stream into my executable for processing, with no saving to disk (or perhaps minimal saving to disk as a cache for P-frames). I've got the impression that this is impossible though, so another acceptable solution is to save the stream (live) to disk, and read it (live) from my executable.
Also, somewhat relatedly, I would like to secure the video feed as much as possible. Eventually I'll use SSL to secure it in transit, but for now I have the more serious problem of FMS sharing every stream it recieves. I'm sure there's an option to disable this somewhere, but I just can't find it.
How I can use ip/network camera with Flash Media Server.I am not able to configure it.Is there any other way of configuring such cameras ?I want to publish live stream data captured by ip camera on Flash Media Server.
I've been chased a lot of things, but so far not found anything very practical, such an example that makes me understand the workings of Flash Media Server, something like transmit, share my webcam or camera!
Sets the camera capture mode to the native mode that best meets the specified requirements. If the camera does not have a native mode that matches all the parameters you pass, Flash Player selects a capture mode that most closely synthesizes the requested mode. This manipulation may involve cropping the image and dropping frames.Does it mean that the frames captured is not necessarily the same as required by setMode(width:int, height:int, fps:Number, favorArea:Boolean = true) ? The actual frame captured may be either wider/ narrower,higher/lower in size and have higher/lower frequency than required ,is this true?
i have problem with publish with fms.i have a song play in back i have microphone to sing and camera when publish in mp4 the voice is ok and camera good. But the song play in background dont have good quality.
1- can i separate voice and song in my microphone? if yes do you have an exemple? because the sound o f voice is ok but song in background not good
2- else can mix my original song mp4 to camera and microphone? and publish record all with fms?
i am building an live streaming website and i am use
1- FMS 2- Apache webserver
i have made the subcriber and i works well .but i need to build the publisher to allow users to broadcast thier stream from thier cameras.i have tested a publisher which was build using actionscript and it didn't work untill i have installed the fms on my localhost , but i need a publisher which any user can use in his web browser .
I'm new to AS3 and attempting to learn it by doing a small task, I am trying to get a live local video feed from USB webcam and want to be able to capture and save screenshots from it.I have the live feed working fine, but I am struggling to capture and save a screenshot.Someone from adobe forums suggested I should use BitmapData and JPGEncoder, which I have alongside as3corelib's.[code]I think I am doing something wrong when I am capturing the binary data, or when I am encoding it to JPG.Some assistance from somebody familiar with these methods will be a huge help as I am stuck now, and I am not getting any replies from adobe forums.Both the resolution of the videoInstance object and camera feed are 640x480, and the video is displayed fine.
I have a delicate question for you which i cant find an answer to.Problem:I want to add an object (lets say it's a logo) to a camera stream before i save it to the FMIS. So the logo is integrated with the camera as it would have been placed in a layer above.Is that possible and if it is, how do i do it? It is quite urgent so any quick tips or solutions are more than welcome
I want to publish camera over FMS. But I don't want to use adobe security panel to be displayed(not even once) with setting as "access" by default.I wan't to set it allowed programatically.
When user first time log into site camera start the capturing and stream is saved in application folder with the streamname (straemname.flv). then user logout and then logged in then i have to append that stream in same straemname.flv file .By this code i have done that task
camera = Camera.getCamera(); camera.setQuality(0,80); camera.setMode(videoWidth,videoheight,15,true);camera.setLoopback(false); video.attachCamera(camera);ns.attachCamera(camera);ns.publish(stream, "append");
when user again logged in ,the video which is saved previously start playing,I dont want start that video.I just want to show camera control
I learned from some previous post that the default encoding format used when you publish the camera from a Flash player is Soreson. I'm wondering whether it is possible to select to use other codec such as VP2 or H.264?
I know the FMS is good for streaming things video to clients and for clients to publish videos from their attached camera. Is it possible for a client to publish something other than a stream from a connected camera, for example, the contents of my swf stage? Is it possible to draw something like a text overlay onto a published stream?
I have been working on a Flex application that sends a feed from my webcam to the Flash Media server. The application connects to the server fine but for some reason the camera is not sending anything to the Media Server. I am pretty sure that the answer to my question is really simple but I need another set of eyes to llok at my code and tell me what I am doing wrong.
I'm having trouble with Camera.getCamera() in Flash AS3. If there are multiple camera drivers, it does not necessarily pick the correct one. If the default is the correct, the program works fine; however, if another driver is selected as the default then LED on the webcam does not come on and no video is captured. I can cycle through the available cams and select an arbitrary one. I don't, however, know how to determine if the selected camera is correct. It is not returning null. I considered using camera.currentFPS, but it's not clear when this gets updated. I'm really struggling to determine dynamically which camera will do the trick!