Home » Android » Streaming video from Android camera to server

Streaming video from Android camera to server

Posted by: admin March 10, 2020 Leave a comment


I’ve seen plenty of info about how to stream video from the server to an android device, but not much about the other way, ala Qik. Could someone point me in the right direction here, or give me some advice on how to approach this?

How to&Answers:

I have hosted an open-source project to enable Android phone to IP camera:


Raw video data is fetched from LocalSocket, and the MDAT MOOV of MP4 was checked first before streaming. The live video is packed in FLV format, and can be played via Flash video player with a build in web server 🙂


Took me some time, but I finally manage do make an app that does just that. Check out the google code page if you’re interested: http://code.google.com/p/spydroid-ipcamera/
I added loads of comments in my code (mainly, look at CameraStreamer.java), so it should be pretty self-explanatory.
The hard part was actually to understand the RFC 3984 and implement a proper algorithm for the packetization process. (This algorithm actually turns the mpeg4/h.264 stream produced by the MediaRecorder into a nice rtp stream, according to the rfc)



I’m looking into this as well, and while I don’t have a good solution for you I did manage to dig up SIPDroid’s video code:



I’ve built an open-source SDK called Kickflip to make streaming video from Android a painless experience.

The SDK demonstrates use of Android 4.3’s MediaCodec API to direct the device hardware encoder’s packets directly to FFmpeg for RTMP (with librtmp) or HLS streaming of H.264 / AAC. It also demonstrates realtime OpenGL Effects (titling, chroma key, fades) and background recording.

Thanks SO, and especially, fadden.


Here is complete article about streaming android camera video to a webpage.

Android Streaming Live Camera Video to Web Page

  1. Used libstreaming on android app
  2. On server side Wowza Media Engine is used to decode the video stream
  3. Finally jWplayer is used to play the video on a webpage.


I am able to send the live camera video from mobile to my server.using this link
see the link

Refer the above link.there is a sample application in that link. Just you need to set your service url in RecordActivity.class.

Example as:

we can able to send H263 and H264 type videos using that link.


This is hardly a full answer, but webRTC may be what you’re looking for. Here’s some quick examples of webRTC in action: http://www.webrtc.org/reference-apps

If you want Android specific code, it exists! http://www.webrtc.org/native-code/android


Check Yasea library

Yasea is an Android streaming client. It encodes YUV and PCM data from
camera and microphone to H.264/AAC, encapsulates in FLV and transmits
over RTMP.


  1. Android mini API 16.
  2. H.264/AAC hard encoding.
  3. H.264 soft encoding.
  4. RTMP streaming with state callback handler.
  5. Portrait and landscape dynamic orientation.
  6. Front and back cameras hot switch.
  7. Recording to MP4 while streaming.


Mux (my company) has an open source android app that streams RTMP to a server, including setting up the camera and user interactions. It’s built to stream to Mux’s live streaming API but can easily stream to any RTMP entrypoint.


Depending by your budget, you can use a Raspberry Pi Camera that can send images to a server. I add here two tutorials where you can find many more details:

This tutorial show you how to use a Raspberry Pi Camera and display images on Android device

This is the second tutorial where you can find a series of tutorial about real-time video streaming between camera and android device