/* --------------------------------------------------------------------------- ** This software is in the public domain, furnished "as is", without technical ** support, and with no warranty, express or implied, as to its usefulness for ** any purpose. ** ** main.cpp ** ** V4L2 RTSP streamer ** ** H264 capture using V4L2 ** RTSP using live555 ** ** -------------------------------------------------------------------------*/ #include #include #include #include #include #include #include #include // libv4l2 #include // project #include "logger.h" #include "V4l2Device.h" #include "V4l2Output.h" #include "DeviceSourceFactory.h" #include "V4l2RTSPServer.h" // ----------------------------------------- // signal handler // ----------------------------------------- char quit = 0; void sighandler(int n) { printf("SIGINT\n"); quit =1; } // ------------------------------------------------------- // split video,audio device // ------------------------------------------------------- void decodeDevice(const std::string & device, std::string & videoDev, std::string & audioDev) { std::istringstream is(device); getline(is, videoDev, ','); getline(is, audioDev); } std::string getDeviceName(const std::string & devicePath) { std::string deviceName(devicePath); size_t pos = deviceName.find_last_of('/'); if (pos != std::string::npos) { deviceName.erase(0,pos+1); } return deviceName; } // ----------------------------------------- // entry point // ----------------------------------------- int main(int argc, char** argv) { // default parameters const char *dev_name = "/dev/video0,/dev/video0"; unsigned int format = ~0; std::list videoformatList; int width = 0; int height = 0; int queueSize = 5; int fps = 25; unsigned short rtspPort = 8554; unsigned short rtspOverHTTPPort = 0; bool multicast = false; int verbose = 0; std::string outputFile; V4l2IoType ioTypeIn = IOTYPE_MMAP; V4l2IoType ioTypeOut = IOTYPE_MMAP; int openflags = O_RDWR | O_NONBLOCK; std::string url = "unicast"; std::string murl = "multicast"; std::string tsurl = "ts"; V4L2DeviceSource::CaptureMode captureMode = V4L2DeviceSource::CAPTURE_INTERNAL_THREAD; std::string maddr; bool repeatConfig = true; int timeout = 65; int defaultHlsSegment = 2; unsigned int hlsSegment = 0; const char* sslKeyCert = NULL; const char* realm = NULL; std::list userPasswordList; std::string webroot; #ifdef HAVE_ALSA int audioFreq = 44100; int audioNbChannels = 2; std::list audioFmtList; snd_pcm_format_t audioFmt = SND_PCM_FORMAT_UNKNOWN; #endif const char* defaultPort = getenv("PORT"); if (defaultPort != NULL) { rtspPort = atoi(defaultPort); } // decode parameters int c = 0; while ((c = getopt (argc, argv, "v::Q:O:b:" "I:P:p:m::u:M::ct:S::x:" "R:U:" "rwBsf::F:W:H:G:" "A:C:a:" "Vh")) != -1) { switch (c) { case 'v': verbose = 1; if (optarg && *optarg=='v') verbose++; break; case 'Q': queueSize = atoi(optarg); break; case 'O': outputFile = optarg; break; case 'b': webroot = optarg; break; // RTSP/RTP case 'I': ReceivingInterfaceAddr = inet_addr(optarg); break; case 'P': rtspPort = atoi(optarg); break; case 'p': rtspOverHTTPPort = atoi(optarg); break; case 'u': url = optarg; break; case 'm': multicast = true; murl = optarg ? optarg : murl; break; case 'M': multicast = true; maddr = optarg ? optarg : maddr; break; case 'c': repeatConfig = false; break; case 't': timeout = atoi(optarg); break; case 'S': hlsSegment = optarg ? atoi(optarg) : defaultHlsSegment; break; case 'x': sslKeyCert = optarg; break; // users case 'R': realm = optarg; break; case 'U': userPasswordList.push_back(optarg); break; // V4L2 case 'r': ioTypeIn = IOTYPE_READWRITE; break; case 'w': ioTypeOut = IOTYPE_READWRITE; break; case 'B': openflags = O_RDWR; break; case 's': captureMode = V4L2DeviceSource::CAPTURE_LIVE555_THREAD; break; case 'f': format = V4l2Device::fourcc(optarg); if (format) {videoformatList.push_back(format);}; break; case 'F': fps = atoi(optarg); break; case 'W': width = atoi(optarg); break; case 'H': height = atoi(optarg); break; case 'G': sscanf(optarg,"%dx%dx%d", &width, &height, &fps); break; // ALSA #ifdef HAVE_ALSA case 'A': audioFreq = atoi(optarg); break; case 'C': audioNbChannels = atoi(optarg); break; case 'a': audioFmt = V4l2RTSPServer::decodeAudioFormat(optarg); if (audioFmt != SND_PCM_FORMAT_UNKNOWN) {audioFmtList.push_back(audioFmt);} ; break; #endif // version case 'V': std::cout << VERSION << std::endl; exit(0); break; // help case 'h': default: { std::cout << argv[0] << " [-v[v]] [-Q queueSize] [-O file]" << std::endl; std::cout << "\t [-I interface] [-P RTSP port] [-p RTSP/HTTP port] [-m multicast url] [-u unicast url] [-M multicast addr] [-c] [-t timeout] [-T] [-S[duration]]" << std::endl; std::cout << "\t [-r] [-w] [-s] [-f[format] [-W width] [-H height] [-F fps] [device] [device]" << std::endl; std::cout << "\t -v : verbose" << std::endl; std::cout << "\t -vv : very verbose" << std::endl; std::cout << "\t -Q : Number of frame queue (default "<< queueSize << ")" << std::endl; std::cout << "\t -O : Copy captured frame to a file or a V4L2 device" << std::endl; std::cout << "\t -b : path to webroot" << std::endl; std::cout << "\t RTSP/RTP options" << std::endl; std::cout << "\t -I : RTSP interface (default autodetect)" << std::endl; std::cout << "\t -P : RTSP port (default "<< rtspPort << ")" << std::endl; std::cout << "\t -p : RTSP over HTTP port (default "<< rtspOverHTTPPort << ")" << std::endl; std::cout << "\t -U : : RTSP user and password" << std::endl; std::cout << "\t -R : use md5 password 'md5(::')" << std::endl; std::cout << "\t -u : unicast url (default " << url << ")" << std::endl; std::cout << "\t -m : multicast url (default " << murl << ")" << std::endl; std::cout << "\t -M : multicast group:port (default is random_address:20000)" << std::endl; std::cout << "\t -c : don't repeat config (default repeat config before IDR frame)" << std::endl; std::cout << "\t -t : RTCP expiration timeout in seconds (default " << timeout << ")" << std::endl; std::cout << "\t -S[] : enable HLS & MPEG-DASH with segment duration in seconds (default " << defaultHlsSegment << ")" << std::endl; std::cout << "\t -x : enable RTSPS & SRTP" << std::endl; std::cout << "\t V4L2 options" << std::endl; std::cout << "\t -r : V4L2 capture using read interface (default use memory mapped buffers)" << std::endl; std::cout << "\t -w : V4L2 capture using write interface (default use memory mapped buffers)" << std::endl; std::cout << "\t -B : V4L2 capture using blocking mode (default use non-blocking mode)" << std::endl; std::cout << "\t -s : V4L2 capture using live555 mainloop (default use a reader thread)" << std::endl; std::cout << "\t -f : V4L2 capture using current capture format (-W,-H,-F are ignored)" << std::endl; std::cout << "\t -f : V4L2 capture using format (-W,-H,-F are used)" << std::endl; std::cout << "\t -W : V4L2 capture width (default "<< width << ")" << std::endl; std::cout << "\t -H : V4L2 capture height (default "<< height << ")" << std::endl; std::cout << "\t -F : V4L2 capture framerate (default "<< fps << ")" << std::endl; std::cout << "\t -G x[x] : V4L2 capture format (default "<< width << "x" << height << "x" << fps << ")" << std::endl; #ifdef HAVE_ALSA std::cout << "\t ALSA options" << std::endl; std::cout << "\t -A freq : ALSA capture frequency and channel (default " << audioFreq << ")" << std::endl; std::cout << "\t -C channels : ALSA capture channels (default " << audioNbChannels << ")" << std::endl; std::cout << "\t -a fmt : ALSA capture audio format (default S16_BE)" << std::endl; #endif std::cout << "\t Devices :" << std::endl; std::cout << "\t [V4L2 device][,ALSA device] : V4L2 capture device or/and ALSA capture device (default "<< dev_name << ")" << std::endl; exit(0); } } } std::list devList; while (optind outList; int nbSource = 0; std::list::iterator devIt; for ( devIt=devList.begin() ; devIt!=devList.end() ; ++devIt) { std::string deviceName(*devIt); std::string videoDev; std::string audioDev; decodeDevice(deviceName, videoDev, audioDev); std::string baseUrl; std::string output(outputFile); if (devList.size() > 1) { baseUrl = getDeviceName(videoDev); baseUrl.append("_"); // output is not compatible with multiple device output.assign(""); } V4l2Output* out = NULL; V4L2DeviceParameters inParam(videoDev.c_str(), videoformatList, width, height, fps, ioTypeIn, verbose, openflags); StreamReplicator* videoReplicator = rtspServer.CreateVideoReplicator( inParam, queueSize, captureMode, repeatConfig, output, ioTypeOut, out); if (out != NULL) { outList.push_back(out); } // Init Audio Capture StreamReplicator* audioReplicator = NULL; #ifdef HAVE_ALSA audioReplicator = rtspServer.CreateAudioReplicator( audioDev, audioFmtList, audioFreq, audioNbChannels, verbose, queueSize, captureMode); #endif // Create Multicast Session if (multicast) { ServerMediaSession* sms = rtspServer.AddMulticastSession(baseUrl+murl, destinationAddress, rtpPortNum, rtcpPortNum, videoReplicator, audioReplicator); if (sms) { nbSource += sms->numSubsessions(); } } // Create HLS Session if (hlsSegment > 0) { ServerMediaSession* sms = rtspServer.AddHlsSession(baseUrl+tsurl, hlsSegment, videoReplicator, audioReplicator); if (sms) { nbSource += sms->numSubsessions(); } } // Create Unicast Session ServerMediaSession* sms = rtspServer.AddUnicastSession(baseUrl+url, videoReplicator, audioReplicator); if (sms) { nbSource += sms->numSubsessions(); } } if (nbSource>0) { // main loop signal(SIGINT,sighandler); rtspServer.eventLoop(&quit); LOG(NOTICE) << "Exiting...."; } while (!outList.empty()) { V4l2Output* out = outList.back(); delete out; outList.pop_back(); } } return 0; }