mirror of
https://github.com/mpromonet/v4l2rtspserver
synced 2024-11-02 03:40:13 +00:00
fix build with latest live555 that removed TCPStreamSink
This commit is contained in:
parent
eaf0608aba
commit
be67798772
@ -5,7 +5,7 @@ project(v4l2rtspserver)
|
||||
option(COVERAGE "Coverage" OFF)
|
||||
|
||||
set(ALSA ON CACHE BOOL "use ALSA is available")
|
||||
set(LIVE555URL https://download.videolan.org/pub/contrib/live555/live.2020.03.06.tar.gz CACHE STRING "live555 url")
|
||||
set(LIVE555URL http://www.live555.com/liveMedia/public/live555-latest.tar.gz CACHE STRING "live555 url")
|
||||
set(LIVE555CFLAGS -DBSD=1 -DSOCKLEN_T=socklen_t -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -DALLOW_RTSP_SERVER_PORT_REUSE=1 -DNO_OPENSSL=1 CACHE STRING "live555 CFGLAGS")
|
||||
|
||||
set(CMAKE_CXX_STANDARD 11)
|
||||
|
109
inc/HTTPServer.h
109
inc/HTTPServer.h
@ -14,6 +14,104 @@
|
||||
|
||||
#include "RTSPServer.hh"
|
||||
#include "RTSPCommon.hh"
|
||||
#include <GroupsockHelper.hh> // for "ignoreSigPipeOnSocket()"
|
||||
|
||||
|
||||
#define TCP_STREAM_SINK_MIN_READ_SIZE 1000
|
||||
#define TCP_STREAM_SINK_BUFFER_SIZE 10000
|
||||
|
||||
class TCPStreamSink: public MediaSink {
|
||||
public:
|
||||
TCPStreamSink(UsageEnvironment& env, int socketNum) : MediaSink(env), fUnwrittenBytesStart(0), fUnwrittenBytesEnd(0), fInputSourceIsOpen(False), fOutputSocketIsWritable(True),fOutputSocketNum(socketNum) {
|
||||
ignoreSigPipeOnSocket(socketNum);
|
||||
}
|
||||
|
||||
protected:
|
||||
virtual ~TCPStreamSink() {
|
||||
envir().taskScheduler().disableBackgroundHandling(fOutputSocketNum);
|
||||
if (fSource != NULL) {
|
||||
Medium::close(fSource);
|
||||
}
|
||||
}
|
||||
|
||||
protected:
|
||||
virtual Boolean continuePlaying() {
|
||||
fInputSourceIsOpen = fSource != NULL;
|
||||
processBuffer();
|
||||
return True;
|
||||
}
|
||||
|
||||
private:
|
||||
void processBuffer(){
|
||||
// First, try writing data to our output socket, if we can:
|
||||
if (fOutputSocketIsWritable && numUnwrittenBytes() > 0) {
|
||||
int numBytesWritten = send(fOutputSocketNum, (const char*)&fBuffer[fUnwrittenBytesStart], numUnwrittenBytes(), 0);
|
||||
if (numBytesWritten < (int)numUnwrittenBytes()) {
|
||||
// The output socket is no longer writable. Set a handler to be called when it becomes writable again.
|
||||
fOutputSocketIsWritable = False;
|
||||
if (envir().getErrno() != EPIPE) { // on this error, the socket might still be writable, but no longer usable
|
||||
envir().taskScheduler().setBackgroundHandling(fOutputSocketNum, SOCKET_WRITABLE, socketWritableHandler, this);
|
||||
}
|
||||
}
|
||||
if (numBytesWritten > 0) {
|
||||
// We wrote at least some of our data. Update our buffer pointers:
|
||||
fUnwrittenBytesStart += numBytesWritten;
|
||||
if (fUnwrittenBytesStart > fUnwrittenBytesEnd) fUnwrittenBytesStart = fUnwrittenBytesEnd; // sanity check
|
||||
if (fUnwrittenBytesStart == fUnwrittenBytesEnd && (!fInputSourceIsOpen || !fSource->isCurrentlyAwaitingData())) {
|
||||
fUnwrittenBytesStart = fUnwrittenBytesEnd = 0; // reset the buffer to empty
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Then, read from our input source, if we can (& we're not already reading from it):
|
||||
if (fInputSourceIsOpen && freeBufferSpace() >= TCP_STREAM_SINK_MIN_READ_SIZE && !fSource->isCurrentlyAwaitingData()) {
|
||||
fSource->getNextFrame(&fBuffer[fUnwrittenBytesEnd], freeBufferSpace(), afterGettingFrame, this, ourOnSourceClosure, this);
|
||||
} else if (!fInputSourceIsOpen && numUnwrittenBytes() == 0) {
|
||||
// We're now done:
|
||||
onSourceClosure();
|
||||
}
|
||||
}
|
||||
|
||||
static void socketWritableHandler(void* clientData, int mask) { ((TCPStreamSink*)clientData)->socketWritableHandler(); }
|
||||
void socketWritableHandler() {
|
||||
envir().taskScheduler().disableBackgroundHandling(fOutputSocketNum); // disable this handler until the next time it's needed
|
||||
fOutputSocketIsWritable = True;
|
||||
processBuffer();
|
||||
}
|
||||
|
||||
static void afterGettingFrame(void* clientData, unsigned frameSize, unsigned numTruncatedBytes,
|
||||
struct timeval presentationTime, unsigned durationInMicroseconds) {
|
||||
((TCPStreamSink*)clientData)->afterGettingFrame(frameSize, numTruncatedBytes);
|
||||
}
|
||||
|
||||
|
||||
void afterGettingFrame(unsigned frameSize, unsigned numTruncatedBytes) {
|
||||
if (numTruncatedBytes > 0) {
|
||||
envir() << "TCPStreamSink::afterGettingFrame(): The input frame data was too large for our buffer. "
|
||||
<< numTruncatedBytes
|
||||
<< " bytes of trailing data was dropped! Correct this by increasing the definition of \"TCP_STREAM_SINK_BUFFER_SIZE\" in \"include/TCPStreamSink.hh\".\n";
|
||||
}
|
||||
fUnwrittenBytesEnd += frameSize;
|
||||
processBuffer();
|
||||
}
|
||||
|
||||
static void ourOnSourceClosure(void* clientData) { ((TCPStreamSink*)clientData)->ourOnSourceClosure(); }
|
||||
|
||||
void ourOnSourceClosure() {
|
||||
// The input source has closed:
|
||||
fInputSourceIsOpen = False;
|
||||
processBuffer();
|
||||
}
|
||||
|
||||
unsigned numUnwrittenBytes() const { return fUnwrittenBytesEnd - fUnwrittenBytesStart; }
|
||||
unsigned freeBufferSpace() const { return TCP_STREAM_SINK_BUFFER_SIZE - fUnwrittenBytesEnd; }
|
||||
|
||||
private:
|
||||
unsigned char fBuffer[TCP_STREAM_SINK_BUFFER_SIZE];
|
||||
unsigned fUnwrittenBytesStart, fUnwrittenBytesEnd;
|
||||
Boolean fInputSourceIsOpen, fOutputSocketIsWritable;
|
||||
int fOutputSocketNum;
|
||||
};
|
||||
|
||||
// ---------------------------------------------------------
|
||||
// Extend RTSP server to add support for HLS and MPEG-DASH
|
||||
@ -25,7 +123,7 @@ class HTTPServer : public RTSPServer
|
||||
{
|
||||
public:
|
||||
HTTPClientConnection(RTSPServer& ourServer, int clientSocket, struct sockaddr_in clientAddr)
|
||||
: RTSPServer::RTSPClientConnection(ourServer, clientSocket, clientAddr), fTCPSink(NULL), fStreamToken(NULL), fSubsession(NULL), fSource(NULL) {
|
||||
: RTSPServer::RTSPClientConnection(ourServer, clientSocket, clientAddr), m_TCPSink(NULL), m_StreamToken(NULL), m_Subsession(NULL) {
|
||||
}
|
||||
virtual ~HTTPClientConnection();
|
||||
|
||||
@ -43,11 +141,10 @@ class HTTPServer : public RTSPServer
|
||||
static void afterStreaming(void* clientData);
|
||||
|
||||
private:
|
||||
static u_int32_t fClientSessionId;
|
||||
TCPStreamSink* fTCPSink;
|
||||
void* fStreamToken;
|
||||
ServerMediaSubsession* fSubsession;
|
||||
FramedSource* fSource;
|
||||
static u_int32_t m_ClientSessionId;
|
||||
TCPStreamSink* m_TCPSink;
|
||||
void* m_StreamToken;
|
||||
ServerMediaSubsession* m_Subsession;
|
||||
};
|
||||
|
||||
public:
|
||||
|
@ -19,13 +19,13 @@
|
||||
class MemoryBufferSink : public MediaSink
|
||||
{
|
||||
public:
|
||||
static MemoryBufferSink* createNew(UsageEnvironment& env, unsigned int bufferSize, unsigned int sliceDuration)
|
||||
static MemoryBufferSink* createNew(UsageEnvironment& env, unsigned int bufferSize, unsigned int sliceDuration, unsigned int nbSlices = 5)
|
||||
{
|
||||
return new MemoryBufferSink(env, bufferSize, sliceDuration);
|
||||
return new MemoryBufferSink(env, bufferSize, sliceDuration, nbSlices);
|
||||
}
|
||||
|
||||
protected:
|
||||
MemoryBufferSink(UsageEnvironment& env, unsigned bufferSize, unsigned int sliceDuration);
|
||||
MemoryBufferSink(UsageEnvironment& env, unsigned bufferSize, unsigned int sliceDuration, unsigned int nbSlices);
|
||||
virtual ~MemoryBufferSink();
|
||||
|
||||
virtual Boolean continuePlaying();
|
||||
@ -53,5 +53,6 @@ class MemoryBufferSink : public MediaSink
|
||||
std::map<unsigned int,std::string> m_outputBuffers;
|
||||
unsigned int m_refTime;
|
||||
unsigned int m_sliceDuration;
|
||||
unsigned int m_nbSlices;
|
||||
};
|
||||
|
||||
|
@ -102,6 +102,7 @@ void* V4L2DeviceSource::thread()
|
||||
{
|
||||
if (FD_ISSET(fd, &fdset))
|
||||
{
|
||||
LOG(DEBUG) << "waitingFrame\tdelay:" << (1000-(tv.tv_usec/1000)) << "ms";
|
||||
if (this->getNextFrame() <= 0)
|
||||
{
|
||||
LOG(ERROR) << "error:" << strerror(errno);
|
||||
@ -213,6 +214,7 @@ int V4L2DeviceSource::getNextFrame()
|
||||
timersub(&tv,&ref,&diff);
|
||||
m_in.notify(tv.tv_sec, frameSize);
|
||||
LOG(DEBUG) << "getNextFrame\ttimestamp:" << ref.tv_sec << "." << ref.tv_usec << "\tsize:" << frameSize <<"\tdiff:" << (diff.tv_sec*1000+diff.tv_usec/1000) << "ms";
|
||||
|
||||
processFrame(buffer,frameSize,ref);
|
||||
if (m_outfd != -1)
|
||||
{
|
||||
@ -238,9 +240,9 @@ void V4L2DeviceSource::processFrame(char * frame, int frameSize, const timeval &
|
||||
char* buf = new char[size];
|
||||
memcpy(buf, frame.first, size);
|
||||
queueFrame(buf,size,ref);
|
||||
frameList.pop_front();
|
||||
|
||||
LOG(DEBUG) << "queueFrame\ttimestamp:" << ref.tv_sec << "." << ref.tv_usec << "\tsize:" << size <<"\tdiff:" << (diff.tv_sec*1000+diff.tv_usec/1000) << "ms";
|
||||
frameList.pop_front();
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -20,11 +20,10 @@
|
||||
#include "RTSPCommon.hh"
|
||||
#include <time.h>
|
||||
#include "ByteStreamMemoryBufferSource.hh"
|
||||
#include "TCPStreamSink.hh"
|
||||
|
||||
#include "HTTPServer.h"
|
||||
|
||||
u_int32_t HTTPServer::HTTPClientConnection::fClientSessionId = 0;
|
||||
u_int32_t HTTPServer::HTTPClientConnection::m_ClientSessionId = 0;
|
||||
|
||||
void HTTPServer::HTTPClientConnection::sendHeader(const char* contentType, unsigned int contentLength)
|
||||
{
|
||||
@ -56,22 +55,16 @@ void HTTPServer::HTTPClientConnection::streamSource(const std::string & content)
|
||||
|
||||
void HTTPServer::HTTPClientConnection::streamSource(FramedSource* source)
|
||||
{
|
||||
if (fTCPSink != NULL)
|
||||
if (m_TCPSink != NULL)
|
||||
{
|
||||
fTCPSink->stopPlaying();
|
||||
Medium::close(fTCPSink);
|
||||
fTCPSink = NULL;
|
||||
}
|
||||
if (fSource != NULL)
|
||||
{
|
||||
Medium::close(fSource);
|
||||
fSource = NULL;
|
||||
m_TCPSink->stopPlaying();
|
||||
Medium::close(m_TCPSink);
|
||||
m_TCPSink = NULL;
|
||||
}
|
||||
if (source != NULL)
|
||||
{
|
||||
fTCPSink = TCPStreamSink::createNew(envir(), fClientOutputSocket);
|
||||
fTCPSink->startPlaying(*source, afterStreaming, this);
|
||||
fSource = source; // we need to keep tracking of source, because sink do not release it
|
||||
m_TCPSink = new TCPStreamSink(envir(), fClientOutputSocket);
|
||||
m_TCPSink->startPlaying(*source, afterStreaming, this);
|
||||
}
|
||||
}
|
||||
|
||||
@ -322,17 +315,17 @@ void HTTPServer::HTTPClientConnection::handleHTTPCmd_StreamingGET(char const* ur
|
||||
|
||||
// Call "getStreamParameters()" to create the stream's source. (Because we're not actually streaming via RTP/RTCP, most
|
||||
// of the parameters to the call are dummy.)
|
||||
++fClientSessionId;
|
||||
++m_ClientSessionId;
|
||||
Port clientRTPPort(0), clientRTCPPort(0), serverRTPPort(0), serverRTCPPort(0);
|
||||
netAddressBits destinationAddress = 0;
|
||||
u_int8_t destinationTTL = 0;
|
||||
Boolean isMulticast = False;
|
||||
subsession->getStreamParameters(fClientSessionId, 0, clientRTPPort,clientRTCPPort, -1,0,0, destinationAddress,destinationTTL, isMulticast, serverRTPPort,serverRTCPPort, fStreamToken);
|
||||
subsession->getStreamParameters(m_ClientSessionId, 0, clientRTPPort,clientRTCPPort, -1,0,0, destinationAddress,destinationTTL, isMulticast, serverRTPPort,serverRTCPPort, m_StreamToken);
|
||||
|
||||
// Seek the stream source to the desired place, with the desired duration, and (as a side effect) get the number of bytes:
|
||||
double dOffsetInSeconds = (double)offsetInSeconds;
|
||||
u_int64_t numBytes = 0;
|
||||
subsession->seekStream(fClientSessionId, fStreamToken, dOffsetInSeconds, 0.0, numBytes);
|
||||
subsession->seekStream(m_ClientSessionId, m_StreamToken, dOffsetInSeconds, 0.0, numBytes);
|
||||
|
||||
if (numBytes == 0)
|
||||
{
|
||||
@ -346,10 +339,10 @@ void HTTPServer::HTTPClientConnection::handleHTTPCmd_StreamingGET(char const* ur
|
||||
this->sendHeader("video/mp2t", numBytes);
|
||||
|
||||
// stream body
|
||||
this->streamSource(subsession->getStreamSource(fStreamToken));
|
||||
this->streamSource(subsession->getStreamSource(m_StreamToken));
|
||||
|
||||
// pointer to subsession to close it
|
||||
fSubsession = subsession;
|
||||
m_Subsession = subsession;
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -385,7 +378,7 @@ HTTPServer::HTTPClientConnection::~HTTPClientConnection()
|
||||
{
|
||||
this->streamSource(NULL);
|
||||
|
||||
if (fSubsession) {
|
||||
fSubsession->deleteStream(fClientSessionId, fStreamToken);
|
||||
if (m_Subsession) {
|
||||
m_Subsession->deleteStream(m_ClientSessionId, m_StreamToken);
|
||||
}
|
||||
}
|
||||
|
@ -12,7 +12,7 @@
|
||||
// -----------------------------------------
|
||||
// MemoryBufferSink
|
||||
// -----------------------------------------
|
||||
MemoryBufferSink::MemoryBufferSink(UsageEnvironment& env, unsigned bufferSize, unsigned int sliceDuration) : MediaSink(env), m_bufferSize(bufferSize), m_refTime(0), m_sliceDuration(sliceDuration)
|
||||
MemoryBufferSink::MemoryBufferSink(UsageEnvironment& env, unsigned bufferSize, unsigned int sliceDuration, unsigned int nbSlices) : MediaSink(env), m_bufferSize(bufferSize), m_refTime(0), m_sliceDuration(sliceDuration), m_nbSlices(nbSlices)
|
||||
{
|
||||
m_buffer = new unsigned char[m_bufferSize];
|
||||
}
|
||||
@ -59,7 +59,7 @@ void MemoryBufferSink::afterGettingFrame(unsigned frameSize, unsigned numTruncat
|
||||
outputBuffer.append((const char*)m_buffer, frameSize);
|
||||
|
||||
// remove old buffers
|
||||
while (m_outputBuffers.size()>3)
|
||||
while (m_outputBuffers.size()>m_nbSlices)
|
||||
{
|
||||
m_outputBuffers.erase(m_outputBuffers.begin());
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user