fix build with latest live555 that removed TCPStreamSink

pull/191/head
mpromonet 4 years ago
parent eaf0608aba
commit be67798772

@ -5,7 +5,7 @@ project(v4l2rtspserver)
option(COVERAGE "Coverage" OFF) option(COVERAGE "Coverage" OFF)
set(ALSA ON CACHE BOOL "use ALSA is available") set(ALSA ON CACHE BOOL "use ALSA is available")
set(LIVE555URL https://download.videolan.org/pub/contrib/live555/live.2020.03.06.tar.gz CACHE STRING "live555 url") set(LIVE555URL http://www.live555.com/liveMedia/public/live555-latest.tar.gz CACHE STRING "live555 url")
set(LIVE555CFLAGS -DBSD=1 -DSOCKLEN_T=socklen_t -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -DALLOW_RTSP_SERVER_PORT_REUSE=1 -DNO_OPENSSL=1 CACHE STRING "live555 CFGLAGS") set(LIVE555CFLAGS -DBSD=1 -DSOCKLEN_T=socklen_t -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE=1 -DALLOW_RTSP_SERVER_PORT_REUSE=1 -DNO_OPENSSL=1 CACHE STRING "live555 CFGLAGS")
set(CMAKE_CXX_STANDARD 11) set(CMAKE_CXX_STANDARD 11)

@ -14,6 +14,104 @@
#include "RTSPServer.hh" #include "RTSPServer.hh"
#include "RTSPCommon.hh" #include "RTSPCommon.hh"
#include <GroupsockHelper.hh> // for "ignoreSigPipeOnSocket()"
#define TCP_STREAM_SINK_MIN_READ_SIZE 1000
#define TCP_STREAM_SINK_BUFFER_SIZE 10000
class TCPStreamSink: public MediaSink {
public:
TCPStreamSink(UsageEnvironment& env, int socketNum) : MediaSink(env), fUnwrittenBytesStart(0), fUnwrittenBytesEnd(0), fInputSourceIsOpen(False), fOutputSocketIsWritable(True),fOutputSocketNum(socketNum) {
ignoreSigPipeOnSocket(socketNum);
}
protected:
virtual ~TCPStreamSink() {
envir().taskScheduler().disableBackgroundHandling(fOutputSocketNum);
if (fSource != NULL) {
Medium::close(fSource);
}
}
protected:
virtual Boolean continuePlaying() {
fInputSourceIsOpen = fSource != NULL;
processBuffer();
return True;
}
private:
void processBuffer(){
// First, try writing data to our output socket, if we can:
if (fOutputSocketIsWritable && numUnwrittenBytes() > 0) {
int numBytesWritten = send(fOutputSocketNum, (const char*)&fBuffer[fUnwrittenBytesStart], numUnwrittenBytes(), 0);
if (numBytesWritten < (int)numUnwrittenBytes()) {
// The output socket is no longer writable. Set a handler to be called when it becomes writable again.
fOutputSocketIsWritable = False;
if (envir().getErrno() != EPIPE) { // on this error, the socket might still be writable, but no longer usable
envir().taskScheduler().setBackgroundHandling(fOutputSocketNum, SOCKET_WRITABLE, socketWritableHandler, this);
}
}
if (numBytesWritten > 0) {
// We wrote at least some of our data. Update our buffer pointers:
fUnwrittenBytesStart += numBytesWritten;
if (fUnwrittenBytesStart > fUnwrittenBytesEnd) fUnwrittenBytesStart = fUnwrittenBytesEnd; // sanity check
if (fUnwrittenBytesStart == fUnwrittenBytesEnd && (!fInputSourceIsOpen || !fSource->isCurrentlyAwaitingData())) {
fUnwrittenBytesStart = fUnwrittenBytesEnd = 0; // reset the buffer to empty
}
}
}
// Then, read from our input source, if we can (& we're not already reading from it):
if (fInputSourceIsOpen && freeBufferSpace() >= TCP_STREAM_SINK_MIN_READ_SIZE && !fSource->isCurrentlyAwaitingData()) {
fSource->getNextFrame(&fBuffer[fUnwrittenBytesEnd], freeBufferSpace(), afterGettingFrame, this, ourOnSourceClosure, this);
} else if (!fInputSourceIsOpen && numUnwrittenBytes() == 0) {
// We're now done:
onSourceClosure();
}
}
static void socketWritableHandler(void* clientData, int mask) { ((TCPStreamSink*)clientData)->socketWritableHandler(); }
void socketWritableHandler() {
envir().taskScheduler().disableBackgroundHandling(fOutputSocketNum); // disable this handler until the next time it's needed
fOutputSocketIsWritable = True;
processBuffer();
}
static void afterGettingFrame(void* clientData, unsigned frameSize, unsigned numTruncatedBytes,
struct timeval presentationTime, unsigned durationInMicroseconds) {
((TCPStreamSink*)clientData)->afterGettingFrame(frameSize, numTruncatedBytes);
}
void afterGettingFrame(unsigned frameSize, unsigned numTruncatedBytes) {
if (numTruncatedBytes > 0) {
envir() << "TCPStreamSink::afterGettingFrame(): The input frame data was too large for our buffer. "
<< numTruncatedBytes
<< " bytes of trailing data was dropped! Correct this by increasing the definition of \"TCP_STREAM_SINK_BUFFER_SIZE\" in \"include/TCPStreamSink.hh\".\n";
}
fUnwrittenBytesEnd += frameSize;
processBuffer();
}
static void ourOnSourceClosure(void* clientData) { ((TCPStreamSink*)clientData)->ourOnSourceClosure(); }
void ourOnSourceClosure() {
// The input source has closed:
fInputSourceIsOpen = False;
processBuffer();
}
unsigned numUnwrittenBytes() const { return fUnwrittenBytesEnd - fUnwrittenBytesStart; }
unsigned freeBufferSpace() const { return TCP_STREAM_SINK_BUFFER_SIZE - fUnwrittenBytesEnd; }
private:
unsigned char fBuffer[TCP_STREAM_SINK_BUFFER_SIZE];
unsigned fUnwrittenBytesStart, fUnwrittenBytesEnd;
Boolean fInputSourceIsOpen, fOutputSocketIsWritable;
int fOutputSocketNum;
};
// --------------------------------------------------------- // ---------------------------------------------------------
// Extend RTSP server to add support for HLS and MPEG-DASH // Extend RTSP server to add support for HLS and MPEG-DASH
@ -25,7 +123,7 @@ class HTTPServer : public RTSPServer
{ {
public: public:
HTTPClientConnection(RTSPServer& ourServer, int clientSocket, struct sockaddr_in clientAddr) HTTPClientConnection(RTSPServer& ourServer, int clientSocket, struct sockaddr_in clientAddr)
: RTSPServer::RTSPClientConnection(ourServer, clientSocket, clientAddr), fTCPSink(NULL), fStreamToken(NULL), fSubsession(NULL), fSource(NULL) { : RTSPServer::RTSPClientConnection(ourServer, clientSocket, clientAddr), m_TCPSink(NULL), m_StreamToken(NULL), m_Subsession(NULL) {
} }
virtual ~HTTPClientConnection(); virtual ~HTTPClientConnection();
@ -43,11 +141,10 @@ class HTTPServer : public RTSPServer
static void afterStreaming(void* clientData); static void afterStreaming(void* clientData);
private: private:
static u_int32_t fClientSessionId; static u_int32_t m_ClientSessionId;
TCPStreamSink* fTCPSink; TCPStreamSink* m_TCPSink;
void* fStreamToken; void* m_StreamToken;
ServerMediaSubsession* fSubsession; ServerMediaSubsession* m_Subsession;
FramedSource* fSource;
}; };
public: public:

@ -19,13 +19,13 @@
class MemoryBufferSink : public MediaSink class MemoryBufferSink : public MediaSink
{ {
public: public:
static MemoryBufferSink* createNew(UsageEnvironment& env, unsigned int bufferSize, unsigned int sliceDuration) static MemoryBufferSink* createNew(UsageEnvironment& env, unsigned int bufferSize, unsigned int sliceDuration, unsigned int nbSlices = 5)
{ {
return new MemoryBufferSink(env, bufferSize, sliceDuration); return new MemoryBufferSink(env, bufferSize, sliceDuration, nbSlices);
} }
protected: protected:
MemoryBufferSink(UsageEnvironment& env, unsigned bufferSize, unsigned int sliceDuration); MemoryBufferSink(UsageEnvironment& env, unsigned bufferSize, unsigned int sliceDuration, unsigned int nbSlices);
virtual ~MemoryBufferSink(); virtual ~MemoryBufferSink();
virtual Boolean continuePlaying(); virtual Boolean continuePlaying();
@ -53,5 +53,6 @@ class MemoryBufferSink : public MediaSink
std::map<unsigned int,std::string> m_outputBuffers; std::map<unsigned int,std::string> m_outputBuffers;
unsigned int m_refTime; unsigned int m_refTime;
unsigned int m_sliceDuration; unsigned int m_sliceDuration;
unsigned int m_nbSlices;
}; };

@ -102,6 +102,7 @@ void* V4L2DeviceSource::thread()
{ {
if (FD_ISSET(fd, &fdset)) if (FD_ISSET(fd, &fdset))
{ {
LOG(DEBUG) << "waitingFrame\tdelay:" << (1000-(tv.tv_usec/1000)) << "ms";
if (this->getNextFrame() <= 0) if (this->getNextFrame() <= 0)
{ {
LOG(ERROR) << "error:" << strerror(errno); LOG(ERROR) << "error:" << strerror(errno);
@ -213,6 +214,7 @@ int V4L2DeviceSource::getNextFrame()
timersub(&tv,&ref,&diff); timersub(&tv,&ref,&diff);
m_in.notify(tv.tv_sec, frameSize); m_in.notify(tv.tv_sec, frameSize);
LOG(DEBUG) << "getNextFrame\ttimestamp:" << ref.tv_sec << "." << ref.tv_usec << "\tsize:" << frameSize <<"\tdiff:" << (diff.tv_sec*1000+diff.tv_usec/1000) << "ms"; LOG(DEBUG) << "getNextFrame\ttimestamp:" << ref.tv_sec << "." << ref.tv_usec << "\tsize:" << frameSize <<"\tdiff:" << (diff.tv_sec*1000+diff.tv_usec/1000) << "ms";
processFrame(buffer,frameSize,ref); processFrame(buffer,frameSize,ref);
if (m_outfd != -1) if (m_outfd != -1)
{ {
@ -238,9 +240,9 @@ void V4L2DeviceSource::processFrame(char * frame, int frameSize, const timeval &
char* buf = new char[size]; char* buf = new char[size];
memcpy(buf, frame.first, size); memcpy(buf, frame.first, size);
queueFrame(buf,size,ref); queueFrame(buf,size,ref);
frameList.pop_front();
LOG(DEBUG) << "queueFrame\ttimestamp:" << ref.tv_sec << "." << ref.tv_usec << "\tsize:" << size <<"\tdiff:" << (diff.tv_sec*1000+diff.tv_usec/1000) << "ms"; LOG(DEBUG) << "queueFrame\ttimestamp:" << ref.tv_sec << "." << ref.tv_usec << "\tsize:" << size <<"\tdiff:" << (diff.tv_sec*1000+diff.tv_usec/1000) << "ms";
frameList.pop_front();
} }
} }

@ -20,11 +20,10 @@
#include "RTSPCommon.hh" #include "RTSPCommon.hh"
#include <time.h> #include <time.h>
#include "ByteStreamMemoryBufferSource.hh" #include "ByteStreamMemoryBufferSource.hh"
#include "TCPStreamSink.hh"
#include "HTTPServer.h" #include "HTTPServer.h"
u_int32_t HTTPServer::HTTPClientConnection::fClientSessionId = 0; u_int32_t HTTPServer::HTTPClientConnection::m_ClientSessionId = 0;
void HTTPServer::HTTPClientConnection::sendHeader(const char* contentType, unsigned int contentLength) void HTTPServer::HTTPClientConnection::sendHeader(const char* contentType, unsigned int contentLength)
{ {
@ -56,22 +55,16 @@ void HTTPServer::HTTPClientConnection::streamSource(const std::string & content)
void HTTPServer::HTTPClientConnection::streamSource(FramedSource* source) void HTTPServer::HTTPClientConnection::streamSource(FramedSource* source)
{ {
if (fTCPSink != NULL) if (m_TCPSink != NULL)
{ {
fTCPSink->stopPlaying(); m_TCPSink->stopPlaying();
Medium::close(fTCPSink); Medium::close(m_TCPSink);
fTCPSink = NULL; m_TCPSink = NULL;
}
if (fSource != NULL)
{
Medium::close(fSource);
fSource = NULL;
} }
if (source != NULL) if (source != NULL)
{ {
fTCPSink = TCPStreamSink::createNew(envir(), fClientOutputSocket); m_TCPSink = new TCPStreamSink(envir(), fClientOutputSocket);
fTCPSink->startPlaying(*source, afterStreaming, this); m_TCPSink->startPlaying(*source, afterStreaming, this);
fSource = source; // we need to keep tracking of source, because sink do not release it
} }
} }
@ -322,17 +315,17 @@ void HTTPServer::HTTPClientConnection::handleHTTPCmd_StreamingGET(char const* ur
// Call "getStreamParameters()" to create the stream's source. (Because we're not actually streaming via RTP/RTCP, most // Call "getStreamParameters()" to create the stream's source. (Because we're not actually streaming via RTP/RTCP, most
// of the parameters to the call are dummy.) // of the parameters to the call are dummy.)
++fClientSessionId; ++m_ClientSessionId;
Port clientRTPPort(0), clientRTCPPort(0), serverRTPPort(0), serverRTCPPort(0); Port clientRTPPort(0), clientRTCPPort(0), serverRTPPort(0), serverRTCPPort(0);
netAddressBits destinationAddress = 0; netAddressBits destinationAddress = 0;
u_int8_t destinationTTL = 0; u_int8_t destinationTTL = 0;
Boolean isMulticast = False; Boolean isMulticast = False;
subsession->getStreamParameters(fClientSessionId, 0, clientRTPPort,clientRTCPPort, -1,0,0, destinationAddress,destinationTTL, isMulticast, serverRTPPort,serverRTCPPort, fStreamToken); subsession->getStreamParameters(m_ClientSessionId, 0, clientRTPPort,clientRTCPPort, -1,0,0, destinationAddress,destinationTTL, isMulticast, serverRTPPort,serverRTCPPort, m_StreamToken);
// Seek the stream source to the desired place, with the desired duration, and (as a side effect) get the number of bytes: // Seek the stream source to the desired place, with the desired duration, and (as a side effect) get the number of bytes:
double dOffsetInSeconds = (double)offsetInSeconds; double dOffsetInSeconds = (double)offsetInSeconds;
u_int64_t numBytes = 0; u_int64_t numBytes = 0;
subsession->seekStream(fClientSessionId, fStreamToken, dOffsetInSeconds, 0.0, numBytes); subsession->seekStream(m_ClientSessionId, m_StreamToken, dOffsetInSeconds, 0.0, numBytes);
if (numBytes == 0) if (numBytes == 0)
{ {
@ -346,10 +339,10 @@ void HTTPServer::HTTPClientConnection::handleHTTPCmd_StreamingGET(char const* ur
this->sendHeader("video/mp2t", numBytes); this->sendHeader("video/mp2t", numBytes);
// stream body // stream body
this->streamSource(subsession->getStreamSource(fStreamToken)); this->streamSource(subsession->getStreamSource(m_StreamToken));
// pointer to subsession to close it // pointer to subsession to close it
fSubsession = subsession; m_Subsession = subsession;
} }
} }
} }
@ -385,7 +378,7 @@ HTTPServer::HTTPClientConnection::~HTTPClientConnection()
{ {
this->streamSource(NULL); this->streamSource(NULL);
if (fSubsession) { if (m_Subsession) {
fSubsession->deleteStream(fClientSessionId, fStreamToken); m_Subsession->deleteStream(m_ClientSessionId, m_StreamToken);
} }
} }

@ -12,7 +12,7 @@
// ----------------------------------------- // -----------------------------------------
// MemoryBufferSink // MemoryBufferSink
// ----------------------------------------- // -----------------------------------------
MemoryBufferSink::MemoryBufferSink(UsageEnvironment& env, unsigned bufferSize, unsigned int sliceDuration) : MediaSink(env), m_bufferSize(bufferSize), m_refTime(0), m_sliceDuration(sliceDuration) MemoryBufferSink::MemoryBufferSink(UsageEnvironment& env, unsigned bufferSize, unsigned int sliceDuration, unsigned int nbSlices) : MediaSink(env), m_bufferSize(bufferSize), m_refTime(0), m_sliceDuration(sliceDuration), m_nbSlices(nbSlices)
{ {
m_buffer = new unsigned char[m_bufferSize]; m_buffer = new unsigned char[m_bufferSize];
} }
@ -59,7 +59,7 @@ void MemoryBufferSink::afterGettingFrame(unsigned frameSize, unsigned numTruncat
outputBuffer.append((const char*)m_buffer, frameSize); outputBuffer.append((const char*)m_buffer, frameSize);
// remove old buffers // remove old buffers
while (m_outputBuffers.size()>3) while (m_outputBuffers.size()>m_nbSlices)
{ {
m_outputBuffers.erase(m_outputBuffers.begin()); m_outputBuffers.erase(m_outputBuffers.begin());
} }

Loading…
Cancel
Save