Code cleanup and cpplint

pull/2653/head
Isaac Connor 2019-07-02 11:14:24 -04:00
parent 268bc2264b
commit 2dc60196c5
1 changed files with 237 additions and 213 deletions

View File

@ -1,5 +1,5 @@
// //
// ZoneMinder Ffmpeg Camera Class Implementation, $Date: 2009-01-16 12:18:50 +0000 (Fri, 16 Jan 2009) $, $Revision: 2713 $ // ZoneMinder Ffmpeg Camera Class Implementation
// Copyright (C) 2001-2008 Philip Coombes // Copyright (C) 2001-2008 Philip Coombes
// //
// This program is free software; you can redistribute it and/or // This program is free software; you can redistribute it and/or
@ -39,6 +39,7 @@ extern "C" {
#include <signal.h> #include <signal.h>
#include <pthread.h> #include <pthread.h>
#endif #endif
#include <string>
#if HAVE_LIBAVUTIL_HWCONTEXT_H #if HAVE_LIBAVUTIL_HWCONTEXT_H
@ -54,9 +55,11 @@ static enum AVPixelFormat get_hw_format(
return *p; return *p;
} }
Error("Failed to get HW surface format for %s.", av_get_pix_fmt_name(hw_pix_fmt)); Error("Failed to get HW surface format for %s.",
av_get_pix_fmt_name(hw_pix_fmt));
for ( p = pix_fmts; *p != -1; p++ ) for ( p = pix_fmts; *p != -1; p++ )
Error("Available HW surface format was %s.", av_get_pix_fmt_name(*p)); Error("Available HW surface format was %s.",
av_get_pix_fmt_name(*p));
return AV_PIX_FMT_NONE; return AV_PIX_FMT_NONE;
} }
@ -106,8 +109,7 @@ FfmpegCamera::FfmpegCamera(
bool p_capture, bool p_capture,
bool p_record_audio, bool p_record_audio,
const std::string &p_hwaccel_name, const std::string &p_hwaccel_name,
const std::string &p_hwaccel_device const std::string &p_hwaccel_device) :
) :
Camera( Camera(
p_id, p_id,
FFMPEG_SRC, FFMPEG_SRC,
@ -157,7 +159,8 @@ FfmpegCamera::FfmpegCamera(
#if HAVE_LIBSWSCALE #if HAVE_LIBSWSCALE
mConvertContext = NULL; mConvertContext = NULL;
#endif #endif
/* Has to be located inside the constructor so other components such as zma will receive correct colours and subpixel order */ /* Has to be located inside the constructor so other components such as zma
* will receive correct colours and subpixel order */
if ( colours == ZM_COLOUR_RGB32 ) { if ( colours == ZM_COLOUR_RGB32 ) {
subpixelorder = ZM_SUBPIX_ORDER_RGBA; subpixelorder = ZM_SUBPIX_ORDER_RGBA;
imagePixFormat = AV_PIX_FMT_RGBA; imagePixFormat = AV_PIX_FMT_RGBA;
@ -173,7 +176,6 @@ FfmpegCamera::FfmpegCamera(
} // FfmpegCamera::FfmpegCamera } // FfmpegCamera::FfmpegCamera
FfmpegCamera::~FfmpegCamera() { FfmpegCamera::~FfmpegCamera() {
Close(); Close();
if ( capture ) { if ( capture ) {
@ -203,7 +205,7 @@ int FfmpegCamera::PrimeCapture() {
int FfmpegCamera::PreCapture() { int FfmpegCamera::PreCapture() {
// If Reopen was called, then ffmpeg is closed and we need to reopen it. // If Reopen was called, then ffmpeg is closed and we need to reopen it.
if ( ! mCanCapture ) if ( !mCanCapture )
return OpenFfmpeg(); return OpenFfmpeg();
// Nothing to do here // Nothing to do here
return 0; return 0;
@ -215,15 +217,20 @@ int FfmpegCamera::Capture(Image &image) {
} }
int ret; int ret;
// If the reopen thread has a value, but mCanCapture != 0, then we have just reopened the connection to the ffmpeg device, and we can clean up the thread. // If the reopen thread has a value, but mCanCapture != 0, then we have just
// reopened the connection to the device, and we can clean up the thread.
int frameComplete = false; int frameComplete = false;
while ( !frameComplete && !zm_terminate) { while ( !frameComplete && !zm_terminate ) {
ret = av_read_frame(mFormatContext, &packet); ret = av_read_frame(mFormatContext, &packet);
if ( ret < 0 ) { if ( ret < 0 ) {
if ( if (
// Check if EOF. // Check if EOF.
(ret == AVERROR_EOF || (mFormatContext->pb && mFormatContext->pb->eof_reached)) || (
ret == AVERROR_EOF
||
(mFormatContext->pb && mFormatContext->pb->eof_reached)
) ||
// Check for Connection failure. // Check for Connection failure.
(ret == -110) (ret == -110)
) { ) {
@ -244,7 +251,11 @@ int FfmpegCamera::Capture(Image &image) {
Debug(5, "Got packet from stream %d dts (%d) pts(%d)", Debug(5, "Got packet from stream %d dts (%d) pts(%d)",
packet.stream_index, packet.pts, packet.dts); packet.stream_index, packet.pts, packet.dts);
// What about audio stream? Maybe someday we could do sound detection... // What about audio stream? Maybe someday we could do sound detection...
if ( ( packet.stream_index == mVideoStreamId ) && ( keyframe || have_video_keyframe ) ) { if (
(packet.stream_index == mVideoStreamId)
&&
(keyframe || have_video_keyframe)
) {
ret = zm_receive_frame(mVideoCodecContext, mRawFrame, packet); ret = zm_receive_frame(mVideoCodecContext, mRawFrame, packet);
if ( ret < 0 ) { if ( ret < 0 ) {
Error("Unable to get frame at frame %d: %s, continuing", Error("Unable to get frame at frame %d: %s, continuing",
@ -276,7 +287,6 @@ int FfmpegCamera::PostCapture() {
} }
int FfmpegCamera::OpenFfmpeg() { int FfmpegCamera::OpenFfmpeg() {
int ret; int ret;
have_video_keyframe = false; have_video_keyframe = false;
@ -290,7 +300,7 @@ int FfmpegCamera::OpenFfmpeg() {
AVDictionary *opts = 0; AVDictionary *opts = 0;
ret = av_dict_parse_string(&opts, Options().c_str(), "=", ",", 0); ret = av_dict_parse_string(&opts, Options().c_str(), "=", ",", 0);
if ( ret < 0 ) { if ( ret < 0 ) {
Warning("Could not parse ffmpeg input options list '%s'\n", Options().c_str()); Warning("Could not parse ffmpeg input options '%s'", Options().c_str());
} }
// Set transport method as specified by method field, rtpUni is default // Set transport method as specified by method field, rtpUni is default
@ -306,7 +316,7 @@ int FfmpegCamera::OpenFfmpeg() {
} else { } else {
Warning("Unknown method (%s)", method.c_str()); Warning("Unknown method (%s)", method.c_str());
} }
//#av_dict_set(&opts, "timeout", "10000000", 0); // in microseconds. // #av_dict_set(&opts, "timeout", "10000000", 0); // in microseconds.
if ( ret < 0 ) { if ( ret < 0 ) {
Warning("Could not set rtsp_transport method '%s'", method.c_str()); Warning("Could not set rtsp_transport method '%s'", method.c_str());
@ -314,11 +324,11 @@ int FfmpegCamera::OpenFfmpeg() {
Debug(1, "Calling avformat_open_input for %s", mPath.c_str()); Debug(1, "Calling avformat_open_input for %s", mPath.c_str());
mFormatContext = avformat_alloc_context( ); mFormatContext = avformat_alloc_context();
// Speed up find_stream_info // Speed up find_stream_info
//FIXME can speed up initial analysis but need sensible parameters... // FIXME can speed up initial analysis but need sensible parameters...
//mFormatContext->probesize = 32; // mFormatContext->probesize = 32;
//mFormatContext->max_analyze_duration = 32; // mFormatContext->max_analyze_duration = 32;
mFormatContext->interrupt_callback.callback = FfmpegInterruptCallback; mFormatContext->interrupt_callback.callback = FfmpegInterruptCallback;
mFormatContext->interrupt_callback.opaque = this; mFormatContext->interrupt_callback.opaque = this;
@ -326,7 +336,8 @@ int FfmpegCamera::OpenFfmpeg() {
if ( ret != 0 ) if ( ret != 0 )
#endif #endif
{ {
Error("Unable to open input %s due to: %s", mPath.c_str(), strerror(ret)); Error("Unable to open input %s due to: %s", mPath.c_str(),
av_make_error_string(ret).c_str());
#if !LIBAVFORMAT_VERSION_CHECK(53, 17, 0, 25, 0) #if !LIBAVFORMAT_VERSION_CHECK(53, 17, 0, 25, 0)
av_close_input_file(mFormatContext); av_close_input_file(mFormatContext);
#else #else
@ -339,7 +350,7 @@ int FfmpegCamera::OpenFfmpeg() {
return -1; return -1;
} }
AVDictionaryEntry *e=NULL; AVDictionaryEntry *e = NULL;
while ( (e = av_dict_get(opts, "", e, AV_DICT_IGNORE_SUFFIX)) != NULL ) { while ( (e = av_dict_get(opts, "", e, AV_DICT_IGNORE_SUFFIX)) != NULL ) {
Warning("Option %s not recognized by ffmpeg", e->key); Warning("Option %s not recognized by ffmpeg", e->key);
} }
@ -348,94 +359,76 @@ int FfmpegCamera::OpenFfmpeg() {
Info("Stream open %s, parsing streams...", mPath.c_str()); Info("Stream open %s, parsing streams...", mPath.c_str());
#if !LIBAVFORMAT_VERSION_CHECK(53, 6, 0, 6, 0) #if !LIBAVFORMAT_VERSION_CHECK(53, 6, 0, 6, 0)
Debug(4, "Calling av_find_stream_info"); ret = av_find_stream_info(mFormatContext);
if ( av_find_stream_info(mFormatContext) < 0 )
#else #else
Debug(4, "Calling avformat_find_stream_info"); ret = avformat_find_stream_info(mFormatContext, 0);
if ( avformat_find_stream_info(mFormatContext, 0) < 0 )
#endif #endif
{ if ( ret < 0 ) {
Error("Unable to find stream info from %s due to: %s", mPath.c_str(), strerror(errno)); Error("Unable to find stream info from %s due to: %s", mPath.c_str(),
av_make_error_string(ret).c_str());
return -1; return -1;
} }
Debug(4, "Got stream info");
// Find first video stream present // Find first video stream present
// The one we want Might not be the first // The one we want Might not be the first
mVideoStreamId = -1; mVideoStreamId = -1;
mAudioStreamId = -1; mAudioStreamId = -1;
for ( unsigned int i=0; i < mFormatContext->nb_streams; i++ ) { for ( unsigned int i=0; i < mFormatContext->nb_streams; i++ ) {
#if LIBAVCODEC_VERSION_CHECK(57, 64, 0, 64, 0) AVStream *stream = mFormatContext->streams[i];
if ( mFormatContext->streams[i]->codecpar->codec_type == AVMEDIA_TYPE_VIDEO ) { if ( is_video_stream(stream) ) {
#else
#if (LIBAVCODEC_VERSION_CHECK(52, 64, 0, 64, 0) || LIBAVUTIL_VERSION_CHECK(50, 14, 0, 14, 0))
if ( mFormatContext->streams[i]->codec->codec_type == AVMEDIA_TYPE_VIDEO ) {
#else
if ( mFormatContext->streams[i]->codec->codec_type == CODEC_TYPE_VIDEO ) {
#endif
#endif
if ( mVideoStreamId == -1 ) { if ( mVideoStreamId == -1 ) {
mVideoStreamId = i; mVideoStreamId = i;
// if we break, then we won't find the audio stream // if we break, then we won't find the audio stream
continue; continue;
} else { } else {
Debug(2, "Have another video stream." ); Debug(2, "Have another video stream.");
} }
} } else if ( is_audio_stream(stream) ) {
#if LIBAVCODEC_VERSION_CHECK(57, 64, 0, 64, 0)
if ( mFormatContext->streams[i]->codecpar->codec_type == AVMEDIA_TYPE_AUDIO ) {
#else
#if (LIBAVCODEC_VERSION_CHECK(52, 64, 0, 64, 0) || LIBAVUTIL_VERSION_CHECK(50, 14, 0, 14, 0))
if ( mFormatContext->streams[i]->codec->codec_type == AVMEDIA_TYPE_AUDIO ) {
#else
if ( mFormatContext->streams[i]->codec->codec_type == CODEC_TYPE_AUDIO ) {
#endif
#endif
if ( mAudioStreamId == -1 ) { if ( mAudioStreamId == -1 ) {
mAudioStreamId = i; mAudioStreamId = i;
} else { } else {
Debug(2, "Have another audio stream." ); Debug(2, "Have another audio stream.");
} }
} }
} // end foreach stream } // end foreach stream
if ( mVideoStreamId == -1 ) if ( mVideoStreamId == -1 )
Fatal("Unable to locate video stream in %s", mPath.c_str()); Fatal("Unable to locate video stream in %s", mPath.c_str());
if ( mAudioStreamId == -1 )
Debug(3, "Unable to locate audio stream in %s", mPath.c_str());
Debug(3, "Found video stream at index %d", mVideoStreamId); Debug(3, "Found video stream at index %d, audio stream at index %d",
Debug(3, "Found audio stream at index %d", mAudioStreamId); mVideoStreamId, mAudioStreamId);
packetqueue = new zm_packetqueue( mVideoStreamId > mAudioStreamId ? mVideoStreamId : mAudioStreamId ); packetqueue = new zm_packetqueue(
(mVideoStreamId > mAudioStreamId) ? mVideoStreamId : mAudioStreamId);
#if LIBAVCODEC_VERSION_CHECK(57, 64, 0, 64, 0) #if LIBAVCODEC_VERSION_CHECK(57, 64, 0, 64, 0)
//mVideoCodecContext = avcodec_alloc_context3(NULL); // mVideoCodecContext = avcodec_alloc_context3(NULL);
//avcodec_parameters_to_context( mVideoCodecContext, mFormatContext->streams[mVideoStreamId]->codecpar ); // avcodec_parameters_to_context(mVideoCodecContext,
// mFormatContext->streams[mVideoStreamId]->codecpar);
// this isn't copied. // this isn't copied.
//mVideoCodecContext->time_base = mFormatContext->streams[mVideoStreamId]->codec->time_base; // mVideoCodecContext->time_base =
// mFormatContext->streams[mVideoStreamId]->codec->time_base;
#else #else
#endif #endif
mVideoCodecContext = mFormatContext->streams[mVideoStreamId]->codec; mVideoCodecContext = mFormatContext->streams[mVideoStreamId]->codec;
// STolen from ispy
//this fixes issues with rtsp streams!! woot.
//mVideoCodecContext->flags2 |= CODEC_FLAG2_FAST | CODEC_FLAG2_CHUNKS | CODEC_FLAG_LOW_DELAY; // Enable faster H264 decode.
#ifdef CODEC_FLAG2_FAST #ifdef CODEC_FLAG2_FAST
mVideoCodecContext->flags2 |= CODEC_FLAG2_FAST | CODEC_FLAG_LOW_DELAY; mVideoCodecContext->flags2 |= CODEC_FLAG2_FAST | CODEC_FLAG_LOW_DELAY;
#endif #endif
if ( mVideoCodecContext->codec_id == AV_CODEC_ID_H264 ) { if ( mVideoCodecContext->codec_id == AV_CODEC_ID_H264 ) {
if ( (mVideoCodec = avcodec_find_decoder_by_name("h264_mmal")) == NULL ) { if ( (mVideoCodec = avcodec_find_decoder_by_name("h264_mmal")) == NULL ) {
Debug(1, "Failed to find decoder (h264_mmal)" ); Debug(1, "Failed to find decoder (h264_mmal)");
} else { } else {
Debug(1, "Success finding decoder (h264_mmal)" ); Debug(1, "Success finding decoder (h264_mmal)");
} }
} }
if ( (!mVideoCodec) and ( (mVideoCodec = avcodec_find_decoder(mVideoCodecContext->codec_id)) == NULL ) ) { if ( !mVideoCodec ) {
mVideoCodec = avcodec_find_decoder(mVideoCodecContext->codec_id);
if ( !mVideoCodec ) {
// Try and get the codec from the codec context // Try and get the codec from the codec context
Error("Can't find codec for video stream from %s", mPath.c_str()); Error("Can't find codec for video stream from %s", mPath.c_str());
return -1; return -1;
} }
}
Debug(1, "Video Found decoder %s", mVideoCodec->name); Debug(1, "Video Found decoder %s", mVideoCodec->name);
zm_dump_stream_format(mFormatContext, mVideoStreamId, 0, 0); zm_dump_stream_format(mFormatContext, mVideoStreamId, 0, 0);
@ -464,7 +457,7 @@ int FfmpegCamera::OpenFfmpeg() {
mVideoCodec->name, av_hwdevice_get_type_name(type)); mVideoCodec->name, av_hwdevice_get_type_name(type));
return -1; return -1;
} }
if ( (config->methods & AV_CODEC_HW_CONFIG_METHOD_HW_DEVICE_CTX ) if ( (config->methods & AV_CODEC_HW_CONFIG_METHOD_HW_DEVICE_CTX)
&& (config->device_type == type) && (config->device_type == type)
) { ) {
hw_pix_fmt = config->pix_fmt; hw_pix_fmt = config->pix_fmt;
@ -474,13 +467,17 @@ int FfmpegCamera::OpenFfmpeg() {
#else #else
hw_pix_fmt = find_fmt_by_hw_type(type); hw_pix_fmt = find_fmt_by_hw_type(type);
#endif #endif
Debug(1, "Selected gw_pix_fmt %d %s", hw_pix_fmt, av_get_pix_fmt_name(hw_pix_fmt)); Debug(1, "Selected gw_pix_fmt %d %s",
hw_pix_fmt,
av_get_pix_fmt_name(hw_pix_fmt));
mVideoCodecContext->get_format = get_hw_format; mVideoCodecContext->get_format = get_hw_format;
Debug(1, "Creating hwdevice for %s", (hwaccel_device != "" ? hwaccel_device.c_str() : "")); Debug(1, "Creating hwdevice for %s",
if ((ret = av_hwdevice_ctx_create(&hw_device_ctx, type, (hwaccel_device != "" ? hwaccel_device.c_str() : ""));
(hwaccel_device != "" ? hwaccel_device.c_str(): NULL), NULL, 0)) < 0) { ret = av_hwdevice_ctx_create(&hw_device_ctx, type,
(hwaccel_device != "" ? hwaccel_device.c_str(): NULL), NULL, 0);
if ( ret < 0 ) {
Error("Failed to create specified HW device."); Error("Failed to create specified HW device.");
return -1; return -1;
} }
@ -501,7 +498,7 @@ Debug(1, "Selected gw_pix_fmt %d %s", hw_pix_fmt, av_get_pix_fmt_name(hw_pix_fmt
#endif #endif
e = NULL; e = NULL;
while ( (e = av_dict_get(opts, "", e, AV_DICT_IGNORE_SUFFIX)) != NULL ) { while ( (e = av_dict_get(opts, "", e, AV_DICT_IGNORE_SUFFIX)) != NULL ) {
Warning( "Option %s not recognized by ffmpeg", e->key); Warning("Option %s not recognized by ffmpeg", e->key);
} }
if ( ret < 0 ) { if ( ret < 0 ) {
Error("Unable to open codec for video stream from %s", mPath.c_str()); Error("Unable to open codec for video stream from %s", mPath.c_str());
@ -528,23 +525,23 @@ Debug(1, "Selected gw_pix_fmt %d %s", hw_pix_fmt, av_get_pix_fmt_name(hw_pix_fmt
} else { } else {
#if LIBAVCODEC_VERSION_CHECK(57, 64, 0, 64, 0) #if LIBAVCODEC_VERSION_CHECK(57, 64, 0, 64, 0)
mAudioCodecContext = avcodec_alloc_context3(mAudioCodec); mAudioCodecContext = avcodec_alloc_context3(mAudioCodec);
avcodec_parameters_to_context( mAudioCodecContext, mFormatContext->streams[mAudioStreamId]->codecpar ); avcodec_parameters_to_context(
mAudioCodecContext,
mFormatContext->streams[mAudioStreamId]->codecpar
);
#else #else
mAudioCodecContext = mFormatContext->streams[mAudioStreamId]->codec; mAudioCodecContext = mFormatContext->streams[mAudioStreamId]->codec;
// = avcodec_alloc_context3(mAudioCodec); // = avcodec_alloc_context3(mAudioCodec);
#endif #endif
Debug(1, "Audio Found decoder");
zm_dump_stream_format(mFormatContext, mAudioStreamId, 0, 0); zm_dump_stream_format(mFormatContext, mAudioStreamId, 0, 0);
// Open the codec // Open the codec
#if !LIBAVFORMAT_VERSION_CHECK(53, 8, 0, 8, 0) #if !LIBAVFORMAT_VERSION_CHECK(53, 8, 0, 8, 0)
Debug(1, "Calling avcodec_open");
if ( avcodec_open(mAudioCodecContext, mAudioCodec) < 0 ) { if ( avcodec_open(mAudioCodecContext, mAudioCodec) < 0 ) {
#else #else
Debug(1, "Calling avcodec_open2" );
if ( avcodec_open2(mAudioCodecContext, mAudioCodec, 0) < 0 ) { if ( avcodec_open2(mAudioCodecContext, mAudioCodec, 0) < 0 ) {
#endif #endif
Error("Unable to open codec for audio stream from %s", mPath.c_str() ); Error("Unable to open codec for audio stream from %s", mPath.c_str());
return -1; return -1;
} }
zm_dump_codec(mAudioCodecContext); zm_dump_codec(mAudioCodecContext);
@ -562,23 +559,18 @@ Debug(1, "Selected gw_pix_fmt %d %s", hw_pix_fmt, av_get_pix_fmt_name(hw_pix_fmt
return -1; return -1;
} }
Debug( 3, "Allocated frames");
#if LIBAVUTIL_VERSION_CHECK(54, 6, 0, 6, 0) #if LIBAVUTIL_VERSION_CHECK(54, 6, 0, 6, 0)
int pSize = av_image_get_buffer_size( imagePixFormat, width, height,1 ); int pSize = av_image_get_buffer_size(imagePixFormat, width, height, 1);
#else #else
int pSize = avpicture_get_size( imagePixFormat, width, height ); int pSize = avpicture_get_size(imagePixFormat, width, height);
#endif #endif
if ( (unsigned int)pSize != imagesize ) { if ( (unsigned int)pSize != imagesize ) {
Error("Image size mismatch. Required: %d Available: %d",pSize,imagesize); Error("Image size mismatch. Required: %d Available: %d", pSize, imagesize);
return -1; return -1;
} }
Debug(4, "Validated imagesize");
#if HAVE_LIBSWSCALE #if HAVE_LIBSWSCALE
Debug(1, "Calling sws_isSupportedInput");
if ( !sws_isSupportedInput(mVideoCodecContext->pix_fmt) ) { if ( !sws_isSupportedInput(mVideoCodecContext->pix_fmt) ) {
Error("swscale does not support the codec format: %c%c%c%c", Error("swscale does not support the codec format: %c%c%c%c",
(mVideoCodecContext->pix_fmt)&0xff, (mVideoCodecContext->pix_fmt)&0xff,
@ -600,7 +592,7 @@ Debug(1, "Selected gw_pix_fmt %d %s", hw_pix_fmt, av_get_pix_fmt_name(hw_pix_fmt
} }
# if 0 # if 0
// Have to get a frame first to find out the actual format returned by decoding // Must get a frame first to find out the actual format returned by decoding
mConvertContext = sws_getContext( mConvertContext = sws_getContext(
mVideoCodecContext->width, mVideoCodecContext->width,
mVideoCodecContext->height, mVideoCodecContext->height,
@ -609,12 +601,13 @@ Debug(1, "Selected gw_pix_fmt %d %s", hw_pix_fmt, av_get_pix_fmt_name(hw_pix_fmt
imagePixFormat, SWS_BICUBIC, NULL, imagePixFormat, SWS_BICUBIC, NULL,
NULL, NULL); NULL, NULL);
if ( mConvertContext == NULL ) { if ( mConvertContext == NULL ) {
Error( "Unable to create conversion context for %s", mPath.c_str() ); Error("Unable to create conversion context for %s", mPath.c_str());
return -1; return -1;
} }
#endif #endif
#else // HAVE_LIBSWSCALE #else // HAVE_LIBSWSCALE
Fatal("You must compile ffmpeg with the --enable-swscale option to use ffmpeg cameras"); Fatal("You must compile ffmpeg with the --enable-swscale "
"option to use ffmpeg cameras");
#endif // HAVE_LIBSWSCALE #endif // HAVE_LIBSWSCALE
if ( if (
@ -623,8 +616,7 @@ Debug(1, "Selected gw_pix_fmt %d %s", hw_pix_fmt, av_get_pix_fmt_name(hw_pix_fmt
((unsigned int)mVideoCodecContext->height != height) ((unsigned int)mVideoCodecContext->height != height)
) { ) {
Warning("Monitor dimensions are %dx%d but camera is sending %dx%d", Warning("Monitor dimensions are %dx%d but camera is sending %dx%d",
width, height, mVideoCodecContext->width, mVideoCodecContext->height width, height, mVideoCodecContext->width, mVideoCodecContext->height);
);
} }
mCanCapture = true; mCanCapture = true;
@ -633,7 +625,6 @@ Debug(1, "Selected gw_pix_fmt %d %s", hw_pix_fmt, av_get_pix_fmt_name(hw_pix_fmt
} // int FfmpegCamera::OpenFfmpeg() } // int FfmpegCamera::OpenFfmpeg()
int FfmpegCamera::Close() { int FfmpegCamera::Close() {
Debug(2, "CloseFfmpeg called."); Debug(2, "CloseFfmpeg called.");
mCanCapture = false; mCanCapture = false;
@ -661,9 +652,8 @@ int FfmpegCamera::Close() {
if ( mVideoCodecContext ) { if ( mVideoCodecContext ) {
avcodec_close(mVideoCodecContext); avcodec_close(mVideoCodecContext);
Debug(1,"After codec close");
#if LIBAVCODEC_VERSION_CHECK(57, 64, 0, 64, 0) #if LIBAVCODEC_VERSION_CHECK(57, 64, 0, 64, 0)
//avcodec_free_context(&mVideoCodecContext); // avcodec_free_context(&mVideoCodecContext);
#endif #endif
mVideoCodecContext = NULL; // Freed by av_close_input_file mVideoCodecContext = NULL; // Freed by av_close_input_file
} }
@ -692,8 +682,12 @@ int FfmpegCamera::Close() {
return 0; return 0;
} // end FfmpegCamera::Close } // end FfmpegCamera::Close
//Function to handle capture and store // Function to handle capture and store
int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event_file ) { int FfmpegCamera::CaptureAndRecord(
Image &image,
timeval recording,
char* event_file
) {
if ( !mCanCapture ) { if ( !mCanCapture ) {
return -1; return -1;
} }
@ -707,7 +701,10 @@ int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event
if ( ret < 0 ) { if ( ret < 0 ) {
if ( if (
// Check if EOF. // Check if EOF.
(ret == AVERROR_EOF || (mFormatContext->pb && mFormatContext->pb->eof_reached)) || (
(ret == AVERROR_EOF) ||
(mFormatContext->pb && mFormatContext->pb->eof_reached)
) ||
// Check for Connection failure. // Check for Connection failure.
(ret == -110) (ret == -110)
) { ) {
@ -721,9 +718,14 @@ int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event
} }
if ( (packet.pts != AV_NOPTS_VALUE) && (packet.pts < -100000) ) { if ( (packet.pts != AV_NOPTS_VALUE) && (packet.pts < -100000) ) {
// Ignore packets that have crazy negative pts. They aren't supposed to happen. // Ignore packets that have crazy negative pts.
Warning("Ignore packet because pts %" PRId64 " is massively negative. Error count is %d", packet.pts, error_count); // They aren't supposed to happen.
dumpPacket(mFormatContext->streams[packet.stream_index], &packet,"Ignored packet"); Warning("Ignore packet because pts %" PRId64 " is massively negative."
" Error count is %d", packet.pts, error_count);
dumpPacket(
mFormatContext->streams[packet.stream_index],
&packet,
"Ignored packet");
if ( error_count > 100 ) { if ( error_count > 100 ) {
Error("Bad packet count over 100, going to close and re-open stream"); Error("Bad packet count over 100, going to close and re-open stream");
return -1; return -1;
@ -736,14 +738,16 @@ int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event
int keyframe = packet.flags & AV_PKT_FLAG_KEY; int keyframe = packet.flags & AV_PKT_FLAG_KEY;
bytes += packet.size; bytes += packet.size;
dumpPacket(mFormatContext->streams[packet.stream_index], &packet, "Captured Packet"); dumpPacket(
mFormatContext->streams[packet.stream_index],
&packet,
"Captured Packet");
if ( packet.dts == AV_NOPTS_VALUE ) { if ( packet.dts == AV_NOPTS_VALUE ) {
packet.dts = packet.pts; packet.dts = packet.pts;
} }
// Video recording // Video recording
if ( recording.tv_sec ) { if ( recording.tv_sec ) {
uint32_t last_event_id = monitor->GetLastEventId(); uint32_t last_event_id = monitor->GetLastEventId();
uint32_t video_writer_event_id = monitor->GetVideoWriterEventId(); uint32_t video_writer_event_id = monitor->GetVideoWriterEventId();
@ -754,12 +758,13 @@ int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event
if ( videoStore ) { if ( videoStore ) {
Info("Re-starting video storage module"); Info("Re-starting video storage module");
// I don't know if this is important or not... but I figure we might as well write this last packet out to the store before closing it. // I don't know if this is important or not... but I figure we might
// as well write this last packet out to the store before closing it.
// Also don't know how much it matters for audio. // Also don't know how much it matters for audio.
if ( packet.stream_index == mVideoStreamId ) { if ( packet.stream_index == mVideoStreamId ) {
//Write the packet to our video store // Write the packet to our video store
int ret = videoStore->writeVideoFramePacket(&packet); int ret = videoStore->writeVideoFramePacket(&packet);
if ( ret < 0 ) { //Less than zero and we skipped a frame if ( ret < 0 ) { // Less than zero and we skipped a frame
Warning("Error writing last packet to videostore."); Warning("Error writing last packet to videostore.");
} }
} // end if video } // end if video
@ -772,8 +777,8 @@ int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event
} // end if videoStore } // end if videoStore
} // end if end of recording } // end if end of recording
if ( last_event_id and !videoStore ) { if ( last_event_id && !videoStore ) {
//Instantiate the video storage module // Instantiate the video storage module
packetqueue->dumpQueue(); packetqueue->dumpQueue();
if ( record_audio ) { if ( record_audio ) {
@ -801,7 +806,7 @@ int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event
this->getMonitor()); this->getMonitor());
} // end if record_audio } // end if record_audio
if ( ! videoStore->open() ) { if ( !videoStore->open() ) {
delete videoStore; delete videoStore;
videoStore = NULL; videoStore = NULL;
@ -809,27 +814,33 @@ int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event
monitor->SetVideoWriterEventId(last_event_id); monitor->SetVideoWriterEventId(last_event_id);
// Need to write out all the frames from the last keyframe? // Need to write out all the frames from the last keyframe?
// No... need to write out all frames from when the event began. Due to PreEventFrames, this could be more than since the last keyframe. // No... need to write out all frames from when the event began.
// Due to PreEventFrames, this could be more than
// since the last keyframe.
unsigned int packet_count = 0; unsigned int packet_count = 0;
ZMPacket *queued_packet; ZMPacket *queued_packet;
// Clear all packets that predate the moment when the recording began // Clear all packets that predate the moment when the recording began
packetqueue->clear_unwanted_packets(&recording, monitor->GetPreEventCount(), mVideoStreamId); packetqueue->clear_unwanted_packets(
&recording, monitor->GetPreEventCount(), mVideoStreamId);
while ( ( queued_packet = packetqueue->popPacket() ) ) { while ( (queued_packet = packetqueue->popPacket()) ) {
AVPacket *avp = queued_packet->av_packet(); AVPacket *avp = queued_packet->av_packet();
packet_count += 1; packet_count += 1;
//Write the packet to our video store // Write the packet to our video store
Debug(2, "Writing queued packet stream: %d KEY %d, remaining (%d)", Debug(2, "Writing queued packet stream: %d KEY %d, remaining (%d)",
avp->stream_index, avp->flags & AV_PKT_FLAG_KEY, packetqueue->size()); avp->stream_index,
avp->flags & AV_PKT_FLAG_KEY,
packetqueue->size());
if ( avp->stream_index == mVideoStreamId ) { if ( avp->stream_index == mVideoStreamId ) {
ret = videoStore->writeVideoFramePacket(avp); ret = videoStore->writeVideoFramePacket(avp);
have_video_keyframe = true; have_video_keyframe = true;
} else if ( avp->stream_index == mAudioStreamId ) { } else if ( avp->stream_index == mAudioStreamId ) {
ret = videoStore->writeAudioFramePacket(avp); ret = videoStore->writeAudioFramePacket(avp);
} else { } else {
Warning("Unknown stream id in queued packet (%d)", avp->stream_index); Warning("Unknown stream id in queued packet (%d)",
avp->stream_index);
ret = -1; ret = -1;
} }
if ( ret < 0 ) { if ( ret < 0 ) {
@ -845,7 +856,7 @@ int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event
// Not recording // Not recording
if ( videoStore ) { if ( videoStore ) {
Debug(1,"Deleting videoStore instance"); Debug(1, "Deleting videoStore instance");
delete videoStore; delete videoStore;
videoStore = NULL; videoStore = NULL;
have_video_keyframe = false; have_video_keyframe = false;
@ -858,11 +869,17 @@ int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event
if ( packet.stream_index == mVideoStreamId ) { if ( packet.stream_index == mVideoStreamId ) {
if ( keyframe ) { if ( keyframe ) {
Debug(3, "Clearing queue"); Debug(3, "Clearing queue");
if ( packetqueue->packet_count(mVideoStreamId) >= monitor->GetImageBufferCount() ) { if (
Warning("ImageBufferCount %d is too small. Needs to be at least %d. Either increase it or decrease time between keyframes", packetqueue->packet_count(mVideoStreamId)
>=
monitor->GetImageBufferCount()
) {
Warning(
"ImageBufferCount %d is too small. "
"Needs to be at least %d. "
"Either increase it or decrease time between keyframes",
monitor->GetImageBufferCount(), monitor->GetImageBufferCount(),
packetqueue->packet_count(mVideoStreamId)+1 packetqueue->packet_count(mVideoStreamId)+1);
);
} }
packetqueue->clearQueue(monitor->GetPreEventCount(), mVideoStreamId); packetqueue->clearQueue(monitor->GetPreEventCount(), mVideoStreamId);
@ -872,34 +889,27 @@ int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event
packetqueue->queuePacket(&packet); packetqueue->queuePacket(&packet);
} }
} else if ( packet.stream_index == mAudioStreamId ) { } else if ( packet.stream_index == mAudioStreamId ) {
// The following lines should ensure that the queue always begins with a video keyframe // Ensure that the queue always begins with a video keyframe
//Debug(2, "Have audio packet, reocrd_audio is (%d) and packetqueue.size is (%d)", record_audio, packetqueue.size() );
if ( record_audio && packetqueue->size() ) { if ( record_audio && packetqueue->size() ) {
// if it's audio, and we are doing audio, and there is already something in the queue
packetqueue->queuePacket(&packet); packetqueue->queuePacket(&packet);
} }
} // end if packet type } // end if packet type
if ( packet.stream_index == mVideoStreamId ) { if ( packet.stream_index == mVideoStreamId ) {
// only do decode if we have had a keyframe, should save a few cycles. if ( (have_video_keyframe || keyframe) && videoStore ) {
if ( have_video_keyframe || keyframe ) {
if ( videoStore ) {
//Write the packet to our video store
int ret = videoStore->writeVideoFramePacket(&packet); int ret = videoStore->writeVideoFramePacket(&packet);
if ( ret < 0 ) { //Less than zero and we skipped a frame if ( ret < 0 ) {
zm_av_packet_unref(&packet); // Less than zero and we skipped a frame
return 0; Error("Unable to write video packet %d: %s",
} frameCount, av_make_error_string(ret).c_str());
} else {
have_video_keyframe = true; have_video_keyframe = true;
} }
} // end if keyframe or have_video_keyframe } // end if keyframe or have_video_keyframe
Debug(4, "about to decode video");
ret = zm_receive_frame(mVideoCodecContext, mRawFrame, packet); ret = zm_receive_frame(mVideoCodecContext, mRawFrame, packet);
if ( ret < 0 ) { if ( ret < 0 ) {
Warning("Unable to receive frame %d: %s, continuing. error count is %d", Warning("Unable to receive frame %d: %s. error count is %d",
frameCount, av_make_error_string(ret).c_str(), error_count); frameCount, av_make_error_string(ret).c_str(), error_count);
error_count += 1; error_count += 1;
if ( error_count > 100 ) { if ( error_count > 100 ) {
@ -909,10 +919,14 @@ int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event
zm_av_packet_unref(&packet); zm_av_packet_unref(&packet);
continue; continue;
} }
if ( error_count > 0 ) error_count --; if ( error_count > 0 ) error_count--;
zm_dump_video_frame(mRawFrame); zm_dump_video_frame(mRawFrame);
#if HAVE_LIBAVUTIL_HWCONTEXT_H #if HAVE_LIBAVUTIL_HWCONTEXT_H
if ( (hw_pix_fmt != AV_PIX_FMT_NONE) && (mRawFrame->format == hw_pix_fmt) ) { if (
(hw_pix_fmt != AV_PIX_FMT_NONE)
&&
(mRawFrame->format == hw_pix_fmt)
) {
/* retrieve data from GPU to CPU */ /* retrieve data from GPU to CPU */
ret = av_hwframe_transfer_data(hwFrame, mRawFrame, 0); ret = av_hwframe_transfer_data(hwFrame, mRawFrame, 0);
if ( ret < 0 ) { if ( ret < 0 ) {
@ -940,23 +954,23 @@ int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event
frameComplete = 1; frameComplete = 1;
frameCount++; frameCount++;
} else if ( packet.stream_index == mAudioStreamId ) { //FIXME best way to copy all other streams } else if ( packet.stream_index == mAudioStreamId ) {
// FIXME best way to copy all other streams
frameComplete = 1; frameComplete = 1;
if ( videoStore ) { if ( videoStore ) {
if ( record_audio ) { if ( record_audio ) {
if ( have_video_keyframe ) { if ( have_video_keyframe ) {
Debug(3, "Recording audio packet streamindex(%d) packetstreamindex(%d)", // Write the packet to our video store
mAudioStreamId, packet.stream_index); // FIXME no relevance of last key frame
//Write the packet to our video store
//FIXME no relevance of last key frame
int ret = videoStore->writeAudioFramePacket(&packet); int ret = videoStore->writeAudioFramePacket(&packet);
if ( ret < 0 ) {//Less than zero and we skipped a frame if ( ret < 0 ) {
// Less than zero and we skipped a frame
Warning("Failure to write audio packet."); Warning("Failure to write audio packet.");
zm_av_packet_unref(&packet); zm_av_packet_unref(&packet);
return 0; return 0;
} }
} else { } else {
Debug(3, "Not recording audio yet because we don't have a video keyframe yet"); Debug(3, "Not recording audio because no video keyframe");
} }
} else { } else {
Debug(4, "Not doing recording of audio packet"); Debug(4, "Not doing recording of audio packet");
@ -970,20 +984,27 @@ int FfmpegCamera::CaptureAndRecord( Image &image, timeval recording, char* event
#if LIBAVUTIL_VERSION_CHECK(56, 23, 0, 23, 0) #if LIBAVUTIL_VERSION_CHECK(56, 23, 0, 23, 0)
Debug(3, "Some other stream index %d, %s", Debug(3, "Some other stream index %d, %s",
packet.stream_index, packet.stream_index,
av_get_media_type_string(mFormatContext->streams[packet.stream_index]->codecpar->codec_type) av_get_media_type_string(
mFormatContext->streams[packet.stream_index]->codecpar->codec_type)
); );
#else #else
Debug(3, "Some other stream index %d", packet.stream_index); Debug(3, "Some other stream index %d", packet.stream_index);
#endif #endif
} // end if is video or audio or something else } // end if is video or audio or something else
// the packet contents are ref counted... when queuing, we allocate another packet and reference it with that one, so we should always need to unref here, which should not affect the queued version. // the packet contents are ref counted... when queuing, we allocate another
// packet and reference it with that one, so we should always need to unref
// here, which should not affect the queued version.
zm_av_packet_unref(&packet); zm_av_packet_unref(&packet);
} // end while ! frameComplete } // end while ! frameComplete
return frameCount; return frameCount;
} // end FfmpegCamera::CaptureAndRecord } // end FfmpegCamera::CaptureAndRecord
int FfmpegCamera::transfer_to_image(Image &image, AVFrame *output_frame, AVFrame *input_frame) { int FfmpegCamera::transfer_to_image(
Image &image,
AVFrame *output_frame,
AVFrame *input_frame
) {
uint8_t* directbuffer; uint8_t* directbuffer;
/* Request a writeable buffer of the target image */ /* Request a writeable buffer of the target image */
@ -1018,9 +1039,11 @@ int FfmpegCamera::transfer_to_image(Image &image, AVFrame *output_frame, AVFrame
} }
} }
if ( sws_scale(mConvertContext, input_frame->data, input_frame->linesize, if ( sws_scale(
0, mVideoCodecContext->height, output_frame->data, output_frame->linesize) <= 0 ) { mConvertContext, input_frame->data, input_frame->linesize,
Error("Unable to convert raw format %u to target format %u at frame %d codec %u ", 0, mVideoCodecContext->height,
output_frame->data, output_frame->linesize) <= 0 ) {
Error("Unable to convert format %u to format %u at frame %d codec %u",
input_frame->format, input_frame->format,
imagePixFormat, frameCount, imagePixFormat, frameCount,
mVideoCodecContext->pix_fmt mVideoCodecContext->pix_fmt
@ -1028,14 +1051,15 @@ int FfmpegCamera::transfer_to_image(Image &image, AVFrame *output_frame, AVFrame
return -1; return -1;
} }
#else // HAVE_LIBSWSCALE #else // HAVE_LIBSWSCALE
Fatal("You must compile ffmpeg with the --enable-swscale option to use ffmpeg cameras"); Fatal("You must compile ffmpeg with the --enable-swscale "
"option to use ffmpeg cameras");
#endif // HAVE_LIBSWSCALE #endif // HAVE_LIBSWSCALE
return 0; return 0;
} // end int FfmpegCamera::transfer_to_image(Image &i, AVFrame *output_frame, AVFrame input_frame) } // end int FfmpegCamera::transfer_to_image
int FfmpegCamera::FfmpegInterruptCallback(void *ctx) { int FfmpegCamera::FfmpegInterruptCallback(void *ctx) {
//FfmpegCamera* camera = reinterpret_cast<FfmpegCamera*>(ctx); // FfmpegCamera* camera = reinterpret_cast<FfmpegCamera*>(ctx);
//Debug(4, "FfmpegInterruptCallback"); // Debug(4, "FfmpegInterruptCallback");
return zm_terminate; return zm_terminate;
} }