前篇
第四篇
RTP工具改进(四) - rtmp协议推送
前面使用的工具一直为mfc,今天将使用qt 来做界面,使用qt 来进行程序和协议的编写,qt部分目前还不包括rtp ps流和rtmp,暂时只有rtp 直接传输,关于rtmp协议和ps流协议,先使用vs的mfc。增加和改变的模块为rtp,和 rtp_recv,如下图,以前的vs MFC版本都放到vs下面,有关于qt的 gb28181 的sip server 和 rtp 发送接收等都放到qt下面,所有可执行都放到外层的bin下面
代码地址
https://gitee.com/guanzhi0319/rtp
QT 加入
除了gb28181 的可视化界面,增加了两个程序,一个qt_rtp,一个qt_rtp_recv,打开后如下所示
2.1 发送端
目前制作还是比较简陋,先以能执行为主,使用qt 5.14,mingw,不依赖于vs,所以读者可以不安装vs就可以使用该代码,首先制作一个CameraVideoSurface类,用来读摄像头
#include "c_cameravideo.h"
#include <QDebug>
//c_cameravideo::c_cameravideo(QObject * parent)
//{
//}
CameraVideoSurface::CameraVideoSurface(QObject *parent)
: QAbstractVideoSurface(parent)
{
//this->InitEncoder();
}
CameraVideoSurface::~CameraVideoSurface()
{
// avformat_close_input(&pOutputFormatCtx);
// av_frame_free(&yuvFrame);
// av_packet_free(&packet);
// avcodec_close(pCodecCtx);
}
void CameraVideoSurface::setCameraResolution(const QSize &size)
{
this->setNativeResolution(size);
}
QList<QVideoFrame::PixelFormat> CameraVideoSurface::supportedPixelFormats
(QAbstractVideoBuffer::HandleType handleType) const
{
QList<QVideoFrame::PixelFormat > pixelFormats;
pixelFormats.append(QVideoFrame::Format_BGR24);
pixelFormats.append(QVideoFrame::Format_RGB32);
pixelFormats.append(QVideoFrame::Format_YUV420P);
return pixelFormats;
}
bool CameraVideoSurface::present(const QVideoFrame &frame)
{
if (frame.isValid())
{
QVideoFrame cloneFrame(frame);
cloneFrame.map(QAbstractVideoBuffer::ReadOnly);
QImage image(cloneFrame.bits(), cloneFrame.width(), cloneFrame.height(),
QVideoFrame::imageFormatFromPixelFormat(frame.pixelFormat()));
// QImage image2(cloneFrame.bits(), cloneFrame.width(), cloneFrame.height(),
// QVideoFrame::imageFormatFromPixelFormat(QVideoFrame::Format_BGR24));
image = image.mirrored(true, true);
// rgb 转 yuv
// uint8_t *data[AV_NUM_DATA_POINTERS] = {0};
// data[0] = (uint8_t *)image.constBits();
// int linesize[AV_NUM_DATA_POINTERS] = {0};
// linesize[0] = pCodecCtx->width * 4;
// sws_scale(image_convert_ctx, data, linesize, 0, pCodecCtx->height,
// yuvFrame->data, yuvFrame->linesize);
// // 编码
// this->Encode(yuvFrame);
emit showFrame(image);
cloneFrame.unmap();
return true;
}
return false;
}
#if 0
void CameraVideoSurface::InitEncoder()
{
//av_register_all();
avformat_network_init();
avcodec_register_all();
QString outputFileName = "output.h264";
QString encoderName = "libx264";
//QString rtmpAddress = "rtmp://192.168.1.111/live/livestream";
pCodec = avcodec_find_encoder_by_name(encoderName.toStdString().c_str());
if(NULL == pCodec)
{
qDebug() <<"查找视频编码器失败!";
return;
}
pCodecCtx = avcodec_alloc_context3(pCodec);
if(NULL == pCodecCtx)
{
qDebug() <<"开辟编解码器上下文";
return;
}
// 输入样本参数
pCodecCtx->bit_rate = 400000;
pCodecCtx->width = 1280;
pCodecCtx->height = 720;
pCodecCtx->time_base = {1, 25};
pCodecCtx->framerate = {25, 1};
pCodecCtx->gop_size = 10;
pCodecCtx->max_b_frames = 1;
pCodecCtx->qmin = 10;
pCodecCtx->qmax = 51;
pCodecCtx->pix_fmt = AV_PIX_FMT_YUV420P;
if(AV_CODEC_ID_H264 == pCodecCtx->codec_id)
{
av_opt_set(pCodecCtx->priv_data, "preset", "slow", 0);
av_opt_set(pCodecCtx->priv_data, "tune", "zerolatency", 0);
}
// 打开编码器
if(avcodec_open2(pCodecCtx, pCodec, NULL) < 0)
{
qDebug() <<"打开编码器失败 !";
return;
}
pOutputFormatCtx = avformat_alloc_context();
if(NULL == pOutputFormatCtx)
{
qDebug() <<"视频封装器开辟失败!";
return;
}
AVOutputFormat *outputFormat = av_guess_format(NULL, outputFileName.toStdString().c_str(), NULL);
if(NULL == outputFormat)
{
qDebug() <<"猜测outputformat失败 !";
return;
}
pOutputFormatCtx->oformat = outputFormat;
// oprn url
if(avio_open(&pOutputFormatCtx->pb, outputFileName.toStdString().c_str(), AVIO_FLAG_READ_WRITE) < 0)
{
qDebug() <<"打开输出文件失败!";
return;
}
pOutputStream = avformat_new_stream(pOutputFormatCtx, NULL);
if(NULL == pOutputStream)
{
qDebug() <<"新建输出流失败 !";
return;
}
// 输出详细信息
av_dump_format(pOutputFormatCtx, 0, outputFileName.toStdString().c_str(), 1);
// 新建数据包
packet = av_packet_alloc();
if(NULL == packet)
{
qDebug() <<"新建数据包失败 !";
return;
}
// yuvFrame 初始化
yuvFrame = av_frame_alloc();
if(NULL == yuvFrame)
{
qDebug() <<"开辟AVFrame失败 !";
return;
}
yuvFrame->width = pCodecCtx->width;
yuvFrame->height = pCodecCtx->height;
yuvFrame->format = pCodecCtx->pix_fmt;
// 初始化 image 空间
av_image_alloc(yuvFrame->data, yuvFrame->linesize, yuvFrame->width, yuvFrame->height,
pCodecCtx->pix_fmt, 32);
// 转换上下文
image_convert_ctx = sws_getContext(pCodecCtx->width, pCodecCtx->height, AV_PIX_FMT_RGB32, pCodecCtx->width,
pCodecCtx->height, AV_PIX_FMT_YUV420P,
SWS_BICUBIC, NULL, NULL, NULL);
if(NULL == image_convert_ctx)
{
qDebug() <<"转换上下文失败 !";
return;
}
// 写封装头
if(avformat_write_header(pOutputFormatCtx, NULL) < 0)
{
qDebug() <<"视频封装头写失败 !";
return;
}
}
// 编码为 h.264
void CameraVideoSurface::Encode(AVFrame *frame)
{
static int index = 0;
frame->pts = index++;
int ret = 0;
if((ret = avcodec_send_frame(pCodecCtx, frame)) < 0)
{
qDebug() <<"avcodec_send_frame 失败 !";
return;
}
while(ret >= 0)
{
ret = avcodec_receive_packet(pCodecCtx, packet);
if (ret == AVERROR(EAGAIN) || ret == AVERROR_EOF)
{
return;
}
else if (ret < 0)
{
qDebug() << "编码时出错";
return;
}
packet->stream_index = 0;
av_interleaved_write_frame(pOutputFormatCtx, packet); // write frame
av_packet_unref(packet);
}
}
#endif
void CameraVideoSurface::cameraStopSlot()
{
qDebug()<<"关闭close";
// av_write_trailer(pOutputFormatCtx);
}
然后就需要使用udp进行发送,使用 QUdpSocket 类来发送,把c_rtp类从QObject类去继承
class c_rtp:public QObject
{
Q_OBJECT
private:
QUdpSocket *v_udpSocket = NULL;
QHostAddress v_host;
//QString v_host_str;
quint16 v_port;
private:
unsigned short _seq_num = 0 ;
unsigned char sendbuf[MAX_LENGTH];
unsigned int v_ssrc = 1001;//the default value must hash("live/1001")
public:
c_rtp()
{}
~c_rtp()
{
if(v_udpSocket!=NULL)
{
v_udpSocket->abort();
delete v_udpSocket;
}
}
void func_init(const char *ip,uint16_t port);
int func_send_video(uint8_t * data, int len);
//发送ps流
int func_send_video_ps(uint8_t* data, int len);
};
2.2 接收端
接收端主要使用IO_Thread 来接收包,拼接包,也就是意味着可以有多个接收,同时显示
class MainWindow : public QMainWindow
{
Q_OBJECT
public:
MainWindow(QWidget *parent = nullptr);
~MainWindow();
private slots:
void on_pushButton_clicked();
private:
//ssrc-->s_rtp_context
std::unordered_map<uint32_t, IO_Thread*> v_ctxs;
Ui::MainWindow *ui;
QUdpSocket *v_udpSocket = NULL;
//IO_Thread* v_iothd = NULL;
IO_Thread * getctx(uint32_t ssrc);
QTimer* v_timer = NULL;
//int v_id1; //定时器1的唯一标示
};
实现使用QUdpSocket 去产生一个服务端来接收数据,根据不同的ssrc来分发,显示在不同的QLabel组件上
#include "mainwindow.h"
#include "ui_mainwindow.h"
MainWindow::MainWindow(QWidget *parent)
: QMainWindow(parent)
, ui(new Ui::MainWindow)
{
ui->setupUi(this);
ui->iptext->setText("127.0.0.1");
ui->porttext->setText("6000");
v_udpSocket = new QUdpSocket(this);
//v_udpSocket->bind(QHostAddress::LocalHost, 6000);
v_udpSocket->bind(6000);
//收包程序
connect(v_udpSocket,&QUdpSocket::readyRead,[=](){
// 获取报文长度大小
qint64 size = v_udpSocket->pendingDatagramSize();
// 读取报文
QByteArray array = QByteArray(size,0);
v_udpSocket->readDatagram(array.data(),size);
uint8_t* data = (uint8_t*)array.data();
int inlen = (int)size;
int outlen =0;
uint32_t last_ts,ssrc;
uint16_t seq;
uint8_t payloadtype;
uint8_t *buffer = rtp_payload(data, inlen, &outlen, last_ts, ssrc,
seq, payloadtype);
IO_Thread* ctx = getctx(ssrc);
ctx->v_last_ts = last_ts;
ctx->v_seq = seq;
ctx->v_payloadtype = payloadtype;
//ctx->v_ssrc = ssrc;
live_rtp_unpack(data,ctx,buffer,outlen);
});
v_timer = new QTimer(this);
//启动定时器
v_timer->start(40);
connect(v_timer,&QTimer::timeout,[=](){
//static int num = 1;
auto iter = v_ctxs.begin();
if(iter!= v_ctxs.end())
{
IO_Thread* ctx = iter->second;
ctx->LockBuffer();
//显示
uint8_t* rgbBuffer = ctx->out_buffer_out;
int w = ctx->m_w;
int h = ctx->m_h;
if(rgbBuffer!=NULL && w>0 && h >0)
{
int iw = ui->label_showv->width();
int ih = ui->label_showv->height();
QImage tmpImg(rgbBuffer,w,h,QImage::Format_RGB32);
QImage imageScale = tmpImg.scaled(QSize(iw,ih));
QImage image = imageScale.mirrored(true, false);
QPixmap pixmap = QPixmap::fromImage(image);
ui->label_showv->setPixmap(pixmap);
}
ctx->UnLockBuffer();
}
});
}
MainWindow::~MainWindow()
{
delete ui;
}
IO_Thread * MainWindow::getctx(uint32_t ssrc)
{
auto it = v_ctxs.find(ssrc);
if (it != v_ctxs.end())
return it->second;
else
{
IO_Thread *sc = new IO_Thread();
v_ctxs[ssrc] = sc;
sc->Start();
return sc;
}
}
后续
改进的地方为:
1 MFC 版本增加音频,包括rtp的音频和rtmp协议的音频
2 QT 版本修改为player,将会丰富QT版本
3 增加rtmp服务器,有想法的同志可以加入一起做。