达芬奇TI DVSDK之视频数据流过程分析
博客:http://www.cnblogs.com/tinz
1 /****************************************************************************** 2 * Capture_create 3 ******************************************************************************/ 4 Capture_Handle Capture_create(BufTab_Handle hBufTab, Capture_Attrs *attrs) 5 { 6 struct v4l2_capability cap; 7 struct v4l2_cropcap cropCap; 8 struct v4l2_crop crop; 9 struct v4l2_format fmt; 10 enum v4l2_buf_type type; 11 Capture_Handle hCapture; 12 VideoStd_Type videoStd; 13 Int32 width, height; 14 Uint32 pixelFormat; 15 16 assert(attrs); 17 Dmai_clear(fmt); 18 19 /* Allocate space for state object */ 20 hCapture = calloc(1, sizeof(Capture_Object)); 21 22 if (hCapture == NULL) { 23 Dmai_err0("Failed to allocate space for Capture Object\n"); 24 return NULL; 25 } 26 27 /* User allocated buffers by default */ 28 hCapture->userAlloc = TRUE; 29 30 /* Open video capture device */ 31 /* 打开V4L2视频输入设备 */ 32 hCapture->fd = open(attrs->captureDevice, O_RDWR, 0); 33 34 if (hCapture->fd == -1) { 35 Dmai_err2("Cannot open %s (%s)\n", attrs->captureDevice, 36 strerror(errno)); 37 cleanup(hCapture); 38 return NULL; 39 } 40 41 /* See if an input is connected, and if so which standard */ 42 /* 检测V4L2设备当前是否有视频信号输入,输入信号的标准 */ 43 if (Capture_detectVideoStd(hCapture, &videoStd, attrs) < 0) { 44 cleanup(hCapture); 45 return NULL; 46 } 47 48 hCapture->videoStd = videoStd; 49 50 if (VideoStd_getResolution(videoStd, &width, &height) < 0) { 51 cleanup(hCapture); 52 Dmai_err0("Failed to get resolution of capture video standard\n"); 53 return NULL; 54 } 55 56 /* Query for capture device capabilities */ 57 /* 这里就是调用ioctl直接操作v4l2设备了,查询设备的特性*/ 58 if (ioctl(hCapture->fd, VIDIOC_QUERYCAP, &cap) == -1) { 59 cleanup(hCapture); 60 if (errno == EINVAL) { 61 Dmai_err1("%s is no V4L2 device\n", attrs->captureDevice); 62 cleanup(hCapture); 63 return NULL; 64 } 65 Dmai_err2("Failed VIDIOC_QUERYCAP on %s (%s)\n", attrs->captureDevice, 66 strerror(errno)); 67 cleanup(hCapture); 68 return NULL; 69 } 70 71 if (!(cap.capabilities & V4L2_CAP_VIDEO_CAPTURE)) { 72 Dmai_err1("%s is not a video capture device\n", attrs->captureDevice); 73 cleanup(hCapture); 74 return NULL; 75 } 76 77 if (!(cap.capabilities & V4L2_CAP_STREAMING)) { 78 Dmai_err1("%s does not support streaming i/o\n", attrs->captureDevice); 79 cleanup(hCapture); 80 return NULL; 81 } 82 83 fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; 84 /* 获取V4L2设备的帧格式 */ 85 if (ioctl(hCapture->fd, VIDIOC_G_FMT, &fmt) == -1) { 86 Dmai_err2("Failed VIDIOC_G_FMT on %s (%s)\n", attrs->captureDevice, 87 strerror(errno)); 88 cleanup(hCapture); 89 return NULL; 90 } 91 92 fmt.fmt.pix.width = width; 93 fmt.fmt.pix.height = height; 94 fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; 95 96 switch(attrs->colorSpace) { 97 case ColorSpace_UYVY: 98 fmt.fmt.pix.pixelformat = V4L2_PIX_FMT_UYVY; 99 break; 100 case ColorSpace_YUV420PSEMI: 101 fmt.fmt.pix.pixelformat = V4L2_PIX_FMT_NV12; 102 break; 103 case ColorSpace_YUV422PSEMI: 104 fmt.fmt.pix.pixelformat = V4L2_PIX_FMT_NV16; 105 break; 106 default: 107 Dmai_err1("Unsupported color format %g\n", attrs->colorSpace); 108 cleanup(hCapture); 109 return NULL; 110 } 111 112 if ((videoStd == VideoStd_BAYER_CIF) || (videoStd == VideoStd_BAYER_VGA) || 113 (videoStd == VideoStd_BAYER_1280)) { 114 fmt.fmt.pix.pixelformat = V4L2_PIX_FMT_SBGGR8; 115 } 116 117 fmt.fmt.pix.bytesperline = BufferGfx_calcLineLength(fmt.fmt.pix.width, 118 attrs->colorSpace); 119 fmt.fmt.pix.sizeimage = BufferGfx_calcSize(attrs->videoStd, attrs->colorSpace); 120 121 //printf("DMAI: pix.bytesperline= %d, pix.sizeimage= %d\r\n",fmt.fmt.pix.bytesperline,fmt.fmt.pix.sizeimage); 122 pixelFormat = fmt.fmt.pix.pixelformat; 123 124 if ((videoStd == VideoStd_CIF) || (videoStd == VideoStd_SIF_PAL) || 125 (videoStd == VideoStd_SIF_NTSC) || (videoStd == VideoStd_D1_PAL) || 126 (videoStd == VideoStd_D1_NTSC) || (videoStd == VideoStd_1080I_30) || 127 (videoStd == VideoStd_1080I_25)) { 128 fmt.fmt.pix.field = V4L2_FIELD_INTERLACED; 129 } else { 130 fmt.fmt.pix.field = V4L2_FIELD_NONE; 131 } 132 133 /* 设置V4L2输入设备的帧格式 */ 134 if (ioctl(hCapture->fd, VIDIOC_S_FMT, &fmt) == -1) { 135 printf("Failed VIDIOC_S_FMT on %s (%s)\n", attrs->captureDevice, 136 strerror(errno)); 137 cleanup(hCapture); 138 return NULL; 139 } 140 141 if ((fmt.fmt.pix.width != width) || (fmt.fmt.pix.height != height)) { 142 Dmai_err4("Failed to set resolution %d x %d (%d x %d)\n", width, 143 height, fmt.fmt.pix.width, fmt.fmt.pix.height); 144 cleanup(hCapture); 145 return NULL; 146 } 147 148 if (pixelFormat != fmt.fmt.pix.pixelformat) { 149 Dmai_err2("Pixel format 0x%x not supported. Received 0x%x\n", 150 pixelFormat, fmt.fmt.pix.pixelformat); 151 cleanup(hCapture); 152 return NULL; 153 } 154 155 Dmai_dbg3("Video input connected size %dx%d pitch %d\n", 156 fmt.fmt.pix.width, fmt.fmt.pix.height, fmt.fmt.pix.bytesperline); 157 158 /* Query for video input cropping capability */ 159 160 if (attrs->cropWidth > 0 && attrs->cropHeight > 0) { 161 162 cropCap.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; 163 if (ioctl(hCapture->fd, VIDIOC_CROPCAP, &cropCap) == -1) { 164 Dmai_err2("VIDIOC_CROPCAP failed on %s (%s)\n", attrs->captureDevice, 165 strerror(errno)); 166 cleanup(hCapture); 167 return NULL; 168 } 169 170 if (attrs->cropX & 0x1) { 171 Dmai_err1("Crop width (%ld) needs to be even\n", attrs->cropX); 172 cleanup(hCapture); 173 return NULL; 174 } 175 176 crop.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; 177 crop.c.left = attrs->cropX; 178 crop.c.top = attrs->cropY; 179 crop.c.width = attrs->cropWidth; 180 crop.c.height = hCapture->topOffset ? attrs->cropHeight + 4 + 2 : 181 attrs->cropHeight; 182 183 Dmai_dbg4("Setting capture cropping at %dx%d size %dx%d\n", 184 crop.c.left, crop.c.top, crop.c.width, crop.c.height); 185 186 /* Crop the image depending on requested image size */ 187 if (ioctl(hCapture->fd, VIDIOC_S_CROP, &crop) == -1) { 188 Dmai_err2("VIDIOC_S_CROP failed on %s (%s)\n", attrs->captureDevice, 189 strerror(errno)); 190 cleanup(hCapture); 191 return NULL; 192 } 193 } 194 195 if (hBufTab == NULL) { 196 hCapture->userAlloc = FALSE; 197 198 /* The driver allocates the buffers */ 199 /* 调用CMEM创建要用到的视频缓冲区 */ 200 if (_Dmai_v4l2DriverAlloc(hCapture->fd, 201 attrs->numBufs, 202 V4L2_BUF_TYPE_VIDEO_CAPTURE, 203 &hCapture->bufDescs, 204 &hBufTab, 205 hCapture->topOffset, 206 attrs->colorSpace) < 0) { 207 Dmai_err1("Failed to allocate capture driver buffers on %s\n", 208 attrs->captureDevice); 209 cleanup(hCapture); 210 return NULL; 211 } 212 } 213 else { 214 /* Make the driver use the user supplied buffers */ 215 /* 如果调用者已经创建好缓冲区,那只需要加入相应的队列管理就可以 */ 216 if (_Dmai_v4l2UserAlloc(hCapture->fd, 217 attrs->numBufs, 218 V4L2_BUF_TYPE_VIDEO_CAPTURE, 219 &hCapture->bufDescs, 220 hBufTab, 221 0, attrs->colorSpace) < 0) { 222 Dmai_err1("Failed to intialize capture driver buffers on %s\n", 223 attrs->captureDevice); 224 cleanup(hCapture); 225 return NULL; 226 } 227 } 228 229 hCapture->hBufTab = hBufTab; 230 231 /* Start the video streaming */ 232 type = V4L2_BUF_TYPE_VIDEO_CAPTURE; 233 234 /* 配置完后,启动V4L2设备 */ 235 if (ioctl(hCapture->fd, VIDIOC_STREAMON, &type) == -1) { 236 Dmai_err2("VIDIOC_STREAMON failed on %s (%s)\n", attrs->captureDevice, 237 strerror(errno)); 238 cleanup(hCapture); 239 return NULL; 240 } 241 242 hCapture->started = TRUE; 243 244 return hCapture; 245 }
从上面的代码可以看出,DMAI的函数将大量的对V4L2的控制操作都封装起来了,使用者只需要传入缓冲区列表和视频采集参数就可以创建一个视频采集设备。设备采集到
数据后,会将数据填充到使用者提供的缓冲区,使用者只需要取出相应的缓冲区就可以得到视频数据。在上面的函数中,调用了一个_Dmai_v4l2DriverAlloc()函数,这个函数
的作用是分配视频缓冲区,它最终是调用CMEM进行内存分配的,下面一起看看这个函数的具体实现。
2.DMAI与CMEM的交互。
查看dmai_2_20_00_15/packages/ti/sdo/dmai/linux/dm6467/_VideoBuf.c
1 /****************************************************************************** 2 * _Dmai_v4l2DriverAlloc 3 ******************************************************************************/ 4 Int _Dmai_v4l2DriverAlloc(Int fd, Int numBufs, enum v4l2_buf_type type, 5 struct _VideoBufDesc **bufDescsPtr, 6 BufTab_Handle *hBufTabPtr, Int topOffset, 7 ColorSpace_Type colorSpace) 8 { 9 BufferGfx_Attrs gfxAttrs = BufferGfx_Attrs_DEFAULT; 10 struct v4l2_requestbuffers req; 11 struct v4l2_format fmt; 12 _VideoBufDesc *bufDesc; 13 Buffer_Handle hBuf; 14 Int bufIdx; 15 Int8 *virtPtr; 16 17 Dmai_clear(fmt); 18 fmt.type = type; 19 20 if (ioctl(fd, VIDIOC_G_FMT, &fmt) == -1) { 21 Dmai_err1("VIDIOC_G_FMT failed (%s)\n", strerror(errno)); 22 return Dmai_EFAIL; 23 } 24 25 Dmai_clear(req); 26 req.count = numBufs; 27 req.type = type; 28 req.memory = V4L2_MEMORY_MMAP; 29 30 /* Allocate buffers in the capture device driver */ 31 /* 申请建立V4L2的视频缓冲区管理队列*/ 32 if (ioctl(fd, VIDIOC_REQBUFS, &req) == -1) { 33 Dmai_err1("VIDIOC_REQBUFS failed (%s)\n", strerror(errno)); 34 return Dmai_ENOMEM; 35 } 36 37 if (req.count < numBufs || !req.count) { 38 Dmai_err0("Insufficient device driver buffer memory\n"); 39 return Dmai_ENOMEM; 40 } 41 42 /* Allocate space for buffer descriptors */ 43 *bufDescsPtr = calloc(numBufs, sizeof(_VideoBufDesc)); 44 45 if (*bufDescsPtr == NULL) { 46 Dmai_err0("Failed to allocate space for buffer descriptors\n"); 47 return Dmai_ENOMEM; 48 } 49 50 gfxAttrs.dim.width = fmt.fmt.pix.width; 51 gfxAttrs.dim.height = fmt.fmt.pix.height; 52 gfxAttrs.dim.lineLength = fmt.fmt.pix.bytesperline; 53 gfxAttrs.colorSpace = colorSpace; 54 gfxAttrs.bAttrs.reference = TRUE; 55 56 /* 调用CMEM建立缓冲区 */ 57 *hBufTabPtr = BufTab_create(numBufs, fmt.fmt.pix.sizeimage, 58 BufferGfx_getBufferAttrs(&gfxAttrs)); 59 60 if (*hBufTabPtr == NULL) { 61 return Dmai_ENOMEM; 62 } 63 64 /* 将建立好的缓冲区放到队列中并且配置好相关属性 */ 65 for (bufIdx = 0; bufIdx < numBufs; bufIdx++) { 66 bufDesc = &(*bufDescsPtr)[bufIdx]; 67 68 /* Ask for information about the driver buffer */ 69 Dmai_clear(bufDesc->v4l2buf); 70 bufDesc->v4l2buf.type = type; 71 bufDesc->v4l2buf.memory = V4L2_MEMORY_MMAP; 72 bufDesc->v4l2buf.index = bufIdx; 73 74 /* 查找队列中的缓冲区 */ 75 if (ioctl(fd, VIDIOC_QUERYBUF, &bufDesc->v4l2buf) == -1) { 76 Dmai_err1("Failed VIDIOC_QUERYBUF (%s)\n", strerror(errno)); 77 return Dmai_EFAIL; 78 } 79 80 /* 修改缓冲区的属性 */ 81 82 /* Map the driver buffer to user space */ 83 virtPtr = mmap(NULL, 84 bufDesc->v4l2buf.length, 85 PROT_READ | PROT_WRITE, 86 MAP_SHARED, 87 fd, 88 bufDesc->v4l2buf.m.offset) + topOffset; 89 90 if (virtPtr == MAP_FAILED) { 91 Dmai_err1("Failed to mmap buffer (%s)\n", strerror(errno)); 92 return Dmai_EFAIL; 93 } 94 95 /* Initialize the Buffer with driver buffer information */ 96 hBuf = BufTab_getBuf(*hBufTabPtr, bufIdx); 97 98 Buffer_setNumBytesUsed(hBuf, fmt.fmt.pix.bytesperline * 99 fmt.fmt.pix.height); 100 Buffer_setUseMask(hBuf, gfxAttrs.bAttrs.useMask); 101 Buffer_setUserPtr(hBuf, virtPtr); 102 103 /* Initialize buffer to black */ 104 _Dmai_blackFill(hBuf); 105 106 Dmai_dbg3("Driver buffer %d mapped to %#x has physical address " 107 "%#lx\n", bufIdx, (Int) virtPtr, Buffer_getPhysicalPtr(hBuf)); 108 109 bufDesc->hBuf = hBuf; 110 111 /* Queue buffer in device driver */ 112 /* 将缓冲区放回队列中 */ 113 if (ioctl(fd, VIDIOC_QBUF, &bufDesc->v4l2buf) == -1) { 114 Dmai_err1("VIODIOC_QBUF failed (%s)\n", strerror(errno)); 115 return Dmai_EFAIL; 116 } 117 } 118 119 return Dmai_EOK; 120 }
继续分析Buffer_create()函数
查看dmai_2_20_00_15/packages/ti/sdo/dmai/Buffer.c。
1 /****************************************************************************** 2 * Buffer_create 3 ******************************************************************************/ 4 Buffer_Handle Buffer_create(Int32 size, Buffer_Attrs *attrs) 5 { 6 Buffer_Handle hBuf; 7 UInt32 objSize; 8 9 if (attrs == NULL) { 10 Dmai_err0("Must provide attrs\n"); 11 return NULL; 12 } 13 14 if (attrs->type != Buffer_Type_BASIC && 15 attrs->type != Buffer_Type_GRAPHICS) { 16 17 Dmai_err1("Unknown Buffer type (%d)\n", attrs->type); 18 return NULL; 19 } 20 21 objSize = attrs->type == Buffer_Type_GRAPHICS ? sizeof(_BufferGfx_Object) : 22 sizeof(_Buffer_Object); 23 24 hBuf = (Buffer_Handle) calloc(1, objSize); 25 26 if (hBuf == NULL) { 27 Dmai_err0("Failed to allocate space for Buffer Object\n"); 28 return NULL; 29 } 30 31 _Buffer_init(hBuf, size, attrs); 32 33 if (!attrs->reference) { 34 35 /* 这里就是调用了CMEM的接口进行缓冲区的创建 */ 36 hBuf->userPtr = (Int8*)Memory_alloc(size, &attrs->memParams); 37 38 if (hBuf->userPtr == NULL) { 39 printf("Failed to allocate memory.\n"); 40 free(hBuf); 41 return NULL; 42 } 43 44 /* 获取缓冲区的物理地址 */ 45 hBuf->physPtr = Memory_getBufferPhysicalAddress(hBuf->userPtr, 46 size, NULL); 47 48 Dmai_dbg3("Alloc Buffer of size %u at 0x%x (0x%x phys)\n", 49 (Uns) size, (Uns) hBuf->userPtr, (Uns) hBuf->physPtr); 50 } 51 52 hBuf->reference = attrs->reference; 53 54 return hBuf; 55 }
数据已经得到了,它们就放在通过CMEM创建的缓冲区里,那么验证或使用这些数据的一种最简单又有效的方式就是将它们直接
显示出来。再来看一下DVSDK的框图:
框图的两条红线表示数据的流向,需要将数据显示出来只需要两部,第一步创建V4L2的显示出输出设备,第二创建显示缓冲区并将采集到的数据放到缓冲区内。
创建V4L2的显示设备和创建VL42的采集设备很相似,请查看dmai_2_20_00_15/packages/ti/sdo/dmai/linux/dm6467/Display_v4l2.c
1 /****************************************************************************** 2 * Display_v4l2_create 3 ******************************************************************************/ 4 Display_Handle Display_v4l2_create(BufTab_Handle hBufTab, Display_Attrs *attrs) 5 { 6 struct v4l2_format fmt; 7 enum v4l2_buf_type type; 8 Display_Handle hDisplay; 9 10 assert(attrs); 11 12 Dmai_clear(fmt); 13 14 /* delayStreamon not supported for this platform */ 15 if (attrs->delayStreamon == TRUE) { 16 Dmai_err0("Support for delayed VIDIOC_STREAMON not implemented\n"); 17 return NULL; 18 } 19 20 /* Allocate space for state object */ 21 hDisplay = calloc(1, sizeof(Display_Object)); 22 23 if (hDisplay == NULL) { 24 Dmai_err0("Failed to allocate space for Display Object\n"); 25 return NULL; 26 } 27 28 hDisplay->userAlloc = TRUE; 29 30 /* Open video capture device */ 31 hDisplay->fd = open(attrs->displayDevice, O_RDWR, 0); 32 33 if (hDisplay->fd == -1) { 34 Dmai_err2("Cannot open %s (%s)\n", 35 attrs->displayDevice, strerror(errno)); 36 cleanup(hDisplay); 37 return NULL; 38 } 39 40 if(Display_detectVideoStd(hDisplay, attrs) != Dmai_EOK) { 41 Dmai_err0("Display_detectVideoStd Failed\n"); 42 cleanup(hDisplay); 43 return NULL; 44 } 45 /* Determine the video image dimensions */ 46 fmt.type = V4L2_BUF_TYPE_VIDEO_OUTPUT; 47 48 if (ioctl(hDisplay->fd, VIDIOC_G_FMT, &fmt) == -1) { 49 Dmai_err0("Failed to determine video display format\n"); 50 cleanup(hDisplay); 51 return NULL; 52 } 53 54 fmt.type = V4L2_BUF_TYPE_VIDEO_OUTPUT; 55 switch(attrs->colorSpace) { 56 case ColorSpace_UYVY: 57 fmt.fmt.pix.pixelformat = V4L2_PIX_FMT_UYVY; 58 break; 59 case ColorSpace_YUV420PSEMI: 60 fmt.fmt.pix.pixelformat = V4L2_PIX_FMT_NV12; 61 break; 62 case ColorSpace_YUV422PSEMI: 63 fmt.fmt.pix.pixelformat = V4L2_PIX_FMT_NV16; 64 break; 65 default: 66 Dmai_err1("Unsupported color format %g\n", attrs->colorSpace); 67 cleanup(hDisplay); 68 return NULL; 69 }; 70 71 if (hBufTab == NULL) { 72 fmt.fmt.pix.bytesperline = Dmai_roundUp(BufferGfx_calcLineLength(fmt.fmt.pix.width, 73 attrs->colorSpace), 32); 74 fmt.fmt.pix.sizeimage = BufferGfx_calcSize(attrs->videoStd, attrs->colorSpace); 75 #if 1 76 } else { 77 /* This will help user to pass lineLength to display driver. */ 78 Buffer_Handle hBuf; 79 BufferGfx_Dimensions dim; 80 81 hBuf = BufTab_getBuf(hBufTab, 0); 82 BufferGfx_getDimensions(hBuf, &dim); 83 if((dim.height > fmt.fmt.pix.height) || 84 (dim.width > fmt.fmt.pix.width)) { 85 Dmai_err2("User buffer size check failed %dx%d\n", 86 dim.height, dim.width); 87 cleanup(hDisplay); 88 return NULL; 89 } 90 fmt.fmt.pix.bytesperline = dim.lineLength; 91 fmt.fmt.pix.sizeimage = Buffer_getSize(hBuf); 92 } 93 #endif 94 Dmai_dbg4("Video output set to size %dx%d pitch %d imageSize %d\n", 95 fmt.fmt.pix.width, fmt.fmt.pix.height, 96 fmt.fmt.pix.bytesperline, fmt.fmt.pix.sizeimage); 97 98 if ((attrs->videoStd == VideoStd_CIF) || (attrs->videoStd == VideoStd_SIF_PAL) || 99 (attrs->videoStd == VideoStd_SIF_NTSC) || (attrs->videoStd == VideoStd_D1_PAL) || 100 (attrs->videoStd == VideoStd_D1_NTSC) || (attrs->videoStd == VideoStd_1080I_30) || 101 (attrs->videoStd == VideoStd_1080I_25)) { 102 fmt.fmt.pix.field = V4L2_FIELD_INTERLACED; 103 } else { 104 fmt.fmt.pix.field = V4L2_FIELD_NONE; 105 } 106 107 if (ioctl(hDisplay->fd, VIDIOC_S_FMT, &fmt) == -1) { 108 Dmai_err2("Failed VIDIOC_S_FMT on %s (%s)\n", attrs->displayDevice, 109 strerror(errno)); 110 cleanup(hDisplay); 111 return NULL; 112 } 113 114 /* Should the device driver allocate the display buffers? */ 115 if (hBufTab == NULL) { 116 hDisplay->userAlloc = FALSE; 117 118 if (_Dmai_v4l2DriverAlloc(hDisplay->fd, 119 attrs->numBufs, 120 V4L2_BUF_TYPE_VIDEO_OUTPUT, 121 &hDisplay->bufDescs, 122 &hBufTab, 123 0, attrs->colorSpace) < 0) { 124 Dmai_err1("Failed to allocate display driver buffers on %s\n", 125 attrs->displayDevice); 126 cleanup(hDisplay); 127 return NULL; 128 } 129 } 130 else { 131 hDisplay->userAlloc = TRUE; 132 133 if (_Dmai_v4l2UserAlloc(hDisplay->fd, 134 attrs->numBufs, 135 V4L2_BUF_TYPE_VIDEO_OUTPUT, 136 &hDisplay->bufDescs, 137 hBufTab, 138 0, attrs->colorSpace) < 0) { 139 Dmai_err1("Failed to intialize display driver buffers on %s\n", 140 attrs->displayDevice); 141 cleanup(hDisplay); 142 return NULL; 143 } 144 } 145 146 /* Start the video streaming */ 147 type = V4L2_BUF_TYPE_VIDEO_OUTPUT; 148 149 if (ioctl(hDisplay->fd, VIDIOC_STREAMON, &type) == -1) { 150 Dmai_err2("VIDIOC_STREAMON failed on %s (%s)\n", attrs->displayDevice, 151 strerror(errno)); 152 cleanup(hDisplay); 153 return NULL; 154 } 155 156 hDisplay->started = TRUE; 157 hDisplay->hBufTab = hBufTab; 158 hDisplay->displayStd = Display_Std_V4L2; 159 160 return hDisplay; 161 }
创建显示设备不需要检查视频标准,因为输出的标准是有调用者决定的。
显示设备创建好了,那怎样将采集到的数据显示出来呢?TI提供了一个叫encode的demo程序,大家可以在dvsdk_demos_3_10_00_16中找到。
这个demo中很巧妙的将视频采集缓冲区和视频显示缓冲区管理起来,接下来一下分析一下。
因为单个显示缓冲和采集缓冲区的大小是一样的,所以可以直接进行交互。
下面截取demo中的代码片段,看一下这个缓冲区的交互的具体实现。
1 /* 采集及显示线程中的循环 */ 2 while (!gblGetQuit()) { 3 4 /* 获取采集缓冲区 */ 5 if (Capture_get(hCapture, &hCapBuf) < 0) { 6 ERR("Failed to get capture buffer\n"); 7 cleanup(THREAD_FAILURE); 8 } 9 10 /* 获取显示缓冲区 */ 11 if (Display_get(hDisplay, &hDisBuf) < 0) { 12 ERR("Failed to get display buffer\n"); 13 cleanup(THREAD_FAILURE); 14 } 15 16 /* 叠加字幕 */ 17 if (envp->osd) { 18 /* Get the current transparency */ 19 trans = UI_getTransparency(envp->hUI); 20 21 if (trans != oldTrans) { 22 /* Change the transparency in the palette */ 23 for (i = 0; i < 4; i++) { 24 bConfigParams.palette[i][3] = trans; 25 } 26 27 /* Reconfigure the blending job if transparency has changed */ 28 if (Blend_config(hBlend, NULL, hBmpBuf, hCapBuf, hCapBuf, 29 &bConfigParams) < 0) { 30 ERR("Failed to configure blending job\n"); 31 cleanup(THREAD_FAILURE); 32 } 33 } 34 35 /* 36 * Because the whole screen is shown even if -r is used, 37 * reset the dimensions while Blending to make sure the OSD 38 * always ends up in the same place. After blending, restore 39 * the real dimensions. 40 */ 41 BufferGfx_getDimensions(hCapBuf, &srcDim); 42 BufferGfx_resetDimensions(hCapBuf); 43 44 /* 45 * Lock the screen making sure no changes are done to 46 * the bitmap while we render it. 47 */ 48 hBmpBuf = UI_lockScreen(envp->hUI); 49 50 /* Execute the blending job to draw the OSD */ 51 /* 直接叠加在采集缓冲区的数据上 */ 52 if (Blend_execute(hBlend, hBmpBuf, hCapBuf, hCapBuf) < 0) { 53 ERR("Failed to execute blending job\n"); 54 cleanup(THREAD_FAILURE); 55 } 56 57 UI_unlockScreen(envp->hUI); 58 59 BufferGfx_setDimensions(hCapBuf, &srcDim); 60 } 61 62 /* Color convert the captured buffer from 422Psemi to 420Psemi */ 63 /* 在进行H264编码前需要进行 422 到 420的颜色空间转换,hDstBuf是从视频编码缓冲区队列中得到的一个缓冲区 */ 64 if (Ccv_execute(hCcv, hCapBuf, hDstBuf) < 0) { 65 ERR("Failed to execute color conversion job\n"); 66 cleanup(THREAD_FAILURE); 67 } 68 69 /* Send color converted buffer to video thread for encoding */ 70 /* 转换后将这个缓冲区放回视频编码缓冲区队列中 */ 71 if (Fifo_put(envp->hOutFifo, hDstBuf) < 0) { 72 ERR("Failed to send buffer to display thread\n"); 73 cleanup(THREAD_FAILURE); 74 } 75 76 BufferGfx_resetDimensions(hCapBuf); 77 78 /* Send the preview to the display device driver */ 79 /* 将采集缓冲区放到显示缓冲区队列中 */ 80 if (Display_put(hDisplay, hCapBuf) < 0) { 81 ERR("Failed to put display buffer\n"); 82 cleanup(THREAD_FAILURE); 83 } 84 85 BufferGfx_resetDimensions(hDisBuf); 86 87 /* Return a buffer to the capture driver */ 88 /* 将显示缓冲区放到采集缓冲区队列中 */ 89 if (Capture_put(hCapture, hDisBuf) < 0) { 90 ERR("Failed to put capture buffer\n"); 91 cleanup(THREAD_FAILURE); 92 } 93 94 /* Incremement statistics for the user interface */ 95 /* 帧计数加1 */ 96 gblIncFrames(); 97 98 /* Get a buffer from the video thread */ 99 /* 从视频编码缓冲区队列中得到一个缓冲区,用于下一次颜色空间转换使用 */ 100 fifoRet = Fifo_get(envp->hInFifo, &hDstBuf); 101 102 if (fifoRet < 0) { 103 ERR("Failed to get buffer from video thread\n"); 104 cleanup(THREAD_FAILURE); 105 } 106 107 /* Did the video thread flush the fifo? */ 108 if (fifoRet == Dmai_EFLUSH) { 109 cleanup(THREAD_SUCCESS); 110 } 111 }
到此为止,已经可以知道了数据是从哪里来到哪里去了,但是数据来了,肯定没那么容易就放它走,下一章将会讲到将采集到的数据如何编码并且保存。再加点预告,后面会讲到将编码后的数据通过live555发送出去,实现rtsp视频服务器。