/*M/////////////////////////////////////////////////////////////////////////////////////// // // IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING. // // By downloading, copying, installing or using the software you agree to this license. // If you do not agree to this license, do not download, install, // copy or use the software. // // // Intel License Agreement // For Open Source Computer Vision Library // // Copyright (C) 2000, Intel Corporation, all rights reserved. // Third party copyrights are property of their respective owners. // // Redistribution and use in source and binary forms, with or without modification, // are permitted provided that the following conditions are met: // // * Redistribution's of source code must retain the above copyright notice, // this list of conditions and the following disclaimer. // // * Redistribution's in binary form must reproduce the above copyright notice, // this list of conditions and the following disclaimer in the documentation // and/or other materials provided with the distribution. // // * The name of Intel Corporation may not be used to endorse or promote products // derived from this software without specific prior written permission. // // This software is provided by the copyright holders and contributors "as is" and // any express or implied warranties, including, but not limited to, the implied // warranties of merchantability and fitness for a particular purpose are disclaimed. // In no event shall the Intel Corporation or contributors be liable for any direct, // indirect, incidental, special, exemplary, or consequential damages // (including, but not limited to, procurement of substitute goods or services; // loss of use, data, or profits; or business interruption) however caused // and on any theory of liability, whether in contract, strict liability, // or tort (including negligence or otherwise) arising in any way out of // the use of this software, even if advised of the possibility of such damage. // //M*/ #include "_highgui.h" #ifdef WIN32 /****************** Capturing video from camera via CMU lib *******************/ #if defined(HAVE_CMU1394) // This firewire capability added by Philip Gruebele (pgruebele@cox.net). // For this to work you need to install the CMU firewire DCAM drivers, // located at http://www-2.cs.cmu.edu/~iwan/1394/. #include "1394camera.h" typedef struct CvCaptureCAM_CMU { CvCaptureVTable* vtable; int fps; // 0-5 int mode; // 0-7 int format; // 0-2, 7 ? int index; IplImage * image; IplImage* rgb_frame; } CvCaptureCAM_CMU; int icvOpenCAM_CMU(CvCaptureCAM_CMU * capture, int wIndex ); int icvSetPropertyCAM_CMU( CvCaptureCAM_CMU* capture, int property_id, double value ); void icvCloseCAM_CMU( CvCaptureCAM_CMU* capture ); int icvGrabFrameCAM_CMU( CvCaptureCAM_CMU* capture ); IplImage* icvRetrieveFrameCAM_CMU( CvCaptureCAM_CMU* capture ); double icvGetPropertyCAM_CMU( CvCaptureCAM_CMU* capture, int property_id ); int icvSetPropertyCAM_CMU( CvCaptureCAM_CMU* capture, int property_id, double value ); static CvCaptureVTable captureCAM_CMU_vtable = { 6, (CvCaptureCloseFunc)icvCloseCAM_CMU, (CvCaptureGrabFrameFunc)icvGrabFrameCAM_CMU, (CvCaptureRetrieveFrameFunc)icvRetrieveFrameCAM_CMU, (CvCaptureGetPropertyFunc)icvGetPropertyCAM_CMU, (CvCaptureSetPropertyFunc)icvSetPropertyCAM_CMU, (CvCaptureGetDescriptionFunc)0 }; // CMU 1394 camera stuff. // This firewire capability added by Philip Gruebele (pgruebele@cox.net) // and modified by Roman Stanchak (rstanchak@yahoo.com). // For this to work you need to install the CMU firewire DCAM drivers, // located at http://www-2.cs.cmu.edu/~iwan/1394/. #define CMU_MAX_CAMERAS 20 int CMU_numCameras = 0; int CMU_numActiveCameras = 0; bool CMU_useCameraFlags[CMU_MAX_CAMERAS]; C1394Camera *CMU_theCamera = 0; // stupid defines for mode, format, FPS #define CV_CAP_IEEE1394_FPS_1_875 0 #define CV_CAP_IEEE1394_FPS_3_75 1 #define CV_CAP_IEEE1394_FPS_7_5 2 #define CV_CAP_IEEE1394_FPS_15 3 #define CV_CAP_IEEE1394_FPS_30 4 #define CV_CAP_IEEE1394_FPS_60 5 // index by size, color #define CV_CAP_IEEE1394_COLOR_MONO 0 #define CV_CAP_IEEE1394_COLOR_MONO16 1 #define CV_CAP_IEEE1394_COLOR_YUV444 2 #define CV_CAP_IEEE1394_COLOR_YUV422 3 #define CV_CAP_IEEE1394_COLOR_YUV411 4 #define CV_CAP_IEEE1394_COLOR_RGB 5 #define CV_CAP_IEEE1394_SIZE_160X120 0 #define CV_CAP_IEEE1394_SIZE_320X240 1 #define CV_CAP_IEEE1394_SIZE_640X480 2 #define CV_CAP_IEEE1394_SIZE_800X600 3 #define CV_CAP_IEEE1394_SIZE_1024X768 4 #define CV_CAP_IEEE1394_SIZE_1280X960 5 #define CV_CAP_IEEE1394_SIZE_1600X1200 6 // given color, size, output format // 1 16 444 422 411 RGB char CV_CAP_IEEE1394_FORMAT[7][6] = { {-1, -1, 0, -1, -1, -1}, // 160x120 {-1, -1, -1, 0, -1, -1}, // 320x240 { 0, 0, -1, 0, 0, 0}, // 640x480 { 1, 1, -1, 1, -1, 1}, // 800x600 { 1, 1, -1, 1, -1, 1}, // 1024x768 { 2, 2, -1, 2, -1, 2}, // 1280x960 { 2, 2, -1, 2, -1, 2}}; // 1600x1200 // given color, size, output corresponding mode char CV_CAP_IEEE1394_MODE[7][6] = { {-1, -1, 0, -1, -1, -1}, // 160x120 {-1, -1, -1, 1, -1, -1}, // 320x240 { 5, 6, -1, 3, 2, 4}, // 640x480 { 2, 6, -1, 0, -1, 1}, // 800x600 { 5, 7, -1, 3, -1, 4}, // 1024x768 { 2, 6, -1, 0, -1, 1}, // 1280x960 { 5, 7, -1, 3, -1, 4}}; // 1600x1200 // given format, mode, return COLOR char CV_CAP_IEEE1394_COLOR[2][8] = { { CV_CAP_IEEE1394_COLOR_YUV444, CV_CAP_IEEE1394_COLOR_YUV422, CV_CAP_IEEE1394_COLOR_YUV411, CV_CAP_IEEE1394_COLOR_YUV422, CV_CAP_IEEE1394_COLOR_RGB, CV_CAP_IEEE1394_COLOR_MONO, CV_CAP_IEEE1394_COLOR_MONO16 }, { CV_CAP_IEEE1394_COLOR_YUV422, CV_CAP_IEEE1394_COLOR_RGB, CV_CAP_IEEE1394_COLOR_MONO, CV_CAP_IEEE1394_COLOR_YUV422, CV_CAP_IEEE1394_COLOR_RGB, CV_CAP_IEEE1394_COLOR_MONO, CV_CAP_IEEE1394_COLOR_MONO16, CV_CAP_IEEE1394_COLOR_MONO16 } }; // convert frame rate to suitable enum /*static int icvFrameRateToIndex_CMU(double framerate){ if(framerate > 30) return CV_CAP_IEEE1394_FPS_60; else if(framerate > 15) return CV_CAP_IEEE1394_FPS_30; else if(framerate > 7.5) return CV_CAP_IEEE1394_FPS_15; else if(framerate > 3.75) return CV_CAP_IEEE1394_FPS_7_5; else if(framerate > 1.875) return CV_CAP_IEEE1394_FPS_3_75; return CV_CAP_IEEE1394_FPS_1_875; }*/ #if _MSC_VER >= 1200 #pragma comment(lib,"1394camera.lib") #endif // return the size of the image CvSize icvGetSize_CMU( C1394Camera * cmucam ){ //int format = cmucam->GetVideoFormat(); //int mode = cmucam->GetVideoMode(); // irrelvant to depth // if( format > 1 ) format=1; // for YUV, should we return a weirdly sized image? //switch(CV_CAP_IEEE1394_COLOR[format][mode]){ //case CV_CAP_IEEE1394_COLOR_MONO16: // return cvSize(cmucam->m_width*2, cmucam->m_height); //default: // return cvSize(cmucam->m_width, cmucam->m_height); //} unsigned long width = 0, height = 0; cmucam->GetVideoFrameDimensions( &width, &height ); return cvSize((int)width, (int)height); } // return the opencv depth flag corresponding to the camera format int icvGetDepth_CMU( C1394Camera * cmucam ){ int format = cmucam->GetVideoFormat(); int mode = cmucam->GetVideoMode(); // TODO if( format==7 ) { assert(0); } // irrelvant to depth if( format > 1 ) format=1; if(CV_CAP_IEEE1394_COLOR[format][mode]==CV_CAP_IEEE1394_COLOR_MONO16){ return IPL_DEPTH_16S; } return IPL_DEPTH_8U; } // return the number of channels for camera int icvGetNChannels_CMU( C1394Camera * cmucam ){ int format = cmucam->GetVideoFormat(); int mode = cmucam->GetVideoMode(); if( format==7 ){ assert(0); } // irrelvant to nchannels if( format > 1 ) format=1; switch(CV_CAP_IEEE1394_COLOR[format][mode]){ case CV_CAP_IEEE1394_COLOR_RGB: return 3; case CV_CAP_IEEE1394_COLOR_MONO: case CV_CAP_IEEE1394_COLOR_MONO16: return 1; case CV_CAP_IEEE1394_COLOR_YUV422: case CV_CAP_IEEE1394_COLOR_YUV444: case CV_CAP_IEEE1394_COLOR_YUV411: return 3; default: ; } return -1; } int icvOpenCAM_CMU(CvCaptureCAM_CMU * capture, int index ) { // if first time, then allocate all available cameras if( CMU_numCameras == 0 ) { CMU_numActiveCameras = 0; CMU_theCamera = new C1394Camera[CMU_MAX_CAMERAS]; //////////////////////////////////////////////////////////////////////////////////////////////////////// // create all cameras try { // create camera0 if( CMU_theCamera[0].CheckLink() != CAM_SUCCESS ) throw 1; // we have one pin per camera CMU_numCameras = CMU_theCamera[0].GetNumberCameras(); // allocate remaining cameras for(int i = 1; i < CMU_numCameras && i=0 && !found_format; rate--){ for (int color=CV_CAP_IEEE1394_COLOR_RGB; color>=0 && !found_format; color--){ for (int size=CV_CAP_IEEE1394_SIZE_1600X1200; size>=0 && !found_format; size--){ int format = CV_CAP_IEEE1394_FORMAT[size][color]; int mode = CV_CAP_IEEE1394_MODE[size][color]; if (format!=-1 && mode!=-1 && CMU_theCamera[index].HasVideoFrameRate(format,mode,rate)){ CMU_theCamera[index].SetVideoFormat(format); CMU_theCamera[index].SetVideoMode(mode); CMU_theCamera[index].SetVideoFrameRate(rate); found_format = (CMU_theCamera[index].StartImageAcquisition() == CAM_SUCCESS); } } } } // try format 7 if(!found_format){ CMU_theCamera[index].SetVideoFormat(7); CMU_theCamera[index].SetVideoMode(0); if(CMU_theCamera[index].StartImageAcquisition() != CAM_SUCCESS){ // no format found throw 9; } } // allocate image frame capture->image = cvCreateImage(icvGetSize_CMU(&(CMU_theCamera[index])), icvGetDepth_CMU(&(CMU_theCamera[index])), icvGetNChannels_CMU(&(CMU_theCamera[index]))); if (capture->image) { // successfully activated camera capture->index = index; //CMU_numActiveCameras; CMU_numActiveCameras++; CMU_useCameraFlags[index] = true; } } catch ( int ) { return 0; } return 1; } void icvCloseCAM_CMU( CvCaptureCAM_CMU* capture ) { if( capture && capture->image ) { cvReleaseImage( &capture->image ); CMU_theCamera[capture->index].StopImageAcquisition(); CMU_useCameraFlags[capture->index] = false; CMU_numActiveCameras--; if (!CMU_numActiveCameras) { delete[] CMU_theCamera; CMU_theCamera = 0; CMU_numCameras = 0; } } } int icvGrabFrameCAM_CMU( CvCaptureCAM_CMU* capture ) { return capture->image && CMU_theCamera && CMU_theCamera[capture->index].AcquireImage() == CAM_SUCCESS; } static void swapRedBlue(IplImage * im){ uchar * ptr = (uchar *) im->imageData; uchar t; for(int i=0; iheight; i++){ ptr = (uchar *) im->imageData+im->widthStep*i; for(int j=0; jwidth; j++){ t = ptr[0]; ptr[0] = ptr[2]; ptr[2] = t; ptr+=3; } } } IplImage* icvRetrieveFrameCAM_CMU( CvCaptureCAM_CMU* capture ) { if( capture->image && CMU_theCamera ) { //capture->image->imageData = (char *) CMU_theCamera[capture->index].GetRawData(0); CMU_theCamera[capture->index].getRGB((unsigned char*)capture->image->imageData, capture->image->imageSize); swapRedBlue( capture->image ); return capture->image; } return 0; } double icvGetPropertyCAM_CMU( CvCaptureCAM_CMU* capture, int property_id ) { switch( property_id ) { case CV_CAP_PROP_FRAME_WIDTH: return capture->image->width; case CV_CAP_PROP_FRAME_HEIGHT: return capture->image->height; case CV_CAP_PROP_FPS: return CMU_theCamera[capture->index].GetVideoFrameRate(); case CV_CAP_PROP_MODE: return CMU_theCamera[capture->index].GetVideoMode(); case CV_CAP_PROP_FORMAT: return CMU_theCamera[capture->index].GetVideoFormat(); } return 0; } static int icvSetVideoSize( CvCaptureCAM_CMU* capture, int, int) { if (capture==0) return 0; // change to closest size return -1; } static int icvSetMode(CvCaptureCAM_CMU * capture, int mode){ int format = CMU_theCamera[capture->index].GetVideoFormat(); if(mode < 0 || mode > 7){ return -1; } if(CMU_theCamera[capture->index].HasVideoMode(format, mode)){ CMU_theCamera[capture->index].StopImageAcquisition(); CMU_theCamera[capture->index].SetVideoMode(mode); CMU_theCamera[capture->index].StartImageAcquisition(); return 0; } return -1; } static int icvSetFrameRate(CvCaptureCAM_CMU * capture, int rate){ int mode = CMU_theCamera[capture->index].GetVideoMode(); int format = CMU_theCamera[capture->index].GetVideoFormat(); if(rate < 0 || rate > 5){ return -1; } if(CMU_theCamera[capture->index].HasVideoFrameRate(format, mode, rate)){ CMU_theCamera[capture->index].StopImageAcquisition(); CMU_theCamera[capture->index].SetVideoFrameRate(rate); CMU_theCamera[capture->index].StartImageAcquisition(); return 0; } return -1; } static int icvSetFormat(CvCaptureCAM_CMU * capture, int format){ if(format < 0 || format > 2){ return -1; } if(CMU_theCamera[capture->index].HasVideoFormat(format)){ CMU_theCamera[capture->index].StopImageAcquisition(); CMU_theCamera[capture->index].SetVideoFormat(format); CMU_theCamera[capture->index].StartImageAcquisition(); return 0; } return -1; } int icvSetPropertyCAM_CMU( CvCaptureCAM_CMU* capture, int property_id, double value ){ int retval = -1; int ival= cvRound(value); switch (property_id) { case CV_CAP_PROP_FRAME_WIDTH: case CV_CAP_PROP_FRAME_HEIGHT: { int width, height; if (property_id == CV_CAP_PROP_FRAME_WIDTH) { width = ival; height = width*3/4; } else { height = ival; width = height*4/3; } retval = icvSetVideoSize(capture, width, height); } break; case CV_CAP_PROP_FPS: retval =icvSetFrameRate(capture, ival); break; case CV_CAP_PROP_MODE: retval =icvSetMode(capture, ival); break; case CV_CAP_PROP_FORMAT: retval =icvSetFormat(capture, ival); break; } // resize image if its not the right size anymore CvSize size = icvGetSize_CMU( &(CMU_theCamera[capture->index]) ); int depth = icvGetDepth_CMU( &(CMU_theCamera[capture->index]) ); int nch = icvGetNChannels_CMU( &(CMU_theCamera[capture->index]) ); if(size.width != capture->image->width || size.height != capture->image->height || depth != capture->image->depth || nch != capture->image->nChannels ) { cvReleaseImage(&capture->image); capture->image = cvCreateImage(size, depth, nch); } return retval; } CvCapture * cvCaptureFromCAM_CMU (int index) { CvCaptureCAM_CMU* capture = (CvCaptureCAM_CMU*)cvAlloc( sizeof(*capture)); memset( capture, 0, sizeof(*capture)); capture->vtable = &captureCAM_CMU_vtable; if( icvOpenCAM_CMU( capture, index )) return (CvCapture*)capture; cvReleaseCapture( (CvCapture**)&capture ); return 0; } #endif // CMU #endif // WIN32