Visual Servoing Platform  version 3.1.0
mbtGenericTracking2.cpp

Example of tracking with vpGenericTracker on an image sequence containing a cube.

/****************************************************************************
*
* This file is part of the ViSP software.
* Copyright (C) 2005 - 2017 by Inria. All rights reserved.
*
* This software is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
* See the file LICENSE.txt at the root directory of this source
* distribution for additional information about the GNU GPL.
*
* For using ViSP with software that can not be combined with the GNU
* GPL, please contact Inria about acquiring a ViSP Professional
* Edition License.
*
* See http://visp.inria.fr for more information.
*
* This software was developed at:
* Inria Rennes - Bretagne Atlantique
* Campus Universitaire de Beaulieu
* 35042 Rennes Cedex
* France
*
* If you have questions regarding the use of this file, please contact
* Inria at visp@inria.fr
*
* This file is provided AS IS with NO WARRANTY OF ANY KIND, INCLUDING THE
* WARRANTY OF DESIGN, MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE.
*
* Description:
* Example of Hybrid Tracking of MBT and MBT KTL.
*
* Authors:
* Aurelien Yol
* Souriya Trinh
*
*****************************************************************************/
#include <iostream>
#include <visp3/core/vpConfig.h>
#if defined(VISP_HAVE_MODULE_MBT) && defined(VISP_HAVE_DISPLAY)
#include <visp3/core/vpDebug.h>
#include <visp3/core/vpHomogeneousMatrix.h>
#include <visp3/core/vpIoTools.h>
#include <visp3/core/vpMath.h>
#include <visp3/gui/vpDisplayD3D.h>
#include <visp3/gui/vpDisplayGDI.h>
#include <visp3/gui/vpDisplayGTK.h>
#include <visp3/gui/vpDisplayOpenCV.h>
#include <visp3/gui/vpDisplayX.h>
#include <visp3/io/vpImageIo.h>
#include <visp3/io/vpParseArgv.h>
#include <visp3/io/vpVideoReader.h>
#include <visp3/mbt/vpMbGenericTracker.h>
#define GETOPTARGS "x:m:i:n:de:chtfColwvpT:"
#define USE_XML 0
void usage(const char *name, const char *badparam)
{
fprintf(stdout, "\n\
Example of tracking based on the 3D model.\n\
\n\
SYNOPSIS\n\
%s [-i <test image path>] [-x <config file>]\n\
[-m <model name>] [-n <initialisation file base name>] [-e <last frame index>]\n\
[-t] [-c] [-d] [-h] [-f] [-C] [-o] [-w] [-l] [-v] [-p]\n\
[-T <tracker type>]\n", name);
fprintf(stdout, "\n\
OPTIONS: \n\
-i <input image path> \n\
Set image input path.\n\
From this path read images \n\
\"mbt/cube/image%%04d.ppm\". These \n\
images come from ViSP-images-x.y.z.tar.gz available \n\
on the ViSP website.\n\
Setting the VISP_INPUT_IMAGE_PATH environment\n\
variable produces the same behavior than using\n\
this option.\n\
\n\
-x <config file> \n\
Set the config file (the xml file) to use.\n\
The config file is used to specify the parameters of the tracker.\n\
\n\
-m <model name> \n\
Specify the name of the file of the model\n\
The model can either be a vrml model (.wrl) or a .cao file.\n\
\n\
-e <last frame index> \n\
Specify the index of the last frame. Once reached, the tracking is stopped\n\
\n\
-f \n\
Do not use the vrml model, use the .cao one. These two models are \n\
equivalent and comes from ViSP-images-x.y.z.tar.gz available on the ViSP\n\
website. However, the .cao model allows to use the 3d model based tracker \n\
without Coin.\n\
\n\
-C \n\
Track only the cube (not the cylinder). In this case the models files are\n\
cube.cao or cube.wrl instead of cube_and_cylinder.cao and \n\
cube_and_cylinder.wrl.\n\
\n\
-n <initialisation file base name> \n\
Base name of the initialisation file. The file will be 'base_name'.init .\n\
This base name is also used for the optional picture specifying where to \n\
click (a .ppm picture).\n\
\n\
-t \n\
Turn off the display of the the moving edges and Klt points. \n\
\n\
-d \n\
Turn off the display.\n\
\n\
-c\n\
Disable the mouse click. Useful to automate the \n\
execution of this program without human intervention.\n\
\n\
-o\n\
Use Ogre3D for visibility tests\n\
\n\
-w\n\
When Ogre3D is enable [-o] show Ogre3D configuration dialog that allows to set the renderer.\n\
\n\
-l\n\
Use the scanline for visibility tests.\n\
\n\
-v\n\
Compute covariance matrix.\n\
\n\
-p\n\
Compute gradient projection error.\n\
\n\
-T <tracker type>\n\
Set tracker type (<1 (Edge)>, <2 (KLT)>, <3 (EdgeKlt)>).\n\
\n\
-h \n\
Print the help.\n\n");
if (badparam)
fprintf(stdout, "\nERROR: Bad parameter [%s]\n", badparam);
}
bool getOptions(int argc, const char **argv, std::string &ipath, std::string &configFile, std::string &modelFile,
std::string &initFile, long &lastFrame, bool &displayFeatures, bool &click_allowed, bool &display,
bool &cao3DModel, bool &trackCylinder, bool &useOgre, bool &showOgreConfigDialog, bool &useScanline,
bool &computeCovariance, bool &projectionError, int &trackerType)
{
const char *optarg_;
int c;
while ((c = vpParseArgv::parse(argc, argv, GETOPTARGS, &optarg_)) > 1) {
switch (c) {
case 'e':
lastFrame = atol(optarg_);
break;
case 'i':
ipath = optarg_;
break;
case 'x':
configFile = optarg_;
break;
case 'm':
modelFile = optarg_;
break;
case 'n':
initFile = optarg_;
break;
case 't':
displayFeatures = false;
break;
case 'f':
cao3DModel = true;
break;
case 'c':
click_allowed = false;
break;
case 'd':
display = false;
break;
case 'C':
trackCylinder = false;
break;
case 'o':
useOgre = true;
break;
case 'l':
useScanline = true;
break;
case 'w':
showOgreConfigDialog = true;
break;
case 'v':
computeCovariance = true;
break;
case 'p':
projectionError = true;
break;
case 'T':
trackerType = atoi(optarg_);
break;
case 'h':
usage(argv[0], NULL);
return false;
break;
default:
usage(argv[0], optarg_);
return false;
break;
}
}
if ((c == 1) || (c == -1)) {
// standalone param or error
usage(argv[0], NULL);
std::cerr << "ERROR: " << std::endl;
std::cerr << " Bad argument " << optarg_ << std::endl << std::endl;
return false;
}
return true;
}
int main(int argc, const char **argv)
{
try {
std::string env_ipath;
std::string opt_ipath;
std::string ipath;
std::string opt_configFile;
std::string opt_modelFile;
std::string modelFile;
std::string opt_initFile;
std::string initFile;
long opt_lastFrame = -1;
bool displayFeatures = true;
bool opt_click_allowed = true;
bool opt_display = true;
bool cao3DModel = false;
bool trackCylinder = true;
bool useOgre = false;
bool showOgreConfigDialog = false;
bool useScanline = false;
bool computeCovariance = false;
bool projectionError = false;
// Get the visp-images-data package path or VISP_INPUT_IMAGE_PATH
// environment variable value
// Set the default input path
if (!env_ipath.empty())
ipath = env_ipath;
// Read the command line options
if (!getOptions(argc, argv, opt_ipath, opt_configFile, opt_modelFile, opt_initFile, opt_lastFrame, displayFeatures,
opt_click_allowed, opt_display, cao3DModel, trackCylinder, useOgre, showOgreConfigDialog,
useScanline, computeCovariance, projectionError, trackerType)) {
return EXIT_FAILURE;
}
// Test if an input path is set
if (opt_ipath.empty() && env_ipath.empty()) {
usage(argv[0], NULL);
std::cerr << std::endl << "ERROR:" << std::endl;
std::cerr << " Use -i <visp image path> option or set VISP_INPUT_IMAGE_PATH " << std::endl
<< " environment variable to specify the location of the " << std::endl
<< " image path where test images are located." << std::endl
<< std::endl;
return EXIT_FAILURE;
}
// Get the option values
if (!opt_ipath.empty())
ipath = vpIoTools::createFilePath(opt_ipath, "mbt/cube/image%04d.pgm");
else
ipath = vpIoTools::createFilePath(env_ipath, "mbt/cube/image%04d.pgm");
#if defined(VISP_HAVE_XML2) && USE_XML
std::string configFile;
if (!opt_configFile.empty())
configFile = opt_configFile;
else if (!opt_ipath.empty())
configFile = vpIoTools::createFilePath(opt_ipath, "mbt/cube.xml");
else
configFile = vpIoTools::createFilePath(env_ipath, "mbt/cube.xml");
#endif
if (!opt_modelFile.empty()) {
modelFile = opt_modelFile;
} else {
std::string modelFileCao;
std::string modelFileWrl;
if (trackCylinder) {
modelFileCao = "mbt/cube_and_cylinder.cao";
modelFileWrl = "mbt/cube_and_cylinder.wrl";
} else {
modelFileCao = "mbt/cube.cao";
modelFileWrl = "mbt/cube.wrl";
}
if (!opt_ipath.empty()) {
if (cao3DModel) {
modelFile = vpIoTools::createFilePath(opt_ipath, modelFileCao);
} else {
#ifdef VISP_HAVE_COIN3D
modelFile = vpIoTools::createFilePath(opt_ipath, modelFileWrl);
#else
std::cerr << "Coin is not detected in ViSP. Use the .cao model instead." << std::endl;
modelFile = vpIoTools::createFilePath(opt_ipath, modelFileCao);
#endif
}
} else {
if (cao3DModel) {
modelFile = vpIoTools::createFilePath(env_ipath, modelFileCao);
} else {
#ifdef VISP_HAVE_COIN3D
modelFile = vpIoTools::createFilePath(env_ipath, modelFileWrl);
#else
std::cerr << "Coin is not detected in ViSP. Use the .cao model instead." << std::endl;
modelFile = vpIoTools::createFilePath(env_ipath, modelFileCao);
#endif
}
}
}
if (!opt_initFile.empty())
initFile = opt_initFile;
else if (!opt_ipath.empty())
initFile = vpIoTools::createFilePath(opt_ipath, "mbt/cube");
else
initFile = vpIoTools::createFilePath(env_ipath, "mbt/cube");
std::map<std::string, const vpImage<unsigned char> *> mapOfImages;
vpVideoReader reader;
reader.setFileName(ipath);
try {
reader.open(I1);
I2 = I1;
I3 = I1;
} catch (...) {
std::cerr << "Cannot open sequence: " << ipath << std::endl;
return EXIT_FAILURE;
}
if (opt_lastFrame > 1 && opt_lastFrame < reader.getLastFrameIndex())
reader.setLastFrameIndex(opt_lastFrame);
reader.acquire(I1);
I2 = I1;
I3 = I1;
mapOfImages["Camera1"] = &I1;
mapOfImages["Camera2"] = &I2;
mapOfImages["Camera3"] = &I3;
// initialise a display
#if defined VISP_HAVE_X11
vpDisplayX display1, display2, display3;
#elif defined VISP_HAVE_GDI
vpDisplayGDI display1, display2, display3;
#elif defined VISP_HAVE_OPENCV
vpDisplayOpenCV display1, display2, display3;
#elif defined VISP_HAVE_D3D9
vpDisplayD3D display1, display2, display3;
#elif defined VISP_HAVE_GTK
vpDisplayGTK display1, display2, display3;
#else
opt_display = false;
#endif
if (opt_display) {
#if (defined VISP_HAVE_DISPLAY)
display1.init(I1, 100, 100, "Test tracking (Cam1)");
display2.init(I2, (int)(I1.getWidth() / vpDisplay::getDownScalingFactor(I1)) + 110, 100, "Test tracking (Cam2)");
display3.init(I3, 100, (int)(I1.getHeight() / vpDisplay::getDownScalingFactor(I1)) + 110, "Test tracking (Cam3)");
#endif
}
vpMbTracker *tracker = new vpMbGenericTracker(3, trackerType);
std::map<std::string, vpHomogeneousMatrix> mapOfCameraPoses;
std::map<std::string, vpCameraParameters> mapOfCameraParams;
// Initialise the tracker: camera parameters, moving edge and KLT settings
#if defined(VISP_HAVE_XML2) && USE_XML
// From the xml file
std::map<std::string, std::string> mapOfConfigFiles;
mapOfConfigFiles["Camera1"] = configFile;
mapOfConfigFiles["Camera2"] = configFile;
mapOfConfigFiles["Camera3"] = configFile;
dynamic_cast<vpMbGenericTracker *>(tracker)->loadConfigFile(mapOfConfigFiles);
#else
// By setting the parameters:
cam.initPersProjWithoutDistortion(547, 542, 338, 234);
mapOfCameraParams["Camera1"] = cam;
mapOfCameraParams["Camera2"] = cam;
mapOfCameraParams["Camera3"] = cam;
vpMe me;
me.setMaskSize(5);
me.setMaskNumber(180);
me.setRange(7);
me.setThreshold(5000);
me.setMu1(0.5);
me.setMu2(0.5);
std::map<std::string, vpMe> mapOfMe;
mapOfMe["Camera1"] = me;
mapOfMe["Camera2"] = me;
mapOfMe["Camera3"] = me;
#if defined(VISP_HAVE_MODULE_KLT) && (defined(VISP_HAVE_OPENCV) && (VISP_HAVE_OPENCV_VERSION >= 0x020100))
klt.setMaxFeatures(10000);
klt.setWindowSize(5);
klt.setQuality(0.01);
klt.setBlockSize(3);
std::map<std::string, vpKltOpencv> mapOfKlt;
mapOfKlt["Camera1"] = klt;
mapOfKlt["Camera2"] = klt;
mapOfKlt["Camera3"] = klt;
dynamic_cast<vpMbGenericTracker *>(tracker)->setKltOpencv(mapOfKlt);
dynamic_cast<vpMbGenericTracker *>(tracker)->setKltMaskBorder(5);
#endif
dynamic_cast<vpMbGenericTracker *>(tracker)->setCameraParameters(mapOfCameraParams);
dynamic_cast<vpMbGenericTracker *>(tracker)->setMovingEdge(mapOfMe);
tracker->setAngleAppear(vpMath::rad(65));
// Specify the clipping to
tracker->setNearClippingDistance(0.01);
tracker->setFarClippingDistance(0.90);
std::map<std::string, unsigned int> mapOfClippingFlags;
dynamic_cast<vpMbGenericTracker *>(tracker)->getClipping(mapOfClippingFlags);
for (std::map<std::string, unsigned int>::iterator it = mapOfClippingFlags.begin(); it != mapOfClippingFlags.end();
++it) {
it->second = (it->second | vpMbtPolygon::FOV_CLIPPING);
}
dynamic_cast<vpMbGenericTracker *>(tracker)->setClipping(mapOfClippingFlags);
// tracker->setClipping(tracker->getClipping() | vpMbtPolygon::LEFT_CLIPPING
// | vpMbtPolygon::RIGHT_CLIPPING | vpMbtPolygon::UP_CLIPPING |
// vpMbtPolygon::DOWN_CLIPPING); // Equivalent to FOV_CLIPPING
#endif
// Display the moving edges, and the Klt points
tracker->setDisplayFeatures(displayFeatures);
// Tells if the tracker has to use Ogre3D for visibility tests
tracker->setOgreVisibilityTest(useOgre);
if (useOgre)
tracker->setOgreShowConfigDialog(showOgreConfigDialog);
// Tells if the tracker has to use the scanline visibility tests
tracker->setScanLineVisibilityTest(useScanline);
// Tells if the tracker has to compute the covariance matrix
tracker->setCovarianceComputation(computeCovariance);
// Tells if the tracker has to compute the projection error
tracker->setProjectionErrorComputation(projectionError);
// Retrieve the camera parameters from the tracker
dynamic_cast<vpMbGenericTracker *>(tracker)->getCameraParameters(mapOfCameraParams);
// Loop to position the cube
if (opt_display && opt_click_allowed) {
while (!vpDisplay::getClick(I1, false)) {
vpDisplay::displayText(I1, 15, 10, "click after positioning the object", vpColor::red);
}
}
// Load the 3D model (either a vrml file or a .cao file)
tracker->loadModel(modelFile);
// Initialise the tracker by clicking on the image
// This function looks for
// - a ./cube/cube.init file that defines the 3d coordinates (in meter,
// in the object basis) of the points used for the initialisation
// - a ./cube/cube.ppm file to display where the user have to click
// (optionnal, set by the third parameter)
if (opt_display && opt_click_allowed) {
std::map<std::string, std::string> mapOfInitFiles;
mapOfInitFiles["Camera1"] = initFile;
dynamic_cast<vpMbGenericTracker *>(tracker)->initClick(mapOfImages, mapOfInitFiles, true);
dynamic_cast<vpMbGenericTracker *>(tracker)->getPose(mapOfCameraPoses);
// display the 3D model at the given pose
dynamic_cast<vpMbGenericTracker *>(tracker)->display(mapOfImages, mapOfCameraPoses, mapOfCameraParams,
} else {
vpHomogeneousMatrix c1Moi(0.02044769891, 0.1101505452, 0.5078963719, 2.063603907, 1.110231561, -0.4392789872);
std::map<std::string, vpHomogeneousMatrix> mapOfInitPoses;
mapOfInitPoses["Camera1"] = c1Moi;
dynamic_cast<vpMbGenericTracker *>(tracker)->initFromPose(mapOfImages, mapOfInitPoses);
}
// track the model
dynamic_cast<vpMbGenericTracker *>(tracker)->track(mapOfImages);
dynamic_cast<vpMbGenericTracker *>(tracker)->getPose(mapOfCameraPoses);
if (opt_display) {
}
bool quit = false, click = false;
while (!reader.end() && !quit) {
// acquire a new image
reader.acquire(I1);
I2 = I1;
I3 = I1;
mapOfImages["Camera1"] = &I1;
mapOfImages["Camera2"] = &I2;
mapOfImages["Camera3"] = &I3;
// display the image
if (opt_display) {
std::stringstream ss;
ss << "Num frame: " << reader.getFrameIndex() << "/" << reader.getLastFrameIndex();
vpDisplay::displayText(I1, 40, 20, ss.str(), vpColor::red);
}
// Test to reset the tracker
if (reader.getFrameIndex() == reader.getFirstFrameIndex() + 10) {
std::cout << "----------Test reset tracker----------" << std::endl;
if (opt_display) {
}
tracker->resetTracker();
#if defined(VISP_HAVE_XML2) && USE_XML
dynamic_cast<vpMbGenericTracker *>(tracker)->loadConfigFile(mapOfConfigFiles);
#else
// By setting the parameters:
cam.initPersProjWithoutDistortion(547, 542, 338, 234);
mapOfCameraParams["Camera1"] = cam;
mapOfCameraParams["Camera2"] = cam;
mapOfCameraParams["Camera3"] = cam;
vpMe me;
me.setMaskSize(5);
me.setMaskNumber(180);
me.setRange(7);
me.setThreshold(5000);
me.setMu1(0.5);
me.setMu2(0.5);
std::map<std::string, vpMe> mapOfMe;
mapOfMe["Camera1"] = me;
mapOfMe["Camera2"] = me;
mapOfMe["Camera3"] = me;
#if defined(VISP_HAVE_MODULE_KLT) && (defined(VISP_HAVE_OPENCV) && (VISP_HAVE_OPENCV_VERSION >= 0x020100))
klt.setMaxFeatures(10000);
klt.setWindowSize(5);
klt.setQuality(0.01);
klt.setBlockSize(3);
std::map<std::string, vpKltOpencv> mapOfKlt;
mapOfKlt["Camera1"] = klt;
mapOfKlt["Camera2"] = klt;
mapOfKlt["Camera3"] = klt;
dynamic_cast<vpMbGenericTracker *>(tracker)->setKltOpencv(mapOfKlt);
dynamic_cast<vpMbGenericTracker *>(tracker)->setKltMaskBorder(5);
#endif
dynamic_cast<vpMbGenericTracker *>(tracker)->setCameraParameters(mapOfCameraParams);
dynamic_cast<vpMbGenericTracker *>(tracker)->setMovingEdge(mapOfMe);
tracker->setAngleAppear(vpMath::rad(65));
// Specify the clipping to
tracker->setNearClippingDistance(0.01);
tracker->setFarClippingDistance(0.90);
std::map<std::string, unsigned int> mapOfClippingFlags;
dynamic_cast<vpMbGenericTracker *>(tracker)->getClipping(mapOfClippingFlags);
for (std::map<std::string, unsigned int>::iterator it = mapOfClippingFlags.begin();
it != mapOfClippingFlags.end(); ++it) {
it->second = (it->second | vpMbtPolygon::FOV_CLIPPING);
}
dynamic_cast<vpMbGenericTracker *>(tracker)->setClipping(mapOfClippingFlags);
// tracker->setClipping(tracker->getClipping() | vpMbtPolygon::LEFT_CLIPPING
// | vpMbtPolygon::RIGHT_CLIPPING | vpMbtPolygon::UP_CLIPPING |
// vpMbtPolygon::DOWN_CLIPPING); // Equivalent to FOV_CLIPPING
#endif
tracker->loadModel(modelFile);
dynamic_cast<vpMbGenericTracker *>(tracker)->setCameraParameters(mapOfCameraParams);
tracker->setOgreVisibilityTest(useOgre);
tracker->setScanLineVisibilityTest(useScanline);
tracker->setCovarianceComputation(computeCovariance);
tracker->setProjectionErrorComputation(projectionError);
dynamic_cast<vpMbGenericTracker *>(tracker)->initFromPose(mapOfImages, mapOfCameraPoses);
}
// Test to set an initial pose
if (reader.getFrameIndex() == reader.getFirstFrameIndex() + 50) {
c1Moi.buildFrom(0.0439540832, 0.0845870108, 0.5477322481, 2.179498458, 0.8611798108, -0.3491961946);
std::map<std::string, vpHomogeneousMatrix> mapOfSetPoses;
mapOfSetPoses["Camera1"] = c1Moi;
std::cout << "Test set pose" << std::endl;
dynamic_cast<vpMbGenericTracker *>(tracker)->setPose(mapOfImages, mapOfSetPoses);
}
// track the object: stop tracking from frame 40 to 50
if (reader.getFrameIndex() - reader.getFirstFrameIndex() < 40 ||
reader.getFrameIndex() - reader.getFirstFrameIndex() >= 50) {
dynamic_cast<vpMbGenericTracker *>(tracker)->track(mapOfImages);
dynamic_cast<vpMbGenericTracker *>(tracker)->getPose(mapOfCameraPoses);
if (opt_display) {
// display the 3D model
if (reader.getFrameIndex() - reader.getFirstFrameIndex() >= 50) {
std::map<std::string, const vpImage<unsigned char> *> mapOfSubImages;
mapOfSubImages["Camera1"] = &I1;
mapOfSubImages["Camera2"] = &I2;
dynamic_cast<vpMbGenericTracker *>(tracker)->display(mapOfSubImages, mapOfCameraPoses, mapOfCameraParams,
} else {
dynamic_cast<vpMbGenericTracker *>(tracker)->display(mapOfImages, mapOfCameraPoses, mapOfCameraParams,
}
// display the frame
vpDisplay::displayFrame(I1, mapOfCameraPoses["Camera1"], mapOfCameraParams["Camera1"], 0.05);
vpDisplay::displayFrame(I2, mapOfCameraPoses["Camera2"], mapOfCameraParams["Camera2"], 0.05);
vpDisplay::displayFrame(I3, mapOfCameraPoses["Camera3"], mapOfCameraParams["Camera3"], 0.05);
}
}
if (opt_click_allowed) {
vpDisplay::displayText(I1, 10, 10, "Click to quit", vpColor::red);
if (vpDisplay::getClick(I1, button, click)) {
switch (button) {
quit = !click;
break;
click = !click;
break;
default:
break;
}
}
}
if (computeCovariance) {
std::cout << "Covariance matrix: \n" << tracker->getCovarianceMatrix() << std::endl << std::endl;
}
if (projectionError) {
std::cout << "Projection error: " << tracker->getProjectionError() << std::endl << std::endl;
}
}
std::cout << "Reached last frame: " << reader.getFrameIndex() << std::endl;
std::cout << "\nFinal poses, c1Mo:\n"
<< mapOfCameraPoses["Camera1"] << "\nc2Mo:\n"
<< mapOfCameraPoses["Camera2"] << "\nc3Mo:\n"
<< mapOfCameraPoses["Camera3"] << std::endl;
if (opt_click_allowed && !quit) {
}
reader.close();
delete tracker;
tracker = NULL;
#if defined(VISP_HAVE_XML2) && USE_XML
// Cleanup memory allocated by xml library used to parse the xml config
// file in vpMbGenericTracker::loadConfigFile()
#endif
#if defined(VISP_HAVE_COIN3D) && (COIN_MAJOR_VERSION == 2 || COIN_MAJOR_VERSION == 3)
// Cleanup memory allocated by Coin library used to load a vrml model in
// vpMbGenericTracker::loadModel() We clean only if Coin was used.
if (!cao3DModel)
SoDB::finish();
#endif
return EXIT_SUCCESS;
} catch (vpException &e) {
std::cout << "Catch an exception: " << e << std::endl;
return EXIT_FAILURE;
}
}
#else
int main()
{
std::cout << "visp_mbt, visp_gui modules and OpenCV are required to run "
"this example."
<< std::endl;
return 0;
}
#endif