Visual Servoing Platform version 3.7.0
Loading...
Searching...
No Matches
tutorial-detection-object-mbt2.cpp
1
2#include <visp3/core/vpConfig.h>
3#include <visp3/core/vpIoTools.h>
4#include <visp3/gui/vpDisplayFactory.h>
5#include <visp3/io/vpVideoReader.h>
6#include <visp3/mbt/vpMbGenericTracker.h>
7#include <visp3/vision/vpKeyPoint.h>
8
9#ifdef ENABLE_VISP_NAMESPACE
10using namespace VISP_NAMESPACE_NAME;
11#endif
12
13#if defined(VISP_HAVE_OPENCV) && defined(HAVE_OPENCV_IMGPROC) && defined(VISP_HAVE_DISPLAY) && \
14 (((VISP_HAVE_OPENCV_VERSION < 0x050000) && defined(HAVE_OPENCV_CALIB3D) && defined(HAVE_OPENCV_FEATURES2D)) || \
15 ((VISP_HAVE_OPENCV_VERSION >= 0x050000) && defined(HAVE_OPENCV_3D) && defined(HAVE_OPENCV_FEATURES)))
16
17void learnCube(const vpImage<unsigned char> &I, vpMbGenericTracker &tracker, vpKeyPoint &keypoint_learning, int id);
18
19void learnCube(const vpImage<unsigned char> &I, vpMbGenericTracker &tracker, vpKeyPoint &keypoint_learning, int id)
20{
22 std::vector<cv::KeyPoint> trainKeyPoints;
23 double elapsedTime;
24 keypoint_learning.detect(I, trainKeyPoints, elapsedTime);
26
28 std::vector<vpPolygon> polygons;
29 std::vector<std::vector<vpPoint> > roisPt;
30 std::pair<std::vector<vpPolygon>, std::vector<std::vector<vpPoint> > > pair = tracker.getPolygonFaces();
31 polygons = pair.first;
32 roisPt = pair.second;
33
34 std::vector<cv::Point3f> points3f;
36 tracker.getPose(cMo);
38 tracker.getCameraParameters(cam);
39 vpKeyPoint::compute3DForPointsInPolygons(cMo, cam, trainKeyPoints, polygons, roisPt, points3f);
41
43 keypoint_learning.buildReference(I, trainKeyPoints, points3f, true, id);
45
47 for (std::vector<cv::KeyPoint>::const_iterator it = trainKeyPoints.begin(); it != trainKeyPoints.end(); ++it) {
48 vpDisplay::displayCross(I, static_cast<int>(it->pt.y), static_cast<int>(it->pt.x), 4, vpColor::red);
49 }
51}
52#endif
53
54int main(int argc, char **argv)
55{
56#if defined(HAVE_OPENCV_IMGPROC) && defined(VISP_HAVE_OPENCV) && \
57 (((VISP_HAVE_OPENCV_VERSION < 0x050000) && defined(HAVE_OPENCV_FEATURES2D)) || \
58 ((VISP_HAVE_OPENCV_VERSION >= 0x050000) && defined(HAVE_OPENCV_FEATURES)))
59
61#if (VISP_CXX_STANDARD >= VISP_CXX_STANDARD_11)
62 std::shared_ptr<vpDisplay> display, display2;
63#else
64 vpDisplay *display = nullptr;
65 vpDisplay *display2 = nullptr;
66#endif
67 try {
68 std::string videoname = "cube.mp4";
69
70 for (int i = 1; i < argc; i++) {
71 if (std::string(argv[i]) == "--name" && i + 1 < argc)
72 videoname = std::string(argv[++i]);
73 else if (std::string(argv[i]) == "--help" || std::string(argv[i]) == "-h") {
74 std::cout << "\nUsage: " << argv[0]
75 << " [--name <video name>]"
76 << " [--help] [-h]\n" << std::endl;
77 return EXIT_SUCCESS;
78 }
79 }
80 std::string parentname = vpIoTools::getParent(videoname);
81 std::string objectname = vpIoTools::getNameWE(videoname);
82
83 if (!parentname.empty())
84 objectname = parentname + "/" + objectname;
85
86 std::cout << "Video name: " << videoname << std::endl;
87 std::cout << "Tracker requested config files: " << objectname << ".[init,"
88 << "xml,"
89 << "cao or wrl]" << std::endl;
90 std::cout << "Tracker optional config files: " << objectname << ".[ppm]" << std::endl;
91
95
97 bool usexml = false;
98#if defined(VISP_HAVE_PUGIXML)
99 if (vpIoTools::checkFilename(objectname + ".xml")) {
100 tracker.loadConfigFile(objectname + ".xml");
101 tracker.getCameraParameters(cam);
102 usexml = true;
103 }
104#endif
105 if (!usexml) {
106 vpMe me;
107 me.setMaskSize(5);
108 me.setMaskNumber(180);
109 me.setRange(7);
111 me.setThreshold(20);
112 me.setMu1(0.5);
113 me.setMu2(0.5);
114 me.setSampleStep(4);
115 me.setNbTotalSample(250);
116 tracker.setMovingEdge(me);
117 cam.initPersProjWithoutDistortion(547, 542, 339, 235);
118 tracker.setCameraParameters(cam);
119 tracker.setAngleAppear(vpMath::rad(89));
120 tracker.setAngleDisappear(vpMath::rad(89));
121 tracker.setNearClippingDistance(0.01);
122 tracker.setFarClippingDistance(10.0);
123 tracker.setClipping(tracker.getClipping() | vpMbtPolygon::FOV_CLIPPING);
124 }
125
126 tracker.setOgreVisibilityTest(false);
127 if (vpIoTools::checkFilename(objectname + ".cao")) {
128 tracker.loadModel(objectname + ".cao");
129 }
130 else if (vpIoTools::checkFilename(objectname + ".wrl")) {
131 tracker.loadModel(objectname + ".wrl");
132 }
133 tracker.setDisplayFeatures(true);
135
137 vpKeyPoint keypoint_learning("ORB", "ORB", "BruteForce-Hamming");
138#if ((VISP_HAVE_OPENCV_VERSION < 0x050000) && defined(HAVE_OPENCV_FEATURES2D)) || ((VISP_HAVE_OPENCV_VERSION >= 0x050000) && defined(HAVE_OPENCV_FEATURES))
139#if (VISP_HAVE_OPENCV_VERSION < 0x030000)
140 keypoint_learning.setDetectorParameter("ORB", "nLevels", 1);
141#else
142 cv::Ptr<cv::ORB> orb_learning = keypoint_learning.getDetector("ORB").dynamicCast<cv::ORB>();
143 if (orb_learning) {
144 orb_learning->setNLevels(1);
145 }
146#endif
147#endif
149
150 /*
151 * Start the part of the code dedicated to object learning from 3 images
152 */
153 std::string imageName[] = { "cube0001.png", "cube0150.png", "cube0200.png" };
154 vpHomogeneousMatrix initPoseTab[] = {
155 vpHomogeneousMatrix(0.02143385294, 0.1098083886, 0.5127439561, 2.087159614, 1.141775176, -0.4701291124),
156 vpHomogeneousMatrix(0.02651282185, -0.03713587374, 0.6873765919, 2.314744454, 0.3492296488, -0.1226054828),
157 vpHomogeneousMatrix(0.02965448956, -0.07283091786, 0.7253526051, 2.300529617, -0.4286674806, 0.1788761025) };
158 for (int i = 0; i < 3; i++) {
159 vpImageIo::read(I, imageName[i]);
160 if (i == 0) {
161#if (VISP_CXX_STANDARD >= VISP_CXX_STANDARD_11)
162 display = vpDisplayFactory::createDisplay(I, 10, 10);
163#else
164 display = vpDisplayFactory::allocateDisplay(I, 10, 10);
165#endif
166 }
167 std::stringstream title;
168 title << "Learning cube on image: " << imageName[i];
169 vpDisplay::setTitle(I, title.str().c_str());
170
172
174 tracker.setPose(I, initPoseTab[i]);
176
178 tracker.track(I);
180
182 tracker.getPose(cMo);
183 tracker.display(I, cMo, cam, vpColor::red);
185
187 learnCube(I, tracker, keypoint_learning, i);
189
190 vpDisplay::displayText(I, 10, 10, "Learning step: keypoints are detected on visible cube faces", vpColor::red);
191 if (i < 2) {
192 vpDisplay::displayText(I, 30, 10, "Click to continue the learning...", vpColor::red);
193 }
194 else {
195 vpDisplay::displayText(I, 30, 10, "Click to continue with the detection...", vpColor::red);
196 }
197
199 vpDisplay::getClick(I, true);
200 }
201
203 keypoint_learning.saveLearningData("cube_learning_data.bin", true);
205
206 /*
207 * Start the part of the code dedicated to detection and localization
208 */
210 vpKeyPoint keypoint_detection("ORB", "ORB", "BruteForce-Hamming");
211#if ((VISP_HAVE_OPENCV_VERSION < 0x050000) && defined(HAVE_OPENCV_FEATURES2D)) || ((VISP_HAVE_OPENCV_VERSION >= 0x050000) && defined(HAVE_OPENCV_FEATURES))
212#if (VISP_HAVE_OPENCV_VERSION < 0x030000)
213 keypoint_detection.setDetectorParameter("ORB", "nLevels", 1);
214#else
215 cv::Ptr<cv::ORB> orb_detector = keypoint_detection.getDetector("ORB").dynamicCast<cv::ORB>();
216 orb_detector = keypoint_detection.getDetector("ORB").dynamicCast<cv::ORB>();
217 if (orb_detector) {
218 orb_detector->setNLevels(1);
219 }
220#endif
221#endif
223
225 keypoint_detection.loadLearningData("cube_learning_data.bin", true);
227
229 vpImage<unsigned char> IMatching;
230 keypoint_detection.createImageMatching(I, IMatching);
232
234 g.setFileName(videoname);
235 g.open(I);
236
237#if (VISP_CXX_STANDARD >= VISP_CXX_STANDARD_11)
238 display2 = vpDisplayFactory::createDisplay(IMatching, 50, 50, "Display matching between learned and current images");
239#else
240 display2 = vpDisplayFactory::allocateDisplay(IMatching, 50, 50, "Display matching between learned and current images");
241#endif
242 vpDisplay::setTitle(I, "Cube detection and localization");
243
244 double error;
245 bool click_done = false;
246
247 while (!g.end()) {
248 g.acquire(I);
250
252 keypoint_detection.insertImageMatching(I, IMatching);
254
255 vpDisplay::display(IMatching);
256 vpDisplay::displayText(I, 10, 10, "Detection and localization in process...", vpColor::red);
257
258 double elapsedTime;
260 if (keypoint_detection.matchPoint(I, cam, cMo, error, elapsedTime)) {
262
264 tracker.setPose(I, cMo);
266
268 tracker.display(I, cMo, cam, vpColor::red, 2);
269 vpDisplay::displayFrame(I, cMo, cam, 0.05, vpColor::none, 3);
271
272 keypoint_detection.displayMatching(I, IMatching);
273
275 std::vector<vpImagePoint> ransacInliers = keypoint_detection.getRansacInliers();
276 std::vector<vpImagePoint> ransacOutliers = keypoint_detection.getRansacOutliers();
278
280 for (std::vector<vpImagePoint>::const_iterator it = ransacInliers.begin(); it != ransacInliers.end(); ++it) {
282 vpImagePoint imPt(*it);
283 imPt.set_u(imPt.get_u() + I.getWidth());
284 imPt.set_v(imPt.get_v() + I.getHeight());
285 vpDisplay::displayCircle(IMatching, imPt, 4, vpColor::green);
286 }
288
290 for (std::vector<vpImagePoint>::const_iterator it = ransacOutliers.begin(); it != ransacOutliers.end(); ++it) {
292 vpImagePoint imPt(*it);
293 imPt.set_u(imPt.get_u() + I.getWidth());
294 imPt.set_v(imPt.get_v() + I.getHeight());
295 vpDisplay::displayCircle(IMatching, imPt, 4, vpColor::red);
296 }
298
300 keypoint_detection.displayMatching(I, IMatching);
302
305 cam2.initPersProjWithoutDistortion(cam.get_px(), cam.get_py(), cam.get_u0() + I.getWidth(),
306 cam.get_v0() + I.getHeight());
307 tracker.setCameraParameters(cam2);
308 tracker.setPose(IMatching, cMo);
309 tracker.display(IMatching, cMo, cam2, vpColor::red, 2);
310 vpDisplay::displayFrame(IMatching, cMo, cam2, 0.05, vpColor::none, 3);
312 }
313
315 vpDisplay::displayText(IMatching, 30, 10, "A click to exit.", vpColor::red);
316 vpDisplay::flush(IMatching);
317 if (vpDisplay::getClick(I, false)) {
318 click_done = true;
319 break;
320 }
321 if (vpDisplay::getClick(IMatching, false)) {
322 click_done = true;
323 break;
324 }
325 }
326
327 if (!click_done)
328 vpDisplay::getClick(IMatching);
329 }
330 catch (const vpException &e) {
331 std::cout << "Catch an exception: " << e << std::endl;
332 }
333#if (VISP_CXX_STANDARD < VISP_CXX_STANDARD_11)
334 if (display != nullptr) {
335 delete display;
336 }
337
338 if (display2 != nullptr) {
339 delete display2;
340 }
341#endif
342#else
343 (void)argc;
344 (void)argv;
345 std::cout << "Install OpenCV and rebuild ViSP to use this example." << std::endl;
346#endif
347
348 return EXIT_SUCCESS;
349}
Generic class defining intrinsic camera parameters.
void initPersProjWithoutDistortion(double px, double py, double u0, double v0)
static const vpColor red
Definition vpColor.h:198
static const vpColor none
Definition vpColor.h:210
static const vpColor green
Definition vpColor.h:201
Class that defines generic functionalities for display.
Definition vpDisplay.h:171
static bool getClick(const vpImage< unsigned char > &I, bool blocking=true)
static void displayCircle(const vpImage< unsigned char > &I, const vpImageCircle &circle, const vpColor &color, bool fill=false, unsigned int thickness=1)
static void display(const vpImage< unsigned char > &I)
static void displayFrame(const vpImage< unsigned char > &I, const vpHomogeneousMatrix &cMo, const vpCameraParameters &cam, double size, const vpColor &color=vpColor::none, unsigned int thickness=1, const vpImagePoint &offset=vpImagePoint(0, 0), const std::string &frameName="", const vpColor &textColor=vpColor::black, const vpImagePoint &textOffset=vpImagePoint(15, 15))
static void displayCross(const vpImage< unsigned char > &I, const vpImagePoint &ip, unsigned int size, const vpColor &color, unsigned int thickness=1)
static void setTitle(const vpImage< unsigned char > &I, const std::string &windowtitle)
static void flush(const vpImage< unsigned char > &I)
static void displayText(const vpImage< unsigned char > &I, const vpImagePoint &ip, const std::string &s, const vpColor &color)
error that can be emitted by ViSP classes.
Definition vpException.h:60
Implementation of an homogeneous matrix and operations on such kind of matrices.
static void read(vpImage< unsigned char > &I, const std::string &filename, int backend=IO_DEFAULT_BACKEND)
Class that defines a 2D point in an image. This class is useful for image processing and stores only ...
Definition of the vpImage class member functions.
Definition vpImage.h:131
static bool checkFilename(const std::string &filename)
static std::string getNameWE(const std::string &pathname)
static std::string getParent(const std::string &pathname)
Class that allows keypoints 2D features detection (and descriptors extraction) and matching thanks to...
Definition vpKeyPoint.h:274
unsigned int buildReference(const vpImage< unsigned char > &I) VP_OVERRIDE
void detect(const vpImage< unsigned char > &I, std::vector< cv::KeyPoint > &keyPoints, const vpRect &rectangle=vpRect())
static void compute3DForPointsInPolygons(const vpHomogeneousMatrix &cMo, const vpCameraParameters &cam, std::vector< cv::KeyPoint > &candidates, const std::vector< vpPolygon > &polygons, const std::vector< std::vector< vpPoint > > &roisPt, std::vector< cv::Point3f > &points, cv::Mat *descriptors=nullptr)
void saveLearningData(const std::string &filename, bool binaryMode=false, bool saveTrainingImages=true)
cv::Ptr< cv::FeatureDetector > getDetector(const vpFeatureDetectorType &type) const
static double rad(double deg)
Definition vpMath.h:129
Real-time 6D object pose tracking using its CAD model.
Definition vpMe.h:143
void setMu1(const double &mu_1)
Definition vpMe.h:408
void setRange(const unsigned int &range)
Definition vpMe.h:438
void setLikelihoodThresholdType(const vpLikelihoodThresholdType likelihood_threshold_type)
Definition vpMe.h:531
void setNbTotalSample(const int &ntotal_sample)
Definition vpMe.h:422
void setMaskNumber(const unsigned int &mask_number)
Definition vpMe.cpp:555
void setThreshold(const double &threshold)
Definition vpMe.h:489
void setSampleStep(const double &sample_step)
Definition vpMe.h:445
void setMaskSize(const unsigned int &mask_size)
Definition vpMe.cpp:563
void setMu2(const double &mu_2)
Definition vpMe.h:415
@ NORMALIZED_THRESHOLD
Definition vpMe.h:154
Class that enables to manipulate easily a video file or a sequence of images. As it inherits from the...
void open(vpImage< vpRGBa > &I) VP_OVERRIDE
void setFileName(const std::string &filename)
void acquire(vpImage< vpRGBa > &I) VP_OVERRIDE
std::shared_ptr< vpDisplay > createDisplay()
Return a smart pointer vpDisplay specialization if a GUI library is available or nullptr otherwise.
vpDisplay * allocateDisplay()
Return a newly allocated vpDisplay specialization if a GUI library is available or nullptr otherwise.