OpenShot Library | libopenshot 0.5.0
Loading...
Searching...
No Matches
CVTracker.cpp
Go to the documentation of this file.
1
9
10// Copyright (c) 2008-2019 OpenShot Studios, LLC
11//
12// SPDX-License-Identifier: LGPL-3.0-or-later
13
14#include <fstream>
15#include <iomanip>
16#include <iostream>
17#include <cmath>
18#include <algorithm>
19
20#include <google/protobuf/util/time_util.h>
21
22#include "OpenCVUtilities.h"
23#include "CVTracker.h"
24#include "trackerdata.pb.h"
25#include "Exceptions.h"
26
27using namespace openshot;
28using google::protobuf::util::TimeUtil;
29
30// Clamp a rectangle to image bounds and ensure a minimal size
31static inline void clampRect(cv::Rect2d &r, int width, int height)
32{
33 r.x = std::clamp(r.x, 0.0, double(width - 1));
34 r.y = std::clamp(r.y, 0.0, double(height - 1));
35 r.width = std::clamp(r.width, 1.0, double(width - r.x));
36 r.height = std::clamp(r.height, 1.0, double(height - r.y));
37}
38
39// Constructor
40CVTracker::CVTracker(std::string processInfoJson, ProcessingController &processingController)
41: processingController(&processingController), json_interval(false){
42 SetJson(processInfoJson);
43 start = 1;
44 end = 1;
45 lostCount = 0;
46}
47
48// Set desirable tracker method
49cv::Ptr<OPENCV_TRACKER_TYPE> CVTracker::selectTracker(std::string trackerType){
50
51 if (trackerType == "BOOSTING")
52 return OPENCV_TRACKER_NS::TrackerBoosting::create();
53 if (trackerType == "MIL")
54 return OPENCV_TRACKER_NS::TrackerMIL::create();
55 if (trackerType == "KCF")
56 return OPENCV_TRACKER_NS::TrackerKCF::create();
57 if (trackerType == "TLD")
58 return OPENCV_TRACKER_NS::TrackerTLD::create();
59 if (trackerType == "MEDIANFLOW")
60 return OPENCV_TRACKER_NS::TrackerMedianFlow::create();
61 if (trackerType == "MOSSE")
62 return OPENCV_TRACKER_NS::TrackerMOSSE::create();
63 if (trackerType == "CSRT")
64 return OPENCV_TRACKER_NS::TrackerCSRT::create();
65
66 return nullptr;
67}
68
69// Track object in the whole clip or in a given interval
71 size_t _start,
72 size_t _end,
73 bool process_interval)
74{
75 video.Open();
76 if (!json_interval) {
77 start = _start; end = _end;
78 if (!process_interval || end <= 1 || end - start == 0) {
79 start = int(video.Start() * video.Reader()->info.fps.ToFloat()) + 1;
80 end = int(video.End() * video.Reader()->info.fps.ToFloat()) + 1;
81 }
82 } else {
83 start = int(start + video.Start() * video.Reader()->info.fps.ToFloat()) + 1;
84 end = int(video.End() * video.Reader()->info.fps.ToFloat()) + 1;
85 }
86 if (error) return;
87 processingController->SetError(false, "");
88
89 bool trackerInit = false;
90 lostCount = 0; // reset lost counter once at the start
91
92 for (size_t frame = start; frame <= end; ++frame) {
93 if (processingController->ShouldStop()) return;
94
95 auto f = video.GetFrame(frame);
96 cv::Mat img = f->GetImageCV();
97
98 if (frame == start) {
99 bbox = cv::Rect2d(
100 int(bbox.x * img.cols),
101 int(bbox.y * img.rows),
102 int(bbox.width * img.cols),
103 int(bbox.height * img.rows)
104 );
105 }
106
107 if (!trackerInit) {
108 initTracker(img, frame);
109 trackerInit = true;
110 lostCount = 0;
111 }
112 else {
113 // trackFrame now manages lostCount and will re-init internally
114 trackFrame(img, frame);
115
116 // record whatever bbox we have now
117 FrameData fd = GetTrackedData(frame);
118 }
119
120 processingController->SetProgress(
121 uint(100 * (frame - start) / (end - start))
122 );
123 }
124}
125
126// Initialize the tracker
127bool CVTracker::initTracker(cv::Mat &frame, size_t frameId)
128{
129 // Create new tracker object
130 tracker = selectTracker(trackerType);
131
132 // Correct negative width/height
133 if (bbox.width < 0) {
134 bbox.x -= bbox.width;
135 bbox.width = -bbox.width;
136 }
137 if (bbox.height < 0) {
138 bbox.y -= bbox.height;
139 bbox.height = -bbox.height;
140 }
141
142 // Clamp to frame bounds
143 clampRect(bbox, frame.cols, frame.rows);
144
145 // Initialize tracker
146 tracker->init(frame, bbox);
147
148 float fw = float(frame.cols), fh = float(frame.rows);
149
150 // record original pixel size
151 origWidth = bbox.width;
152 origHeight = bbox.height;
153
154 // initialize sub-pixel smoother at true center
155 smoothC_x = bbox.x + bbox.width * 0.5;
156 smoothC_y = bbox.y + bbox.height * 0.5;
157
158 // Add new frame data
159 trackedDataById[frameId] = FrameData(
160 frameId, 0,
161 bbox.x / fw,
162 bbox.y / fh,
163 (bbox.x + bbox.width) / fw,
164 (bbox.y + bbox.height) / fh
165 );
166
167 return true;
168}
169
170// Update the object tracker according to frame
171// returns true if KLT succeeded, false otherwise
172bool CVTracker::trackFrame(cv::Mat &frame, size_t frameId)
173{
174 const int W = frame.cols, H = frame.rows;
175 const auto& prev = trackedDataById[frameId - 1];
176
177 // Reconstruct last-known box in pixel coords
178 cv::Rect2d lastBox(
179 prev.x1 * W, prev.y1 * H,
180 (prev.x2 - prev.x1) * W,
181 (prev.y2 - prev.y1) * H
182 );
183
184 // Convert to grayscale
185 cv::Mat gray;
186 cv::cvtColor(frame, gray, cv::COLOR_BGR2GRAY);
187
188 cv::Rect2d cand;
189 bool didKLT = false;
190
191 // Try KLT-based drift
192 if (!prevGray.empty() && !prevPts.empty()) {
193 std::vector<cv::Point2f> currPts;
194 std::vector<uchar> status;
195 std::vector<float> err;
196 cv::calcOpticalFlowPyrLK(
197 prevGray, gray,
198 prevPts, currPts,
199 status, err,
200 cv::Size(21,21), 3,
201 cv::TermCriteria{cv::TermCriteria::COUNT|cv::TermCriteria::EPS,30,0.01},
202 cv::OPTFLOW_LK_GET_MIN_EIGENVALS, 1e-4
203 );
204
205 // collect per-point displacements
206 std::vector<double> dx, dy;
207 for (size_t i = 0; i < status.size(); ++i) {
208 if (status[i] && err[i] < 12.0) {
209 dx.push_back(currPts[i].x - prevPts[i].x);
210 dy.push_back(currPts[i].y - prevPts[i].y);
211 }
212 }
213
214 if ((int)dx.size() >= minKltPts) {
215 auto median = [&](auto &v){
216 std::nth_element(v.begin(), v.begin()+v.size()/2, v.end());
217 return v[v.size()/2];
218 };
219 double mdx = median(dx), mdy = median(dy);
220
221 cand = lastBox;
222 cand.x += mdx;
223 cand.y += mdy;
224 cand.width = origWidth;
225 cand.height = origHeight;
226
227 lostCount = 0;
228 didKLT = true;
229 }
230 }
231
232 // Fallback to whole-frame flow if KLT failed
233 if (!didKLT) {
234 ++lostCount;
235 cand = lastBox;
236 if (!fullPrevGray.empty()) {
237 cv::Mat flow;
238 cv::calcOpticalFlowFarneback(
239 fullPrevGray, gray, flow,
240 0.5,3,15,3,5,1.2,0
241 );
242 cv::Scalar avg = cv::mean(flow);
243 cand.x += avg[0];
244 cand.y += avg[1];
245 }
246 cand.width = origWidth;
247 cand.height = origHeight;
248
249 if (lostCount >= 10) {
250 initTracker(frame, frameId);
251 cand = bbox;
252 lostCount = 0;
253 }
254 }
255
256 // Dead-zone sub-pixel smoothing
257 {
258 constexpr double JITTER_THRESH = 1.0;
259 double measCx = cand.x + cand.width * 0.5;
260 double measCy = cand.y + cand.height * 0.5;
261 double dx = measCx - smoothC_x;
262 double dy = measCy - smoothC_y;
263
264 if (std::abs(dx) > JITTER_THRESH || std::abs(dy) > JITTER_THRESH) {
265 smoothC_x = measCx;
266 smoothC_y = measCy;
267 }
268
269 cand.x = smoothC_x - cand.width * 0.5;
270 cand.y = smoothC_y - cand.height * 0.5;
271 }
272
273
274 // Candidate box may now lie outside frame; ROI for KLT is clamped below
275 // Re-seed KLT features
276 {
277 // Clamp ROI to frame bounds and avoid negative width/height
278 int roiX = int(std::clamp(cand.x, 0.0, double(W - 1)));
279 int roiY = int(std::clamp(cand.y, 0.0, double(H - 1)));
280 int roiW = int(std::min(cand.width, double(W - roiX)));
281 int roiH = int(std::min(cand.height, double(H - roiY)));
282 roiW = std::max(0, roiW);
283 roiH = std::max(0, roiH);
284
285 if (roiW > 0 && roiH > 0) {
286 cv::Rect roi(roiX, roiY, roiW, roiH);
287 cv::goodFeaturesToTrack(
288 gray(roi), prevPts,
289 kltMaxCorners, kltQualityLevel,
290 kltMinDist, cv::Mat(), kltBlockSize
291 );
292 for (auto &pt : prevPts)
293 pt += cv::Point2f(float(roi.x), float(roi.y));
294 } else {
295 prevPts.clear();
296 }
297 }
298
299 // Commit state
300 fullPrevGray = gray.clone();
301 prevGray = gray.clone();
302 bbox = cand;
303 float fw = float(W), fh = float(H);
304 trackedDataById[frameId] = FrameData(
305 frameId, 0,
306 cand.x / fw,
307 cand.y / fh,
308 (cand.x + cand.width) / fw,
309 (cand.y + cand.height) / fh
310 );
311
312 return didKLT;
313}
314
316 using std::ios;
317
318 // Create tracker message
319 pb_tracker::Tracker trackerMessage;
320
321 // Iterate over all frames data and save in protobuf message
322 for(std::map<size_t,FrameData>::iterator it=trackedDataById.begin(); it!=trackedDataById.end(); ++it){
323 FrameData fData = it->second;
324 pb_tracker::Frame* pbFrameData;
325 AddFrameDataToProto(trackerMessage.add_frame(), fData);
326 }
327
328 // Add timestamp
329 *trackerMessage.mutable_last_updated() = TimeUtil::SecondsToTimestamp(time(NULL));
330
331 {
332 // Write the new message to disk.
333 std::fstream output(protobuf_data_path, ios::out | ios::trunc | ios::binary);
334 if (!trackerMessage.SerializeToOstream(&output)) {
335 std::cerr << "Failed to write protobuf message." << std::endl;
336 return false;
337 }
338 }
339
340 // Delete all global objects allocated by libprotobuf.
341 google::protobuf::ShutdownProtobufLibrary();
342
343 return true;
344
345}
346
347// Add frame tracked data into protobuf message.
348void CVTracker::AddFrameDataToProto(pb_tracker::Frame* pbFrameData, FrameData& fData) {
349
350 // Save frame number and rotation
351 pbFrameData->set_id(fData.frame_id);
352 pbFrameData->set_rotation(0);
353
354 pb_tracker::Frame::Box* box = pbFrameData->mutable_bounding_box();
355 // Save bounding box data
356 box->set_x1(fData.x1);
357 box->set_y1(fData.y1);
358 box->set_x2(fData.x2);
359 box->set_y2(fData.y2);
360}
361
362// Get tracker info for the desired frame
364
365 // Check if the tracker info for the requested frame exists
366 if ( trackedDataById.find(frameId) == trackedDataById.end() ) {
367
368 return FrameData();
369 } else {
370
371 return trackedDataById[frameId];
372 }
373
374}
375
376// Load JSON string into this object
377void CVTracker::SetJson(const std::string value) {
378 // Parse JSON string into JSON objects
379 try
380 {
381 const Json::Value root = openshot::stringToJson(value);
382 // Set all values that match
383
384 SetJsonValue(root);
385 }
386 catch (const std::exception& e)
387 {
388 // Error parsing JSON (or missing keys)
389 throw openshot::InvalidJSON("JSON is invalid (missing keys or invalid data types)");
390 }
391}
392
393// Load Json::Value into this object
394void CVTracker::SetJsonValue(const Json::Value root) {
395
396 // Set data from Json (if key is found)
397 if (!root["protobuf_data_path"].isNull()){
398 protobuf_data_path = (root["protobuf_data_path"].asString());
399 }
400 if (!root["tracker-type"].isNull()){
401 trackerType = (root["tracker-type"].asString());
402 }
403
404 if (!root["region"].isNull()){
405 double x = root["region"]["normalized_x"].asDouble();
406 double y = root["region"]["normalized_y"].asDouble();
407 double w = root["region"]["normalized_width"].asDouble();
408 double h = root["region"]["normalized_height"].asDouble();
409 cv::Rect2d prev_bbox(x,y,w,h);
410 bbox = prev_bbox;
411
412 if (!root["region"]["first-frame"].isNull()){
413 start = root["region"]["first-frame"].asInt64();
414 json_interval = true;
415 }
416 else{
417 processingController->SetError(true, "No first-frame");
418 error = true;
419 }
420
421 }
422 else{
423 processingController->SetError(true, "No initial bounding box selected");
424 error = true;
425 }
426
427}
428
429/*
430||||||||||||||||||||||||||||||||||||||||||||||||||
431 ONLY FOR MAKE TEST
432||||||||||||||||||||||||||||||||||||||||||||||||||
433*/
434
435// Load protobuf data file
437 using std::ios;
438
439 // Create tracker message
440 pb_tracker::Tracker trackerMessage;
441
442 {
443 // Read the existing tracker message.
444 std::fstream input(protobuf_data_path, ios::in | ios::binary);
445 if (!trackerMessage.ParseFromIstream(&input)) {
446 std::cerr << "Failed to parse protobuf message." << std::endl;
447 return false;
448 }
449 }
450
451 // Make sure the trackedData is empty
452 trackedDataById.clear();
453
454 // Iterate over all frames of the saved message
455 for (size_t i = 0; i < trackerMessage.frame_size(); i++) {
456 const pb_tracker::Frame& pbFrameData = trackerMessage.frame(i);
457
458 // Load frame and rotation data
459 size_t id = pbFrameData.id();
460 float rotation = pbFrameData.rotation();
461
462 // Load bounding box data
463 const pb_tracker::Frame::Box& box = pbFrameData.bounding_box();
464 float x1 = box.x1();
465 float y1 = box.y1();
466 float x2 = box.x2();
467 float y2 = box.y2();
468
469 // Assign data to tracker map
470 trackedDataById[id] = FrameData(id, rotation, x1, y1, x2, y2);
471 }
472
473 // Delete all global objects allocated by libprotobuf.
474 google::protobuf::ShutdownProtobufLibrary();
475
476 return true;
477}
Track an object selected by the user.
Header file for all Exception classes.
Header file for OpenCVUtilities (set some common macros).
void trackClip(openshot::Clip &video, size_t _start=0, size_t _end=0, bool process_interval=false)
Definition CVTracker.cpp:70
CVTracker(std::string processInfoJson, ProcessingController &processingController)
Definition CVTracker.cpp:40
bool SaveTrackedData()
Save protobuf file.
void SetJsonValue(const Json::Value root)
Load Json::Value into this object.
void SetJson(const std::string value)
Load JSON string into this object.
bool trackFrame(cv::Mat &frame, size_t frameId)
void AddFrameDataToProto(pb_tracker::Frame *pbFrameData, FrameData &fData)
Add frame tracked data into protobuf message.
FrameData GetTrackedData(size_t frameId)
Get tracked data for a given frame.
cv::Ptr< OPENCV_TRACKER_TYPE > selectTracker(std::string trackerType)
Definition CVTracker.cpp:49
bool initTracker(cv::Mat &frame, size_t frameId)
float Start() const
Get start position (in seconds) of clip (trim start of video).
Definition ClipBase.h:88
This class represents a clip (used to arrange readers on the timeline).
Definition Clip.h:89
void Open() override
Open the internal reader.
Definition Clip.cpp:382
float End() const override
Get end position (in seconds) of clip (trim end of video), which can be affected by the time curve.
Definition Clip.cpp:418
std::shared_ptr< openshot::Frame > GetFrame(int64_t clip_frame_number) override
Get an openshot::Frame object for a specific frame number of this clip. The image size and number of ...
Definition Clip.cpp:453
void Reader(openshot::ReaderBase *new_reader)
Set the current reader.
Definition Clip.cpp:336
Exception for invalid JSON.
Definition Exceptions.h:218
This namespace is the default namespace for all code in the openshot library.
Definition Compressor.h:29
const Json::Value stringToJson(const std::string value)
Definition Json.cpp:16