forked from stephanecharette/DarkHelp
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathDarkHelpThreads.hpp
309 lines (268 loc) · 12.8 KB
/
DarkHelpThreads.hpp
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
/* DarkHelpOnThreads - C++ threads helper class for DarkHelp and Darknet.
* Copyright 2024 Stephane Charette <[email protected]>
* MIT license applies. See "license.txt" for details.
*/
#include "DarkHelp.hpp"
#include <atomic>
#include <condition_variable>
#include <filesystem>
#include <mutex>
#include <thread>
/** @file
* %DarkHelp's class to load multiple identical neural networks.
*/
namespace DarkHelp
{
/** This class allows you to easily run multiple identical copies of a neural network to process many files at once.
*
* If you have enough vram, or if you're running the CPU-only version of Darknet on a device with many cores, then you
* can use this class to load multiple copies of a neural network. Each copy will run on a different thread, and can
* process images independently of each other. This works especially well when the network is loaded once and re-used
* many times to process a large set of images.
*
* For example, on a RTX 3050 with 4 GiB of vram, a YOLOv4-tiny network measuring 768x576 takes 29.4 seconds to process
* 1540 image files.
*
* With this class, the same set of 1540 images only takes 8.8 seconds to process using 13 identical instances of the
* network loaded on the GPU.
*
* (Each instance of the network consumes 289 MiB of vram, which is why 13 copies can be loaded at once on a GPU with
* 4 GiB of vram.)
*
* Note this header file is not included by @p DarkHelp.hpp. To use this functionality you'll need to explicitely
* include this header file.
*
* @since 2024-03-26
*/
class DHThreads final
{
public:
/** The key is the input image filename, while the value contains a copy of the prediction results.
*
* @see @ref DarkHelp::PredictionResults (a vector of results)
* @see @ref DarkHelp::PredictionResult (an individual result)
*
* @since 2024-03-26
*/
using ResultsMap = std::map<std::string, DarkHelp::PredictionResults>;
/** Constructor. No worker threads are started with this constructor. You'll need to manually call @ref init().
*
* @since 2024-03-26
*/
DHThreads();
/** Constructor.
*
* @param [in] c The %DarkHelp configuration that each thread should use to load the neural network.
* @param [in] workers The number of worker threads to create. Each of these threads will also load a copy of the
* neural network, so the number you use here should depend on the amount of vram available and the neural network
* dimensions.
* @param [in] output_directory The directory where the output files will be saved when @ref annotate_output_images
* is enabled. If the directory does not exist, then it will be created. If it already exists, then it is left
* as-is, meaning existing files will remain.
*
* The @p %DHThreads constructor will automatically call @ref init() to ensure all the threads and the neural networks
* are running.
*
* Call @ref add_image() or @ref add_images() to get the threads to start processing images. Then call
* @ref wait_for_results() or @ref get_results().
*
* @since 2024-03-26
*/
DHThreads(const DarkHelp::Config & c, const size_t workers, const std::filesystem::path & output_directory = ".");
/** Constructor.
*
* This is similar to the other constructor, but uses the name of the "bundle" file introduced in %DarkHelp v1.8.
* Run the CLI tool @p DarkHelpCombine to create a @p .dh bundle file.
*
* Unlike the other @p DHThreads constructor, this one will wait until all the worker threads have fully loaded the
* neural network before returning to the caller. (This is required due to the use of the obfuscated bundle file.)
*
* @see @ref DarkHelp::combine()
*
* @since 2024-04-16
*/
DHThreads(const std::filesystem::path & filename, const std::string & key, const size_t workers, const std::filesystem::path & output_directory = ".", const DarkHelp::EDriver & driver = DarkHelp::EDriver::kDarknet);
/// Destructor. This will immediately stop all the worker threads and discard any results.
~DHThreads();
/** Mostly for use with the default constructor. Parameters are the same as the other constructor.
*
* @since 2024-03-26
*/
DHThreads & init(const DarkHelp::Config & c, const size_t workers, const std::filesystem::path & output_directory = ".");
/** Similar to the other @ref init() call, but uses the name of the "bundle" file introduced in %DarkHelp v1.8.
*
* @note This will @em not delete the obfuscated bundle file from disk.
*
* @since 2024-04-16
*/
DHThreads & init(const std::filesystem::path & filename, const std::string & key, const size_t workers, const std::filesystem::path & output_directory = ".", const DarkHelp::EDriver & driver = DarkHelp::EDriver::kDarknet);
/** Starts all of the processing threads. This is automatically called by @ref init(), but may also be called
* manually if @ref stop() was called. Calling @p restart() when the threads are already running will cause the
* existing threads to stop, input files to be cleared, and existing results to be reset.
*
* @warning If a "bundle" file was used when loading the neural network, then the @p .cfg, @p .names, and @p .weights
* files do not exist on disk and the call to @p restart() will fail since the neural network cannot be loaded.
*
* @see @ref init()
*
* @since 2024-03-26
*/
DHThreads & restart();
/** Causes the threads to stop processing and exit. All results and input files are cleared. Does not return until
* all threads have joined. Worker threads can be restarted by calling either @ref init() or @ref restart().
* @see @ref purge()
*
* @since 2024-03-26
*/
DHThreads & stop();
/** Reset to zero the image index used by @ref add_image() to generate the image filename.
*
* @since 2024-04-01
*/
DHThreads & reset_image_index();
/** Add a single image to be processed. Unlike @ref add_images(), this does not require the image to be written to
* disk. The @p cv::Mat must be in standard OpenCV @p BGR format, and the lifetime of the image data must extend to
* when the image will be picked up and processed by one of the worker threads executing @ref run().
*
* @return A "virtual" filename will be created and returned to represent the OpenCV image. This filename is
* needed to correctly interpret the results from @ref wait_for_results(). The filename generated contains a
* numerical value which is assigned in increasing sequential order, until one of @ref purge(), @ref restart()
* or @ref reset_image_index() are called.
*
* @note Images added via @ref add_image() will be processed before filenames added via @ref add_images(). This
* is done to ensure that memory is freed up as quickly as possible (filenames barely take any memory).
*
* @see @ref add_images()
* @see @ref reset_image_index()
*
* @since 2024-04-01
*/
std::string add_image(cv::Mat image);
/** Can be used to add a single image, or a subdirectory. If a subdirectory, then recurse looking for all images.
* Call this as many times as necessary until all images have been added. Image processing by the worker threads
* will start immediately. Additional images can be added at any time, even while the worker threads have already
* started processing the first set of images.
*
* @note Images added via @ref add_image() will be processed before filenames added via @ref add_images(). This
* is done to ensure that memory is freed up as quickly as possible.
*
* @see @ref add_image()
*
* @since 2024-03-26
*/
DHThreads & add_images(const std::filesystem::path & dir);
/** Removes all input files, waits for all worker threads to finish processing, clears out any results, and resets
* the image index (similar to @ref reset_image_index()).
*
* Unlike the call to @ref stop(), calling @ref purge() does not terminate the worker threads or unload the neural
* networks, so you can immediately call @ref add_image() or @ref add_images() without needing to @ref restart().
*
* @since 2024-03-26
*/
DHThreads & purge();
/** Get the number of images and files that have not yet been processed. This is the number of unprocessed images
* and files combined, plus the images which are in the middle of being processed.
*
* @since 2024-03-26
*/
size_t files_remaining() const
{
return input_images.size() + input_files.size() + files_processing;
}
/** Get the number of worker threads which have loaded a copy of the neural network.
*
* @since 2024-03-26
*/
size_t networks_loaded() const
{
return threads_ready;
}
/** Returns when there are zero input files remaining and all worker threads have finished processing images.
* Note this will clear out the results since it internally calls @ref get_results().
*
* The difference between @ref wait_for_results() and @ref get_results() is that @p wait_for_results() will
* wait until @em all the results are available, while @p get_results() will immediately return with whatever
* results are available at this point in time.
*
* @since 2024-03-26
*/
ResultsMap wait_for_results();
/** Gain access to the neural network for the given worker thread. This will return @p nullptr if the given worker
* thread has not loaded a neural network.
*
* @since 2024-03-26
*/
DarkHelp::NN * get_nn(const size_t idx);
/** Get all of the available prediction results. You may call this as often you like to get intermediate results
* (if any are available), or you can indirectly call this via @ref wait_for_results().
*
* @note Once the results are read and returned, the structure in which the results are stored internally is cleared.
*
* The difference between @ref wait_for_results() and @ref get_results() is that @p wait_for_results() will
* wait until @em all the results are available, while @p get_results() will immediately return with whatever
* results are available, even if the results are empty.
*
* @since 2024-03-26
*/
ResultsMap get_results();
/** A copy of the configuration to use when instantiating each of the @ref DarkHelp::NN objects. This is only
* referenced by @ref restart(). Meaning if you change @p cfg after the @ref DHThreads() consructor or @ref init(),
* you'll need to call @ref stop() or @ref restart().
*
* @since 2024-03-26
*/
DarkHelp::Config cfg;
/** Determines whether the input images are deleted once they are processed. Default value is @p false, meaning
* files are not deleted.
*
* @since 2024-03-26
*/
std::atomic<bool> detele_input_file_after_processing;
/** Determines if annotated output images are created. Default value is @p false.
*
* @since 2024-03-26
*/
std::atomic<bool> annotate_output_images;
private:
/// The method that each worker thread runs to process images. @see @ref restart()
void run(const size_t id);
/// If the threads need to stop, set this variable to @p true. @see @ref stop()
std::atomic<bool> stop_requested;
/// The number of threads to start.
size_t worker_threads_to_start;
/// The directory where the output will be saved.
std::filesystem::path output_dir;
/// The threads that were started. @see @ref worker_threads_to_start
std::vector<std::thread> threads;
/// Address to the worker thread neural networks. @see @ref get_nn()
std::vector<DarkHelp::NN *> networks;
/// @{ Used to signal the worker threads when more work becomes available.
std::condition_variable trigger;
std::mutex trigger_lock;
/// @}
/** Used to keep track of all the input @em files remaining to be processed.
* @see @ref add_image()
* @see @ref add_images()
* @see @ref input_image_and_file_lock
*/
std::deque<std::string> input_files;
/** Used to keep track of all the input @em images remaining to be processed.
* @see @ref add_image()
* @see @ref add_images()
* @see @ref input_image_and_file_lock
*/
std::map<std::string, cv::Mat> input_images;
/// Used by @ref add_image() to generate an image filename.
std::atomic<size_t> input_image_index;
/// Lock used to protect access to @em both @ref input_files and @ref input_images.
std::mutex input_image_and_file_lock;
/// @{ The prediction results for all the image file which have been processed. @see @ref get_results()
ResultsMap all_results;
std::mutex results_lock;
/// @}
/// Track the number of worker threads which have loaded the neural network.
std::atomic<size_t> threads_ready;
/// The number of worker threads which are currently processing an image.
std::atomic<size_t> files_processing;
};
}