|
#include <iostream> |
|
#include <opencv2/opencv.hpp> |
|
#include <opencv2/wechat_qrcode.hpp> |
|
#include <string> |
|
#include <vector> |
|
|
|
class WeChatQRCode { |
|
public: |
|
WeChatQRCode(const std::string& detect_prototxt, |
|
const std::string& detect_model, |
|
const std::string& sr_prototxt, const std::string& sr_model, |
|
int backend_target_index) |
|
: backend_target_index_(backend_target_index) { |
|
|
|
const std::vector<std::pair<int, int>> backend_target_pairs = { |
|
{cv::dnn::DNN_BACKEND_OPENCV, cv::dnn::DNN_TARGET_CPU}, |
|
{cv::dnn::DNN_BACKEND_CUDA, cv::dnn::DNN_TARGET_CUDA}, |
|
{cv::dnn::DNN_BACKEND_CUDA, cv::dnn::DNN_TARGET_CUDA_FP16}, |
|
{cv::dnn::DNN_BACKEND_TIMVX, cv::dnn::DNN_TARGET_NPU}, |
|
{cv::dnn::DNN_BACKEND_CANN, cv::dnn::DNN_TARGET_NPU}}; |
|
|
|
if (backend_target_index_ < 0 || |
|
backend_target_index_ >= backend_target_pairs.size()) { |
|
throw std::invalid_argument("Invalid backend-target index"); |
|
} |
|
|
|
|
|
detector_ = cv::makePtr<cv::wechat_qrcode::WeChatQRCode>( |
|
detect_prototxt, detect_model, sr_prototxt, sr_model); |
|
} |
|
|
|
std::pair<std::vector<std::string>, std::vector<cv::Mat>> detect( |
|
const cv::Mat& image) { |
|
std::vector<std::string> results; |
|
std::vector<cv::Mat> points; |
|
results = detector_->detectAndDecode(image, points); |
|
return {results, points}; |
|
} |
|
|
|
cv::Mat visualize(const cv::Mat& image, |
|
const std::vector<std::string>& results, |
|
const std::vector<cv::Mat>& points, |
|
cv::Scalar points_color = cv::Scalar(0, 255, 0), |
|
cv::Scalar text_color = cv::Scalar(0, 255, 0), |
|
double fps = -1) const { |
|
cv::Mat output = image.clone(); |
|
|
|
if (fps >= 0) { |
|
cv::putText(output, "FPS: " + std::to_string(fps), cv::Point(0, 15), |
|
cv::FONT_HERSHEY_SIMPLEX, 0.5, text_color); |
|
} |
|
|
|
double fontScale = 0.5; |
|
int fontSize = 1; |
|
|
|
for (size_t i = 0; i < results.size(); ++i) { |
|
const auto& p = points[i]; |
|
|
|
for (int r = 0; r < p.rows; ++r) { |
|
cv::Point point(p.at<float>(r, 0), p.at<float>(r, 1)); |
|
cv::circle(output, point, 10, points_color, -1); |
|
} |
|
|
|
int qrcode_center_x = (p.at<float>(0, 0) + p.at<float>(2, 0)) / 2; |
|
int qrcode_center_y = (p.at<float>(0, 1) + p.at<float>(2, 1)) / 2; |
|
|
|
int baseline = 0; |
|
cv::Size text_size = |
|
cv::getTextSize(results[i], cv::FONT_HERSHEY_DUPLEX, fontScale, |
|
fontSize, &baseline); |
|
|
|
cv::Point text_pos(qrcode_center_x - text_size.width / 2, |
|
qrcode_center_y + text_size.height / 2); |
|
|
|
cv::putText(output, results[i], text_pos, cv::FONT_HERSHEY_DUPLEX, |
|
fontScale, text_color, fontSize); |
|
} |
|
|
|
return output; |
|
} |
|
|
|
private: |
|
int backend_target_index_; |
|
cv::Ptr<cv::wechat_qrcode::WeChatQRCode> detector_; |
|
}; |
|
|
|
int main(int argc, char** argv) { |
|
|
|
cv::CommandLineParser parser( |
|
argc, argv, |
|
"{help h | | Show this help message.}" |
|
"{input i | | Set path to the input image. Omit for using default camera.}" |
|
"{detect_prototxt_path | detect_2021nov.prototxt | Set path to detect.prototxt.}" |
|
"{detect_model_path | detect_2021nov.caffemodel | Set path to detect.caffemodel.}" |
|
"{sr_prototxt_path | sr_2021nov.prototxt | Set path to sr.prototxt.}" |
|
"{sr_model_path | sr_2021nov.caffemodel | Set path to sr.caffemodel.}" |
|
"{backend_target bt | 0 | Choose one of the backend-target pairs to run this demo.}" |
|
"{save s | false | Specify to save file with results.}" |
|
"{vis v | false | Specify to open a new window to show results.}"); |
|
|
|
if (parser.has("help")) { |
|
parser.printMessage(); |
|
return 0; |
|
} |
|
|
|
|
|
std::string detect_prototxt = parser.get<std::string>("detect_prototxt_path"); |
|
std::string detect_model = parser.get<std::string>("detect_model_path"); |
|
std::string sr_prototxt = parser.get<std::string>("sr_prototxt_path"); |
|
std::string sr_model = parser.get<std::string>("sr_model_path"); |
|
int backend_target_index = parser.get<int>("backend_target"); |
|
|
|
|
|
std::string input_path = parser.get<std::string>("input"); |
|
bool save_result = parser.get<bool>("save"); |
|
bool visualize_result = parser.get<bool>("vis"); |
|
|
|
try { |
|
WeChatQRCode qrDetector(detect_prototxt, detect_model, sr_prototxt, |
|
sr_model, backend_target_index); |
|
|
|
if (!input_path.empty()) { |
|
|
|
cv::Mat image = cv::imread(input_path); |
|
if (image.empty()) { |
|
std::cerr << "Could not read the image" << std::endl; |
|
return -1; |
|
} |
|
|
|
std::pair<std::vector<std::string>, std::vector<cv::Mat>> detectionResult = qrDetector.detect(image); |
|
auto& results = detectionResult.first; |
|
auto& points = detectionResult.second; |
|
|
|
for (const auto& result : results) { |
|
std::cout << result << std::endl; |
|
} |
|
|
|
cv::Mat result_image = qrDetector.visualize(image, results, points); |
|
|
|
if (save_result) { |
|
cv::imwrite("result.jpg", result_image); |
|
std::cout << "Results saved to result.jpg" << std::endl; |
|
} |
|
|
|
if (visualize_result) { |
|
cv::imshow(input_path, result_image); |
|
cv::waitKey(0); |
|
} |
|
} else { |
|
|
|
cv::VideoCapture cap(0); |
|
if (!cap.isOpened()) { |
|
std::cerr << "Error opening camera" << std::endl; |
|
return -1; |
|
} |
|
|
|
cv::Mat frame; |
|
cv::TickMeter tm; |
|
|
|
while (true) { |
|
cap >> frame; |
|
if (frame.empty()) { |
|
std::cout << "No frames grabbed" << std::endl; |
|
break; |
|
} |
|
|
|
std::pair<std::vector<std::string>, std::vector<cv::Mat>> detectionResult = qrDetector.detect(frame); |
|
auto& results = detectionResult.first; |
|
auto& points = detectionResult.second; |
|
|
|
tm.start(); |
|
double fps = tm.getFPS(); |
|
tm.stop(); |
|
|
|
cv::Mat result_frame = qrDetector.visualize( |
|
frame, results, points, cv::Scalar(0, 255, 0), |
|
cv::Scalar(0, 255, 0), fps); |
|
cv::imshow("WeChatQRCode Demo", result_frame); |
|
|
|
tm.reset(); |
|
|
|
if (cv::waitKey(1) >= 0) break; |
|
} |
|
} |
|
|
|
} catch (const std::exception& ex) { |
|
std::cerr << "Error: " << ex.what() << std::endl; |
|
return -1; |
|
} |
|
|
|
return 0; |
|
} |
|
|