diff --git a/packages/dartcv/lib/src/imgproc/imgproc_async.dart b/packages/dartcv/lib/src/imgproc/imgproc_async.dart index c43d2aef..064986d8 100644 --- a/packages/dartcv/lib/src/imgproc/imgproc_async.dart +++ b/packages/dartcv/lib/src/imgproc/imgproc_async.dart @@ -61,8 +61,12 @@ Future arcLengthAsync(VecPoint curve, bool closed) async { /// /// For further details, please see: /// https:///docs.opencv.org/master/d3/dc0/group__imgproc__shape.html#ga014b28e56cb8854c0de4a211cb2be656 -Future convexHullAsync(VecPoint points, - {Mat? hull, bool clockwise = false, bool returnPoints = true}) async { +Future convexHullAsync( + VecPoint points, { + Mat? hull, + bool clockwise = false, + bool returnPoints = true, +}) async { hull ??= Mat.empty(); return cvRunAsync0( (callback) => cimgproc.cv_convexHull(points.ref, hull!.ref, clockwise, returnPoints, callback), @@ -218,8 +222,13 @@ Future<(bool, Point, Point)> clipLineAsync(Rect imgRect, Point pt1, Point pt2) a /// /// For further details, please see: /// https:///docs.opencv.org/master/d4/d86/group__imgproc__filter.html#ga9d7064d478c95d60003cf839430737ed -Future bilateralFilterAsync(Mat src, int diameter, double sigmaColor, double sigmaSpace, - {Mat? dst}) async { +Future bilateralFilterAsync( + Mat src, + int diameter, + double sigmaColor, + double sigmaSpace, { + Mat? dst, +}) async { dst ??= Mat.empty(); return cvRunAsync0( (callback) => cimgproc.cv_bilateralFilter(src.ref, dst!.ref, diameter, sigmaColor, sigmaSpace, callback), @@ -1720,8 +1729,11 @@ Future getPerspectiveTransformAsync(VecPoint src, VecPoint dst, [int solveM /// /// For further details, please see: /// https:///docs.opencv.org/master/da/d54/group__imgproc__transform.html#ga8c1ae0e3589a9d77fffc962c49b22043 -Future getPerspectiveTransform2fAsync(VecPoint2f src, VecPoint2f dst, - [int solveMethod = DECOMP_LU]) async { +Future getPerspectiveTransform2fAsync( + VecPoint2f src, + VecPoint2f dst, [ + int solveMethod = DECOMP_LU, +]) async { final mat = Mat.empty(); return cvRunAsync0( (callback) => cimgproc.cv_getPerspectiveTransform2f(src.ref, dst.ref, mat.ptr, solveMethod, callback), @@ -2075,8 +2087,12 @@ Future accumulateProductAsync( /// /// For further details, please see: /// https:///docs.opencv.org/master/d7/df3/group__imgproc__motion.html#ga4f9552b541187f61f6818e8d2d826bc7 -Future accumulateWeightedAsync(InputArray src, InputOutputArray dst, double alpha, - {InputArray? mask}) async { +Future accumulateWeightedAsync( + InputArray src, + InputOutputArray dst, + double alpha, { + InputArray? mask, +}) async { if (mask == null) { return cvRunAsync0( (callback) => cimgproc.cv_accumulatedWeighted(src.ref, dst.ref, alpha, callback), diff --git a/packages/dartcv/test/dnn/dnn_test.dart b/packages/dartcv/test/dnn/dnn_test.dart index 5536a93c..dc82379a 100644 --- a/packages/dartcv/test/dnn/dnn_test.dart +++ b/packages/dartcv/test/dnn/dnn_test.dart @@ -205,7 +205,7 @@ void main() async { test('cv.Net.fromTFLite', skip: true, () { final model = cv.Net.fromTFLite("test/models/face_landmark.tflite"); - print(model.getUnconnectedOutLayersNames()); + expect(model.getUnconnectedOutLayersNames(), isNotEmpty); checkTflite(model); });