diff --git a/.github/ISSUE_TEMPLATE/bug-or-crash-report.md b/.github/ISSUE_TEMPLATE/bug-or-crash-report.md index 66ac50ca..6e1e643d 100644 --- a/.github/ISSUE_TEMPLATE/bug-or-crash-report.md +++ b/.github/ISSUE_TEMPLATE/bug-or-crash-report.md @@ -7,10 +7,11 @@ assignees: '' --- -**Describe the bug** +### Describe the bug -**To Reproduce** +### To Reproduce + Steps to reproduce the behavior: 1. Go to '...' @@ -18,27 +19,28 @@ Steps to reproduce the behavior: 3. Scroll down to '....' 4. See error -**Expected behavior** +### Expected behavior -**Screenshots and recordings** +### Screenshots and recordings -**Traceback or Crash Report** +### Traceback or Crash Report + If AutoSplit showed an exception traceback, please paste it here: -``` +```py ``` -**Version (please complete the following information):** +### Version (please complete the following information) - OS: [e.g. Windows 10.0.19045] - AutoSplit: [e.g. v2.0.0] -**AutoSplit Profile and Split Images** +### AutoSplit Profile and Split Images -**Additional context** +### Additional context diff --git a/.github/ISSUE_TEMPLATE/feature_request.md b/.github/ISSUE_TEMPLATE/feature_request.md index 546da8b0..7826d5a1 100644 --- a/.github/ISSUE_TEMPLATE/feature_request.md +++ b/.github/ISSUE_TEMPLATE/feature_request.md @@ -7,15 +7,15 @@ assignees: '' --- -**Is your feature request related to a problem? Please describe.** +### Is your feature request related to a problem? Please describe -**Describe the solution you'd like** +### Describe the solution you'd like -**Describe alternatives you've considered** +### Describe alternatives you've considered -**Additional context** +### Additional context diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml index d8c6f3d4..71be36f4 100644 --- a/.pre-commit-config.yaml +++ b/.pre-commit-config.yaml @@ -17,7 +17,7 @@ repos: - id: pretty-format-ini args: [--autofix] - repo: https://github.com/astral-sh/ruff-pre-commit - rev: "v0.1.11" # Must match requirements-dev.txt + rev: "v0.3.2" # Must match requirements-dev.txt hooks: - id: ruff args: [--fix] diff --git a/docs/CONTRIBUTING.md b/docs/CONTRIBUTING.md index d6066670..7f22d1e8 100644 --- a/docs/CONTRIBUTING.md +++ b/docs/CONTRIBUTING.md @@ -8,8 +8,9 @@ Refer to the [build instructions](/docs/build%20instructions.md) if you're inter ## Linting and formatting -The project is setup to automatically configure VSCode witht he proper extensions and settings. Linters and formatters will be run on save. +The project is setup to automatically configure VSCode with the proper extensions and settings. Fixers and formatters will be run on save. If you use a different IDE or for some reason cannot / don't want to use the recommended extensions, you can run `scripts/lint.ps1`. +Project configurations for other IDEs are welcome. If you like to use pre-commit hooks, `.pre-commit-config.yaml` is setup for such uses. @@ -28,6 +29,16 @@ Your Pull Request has to pass all checks ot be accepted. If it is still a work-i Most coding standards will be enforced by automated tooling. As time goes on, project-specific standards and "gotchas" in the frameworks we use will be listed here. +### Keep shipped dependencies and bundle size low + +The bigger the bundle, the longer it takes to boot single-file executables. That is because we need to ship everything and the bootloader basically has to extract it all. +Our main use case is a single-file that is as easy to use as possible for the end user. +Keeping install time, build time and bandwith as low as possible is also a nice-to-have. + +You should also consider whether the work the dependency is doing is simple enough that you could implement it yourself. + +For these reasons, it's important to consider the impacts of adding any new dependency bundled with AutoSplit. + ### Magic numbers Please avoid using magic numbers and prefer constants and enums that have a meaningful name when possible. @@ -38,6 +49,10 @@ For image shape and channels, please use `utils.ImageShape` and `utils.ColorChan To avoid image shape mismatch issues, and to keep code simpler, we standardize the image color format to BGRA. This should always be done early in the pipeline, so whatever functionality takes care of obtaining an image should also ensure its color format. You can do so with `cv2.cvtColor` (ie: `cv2.cvtColor(image, cv2.COLOR_RGBA2BGRA)` or `cv2.cvtColor(image, cv2.COLOR_BGR2BGRA)`). +### Split-specific setting overrides + +Whenever a split image overrides a default global setting, we add a getter that handles the logic of checking for a split-specific override, then falling back to globals. This avoids repeating the fallback logic in multiple places. See `AutoSpitImage.get_*` methods for examples. + ## Testing -None 😦 Please help us create test suites, we lack the time, but we really want (need!) them. +None 😦 Please help us create test suites, we lack the time, but we really want (*need!*) them. diff --git a/pyproject.toml b/pyproject.toml index 0b80a23c..a0b91430 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -2,8 +2,10 @@ [tool.ruff] target-version = "py310" line-length = 120 -select = ["ALL"] preview = true + +[tool.ruff.lint] +select = ["ALL"] # https://docs.astral.sh/ruff/rules/ ignore = [ ### @@ -58,14 +60,15 @@ ignore = [ # Python 3.11, introduced "zero cost" exception handling "PERF203", # try-except-in-loop - ### FIXME/TODO (no warnings in Ruff yet: https://github.com/astral-sh/ruff/issues/1256 & https://github.com/astral-sh/ruff/issues/1774): + ### FIXME/TODO: I'd normally set them as temporarily warnings, but no warnings in Ruff yet: + ### https://github.com/astral-sh/ruff/issues/1256 & https://github.com/astral-sh/ruff/issues/1774): "CPY001", # flake8-copyright "PTH", # flake8-use-pathlib # Ignore until linux support "EXE", # flake8-executable ] -[tool.ruff.per-file-ignores] +[tool.ruff.lint.per-file-ignores] "typings/**/*.pyi" = [ "F811", # Re-exports false positives "F821", # https://github.com/astral-sh/ruff/issues/3011 @@ -73,6 +76,7 @@ ignore = [ "A", # Shadowing builtin names "ICN001", # unconventional-import-alias "N8", # Naming conventions + "PLC2701", # Private name import "PLR0904", # Too many public methods "PLR0913", # Argument count "PLR0917", # Too many positional arguments @@ -81,11 +85,11 @@ ignore = [ ] # https://docs.astral.sh/ruff/settings/#flake8-implicit-str-concat -[tool.ruff.flake8-implicit-str-concat] +[tool.ruff.lint.flake8-implicit-str-concat] allow-multiline = false # https://docs.astral.sh/ruff/settings/#isort -[tool.ruff.isort] +[tool.ruff.lint.isort] combine-as-imports = true split-on-trailing-comma = false # Unlike isort, Ruff only counts relative imports as local-folder by default for know. @@ -107,13 +111,13 @@ known-local-folder = [ ] # https://docs.astral.sh/ruff/settings/#mccabe -[tool.ruff.mccabe] +[tool.ruff.lint.mccabe] # Hard limit, arbitrary to 4 bytes max-complexity = 31 # Arbitrary to 2 bytes, same as SonarLint # max-complexity = 15 -[tool.ruff.pylint] +[tool.ruff.lint.pylint] # Arbitrary to 1 byte, same as SonarLint max-args = 7 # At least same as max-complexity @@ -132,6 +136,7 @@ ignore = [ # Autofixed by Ruff # Check for the "Fix" flag https://docs.astral.sh/ruff/rules/#pycodestyle-e-w "E2", # Whitespace + "E3", # Blank lines "E703", # useless-semicolon "E71", # Statement (comparisons) "E731", # lambda-assignment diff --git a/scripts/compile_resources.ps1 b/scripts/compile_resources.ps1 index 8749068e..e2abdb9a 100644 --- a/scripts/compile_resources.ps1 +++ b/scripts/compile_resources.ps1 @@ -29,7 +29,10 @@ Else { $GITHUB_REPOSITORY = $Env:GITHUB_HEAD_REPOSITORY If (-not $GITHUB_REPOSITORY) { $repo_url = git config --get remote.origin.url - $GITHUB_REPOSITORY = $repo_url.substring(19, $repo_url.length - 19) -replace '\.git', '' + # Validate in case the repo was downloaded rather than cloned + If ($repo_url) { + $GITHUB_REPOSITORY = $repo_url.substring(19, $repo_url.length - 19) -replace '\.git', '' + } } If (-not $GITHUB_REPOSITORY) { $GITHUB_REPOSITORY = 'Toufool/AutoSplit' diff --git a/scripts/install.ps1 b/scripts/install.ps1 index 6072c5a3..9be64cbc 100644 --- a/scripts/install.ps1 +++ b/scripts/install.ps1 @@ -40,6 +40,9 @@ If ($IsLinux) { } # Ensures installation tools are up to date. This also aliases pip to pip3 on MacOS. &"$python" -m pip install wheel pip setuptools --upgrade +# Upgrading QT to 6.6.2 w/o first uninstalling shiboken6 can lead to issues +# https://bugreports.qt.io/browse/PYSIDE-2616?focusedId=777285&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-777285 +&"$python" -m pip uninstall shiboken6 &"$python" -m pip install -r "$PSScriptRoot/requirements$dev.txt" --upgrade # These libraries install extra requirements we don't want # Open suggestion for support in requirements files: https://github.com/pypa/pip/issues/9948 & https://github.com/pypa/pip/pull/10837 @@ -72,11 +75,11 @@ If ($IsLinux) { Set-Content "$libPath/QtWidgets.pyi" } # Uninstall optional dependencies if PyAutoGUI or D3DShot was installed outside this script -# pyscreeze -> pyscreenshot -> mss deps call SetProcessDpiAwareness, used to be installed on Windows +# PyScreeze -> pyscreenshot -> mss deps call SetProcessDpiAwareness, used to be installed on Windows # Pillow, pygetwindow, pymsgbox, pytweening, MouseInfo are picked up by PySide6 # (also --exclude from build script, but more consistent with unfrozen run) &"$python" -m pip uninstall pyscreenshot mss pygetwindow pymsgbox pytweening MouseInfo -y -If ($IsWindows) { &"$python" -m pip uninstall pyscreeze Pillow -y } +If ($IsWindows) { &"$python" -m pip uninstall PyScreeze Pillow -y } # Don't compile resources on the Build CI job as it'll do so in build script If ($dev) { diff --git a/scripts/requirements-dev.txt b/scripts/requirements-dev.txt index ab8a67f1..5e242a22 100644 --- a/scripts/requirements-dev.txt +++ b/scripts/requirements-dev.txt @@ -14,7 +14,7 @@ # Linters & Formatters add-trailing-comma>=3.1.0 # Must match .pre-commit-config.yaml autopep8>=2.0.4 # Must match .pre-commit-config.yaml -ruff>=0.1.11 # New checks # Must match .pre-commit-config.yaml +ruff>=0.3.2 # New checks # Must match .pre-commit-config.yaml # # Types types-D3DShot ; sys_platform == 'win32' @@ -23,5 +23,5 @@ types-psutil types-PyAutoGUI types-pyinstaller types-python-xlib ; sys_platform == 'linux' -types-pywin32>=306.0.0.8 ; sys_platform == 'win32' +types-pywin32>=306.0.0.20240130 ; sys_platform == 'win32' types-toml diff --git a/scripts/requirements.txt b/scripts/requirements.txt index 9391da99..e89669ef 100644 --- a/scripts/requirements.txt +++ b/scripts/requirements.txt @@ -5,7 +5,7 @@ # Dependencies: git+https://github.com/boppreh/keyboard.git#egg=keyboard # Fix install on macos and linux-ci https://github.com/boppreh/keyboard/pull/568 numpy>=1.26 # Python 3.12 support -opencv-python-headless>=4.8.1.78 # Typing fixes +opencv-python-headless>=4.9.0.80 # Typing fixes packaging psutil>=5.9.6 # Python 3.12 fixes # PyAutoGUI # See install.ps1 @@ -29,5 +29,5 @@ winsdk>=1.0.0b10 ; sys_platform == 'win32' # Python 3.12 support # D3DShot # See install.ps1 # # Linux-only dependencies -pyscreeze ; sys_platform == 'linux' +PyScreeze ; sys_platform == 'linux' python-xlib ; sys_platform == 'linux' diff --git a/src/AutoSplit.py b/src/AutoSplit.py index 85228c4c..cec4fde7 100644 --- a/src/AutoSplit.py +++ b/src/AutoSplit.py @@ -895,7 +895,7 @@ def set_preview_image(qlabel: QLabel, image: MatLike | None): capture = image qimage = QtGui.QImage( - capture.data, # pyright: ignore[reportGeneralTypeIssues] # https://bugreports.qt.io/browse/PYSIDE-2476 + capture.data, width, height, width * channels, diff --git a/src/capture_method/CaptureMethodBase.py b/src/capture_method/CaptureMethodBase.py index 31c9e16d..4926da12 100644 --- a/src/capture_method/CaptureMethodBase.py +++ b/src/capture_method/CaptureMethodBase.py @@ -29,7 +29,7 @@ def __init__(self, autosplit: "AutoSplit"): def reinitialize(self): self.close() - self.__init__(self._autosplit_ref) # type: ignore[misc] + self.__init__(self._autosplit_ref) # type: ignore[misc] # noqa: PLC2801 def close(self): # Some capture methods don't need any cleanup diff --git a/src/capture_method/VideoCaptureDeviceCaptureMethod.py b/src/capture_method/VideoCaptureDeviceCaptureMethod.py index 6725ac59..7f778e41 100644 --- a/src/capture_method/VideoCaptureDeviceCaptureMethod.py +++ b/src/capture_method/VideoCaptureDeviceCaptureMethod.py @@ -16,7 +16,7 @@ if TYPE_CHECKING: from AutoSplit import AutoSplit -OBS_VIRTUALCAM_PLUGIN_BLANK_PIXEL = [127, 129, 128] +OBS_VIRTUALCAM_PLUGIN_BLANK_PIXEL = (127, 129, 128) def is_blank(image: MatLike): diff --git a/src/capture_method/__init__.py b/src/capture_method/__init__.py index 0e26596e..3c5c28b1 100644 --- a/src/capture_method/__init__.py +++ b/src/capture_method/__init__.py @@ -1,5 +1,3 @@ -# pyright: reportUnknownMemberType=false -# Complete type for PIL.features.check_feature upstream import asyncio import os import sys @@ -16,7 +14,7 @@ from utils import WGC_MIN_BUILD, WINDOWS_BUILD_NUMBER, first, try_get_direct3d_device if sys.platform == "win32": - from _ctypes import COMError + from _ctypes import COMError # noqa: PLC2701 from pygrabber.dshow_graph import FilterGraph from capture_method.BitBltCaptureMethod import BitBltCaptureMethod @@ -120,7 +118,7 @@ def __getitem__( # type:ignore[override] # pyright: ignore[reportIncompatibleMe return super().__getitem__(__key) @override - def get(self, key: CaptureMethodEnum, __default: object = None): + def get(self, key: CaptureMethodEnum, default: object = None, /): """ Returns the `CaptureMethodBase` subclass for `CaptureMethodEnum` if `CaptureMethodEnum` is available, else defaults to the first available `CaptureMethodEnum`. diff --git a/src/compare.py b/src/compare.py index 49bec7e6..f93788da 100644 --- a/src/compare.py +++ b/src/compare.py @@ -8,9 +8,9 @@ from utils import BGRA_CHANNEL_COUNT, MAXBYTE, ColorChannel, ImageShape, is_valid_image MAXRANGE = MAXBYTE + 1 -CHANNELS = [ColorChannel.Red.value, ColorChannel.Green.value, ColorChannel.Blue.value] -HISTOGRAM_SIZE = [8, 8, 8] -RANGES = [0, MAXRANGE, 0, MAXRANGE, 0, MAXRANGE] +CHANNELS = (ColorChannel.Red.value, ColorChannel.Green.value, ColorChannel.Blue.value) +HISTOGRAM_SIZE = (8, 8, 8) +RANGES = (0, MAXRANGE, 0, MAXRANGE, 0, MAXRANGE) MASK_SIZE_MULTIPLIER = ColorChannel.Alpha * MAXBYTE * MAXBYTE diff --git a/src/hotkeys.py b/src/hotkeys.py index ea83651b..8b6f02ba 100644 --- a/src/hotkeys.py +++ b/src/hotkeys.py @@ -31,7 +31,7 @@ Commands = Literal["split", "start", "pause", "reset", "skip", "undo"] Hotkey = Literal["split", "reset", "skip_split", "undo_split", "pause", "screenshot", "toggle_auto_reset_image"] -HOTKEYS: list[Hotkey] = ["split", "reset", "skip_split", "undo_split", "pause", "screenshot", "toggle_auto_reset_image"] +HOTKEYS = ("split", "reset", "skip_split", "undo_split", "pause", "screenshot", "toggle_auto_reset_image") def remove_all_hotkeys(): diff --git a/src/menu_bar.py b/src/menu_bar.py index 8ae61121..7811573c 100644 --- a/src/menu_bar.py +++ b/src/menu_bar.py @@ -2,6 +2,7 @@ import json import sys import webbrowser +from functools import partial from typing import TYPE_CHECKING, Any, cast from urllib.error import URLError from urllib.request import urlopen @@ -23,7 +24,7 @@ get_all_video_capture_devices, ) from gen import about, design, settings as settings_ui, update_checker -from hotkeys import HOTKEYS, Hotkey, set_hotkey +from hotkeys import HOTKEYS, set_hotkey from utils import AUTOSPLIT_VERSION, GITHUB_REPOSITORY, decimal, fire_and_forget if TYPE_CHECKING: @@ -286,16 +287,13 @@ def __select_screenshot_directory(self): self.screenshot_directory_input.setText(self._autosplit_ref.settings_dict["screenshot_directory"]) def __setup_bindings(self): - # Hotkey initial values and bindings - def hotkey_connect(hotkey: Hotkey): - return lambda: set_hotkey(self._autosplit_ref, hotkey) - + """Hotkey initial values and bindings.""" for hotkey in HOTKEYS: hotkey_input: QtWidgets.QLineEdit = getattr(self, f"{hotkey}_input") set_hotkey_hotkey_button: QtWidgets.QPushButton = getattr(self, f"set_{hotkey}_hotkey_button") hotkey_input.setText(self._autosplit_ref.settings_dict.get(f"{hotkey}_hotkey", "")) - set_hotkey_hotkey_button.clicked.connect(hotkey_connect(hotkey)) + set_hotkey_hotkey_button.clicked.connect(partial(set_hotkey, hotkey=hotkey)) # Make it very clear that hotkeys are not used when auto-controlled if self._autosplit_ref.is_auto_controlled and hotkey != "toggle_auto_reset_image": set_hotkey_hotkey_button.setEnabled(False) diff --git a/src/region_selection.py b/src/region_selection.py index 51431284..c53fad07 100644 --- a/src/region_selection.py +++ b/src/region_selection.py @@ -26,7 +26,7 @@ import win32api import win32gui from win32con import SM_CXVIRTUALSCREEN, SM_CYVIRTUALSCREEN, SM_XVIRTUALSCREEN, SM_YVIRTUALSCREEN - from winsdk._winrt import initialize_with_window + from winsdk._winrt import initialize_with_window # noqa: PLC2701 from winsdk.windows.foundation import AsyncStatus, IAsyncOperation from winsdk.windows.graphics.capture import GraphicsCaptureItem, GraphicsCapturePicker @@ -45,7 +45,7 @@ GNOME_DESKTOP_ICONS_EXTENSION = "@!0,0;BDHF" ALIGN_REGION_THRESHOLD = 0.9 BORDER_WIDTH = 2 -SUPPORTED_IMREAD_FORMATS = [ +SUPPORTED_IMREAD_FORMATS = ( ("Windows bitmaps", "*.bmp *.dib"), ("JPEG files", "*.jpeg *.jpg *.jpe"), ("JPEG 2000 files", "*.jp2"), @@ -58,7 +58,7 @@ ("TIFF files", "*.tiff *.tif"), ("OpenEXR Image files", "*.exr"), ("Radiance HDR", "*.hdr *.pic"), -] +) """https://docs.opencv.org/4.8.0/d4/da8/group__imgcodecs.html#imread""" IMREAD_EXT_FILTER = ( "All Files (" diff --git a/src/split_parser.py b/src/split_parser.py index acdf6dd7..7c2ab55c 100644 --- a/src/split_parser.py +++ b/src/split_parser.py @@ -1,5 +1,6 @@ import os from collections.abc import Callable +from functools import partial from typing import TYPE_CHECKING, TypeVar import error_messages @@ -9,12 +10,12 @@ if TYPE_CHECKING: from AutoSplit import AutoSplit -[ +( DUMMY_FLAG, BELOW_FLAG, PAUSE_FLAG, *_, -] = [1 << i for i in range(31)] # 32 bits of flags +) = tuple(1 << i for i in range(31)) # 32 bits of flags T = TypeVar("T", str, int, float) @@ -209,11 +210,7 @@ def parse_and_validate_images(autosplit: "AutoSplit"): for image in split_images: # Test for image without transparency if not is_valid_image(image.byte_array): - - def image_validity(filename: str): - return lambda: error_messages.image_validity(filename) - - error_message = image_validity(image.filename) + error_message = partial(error_messages.image_validity, image.filename) break # error out if there is a {p} flag but no pause hotkey set and is not auto controlled. diff --git a/src/utils.py b/src/utils.py index dfaf9486..8b4c3c5e 100644 --- a/src/utils.py +++ b/src/utils.py @@ -1,6 +1,5 @@ import asyncio import os -import subprocess import sys from collections.abc import Callable, Iterable from enum import IntEnum @@ -129,8 +128,10 @@ def open_file(file_path: str | bytes | os.PathLike[str] | os.PathLike[bytes]): if sys.platform == "win32": os.startfile(file_path) # noqa: S606 else: + import subprocess # noqa: PLC0415, S404 + opener = "xdg-open" if sys.platform == "linux" else "open" - subprocess.call([opener, file_path]) # noqa: S603 + subprocess.call([opener, file_path]) # noqa: S603 def get_or_create_eventloop(): diff --git a/typings/cv2/__init__.pyi b/typings/cv2/__init__.pyi deleted file mode 100644 index 28ef8445..00000000 --- a/typings/cv2/__init__.pyi +++ /dev/null @@ -1,10430 +0,0 @@ -import typing - -import cv2.aruco -import cv2.cuda -import cv2.gapi -import cv2.gapi.streaming -import cv2.typing -import numpy -from cv2 import ( - Error as Error, - aruco as aruco, - barcode as barcode, - cuda as cuda, - detail as detail, - dnn as dnn, - fisheye as fisheye, - flann as flann, - gapi as gapi, - ipp as ipp, - ml as ml, - ocl as ocl, - ogl as ogl, - parallel as parallel, - samples as samples, - segmentation as segmentation, - utils as utils, - videoio_registry as videoio_registry, -) -from cv2.mat_wrapper import Mat as Mat - -# Enumerations -SORT_EVERY_ROW: int -SORT_EVERY_COLUMN: int -SORT_ASCENDING: int -SORT_DESCENDING: int -SortFlags = int -"""One of [SORT_EVERY_ROW, SORT_EVERY_COLUMN, SORT_ASCENDING, SORT_DESCENDING]""" - -COVAR_SCRAMBLED: int -COVAR_NORMAL: int -COVAR_USE_AVG: int -COVAR_SCALE: int -COVAR_ROWS: int -COVAR_COLS: int -CovarFlags = int -"""One of [COVAR_SCRAMBLED, COVAR_NORMAL, COVAR_USE_AVG, COVAR_SCALE, COVAR_ROWS, COVAR_COLS]""" - -KMEANS_RANDOM_CENTERS: int -KMEANS_PP_CENTERS: int -KMEANS_USE_INITIAL_LABELS: int -KmeansFlags = int -"""One of [KMEANS_RANDOM_CENTERS, KMEANS_PP_CENTERS, KMEANS_USE_INITIAL_LABELS]""" - -REDUCE_SUM: int -REDUCE_AVG: int -REDUCE_MAX: int -REDUCE_MIN: int -REDUCE_SUM2: int -ReduceTypes = int -"""One of [REDUCE_SUM, REDUCE_AVG, REDUCE_MAX, REDUCE_MIN, REDUCE_SUM2]""" - -ROTATE_90_CLOCKWISE: int -ROTATE_180: int -ROTATE_90_COUNTERCLOCKWISE: int -RotateFlags = int -"""One of [ROTATE_90_CLOCKWISE, ROTATE_180, ROTATE_90_COUNTERCLOCKWISE]""" - -Param_INT: int -PARAM_INT: int -Param_BOOLEAN: int -PARAM_BOOLEAN: int -Param_REAL: int -PARAM_REAL: int -Param_STRING: int -PARAM_STRING: int -Param_MAT: int -PARAM_MAT: int -Param_MAT_VECTOR: int -PARAM_MAT_VECTOR: int -Param_ALGORITHM: int -PARAM_ALGORITHM: int -Param_FLOAT: int -PARAM_FLOAT: int -Param_UNSIGNED_INT: int -PARAM_UNSIGNED_INT: int -Param_UINT64: int -PARAM_UINT64: int -Param_UCHAR: int -PARAM_UCHAR: int -Param_SCALAR: int -PARAM_SCALAR: int -Param = int -"""One of [Param_INT, PARAM_INT, Param_BOOLEAN, PARAM_BOOLEAN, Param_REAL, PARAM_REAL, Param_STRING, PARAM_STRING, -Param_MAT, PARAM_MAT, Param_MAT_VECTOR, PARAM_MAT_VECTOR, Param_ALGORITHM, PARAM_ALGORITHM, Param_FLOAT, PARAM_FLOAT, -Param_UNSIGNED_INT, PARAM_UNSIGNED_INT, Param_UINT64, PARAM_UINT64, Param_UCHAR, PARAM_UCHAR, Param_SCALAR, -PARAM_SCALAR]""" - -DECOMP_LU: int -DECOMP_SVD: int -DECOMP_EIG: int -DECOMP_CHOLESKY: int -DECOMP_QR: int -DECOMP_NORMAL: int -DecompTypes = int -"""One of [DECOMP_LU, DECOMP_SVD, DECOMP_EIG, DECOMP_CHOLESKY, DECOMP_QR, DECOMP_NORMAL]""" - -NORM_INF: int -NORM_L1: int -NORM_L2: int -NORM_L2SQR: int -NORM_HAMMING: int -NORM_HAMMING2: int -NORM_TYPE_MASK: int -NORM_RELATIVE: int -NORM_MINMAX: int -NormTypes = int -"""One of [NORM_INF, NORM_L1, NORM_L2, NORM_L2SQR, NORM_HAMMING, NORM_HAMMING2, NORM_TYPE_MASK, NORM_RELATIVE, -NORM_MINMAX]""" - -CMP_EQ: int -CMP_GT: int -CMP_GE: int -CMP_LT: int -CMP_LE: int -CMP_NE: int -CmpTypes = int -"""One of [CMP_EQ, CMP_GT, CMP_GE, CMP_LT, CMP_LE, CMP_NE]""" - -GEMM_1_T: int -GEMM_2_T: int -GEMM_3_T: int -GemmFlags = int -"""One of [GEMM_1_T, GEMM_2_T, GEMM_3_T]""" - -DFT_INVERSE: int -DFT_SCALE: int -DFT_ROWS: int -DFT_COMPLEX_OUTPUT: int -DFT_REAL_OUTPUT: int -DFT_COMPLEX_INPUT: int -DCT_INVERSE: int -DCT_ROWS: int -DftFlags = int -"""One of [DFT_INVERSE, DFT_SCALE, DFT_ROWS, DFT_COMPLEX_OUTPUT, DFT_REAL_OUTPUT, DFT_COMPLEX_INPUT, DCT_INVERSE, -DCT_ROWS]""" - -BORDER_CONSTANT: int -BORDER_REPLICATE: int -BORDER_REFLECT: int -BORDER_WRAP: int -BORDER_REFLECT_101: int -BORDER_TRANSPARENT: int -BORDER_REFLECT101: int -BORDER_DEFAULT: int -BORDER_ISOLATED: int -BorderTypes = int -"""One of [BORDER_CONSTANT, BORDER_REPLICATE, BORDER_REFLECT, BORDER_WRAP, BORDER_REFLECT_101, BORDER_TRANSPARENT, -BORDER_REFLECT101, BORDER_DEFAULT, BORDER_ISOLATED]""" - -ACCESS_READ: int -ACCESS_WRITE: int -ACCESS_RW: int -ACCESS_MASK: int -ACCESS_FAST: int -AccessFlag = int -"""One of [ACCESS_READ, ACCESS_WRITE, ACCESS_RW, ACCESS_MASK, ACCESS_FAST]""" - -USAGE_DEFAULT: int -USAGE_ALLOCATE_HOST_MEMORY: int -USAGE_ALLOCATE_DEVICE_MEMORY: int -USAGE_ALLOCATE_SHARED_MEMORY: int -__UMAT_USAGE_FLAGS_32BIT: int -UMatUsageFlags = int -"""One of [USAGE_DEFAULT, USAGE_ALLOCATE_HOST_MEMORY, USAGE_ALLOCATE_DEVICE_MEMORY, USAGE_ALLOCATE_SHARED_MEMORY, -__UMAT_USAGE_FLAGS_32BIT]""" - -SOLVELP_LOST: int -SOLVELP_UNBOUNDED: int -SOLVELP_UNFEASIBLE: int -SOLVELP_SINGLE: int -SOLVELP_MULTI: int -SolveLPResult = int -"""One of [SOLVELP_LOST, SOLVELP_UNBOUNDED, SOLVELP_UNFEASIBLE, SOLVELP_SINGLE, SOLVELP_MULTI]""" - -QUAT_ASSUME_NOT_UNIT: int -QUAT_ASSUME_UNIT: int -QuatAssumeType = int -"""One of [QUAT_ASSUME_NOT_UNIT, QUAT_ASSUME_UNIT]""" - -FILTER_SCHARR: int -SpecialFilter = int -"""One of [FILTER_SCHARR]""" - -MORPH_ERODE: int -MORPH_DILATE: int -MORPH_OPEN: int -MORPH_CLOSE: int -MORPH_GRADIENT: int -MORPH_TOPHAT: int -MORPH_BLACKHAT: int -MORPH_HITMISS: int -MorphTypes = int -"""One of [MORPH_ERODE, MORPH_DILATE, MORPH_OPEN, MORPH_CLOSE, MORPH_GRADIENT, MORPH_TOPHAT, MORPH_BLACKHAT, -MORPH_HITMISS]""" - -MORPH_RECT: int -MORPH_CROSS: int -MORPH_ELLIPSE: int -MorphShapes = int -"""One of [MORPH_RECT, MORPH_CROSS, MORPH_ELLIPSE]""" - -INTER_NEAREST: int -INTER_LINEAR: int -INTER_CUBIC: int -INTER_AREA: int -INTER_LANCZOS4: int -INTER_LINEAR_EXACT: int -INTER_NEAREST_EXACT: int -INTER_MAX: int -WARP_FILL_OUTLIERS: int -WARP_INVERSE_MAP: int -InterpolationFlags = int -"""One of [INTER_NEAREST, INTER_LINEAR, INTER_CUBIC, INTER_AREA, INTER_LANCZOS4, INTER_LINEAR_EXACT, -INTER_NEAREST_EXACT, INTER_MAX, WARP_FILL_OUTLIERS, WARP_INVERSE_MAP]""" - -WARP_POLAR_LINEAR: int -WARP_POLAR_LOG: int -WarpPolarMode = int -"""One of [WARP_POLAR_LINEAR, WARP_POLAR_LOG]""" - -INTER_BITS: int -INTER_BITS2: int -INTER_TAB_SIZE: int -INTER_TAB_SIZE2: int -InterpolationMasks = int -"""One of [INTER_BITS, INTER_BITS2, INTER_TAB_SIZE, INTER_TAB_SIZE2]""" - -DIST_USER: int -DIST_L1: int -DIST_L2: int -DIST_C: int -DIST_L12: int -DIST_FAIR: int -DIST_WELSCH: int -DIST_HUBER: int -DistanceTypes = int -"""One of [DIST_USER, DIST_L1, DIST_L2, DIST_C, DIST_L12, DIST_FAIR, DIST_WELSCH, DIST_HUBER]""" - -DIST_MASK_3: int -DIST_MASK_5: int -DIST_MASK_PRECISE: int -DistanceTransformMasks = int -"""One of [DIST_MASK_3, DIST_MASK_5, DIST_MASK_PRECISE]""" - -THRESH_BINARY: int -THRESH_BINARY_INV: int -THRESH_TRUNC: int -THRESH_TOZERO: int -THRESH_TOZERO_INV: int -THRESH_MASK: int -THRESH_OTSU: int -THRESH_TRIANGLE: int -ThresholdTypes = int -"""One of [THRESH_BINARY, THRESH_BINARY_INV, THRESH_TRUNC, THRESH_TOZERO, THRESH_TOZERO_INV, THRESH_MASK, THRESH_OTSU, -THRESH_TRIANGLE]""" - -ADAPTIVE_THRESH_MEAN_C: int -ADAPTIVE_THRESH_GAUSSIAN_C: int -AdaptiveThresholdTypes = int -"""One of [ADAPTIVE_THRESH_MEAN_C, ADAPTIVE_THRESH_GAUSSIAN_C]""" - -GC_BGD: int -GC_FGD: int -GC_PR_BGD: int -GC_PR_FGD: int -GrabCutClasses = int -"""One of [GC_BGD, GC_FGD, GC_PR_BGD, GC_PR_FGD]""" - -GC_INIT_WITH_RECT: int -GC_INIT_WITH_MASK: int -GC_EVAL: int -GC_EVAL_FREEZE_MODEL: int -GrabCutModes = int -"""One of [GC_INIT_WITH_RECT, GC_INIT_WITH_MASK, GC_EVAL, GC_EVAL_FREEZE_MODEL]""" - -DIST_LABEL_CCOMP: int -DIST_LABEL_PIXEL: int -DistanceTransformLabelTypes = int -"""One of [DIST_LABEL_CCOMP, DIST_LABEL_PIXEL]""" - -FLOODFILL_FIXED_RANGE: int -FLOODFILL_MASK_ONLY: int -FloodFillFlags = int -"""One of [FLOODFILL_FIXED_RANGE, FLOODFILL_MASK_ONLY]""" - -CC_STAT_LEFT: int -CC_STAT_TOP: int -CC_STAT_WIDTH: int -CC_STAT_HEIGHT: int -CC_STAT_AREA: int -CC_STAT_MAX: int -ConnectedComponentsTypes = int -"""One of [CC_STAT_LEFT, CC_STAT_TOP, CC_STAT_WIDTH, CC_STAT_HEIGHT, CC_STAT_AREA, CC_STAT_MAX]""" - -CCL_DEFAULT: int -CCL_WU: int -CCL_GRANA: int -CCL_BOLELLI: int -CCL_SAUF: int -CCL_BBDT: int -CCL_SPAGHETTI: int -ConnectedComponentsAlgorithmsTypes = int -"""One of [CCL_DEFAULT, CCL_WU, CCL_GRANA, CCL_BOLELLI, CCL_SAUF, CCL_BBDT, CCL_SPAGHETTI]""" - -RETR_EXTERNAL: int -RETR_LIST: int -RETR_CCOMP: int -RETR_TREE: int -RETR_FLOODFILL: int -RetrievalModes = int -"""One of [RETR_EXTERNAL, RETR_LIST, RETR_CCOMP, RETR_TREE, RETR_FLOODFILL]""" - -CHAIN_APPROX_NONE: int -CHAIN_APPROX_SIMPLE: int -CHAIN_APPROX_TC89_L1: int -CHAIN_APPROX_TC89_KCOS: int -ContourApproximationModes = int -"""One of [CHAIN_APPROX_NONE, CHAIN_APPROX_SIMPLE, CHAIN_APPROX_TC89_L1, CHAIN_APPROX_TC89_KCOS]""" - -CONTOURS_MATCH_I1: int -CONTOURS_MATCH_I2: int -CONTOURS_MATCH_I3: int -ShapeMatchModes = int -"""One of [CONTOURS_MATCH_I1, CONTOURS_MATCH_I2, CONTOURS_MATCH_I3]""" - -HOUGH_STANDARD: int -HOUGH_PROBABILISTIC: int -HOUGH_MULTI_SCALE: int -HOUGH_GRADIENT: int -HOUGH_GRADIENT_ALT: int -HoughModes = int -"""One of [HOUGH_STANDARD, HOUGH_PROBABILISTIC, HOUGH_MULTI_SCALE, HOUGH_GRADIENT, HOUGH_GRADIENT_ALT]""" - -LSD_REFINE_NONE: int -LSD_REFINE_STD: int -LSD_REFINE_ADV: int -LineSegmentDetectorModes = int -"""One of [LSD_REFINE_NONE, LSD_REFINE_STD, LSD_REFINE_ADV]""" - -HISTCMP_CORREL: int -HISTCMP_CHISQR: int -HISTCMP_INTERSECT: int -HISTCMP_BHATTACHARYYA: int -HISTCMP_HELLINGER: int -HISTCMP_CHISQR_ALT: int -HISTCMP_KL_DIV: int -HistCompMethods = int -"""One of [HISTCMP_CORREL, HISTCMP_CHISQR, HISTCMP_INTERSECT, HISTCMP_BHATTACHARYYA, HISTCMP_HELLINGER, -HISTCMP_CHISQR_ALT, HISTCMP_KL_DIV]""" - -COLOR_BGR2BGRA: int -COLOR_RGB2RGBA: int -COLOR_BGRA2BGR: int -COLOR_RGBA2RGB: int -COLOR_BGR2RGBA: int -COLOR_RGB2BGRA: int -COLOR_RGBA2BGR: int -COLOR_BGRA2RGB: int -COLOR_BGR2RGB: int -COLOR_RGB2BGR: int -COLOR_BGRA2RGBA: int -COLOR_RGBA2BGRA: int -COLOR_BGR2GRAY: int -COLOR_RGB2GRAY: int -COLOR_GRAY2BGR: int -COLOR_GRAY2RGB: int -COLOR_GRAY2BGRA: int -COLOR_GRAY2RGBA: int -COLOR_BGRA2GRAY: int -COLOR_RGBA2GRAY: int -COLOR_BGR2BGR565: int -COLOR_RGB2BGR565: int -COLOR_BGR5652BGR: int -COLOR_BGR5652RGB: int -COLOR_BGRA2BGR565: int -COLOR_RGBA2BGR565: int -COLOR_BGR5652BGRA: int -COLOR_BGR5652RGBA: int -COLOR_GRAY2BGR565: int -COLOR_BGR5652GRAY: int -COLOR_BGR2BGR555: int -COLOR_RGB2BGR555: int -COLOR_BGR5552BGR: int -COLOR_BGR5552RGB: int -COLOR_BGRA2BGR555: int -COLOR_RGBA2BGR555: int -COLOR_BGR5552BGRA: int -COLOR_BGR5552RGBA: int -COLOR_GRAY2BGR555: int -COLOR_BGR5552GRAY: int -COLOR_BGR2XYZ: int -COLOR_RGB2XYZ: int -COLOR_XYZ2BGR: int -COLOR_XYZ2RGB: int -COLOR_BGR2YCrCb: int -COLOR_BGR2YCR_CB: int -COLOR_RGB2YCrCb: int -COLOR_RGB2YCR_CB: int -COLOR_YCrCb2BGR: int -COLOR_YCR_CB2BGR: int -COLOR_YCrCb2RGB: int -COLOR_YCR_CB2RGB: int -COLOR_BGR2HSV: int -COLOR_RGB2HSV: int -COLOR_BGR2Lab: int -COLOR_BGR2LAB: int -COLOR_RGB2Lab: int -COLOR_RGB2LAB: int -COLOR_BGR2Luv: int -COLOR_BGR2LUV: int -COLOR_RGB2Luv: int -COLOR_RGB2LUV: int -COLOR_BGR2HLS: int -COLOR_RGB2HLS: int -COLOR_HSV2BGR: int -COLOR_HSV2RGB: int -COLOR_Lab2BGR: int -COLOR_LAB2BGR: int -COLOR_Lab2RGB: int -COLOR_LAB2RGB: int -COLOR_Luv2BGR: int -COLOR_LUV2BGR: int -COLOR_Luv2RGB: int -COLOR_LUV2RGB: int -COLOR_HLS2BGR: int -COLOR_HLS2RGB: int -COLOR_BGR2HSV_FULL: int -COLOR_RGB2HSV_FULL: int -COLOR_BGR2HLS_FULL: int -COLOR_RGB2HLS_FULL: int -COLOR_HSV2BGR_FULL: int -COLOR_HSV2RGB_FULL: int -COLOR_HLS2BGR_FULL: int -COLOR_HLS2RGB_FULL: int -COLOR_LBGR2Lab: int -COLOR_LBGR2LAB: int -COLOR_LRGB2Lab: int -COLOR_LRGB2LAB: int -COLOR_LBGR2Luv: int -COLOR_LBGR2LUV: int -COLOR_LRGB2Luv: int -COLOR_LRGB2LUV: int -COLOR_Lab2LBGR: int -COLOR_LAB2LBGR: int -COLOR_Lab2LRGB: int -COLOR_LAB2LRGB: int -COLOR_Luv2LBGR: int -COLOR_LUV2LBGR: int -COLOR_Luv2LRGB: int -COLOR_LUV2LRGB: int -COLOR_BGR2YUV: int -COLOR_RGB2YUV: int -COLOR_YUV2BGR: int -COLOR_YUV2RGB: int -COLOR_YUV2RGB_NV12: int -COLOR_YUV2BGR_NV12: int -COLOR_YUV2RGB_NV21: int -COLOR_YUV2BGR_NV21: int -COLOR_YUV420sp2RGB: int -COLOR_YUV420SP2RGB: int -COLOR_YUV420sp2BGR: int -COLOR_YUV420SP2BGR: int -COLOR_YUV2RGBA_NV12: int -COLOR_YUV2BGRA_NV12: int -COLOR_YUV2RGBA_NV21: int -COLOR_YUV2BGRA_NV21: int -COLOR_YUV420sp2RGBA: int -COLOR_YUV420SP2RGBA: int -COLOR_YUV420sp2BGRA: int -COLOR_YUV420SP2BGRA: int -COLOR_YUV2RGB_YV12: int -COLOR_YUV2BGR_YV12: int -COLOR_YUV2RGB_IYUV: int -COLOR_YUV2BGR_IYUV: int -COLOR_YUV2RGB_I420: int -COLOR_YUV2BGR_I420: int -COLOR_YUV420p2RGB: int -COLOR_YUV420P2RGB: int -COLOR_YUV420p2BGR: int -COLOR_YUV420P2BGR: int -COLOR_YUV2RGBA_YV12: int -COLOR_YUV2BGRA_YV12: int -COLOR_YUV2RGBA_IYUV: int -COLOR_YUV2BGRA_IYUV: int -COLOR_YUV2RGBA_I420: int -COLOR_YUV2BGRA_I420: int -COLOR_YUV420p2RGBA: int -COLOR_YUV420P2RGBA: int -COLOR_YUV420p2BGRA: int -COLOR_YUV420P2BGRA: int -COLOR_YUV2GRAY_420: int -COLOR_YUV2GRAY_NV21: int -COLOR_YUV2GRAY_NV12: int -COLOR_YUV2GRAY_YV12: int -COLOR_YUV2GRAY_IYUV: int -COLOR_YUV2GRAY_I420: int -COLOR_YUV420sp2GRAY: int -COLOR_YUV420SP2GRAY: int -COLOR_YUV420p2GRAY: int -COLOR_YUV420P2GRAY: int -COLOR_YUV2RGB_UYVY: int -COLOR_YUV2BGR_UYVY: int -COLOR_YUV2RGB_Y422: int -COLOR_YUV2BGR_Y422: int -COLOR_YUV2RGB_UYNV: int -COLOR_YUV2BGR_UYNV: int -COLOR_YUV2RGBA_UYVY: int -COLOR_YUV2BGRA_UYVY: int -COLOR_YUV2RGBA_Y422: int -COLOR_YUV2BGRA_Y422: int -COLOR_YUV2RGBA_UYNV: int -COLOR_YUV2BGRA_UYNV: int -COLOR_YUV2RGB_YUY2: int -COLOR_YUV2BGR_YUY2: int -COLOR_YUV2RGB_YVYU: int -COLOR_YUV2BGR_YVYU: int -COLOR_YUV2RGB_YUYV: int -COLOR_YUV2BGR_YUYV: int -COLOR_YUV2RGB_YUNV: int -COLOR_YUV2BGR_YUNV: int -COLOR_YUV2RGBA_YUY2: int -COLOR_YUV2BGRA_YUY2: int -COLOR_YUV2RGBA_YVYU: int -COLOR_YUV2BGRA_YVYU: int -COLOR_YUV2RGBA_YUYV: int -COLOR_YUV2BGRA_YUYV: int -COLOR_YUV2RGBA_YUNV: int -COLOR_YUV2BGRA_YUNV: int -COLOR_YUV2GRAY_UYVY: int -COLOR_YUV2GRAY_YUY2: int -COLOR_YUV2GRAY_Y422: int -COLOR_YUV2GRAY_UYNV: int -COLOR_YUV2GRAY_YVYU: int -COLOR_YUV2GRAY_YUYV: int -COLOR_YUV2GRAY_YUNV: int -COLOR_RGBA2mRGBA: int -COLOR_RGBA2M_RGBA: int -COLOR_mRGBA2RGBA: int -COLOR_M_RGBA2RGBA: int -COLOR_RGB2YUV_I420: int -COLOR_BGR2YUV_I420: int -COLOR_RGB2YUV_IYUV: int -COLOR_BGR2YUV_IYUV: int -COLOR_RGBA2YUV_I420: int -COLOR_BGRA2YUV_I420: int -COLOR_RGBA2YUV_IYUV: int -COLOR_BGRA2YUV_IYUV: int -COLOR_RGB2YUV_YV12: int -COLOR_BGR2YUV_YV12: int -COLOR_RGBA2YUV_YV12: int -COLOR_BGRA2YUV_YV12: int -COLOR_BayerBG2BGR: int -COLOR_BAYER_BG2BGR: int -COLOR_BayerGB2BGR: int -COLOR_BAYER_GB2BGR: int -COLOR_BayerRG2BGR: int -COLOR_BAYER_RG2BGR: int -COLOR_BayerGR2BGR: int -COLOR_BAYER_GR2BGR: int -COLOR_BayerRGGB2BGR: int -COLOR_BAYER_RGGB2BGR: int -COLOR_BayerGRBG2BGR: int -COLOR_BAYER_GRBG2BGR: int -COLOR_BayerBGGR2BGR: int -COLOR_BAYER_BGGR2BGR: int -COLOR_BayerGBRG2BGR: int -COLOR_BAYER_GBRG2BGR: int -COLOR_BayerRGGB2RGB: int -COLOR_BAYER_RGGB2RGB: int -COLOR_BayerGRBG2RGB: int -COLOR_BAYER_GRBG2RGB: int -COLOR_BayerBGGR2RGB: int -COLOR_BAYER_BGGR2RGB: int -COLOR_BayerGBRG2RGB: int -COLOR_BAYER_GBRG2RGB: int -COLOR_BayerBG2RGB: int -COLOR_BAYER_BG2RGB: int -COLOR_BayerGB2RGB: int -COLOR_BAYER_GB2RGB: int -COLOR_BayerRG2RGB: int -COLOR_BAYER_RG2RGB: int -COLOR_BayerGR2RGB: int -COLOR_BAYER_GR2RGB: int -COLOR_BayerBG2GRAY: int -COLOR_BAYER_BG2GRAY: int -COLOR_BayerGB2GRAY: int -COLOR_BAYER_GB2GRAY: int -COLOR_BayerRG2GRAY: int -COLOR_BAYER_RG2GRAY: int -COLOR_BayerGR2GRAY: int -COLOR_BAYER_GR2GRAY: int -COLOR_BayerRGGB2GRAY: int -COLOR_BAYER_RGGB2GRAY: int -COLOR_BayerGRBG2GRAY: int -COLOR_BAYER_GRBG2GRAY: int -COLOR_BayerBGGR2GRAY: int -COLOR_BAYER_BGGR2GRAY: int -COLOR_BayerGBRG2GRAY: int -COLOR_BAYER_GBRG2GRAY: int -COLOR_BayerBG2BGR_VNG: int -COLOR_BAYER_BG2BGR_VNG: int -COLOR_BayerGB2BGR_VNG: int -COLOR_BAYER_GB2BGR_VNG: int -COLOR_BayerRG2BGR_VNG: int -COLOR_BAYER_RG2BGR_VNG: int -COLOR_BayerGR2BGR_VNG: int -COLOR_BAYER_GR2BGR_VNG: int -COLOR_BayerRGGB2BGR_VNG: int -COLOR_BAYER_RGGB2BGR_VNG: int -COLOR_BayerGRBG2BGR_VNG: int -COLOR_BAYER_GRBG2BGR_VNG: int -COLOR_BayerBGGR2BGR_VNG: int -COLOR_BAYER_BGGR2BGR_VNG: int -COLOR_BayerGBRG2BGR_VNG: int -COLOR_BAYER_GBRG2BGR_VNG: int -COLOR_BayerRGGB2RGB_VNG: int -COLOR_BAYER_RGGB2RGB_VNG: int -COLOR_BayerGRBG2RGB_VNG: int -COLOR_BAYER_GRBG2RGB_VNG: int -COLOR_BayerBGGR2RGB_VNG: int -COLOR_BAYER_BGGR2RGB_VNG: int -COLOR_BayerGBRG2RGB_VNG: int -COLOR_BAYER_GBRG2RGB_VNG: int -COLOR_BayerBG2RGB_VNG: int -COLOR_BAYER_BG2RGB_VNG: int -COLOR_BayerGB2RGB_VNG: int -COLOR_BAYER_GB2RGB_VNG: int -COLOR_BayerRG2RGB_VNG: int -COLOR_BAYER_RG2RGB_VNG: int -COLOR_BayerGR2RGB_VNG: int -COLOR_BAYER_GR2RGB_VNG: int -COLOR_BayerBG2BGR_EA: int -COLOR_BAYER_BG2BGR_EA: int -COLOR_BayerGB2BGR_EA: int -COLOR_BAYER_GB2BGR_EA: int -COLOR_BayerRG2BGR_EA: int -COLOR_BAYER_RG2BGR_EA: int -COLOR_BayerGR2BGR_EA: int -COLOR_BAYER_GR2BGR_EA: int -COLOR_BayerRGGB2BGR_EA: int -COLOR_BAYER_RGGB2BGR_EA: int -COLOR_BayerGRBG2BGR_EA: int -COLOR_BAYER_GRBG2BGR_EA: int -COLOR_BayerBGGR2BGR_EA: int -COLOR_BAYER_BGGR2BGR_EA: int -COLOR_BayerGBRG2BGR_EA: int -COLOR_BAYER_GBRG2BGR_EA: int -COLOR_BayerRGGB2RGB_EA: int -COLOR_BAYER_RGGB2RGB_EA: int -COLOR_BayerGRBG2RGB_EA: int -COLOR_BAYER_GRBG2RGB_EA: int -COLOR_BayerBGGR2RGB_EA: int -COLOR_BAYER_BGGR2RGB_EA: int -COLOR_BayerGBRG2RGB_EA: int -COLOR_BAYER_GBRG2RGB_EA: int -COLOR_BayerBG2RGB_EA: int -COLOR_BAYER_BG2RGB_EA: int -COLOR_BayerGB2RGB_EA: int -COLOR_BAYER_GB2RGB_EA: int -COLOR_BayerRG2RGB_EA: int -COLOR_BAYER_RG2RGB_EA: int -COLOR_BayerGR2RGB_EA: int -COLOR_BAYER_GR2RGB_EA: int -COLOR_BayerBG2BGRA: int -COLOR_BAYER_BG2BGRA: int -COLOR_BayerGB2BGRA: int -COLOR_BAYER_GB2BGRA: int -COLOR_BayerRG2BGRA: int -COLOR_BAYER_RG2BGRA: int -COLOR_BayerGR2BGRA: int -COLOR_BAYER_GR2BGRA: int -COLOR_BayerRGGB2BGRA: int -COLOR_BAYER_RGGB2BGRA: int -COLOR_BayerGRBG2BGRA: int -COLOR_BAYER_GRBG2BGRA: int -COLOR_BayerBGGR2BGRA: int -COLOR_BAYER_BGGR2BGRA: int -COLOR_BayerGBRG2BGRA: int -COLOR_BAYER_GBRG2BGRA: int -COLOR_BayerRGGB2RGBA: int -COLOR_BAYER_RGGB2RGBA: int -COLOR_BayerGRBG2RGBA: int -COLOR_BAYER_GRBG2RGBA: int -COLOR_BayerBGGR2RGBA: int -COLOR_BAYER_BGGR2RGBA: int -COLOR_BayerGBRG2RGBA: int -COLOR_BAYER_GBRG2RGBA: int -COLOR_BayerBG2RGBA: int -COLOR_BAYER_BG2RGBA: int -COLOR_BayerGB2RGBA: int -COLOR_BAYER_GB2RGBA: int -COLOR_BayerRG2RGBA: int -COLOR_BAYER_RG2RGBA: int -COLOR_BayerGR2RGBA: int -COLOR_BAYER_GR2RGBA: int -COLOR_COLORCVT_MAX: int -ColorConversionCodes = int -"""One of [COLOR_BGR2BGRA, COLOR_RGB2RGBA, COLOR_BGRA2BGR, COLOR_RGBA2RGB, COLOR_BGR2RGBA, COLOR_RGB2BGRA, -COLOR_RGBA2BGR, COLOR_BGRA2RGB, COLOR_BGR2RGB, COLOR_RGB2BGR, COLOR_BGRA2RGBA, COLOR_RGBA2BGRA, COLOR_BGR2GRAY, -COLOR_RGB2GRAY, COLOR_GRAY2BGR, COLOR_GRAY2RGB, COLOR_GRAY2BGRA, COLOR_GRAY2RGBA, COLOR_BGRA2GRAY, COLOR_RGBA2GRAY, -COLOR_BGR2BGR565, COLOR_RGB2BGR565, COLOR_BGR5652BGR, COLOR_BGR5652RGB, COLOR_BGRA2BGR565, COLOR_RGBA2BGR565, -COLOR_BGR5652BGRA, COLOR_BGR5652RGBA, COLOR_GRAY2BGR565, COLOR_BGR5652GRAY, COLOR_BGR2BGR555, COLOR_RGB2BGR555, -COLOR_BGR5552BGR, COLOR_BGR5552RGB, COLOR_BGRA2BGR555, COLOR_RGBA2BGR555, COLOR_BGR5552BGRA, COLOR_BGR5552RGBA, -COLOR_GRAY2BGR555, COLOR_BGR5552GRAY, COLOR_BGR2XYZ, COLOR_RGB2XYZ, COLOR_XYZ2BGR, COLOR_XYZ2RGB, COLOR_BGR2YCrCb, -COLOR_BGR2YCR_CB, COLOR_RGB2YCrCb, COLOR_RGB2YCR_CB, COLOR_YCrCb2BGR, COLOR_YCR_CB2BGR, COLOR_YCrCb2RGB, -COLOR_YCR_CB2RGB, COLOR_BGR2HSV, COLOR_RGB2HSV, COLOR_BGR2Lab, COLOR_BGR2LAB, COLOR_RGB2Lab, COLOR_RGB2LAB, -COLOR_BGR2Luv, COLOR_BGR2LUV, COLOR_RGB2Luv, COLOR_RGB2LUV, COLOR_BGR2HLS, COLOR_RGB2HLS, COLOR_HSV2BGR, COLOR_HSV2RGB, -COLOR_Lab2BGR, COLOR_LAB2BGR, COLOR_Lab2RGB, COLOR_LAB2RGB, COLOR_Luv2BGR, COLOR_LUV2BGR, COLOR_Luv2RGB, COLOR_LUV2RGB, -COLOR_HLS2BGR, COLOR_HLS2RGB, COLOR_BGR2HSV_FULL, COLOR_RGB2HSV_FULL, COLOR_BGR2HLS_FULL, COLOR_RGB2HLS_FULL, -COLOR_HSV2BGR_FULL, COLOR_HSV2RGB_FULL, COLOR_HLS2BGR_FULL, COLOR_HLS2RGB_FULL, COLOR_LBGR2Lab, COLOR_LBGR2LAB, -COLOR_LRGB2Lab, COLOR_LRGB2LAB, COLOR_LBGR2Luv, COLOR_LBGR2LUV, COLOR_LRGB2Luv, COLOR_LRGB2LUV, COLOR_Lab2LBGR, -COLOR_LAB2LBGR, COLOR_Lab2LRGB, COLOR_LAB2LRGB, COLOR_Luv2LBGR, COLOR_LUV2LBGR, COLOR_Luv2LRGB, COLOR_LUV2LRGB, -COLOR_BGR2YUV, COLOR_RGB2YUV, COLOR_YUV2BGR, COLOR_YUV2RGB, COLOR_YUV2RGB_NV12, COLOR_YUV2BGR_NV12, COLOR_YUV2RGB_NV21, -COLOR_YUV2BGR_NV21, COLOR_YUV420sp2RGB, COLOR_YUV420SP2RGB, COLOR_YUV420sp2BGR, COLOR_YUV420SP2BGR, COLOR_YUV2RGBA_NV12, -COLOR_YUV2BGRA_NV12, COLOR_YUV2RGBA_NV21, COLOR_YUV2BGRA_NV21, COLOR_YUV420sp2RGBA, COLOR_YUV420SP2RGBA, -COLOR_YUV420sp2BGRA, COLOR_YUV420SP2BGRA, COLOR_YUV2RGB_YV12, COLOR_YUV2BGR_YV12, COLOR_YUV2RGB_IYUV, -COLOR_YUV2BGR_IYUV, COLOR_YUV2RGB_I420, COLOR_YUV2BGR_I420, COLOR_YUV420p2RGB, COLOR_YUV420P2RGB, COLOR_YUV420p2BGR, -COLOR_YUV420P2BGR, COLOR_YUV2RGBA_YV12, COLOR_YUV2BGRA_YV12, COLOR_YUV2RGBA_IYUV, COLOR_YUV2BGRA_IYUV, -COLOR_YUV2RGBA_I420, COLOR_YUV2BGRA_I420, COLOR_YUV420p2RGBA, COLOR_YUV420P2RGBA, COLOR_YUV420p2BGRA, -COLOR_YUV420P2BGRA, COLOR_YUV2GRAY_420, COLOR_YUV2GRAY_NV21, COLOR_YUV2GRAY_NV12, COLOR_YUV2GRAY_YV12, -COLOR_YUV2GRAY_IYUV, COLOR_YUV2GRAY_I420, COLOR_YUV420sp2GRAY, COLOR_YUV420SP2GRAY, COLOR_YUV420p2GRAY, -COLOR_YUV420P2GRAY, COLOR_YUV2RGB_UYVY, COLOR_YUV2BGR_UYVY, COLOR_YUV2RGB_Y422, COLOR_YUV2BGR_Y422, COLOR_YUV2RGB_UYNV, -COLOR_YUV2BGR_UYNV, COLOR_YUV2RGBA_UYVY, COLOR_YUV2BGRA_UYVY, COLOR_YUV2RGBA_Y422, COLOR_YUV2BGRA_Y422, -COLOR_YUV2RGBA_UYNV, COLOR_YUV2BGRA_UYNV, COLOR_YUV2RGB_YUY2, COLOR_YUV2BGR_YUY2, COLOR_YUV2RGB_YVYU, -COLOR_YUV2BGR_YVYU, COLOR_YUV2RGB_YUYV, COLOR_YUV2BGR_YUYV, COLOR_YUV2RGB_YUNV, COLOR_YUV2BGR_YUNV, COLOR_YUV2RGBA_YUY2, -COLOR_YUV2BGRA_YUY2, COLOR_YUV2RGBA_YVYU, COLOR_YUV2BGRA_YVYU, COLOR_YUV2RGBA_YUYV, COLOR_YUV2BGRA_YUYV, -COLOR_YUV2RGBA_YUNV, COLOR_YUV2BGRA_YUNV, COLOR_YUV2GRAY_UYVY, COLOR_YUV2GRAY_YUY2, COLOR_YUV2GRAY_Y422, -COLOR_YUV2GRAY_UYNV, COLOR_YUV2GRAY_YVYU, COLOR_YUV2GRAY_YUYV, COLOR_YUV2GRAY_YUNV, COLOR_RGBA2mRGBA, COLOR_RGBA2M_RGBA, -COLOR_mRGBA2RGBA, COLOR_M_RGBA2RGBA, COLOR_RGB2YUV_I420, COLOR_BGR2YUV_I420, COLOR_RGB2YUV_IYUV, COLOR_BGR2YUV_IYUV, -COLOR_RGBA2YUV_I420, COLOR_BGRA2YUV_I420, COLOR_RGBA2YUV_IYUV, COLOR_BGRA2YUV_IYUV, COLOR_RGB2YUV_YV12, -COLOR_BGR2YUV_YV12, COLOR_RGBA2YUV_YV12, COLOR_BGRA2YUV_YV12, COLOR_BayerBG2BGR, COLOR_BAYER_BG2BGR, COLOR_BayerGB2BGR, -COLOR_BAYER_GB2BGR, COLOR_BayerRG2BGR, COLOR_BAYER_RG2BGR, COLOR_BayerGR2BGR, COLOR_BAYER_GR2BGR, COLOR_BayerRGGB2BGR, -COLOR_BAYER_RGGB2BGR, COLOR_BayerGRBG2BGR, COLOR_BAYER_GRBG2BGR, COLOR_BayerBGGR2BGR, COLOR_BAYER_BGGR2BGR, -COLOR_BayerGBRG2BGR, COLOR_BAYER_GBRG2BGR, COLOR_BayerRGGB2RGB, COLOR_BAYER_RGGB2RGB, COLOR_BayerGRBG2RGB, -COLOR_BAYER_GRBG2RGB, COLOR_BayerBGGR2RGB, COLOR_BAYER_BGGR2RGB, COLOR_BayerGBRG2RGB, COLOR_BAYER_GBRG2RGB, -COLOR_BayerBG2RGB, COLOR_BAYER_BG2RGB, COLOR_BayerGB2RGB, COLOR_BAYER_GB2RGB, COLOR_BayerRG2RGB, COLOR_BAYER_RG2RGB, -COLOR_BayerGR2RGB, COLOR_BAYER_GR2RGB, COLOR_BayerBG2GRAY, COLOR_BAYER_BG2GRAY, COLOR_BayerGB2GRAY, COLOR_BAYER_GB2GRAY, -COLOR_BayerRG2GRAY, COLOR_BAYER_RG2GRAY, COLOR_BayerGR2GRAY, COLOR_BAYER_GR2GRAY, COLOR_BayerRGGB2GRAY, -COLOR_BAYER_RGGB2GRAY, COLOR_BayerGRBG2GRAY, COLOR_BAYER_GRBG2GRAY, COLOR_BayerBGGR2GRAY, COLOR_BAYER_BGGR2GRAY, -COLOR_BayerGBRG2GRAY, COLOR_BAYER_GBRG2GRAY, COLOR_BayerBG2BGR_VNG, COLOR_BAYER_BG2BGR_VNG, COLOR_BayerGB2BGR_VNG, -COLOR_BAYER_GB2BGR_VNG, COLOR_BayerRG2BGR_VNG, COLOR_BAYER_RG2BGR_VNG, COLOR_BayerGR2BGR_VNG, COLOR_BAYER_GR2BGR_VNG, -COLOR_BayerRGGB2BGR_VNG, COLOR_BAYER_RGGB2BGR_VNG, COLOR_BayerGRBG2BGR_VNG, COLOR_BAYER_GRBG2BGR_VNG, -COLOR_BayerBGGR2BGR_VNG, COLOR_BAYER_BGGR2BGR_VNG, COLOR_BayerGBRG2BGR_VNG, COLOR_BAYER_GBRG2BGR_VNG, -COLOR_BayerRGGB2RGB_VNG, COLOR_BAYER_RGGB2RGB_VNG, COLOR_BayerGRBG2RGB_VNG, COLOR_BAYER_GRBG2RGB_VNG, -COLOR_BayerBGGR2RGB_VNG, COLOR_BAYER_BGGR2RGB_VNG, COLOR_BayerGBRG2RGB_VNG, COLOR_BAYER_GBRG2RGB_VNG, -COLOR_BayerBG2RGB_VNG, COLOR_BAYER_BG2RGB_VNG, COLOR_BayerGB2RGB_VNG, COLOR_BAYER_GB2RGB_VNG, COLOR_BayerRG2RGB_VNG, -COLOR_BAYER_RG2RGB_VNG, COLOR_BayerGR2RGB_VNG, COLOR_BAYER_GR2RGB_VNG, COLOR_BayerBG2BGR_EA, COLOR_BAYER_BG2BGR_EA, -COLOR_BayerGB2BGR_EA, COLOR_BAYER_GB2BGR_EA, COLOR_BayerRG2BGR_EA, COLOR_BAYER_RG2BGR_EA, COLOR_BayerGR2BGR_EA, -COLOR_BAYER_GR2BGR_EA, COLOR_BayerRGGB2BGR_EA, COLOR_BAYER_RGGB2BGR_EA, COLOR_BayerGRBG2BGR_EA, COLOR_BAYER_GRBG2BGR_EA, -COLOR_BayerBGGR2BGR_EA, COLOR_BAYER_BGGR2BGR_EA, COLOR_BayerGBRG2BGR_EA, COLOR_BAYER_GBRG2BGR_EA, -COLOR_BayerRGGB2RGB_EA, COLOR_BAYER_RGGB2RGB_EA, COLOR_BayerGRBG2RGB_EA, COLOR_BAYER_GRBG2RGB_EA, -COLOR_BayerBGGR2RGB_EA, COLOR_BAYER_BGGR2RGB_EA, COLOR_BayerGBRG2RGB_EA, COLOR_BAYER_GBRG2RGB_EA, COLOR_BayerBG2RGB_EA, -COLOR_BAYER_BG2RGB_EA, COLOR_BayerGB2RGB_EA, COLOR_BAYER_GB2RGB_EA, COLOR_BayerRG2RGB_EA, COLOR_BAYER_RG2RGB_EA, -COLOR_BayerGR2RGB_EA, COLOR_BAYER_GR2RGB_EA, COLOR_BayerBG2BGRA, COLOR_BAYER_BG2BGRA, COLOR_BayerGB2BGRA, -COLOR_BAYER_GB2BGRA, COLOR_BayerRG2BGRA, COLOR_BAYER_RG2BGRA, COLOR_BayerGR2BGRA, COLOR_BAYER_GR2BGRA, -COLOR_BayerRGGB2BGRA, COLOR_BAYER_RGGB2BGRA, COLOR_BayerGRBG2BGRA, COLOR_BAYER_GRBG2BGRA, COLOR_BayerBGGR2BGRA, -COLOR_BAYER_BGGR2BGRA, COLOR_BayerGBRG2BGRA, COLOR_BAYER_GBRG2BGRA, COLOR_BayerRGGB2RGBA, COLOR_BAYER_RGGB2RGBA, -COLOR_BayerGRBG2RGBA, COLOR_BAYER_GRBG2RGBA, COLOR_BayerBGGR2RGBA, COLOR_BAYER_BGGR2RGBA, COLOR_BayerGBRG2RGBA, -COLOR_BAYER_GBRG2RGBA, COLOR_BayerBG2RGBA, COLOR_BAYER_BG2RGBA, COLOR_BayerGB2RGBA, COLOR_BAYER_GB2RGBA, -COLOR_BayerRG2RGBA, COLOR_BAYER_RG2RGBA, COLOR_BayerGR2RGBA, COLOR_BAYER_GR2RGBA, COLOR_COLORCVT_MAX]""" - -INTERSECT_NONE: int -INTERSECT_PARTIAL: int -INTERSECT_FULL: int -RectanglesIntersectTypes = int -"""One of [INTERSECT_NONE, INTERSECT_PARTIAL, INTERSECT_FULL]""" - -FILLED: int -LINE_4: int -LINE_8: int -LINE_AA: int -LineTypes = int -"""One of [FILLED, LINE_4, LINE_8, LINE_AA]""" - -FONT_HERSHEY_SIMPLEX: int -FONT_HERSHEY_PLAIN: int -FONT_HERSHEY_DUPLEX: int -FONT_HERSHEY_COMPLEX: int -FONT_HERSHEY_TRIPLEX: int -FONT_HERSHEY_COMPLEX_SMALL: int -FONT_HERSHEY_SCRIPT_SIMPLEX: int -FONT_HERSHEY_SCRIPT_COMPLEX: int -FONT_ITALIC: int -HersheyFonts = int -"""One of [FONT_HERSHEY_SIMPLEX, FONT_HERSHEY_PLAIN, FONT_HERSHEY_DUPLEX, FONT_HERSHEY_COMPLEX, FONT_HERSHEY_TRIPLEX, -FONT_HERSHEY_COMPLEX_SMALL, FONT_HERSHEY_SCRIPT_SIMPLEX, FONT_HERSHEY_SCRIPT_COMPLEX, FONT_ITALIC]""" - -MARKER_CROSS: int -MARKER_TILTED_CROSS: int -MARKER_STAR: int -MARKER_DIAMOND: int -MARKER_SQUARE: int -MARKER_TRIANGLE_UP: int -MARKER_TRIANGLE_DOWN: int -MarkerTypes = int -"""One of [MARKER_CROSS, MARKER_TILTED_CROSS, MARKER_STAR, MARKER_DIAMOND, MARKER_SQUARE, MARKER_TRIANGLE_UP, -MARKER_TRIANGLE_DOWN]""" - -TM_SQDIFF: int -TM_SQDIFF_NORMED: int -TM_CCORR: int -TM_CCORR_NORMED: int -TM_CCOEFF: int -TM_CCOEFF_NORMED: int -TemplateMatchModes = int -"""One of [TM_SQDIFF, TM_SQDIFF_NORMED, TM_CCORR, TM_CCORR_NORMED, TM_CCOEFF, TM_CCOEFF_NORMED]""" - -COLORMAP_AUTUMN: int -COLORMAP_BONE: int -COLORMAP_JET: int -COLORMAP_WINTER: int -COLORMAP_RAINBOW: int -COLORMAP_OCEAN: int -COLORMAP_SUMMER: int -COLORMAP_SPRING: int -COLORMAP_COOL: int -COLORMAP_HSV: int -COLORMAP_PINK: int -COLORMAP_HOT: int -COLORMAP_PARULA: int -COLORMAP_MAGMA: int -COLORMAP_INFERNO: int -COLORMAP_PLASMA: int -COLORMAP_VIRIDIS: int -COLORMAP_CIVIDIS: int -COLORMAP_TWILIGHT: int -COLORMAP_TWILIGHT_SHIFTED: int -COLORMAP_TURBO: int -COLORMAP_DEEPGREEN: int -ColormapTypes = int -"""One of [COLORMAP_AUTUMN, COLORMAP_BONE, COLORMAP_JET, COLORMAP_WINTER, COLORMAP_RAINBOW, COLORMAP_OCEAN, -COLORMAP_SUMMER, COLORMAP_SPRING, COLORMAP_COOL, COLORMAP_HSV, COLORMAP_PINK, COLORMAP_HOT, COLORMAP_PARULA, -COLORMAP_MAGMA, COLORMAP_INFERNO, COLORMAP_PLASMA, COLORMAP_VIRIDIS, COLORMAP_CIVIDIS, COLORMAP_TWILIGHT, -COLORMAP_TWILIGHT_SHIFTED, COLORMAP_TURBO, COLORMAP_DEEPGREEN]""" - -INPAINT_NS: int -INPAINT_TELEA: int -LDR_SIZE: int -NORMAL_CLONE: int -MIXED_CLONE: int -MONOCHROME_TRANSFER: int -RECURS_FILTER: int -NORMCONV_FILTER: int -CAP_PROP_DC1394_OFF: int -CAP_PROP_DC1394_MODE_MANUAL: int -CAP_PROP_DC1394_MODE_AUTO: int -CAP_PROP_DC1394_MODE_ONE_PUSH_AUTO: int -CAP_PROP_DC1394_MAX: int -CAP_OPENNI_DEPTH_GENERATOR: int -CAP_OPENNI_IMAGE_GENERATOR: int -CAP_OPENNI_IR_GENERATOR: int -CAP_OPENNI_GENERATORS_MASK: int -CAP_PROP_OPENNI_OUTPUT_MODE: int -CAP_PROP_OPENNI_FRAME_MAX_DEPTH: int -CAP_PROP_OPENNI_BASELINE: int -CAP_PROP_OPENNI_FOCAL_LENGTH: int -CAP_PROP_OPENNI_REGISTRATION: int -CAP_PROP_OPENNI_REGISTRATION_ON: int -CAP_PROP_OPENNI_APPROX_FRAME_SYNC: int -CAP_PROP_OPENNI_MAX_BUFFER_SIZE: int -CAP_PROP_OPENNI_CIRCLE_BUFFER: int -CAP_PROP_OPENNI_MAX_TIME_DURATION: int -CAP_PROP_OPENNI_GENERATOR_PRESENT: int -CAP_PROP_OPENNI2_SYNC: int -CAP_PROP_OPENNI2_MIRROR: int -CAP_OPENNI_IMAGE_GENERATOR_PRESENT: int -CAP_OPENNI_IMAGE_GENERATOR_OUTPUT_MODE: int -CAP_OPENNI_DEPTH_GENERATOR_PRESENT: int -CAP_OPENNI_DEPTH_GENERATOR_BASELINE: int -CAP_OPENNI_DEPTH_GENERATOR_FOCAL_LENGTH: int -CAP_OPENNI_DEPTH_GENERATOR_REGISTRATION: int -CAP_OPENNI_DEPTH_GENERATOR_REGISTRATION_ON: int -CAP_OPENNI_IR_GENERATOR_PRESENT: int -CAP_OPENNI_DEPTH_MAP: int -CAP_OPENNI_POINT_CLOUD_MAP: int -CAP_OPENNI_DISPARITY_MAP: int -CAP_OPENNI_DISPARITY_MAP_32F: int -CAP_OPENNI_VALID_DEPTH_MASK: int -CAP_OPENNI_BGR_IMAGE: int -CAP_OPENNI_GRAY_IMAGE: int -CAP_OPENNI_IR_IMAGE: int -CAP_OPENNI_VGA_30HZ: int -CAP_OPENNI_SXGA_15HZ: int -CAP_OPENNI_SXGA_30HZ: int -CAP_OPENNI_QVGA_30HZ: int -CAP_OPENNI_QVGA_60HZ: int -CAP_PROP_GSTREAMER_QUEUE_LENGTH: int -CAP_PROP_PVAPI_MULTICASTIP: int -CAP_PROP_PVAPI_FRAMESTARTTRIGGERMODE: int -CAP_PROP_PVAPI_DECIMATIONHORIZONTAL: int -CAP_PROP_PVAPI_DECIMATIONVERTICAL: int -CAP_PROP_PVAPI_BINNINGX: int -CAP_PROP_PVAPI_BINNINGY: int -CAP_PROP_PVAPI_PIXELFORMAT: int -CAP_PVAPI_FSTRIGMODE_FREERUN: int -CAP_PVAPI_FSTRIGMODE_SYNCIN1: int -CAP_PVAPI_FSTRIGMODE_SYNCIN2: int -CAP_PVAPI_FSTRIGMODE_FIXEDRATE: int -CAP_PVAPI_FSTRIGMODE_SOFTWARE: int -CAP_PVAPI_DECIMATION_OFF: int -CAP_PVAPI_DECIMATION_2OUTOF4: int -CAP_PVAPI_DECIMATION_2OUTOF8: int -CAP_PVAPI_DECIMATION_2OUTOF16: int -CAP_PVAPI_PIXELFORMAT_MONO8: int -CAP_PVAPI_PIXELFORMAT_MONO16: int -CAP_PVAPI_PIXELFORMAT_BAYER8: int -CAP_PVAPI_PIXELFORMAT_BAYER16: int -CAP_PVAPI_PIXELFORMAT_RGB24: int -CAP_PVAPI_PIXELFORMAT_BGR24: int -CAP_PVAPI_PIXELFORMAT_RGBA32: int -CAP_PVAPI_PIXELFORMAT_BGRA32: int -CAP_PROP_XI_DOWNSAMPLING: int -CAP_PROP_XI_DATA_FORMAT: int -CAP_PROP_XI_OFFSET_X: int -CAP_PROP_XI_OFFSET_Y: int -CAP_PROP_XI_TRG_SOURCE: int -CAP_PROP_XI_TRG_SOFTWARE: int -CAP_PROP_XI_GPI_SELECTOR: int -CAP_PROP_XI_GPI_MODE: int -CAP_PROP_XI_GPI_LEVEL: int -CAP_PROP_XI_GPO_SELECTOR: int -CAP_PROP_XI_GPO_MODE: int -CAP_PROP_XI_LED_SELECTOR: int -CAP_PROP_XI_LED_MODE: int -CAP_PROP_XI_MANUAL_WB: int -CAP_PROP_XI_AUTO_WB: int -CAP_PROP_XI_AEAG: int -CAP_PROP_XI_EXP_PRIORITY: int -CAP_PROP_XI_AE_MAX_LIMIT: int -CAP_PROP_XI_AG_MAX_LIMIT: int -CAP_PROP_XI_AEAG_LEVEL: int -CAP_PROP_XI_TIMEOUT: int -CAP_PROP_XI_EXPOSURE: int -CAP_PROP_XI_EXPOSURE_BURST_COUNT: int -CAP_PROP_XI_GAIN_SELECTOR: int -CAP_PROP_XI_GAIN: int -CAP_PROP_XI_DOWNSAMPLING_TYPE: int -CAP_PROP_XI_BINNING_SELECTOR: int -CAP_PROP_XI_BINNING_VERTICAL: int -CAP_PROP_XI_BINNING_HORIZONTAL: int -CAP_PROP_XI_BINNING_PATTERN: int -CAP_PROP_XI_DECIMATION_SELECTOR: int -CAP_PROP_XI_DECIMATION_VERTICAL: int -CAP_PROP_XI_DECIMATION_HORIZONTAL: int -CAP_PROP_XI_DECIMATION_PATTERN: int -CAP_PROP_XI_TEST_PATTERN_GENERATOR_SELECTOR: int -CAP_PROP_XI_TEST_PATTERN: int -CAP_PROP_XI_IMAGE_DATA_FORMAT: int -CAP_PROP_XI_SHUTTER_TYPE: int -CAP_PROP_XI_SENSOR_TAPS: int -CAP_PROP_XI_AEAG_ROI_OFFSET_X: int -CAP_PROP_XI_AEAG_ROI_OFFSET_Y: int -CAP_PROP_XI_AEAG_ROI_WIDTH: int -CAP_PROP_XI_AEAG_ROI_HEIGHT: int -CAP_PROP_XI_BPC: int -CAP_PROP_XI_WB_KR: int -CAP_PROP_XI_WB_KG: int -CAP_PROP_XI_WB_KB: int -CAP_PROP_XI_WIDTH: int -CAP_PROP_XI_HEIGHT: int -CAP_PROP_XI_REGION_SELECTOR: int -CAP_PROP_XI_REGION_MODE: int -CAP_PROP_XI_LIMIT_BANDWIDTH: int -CAP_PROP_XI_SENSOR_DATA_BIT_DEPTH: int -CAP_PROP_XI_OUTPUT_DATA_BIT_DEPTH: int -CAP_PROP_XI_IMAGE_DATA_BIT_DEPTH: int -CAP_PROP_XI_OUTPUT_DATA_PACKING: int -CAP_PROP_XI_OUTPUT_DATA_PACKING_TYPE: int -CAP_PROP_XI_IS_COOLED: int -CAP_PROP_XI_COOLING: int -CAP_PROP_XI_TARGET_TEMP: int -CAP_PROP_XI_CHIP_TEMP: int -CAP_PROP_XI_HOUS_TEMP: int -CAP_PROP_XI_HOUS_BACK_SIDE_TEMP: int -CAP_PROP_XI_SENSOR_BOARD_TEMP: int -CAP_PROP_XI_CMS: int -CAP_PROP_XI_APPLY_CMS: int -CAP_PROP_XI_IMAGE_IS_COLOR: int -CAP_PROP_XI_COLOR_FILTER_ARRAY: int -CAP_PROP_XI_GAMMAY: int -CAP_PROP_XI_GAMMAC: int -CAP_PROP_XI_SHARPNESS: int -CAP_PROP_XI_CC_MATRIX_00: int -CAP_PROP_XI_CC_MATRIX_01: int -CAP_PROP_XI_CC_MATRIX_02: int -CAP_PROP_XI_CC_MATRIX_03: int -CAP_PROP_XI_CC_MATRIX_10: int -CAP_PROP_XI_CC_MATRIX_11: int -CAP_PROP_XI_CC_MATRIX_12: int -CAP_PROP_XI_CC_MATRIX_13: int -CAP_PROP_XI_CC_MATRIX_20: int -CAP_PROP_XI_CC_MATRIX_21: int -CAP_PROP_XI_CC_MATRIX_22: int -CAP_PROP_XI_CC_MATRIX_23: int -CAP_PROP_XI_CC_MATRIX_30: int -CAP_PROP_XI_CC_MATRIX_31: int -CAP_PROP_XI_CC_MATRIX_32: int -CAP_PROP_XI_CC_MATRIX_33: int -CAP_PROP_XI_DEFAULT_CC_MATRIX: int -CAP_PROP_XI_TRG_SELECTOR: int -CAP_PROP_XI_ACQ_FRAME_BURST_COUNT: int -CAP_PROP_XI_DEBOUNCE_EN: int -CAP_PROP_XI_DEBOUNCE_T0: int -CAP_PROP_XI_DEBOUNCE_T1: int -CAP_PROP_XI_DEBOUNCE_POL: int -CAP_PROP_XI_LENS_MODE: int -CAP_PROP_XI_LENS_APERTURE_VALUE: int -CAP_PROP_XI_LENS_FOCUS_MOVEMENT_VALUE: int -CAP_PROP_XI_LENS_FOCUS_MOVE: int -CAP_PROP_XI_LENS_FOCUS_DISTANCE: int -CAP_PROP_XI_LENS_FOCAL_LENGTH: int -CAP_PROP_XI_LENS_FEATURE_SELECTOR: int -CAP_PROP_XI_LENS_FEATURE: int -CAP_PROP_XI_DEVICE_MODEL_ID: int -CAP_PROP_XI_DEVICE_SN: int -CAP_PROP_XI_IMAGE_DATA_FORMAT_RGB32_ALPHA: int -CAP_PROP_XI_IMAGE_PAYLOAD_SIZE: int -CAP_PROP_XI_TRANSPORT_PIXEL_FORMAT: int -CAP_PROP_XI_SENSOR_CLOCK_FREQ_HZ: int -CAP_PROP_XI_SENSOR_CLOCK_FREQ_INDEX: int -CAP_PROP_XI_SENSOR_OUTPUT_CHANNEL_COUNT: int -CAP_PROP_XI_FRAMERATE: int -CAP_PROP_XI_COUNTER_SELECTOR: int -CAP_PROP_XI_COUNTER_VALUE: int -CAP_PROP_XI_ACQ_TIMING_MODE: int -CAP_PROP_XI_AVAILABLE_BANDWIDTH: int -CAP_PROP_XI_BUFFER_POLICY: int -CAP_PROP_XI_LUT_EN: int -CAP_PROP_XI_LUT_INDEX: int -CAP_PROP_XI_LUT_VALUE: int -CAP_PROP_XI_TRG_DELAY: int -CAP_PROP_XI_TS_RST_MODE: int -CAP_PROP_XI_TS_RST_SOURCE: int -CAP_PROP_XI_IS_DEVICE_EXIST: int -CAP_PROP_XI_ACQ_BUFFER_SIZE: int -CAP_PROP_XI_ACQ_BUFFER_SIZE_UNIT: int -CAP_PROP_XI_ACQ_TRANSPORT_BUFFER_SIZE: int -CAP_PROP_XI_BUFFERS_QUEUE_SIZE: int -CAP_PROP_XI_ACQ_TRANSPORT_BUFFER_COMMIT: int -CAP_PROP_XI_RECENT_FRAME: int -CAP_PROP_XI_DEVICE_RESET: int -CAP_PROP_XI_COLUMN_FPN_CORRECTION: int -CAP_PROP_XI_ROW_FPN_CORRECTION: int -CAP_PROP_XI_SENSOR_MODE: int -CAP_PROP_XI_HDR: int -CAP_PROP_XI_HDR_KNEEPOINT_COUNT: int -CAP_PROP_XI_HDR_T1: int -CAP_PROP_XI_HDR_T2: int -CAP_PROP_XI_KNEEPOINT1: int -CAP_PROP_XI_KNEEPOINT2: int -CAP_PROP_XI_IMAGE_BLACK_LEVEL: int -CAP_PROP_XI_HW_REVISION: int -CAP_PROP_XI_DEBUG_LEVEL: int -CAP_PROP_XI_AUTO_BANDWIDTH_CALCULATION: int -CAP_PROP_XI_FFS_FILE_ID: int -CAP_PROP_XI_FFS_FILE_SIZE: int -CAP_PROP_XI_FREE_FFS_SIZE: int -CAP_PROP_XI_USED_FFS_SIZE: int -CAP_PROP_XI_FFS_ACCESS_KEY: int -CAP_PROP_XI_SENSOR_FEATURE_SELECTOR: int -CAP_PROP_XI_SENSOR_FEATURE_VALUE: int -CAP_PROP_ARAVIS_AUTOTRIGGER: int -CAP_PROP_IOS_DEVICE_FOCUS: int -CAP_PROP_IOS_DEVICE_EXPOSURE: int -CAP_PROP_IOS_DEVICE_FLASH: int -CAP_PROP_IOS_DEVICE_WHITEBALANCE: int -CAP_PROP_IOS_DEVICE_TORCH: int -CAP_PROP_GIGA_FRAME_OFFSET_X: int -CAP_PROP_GIGA_FRAME_OFFSET_Y: int -CAP_PROP_GIGA_FRAME_WIDTH_MAX: int -CAP_PROP_GIGA_FRAME_HEIGH_MAX: int -CAP_PROP_GIGA_FRAME_SENS_WIDTH: int -CAP_PROP_GIGA_FRAME_SENS_HEIGH: int -CAP_PROP_INTELPERC_PROFILE_COUNT: int -CAP_PROP_INTELPERC_PROFILE_IDX: int -CAP_PROP_INTELPERC_DEPTH_LOW_CONFIDENCE_VALUE: int -CAP_PROP_INTELPERC_DEPTH_SATURATION_VALUE: int -CAP_PROP_INTELPERC_DEPTH_CONFIDENCE_THRESHOLD: int -CAP_PROP_INTELPERC_DEPTH_FOCAL_LENGTH_HORZ: int -CAP_PROP_INTELPERC_DEPTH_FOCAL_LENGTH_VERT: int -CAP_INTELPERC_DEPTH_GENERATOR: int -CAP_INTELPERC_IMAGE_GENERATOR: int -CAP_INTELPERC_IR_GENERATOR: int -CAP_INTELPERC_GENERATORS_MASK: int -CAP_INTELPERC_DEPTH_MAP: int -CAP_INTELPERC_UVDEPTH_MAP: int -CAP_INTELPERC_IR_MAP: int -CAP_INTELPERC_IMAGE: int -CAP_PROP_GPHOTO2_PREVIEW: int -CAP_PROP_GPHOTO2_WIDGET_ENUMERATE: int -CAP_PROP_GPHOTO2_RELOAD_CONFIG: int -CAP_PROP_GPHOTO2_RELOAD_ON_CHANGE: int -CAP_PROP_GPHOTO2_COLLECT_MSGS: int -CAP_PROP_GPHOTO2_FLUSH_MSGS: int -CAP_PROP_SPEED: int -CAP_PROP_APERTURE: int -CAP_PROP_EXPOSUREPROGRAM: int -CAP_PROP_VIEWFINDER: int -CAP_PROP_IMAGES_BASE: int -CAP_PROP_IMAGES_LAST: int -LMEDS: int -RANSAC: int -RHO: int -USAC_DEFAULT: int -USAC_PARALLEL: int -USAC_FM_8PTS: int -USAC_FAST: int -USAC_ACCURATE: int -USAC_PROSAC: int -USAC_MAGSAC: int -CALIB_CB_ADAPTIVE_THRESH: int -CALIB_CB_NORMALIZE_IMAGE: int -CALIB_CB_FILTER_QUADS: int -CALIB_CB_FAST_CHECK: int -CALIB_CB_EXHAUSTIVE: int -CALIB_CB_ACCURACY: int -CALIB_CB_LARGER: int -CALIB_CB_MARKER: int -CALIB_CB_SYMMETRIC_GRID: int -CALIB_CB_ASYMMETRIC_GRID: int -CALIB_CB_CLUSTERING: int -CALIB_NINTRINSIC: int -CALIB_USE_INTRINSIC_GUESS: int -CALIB_FIX_ASPECT_RATIO: int -CALIB_FIX_PRINCIPAL_POINT: int -CALIB_ZERO_TANGENT_DIST: int -CALIB_FIX_FOCAL_LENGTH: int -CALIB_FIX_K1: int -CALIB_FIX_K2: int -CALIB_FIX_K3: int -CALIB_FIX_K4: int -CALIB_FIX_K5: int -CALIB_FIX_K6: int -CALIB_RATIONAL_MODEL: int -CALIB_THIN_PRISM_MODEL: int -CALIB_FIX_S1_S2_S3_S4: int -CALIB_TILTED_MODEL: int -CALIB_FIX_TAUX_TAUY: int -CALIB_USE_QR: int -CALIB_FIX_TANGENT_DIST: int -CALIB_FIX_INTRINSIC: int -CALIB_SAME_FOCAL_LENGTH: int -CALIB_ZERO_DISPARITY: int -CALIB_USE_LU: int -CALIB_USE_EXTRINSIC_GUESS: int -FM_7POINT: int -FM_8POINT: int -FM_LMEDS: int -FM_RANSAC: int -CASCADE_DO_CANNY_PRUNING: int -CASCADE_SCALE_IMAGE: int -CASCADE_FIND_BIGGEST_OBJECT: int -CASCADE_DO_ROUGH_SEARCH: int -OPTFLOW_USE_INITIAL_FLOW: int -OPTFLOW_LK_GET_MIN_EIGENVALS: int -OPTFLOW_FARNEBACK_GAUSSIAN: int -MOTION_TRANSLATION: int -MOTION_EUCLIDEAN: int -MOTION_AFFINE: int -MOTION_HOMOGRAPHY: int - -DrawMatchesFlags_DEFAULT: int -DRAW_MATCHES_FLAGS_DEFAULT: int -DrawMatchesFlags_DRAW_OVER_OUTIMG: int -DRAW_MATCHES_FLAGS_DRAW_OVER_OUTIMG: int -DrawMatchesFlags_NOT_DRAW_SINGLE_POINTS: int -DRAW_MATCHES_FLAGS_NOT_DRAW_SINGLE_POINTS: int -DrawMatchesFlags_DRAW_RICH_KEYPOINTS: int -DRAW_MATCHES_FLAGS_DRAW_RICH_KEYPOINTS: int -DrawMatchesFlags = int -"""One of [DrawMatchesFlags_DEFAULT, DRAW_MATCHES_FLAGS_DEFAULT, DrawMatchesFlags_DRAW_OVER_OUTIMG, -DRAW_MATCHES_FLAGS_DRAW_OVER_OUTIMG, DrawMatchesFlags_NOT_DRAW_SINGLE_POINTS, DRAW_MATCHES_FLAGS_NOT_DRAW_SINGLE_POINTS, -DrawMatchesFlags_DRAW_RICH_KEYPOINTS, DRAW_MATCHES_FLAGS_DRAW_RICH_KEYPOINTS]""" - -IMREAD_UNCHANGED: int -IMREAD_GRAYSCALE: int -IMREAD_COLOR: int -IMREAD_ANYDEPTH: int -IMREAD_ANYCOLOR: int -IMREAD_LOAD_GDAL: int -IMREAD_REDUCED_GRAYSCALE_2: int -IMREAD_REDUCED_COLOR_2: int -IMREAD_REDUCED_GRAYSCALE_4: int -IMREAD_REDUCED_COLOR_4: int -IMREAD_REDUCED_GRAYSCALE_8: int -IMREAD_REDUCED_COLOR_8: int -IMREAD_IGNORE_ORIENTATION: int -ImreadModes = int -"""One of [IMREAD_UNCHANGED, IMREAD_GRAYSCALE, IMREAD_COLOR, IMREAD_ANYDEPTH, IMREAD_ANYCOLOR, IMREAD_LOAD_GDAL, -IMREAD_REDUCED_GRAYSCALE_2, IMREAD_REDUCED_COLOR_2, IMREAD_REDUCED_GRAYSCALE_4, IMREAD_REDUCED_COLOR_4, -IMREAD_REDUCED_GRAYSCALE_8, IMREAD_REDUCED_COLOR_8, IMREAD_IGNORE_ORIENTATION]""" - -IMWRITE_JPEG_QUALITY: int -IMWRITE_JPEG_PROGRESSIVE: int -IMWRITE_JPEG_OPTIMIZE: int -IMWRITE_JPEG_RST_INTERVAL: int -IMWRITE_JPEG_LUMA_QUALITY: int -IMWRITE_JPEG_CHROMA_QUALITY: int -IMWRITE_JPEG_SAMPLING_FACTOR: int -IMWRITE_PNG_COMPRESSION: int -IMWRITE_PNG_STRATEGY: int -IMWRITE_PNG_BILEVEL: int -IMWRITE_PXM_BINARY: int -IMWRITE_EXR_TYPE: int -IMWRITE_EXR_COMPRESSION: int -IMWRITE_EXR_DWA_COMPRESSION_LEVEL: int -IMWRITE_WEBP_QUALITY: int -IMWRITE_HDR_COMPRESSION: int -IMWRITE_PAM_TUPLETYPE: int -IMWRITE_TIFF_RESUNIT: int -IMWRITE_TIFF_XDPI: int -IMWRITE_TIFF_YDPI: int -IMWRITE_TIFF_COMPRESSION: int -IMWRITE_JPEG2000_COMPRESSION_X1000: int -IMWRITE_AVIF_QUALITY: int -IMWRITE_AVIF_DEPTH: int -IMWRITE_AVIF_SPEED: int -ImwriteFlags = int -"""One of [IMWRITE_JPEG_QUALITY, IMWRITE_JPEG_PROGRESSIVE, IMWRITE_JPEG_OPTIMIZE, IMWRITE_JPEG_RST_INTERVAL, -IMWRITE_JPEG_LUMA_QUALITY, IMWRITE_JPEG_CHROMA_QUALITY, IMWRITE_JPEG_SAMPLING_FACTOR, IMWRITE_PNG_COMPRESSION, -IMWRITE_PNG_STRATEGY, IMWRITE_PNG_BILEVEL, IMWRITE_PXM_BINARY, IMWRITE_EXR_TYPE, IMWRITE_EXR_COMPRESSION, -IMWRITE_EXR_DWA_COMPRESSION_LEVEL, IMWRITE_WEBP_QUALITY, IMWRITE_HDR_COMPRESSION, IMWRITE_PAM_TUPLETYPE, -IMWRITE_TIFF_RESUNIT, IMWRITE_TIFF_XDPI, IMWRITE_TIFF_YDPI, IMWRITE_TIFF_COMPRESSION, -IMWRITE_JPEG2000_COMPRESSION_X1000, IMWRITE_AVIF_QUALITY, IMWRITE_AVIF_DEPTH, IMWRITE_AVIF_SPEED]""" - -IMWRITE_JPEG_SAMPLING_FACTOR_411: int -IMWRITE_JPEG_SAMPLING_FACTOR_420: int -IMWRITE_JPEG_SAMPLING_FACTOR_422: int -IMWRITE_JPEG_SAMPLING_FACTOR_440: int -IMWRITE_JPEG_SAMPLING_FACTOR_444: int -ImwriteJPEGSamplingFactorParams = int -"""One of [IMWRITE_JPEG_SAMPLING_FACTOR_411, IMWRITE_JPEG_SAMPLING_FACTOR_420, IMWRITE_JPEG_SAMPLING_FACTOR_422, -IMWRITE_JPEG_SAMPLING_FACTOR_440, IMWRITE_JPEG_SAMPLING_FACTOR_444]""" - -IMWRITE_EXR_TYPE_HALF: int -IMWRITE_EXR_TYPE_FLOAT: int -ImwriteEXRTypeFlags = int -"""One of [IMWRITE_EXR_TYPE_HALF, IMWRITE_EXR_TYPE_FLOAT]""" - -IMWRITE_EXR_COMPRESSION_NO: int -IMWRITE_EXR_COMPRESSION_RLE: int -IMWRITE_EXR_COMPRESSION_ZIPS: int -IMWRITE_EXR_COMPRESSION_ZIP: int -IMWRITE_EXR_COMPRESSION_PIZ: int -IMWRITE_EXR_COMPRESSION_PXR24: int -IMWRITE_EXR_COMPRESSION_B44: int -IMWRITE_EXR_COMPRESSION_B44A: int -IMWRITE_EXR_COMPRESSION_DWAA: int -IMWRITE_EXR_COMPRESSION_DWAB: int -ImwriteEXRCompressionFlags = int -"""One of [IMWRITE_EXR_COMPRESSION_NO, IMWRITE_EXR_COMPRESSION_RLE, IMWRITE_EXR_COMPRESSION_ZIPS, -IMWRITE_EXR_COMPRESSION_ZIP, IMWRITE_EXR_COMPRESSION_PIZ, IMWRITE_EXR_COMPRESSION_PXR24, IMWRITE_EXR_COMPRESSION_B44, -IMWRITE_EXR_COMPRESSION_B44A, IMWRITE_EXR_COMPRESSION_DWAA, IMWRITE_EXR_COMPRESSION_DWAB]""" - -IMWRITE_PNG_STRATEGY_DEFAULT: int -IMWRITE_PNG_STRATEGY_FILTERED: int -IMWRITE_PNG_STRATEGY_HUFFMAN_ONLY: int -IMWRITE_PNG_STRATEGY_RLE: int -IMWRITE_PNG_STRATEGY_FIXED: int -ImwritePNGFlags = int -"""One of [IMWRITE_PNG_STRATEGY_DEFAULT, IMWRITE_PNG_STRATEGY_FILTERED, IMWRITE_PNG_STRATEGY_HUFFMAN_ONLY, -IMWRITE_PNG_STRATEGY_RLE, IMWRITE_PNG_STRATEGY_FIXED]""" - -IMWRITE_PAM_FORMAT_NULL: int -IMWRITE_PAM_FORMAT_BLACKANDWHITE: int -IMWRITE_PAM_FORMAT_GRAYSCALE: int -IMWRITE_PAM_FORMAT_GRAYSCALE_ALPHA: int -IMWRITE_PAM_FORMAT_RGB: int -IMWRITE_PAM_FORMAT_RGB_ALPHA: int -ImwritePAMFlags = int -"""One of [IMWRITE_PAM_FORMAT_NULL, IMWRITE_PAM_FORMAT_BLACKANDWHITE, IMWRITE_PAM_FORMAT_GRAYSCALE, -IMWRITE_PAM_FORMAT_GRAYSCALE_ALPHA, IMWRITE_PAM_FORMAT_RGB, IMWRITE_PAM_FORMAT_RGB_ALPHA]""" - -IMWRITE_HDR_COMPRESSION_NONE: int -IMWRITE_HDR_COMPRESSION_RLE: int -ImwriteHDRCompressionFlags = int -"""One of [IMWRITE_HDR_COMPRESSION_NONE, IMWRITE_HDR_COMPRESSION_RLE]""" - -CAP_ANY: int -CAP_VFW: int -CAP_V4L: int -CAP_V4L2: int -CAP_FIREWIRE: int -CAP_FIREWARE: int -CAP_IEEE1394: int -CAP_DC1394: int -CAP_CMU1394: int -CAP_QT: int -CAP_UNICAP: int -CAP_DSHOW: int -CAP_PVAPI: int -CAP_OPENNI: int -CAP_OPENNI_ASUS: int -CAP_ANDROID: int -CAP_XIAPI: int -CAP_AVFOUNDATION: int -CAP_GIGANETIX: int -CAP_MSMF: int -CAP_WINRT: int -CAP_INTELPERC: int -CAP_REALSENSE: int -CAP_OPENNI2: int -CAP_OPENNI2_ASUS: int -CAP_OPENNI2_ASTRA: int -CAP_GPHOTO2: int -CAP_GSTREAMER: int -CAP_FFMPEG: int -CAP_IMAGES: int -CAP_ARAVIS: int -CAP_OPENCV_MJPEG: int -CAP_INTEL_MFX: int -CAP_XINE: int -CAP_UEYE: int -CAP_OBSENSOR: int -VideoCaptureAPIs = int -"""One of [CAP_ANY, CAP_VFW, CAP_V4L, CAP_V4L2, CAP_FIREWIRE, CAP_FIREWARE, CAP_IEEE1394, CAP_DC1394, CAP_CMU1394, -CAP_QT, CAP_UNICAP, CAP_DSHOW, CAP_PVAPI, CAP_OPENNI, CAP_OPENNI_ASUS, CAP_ANDROID, CAP_XIAPI, CAP_AVFOUNDATION, -CAP_GIGANETIX, CAP_MSMF, CAP_WINRT, CAP_INTELPERC, CAP_REALSENSE, CAP_OPENNI2, CAP_OPENNI2_ASUS, CAP_OPENNI2_ASTRA, -CAP_GPHOTO2, CAP_GSTREAMER, CAP_FFMPEG, CAP_IMAGES, CAP_ARAVIS, CAP_OPENCV_MJPEG, CAP_INTEL_MFX, CAP_XINE, CAP_UEYE, -CAP_OBSENSOR]""" - -CAP_PROP_POS_MSEC: int -CAP_PROP_POS_FRAMES: int -CAP_PROP_POS_AVI_RATIO: int -CAP_PROP_FRAME_WIDTH: int -CAP_PROP_FRAME_HEIGHT: int -CAP_PROP_FPS: int -CAP_PROP_FOURCC: int -CAP_PROP_FRAME_COUNT: int -CAP_PROP_FORMAT: int -CAP_PROP_MODE: int -CAP_PROP_BRIGHTNESS: int -CAP_PROP_CONTRAST: int -CAP_PROP_SATURATION: int -CAP_PROP_HUE: int -CAP_PROP_GAIN: int -CAP_PROP_EXPOSURE: int -CAP_PROP_CONVERT_RGB: int -CAP_PROP_WHITE_BALANCE_BLUE_U: int -CAP_PROP_RECTIFICATION: int -CAP_PROP_MONOCHROME: int -CAP_PROP_SHARPNESS: int -CAP_PROP_AUTO_EXPOSURE: int -CAP_PROP_GAMMA: int -CAP_PROP_TEMPERATURE: int -CAP_PROP_TRIGGER: int -CAP_PROP_TRIGGER_DELAY: int -CAP_PROP_WHITE_BALANCE_RED_V: int -CAP_PROP_ZOOM: int -CAP_PROP_FOCUS: int -CAP_PROP_GUID: int -CAP_PROP_ISO_SPEED: int -CAP_PROP_BACKLIGHT: int -CAP_PROP_PAN: int -CAP_PROP_TILT: int -CAP_PROP_ROLL: int -CAP_PROP_IRIS: int -CAP_PROP_SETTINGS: int -CAP_PROP_BUFFERSIZE: int -CAP_PROP_AUTOFOCUS: int -CAP_PROP_SAR_NUM: int -CAP_PROP_SAR_DEN: int -CAP_PROP_BACKEND: int -CAP_PROP_CHANNEL: int -CAP_PROP_AUTO_WB: int -CAP_PROP_WB_TEMPERATURE: int -CAP_PROP_CODEC_PIXEL_FORMAT: int -CAP_PROP_BITRATE: int -CAP_PROP_ORIENTATION_META: int -CAP_PROP_ORIENTATION_AUTO: int -CAP_PROP_HW_ACCELERATION: int -CAP_PROP_HW_DEVICE: int -CAP_PROP_HW_ACCELERATION_USE_OPENCL: int -CAP_PROP_OPEN_TIMEOUT_MSEC: int -CAP_PROP_READ_TIMEOUT_MSEC: int -CAP_PROP_STREAM_OPEN_TIME_USEC: int -CAP_PROP_VIDEO_TOTAL_CHANNELS: int -CAP_PROP_VIDEO_STREAM: int -CAP_PROP_AUDIO_STREAM: int -CAP_PROP_AUDIO_POS: int -CAP_PROP_AUDIO_SHIFT_NSEC: int -CAP_PROP_AUDIO_DATA_DEPTH: int -CAP_PROP_AUDIO_SAMPLES_PER_SECOND: int -CAP_PROP_AUDIO_BASE_INDEX: int -CAP_PROP_AUDIO_TOTAL_CHANNELS: int -CAP_PROP_AUDIO_TOTAL_STREAMS: int -CAP_PROP_AUDIO_SYNCHRONIZE: int -CAP_PROP_LRF_HAS_KEY_FRAME: int -CAP_PROP_CODEC_EXTRADATA_INDEX: int -CAP_PROP_FRAME_TYPE: int -CAP_PROP_N_THREADS: int -VideoCaptureProperties = int -"""One of [CAP_PROP_POS_MSEC, CAP_PROP_POS_FRAMES, CAP_PROP_POS_AVI_RATIO, CAP_PROP_FRAME_WIDTH, CAP_PROP_FRAME_HEIGHT, -CAP_PROP_FPS, CAP_PROP_FOURCC, CAP_PROP_FRAME_COUNT, CAP_PROP_FORMAT, CAP_PROP_MODE, CAP_PROP_BRIGHTNESS, -CAP_PROP_CONTRAST, CAP_PROP_SATURATION, CAP_PROP_HUE, CAP_PROP_GAIN, CAP_PROP_EXPOSURE, CAP_PROP_CONVERT_RGB, -CAP_PROP_WHITE_BALANCE_BLUE_U, CAP_PROP_RECTIFICATION, CAP_PROP_MONOCHROME, CAP_PROP_SHARPNESS, CAP_PROP_AUTO_EXPOSURE, -CAP_PROP_GAMMA, CAP_PROP_TEMPERATURE, CAP_PROP_TRIGGER, CAP_PROP_TRIGGER_DELAY, CAP_PROP_WHITE_BALANCE_RED_V, -CAP_PROP_ZOOM, CAP_PROP_FOCUS, CAP_PROP_GUID, CAP_PROP_ISO_SPEED, CAP_PROP_BACKLIGHT, CAP_PROP_PAN, CAP_PROP_TILT, -CAP_PROP_ROLL, CAP_PROP_IRIS, CAP_PROP_SETTINGS, CAP_PROP_BUFFERSIZE, CAP_PROP_AUTOFOCUS, CAP_PROP_SAR_NUM, -CAP_PROP_SAR_DEN, CAP_PROP_BACKEND, CAP_PROP_CHANNEL, CAP_PROP_AUTO_WB, CAP_PROP_WB_TEMPERATURE, -CAP_PROP_CODEC_PIXEL_FORMAT, CAP_PROP_BITRATE, CAP_PROP_ORIENTATION_META, CAP_PROP_ORIENTATION_AUTO, -CAP_PROP_HW_ACCELERATION, CAP_PROP_HW_DEVICE, CAP_PROP_HW_ACCELERATION_USE_OPENCL, CAP_PROP_OPEN_TIMEOUT_MSEC, -CAP_PROP_READ_TIMEOUT_MSEC, CAP_PROP_STREAM_OPEN_TIME_USEC, CAP_PROP_VIDEO_TOTAL_CHANNELS, CAP_PROP_VIDEO_STREAM, -CAP_PROP_AUDIO_STREAM, CAP_PROP_AUDIO_POS, CAP_PROP_AUDIO_SHIFT_NSEC, CAP_PROP_AUDIO_DATA_DEPTH, -CAP_PROP_AUDIO_SAMPLES_PER_SECOND, CAP_PROP_AUDIO_BASE_INDEX, CAP_PROP_AUDIO_TOTAL_CHANNELS, -CAP_PROP_AUDIO_TOTAL_STREAMS, CAP_PROP_AUDIO_SYNCHRONIZE, CAP_PROP_LRF_HAS_KEY_FRAME, CAP_PROP_CODEC_EXTRADATA_INDEX, -CAP_PROP_FRAME_TYPE, CAP_PROP_N_THREADS]""" - -VIDEOWRITER_PROP_QUALITY: int -VIDEOWRITER_PROP_FRAMEBYTES: int -VIDEOWRITER_PROP_NSTRIPES: int -VIDEOWRITER_PROP_IS_COLOR: int -VIDEOWRITER_PROP_DEPTH: int -VIDEOWRITER_PROP_HW_ACCELERATION: int -VIDEOWRITER_PROP_HW_DEVICE: int -VIDEOWRITER_PROP_HW_ACCELERATION_USE_OPENCL: int -VideoWriterProperties = int -"""One of [VIDEOWRITER_PROP_QUALITY, VIDEOWRITER_PROP_FRAMEBYTES, VIDEOWRITER_PROP_NSTRIPES, VIDEOWRITER_PROP_IS_COLOR, -VIDEOWRITER_PROP_DEPTH, VIDEOWRITER_PROP_HW_ACCELERATION, VIDEOWRITER_PROP_HW_DEVICE, -VIDEOWRITER_PROP_HW_ACCELERATION_USE_OPENCL]""" - -VIDEO_ACCELERATION_NONE: int -VIDEO_ACCELERATION_ANY: int -VIDEO_ACCELERATION_D3D11: int -VIDEO_ACCELERATION_VAAPI: int -VIDEO_ACCELERATION_MFX: int -VideoAccelerationType = int -"""One of [VIDEO_ACCELERATION_NONE, VIDEO_ACCELERATION_ANY, VIDEO_ACCELERATION_D3D11, VIDEO_ACCELERATION_VAAPI, -VIDEO_ACCELERATION_MFX]""" - -CAP_OBSENSOR_DEPTH_MAP: int -CAP_OBSENSOR_BGR_IMAGE: int -CAP_OBSENSOR_IR_IMAGE: int -VideoCaptureOBSensorDataType = int -"""One of [CAP_OBSENSOR_DEPTH_MAP, CAP_OBSENSOR_BGR_IMAGE, CAP_OBSENSOR_IR_IMAGE]""" - -CAP_OBSENSOR_DEPTH_GENERATOR: int -CAP_OBSENSOR_IMAGE_GENERATOR: int -CAP_OBSENSOR_IR_GENERATOR: int -CAP_OBSENSOR_GENERATORS_MASK: int -VideoCaptureOBSensorGenerators = int -"""One of [CAP_OBSENSOR_DEPTH_GENERATOR, CAP_OBSENSOR_IMAGE_GENERATOR, CAP_OBSENSOR_IR_GENERATOR, -CAP_OBSENSOR_GENERATORS_MASK]""" - -CAP_PROP_OBSENSOR_INTRINSIC_FX: int -CAP_PROP_OBSENSOR_INTRINSIC_FY: int -CAP_PROP_OBSENSOR_INTRINSIC_CX: int -CAP_PROP_OBSENSOR_INTRINSIC_CY: int -VideoCaptureOBSensorProperties = int -"""One of [CAP_PROP_OBSENSOR_INTRINSIC_FX, CAP_PROP_OBSENSOR_INTRINSIC_FY, CAP_PROP_OBSENSOR_INTRINSIC_CX, -CAP_PROP_OBSENSOR_INTRINSIC_CY]""" - -SOLVEPNP_ITERATIVE: int -SOLVEPNP_EPNP: int -SOLVEPNP_P3P: int -SOLVEPNP_DLS: int -SOLVEPNP_UPNP: int -SOLVEPNP_AP3P: int -SOLVEPNP_IPPE: int -SOLVEPNP_IPPE_SQUARE: int -SOLVEPNP_SQPNP: int -SOLVEPNP_MAX_COUNT: int -SolvePnPMethod = int -"""One of [SOLVEPNP_ITERATIVE, SOLVEPNP_EPNP, SOLVEPNP_P3P, SOLVEPNP_DLS, SOLVEPNP_UPNP, SOLVEPNP_AP3P, SOLVEPNP_IPPE, -SOLVEPNP_IPPE_SQUARE, SOLVEPNP_SQPNP, SOLVEPNP_MAX_COUNT]""" - -CALIB_HAND_EYE_TSAI: int -CALIB_HAND_EYE_PARK: int -CALIB_HAND_EYE_HORAUD: int -CALIB_HAND_EYE_ANDREFF: int -CALIB_HAND_EYE_DANIILIDIS: int -HandEyeCalibrationMethod = int -"""One of [CALIB_HAND_EYE_TSAI, CALIB_HAND_EYE_PARK, CALIB_HAND_EYE_HORAUD, CALIB_HAND_EYE_ANDREFF, -CALIB_HAND_EYE_DANIILIDIS]""" - -CALIB_ROBOT_WORLD_HAND_EYE_SHAH: int -CALIB_ROBOT_WORLD_HAND_EYE_LI: int -RobotWorldHandEyeCalibrationMethod = int -"""One of [CALIB_ROBOT_WORLD_HAND_EYE_SHAH, CALIB_ROBOT_WORLD_HAND_EYE_LI]""" - -SAMPLING_UNIFORM: int -SAMPLING_PROGRESSIVE_NAPSAC: int -SAMPLING_NAPSAC: int -SAMPLING_PROSAC: int -SamplingMethod = int -"""One of [SAMPLING_UNIFORM, SAMPLING_PROGRESSIVE_NAPSAC, SAMPLING_NAPSAC, SAMPLING_PROSAC]""" - -LOCAL_OPTIM_NULL: int -LOCAL_OPTIM_INNER_LO: int -LOCAL_OPTIM_INNER_AND_ITER_LO: int -LOCAL_OPTIM_GC: int -LOCAL_OPTIM_SIGMA: int -LocalOptimMethod = int -"""One of [LOCAL_OPTIM_NULL, LOCAL_OPTIM_INNER_LO, LOCAL_OPTIM_INNER_AND_ITER_LO, LOCAL_OPTIM_GC, LOCAL_OPTIM_SIGMA]""" - -SCORE_METHOD_RANSAC: int -SCORE_METHOD_MSAC: int -SCORE_METHOD_MAGSAC: int -SCORE_METHOD_LMEDS: int -ScoreMethod = int -"""One of [SCORE_METHOD_RANSAC, SCORE_METHOD_MSAC, SCORE_METHOD_MAGSAC, SCORE_METHOD_LMEDS]""" - -NEIGH_FLANN_KNN: int -NEIGH_GRID: int -NEIGH_FLANN_RADIUS: int -NeighborSearchMethod = int -"""One of [NEIGH_FLANN_KNN, NEIGH_GRID, NEIGH_FLANN_RADIUS]""" - -NONE_POLISHER: int -LSQ_POLISHER: int -MAGSAC: int -COV_POLISHER: int -PolishingMethod = int -"""One of [NONE_POLISHER, LSQ_POLISHER, MAGSAC, COV_POLISHER]""" - -PROJ_SPHERICAL_ORTHO: int -PROJ_SPHERICAL_EQRECT: int -UndistortTypes = int -"""One of [PROJ_SPHERICAL_ORTHO, PROJ_SPHERICAL_EQRECT]""" - -WINDOW_NORMAL: int -WINDOW_AUTOSIZE: int -WINDOW_OPENGL: int -WINDOW_FULLSCREEN: int -WINDOW_FREERATIO: int -WINDOW_KEEPRATIO: int -WINDOW_GUI_EXPANDED: int -WINDOW_GUI_NORMAL: int -WindowFlags = int -"""One of [WINDOW_NORMAL, WINDOW_AUTOSIZE, WINDOW_OPENGL, WINDOW_FULLSCREEN, WINDOW_FREERATIO, WINDOW_KEEPRATIO, -WINDOW_GUI_EXPANDED, WINDOW_GUI_NORMAL]""" - -WND_PROP_FULLSCREEN: int -WND_PROP_AUTOSIZE: int -WND_PROP_ASPECT_RATIO: int -WND_PROP_OPENGL: int -WND_PROP_VISIBLE: int -WND_PROP_TOPMOST: int -WND_PROP_VSYNC: int -WindowPropertyFlags = int -"""One of [WND_PROP_FULLSCREEN, WND_PROP_AUTOSIZE, WND_PROP_ASPECT_RATIO, WND_PROP_OPENGL, WND_PROP_VISIBLE, -WND_PROP_TOPMOST, WND_PROP_VSYNC]""" - -EVENT_MOUSEMOVE: int -EVENT_LBUTTONDOWN: int -EVENT_RBUTTONDOWN: int -EVENT_MBUTTONDOWN: int -EVENT_LBUTTONUP: int -EVENT_RBUTTONUP: int -EVENT_MBUTTONUP: int -EVENT_LBUTTONDBLCLK: int -EVENT_RBUTTONDBLCLK: int -EVENT_MBUTTONDBLCLK: int -EVENT_MOUSEWHEEL: int -EVENT_MOUSEHWHEEL: int -MouseEventTypes = int -"""One of [EVENT_MOUSEMOVE, EVENT_LBUTTONDOWN, EVENT_RBUTTONDOWN, EVENT_MBUTTONDOWN, EVENT_LBUTTONUP, EVENT_RBUTTONUP, - EVENT_MBUTTONUP, EVENT_LBUTTONDBLCLK, EVENT_RBUTTONDBLCLK, EVENT_MBUTTONDBLCLK, EVENT_MOUSEWHEEL, EVENT_MOUSEHWHEEL]""" - -EVENT_FLAG_LBUTTON: int -EVENT_FLAG_RBUTTON: int -EVENT_FLAG_MBUTTON: int -EVENT_FLAG_CTRLKEY: int -EVENT_FLAG_SHIFTKEY: int -EVENT_FLAG_ALTKEY: int -MouseEventFlags = int -"""One of [EVENT_FLAG_LBUTTON, EVENT_FLAG_RBUTTON, EVENT_FLAG_MBUTTON, EVENT_FLAG_CTRLKEY, EVENT_FLAG_SHIFTKEY, -EVENT_FLAG_ALTKEY]""" - -QT_FONT_LIGHT: int -QT_FONT_NORMAL: int -QT_FONT_DEMIBOLD: int -QT_FONT_BOLD: int -QT_FONT_BLACK: int -QtFontWeights = int -"""One of [QT_FONT_LIGHT, QT_FONT_NORMAL, QT_FONT_DEMIBOLD, QT_FONT_BOLD, QT_FONT_BLACK]""" - -QT_STYLE_NORMAL: int -QT_STYLE_ITALIC: int -QT_STYLE_OBLIQUE: int -QtFontStyles = int -"""One of [QT_STYLE_NORMAL, QT_STYLE_ITALIC, QT_STYLE_OBLIQUE]""" - -QT_PUSH_BUTTON: int -QT_CHECKBOX: int -QT_RADIOBOX: int -QT_NEW_BUTTONBAR: int -QtButtonTypes = int -"""One of [QT_PUSH_BUTTON, QT_CHECKBOX, QT_RADIOBOX, QT_NEW_BUTTONBAR]""" - -GShape_GMAT: int -GSHAPE_GMAT: int -GShape_GSCALAR: int -GSHAPE_GSCALAR: int -GShape_GARRAY: int -GSHAPE_GARRAY: int -GShape_GOPAQUE: int -GSHAPE_GOPAQUE: int -GShape_GFRAME: int -GSHAPE_GFRAME: int -GShape = int -"""One of [GShape_GMAT, GSHAPE_GMAT, GShape_GSCALAR, GSHAPE_GSCALAR, GShape_GARRAY, GSHAPE_GARRAY, GShape_GOPAQUE, -GSHAPE_GOPAQUE, GShape_GFRAME, GSHAPE_GFRAME]""" - -MediaFormat_BGR: int -MEDIA_FORMAT_BGR: int -MediaFormat_NV12: int -MEDIA_FORMAT_NV12: int -MediaFormat_GRAY: int -MEDIA_FORMAT_GRAY: int -MediaFormat = int -"""One of [MediaFormat_BGR, MEDIA_FORMAT_BGR, MediaFormat_NV12, MEDIA_FORMAT_NV12, MediaFormat_GRAY, -MEDIA_FORMAT_GRAY]""" - -FileStorage_READ: int -FILE_STORAGE_READ: int -FileStorage_WRITE: int -FILE_STORAGE_WRITE: int -FileStorage_APPEND: int -FILE_STORAGE_APPEND: int -FileStorage_MEMORY: int -FILE_STORAGE_MEMORY: int -FileStorage_FORMAT_MASK: int -FILE_STORAGE_FORMAT_MASK: int -FileStorage_FORMAT_AUTO: int -FILE_STORAGE_FORMAT_AUTO: int -FileStorage_FORMAT_XML: int -FILE_STORAGE_FORMAT_XML: int -FileStorage_FORMAT_YAML: int -FILE_STORAGE_FORMAT_YAML: int -FileStorage_FORMAT_JSON: int -FILE_STORAGE_FORMAT_JSON: int -FileStorage_BASE64: int -FILE_STORAGE_BASE64: int -FileStorage_WRITE_BASE64: int -FILE_STORAGE_WRITE_BASE64: int -FileStorage_Mode = int -"""One of [FileStorage_READ, FILE_STORAGE_READ, FileStorage_WRITE, FILE_STORAGE_WRITE, FileStorage_APPEND, -FILE_STORAGE_APPEND, FileStorage_MEMORY, FILE_STORAGE_MEMORY, FileStorage_FORMAT_MASK, FILE_STORAGE_FORMAT_MASK, -FileStorage_FORMAT_AUTO, FILE_STORAGE_FORMAT_AUTO, FileStorage_FORMAT_XML, FILE_STORAGE_FORMAT_XML, -FileStorage_FORMAT_YAML, FILE_STORAGE_FORMAT_YAML, FileStorage_FORMAT_JSON, FILE_STORAGE_FORMAT_JSON, -FileStorage_BASE64, FILE_STORAGE_BASE64, FileStorage_WRITE_BASE64, FILE_STORAGE_WRITE_BASE64]""" - -FileStorage_UNDEFINED: int -FILE_STORAGE_UNDEFINED: int -FileStorage_VALUE_EXPECTED: int -FILE_STORAGE_VALUE_EXPECTED: int -FileStorage_NAME_EXPECTED: int -FILE_STORAGE_NAME_EXPECTED: int -FileStorage_INSIDE_MAP: int -FILE_STORAGE_INSIDE_MAP: int -FileStorage_State = int -"""One of [FileStorage_UNDEFINED, FILE_STORAGE_UNDEFINED, FileStorage_VALUE_EXPECTED, FILE_STORAGE_VALUE_EXPECTED, -FileStorage_NAME_EXPECTED, FILE_STORAGE_NAME_EXPECTED, FileStorage_INSIDE_MAP, FILE_STORAGE_INSIDE_MAP]""" - -FileNode_NONE: int -FILE_NODE_NONE: int -FileNode_INT: int -FILE_NODE_INT: int -FileNode_REAL: int -FILE_NODE_REAL: int -FileNode_FLOAT: int -FILE_NODE_FLOAT: int -FileNode_STR: int -FILE_NODE_STR: int -FileNode_STRING: int -FILE_NODE_STRING: int -FileNode_SEQ: int -FILE_NODE_SEQ: int -FileNode_MAP: int -FILE_NODE_MAP: int -FileNode_TYPE_MASK: int -FILE_NODE_TYPE_MASK: int -FileNode_FLOW: int -FILE_NODE_FLOW: int -FileNode_UNIFORM: int -FILE_NODE_UNIFORM: int -FileNode_EMPTY: int -FILE_NODE_EMPTY: int -FileNode_NAMED: int -FILE_NODE_NAMED: int - -UMat_MAGIC_VAL: int -UMAT_MAGIC_VAL: int -UMat_AUTO_STEP: int -UMAT_AUTO_STEP: int -UMat_CONTINUOUS_FLAG: int -UMAT_CONTINUOUS_FLAG: int -UMat_SUBMATRIX_FLAG: int -UMAT_SUBMATRIX_FLAG: int -UMat_MAGIC_MASK: int -UMAT_MAGIC_MASK: int -UMat_TYPE_MASK: int -UMAT_TYPE_MASK: int -UMat_DEPTH_MASK: int -UMAT_DEPTH_MASK: int - -Subdiv2D_PTLOC_ERROR: int -SUBDIV2D_PTLOC_ERROR: int -Subdiv2D_PTLOC_OUTSIDE_RECT: int -SUBDIV2D_PTLOC_OUTSIDE_RECT: int -Subdiv2D_PTLOC_INSIDE: int -SUBDIV2D_PTLOC_INSIDE: int -Subdiv2D_PTLOC_VERTEX: int -SUBDIV2D_PTLOC_VERTEX: int -Subdiv2D_PTLOC_ON_EDGE: int -SUBDIV2D_PTLOC_ON_EDGE: int -Subdiv2D_NEXT_AROUND_ORG: int -SUBDIV2D_NEXT_AROUND_ORG: int -Subdiv2D_NEXT_AROUND_DST: int -SUBDIV2D_NEXT_AROUND_DST: int -Subdiv2D_PREV_AROUND_ORG: int -SUBDIV2D_PREV_AROUND_ORG: int -Subdiv2D_PREV_AROUND_DST: int -SUBDIV2D_PREV_AROUND_DST: int -Subdiv2D_NEXT_AROUND_LEFT: int -SUBDIV2D_NEXT_AROUND_LEFT: int -Subdiv2D_NEXT_AROUND_RIGHT: int -SUBDIV2D_NEXT_AROUND_RIGHT: int -Subdiv2D_PREV_AROUND_LEFT: int -SUBDIV2D_PREV_AROUND_LEFT: int -Subdiv2D_PREV_AROUND_RIGHT: int -SUBDIV2D_PREV_AROUND_RIGHT: int - -ORB_HARRIS_SCORE: int -ORB_FAST_SCORE: int -ORB_ScoreType = int -"""One of [ORB_HARRIS_SCORE, ORB_FAST_SCORE]""" - -FastFeatureDetector_TYPE_5_8: int -FAST_FEATURE_DETECTOR_TYPE_5_8: int -FastFeatureDetector_TYPE_7_12: int -FAST_FEATURE_DETECTOR_TYPE_7_12: int -FastFeatureDetector_TYPE_9_16: int -FAST_FEATURE_DETECTOR_TYPE_9_16: int -FastFeatureDetector_DetectorType = int -"""One of [FastFeatureDetector_TYPE_5_8, FAST_FEATURE_DETECTOR_TYPE_5_8, FastFeatureDetector_TYPE_7_12, -FAST_FEATURE_DETECTOR_TYPE_7_12, FastFeatureDetector_TYPE_9_16, FAST_FEATURE_DETECTOR_TYPE_9_16]""" - -FastFeatureDetector_THRESHOLD: int -FAST_FEATURE_DETECTOR_THRESHOLD: int -FastFeatureDetector_NONMAX_SUPPRESSION: int -FAST_FEATURE_DETECTOR_NONMAX_SUPPRESSION: int -FastFeatureDetector_FAST_N: int -FAST_FEATURE_DETECTOR_FAST_N: int - -AgastFeatureDetector_AGAST_5_8: int -AGAST_FEATURE_DETECTOR_AGAST_5_8: int -AgastFeatureDetector_AGAST_7_12d: int -AGAST_FEATURE_DETECTOR_AGAST_7_12D: int -AgastFeatureDetector_AGAST_7_12s: int -AGAST_FEATURE_DETECTOR_AGAST_7_12S: int -AgastFeatureDetector_OAST_9_16: int -AGAST_FEATURE_DETECTOR_OAST_9_16: int -AgastFeatureDetector_DetectorType = int -"""One of [AgastFeatureDetector_AGAST_5_8, AGAST_FEATURE_DETECTOR_AGAST_5_8, AgastFeatureDetector_AGAST_7_12d, -AGAST_FEATURE_DETECTOR_AGAST_7_12D, AgastFeatureDetector_AGAST_7_12s, AGAST_FEATURE_DETECTOR_AGAST_7_12S, -AgastFeatureDetector_OAST_9_16, AGAST_FEATURE_DETECTOR_OAST_9_16]""" - -AgastFeatureDetector_THRESHOLD: int -AGAST_FEATURE_DETECTOR_THRESHOLD: int -AgastFeatureDetector_NONMAX_SUPPRESSION: int -AGAST_FEATURE_DETECTOR_NONMAX_SUPPRESSION: int - -KAZE_DIFF_PM_G1: int -KAZE_DIFF_PM_G2: int -KAZE_DIFF_WEICKERT: int -KAZE_DIFF_CHARBONNIER: int -KAZE_DiffusivityType = int -"""One of [KAZE_DIFF_PM_G1, KAZE_DIFF_PM_G2, KAZE_DIFF_WEICKERT, KAZE_DIFF_CHARBONNIER]""" - -AKAZE_DESCRIPTOR_KAZE_UPRIGHT: int -AKAZE_DESCRIPTOR_KAZE: int -AKAZE_DESCRIPTOR_MLDB_UPRIGHT: int -AKAZE_DESCRIPTOR_MLDB: int -AKAZE_DescriptorType = int -"""One of [AKAZE_DESCRIPTOR_KAZE_UPRIGHT, AKAZE_DESCRIPTOR_KAZE, AKAZE_DESCRIPTOR_MLDB_UPRIGHT, -AKAZE_DESCRIPTOR_MLDB]""" - -DescriptorMatcher_FLANNBASED: int -DESCRIPTOR_MATCHER_FLANNBASED: int -DescriptorMatcher_BRUTEFORCE: int -DESCRIPTOR_MATCHER_BRUTEFORCE: int -DescriptorMatcher_BRUTEFORCE_L1: int -DESCRIPTOR_MATCHER_BRUTEFORCE_L1: int -DescriptorMatcher_BRUTEFORCE_HAMMING: int -DESCRIPTOR_MATCHER_BRUTEFORCE_HAMMING: int -DescriptorMatcher_BRUTEFORCE_HAMMINGLUT: int -DESCRIPTOR_MATCHER_BRUTEFORCE_HAMMINGLUT: int -DescriptorMatcher_BRUTEFORCE_SL2: int -DESCRIPTOR_MATCHER_BRUTEFORCE_SL2: int -DescriptorMatcher_MatcherType = int -"""One of [DescriptorMatcher_FLANNBASED, DESCRIPTOR_MATCHER_FLANNBASED, DescriptorMatcher_BRUTEFORCE, -DESCRIPTOR_MATCHER_BRUTEFORCE, DescriptorMatcher_BRUTEFORCE_L1, DESCRIPTOR_MATCHER_BRUTEFORCE_L1, -DescriptorMatcher_BRUTEFORCE_HAMMING, DESCRIPTOR_MATCHER_BRUTEFORCE_HAMMING, DescriptorMatcher_BRUTEFORCE_HAMMINGLUT, -DESCRIPTOR_MATCHER_BRUTEFORCE_HAMMINGLUT, DescriptorMatcher_BRUTEFORCE_SL2, DESCRIPTOR_MATCHER_BRUTEFORCE_SL2]""" - -CirclesGridFinderParameters_SYMMETRIC_GRID: int -CIRCLES_GRID_FINDER_PARAMETERS_SYMMETRIC_GRID: int -CirclesGridFinderParameters_ASYMMETRIC_GRID: int -CIRCLES_GRID_FINDER_PARAMETERS_ASYMMETRIC_GRID: int -CirclesGridFinderParameters_GridType = int -"""One of [CirclesGridFinderParameters_SYMMETRIC_GRID, CIRCLES_GRID_FINDER_PARAMETERS_SYMMETRIC_GRID, -CirclesGridFinderParameters_ASYMMETRIC_GRID, CIRCLES_GRID_FINDER_PARAMETERS_ASYMMETRIC_GRID]""" - -StereoMatcher_DISP_SHIFT: int -STEREO_MATCHER_DISP_SHIFT: int -StereoMatcher_DISP_SCALE: int -STEREO_MATCHER_DISP_SCALE: int - -StereoBM_PREFILTER_NORMALIZED_RESPONSE: int -STEREO_BM_PREFILTER_NORMALIZED_RESPONSE: int -StereoBM_PREFILTER_XSOBEL: int -STEREO_BM_PREFILTER_XSOBEL: int - -StereoSGBM_MODE_SGBM: int -STEREO_SGBM_MODE_SGBM: int -StereoSGBM_MODE_HH: int -STEREO_SGBM_MODE_HH: int -StereoSGBM_MODE_SGBM_3WAY: int -STEREO_SGBM_MODE_SGBM_3WAY: int -StereoSGBM_MODE_HH4: int -STEREO_SGBM_MODE_HH4: int - -HOGDescriptor_L2Hys: int -HOGDESCRIPTOR_L2HYS: int -HOGDescriptor_HistogramNormType = int -"""One of [HOGDescriptor_L2Hys, HOGDESCRIPTOR_L2HYS]""" - -HOGDescriptor_DEFAULT_NLEVELS: int -HOGDESCRIPTOR_DEFAULT_NLEVELS: int - -HOGDescriptor_DESCR_FORMAT_COL_BY_COL: int -HOGDESCRIPTOR_DESCR_FORMAT_COL_BY_COL: int -HOGDescriptor_DESCR_FORMAT_ROW_BY_ROW: int -HOGDESCRIPTOR_DESCR_FORMAT_ROW_BY_ROW: int -HOGDescriptor_DescriptorStorageFormat = int -"""One of [HOGDescriptor_DESCR_FORMAT_COL_BY_COL, HOGDESCRIPTOR_DESCR_FORMAT_COL_BY_COL, -HOGDescriptor_DESCR_FORMAT_ROW_BY_ROW, HOGDESCRIPTOR_DESCR_FORMAT_ROW_BY_ROW]""" - -QRCodeEncoder_MODE_AUTO: int -QRCODE_ENCODER_MODE_AUTO: int -QRCodeEncoder_MODE_NUMERIC: int -QRCODE_ENCODER_MODE_NUMERIC: int -QRCodeEncoder_MODE_ALPHANUMERIC: int -QRCODE_ENCODER_MODE_ALPHANUMERIC: int -QRCodeEncoder_MODE_BYTE: int -QRCODE_ENCODER_MODE_BYTE: int -QRCodeEncoder_MODE_ECI: int -QRCODE_ENCODER_MODE_ECI: int -QRCodeEncoder_MODE_KANJI: int -QRCODE_ENCODER_MODE_KANJI: int -QRCodeEncoder_MODE_STRUCTURED_APPEND: int -QRCODE_ENCODER_MODE_STRUCTURED_APPEND: int -QRCodeEncoder_EncodeMode = int -"""One of [QRCodeEncoder_MODE_AUTO, QRCODE_ENCODER_MODE_AUTO, QRCodeEncoder_MODE_NUMERIC, QRCODE_ENCODER_MODE_NUMERIC, -QRCodeEncoder_MODE_ALPHANUMERIC, QRCODE_ENCODER_MODE_ALPHANUMERIC, QRCodeEncoder_MODE_BYTE, QRCODE_ENCODER_MODE_BYTE, -QRCodeEncoder_MODE_ECI, QRCODE_ENCODER_MODE_ECI, QRCodeEncoder_MODE_KANJI, QRCODE_ENCODER_MODE_KANJI, - QRCodeEncoder_MODE_STRUCTURED_APPEND, QRCODE_ENCODER_MODE_STRUCTURED_APPEND]""" - -QRCodeEncoder_CORRECT_LEVEL_L: int -QRCODE_ENCODER_CORRECT_LEVEL_L: int -QRCodeEncoder_CORRECT_LEVEL_M: int -QRCODE_ENCODER_CORRECT_LEVEL_M: int -QRCodeEncoder_CORRECT_LEVEL_Q: int -QRCODE_ENCODER_CORRECT_LEVEL_Q: int -QRCodeEncoder_CORRECT_LEVEL_H: int -QRCODE_ENCODER_CORRECT_LEVEL_H: int -QRCodeEncoder_CorrectionLevel = int -"""One of [QRCodeEncoder_CORRECT_LEVEL_L, QRCODE_ENCODER_CORRECT_LEVEL_L, QRCodeEncoder_CORRECT_LEVEL_M, -QRCODE_ENCODER_CORRECT_LEVEL_M, QRCodeEncoder_CORRECT_LEVEL_Q, QRCODE_ENCODER_CORRECT_LEVEL_Q, -QRCodeEncoder_CORRECT_LEVEL_H, QRCODE_ENCODER_CORRECT_LEVEL_H]""" - -QRCodeEncoder_ECI_UTF8: int -QRCODE_ENCODER_ECI_UTF8: int -QRCodeEncoder_ECIEncodings = int -"""One of [QRCodeEncoder_ECI_UTF8, QRCODE_ENCODER_ECI_UTF8]""" - -FaceRecognizerSF_FR_COSINE: int -FACE_RECOGNIZER_SF_FR_COSINE: int -FaceRecognizerSF_FR_NORM_L2: int -FACE_RECOGNIZER_SF_FR_NORM_L2: int -FaceRecognizerSF_DisType = int -"""One of [FaceRecognizerSF_FR_COSINE, FACE_RECOGNIZER_SF_FR_COSINE, FaceRecognizerSF_FR_NORM_L2, -FACE_RECOGNIZER_SF_FR_NORM_L2]""" - -Stitcher_OK: int -STITCHER_OK: int -Stitcher_ERR_NEED_MORE_IMGS: int -STITCHER_ERR_NEED_MORE_IMGS: int -Stitcher_ERR_HOMOGRAPHY_EST_FAIL: int -STITCHER_ERR_HOMOGRAPHY_EST_FAIL: int -Stitcher_ERR_CAMERA_PARAMS_ADJUST_FAIL: int -STITCHER_ERR_CAMERA_PARAMS_ADJUST_FAIL: int -Stitcher_Status = int -"""One of [Stitcher_OK, STITCHER_OK, Stitcher_ERR_NEED_MORE_IMGS, STITCHER_ERR_NEED_MORE_IMGS, -Stitcher_ERR_HOMOGRAPHY_EST_FAIL, STITCHER_ERR_HOMOGRAPHY_EST_FAIL, Stitcher_ERR_CAMERA_PARAMS_ADJUST_FAIL, -STITCHER_ERR_CAMERA_PARAMS_ADJUST_FAIL]""" - -Stitcher_PANORAMA: int -STITCHER_PANORAMA: int -Stitcher_SCANS: int -STITCHER_SCANS: int -Stitcher_Mode = int -"""One of [Stitcher_PANORAMA, STITCHER_PANORAMA, Stitcher_SCANS, STITCHER_SCANS]""" - -DISOpticalFlow_PRESET_ULTRAFAST: int -DISOPTICAL_FLOW_PRESET_ULTRAFAST: int -DISOpticalFlow_PRESET_FAST: int -DISOPTICAL_FLOW_PRESET_FAST: int -DISOpticalFlow_PRESET_MEDIUM: int -DISOPTICAL_FLOW_PRESET_MEDIUM: int - -PCA_DATA_AS_ROW: int -PCA_DATA_AS_COL: int -PCA_USE_AVG: int -PCA_Flags = int -"""One of [PCA_DATA_AS_ROW, PCA_DATA_AS_COL, PCA_USE_AVG]""" - -SVD_MODIFY_A: int -SVD_NO_UV: int -SVD_FULL_UV: int -SVD_Flags = int -"""One of [SVD_MODIFY_A, SVD_NO_UV, SVD_FULL_UV]""" - -RNG_UNIFORM: int -RNG_NORMAL: int - -Formatter_FMT_DEFAULT: int -FORMATTER_FMT_DEFAULT: int -Formatter_FMT_MATLAB: int -FORMATTER_FMT_MATLAB: int -Formatter_FMT_CSV: int -FORMATTER_FMT_CSV: int -Formatter_FMT_PYTHON: int -FORMATTER_FMT_PYTHON: int -Formatter_FMT_NUMPY: int -FORMATTER_FMT_NUMPY: int -Formatter_FMT_C: int -FORMATTER_FMT_C: int -Formatter_FormatType = int -"""One of [Formatter_FMT_DEFAULT, FORMATTER_FMT_DEFAULT, Formatter_FMT_MATLAB, FORMATTER_FMT_MATLAB, Formatter_FMT_CSV, -FORMATTER_FMT_CSV, Formatter_FMT_PYTHON, FORMATTER_FMT_PYTHON, Formatter_FMT_NUMPY, FORMATTER_FMT_NUMPY, -Formatter_FMT_C, FORMATTER_FMT_C]""" - -_InputArray_KIND_SHIFT: int -_INPUT_ARRAY_KIND_SHIFT: int -_InputArray_FIXED_TYPE: int -_INPUT_ARRAY_FIXED_TYPE: int -_InputArray_FIXED_SIZE: int -_INPUT_ARRAY_FIXED_SIZE: int -_InputArray_KIND_MASK: int -_INPUT_ARRAY_KIND_MASK: int -_InputArray_NONE: int -_INPUT_ARRAY_NONE: int -_InputArray_MAT: int -_INPUT_ARRAY_MAT: int -_InputArray_MATX: int -_INPUT_ARRAY_MATX: int -_InputArray_STD_VECTOR: int -_INPUT_ARRAY_STD_VECTOR: int -_InputArray_STD_VECTOR_VECTOR: int -_INPUT_ARRAY_STD_VECTOR_VECTOR: int -_InputArray_STD_VECTOR_MAT: int -_INPUT_ARRAY_STD_VECTOR_MAT: int -_InputArray_EXPR: int -_INPUT_ARRAY_EXPR: int -_InputArray_OPENGL_BUFFER: int -_INPUT_ARRAY_OPENGL_BUFFER: int -_InputArray_CUDA_HOST_MEM: int -_INPUT_ARRAY_CUDA_HOST_MEM: int -_InputArray_CUDA_GPU_MAT: int -_INPUT_ARRAY_CUDA_GPU_MAT: int -_InputArray_UMAT: int -_INPUT_ARRAY_UMAT: int -_InputArray_STD_VECTOR_UMAT: int -_INPUT_ARRAY_STD_VECTOR_UMAT: int -_InputArray_STD_BOOL_VECTOR: int -_INPUT_ARRAY_STD_BOOL_VECTOR: int -_InputArray_STD_VECTOR_CUDA_GPU_MAT: int -_INPUT_ARRAY_STD_VECTOR_CUDA_GPU_MAT: int -_InputArray_STD_ARRAY: int -_INPUT_ARRAY_STD_ARRAY: int -_InputArray_STD_ARRAY_MAT: int -_INPUT_ARRAY_STD_ARRAY_MAT: int -_InputArray_KindFlag = int -"""One of [_InputArray_KIND_SHIFT, _INPUT_ARRAY_KIND_SHIFT, _InputArray_FIXED_TYPE, _INPUT_ARRAY_FIXED_TYPE, -_InputArray_FIXED_SIZE, _INPUT_ARRAY_FIXED_SIZE, _InputArray_KIND_MASK, _INPUT_ARRAY_KIND_MASK, _InputArray_NONE, -_INPUT_ARRAY_NONE, _InputArray_MAT, _INPUT_ARRAY_MAT, _InputArray_MATX, _INPUT_ARRAY_MATX, _InputArray_STD_VECTOR, -_INPUT_ARRAY_STD_VECTOR, _InputArray_STD_VECTOR_VECTOR, _INPUT_ARRAY_STD_VECTOR_VECTOR, _InputArray_STD_VECTOR_MAT, -_INPUT_ARRAY_STD_VECTOR_MAT, _InputArray_EXPR, _INPUT_ARRAY_EXPR, _InputArray_OPENGL_BUFFER, _INPUT_ARRAY_OPENGL_BUFFER, -_InputArray_CUDA_HOST_MEM, _INPUT_ARRAY_CUDA_HOST_MEM, _InputArray_CUDA_GPU_MAT, _INPUT_ARRAY_CUDA_GPU_MAT, -_InputArray_UMAT, _INPUT_ARRAY_UMAT, _InputArray_STD_VECTOR_UMAT, _INPUT_ARRAY_STD_VECTOR_UMAT, -_InputArray_STD_BOOL_VECTOR, _INPUT_ARRAY_STD_BOOL_VECTOR, _InputArray_STD_VECTOR_CUDA_GPU_MAT, -_INPUT_ARRAY_STD_VECTOR_CUDA_GPU_MAT, _InputArray_STD_ARRAY, _INPUT_ARRAY_STD_ARRAY, _InputArray_STD_ARRAY_MAT, -_INPUT_ARRAY_STD_ARRAY_MAT]""" - -_OutputArray_DEPTH_MASK_8U: int -_OUTPUT_ARRAY_DEPTH_MASK_8U: int -_OutputArray_DEPTH_MASK_8S: int -_OUTPUT_ARRAY_DEPTH_MASK_8S: int -_OutputArray_DEPTH_MASK_16U: int -_OUTPUT_ARRAY_DEPTH_MASK_16U: int -_OutputArray_DEPTH_MASK_16S: int -_OUTPUT_ARRAY_DEPTH_MASK_16S: int -_OutputArray_DEPTH_MASK_32S: int -_OUTPUT_ARRAY_DEPTH_MASK_32S: int -_OutputArray_DEPTH_MASK_32F: int -_OUTPUT_ARRAY_DEPTH_MASK_32F: int -_OutputArray_DEPTH_MASK_64F: int -_OUTPUT_ARRAY_DEPTH_MASK_64F: int -_OutputArray_DEPTH_MASK_16F: int -_OUTPUT_ARRAY_DEPTH_MASK_16F: int -_OutputArray_DEPTH_MASK_ALL: int -_OUTPUT_ARRAY_DEPTH_MASK_ALL: int -_OutputArray_DEPTH_MASK_ALL_BUT_8S: int -_OUTPUT_ARRAY_DEPTH_MASK_ALL_BUT_8S: int -_OutputArray_DEPTH_MASK_ALL_16F: int -_OUTPUT_ARRAY_DEPTH_MASK_ALL_16F: int -_OutputArray_DEPTH_MASK_FLT: int -_OUTPUT_ARRAY_DEPTH_MASK_FLT: int -_OutputArray_DepthMask = int -"""One of [_OutputArray_DEPTH_MASK_8U, _OUTPUT_ARRAY_DEPTH_MASK_8U, _OutputArray_DEPTH_MASK_8S, -_OUTPUT_ARRAY_DEPTH_MASK_8S, _OutputArray_DEPTH_MASK_16U, _OUTPUT_ARRAY_DEPTH_MASK_16U, _OutputArray_DEPTH_MASK_16S, -_OUTPUT_ARRAY_DEPTH_MASK_16S, _OutputArray_DEPTH_MASK_32S, _OUTPUT_ARRAY_DEPTH_MASK_32S, _OutputArray_DEPTH_MASK_32F, -_OUTPUT_ARRAY_DEPTH_MASK_32F, _OutputArray_DEPTH_MASK_64F, _OUTPUT_ARRAY_DEPTH_MASK_64F, _OutputArray_DEPTH_MASK_16F, -_OUTPUT_ARRAY_DEPTH_MASK_16F, _OutputArray_DEPTH_MASK_ALL, _OUTPUT_ARRAY_DEPTH_MASK_ALL, -_OutputArray_DEPTH_MASK_ALL_BUT_8S, _OUTPUT_ARRAY_DEPTH_MASK_ALL_BUT_8S, _OutputArray_DEPTH_MASK_ALL_16F, -_OUTPUT_ARRAY_DEPTH_MASK_ALL_16F, _OutputArray_DEPTH_MASK_FLT, _OUTPUT_ARRAY_DEPTH_MASK_FLT]""" - -UMatData_COPY_ON_MAP: int -UMAT_DATA_COPY_ON_MAP: int -UMatData_HOST_COPY_OBSOLETE: int -UMAT_DATA_HOST_COPY_OBSOLETE: int -UMatData_DEVICE_COPY_OBSOLETE: int -UMAT_DATA_DEVICE_COPY_OBSOLETE: int -UMatData_TEMP_UMAT: int -UMAT_DATA_TEMP_UMAT: int -UMatData_TEMP_COPIED_UMAT: int -UMAT_DATA_TEMP_COPIED_UMAT: int -UMatData_USER_ALLOCATED: int -UMAT_DATA_USER_ALLOCATED: int -UMatData_DEVICE_MEM_MAPPED: int -UMAT_DATA_DEVICE_MEM_MAPPED: int -UMatData_ASYNC_CLEANUP: int -UMAT_DATA_ASYNC_CLEANUP: int -UMatData_MemoryFlag = int -"""One of [UMatData_COPY_ON_MAP, UMAT_DATA_COPY_ON_MAP, UMatData_HOST_COPY_OBSOLETE, UMAT_DATA_HOST_COPY_OBSOLETE, -UMatData_DEVICE_COPY_OBSOLETE, UMAT_DATA_DEVICE_COPY_OBSOLETE, UMatData_TEMP_UMAT, UMAT_DATA_TEMP_UMAT, -UMatData_TEMP_COPIED_UMAT, UMAT_DATA_TEMP_COPIED_UMAT, UMatData_USER_ALLOCATED, UMAT_DATA_USER_ALLOCATED, -UMatData_DEVICE_MEM_MAPPED, UMAT_DATA_DEVICE_MEM_MAPPED, UMatData_ASYNC_CLEANUP, UMAT_DATA_ASYNC_CLEANUP]""" - -Mat_MAGIC_VAL: int -MAT_MAGIC_VAL: int -Mat_AUTO_STEP: int -MAT_AUTO_STEP: int -Mat_CONTINUOUS_FLAG: int -MAT_CONTINUOUS_FLAG: int -Mat_SUBMATRIX_FLAG: int -MAT_SUBMATRIX_FLAG: int -Mat_MAGIC_MASK: int -MAT_MAGIC_MASK: int -Mat_TYPE_MASK: int -MAT_TYPE_MASK: int -Mat_DEPTH_MASK: int -MAT_DEPTH_MASK: int - -SparseMat_MAGIC_VAL: int -SPARSE_MAT_MAGIC_VAL: int -SparseMat_MAX_DIM: int -SPARSE_MAT_MAX_DIM: int -SparseMat_HASH_SCALE: int -SPARSE_MAT_HASH_SCALE: int -SparseMat_HASH_BIT: int -SPARSE_MAT_HASH_BIT: int - -QuatEnum_INT_XYZ: int -QUAT_ENUM_INT_XYZ: int -QuatEnum_INT_XZY: int -QUAT_ENUM_INT_XZY: int -QuatEnum_INT_YXZ: int -QUAT_ENUM_INT_YXZ: int -QuatEnum_INT_YZX: int -QUAT_ENUM_INT_YZX: int -QuatEnum_INT_ZXY: int -QUAT_ENUM_INT_ZXY: int -QuatEnum_INT_ZYX: int -QUAT_ENUM_INT_ZYX: int -QuatEnum_INT_XYX: int -QUAT_ENUM_INT_XYX: int -QuatEnum_INT_XZX: int -QUAT_ENUM_INT_XZX: int -QuatEnum_INT_YXY: int -QUAT_ENUM_INT_YXY: int -QuatEnum_INT_YZY: int -QUAT_ENUM_INT_YZY: int -QuatEnum_INT_ZXZ: int -QUAT_ENUM_INT_ZXZ: int -QuatEnum_INT_ZYZ: int -QUAT_ENUM_INT_ZYZ: int -QuatEnum_EXT_XYZ: int -QUAT_ENUM_EXT_XYZ: int -QuatEnum_EXT_XZY: int -QUAT_ENUM_EXT_XZY: int -QuatEnum_EXT_YXZ: int -QUAT_ENUM_EXT_YXZ: int -QuatEnum_EXT_YZX: int -QUAT_ENUM_EXT_YZX: int -QuatEnum_EXT_ZXY: int -QUAT_ENUM_EXT_ZXY: int -QuatEnum_EXT_ZYX: int -QUAT_ENUM_EXT_ZYX: int -QuatEnum_EXT_XYX: int -QUAT_ENUM_EXT_XYX: int -QuatEnum_EXT_XZX: int -QUAT_ENUM_EXT_XZX: int -QuatEnum_EXT_YXY: int -QUAT_ENUM_EXT_YXY: int -QuatEnum_EXT_YZY: int -QUAT_ENUM_EXT_YZY: int -QuatEnum_EXT_ZXZ: int -QUAT_ENUM_EXT_ZXZ: int -QuatEnum_EXT_ZYZ: int -QUAT_ENUM_EXT_ZYZ: int -QuatEnum_EULER_ANGLES_MAX_VALUE: int -QUAT_ENUM_EULER_ANGLES_MAX_VALUE: int -QuatEnum_EulerAnglesType = int -"""One of [QuatEnum_INT_XYZ, QUAT_ENUM_INT_XYZ, QuatEnum_INT_XZY, QUAT_ENUM_INT_XZY, QuatEnum_INT_YXZ, -QUAT_ENUM_INT_YXZ, QuatEnum_INT_YZX, QUAT_ENUM_INT_YZX, QuatEnum_INT_ZXY, QUAT_ENUM_INT_ZXY, QuatEnum_INT_ZYX, -QUAT_ENUM_INT_ZYX, QuatEnum_INT_XYX, QUAT_ENUM_INT_XYX, QuatEnum_INT_XZX, QUAT_ENUM_INT_XZX, QuatEnum_INT_YXY, -QUAT_ENUM_INT_YXY, QuatEnum_INT_YZY, QUAT_ENUM_INT_YZY, QuatEnum_INT_ZXZ, QUAT_ENUM_INT_ZXZ, QuatEnum_INT_ZYZ, -QUAT_ENUM_INT_ZYZ, QuatEnum_EXT_XYZ, QUAT_ENUM_EXT_XYZ, QuatEnum_EXT_XZY, QUAT_ENUM_EXT_XZY, QuatEnum_EXT_YXZ, -QUAT_ENUM_EXT_YXZ, QuatEnum_EXT_YZX, QUAT_ENUM_EXT_YZX, QuatEnum_EXT_ZXY, QUAT_ENUM_EXT_ZXY, QuatEnum_EXT_ZYX, -QUAT_ENUM_EXT_ZYX, QuatEnum_EXT_XYX, QUAT_ENUM_EXT_XYX, QuatEnum_EXT_XZX, QUAT_ENUM_EXT_XZX, QuatEnum_EXT_YXY, -QUAT_ENUM_EXT_YXY, QuatEnum_EXT_YZY, QUAT_ENUM_EXT_YZY, QuatEnum_EXT_ZXZ, QUAT_ENUM_EXT_ZXZ, QuatEnum_EXT_ZYZ, -QUAT_ENUM_EXT_ZYZ, QuatEnum_EULER_ANGLES_MAX_VALUE, QUAT_ENUM_EULER_ANGLES_MAX_VALUE]""" - -TermCriteria_COUNT: int -TERM_CRITERIA_COUNT: int -TermCriteria_MAX_ITER: int -TERM_CRITERIA_MAX_ITER: int -TermCriteria_EPS: int -TERM_CRITERIA_EPS: int -TermCriteria_Type = int -"""One of [TermCriteria_COUNT, TERM_CRITERIA_COUNT, TermCriteria_MAX_ITER, TERM_CRITERIA_MAX_ITER, TermCriteria_EPS, -TERM_CRITERIA_EPS]""" - -GFluidKernel_Kind_Filter: int -GFLUID_KERNEL_KIND_FILTER: int -GFluidKernel_Kind_Resize: int -GFLUID_KERNEL_KIND_RESIZE: int -GFluidKernel_Kind_YUV420toRGB: int -GFLUID_KERNEL_KIND_YUV420TO_RGB: int -GFluidKernel_Kind = int -"""One of [GFluidKernel_Kind_Filter, GFLUID_KERNEL_KIND_FILTER, GFluidKernel_Kind_Resize, GFLUID_KERNEL_KIND_RESIZE, -GFluidKernel_Kind_YUV420toRGB, GFLUID_KERNEL_KIND_YUV420TO_RGB]""" - -MediaFrame_Access_R: int -MEDIA_FRAME_ACCESS_R: int -MediaFrame_Access_W: int -MEDIA_FRAME_ACCESS_W: int -MediaFrame_Access = int -"""One of [MediaFrame_Access_R, MEDIA_FRAME_ACCESS_R, MediaFrame_Access_W, MEDIA_FRAME_ACCESS_W]""" - -RMat_Access_R: int -RMAT_ACCESS_R: int -RMat_Access_W: int -RMAT_ACCESS_W: int -RMat_Access = int -"""One of [RMat_Access_R, RMAT_ACCESS_R, RMat_Access_W, RMAT_ACCESS_W]""" - -# Classes -class Algorithm: - # Functions - def clear(self) -> None: ... - @typing.overload - def write(self, fs: FileStorage) -> None: ... - @typing.overload - def write(self, fs: FileStorage, name: str) -> None: ... - def read(self, fn: FileNode) -> None: ... - def empty(self) -> bool: ... - def save(self, filename: str) -> None: ... - def getDefaultName(self) -> str: ... - -class AsyncArray: - # Functions - def __init__(self) -> None: ... - def release(self) -> None: ... - @typing.overload - def get(self, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... - @typing.overload - def get(self, dst: UMat | None = ...) -> UMat: ... - @typing.overload - def get(self, timeoutNs: float, dst: cv2.typing.MatLike | None = ...) -> tuple[bool, cv2.typing.MatLike]: ... - @typing.overload - def get(self, timeoutNs: float, dst: UMat | None = ...) -> tuple[bool, UMat]: ... - def wait_for(self, timeoutNs: float) -> bool: ... - def valid(self) -> bool: ... - -class FileStorage: - # Functions - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def __init__(self, filename: str, flags: int, encoding: str = ...) -> None: ... - def open(self, filename: str, flags: int, encoding: str = ...) -> bool: ... - def isOpened(self) -> bool: ... - def release(self) -> None: ... - def releaseAndGetString(self) -> str: ... - def getFirstTopLevelNode(self) -> FileNode: ... - def root(self, streamidx: int = ...) -> FileNode: ... - def getNode(self, nodename: str) -> FileNode: ... - @typing.overload - def write(self, name: str, val: int) -> None: ... - @typing.overload - def write(self, name: str, val: float) -> None: ... - @typing.overload - def write(self, name: str, val: str) -> None: ... - @typing.overload - def write(self, name: str, val: cv2.typing.MatLike) -> None: ... - @typing.overload - def write(self, name: str, val: typing.Sequence[str]) -> None: ... - def writeComment(self, comment: str, append: bool = ...) -> None: ... - def startWriteStruct(self, name: str, flags: int, typeName: str = ...) -> None: ... - def endWriteStruct(self) -> None: ... - def getFormat(self) -> int: ... - -class FileNode: - # Functions - def __init__(self) -> None: ... - def getNode(self, nodename: str) -> FileNode: ... - def at(self, i: int) -> FileNode: ... - def keys(self) -> typing.Sequence[str]: ... - def type(self) -> int: ... - def empty(self) -> bool: ... - def isNone(self) -> bool: ... - def isSeq(self) -> bool: ... - def isMap(self) -> bool: ... - def isInt(self) -> bool: ... - def isReal(self) -> bool: ... - def isString(self) -> bool: ... - def isNamed(self) -> bool: ... - def name(self) -> str: ... - def size(self) -> int: ... - def rawSize(self) -> int: ... - def real(self) -> float: ... - def string(self) -> str: ... - def mat(self) -> cv2.typing.MatLike: ... - -class RotatedRect: - center: cv2.typing.Point2f - size: cv2.typing.Size2f - angle: float - - # Functions - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def __init__(self, center: cv2.typing.Point2f, size: cv2.typing.Size2f, angle: float) -> None: ... - @typing.overload - def __init__(self, point1: cv2.typing.Point2f, point2: cv2.typing.Point2f, point3: cv2.typing.Point2f) -> None: ... - def points(self) -> typing.Sequence[cv2.typing.Point2f]: ... - def boundingRect(self) -> cv2.typing.Rect: ... - -class KeyPoint: - pt: cv2.typing.Point2f - size: float - angle: float - response: float - octave: int - class_id: int - - # Functions - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def __init__( - self, - x: float, - y: float, - size: float, - angle: float = ..., - response: float = ..., - octave: int = ..., - class_id: int = ..., - ) -> None: ... - @staticmethod - @typing.overload - def convert( - keypoints: typing.Sequence[KeyPoint], - keypointIndexes: typing.Sequence[int] = ..., - ) -> typing.Sequence[cv2.typing.Point2f]: ... - @staticmethod - @typing.overload - def convert( - points2f: typing.Sequence[cv2.typing.Point2f], - size: float = ..., - response: float = ..., - octave: int = ..., - class_id: int = ..., - ) -> typing.Sequence[KeyPoint]: ... - @staticmethod - def overlap(kp1: KeyPoint, kp2: KeyPoint) -> float: ... - -class DMatch: - queryIdx: int - trainIdx: int - imgIdx: int - distance: float - - # Functions - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def __init__(self, _queryIdx: int, _trainIdx: int, _distance: float) -> None: ... - @typing.overload - def __init__(self, _queryIdx: int, _trainIdx: int, _imgIdx: int, _distance: float) -> None: ... - -class TickMeter: - # Functions - def __init__(self) -> None: ... - def start(self) -> None: ... - def stop(self) -> None: ... - def getTimeTicks(self) -> int: ... - def getTimeMicro(self) -> float: ... - def getTimeMilli(self) -> float: ... - def getTimeSec(self) -> float: ... - def getCounter(self) -> int: ... - def getFPS(self) -> float: ... - def getAvgTimeSec(self) -> float: ... - def getAvgTimeMilli(self) -> float: ... - def reset(self) -> None: ... - -class UMat: - offset: int - - # Functions - @typing.overload - def __init__(self, usageFlags: UMatUsageFlags = ...) -> None: ... - @typing.overload - def __init__(self, rows: int, cols: int, type: int, usageFlags: UMatUsageFlags = ...) -> None: ... - @typing.overload - def __init__(self, size: cv2.typing.Size, type: int, usageFlags: UMatUsageFlags = ...) -> None: ... - @typing.overload - def __init__( - self, - rows: int, - cols: int, - type: int, - s: cv2.typing.Scalar, - usageFlags: UMatUsageFlags = ..., - ) -> None: ... - @typing.overload - def __init__( - self, - size: cv2.typing.Size, - type: int, - s: cv2.typing.Scalar, - usageFlags: UMatUsageFlags = ..., - ) -> None: ... - @typing.overload - def __init__(self, m: UMat) -> None: ... - @typing.overload - def __init__(self, m: UMat, rowRange: cv2.typing.Range, colRange: cv2.typing.Range = ...) -> None: ... - @typing.overload - def __init__(self, m: UMat, roi: cv2.typing.Rect) -> None: ... - @typing.overload - def __init__(self, m: UMat, ranges: typing.Sequence[cv2.typing.Range]) -> None: ... - @staticmethod - def queue() -> cv2.typing.IntPointer: ... - @staticmethod - def context() -> cv2.typing.IntPointer: ... - def get(self) -> cv2.typing.MatLike: ... - def isContinuous(self) -> bool: ... - def isSubmatrix(self) -> bool: ... - def handle(self, accessFlags: AccessFlag) -> cv2.typing.IntPointer: ... - -class Subdiv2D: - # Functions - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def __init__(self, rect: cv2.typing.Rect) -> None: ... - def initDelaunay(self, rect: cv2.typing.Rect) -> None: ... - @typing.overload - def insert(self, pt: cv2.typing.Point2f) -> int: ... - @typing.overload - def insert(self, ptvec: typing.Sequence[cv2.typing.Point2f]) -> None: ... - def locate(self, pt: cv2.typing.Point2f) -> tuple[int, int, int]: ... - def findNearest(self, pt: cv2.typing.Point2f) -> tuple[int, cv2.typing.Point2f]: ... - def getEdgeList(self) -> typing.Sequence[cv2.typing.Vec4f]: ... - def getLeadingEdgeList(self) -> typing.Sequence[int]: ... - def getTriangleList(self) -> typing.Sequence[cv2.typing.Vec6f]: ... - def getVoronoiFacetList( - self, - idx: typing.Sequence[int], - ) -> tuple[ - typing.Sequence[typing.Sequence[cv2.typing.Point2f]], - typing.Sequence[cv2.typing.Point2f], - ]: ... - def getVertex(self, vertex: int) -> tuple[cv2.typing.Point2f, int]: ... - def getEdge(self, edge: int, nextEdgeType: int) -> int: ... - def nextEdge(self, edge: int) -> int: ... - def rotateEdge(self, edge: int, rotate: int) -> int: ... - def symEdge(self, edge: int) -> int: ... - def edgeOrg(self, edge: int) -> tuple[int, cv2.typing.Point2f]: ... - def edgeDst(self, edge: int) -> tuple[int, cv2.typing.Point2f]: ... - -class Feature2D: - # Functions - @typing.overload - def detect(self, image: cv2.typing.MatLike, mask: cv2.typing.MatLike | None = ...) -> typing.Sequence[KeyPoint]: ... - @typing.overload - def detect(self, image: UMat, mask: UMat | None = ...) -> typing.Sequence[KeyPoint]: ... - @typing.overload - def detect( - self, - images: typing.Sequence[cv2.typing.MatLike], - masks: typing.Sequence[cv2.typing.MatLike] | None = ..., - ) -> typing.Sequence[typing.Sequence[KeyPoint]]: ... - @typing.overload - def detect( - self, - images: typing.Sequence[UMat], - masks: typing.Sequence[UMat] | None = ..., - ) -> typing.Sequence[typing.Sequence[KeyPoint]]: ... - @typing.overload - def compute( - self, - image: cv2.typing.MatLike, - keypoints: typing.Sequence[KeyPoint], - descriptors: cv2.typing.MatLike | None = ..., - ) -> tuple[ - typing.Sequence[KeyPoint], - cv2.typing.MatLike, - ]: ... - @typing.overload - def compute( - self, - image: UMat, - keypoints: typing.Sequence[KeyPoint], - descriptors: UMat | None = ..., - ) -> tuple[ - typing.Sequence[KeyPoint], - UMat, - ]: ... - @typing.overload - def compute( - self, - images: typing.Sequence[cv2.typing.MatLike], - keypoints: typing.Sequence[typing.Sequence[KeyPoint]], - descriptors: typing.Sequence[cv2.typing.MatLike] | None = ..., - ) -> tuple[ - typing.Sequence[typing.Sequence[KeyPoint]], - typing.Sequence[cv2.typing.MatLike], - ]: ... - @typing.overload - def compute( - self, - images: typing.Sequence[UMat], - keypoints: typing.Sequence[typing.Sequence[KeyPoint]], - descriptors: typing.Sequence[UMat] | None = ..., - ) -> tuple[ - typing.Sequence[typing.Sequence[KeyPoint]], - typing.Sequence[UMat], - ]: ... - @typing.overload - def detectAndCompute( - self, - image: cv2.typing.MatLike, - mask: cv2.typing.MatLike, - descriptors: cv2.typing.MatLike | None = ..., - useProvidedKeypoints: bool = ..., - ) -> tuple[ - typing.Sequence[KeyPoint], - cv2.typing.MatLike, - ]: ... - @typing.overload - def detectAndCompute( - self, - image: UMat, - mask: UMat, - descriptors: UMat | None = ..., - useProvidedKeypoints: bool = ..., - ) -> tuple[ - typing.Sequence[KeyPoint], - UMat, - ]: ... - def descriptorSize(self) -> int: ... - def descriptorType(self) -> int: ... - def defaultNorm(self) -> int: ... - @typing.overload - def write(self, fileName: str) -> None: ... - @typing.overload - def write(self, fs: FileStorage, name: str) -> None: ... - @typing.overload - def read(self, fileName: str) -> None: ... - @typing.overload - def read(self, arg1: FileNode) -> None: ... - def empty(self) -> bool: ... - def getDefaultName(self) -> str: ... - -class BOWTrainer: - # Functions - def add(self, descriptors: cv2.typing.MatLike) -> None: ... - def getDescriptors(self) -> typing.Sequence[cv2.typing.MatLike]: ... - def descriptorsCount(self) -> int: ... - def clear(self) -> None: ... - @typing.overload - def cluster(self) -> cv2.typing.MatLike: ... - @typing.overload - def cluster(self, descriptors: cv2.typing.MatLike) -> cv2.typing.MatLike: ... - -class BOWImgDescriptorExtractor: - # Functions - def __init__(self, dextractor: cv2.typing.DescriptorExtractor, dmatcher: DescriptorMatcher) -> None: ... - def setVocabulary(self, vocabulary: cv2.typing.MatLike) -> None: ... - def getVocabulary(self) -> cv2.typing.MatLike: ... - def compute( - self, - image: cv2.typing.MatLike, - keypoints: typing.Sequence[KeyPoint], - imgDescriptor: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - def descriptorSize(self) -> int: ... - def descriptorType(self) -> int: ... - -class VideoCapture: - # Functions - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def __init__(self, filename: str, apiPreference: int = ...) -> None: ... - @typing.overload - def __init__(self, filename: str, apiPreference: int, params: typing.Sequence[int]) -> None: ... - @typing.overload - def __init__(self, index: int, apiPreference: int = ...) -> None: ... - @typing.overload - def __init__(self, index: int, apiPreference: int, params: typing.Sequence[int]) -> None: ... - @typing.overload - def open(self, filename: str, apiPreference: int = ...) -> bool: ... - @typing.overload - def open(self, filename: str, apiPreference: int, params: typing.Sequence[int]) -> bool: ... - @typing.overload - def open(self, index: int, apiPreference: int = ...) -> bool: ... - @typing.overload - def open(self, index: int, apiPreference: int, params: typing.Sequence[int]) -> bool: ... - def isOpened(self) -> bool: ... - def release(self) -> None: ... - def grab(self) -> bool: ... - @typing.overload - def retrieve(self, image: cv2.typing.MatLike | None = ..., flag: int = ...) -> tuple[bool, cv2.typing.MatLike]: ... - @typing.overload - def retrieve(self, image: UMat | None = ..., flag: int = ...) -> tuple[bool, UMat]: ... - @typing.overload - def read(self, image: cv2.typing.MatLike | None = ...) -> tuple[bool, cv2.typing.MatLike]: ... - @typing.overload - def read(self, image: UMat | None = ...) -> tuple[bool, UMat]: ... - def set(self, propId: int, value: float) -> bool: ... - def get(self, propId: int) -> float: ... - def getBackendName(self) -> str: ... - def setExceptionMode(self, enable: bool) -> None: ... - def getExceptionMode(self) -> bool: ... - @staticmethod - def waitAny(streams: typing.Sequence[VideoCapture], timeoutNs: int = ...) -> tuple[bool, typing.Sequence[int]]: ... - -class VideoWriter: - # Functions - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def __init__( - self, - filename: str, - fourcc: int, - fps: float, - frameSize: cv2.typing.Size, - isColor: bool = ..., - ) -> None: ... - @typing.overload - def __init__( - self, - filename: str, - apiPreference: int, - fourcc: int, - fps: float, - frameSize: cv2.typing.Size, - isColor: bool = ..., - ) -> None: ... - @typing.overload - def __init__( - self, - filename: str, - fourcc: int, - fps: float, - frameSize: cv2.typing.Size, - params: typing.Sequence[int], - ) -> None: ... - @typing.overload - def __init__( - self, - filename: str, - apiPreference: int, - fourcc: int, - fps: float, - frameSize: cv2.typing.Size, - params: typing.Sequence[int], - ) -> None: ... - @typing.overload - def open(self, filename: str, fourcc: int, fps: float, frameSize: cv2.typing.Size, isColor: bool = ...) -> bool: ... - @typing.overload - def open( - self, - filename: str, - apiPreference: int, - fourcc: int, - fps: float, - frameSize: cv2.typing.Size, - isColor: bool = ..., - ) -> bool: ... - @typing.overload - def open( - self, - filename: str, - fourcc: int, - fps: float, - frameSize: cv2.typing.Size, - params: typing.Sequence[int], - ) -> bool: ... - @typing.overload - def open( - self, - filename: str, - apiPreference: int, - fourcc: int, - fps: float, - frameSize: cv2.typing.Size, - params: typing.Sequence[int], - ) -> bool: ... - def isOpened(self) -> bool: ... - def release(self) -> None: ... - @typing.overload - def write(self, image: cv2.typing.MatLike) -> None: ... - @typing.overload - def write(self, image: UMat) -> None: ... - def set(self, propId: int, value: float) -> bool: ... - def get(self, propId: int) -> float: ... - @staticmethod - def fourcc(c1: str, c2: str, c3: str, c4: str) -> int: ... - def getBackendName(self) -> str: ... - -class UsacParams: - confidence: float - isParallel: bool - loIterations: int - loMethod: LocalOptimMethod - loSampleSize: int - maxIterations: int - neighborsSearch: NeighborSearchMethod - randomGeneratorState: int - sampler: SamplingMethod - score: ScoreMethod - threshold: float - final_polisher: PolishingMethod - final_polisher_iterations: int - - # Functions - def __init__(self) -> None: ... - -class CirclesGridFinderParameters: - densityNeighborhoodSize: cv2.typing.Size2f - minDensity: float - kmeansAttempts: int - minDistanceToAddKeypoint: int - keypointScale: int - minGraphConfidence: float - vertexGain: float - vertexPenalty: float - existingVertexGain: float - edgeGain: float - edgePenalty: float - convexHullFactor: float - minRNGEdgeSwitchDist: float - squareSize: float - maxRectifiedDistance: float - - # Functions - def __init__(self) -> None: ... - -class CascadeClassifier: - # Functions - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def __init__(self, filename: str) -> None: ... - def empty(self) -> bool: ... - def load(self, filename: str) -> bool: ... - def read(self, node: FileNode) -> bool: ... - @typing.overload - def detectMultiScale( - self, - image: cv2.typing.MatLike, - scaleFactor: float = ..., - minNeighbors: int = ..., - flags: int = ..., - minSize: cv2.typing.Size = ..., - maxSize: cv2.typing.Size = ..., - ) -> typing.Sequence[cv2.typing.Rect]: ... - @typing.overload - def detectMultiScale( - self, - image: UMat, - scaleFactor: float = ..., - minNeighbors: int = ..., - flags: int = ..., - minSize: cv2.typing.Size = ..., - maxSize: cv2.typing.Size = ..., - ) -> typing.Sequence[cv2.typing.Rect]: ... - @typing.overload - def detectMultiScale2( - self, - image: cv2.typing.MatLike, - scaleFactor: float = ..., - minNeighbors: int = ..., - flags: int = ..., - minSize: cv2.typing.Size = ..., - maxSize: cv2.typing.Size = ..., - ) -> tuple[ - typing.Sequence[cv2.typing.Rect], - typing.Sequence[int], - ]: ... - @typing.overload - def detectMultiScale2( - self, - image: UMat, - scaleFactor: float = ..., - minNeighbors: int = ..., - flags: int = ..., - minSize: cv2.typing.Size = ..., - maxSize: cv2.typing.Size = ..., - ) -> tuple[ - typing.Sequence[cv2.typing.Rect], - typing.Sequence[int], - ]: ... - @typing.overload - def detectMultiScale3( - self, - image: cv2.typing.MatLike, - scaleFactor: float = ..., - minNeighbors: int = ..., - flags: int = ..., - minSize: cv2.typing.Size = ..., - maxSize: cv2.typing.Size = ..., - outputRejectLevels: bool = ..., - ) -> tuple[ - typing.Sequence[cv2.typing.Rect], - typing.Sequence[int], - typing.Sequence[float], - ]: ... - @typing.overload - def detectMultiScale3( - self, - image: UMat, - scaleFactor: float = ..., - minNeighbors: int = ..., - flags: int = ..., - minSize: cv2.typing.Size = ..., - maxSize: cv2.typing.Size = ..., - outputRejectLevels: bool = ..., - ) -> tuple[ - typing.Sequence[cv2.typing.Rect], - typing.Sequence[int], - typing.Sequence[float], - ]: ... - def isOldFormatCascade(self) -> bool: ... - def getOriginalWindowSize(self) -> cv2.typing.Size: ... - def getFeatureType(self) -> int: ... - @staticmethod - def convert(oldcascade: str, newcascade: str) -> bool: ... - -class HOGDescriptor: - @property - def winSize(self) -> cv2.typing.Size: ... - @property - def blockSize(self) -> cv2.typing.Size: ... - @property - def blockStride(self) -> cv2.typing.Size: ... - @property - def cellSize(self) -> cv2.typing.Size: ... - @property - def nbins(self) -> int: ... - @property - def derivAperture(self) -> int: ... - @property - def winSigma(self) -> float: ... - @property - def histogramNormType(self) -> HOGDescriptor_HistogramNormType: ... - @property - def L2HysThreshold(self) -> float: ... - @property - def gammaCorrection(self) -> bool: ... - @property - def svmDetector(self) -> typing.Sequence[float]: ... - @property - def nlevels(self) -> int: ... - @property - def signedGradient(self) -> bool: ... - - # Functions - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def __init__( - self, - _winSize: cv2.typing.Size, - _blockSize: cv2.typing.Size, - _blockStride: cv2.typing.Size, - _cellSize: cv2.typing.Size, - _nbins: int, - _derivAperture: int = ..., - _winSigma: float = ..., - _histogramNormType: HOGDescriptor_HistogramNormType = ..., - _L2HysThreshold: float = ..., - _gammaCorrection: bool = ..., - _nlevels: int = ..., - _signedGradient: bool = ..., - ) -> None: ... - @typing.overload - def __init__(self, filename: str) -> None: ... - def getDescriptorSize(self) -> int: ... - def checkDetectorSize(self) -> bool: ... - def getWinSigma(self) -> float: ... - @typing.overload - def setSVMDetector(self, svmdetector: cv2.typing.MatLike) -> None: ... - @typing.overload - def setSVMDetector(self, svmdetector: UMat) -> None: ... - def load(self, filename: str, objname: str = ...) -> bool: ... - def save(self, filename: str, objname: str = ...) -> None: ... - @typing.overload - def compute( - self, - img: cv2.typing.MatLike, - winStride: cv2.typing.Size = ..., - padding: cv2.typing.Size = ..., - locations: typing.Sequence[cv2.typing.Point] = ..., - ) -> typing.Sequence[float]: ... - @typing.overload - def compute( - self, - img: UMat, - winStride: cv2.typing.Size = ..., - padding: cv2.typing.Size = ..., - locations: typing.Sequence[cv2.typing.Point] = ..., - ) -> typing.Sequence[float]: ... - @typing.overload - def detect( - self, - img: cv2.typing.MatLike, - hitThreshold: float = ..., - winStride: cv2.typing.Size = ..., - padding: cv2.typing.Size = ..., - searchLocations: typing.Sequence[cv2.typing.Point] = ..., - ) -> tuple[ - typing.Sequence[cv2.typing.Point], - typing.Sequence[float], - ]: ... - @typing.overload - def detect( - self, - img: UMat, - hitThreshold: float = ..., - winStride: cv2.typing.Size = ..., - padding: cv2.typing.Size = ..., - searchLocations: typing.Sequence[cv2.typing.Point] = ..., - ) -> tuple[ - typing.Sequence[cv2.typing.Point], - typing.Sequence[float], - ]: ... - @typing.overload - def detectMultiScale( - self, - img: cv2.typing.MatLike, - hitThreshold: float = ..., - winStride: cv2.typing.Size = ..., - padding: cv2.typing.Size = ..., - scale: float = ..., - groupThreshold: float = ..., - useMeanshiftGrouping: bool = ..., - ) -> tuple[ - typing.Sequence[cv2.typing.Rect], - typing.Sequence[float], - ]: ... - @typing.overload - def detectMultiScale( - self, - img: UMat, - hitThreshold: float = ..., - winStride: cv2.typing.Size = ..., - padding: cv2.typing.Size = ..., - scale: float = ..., - groupThreshold: float = ..., - useMeanshiftGrouping: bool = ..., - ) -> tuple[ - typing.Sequence[cv2.typing.Rect], - typing.Sequence[float], - ]: ... - @typing.overload - def computeGradient( - self, - img: cv2.typing.MatLike, - grad: cv2.typing.MatLike, - angleOfs: cv2.typing.MatLike, - paddingTL: cv2.typing.Size = ..., - paddingBR: cv2.typing.Size = ..., - ) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - ]: ... - @typing.overload - def computeGradient( - self, - img: UMat, - grad: UMat, - angleOfs: UMat, - paddingTL: cv2.typing.Size = ..., - paddingBR: cv2.typing.Size = ..., - ) -> tuple[ - UMat, - UMat, - ]: ... - @staticmethod - def getDefaultPeopleDetector() -> typing.Sequence[float]: ... - @staticmethod - def getDaimlerPeopleDetector() -> typing.Sequence[float]: ... - -class QRCodeEncoder: - # Classes - class Params: - version: int - correction_level: QRCodeEncoder_CorrectionLevel - mode: QRCodeEncoder_EncodeMode - structure_number: int - - # Functions - def __init__(self) -> None: ... - - # Functions - - @classmethod - def create(cls, parameters: QRCodeEncoder.Params = ...) -> QRCodeEncoder: ... - @typing.overload - def encode(self, encoded_info: str, qrcode: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... - @typing.overload - def encode(self, encoded_info: str, qrcode: UMat | None = ...) -> UMat: ... - @typing.overload - def encodeStructuredAppend( - self, - encoded_info: str, - qrcodes: typing.Sequence[cv2.typing.MatLike] | None = ..., - ) -> typing.Sequence[cv2.typing.MatLike]: ... - @typing.overload - def encodeStructuredAppend( - self, - encoded_info: str, - qrcodes: typing.Sequence[UMat] | None = ..., - ) -> typing.Sequence[UMat]: ... - -class GraphicalCodeDetector: - # Functions - @typing.overload - def detect( - self, - img: cv2.typing.MatLike, - points: cv2.typing.MatLike | None = ..., - ) -> tuple[ - bool, - cv2.typing.MatLike, - ]: ... - @typing.overload - def detect(self, img: UMat, points: UMat | None = ...) -> tuple[bool, UMat]: ... - @typing.overload - def decode( - self, - img: cv2.typing.MatLike, - points: cv2.typing.MatLike, - straight_code: cv2.typing.MatLike | None = ..., - ) -> tuple[ - str, - cv2.typing.MatLike, - ]: ... - @typing.overload - def decode(self, img: UMat, points: UMat, straight_code: UMat | None = ...) -> tuple[str, UMat]: ... - @typing.overload - def detectAndDecode( - self, - img: cv2.typing.MatLike, - points: cv2.typing.MatLike | None = ..., - straight_code: cv2.typing.MatLike | None = ..., - ) -> tuple[ - str, - cv2.typing.MatLike, - cv2.typing.MatLike, - ]: ... - @typing.overload - def detectAndDecode( - self, - img: UMat, - points: UMat | None = ..., - straight_code: UMat | None = ..., - ) -> tuple[ - str, - UMat, - UMat, - ]: ... - @typing.overload - def detectMulti( - self, - img: cv2.typing.MatLike, - points: cv2.typing.MatLike | None = ..., - ) -> tuple[ - bool, - cv2.typing.MatLike, - ]: ... - @typing.overload - def detectMulti(self, img: UMat, points: UMat | None = ...) -> tuple[bool, UMat]: ... - @typing.overload - def decodeMulti( - self, - img: cv2.typing.MatLike, - points: cv2.typing.MatLike, - straight_code: typing.Sequence[cv2.typing.MatLike] | None = ..., - ) -> tuple[ - bool, - typing.Sequence[str], - typing.Sequence[cv2.typing.MatLike], - ]: ... - @typing.overload - def decodeMulti( - self, - img: UMat, - points: UMat, - straight_code: typing.Sequence[UMat] | None = ..., - ) -> tuple[ - bool, - typing.Sequence[str], - typing.Sequence[UMat], - ]: ... - @typing.overload - def detectAndDecodeMulti( - self, - img: cv2.typing.MatLike, - points: cv2.typing.MatLike | None = ..., - straight_code: typing.Sequence[cv2.typing.MatLike] | None = ..., - ) -> tuple[ - bool, - typing.Sequence[str], - cv2.typing.MatLike, - typing.Sequence[cv2.typing.MatLike], - ]: ... - @typing.overload - def detectAndDecodeMulti( - self, - img: UMat, - points: UMat | None = ..., - straight_code: typing.Sequence[UMat] | None = ..., - ) -> tuple[ - bool, - typing.Sequence[str], - UMat, - typing.Sequence[UMat], - ]: ... - -class FaceDetectorYN: - # Functions - def setInputSize(self, input_size: cv2.typing.Size) -> None: ... - def getInputSize(self) -> cv2.typing.Size: ... - def setScoreThreshold(self, score_threshold: float) -> None: ... - def getScoreThreshold(self) -> float: ... - def setNMSThreshold(self, nms_threshold: float) -> None: ... - def getNMSThreshold(self) -> float: ... - def setTopK(self, top_k: int) -> None: ... - def getTopK(self) -> int: ... - @typing.overload - def detect( - self, - image: cv2.typing.MatLike, - faces: cv2.typing.MatLike | None = ..., - ) -> tuple[ - int, - cv2.typing.MatLike, - ]: ... - @typing.overload - def detect(self, image: UMat, faces: UMat | None = ...) -> tuple[int, UMat]: ... - @classmethod - def create( - cls, - model: str, - config: str, - input_size: cv2.typing.Size, - score_threshold: float = ..., - nms_threshold: float = ..., - top_k: int = ..., - backend_id: int = ..., - target_id: int = ..., - ) -> FaceDetectorYN: ... - -class FaceRecognizerSF: - # Functions - @typing.overload - def alignCrop( - self, - src_img: cv2.typing.MatLike, - face_box: cv2.typing.MatLike, - aligned_img: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def alignCrop(self, src_img: UMat, face_box: UMat, aligned_img: UMat | None = ...) -> UMat: ... - @typing.overload - def feature( - self, - aligned_img: cv2.typing.MatLike, - face_feature: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def feature(self, aligned_img: UMat, face_feature: UMat | None = ...) -> UMat: ... - @typing.overload - def match( - self, - face_feature1: cv2.typing.MatLike, - face_feature2: cv2.typing.MatLike, - dis_type: int = ..., - ) -> float: ... - @typing.overload - def match(self, face_feature1: UMat, face_feature2: UMat, dis_type: int = ...) -> float: ... - @classmethod - def create(cls, model: str, config: str, backend_id: int = ..., target_id: int = ...) -> FaceRecognizerSF: ... - -class Stitcher: - # Functions - @classmethod - def create(cls, mode: Stitcher_Mode = ...) -> Stitcher: ... - def registrationResol(self) -> float: ... - def setRegistrationResol(self, resol_mpx: float) -> None: ... - def seamEstimationResol(self) -> float: ... - def setSeamEstimationResol(self, resol_mpx: float) -> None: ... - def compositingResol(self) -> float: ... - def setCompositingResol(self, resol_mpx: float) -> None: ... - def panoConfidenceThresh(self) -> float: ... - def setPanoConfidenceThresh(self, conf_thresh: float) -> None: ... - def waveCorrection(self) -> bool: ... - def setWaveCorrection(self, flag: bool) -> None: ... - def interpolationFlags(self) -> InterpolationFlags: ... - def setInterpolationFlags(self, interp_flags: InterpolationFlags) -> None: ... - @typing.overload - def estimateTransform( - self, - images: typing.Sequence[cv2.typing.MatLike], - masks: typing.Sequence[cv2.typing.MatLike] | None = ..., - ) -> Stitcher_Status: ... - @typing.overload - def estimateTransform( - self, - images: typing.Sequence[UMat], - masks: typing.Sequence[UMat] | None = ..., - ) -> Stitcher_Status: ... - @typing.overload - def composePanorama(self, pano: cv2.typing.MatLike | None = ...) -> tuple[Stitcher_Status, cv2.typing.MatLike]: ... - @typing.overload - def composePanorama(self, pano: UMat | None = ...) -> tuple[Stitcher_Status, UMat]: ... - @typing.overload - def composePanorama( - self, - images: typing.Sequence[cv2.typing.MatLike], - pano: cv2.typing.MatLike | None = ..., - ) -> tuple[ - Stitcher_Status, - cv2.typing.MatLike, - ]: ... - @typing.overload - def composePanorama( - self, - images: typing.Sequence[UMat], - pano: UMat | None = ..., - ) -> tuple[ - Stitcher_Status, - UMat, - ]: ... - @typing.overload - def stitch( - self, - images: typing.Sequence[cv2.typing.MatLike], - pano: cv2.typing.MatLike | None = ..., - ) -> tuple[ - Stitcher_Status, - cv2.typing.MatLike, - ]: ... - @typing.overload - def stitch(self, images: typing.Sequence[UMat], pano: UMat | None = ...) -> tuple[Stitcher_Status, UMat]: ... - @typing.overload - def stitch( - self, - images: typing.Sequence[cv2.typing.MatLike], - masks: typing.Sequence[cv2.typing.MatLike], - pano: cv2.typing.MatLike | None = ..., - ) -> tuple[ - Stitcher_Status, - cv2.typing.MatLike, - ]: ... - @typing.overload - def stitch( - self, - images: typing.Sequence[UMat], - masks: typing.Sequence[UMat], - pano: UMat | None = ..., - ) -> tuple[ - Stitcher_Status, - UMat, - ]: ... - def workScale(self) -> float: ... - -class PyRotationWarper: - # Functions - @typing.overload - def __init__(self, type: str, scale: float) -> None: ... - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def warpPoint(self, pt: cv2.typing.Point2f, K: cv2.typing.MatLike, R: cv2.typing.MatLike) -> cv2.typing.Point2f: ... - @typing.overload - def warpPoint(self, pt: cv2.typing.Point2f, K: UMat, R: UMat) -> cv2.typing.Point2f: ... - @typing.overload - def warpPointBackward( - self, - pt: cv2.typing.Point2f, - K: cv2.typing.MatLike, - R: cv2.typing.MatLike, - ) -> cv2.typing.Point2f: ... - @typing.overload - def warpPointBackward(self, pt: cv2.typing.Point2f, K: UMat, R: UMat) -> cv2.typing.Point2f: ... - @typing.overload - def warpPointBackward( - self, - pt: cv2.typing.Point2f, - K: cv2.typing.MatLike, - R: cv2.typing.MatLike, - ) -> cv2.typing.Point2f: ... - @typing.overload - def warpPointBackward(self, pt: cv2.typing.Point2f, K: UMat, R: UMat) -> cv2.typing.Point2f: ... - @typing.overload - def buildMaps( - self, - src_size: cv2.typing.Size, - K: cv2.typing.MatLike, - R: cv2.typing.MatLike, - xmap: cv2.typing.MatLike | None = ..., - ymap: cv2.typing.MatLike | None = ..., - ) -> tuple[ - cv2.typing.Rect, - cv2.typing.MatLike, - cv2.typing.MatLike, - ]: ... - @typing.overload - def buildMaps( - self, - src_size: cv2.typing.Size, - K: UMat, - R: UMat, - xmap: UMat | None = ..., - ymap: UMat | None = ..., - ) -> tuple[ - cv2.typing.Rect, - UMat, - UMat, - ]: ... - @typing.overload - def warp( - self, - src: cv2.typing.MatLike, - K: cv2.typing.MatLike, - R: cv2.typing.MatLike, - interp_mode: int, - border_mode: int, - dst: cv2.typing.MatLike | None = ..., - ) -> tuple[ - cv2.typing.Point, - cv2.typing.MatLike, - ]: ... - @typing.overload - def warp( - self, - src: UMat, - K: UMat, - R: UMat, - interp_mode: int, - border_mode: int, - dst: UMat | None = ..., - ) -> tuple[ - cv2.typing.Point, - UMat, - ]: ... - @typing.overload - def warpBackward( - self, - src: cv2.typing.MatLike, - K: cv2.typing.MatLike, - R: cv2.typing.MatLike, - interp_mode: int, - border_mode: int, - dst_size: cv2.typing.Size, - dst: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def warpBackward( - self, - src: UMat, - K: UMat, - R: UMat, - interp_mode: int, - border_mode: int, - dst_size: cv2.typing.Size, - dst: UMat | None = ..., - ) -> UMat: ... - @typing.overload - def warpRoi(self, src_size: cv2.typing.Size, K: cv2.typing.MatLike, R: cv2.typing.MatLike) -> cv2.typing.Rect: ... - @typing.overload - def warpRoi(self, src_size: cv2.typing.Size, K: UMat, R: UMat) -> cv2.typing.Rect: ... - def getScale(self) -> float: ... - def setScale(self, arg1: float) -> None: ... - -class WarperCreator: ... - -class KalmanFilter: - statePre: cv2.typing.MatLike - statePost: cv2.typing.MatLike - transitionMatrix: cv2.typing.MatLike - controlMatrix: cv2.typing.MatLike - measurementMatrix: cv2.typing.MatLike - processNoiseCov: cv2.typing.MatLike - measurementNoiseCov: cv2.typing.MatLike - errorCovPre: cv2.typing.MatLike - gain: cv2.typing.MatLike - errorCovPost: cv2.typing.MatLike - - # Functions - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def __init__(self, dynamParams: int, measureParams: int, controlParams: int = ..., type: int = ...) -> None: ... - def predict(self, control: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... - def correct(self, measurement: cv2.typing.MatLike) -> cv2.typing.MatLike: ... - -class Tracker: - # Functions - @typing.overload - def init(self, image: cv2.typing.MatLike, boundingBox: cv2.typing.Rect) -> None: ... - @typing.overload - def init(self, image: UMat, boundingBox: cv2.typing.Rect) -> None: ... - @typing.overload - def update(self, image: cv2.typing.MatLike) -> tuple[bool, cv2.typing.Rect]: ... - @typing.overload - def update(self, image: UMat) -> tuple[bool, cv2.typing.Rect]: ... - -class GArrayDesc: ... - -class GComputation: - # Functions - @typing.overload - def __init__(self, ins: cv2.typing.GProtoInputArgs, outs: cv2.typing.GProtoOutputArgs) -> None: ... - @typing.overload - def __init__(self, in_: GMat, out: GMat) -> None: ... - @typing.overload - def __init__(self, in_: GMat, out: GScalar) -> None: ... - @typing.overload - def __init__(self, in1: GMat, in2: GMat, out: GMat) -> None: ... - def apply( - self, - callback: cv2.typing.ExtractArgsCallback, - args: typing.Sequence[GCompileArg] = ..., - ) -> typing.Sequence[cv2.typing.GRunArg]: ... - @typing.overload - def compileStreaming( - self, - in_metas: typing.Sequence[cv2.typing.GMetaArg], - args: typing.Sequence[GCompileArg] = ..., - ) -> GStreamingCompiled: ... - @typing.overload - def compileStreaming(self, args: typing.Sequence[GCompileArg] = ...) -> GStreamingCompiled: ... - @typing.overload - def compileStreaming( - self, - callback: cv2.typing.ExtractMetaCallback, - args: typing.Sequence[GCompileArg] = ..., - ) -> GStreamingCompiled: ... - -class GFrame: - # Functions - def __init__(self) -> None: ... - -class GKernelPackage: - # Functions - def size(self) -> int: ... - -class GMat: - # Functions - def __init__(self) -> None: ... - -class GMatDesc: - @property - def depth(self) -> int: ... - @property - def chan(self) -> int: ... - @property - def size(self) -> cv2.typing.Size: ... - @property - def planar(self) -> bool: ... - @property - def dims(self) -> typing.Sequence[int]: ... - - # Functions - @typing.overload - def __init__(self, d: int, c: int, s: cv2.typing.Size, p: bool = ...) -> None: ... - @typing.overload - def __init__(self, d: int, dd: typing.Sequence[int]) -> None: ... - @typing.overload - def __init__(self, d: int, dd: typing.Sequence[int]) -> None: ... - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def withSizeDelta(self, delta: cv2.typing.Size) -> GMatDesc: ... - @typing.overload - def withSizeDelta(self, dx: int, dy: int) -> GMatDesc: ... - def withSize(self, sz: cv2.typing.Size) -> GMatDesc: ... - def withDepth(self, ddepth: int) -> GMatDesc: ... - def withType(self, ddepth: int, dchan: int) -> GMatDesc: ... - @typing.overload - def asPlanar(self) -> GMatDesc: ... - @typing.overload - def asPlanar(self, planes: int) -> GMatDesc: ... - def asInterleaved(self) -> GMatDesc: ... - -class GOpaqueDesc: ... - -class GScalar: - # Functions - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def __init__(self, s: cv2.typing.Scalar) -> None: ... - -class GScalarDesc: ... - -class GStreamingCompiled: - # Functions - def __init__(self) -> None: ... - def setSource(self, callback: cv2.typing.ExtractArgsCallback) -> None: ... - def start(self) -> None: ... - def pull(self) -> tuple[bool, typing.Sequence[cv2.typing.GRunArg] | typing.Sequence[cv2.typing.GOptRunArg]]: ... - def stop(self) -> None: ... - def running(self) -> bool: ... - -class GOpaqueT: - # Functions - def __init__(self, type: cv2.gapi.ArgType) -> None: ... - def type(self) -> cv2.gapi.ArgType: ... - -class GArrayT: - # Functions - def __init__(self, type: cv2.gapi.ArgType) -> None: ... - def type(self) -> cv2.gapi.ArgType: ... - -class GCompileArg: - # Functions - @typing.overload - def __init__(self, arg: GKernelPackage) -> None: ... - @typing.overload - def __init__(self, arg: cv2.gapi.GNetPackage) -> None: ... - @typing.overload - def __init__(self, arg: cv2.gapi.streaming.queue_capacity) -> None: ... - -class GInferInputs: - # Functions - def __init__(self) -> None: ... - @typing.overload - def setInput(self, name: str, value: GMat) -> GInferInputs: ... - @typing.overload - def setInput(self, name: str, value: GFrame) -> GInferInputs: ... - -class GInferListInputs: - # Functions - def __init__(self) -> None: ... - @typing.overload - def setInput(self, name: str, value: GArrayT) -> GInferListInputs: ... - @typing.overload - def setInput(self, name: str, value: GArrayT) -> GInferListInputs: ... - -class GInferOutputs: - # Functions - def __init__(self) -> None: ... - def at(self, name: str) -> GMat: ... - -class GInferListOutputs: - # Functions - def __init__(self) -> None: ... - def at(self, name: str) -> GArrayT: ... - -class GeneralizedHough(Algorithm): - # Functions - @typing.overload - def setTemplate(self, templ: cv2.typing.MatLike, templCenter: cv2.typing.Point = ...) -> None: ... - @typing.overload - def setTemplate(self, templ: UMat, templCenter: cv2.typing.Point = ...) -> None: ... - @typing.overload - def setTemplate( - self, - edges: cv2.typing.MatLike, - dx: cv2.typing.MatLike, - dy: cv2.typing.MatLike, - templCenter: cv2.typing.Point = ..., - ) -> None: ... - @typing.overload - def setTemplate(self, edges: UMat, dx: UMat, dy: UMat, templCenter: cv2.typing.Point = ...) -> None: ... - @typing.overload - def detect( - self, - image: cv2.typing.MatLike, - positions: cv2.typing.MatLike | None = ..., - votes: cv2.typing.MatLike | None = ..., - ) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - ]: ... - @typing.overload - def detect(self, image: UMat, positions: UMat | None = ..., votes: UMat | None = ...) -> tuple[UMat, UMat]: ... - @typing.overload - def detect( - self, - edges: cv2.typing.MatLike, - dx: cv2.typing.MatLike, - dy: cv2.typing.MatLike, - positions: cv2.typing.MatLike | None = ..., - votes: cv2.typing.MatLike | None = ..., - ) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - ]: ... - @typing.overload - def detect( - self, - edges: UMat, - dx: UMat, - dy: UMat, - positions: UMat | None = ..., - votes: UMat | None = ..., - ) -> tuple[ - UMat, - UMat, - ]: ... - def setCannyLowThresh(self, cannyLowThresh: int) -> None: ... - def getCannyLowThresh(self) -> int: ... - def setCannyHighThresh(self, cannyHighThresh: int) -> None: ... - def getCannyHighThresh(self) -> int: ... - def setMinDist(self, minDist: float) -> None: ... - def getMinDist(self) -> float: ... - def setDp(self, dp: float) -> None: ... - def getDp(self) -> float: ... - def setMaxBufferSize(self, maxBufferSize: int) -> None: ... - def getMaxBufferSize(self) -> int: ... - -class CLAHE(Algorithm): - # Functions - @typing.overload - def apply(self, src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... - @typing.overload - def apply(self, src: UMat, dst: UMat | None = ...) -> UMat: ... - def setClipLimit(self, clipLimit: float) -> None: ... - def getClipLimit(self) -> float: ... - def setTilesGridSize(self, tileGridSize: cv2.typing.Size) -> None: ... - def getTilesGridSize(self) -> cv2.typing.Size: ... - def collectGarbage(self) -> None: ... - -class LineSegmentDetector(Algorithm): - # Functions - @typing.overload - def detect( - self, - image: cv2.typing.MatLike, - lines: cv2.typing.MatLike | None = ..., - width: cv2.typing.MatLike | None = ..., - prec: cv2.typing.MatLike | None = ..., - nfa: cv2.typing.MatLike | None = ..., - ) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - ]: ... - @typing.overload - def detect( - self, - image: UMat, - lines: UMat | None = ..., - width: UMat | None = ..., - prec: UMat | None = ..., - nfa: UMat | None = ..., - ) -> tuple[ - UMat, - UMat, - UMat, - UMat, - ]: ... - @typing.overload - def drawSegments(self, image: cv2.typing.MatLike, lines: cv2.typing.MatLike) -> cv2.typing.MatLike: ... - @typing.overload - def drawSegments(self, image: UMat, lines: UMat) -> UMat: ... - @typing.overload - def compareSegments( - self, - size: cv2.typing.Size, - lines1: cv2.typing.MatLike, - lines2: cv2.typing.MatLike, - image: cv2.typing.MatLike | None = ..., - ) -> tuple[ - int, - cv2.typing.MatLike, - ]: ... - @typing.overload - def compareSegments( - self, - size: cv2.typing.Size, - lines1: UMat, - lines2: UMat, - image: UMat | None = ..., - ) -> tuple[ - int, - UMat, - ]: ... - -class Tonemap(Algorithm): - # Functions - @typing.overload - def process(self, src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... - @typing.overload - def process(self, src: UMat, dst: UMat | None = ...) -> UMat: ... - def getGamma(self) -> float: ... - def setGamma(self, gamma: float) -> None: ... - -class AlignExposures(Algorithm): - # Functions - @typing.overload - def process( - self, - src: typing.Sequence[cv2.typing.MatLike], - dst: typing.Sequence[cv2.typing.MatLike], - times: cv2.typing.MatLike, - response: cv2.typing.MatLike, - ) -> None: ... - @typing.overload - def process( - self, - src: typing.Sequence[UMat], - dst: typing.Sequence[cv2.typing.MatLike], - times: UMat, - response: UMat, - ) -> None: ... - -class CalibrateCRF(Algorithm): - # Functions - @typing.overload - def process( - self, - src: typing.Sequence[cv2.typing.MatLike], - times: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def process(self, src: typing.Sequence[UMat], times: UMat, dst: UMat | None = ...) -> UMat: ... - -class MergeExposures(Algorithm): - # Functions - @typing.overload - def process( - self, - src: typing.Sequence[cv2.typing.MatLike], - times: cv2.typing.MatLike, - response: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def process(self, src: typing.Sequence[UMat], times: UMat, response: UMat, dst: UMat | None = ...) -> UMat: ... - -class AffineFeature(Feature2D): - # Functions - @classmethod - def create( - cls, - backend: Feature2D, - maxTilt: int = ..., - minTilt: int = ..., - tiltStep: float = ..., - rotateStepBase: float = ..., - ) -> AffineFeature: ... - def setViewParams(self, tilts: typing.Sequence[float], rolls: typing.Sequence[float]) -> None: ... - def getViewParams(self, tilts: typing.Sequence[float], rolls: typing.Sequence[float]) -> None: ... - def getDefaultName(self) -> str: ... - -class SIFT(Feature2D): - # Functions - @classmethod - @typing.overload - def create( - cls, - nfeatures: int = ..., - nOctaveLayers: int = ..., - contrastThreshold: float = ..., - edgeThreshold: float = ..., - sigma: float = ..., - enable_precise_upscale: bool = ..., - ) -> SIFT: ... - @classmethod - @typing.overload - def create( - cls, - nfeatures: int, - nOctaveLayers: int, - contrastThreshold: float, - edgeThreshold: float, - sigma: float, - descriptorType: int, - enable_precise_upscale: bool = ..., - ) -> SIFT: ... - def getDefaultName(self) -> str: ... - def setNFeatures(self, maxFeatures: int) -> None: ... - def getNFeatures(self) -> int: ... - def setNOctaveLayers(self, nOctaveLayers: int) -> None: ... - def getNOctaveLayers(self) -> int: ... - def setContrastThreshold(self, contrastThreshold: float) -> None: ... - def getContrastThreshold(self) -> float: ... - def setEdgeThreshold(self, edgeThreshold: float) -> None: ... - def getEdgeThreshold(self) -> float: ... - def setSigma(self, sigma: float) -> None: ... - def getSigma(self) -> float: ... - -class BRISK(Feature2D): - # Functions - @classmethod - @typing.overload - def create(cls, thresh: int = ..., octaves: int = ..., patternScale: float = ...) -> BRISK: ... - @classmethod - @typing.overload - def create( - cls, - radiusList: typing.Sequence[float], - numberList: typing.Sequence[int], - dMax: float = ..., - dMin: float = ..., - indexChange: typing.Sequence[int] = ..., - ) -> BRISK: ... - @classmethod - @typing.overload - def create( - cls, - thresh: int, - octaves: int, - radiusList: typing.Sequence[float], - numberList: typing.Sequence[int], - dMax: float = ..., - dMin: float = ..., - indexChange: typing.Sequence[int] = ..., - ) -> BRISK: ... - def getDefaultName(self) -> str: ... - def setThreshold(self, threshold: int) -> None: ... - def getThreshold(self) -> int: ... - def setOctaves(self, octaves: int) -> None: ... - def getOctaves(self) -> int: ... - def setPatternScale(self, patternScale: float) -> None: ... - def getPatternScale(self) -> float: ... - -class ORB(Feature2D): - # Functions - @classmethod - def create( - cls, - nfeatures: int = ..., - scaleFactor: float = ..., - nlevels: int = ..., - edgeThreshold: int = ..., - firstLevel: int = ..., - WTA_K: int = ..., - scoreType: ORB_ScoreType = ..., - patchSize: int = ..., - fastThreshold: int = ..., - ) -> ORB: ... - def setMaxFeatures(self, maxFeatures: int) -> None: ... - def getMaxFeatures(self) -> int: ... - def setScaleFactor(self, scaleFactor: float) -> None: ... - def getScaleFactor(self) -> float: ... - def setNLevels(self, nlevels: int) -> None: ... - def getNLevels(self) -> int: ... - def setEdgeThreshold(self, edgeThreshold: int) -> None: ... - def getEdgeThreshold(self) -> int: ... - def setFirstLevel(self, firstLevel: int) -> None: ... - def getFirstLevel(self) -> int: ... - def setWTA_K(self, wta_k: int) -> None: ... - def getWTA_K(self) -> int: ... - def setScoreType(self, scoreType: ORB_ScoreType) -> None: ... - def getScoreType(self) -> ORB_ScoreType: ... - def setPatchSize(self, patchSize: int) -> None: ... - def getPatchSize(self) -> int: ... - def setFastThreshold(self, fastThreshold: int) -> None: ... - def getFastThreshold(self) -> int: ... - def getDefaultName(self) -> str: ... - -class MSER(Feature2D): - # Functions - @classmethod - def create( - cls, - delta: int = ..., - min_area: int = ..., - max_area: int = ..., - max_variation: float = ..., - min_diversity: float = ..., - max_evolution: int = ..., - area_threshold: float = ..., - min_margin: float = ..., - edge_blur_size: int = ..., - ) -> MSER: ... - @typing.overload - def detectRegions( - self, - image: cv2.typing.MatLike, - ) -> tuple[ - typing.Sequence[typing.Sequence[cv2.typing.Point]], - typing.Sequence[cv2.typing.Rect], - ]: ... - @typing.overload - def detectRegions( - self, - image: UMat, - ) -> tuple[ - typing.Sequence[typing.Sequence[cv2.typing.Point]], - typing.Sequence[cv2.typing.Rect], - ]: ... - def setDelta(self, delta: int) -> None: ... - def getDelta(self) -> int: ... - def setMinArea(self, minArea: int) -> None: ... - def getMinArea(self) -> int: ... - def setMaxArea(self, maxArea: int) -> None: ... - def getMaxArea(self) -> int: ... - def setMaxVariation(self, maxVariation: float) -> None: ... - def getMaxVariation(self) -> float: ... - def setMinDiversity(self, minDiversity: float) -> None: ... - def getMinDiversity(self) -> float: ... - def setMaxEvolution(self, maxEvolution: int) -> None: ... - def getMaxEvolution(self) -> int: ... - def setAreaThreshold(self, areaThreshold: float) -> None: ... - def getAreaThreshold(self) -> float: ... - def setMinMargin(self, min_margin: float) -> None: ... - def getMinMargin(self) -> float: ... - def setEdgeBlurSize(self, edge_blur_size: int) -> None: ... - def getEdgeBlurSize(self) -> int: ... - def setPass2Only(self, f: bool) -> None: ... - def getPass2Only(self) -> bool: ... - def getDefaultName(self) -> str: ... - -class FastFeatureDetector(Feature2D): - # Functions - @classmethod - def create( - cls, - threshold: int = ..., - nonmaxSuppression: bool = ..., - type: FastFeatureDetector_DetectorType = ..., - ) -> FastFeatureDetector: ... - def setThreshold(self, threshold: int) -> None: ... - def getThreshold(self) -> int: ... - def setNonmaxSuppression(self, f: bool) -> None: ... - def getNonmaxSuppression(self) -> bool: ... - def setType(self, type: FastFeatureDetector_DetectorType) -> None: ... - def getType(self) -> FastFeatureDetector_DetectorType: ... - def getDefaultName(self) -> str: ... - -class AgastFeatureDetector(Feature2D): - # Functions - @classmethod - def create( - cls, - threshold: int = ..., - nonmaxSuppression: bool = ..., - type: AgastFeatureDetector_DetectorType = ..., - ) -> AgastFeatureDetector: ... - def setThreshold(self, threshold: int) -> None: ... - def getThreshold(self) -> int: ... - def setNonmaxSuppression(self, f: bool) -> None: ... - def getNonmaxSuppression(self) -> bool: ... - def setType(self, type: AgastFeatureDetector_DetectorType) -> None: ... - def getType(self) -> AgastFeatureDetector_DetectorType: ... - def getDefaultName(self) -> str: ... - -class GFTTDetector(Feature2D): - # Functions - @classmethod - @typing.overload - def create( - cls, - maxCorners: int = ..., - qualityLevel: float = ..., - minDistance: float = ..., - blockSize: int = ..., - useHarrisDetector: bool = ..., - k: float = ..., - ) -> GFTTDetector: ... - @classmethod - @typing.overload - def create( - cls, - maxCorners: int, - qualityLevel: float, - minDistance: float, - blockSize: int, - gradiantSize: int, - useHarrisDetector: bool = ..., - k: float = ..., - ) -> GFTTDetector: ... - def setMaxFeatures(self, maxFeatures: int) -> None: ... - def getMaxFeatures(self) -> int: ... - def setQualityLevel(self, qlevel: float) -> None: ... - def getQualityLevel(self) -> float: ... - def setMinDistance(self, minDistance: float) -> None: ... - def getMinDistance(self) -> float: ... - def setBlockSize(self, blockSize: int) -> None: ... - def getBlockSize(self) -> int: ... - def setGradientSize(self, gradientSize_: int) -> None: ... - def getGradientSize(self) -> int: ... - def setHarrisDetector(self, val: bool) -> None: ... - def getHarrisDetector(self) -> bool: ... - def setK(self, k: float) -> None: ... - def getK(self) -> float: ... - def getDefaultName(self) -> str: ... - -class SimpleBlobDetector(Feature2D): - # Classes - class Params: - thresholdStep: float - minThreshold: float - maxThreshold: float - minRepeatability: int - minDistBetweenBlobs: float - filterByColor: bool - blobColor: int - filterByArea: bool - minArea: float - maxArea: float - filterByCircularity: bool - minCircularity: float - maxCircularity: float - filterByInertia: bool - minInertiaRatio: float - maxInertiaRatio: float - filterByConvexity: bool - minConvexity: float - maxConvexity: float - collectContours: bool - - # Functions - def __init__(self) -> None: ... - - # Functions - - @classmethod - def create(cls, parameters: SimpleBlobDetector.Params = ...) -> SimpleBlobDetector: ... - def setParams(self, params: SimpleBlobDetector.Params) -> None: ... - def getParams(self) -> SimpleBlobDetector.Params: ... - def getDefaultName(self) -> str: ... - def getBlobContours(self) -> typing.Sequence[typing.Sequence[cv2.typing.Point]]: ... - -class KAZE(Feature2D): - # Functions - @classmethod - def create( - cls, - extended: bool = ..., - upright: bool = ..., - threshold: float = ..., - nOctaves: int = ..., - nOctaveLayers: int = ..., - diffusivity: KAZE_DiffusivityType = ..., - ) -> KAZE: ... - def setExtended(self, extended: bool) -> None: ... - def getExtended(self) -> bool: ... - def setUpright(self, upright: bool) -> None: ... - def getUpright(self) -> bool: ... - def setThreshold(self, threshold: float) -> None: ... - def getThreshold(self) -> float: ... - def setNOctaves(self, octaves: int) -> None: ... - def getNOctaves(self) -> int: ... - def setNOctaveLayers(self, octaveLayers: int) -> None: ... - def getNOctaveLayers(self) -> int: ... - def setDiffusivity(self, diff: KAZE_DiffusivityType) -> None: ... - def getDiffusivity(self) -> KAZE_DiffusivityType: ... - def getDefaultName(self) -> str: ... - -class AKAZE(Feature2D): - # Functions - @classmethod - def create( - cls, - descriptor_type: AKAZE_DescriptorType = ..., - descriptor_size: int = ..., - descriptor_channels: int = ..., - threshold: float = ..., - nOctaves: int = ..., - nOctaveLayers: int = ..., - diffusivity: KAZE_DiffusivityType = ..., - ) -> AKAZE: ... - def setDescriptorType(self, dtype: AKAZE_DescriptorType) -> None: ... - def getDescriptorType(self) -> AKAZE_DescriptorType: ... - def setDescriptorSize(self, dsize: int) -> None: ... - def getDescriptorSize(self) -> int: ... - def setDescriptorChannels(self, dch: int) -> None: ... - def getDescriptorChannels(self) -> int: ... - def setThreshold(self, threshold: float) -> None: ... - def getThreshold(self) -> float: ... - def setNOctaves(self, octaves: int) -> None: ... - def getNOctaves(self) -> int: ... - def setNOctaveLayers(self, octaveLayers: int) -> None: ... - def getNOctaveLayers(self) -> int: ... - def setDiffusivity(self, diff: KAZE_DiffusivityType) -> None: ... - def getDiffusivity(self) -> KAZE_DiffusivityType: ... - def getDefaultName(self) -> str: ... - -class DescriptorMatcher(Algorithm): - # Functions - @typing.overload - def add(self, descriptors: typing.Sequence[cv2.typing.MatLike]) -> None: ... - @typing.overload - def add(self, descriptors: typing.Sequence[UMat]) -> None: ... - def getTrainDescriptors(self) -> typing.Sequence[cv2.typing.MatLike]: ... - def clear(self) -> None: ... - def empty(self) -> bool: ... - def isMaskSupported(self) -> bool: ... - def train(self) -> None: ... - @typing.overload - def match( - self, - queryDescriptors: cv2.typing.MatLike, - trainDescriptors: cv2.typing.MatLike, - mask: cv2.typing.MatLike | None = ..., - ) -> typing.Sequence[DMatch]: ... - @typing.overload - def match( - self, - queryDescriptors: UMat, - trainDescriptors: UMat, - mask: UMat | None = ..., - ) -> typing.Sequence[DMatch]: ... - @typing.overload - def match( - self, - queryDescriptors: cv2.typing.MatLike, - masks: typing.Sequence[cv2.typing.MatLike] | None = ..., - ) -> typing.Sequence[DMatch]: ... - @typing.overload - def match(self, queryDescriptors: UMat, masks: typing.Sequence[UMat] | None = ...) -> typing.Sequence[DMatch]: ... - @typing.overload - def knnMatch( - self, - queryDescriptors: cv2.typing.MatLike, - trainDescriptors: cv2.typing.MatLike, - k: int, - mask: cv2.typing.MatLike | None = ..., - compactResult: bool = ..., - ) -> typing.Sequence[typing.Sequence[DMatch]]: ... - @typing.overload - def knnMatch( - self, - queryDescriptors: UMat, - trainDescriptors: UMat, - k: int, - mask: UMat | None = ..., - compactResult: bool = ..., - ) -> typing.Sequence[typing.Sequence[DMatch]]: ... - @typing.overload - def knnMatch( - self, - queryDescriptors: cv2.typing.MatLike, - k: int, - masks: typing.Sequence[cv2.typing.MatLike] | None = ..., - compactResult: bool = ..., - ) -> typing.Sequence[typing.Sequence[DMatch]]: ... - @typing.overload - def knnMatch( - self, - queryDescriptors: UMat, - k: int, - masks: typing.Sequence[UMat] | None = ..., - compactResult: bool = ..., - ) -> typing.Sequence[typing.Sequence[DMatch]]: ... - @typing.overload - def radiusMatch( - self, - queryDescriptors: cv2.typing.MatLike, - trainDescriptors: cv2.typing.MatLike, - maxDistance: float, - mask: cv2.typing.MatLike | None = ..., - compactResult: bool = ..., - ) -> typing.Sequence[typing.Sequence[DMatch]]: ... - @typing.overload - def radiusMatch( - self, - queryDescriptors: UMat, - trainDescriptors: UMat, - maxDistance: float, - mask: UMat | None = ..., - compactResult: bool = ..., - ) -> typing.Sequence[typing.Sequence[DMatch]]: ... - @typing.overload - def radiusMatch( - self, - queryDescriptors: cv2.typing.MatLike, - maxDistance: float, - masks: typing.Sequence[cv2.typing.MatLike] | None = ..., - compactResult: bool = ..., - ) -> typing.Sequence[typing.Sequence[DMatch]]: ... - @typing.overload - def radiusMatch( - self, - queryDescriptors: UMat, - maxDistance: float, - masks: typing.Sequence[UMat] | None = ..., - compactResult: bool = ..., - ) -> typing.Sequence[typing.Sequence[DMatch]]: ... - @typing.overload - def write(self, fileName: str) -> None: ... - @typing.overload - def write(self, fs: FileStorage, name: str) -> None: ... - @typing.overload - def read(self, fileName: str) -> None: ... - @typing.overload - def read(self, arg1: FileNode) -> None: ... - def clone(self, emptyTrainData: bool = ...) -> DescriptorMatcher: ... - @classmethod - @typing.overload - def create(cls, descriptorMatcherType: str) -> DescriptorMatcher: ... - @classmethod - @typing.overload - def create(cls, matcherType: DescriptorMatcher_MatcherType) -> DescriptorMatcher: ... - -class BOWKMeansTrainer(BOWTrainer): - # Functions - def __init__( - self, - clusterCount: int, - termcrit: cv2.typing.TermCriteria = ..., - attempts: int = ..., - flags: int = ..., - ) -> None: ... - @typing.overload - def cluster(self) -> cv2.typing.MatLike: ... - @typing.overload - def cluster(self, descriptors: cv2.typing.MatLike) -> cv2.typing.MatLike: ... - -class StereoMatcher(Algorithm): - # Functions - @typing.overload - def compute( - self, - left: cv2.typing.MatLike, - right: cv2.typing.MatLike, - disparity: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def compute(self, left: UMat, right: UMat, disparity: UMat | None = ...) -> UMat: ... - def getMinDisparity(self) -> int: ... - def setMinDisparity(self, minDisparity: int) -> None: ... - def getNumDisparities(self) -> int: ... - def setNumDisparities(self, numDisparities: int) -> None: ... - def getBlockSize(self) -> int: ... - def setBlockSize(self, blockSize: int) -> None: ... - def getSpeckleWindowSize(self) -> int: ... - def setSpeckleWindowSize(self, speckleWindowSize: int) -> None: ... - def getSpeckleRange(self) -> int: ... - def setSpeckleRange(self, speckleRange: int) -> None: ... - def getDisp12MaxDiff(self) -> int: ... - def setDisp12MaxDiff(self, disp12MaxDiff: int) -> None: ... - -class BaseCascadeClassifier(Algorithm): ... - -class QRCodeDetector(GraphicalCodeDetector): - # Functions - def __init__(self) -> None: ... - def setEpsX(self, epsX: float) -> QRCodeDetector: ... - def setEpsY(self, epsY: float) -> QRCodeDetector: ... - def setUseAlignmentMarkers(self, useAlignmentMarkers: bool) -> QRCodeDetector: ... - @typing.overload - def decodeCurved( - self, - img: cv2.typing.MatLike, - points: cv2.typing.MatLike, - straight_qrcode: cv2.typing.MatLike | None = ..., - ) -> tuple[ - str, - cv2.typing.MatLike, - ]: ... - @typing.overload - def decodeCurved(self, img: UMat, points: UMat, straight_qrcode: UMat | None = ...) -> tuple[str, UMat]: ... - @typing.overload - def detectAndDecodeCurved( - self, - img: cv2.typing.MatLike, - points: cv2.typing.MatLike | None = ..., - straight_qrcode: cv2.typing.MatLike | None = ..., - ) -> tuple[ - str, - cv2.typing.MatLike, - cv2.typing.MatLike, - ]: ... - @typing.overload - def detectAndDecodeCurved( - self, - img: UMat, - points: UMat | None = ..., - straight_qrcode: UMat | None = ..., - ) -> tuple[ - str, - UMat, - UMat, - ]: ... - -class QRCodeDetectorAruco(GraphicalCodeDetector): - # Classes - class Params: - minModuleSizeInPyramid: float - maxRotation: float - maxModuleSizeMismatch: float - maxTimingPatternMismatch: float - maxPenalties: float - maxColorsMismatch: float - scaleTimingPatternScore: float - - # Functions - def __init__(self) -> None: ... - - # Functions - - @typing.overload - def __init__(self) -> None: ... - @typing.overload - def __init__(self, params: QRCodeDetectorAruco.Params) -> None: ... - def getDetectorParameters(self) -> QRCodeDetectorAruco.Params: ... - def setDetectorParameters(self, params: QRCodeDetectorAruco.Params) -> QRCodeDetectorAruco: ... - def getArucoParameters(self) -> cv2.aruco.DetectorParameters: ... - def setArucoParameters(self, params: cv2.aruco.DetectorParameters) -> None: ... - -class BackgroundSubtractor(Algorithm): - # Functions - @typing.overload - def apply( - self, - image: cv2.typing.MatLike, - fgmask: cv2.typing.MatLike | None = ..., - learningRate: float = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def apply(self, image: UMat, fgmask: UMat | None = ..., learningRate: float = ...) -> UMat: ... - @typing.overload - def getBackgroundImage(self, backgroundImage: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... - @typing.overload - def getBackgroundImage(self, backgroundImage: UMat | None = ...) -> UMat: ... - -class DenseOpticalFlow(Algorithm): - # Functions - @typing.overload - def calc(self, I0: cv2.typing.MatLike, I1: cv2.typing.MatLike, flow: cv2.typing.MatLike) -> cv2.typing.MatLike: ... - @typing.overload - def calc(self, I0: UMat, I1: UMat, flow: UMat) -> UMat: ... - def collectGarbage(self) -> None: ... - -class SparseOpticalFlow(Algorithm): - # Functions - @typing.overload - def calc( - self, - prevImg: cv2.typing.MatLike, - nextImg: cv2.typing.MatLike, - prevPts: cv2.typing.MatLike, - nextPts: cv2.typing.MatLike, - status: cv2.typing.MatLike | None = ..., - err: cv2.typing.MatLike | None = ..., - ) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - ]: ... - @typing.overload - def calc( - self, - prevImg: UMat, - nextImg: UMat, - prevPts: UMat, - nextPts: UMat, - status: UMat | None = ..., - err: UMat | None = ..., - ) -> tuple[ - UMat, - UMat, - UMat, - ]: ... - -class TrackerMIL(Tracker): - # Classes - class Params: - samplerInitInRadius: float - samplerInitMaxNegNum: int - samplerSearchWinSize: float - samplerTrackInRadius: float - samplerTrackMaxPosNum: int - samplerTrackMaxNegNum: int - featureSetNumFeatures: int - - # Functions - def __init__(self) -> None: ... - - # Functions - - @classmethod - def create(cls, parameters: TrackerMIL.Params = ...) -> TrackerMIL: ... - -class TrackerGOTURN(Tracker): - # Classes - class Params: - modelTxt: str - modelBin: str - - # Functions - def __init__(self) -> None: ... - - # Functions - - @classmethod - def create(cls, parameters: TrackerGOTURN.Params = ...) -> TrackerGOTURN: ... - -class TrackerDaSiamRPN(Tracker): - # Classes - class Params: - model: str - kernel_cls1: str - kernel_r1: str - backend: int - target: int - - # Functions - def __init__(self) -> None: ... - - # Functions - - @classmethod - def create(cls, parameters: TrackerDaSiamRPN.Params = ...) -> TrackerDaSiamRPN: ... - def getTrackingScore(self) -> float: ... - -class TrackerNano(Tracker): - # Classes - class Params: - backbone: str - neckhead: str - backend: int - target: int - - # Functions - def __init__(self) -> None: ... - - # Functions - - @classmethod - def create(cls, parameters: TrackerNano.Params = ...) -> TrackerNano: ... - def getTrackingScore(self) -> float: ... - -class error(Exception): - code: int - err: str - file: str - func: str - line: int - msg: str - -class GeneralizedHoughBallard(GeneralizedHough): - # Functions - def setLevels(self, levels: int) -> None: ... - def getLevels(self) -> int: ... - def setVotesThreshold(self, votesThreshold: int) -> None: ... - def getVotesThreshold(self) -> int: ... - -class GeneralizedHoughGuil(GeneralizedHough): - # Functions - def setXi(self, xi: float) -> None: ... - def getXi(self) -> float: ... - def setLevels(self, levels: int) -> None: ... - def getLevels(self) -> int: ... - def setAngleEpsilon(self, angleEpsilon: float) -> None: ... - def getAngleEpsilon(self) -> float: ... - def setMinAngle(self, minAngle: float) -> None: ... - def getMinAngle(self) -> float: ... - def setMaxAngle(self, maxAngle: float) -> None: ... - def getMaxAngle(self) -> float: ... - def setAngleStep(self, angleStep: float) -> None: ... - def getAngleStep(self) -> float: ... - def setAngleThresh(self, angleThresh: int) -> None: ... - def getAngleThresh(self) -> int: ... - def setMinScale(self, minScale: float) -> None: ... - def getMinScale(self) -> float: ... - def setMaxScale(self, maxScale: float) -> None: ... - def getMaxScale(self) -> float: ... - def setScaleStep(self, scaleStep: float) -> None: ... - def getScaleStep(self) -> float: ... - def setScaleThresh(self, scaleThresh: int) -> None: ... - def getScaleThresh(self) -> int: ... - def setPosThresh(self, posThresh: int) -> None: ... - def getPosThresh(self) -> int: ... - -class TonemapDrago(Tonemap): - # Functions - def getSaturation(self) -> float: ... - def setSaturation(self, saturation: float) -> None: ... - def getBias(self) -> float: ... - def setBias(self, bias: float) -> None: ... - -class TonemapReinhard(Tonemap): - # Functions - def getIntensity(self) -> float: ... - def setIntensity(self, intensity: float) -> None: ... - def getLightAdaptation(self) -> float: ... - def setLightAdaptation(self, light_adapt: float) -> None: ... - def getColorAdaptation(self) -> float: ... - def setColorAdaptation(self, color_adapt: float) -> None: ... - -class TonemapMantiuk(Tonemap): - # Functions - def getScale(self) -> float: ... - def setScale(self, scale: float) -> None: ... - def getSaturation(self) -> float: ... - def setSaturation(self, saturation: float) -> None: ... - -class AlignMTB(AlignExposures): - # Functions - @typing.overload - def process( - self, - src: typing.Sequence[cv2.typing.MatLike], - dst: typing.Sequence[cv2.typing.MatLike], - times: cv2.typing.MatLike, - response: cv2.typing.MatLike, - ) -> None: ... - @typing.overload - def process( - self, - src: typing.Sequence[UMat], - dst: typing.Sequence[cv2.typing.MatLike], - times: UMat, - response: UMat, - ) -> None: ... - @typing.overload - def process(self, src: typing.Sequence[cv2.typing.MatLike], dst: typing.Sequence[cv2.typing.MatLike]) -> None: ... - @typing.overload - def process(self, src: typing.Sequence[UMat], dst: typing.Sequence[cv2.typing.MatLike]) -> None: ... - @typing.overload - def calculateShift(self, img0: cv2.typing.MatLike, img1: cv2.typing.MatLike) -> cv2.typing.Point: ... - @typing.overload - def calculateShift(self, img0: UMat, img1: UMat) -> cv2.typing.Point: ... - @typing.overload - def shiftMat( - self, - src: cv2.typing.MatLike, - shift: cv2.typing.Point, - dst: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def shiftMat(self, src: UMat, shift: cv2.typing.Point, dst: UMat | None = ...) -> UMat: ... - @typing.overload - def computeBitmaps( - self, - img: cv2.typing.MatLike, - tb: cv2.typing.MatLike | None = ..., - eb: cv2.typing.MatLike | None = ..., - ) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - ]: ... - @typing.overload - def computeBitmaps(self, img: UMat, tb: UMat | None = ..., eb: UMat | None = ...) -> tuple[UMat, UMat]: ... - def getMaxBits(self) -> int: ... - def setMaxBits(self, max_bits: int) -> None: ... - def getExcludeRange(self) -> int: ... - def setExcludeRange(self, exclude_range: int) -> None: ... - def getCut(self) -> bool: ... - def setCut(self, value: bool) -> None: ... - -class CalibrateDebevec(CalibrateCRF): - # Functions - def getLambda(self) -> float: ... - def setLambda(self, lambda_: float) -> None: ... - def getSamples(self) -> int: ... - def setSamples(self, samples: int) -> None: ... - def getRandom(self) -> bool: ... - def setRandom(self, random: bool) -> None: ... - -class CalibrateRobertson(CalibrateCRF): - # Functions - def getMaxIter(self) -> int: ... - def setMaxIter(self, max_iter: int) -> None: ... - def getThreshold(self) -> float: ... - def setThreshold(self, threshold: float) -> None: ... - def getRadiance(self) -> cv2.typing.MatLike: ... - -class MergeDebevec(MergeExposures): - # Functions - @typing.overload - def process( - self, - src: typing.Sequence[cv2.typing.MatLike], - times: cv2.typing.MatLike, - response: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def process(self, src: typing.Sequence[UMat], times: UMat, response: UMat, dst: UMat | None = ...) -> UMat: ... - @typing.overload - def process( - self, - src: typing.Sequence[cv2.typing.MatLike], - times: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def process(self, src: typing.Sequence[UMat], times: UMat, dst: UMat | None = ...) -> UMat: ... - -class MergeMertens(MergeExposures): - # Functions - @typing.overload - def process( - self, - src: typing.Sequence[cv2.typing.MatLike], - times: cv2.typing.MatLike, - response: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def process(self, src: typing.Sequence[UMat], times: UMat, response: UMat, dst: UMat | None = ...) -> UMat: ... - @typing.overload - def process( - self, - src: typing.Sequence[cv2.typing.MatLike], - dst: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def process(self, src: typing.Sequence[UMat], dst: UMat | None = ...) -> UMat: ... - def getContrastWeight(self) -> float: ... - def setContrastWeight(self, contrast_weiht: float) -> None: ... - def getSaturationWeight(self) -> float: ... - def setSaturationWeight(self, saturation_weight: float) -> None: ... - def getExposureWeight(self) -> float: ... - def setExposureWeight(self, exposure_weight: float) -> None: ... - -class MergeRobertson(MergeExposures): - # Functions - @typing.overload - def process( - self, - src: typing.Sequence[cv2.typing.MatLike], - times: cv2.typing.MatLike, - response: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def process(self, src: typing.Sequence[UMat], times: UMat, response: UMat, dst: UMat | None = ...) -> UMat: ... - @typing.overload - def process( - self, - src: typing.Sequence[cv2.typing.MatLike], - times: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def process(self, src: typing.Sequence[UMat], times: UMat, dst: UMat | None = ...) -> UMat: ... - -class BFMatcher(DescriptorMatcher): - # Functions - def __init__(self, normType: int = ..., crossCheck: bool = ...) -> None: ... - @classmethod - def create(cls, normType: int = ..., crossCheck: bool = ...) -> BFMatcher: ... - -class FlannBasedMatcher(DescriptorMatcher): - # Functions - def __init__( - self, - indexParams: cv2.typing.IndexParams = ..., - searchParams: cv2.typing.SearchParams = ..., - ) -> None: ... - @classmethod - def create(cls) -> FlannBasedMatcher: ... - -class StereoBM(StereoMatcher): - # Functions - def getPreFilterType(self) -> int: ... - def setPreFilterType(self, preFilterType: int) -> None: ... - def getPreFilterSize(self) -> int: ... - def setPreFilterSize(self, preFilterSize: int) -> None: ... - def getPreFilterCap(self) -> int: ... - def setPreFilterCap(self, preFilterCap: int) -> None: ... - def getTextureThreshold(self) -> int: ... - def setTextureThreshold(self, textureThreshold: int) -> None: ... - def getUniquenessRatio(self) -> int: ... - def setUniquenessRatio(self, uniquenessRatio: int) -> None: ... - def getSmallerBlockSize(self) -> int: ... - def setSmallerBlockSize(self, blockSize: int) -> None: ... - def getROI1(self) -> cv2.typing.Rect: ... - def setROI1(self, roi1: cv2.typing.Rect) -> None: ... - def getROI2(self) -> cv2.typing.Rect: ... - def setROI2(self, roi2: cv2.typing.Rect) -> None: ... - @classmethod - def create(cls, numDisparities: int = ..., blockSize: int = ...) -> StereoBM: ... - -class StereoSGBM(StereoMatcher): - # Functions - def getPreFilterCap(self) -> int: ... - def setPreFilterCap(self, preFilterCap: int) -> None: ... - def getUniquenessRatio(self) -> int: ... - def setUniquenessRatio(self, uniquenessRatio: int) -> None: ... - def getP1(self) -> int: ... - def setP1(self, P1: int) -> None: ... - def getP2(self) -> int: ... - def setP2(self, P2: int) -> None: ... - def getMode(self) -> int: ... - def setMode(self, mode: int) -> None: ... - @classmethod - def create( - cls, - minDisparity: int = ..., - numDisparities: int = ..., - blockSize: int = ..., - P1: int = ..., - P2: int = ..., - disp12MaxDiff: int = ..., - preFilterCap: int = ..., - uniquenessRatio: int = ..., - speckleWindowSize: int = ..., - speckleRange: int = ..., - mode: int = ..., - ) -> StereoSGBM: ... - -class BackgroundSubtractorMOG2(BackgroundSubtractor): - # Functions - def getHistory(self) -> int: ... - def setHistory(self, history: int) -> None: ... - def getNMixtures(self) -> int: ... - def setNMixtures(self, nmixtures: int) -> None: ... - def getBackgroundRatio(self) -> float: ... - def setBackgroundRatio(self, ratio: float) -> None: ... - def getVarThreshold(self) -> float: ... - def setVarThreshold(self, varThreshold: float) -> None: ... - def getVarThresholdGen(self) -> float: ... - def setVarThresholdGen(self, varThresholdGen: float) -> None: ... - def getVarInit(self) -> float: ... - def setVarInit(self, varInit: float) -> None: ... - def getVarMin(self) -> float: ... - def setVarMin(self, varMin: float) -> None: ... - def getVarMax(self) -> float: ... - def setVarMax(self, varMax: float) -> None: ... - def getComplexityReductionThreshold(self) -> float: ... - def setComplexityReductionThreshold(self, ct: float) -> None: ... - def getDetectShadows(self) -> bool: ... - def setDetectShadows(self, detectShadows: bool) -> None: ... - def getShadowValue(self) -> int: ... - def setShadowValue(self, value: int) -> None: ... - def getShadowThreshold(self) -> float: ... - def setShadowThreshold(self, threshold: float) -> None: ... - @typing.overload - def apply( - self, - image: cv2.typing.MatLike, - fgmask: cv2.typing.MatLike | None = ..., - learningRate: float = ..., - ) -> cv2.typing.MatLike: ... - @typing.overload - def apply(self, image: UMat, fgmask: UMat | None = ..., learningRate: float = ...) -> UMat: ... - -class BackgroundSubtractorKNN(BackgroundSubtractor): - # Functions - def getHistory(self) -> int: ... - def setHistory(self, history: int) -> None: ... - def getNSamples(self) -> int: ... - def setNSamples(self, _nN: int) -> None: ... - def getDist2Threshold(self) -> float: ... - def setDist2Threshold(self, _dist2Threshold: float) -> None: ... - def getkNNSamples(self) -> int: ... - def setkNNSamples(self, _nkNN: int) -> None: ... - def getDetectShadows(self) -> bool: ... - def setDetectShadows(self, detectShadows: bool) -> None: ... - def getShadowValue(self) -> int: ... - def setShadowValue(self, value: int) -> None: ... - def getShadowThreshold(self) -> float: ... - def setShadowThreshold(self, threshold: float) -> None: ... - -class FarnebackOpticalFlow(DenseOpticalFlow): - # Functions - def getNumLevels(self) -> int: ... - def setNumLevels(self, numLevels: int) -> None: ... - def getPyrScale(self) -> float: ... - def setPyrScale(self, pyrScale: float) -> None: ... - def getFastPyramids(self) -> bool: ... - def setFastPyramids(self, fastPyramids: bool) -> None: ... - def getWinSize(self) -> int: ... - def setWinSize(self, winSize: int) -> None: ... - def getNumIters(self) -> int: ... - def setNumIters(self, numIters: int) -> None: ... - def getPolyN(self) -> int: ... - def setPolyN(self, polyN: int) -> None: ... - def getPolySigma(self) -> float: ... - def setPolySigma(self, polySigma: float) -> None: ... - def getFlags(self) -> int: ... - def setFlags(self, flags: int) -> None: ... - @classmethod - def create( - cls, - numLevels: int = ..., - pyrScale: float = ..., - fastPyramids: bool = ..., - winSize: int = ..., - numIters: int = ..., - polyN: int = ..., - polySigma: float = ..., - flags: int = ..., - ) -> FarnebackOpticalFlow: ... - -class VariationalRefinement(DenseOpticalFlow): - # Functions - @typing.overload - def calcUV( - self, - I0: cv2.typing.MatLike, - I1: cv2.typing.MatLike, - flow_u: cv2.typing.MatLike, - flow_v: cv2.typing.MatLike, - ) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - ]: ... - @typing.overload - def calcUV(self, I0: UMat, I1: UMat, flow_u: UMat, flow_v: UMat) -> tuple[UMat, UMat]: ... - def getFixedPointIterations(self) -> int: ... - def setFixedPointIterations(self, val: int) -> None: ... - def getSorIterations(self) -> int: ... - def setSorIterations(self, val: int) -> None: ... - def getOmega(self) -> float: ... - def setOmega(self, val: float) -> None: ... - def getAlpha(self) -> float: ... - def setAlpha(self, val: float) -> None: ... - def getDelta(self) -> float: ... - def setDelta(self, val: float) -> None: ... - def getGamma(self) -> float: ... - def setGamma(self, val: float) -> None: ... - @classmethod - def create(cls) -> VariationalRefinement: ... - -class DISOpticalFlow(DenseOpticalFlow): - # Functions - def getFinestScale(self) -> int: ... - def setFinestScale(self, val: int) -> None: ... - def getPatchSize(self) -> int: ... - def setPatchSize(self, val: int) -> None: ... - def getPatchStride(self) -> int: ... - def setPatchStride(self, val: int) -> None: ... - def getGradientDescentIterations(self) -> int: ... - def setGradientDescentIterations(self, val: int) -> None: ... - def getVariationalRefinementIterations(self) -> int: ... - def setVariationalRefinementIterations(self, val: int) -> None: ... - def getVariationalRefinementAlpha(self) -> float: ... - def setVariationalRefinementAlpha(self, val: float) -> None: ... - def getVariationalRefinementDelta(self) -> float: ... - def setVariationalRefinementDelta(self, val: float) -> None: ... - def getVariationalRefinementGamma(self) -> float: ... - def setVariationalRefinementGamma(self, val: float) -> None: ... - def getUseMeanNormalization(self) -> bool: ... - def setUseMeanNormalization(self, val: bool) -> None: ... - def getUseSpatialPropagation(self) -> bool: ... - def setUseSpatialPropagation(self, val: bool) -> None: ... - @classmethod - def create(cls, preset: int = ...) -> DISOpticalFlow: ... - -class SparsePyrLKOpticalFlow(SparseOpticalFlow): - # Functions - def getWinSize(self) -> cv2.typing.Size: ... - def setWinSize(self, winSize: cv2.typing.Size) -> None: ... - def getMaxLevel(self) -> int: ... - def setMaxLevel(self, maxLevel: int) -> None: ... - def getTermCriteria(self) -> cv2.typing.TermCriteria: ... - def setTermCriteria(self, crit: cv2.typing.TermCriteria) -> None: ... - def getFlags(self) -> int: ... - def setFlags(self, flags: int) -> None: ... - def getMinEigThreshold(self) -> float: ... - def setMinEigThreshold(self, minEigThreshold: float) -> None: ... - @classmethod - def create( - cls, - winSize: cv2.typing.Size = ..., - maxLevel: int = ..., - crit: cv2.typing.TermCriteria = ..., - flags: int = ..., - minEigThreshold: float = ..., - ) -> SparsePyrLKOpticalFlow: ... - -# Functions -@typing.overload -def CamShift( - probImage: cv2.typing.MatLike, - window: cv2.typing.Rect, - criteria: cv2.typing.TermCriteria, -) -> tuple[ - cv2.typing.RotatedRect, - cv2.typing.Rect, -]: ... -@typing.overload -def CamShift( - probImage: UMat, - window: cv2.typing.Rect, - criteria: cv2.typing.TermCriteria, -) -> tuple[ - cv2.typing.RotatedRect, - cv2.typing.Rect, -]: ... -@typing.overload -def Canny( - image: cv2.typing.MatLike, - threshold1: float, - threshold2: float, - edges: cv2.typing.MatLike | None = ..., - apertureSize: int = ..., - L2gradient: bool = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def Canny( - image: UMat, - threshold1: float, - threshold2: float, - edges: UMat | None = ..., - apertureSize: int = ..., - L2gradient: bool = ..., -) -> UMat: ... -@typing.overload -def Canny( - dx: cv2.typing.MatLike, - dy: cv2.typing.MatLike, - threshold1: float, - threshold2: float, - edges: cv2.typing.MatLike | None = ..., - L2gradient: bool = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def Canny( - dx: UMat, - dy: UMat, - threshold1: float, - threshold2: float, - edges: UMat | None = ..., - L2gradient: bool = ..., -) -> UMat: ... -@typing.overload -def EMD( - signature1: cv2.typing.MatLike, - signature2: cv2.typing.MatLike, - distType: int, - cost: cv2.typing.MatLike | None = ..., - lowerBound: float | None = ..., - flow: cv2.typing.MatLike | None = ..., -) -> tuple[ - float, - float, - cv2.typing.MatLike, -]: ... -@typing.overload -def EMD( - signature1: UMat, - signature2: UMat, - distType: int, - cost: UMat | None = ..., - lowerBound: float | None = ..., - flow: UMat | None = ..., -) -> tuple[ - float, - float, - UMat, -]: ... -@typing.overload -def GaussianBlur( - src: cv2.typing.MatLike, - ksize: cv2.typing.Size, - sigmaX: float, - dst: cv2.typing.MatLike | None = ..., - sigmaY: float = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def GaussianBlur( - src: UMat, - ksize: cv2.typing.Size, - sigmaX: float, - dst: UMat | None = ..., - sigmaY: float = ..., - borderType: int = ..., -) -> UMat: ... -@typing.overload -def HoughCircles( - image: cv2.typing.MatLike, - method: int, - dp: float, - minDist: float, - circles: cv2.typing.MatLike | None = ..., - param1: float = ..., - param2: float = ..., - minRadius: int = ..., - maxRadius: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def HoughCircles( - image: UMat, - method: int, - dp: float, - minDist: float, - circles: UMat | None = ..., - param1: float = ..., - param2: float = ..., - minRadius: int = ..., - maxRadius: int = ..., -) -> UMat: ... -@typing.overload -def HoughLines( - image: cv2.typing.MatLike, - rho: float, - theta: float, - threshold: int, - lines: cv2.typing.MatLike | None = ..., - srn: float = ..., - stn: float = ..., - min_theta: float = ..., - max_theta: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def HoughLines( - image: UMat, - rho: float, - theta: float, - threshold: int, - lines: UMat | None = ..., - srn: float = ..., - stn: float = ..., - min_theta: float = ..., - max_theta: float = ..., -) -> UMat: ... -@typing.overload -def HoughLinesP( - image: cv2.typing.MatLike, - rho: float, - theta: float, - threshold: int, - lines: cv2.typing.MatLike | None = ..., - minLineLength: float = ..., - maxLineGap: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def HoughLinesP( - image: UMat, - rho: float, - theta: float, - threshold: int, - lines: UMat | None = ..., - minLineLength: float = ..., - maxLineGap: float = ..., -) -> UMat: ... -@typing.overload -def HoughLinesPointSet( - point: cv2.typing.MatLike, - lines_max: int, - threshold: int, - min_rho: float, - max_rho: float, - rho_step: float, - min_theta: float, - max_theta: float, - theta_step: float, - lines: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def HoughLinesPointSet( - point: UMat, - lines_max: int, - threshold: int, - min_rho: float, - max_rho: float, - rho_step: float, - min_theta: float, - max_theta: float, - theta_step: float, - lines: UMat | None = ..., -) -> UMat: ... -@typing.overload -def HoughLinesWithAccumulator( - image: cv2.typing.MatLike, - rho: float, - theta: float, - threshold: int, - lines: cv2.typing.MatLike | None = ..., - srn: float = ..., - stn: float = ..., - min_theta: float = ..., - max_theta: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def HoughLinesWithAccumulator( - image: UMat, - rho: float, - theta: float, - threshold: int, - lines: UMat | None = ..., - srn: float = ..., - stn: float = ..., - min_theta: float = ..., - max_theta: float = ..., -) -> UMat: ... -@typing.overload -def HuMoments(m: cv2.typing.Moments, hu: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def HuMoments(m: cv2.typing.Moments, hu: UMat | None = ...) -> UMat: ... -@typing.overload -def LUT( - src: cv2.typing.MatLike, - lut: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def LUT(src: UMat, lut: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def Laplacian( - src: cv2.typing.MatLike, - ddepth: int, - dst: cv2.typing.MatLike | None = ..., - ksize: int = ..., - scale: float = ..., - delta: float = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def Laplacian( - src: UMat, - ddepth: int, - dst: UMat | None = ..., - ksize: int = ..., - scale: float = ..., - delta: float = ..., - borderType: int = ..., -) -> UMat: ... -@typing.overload -def Mahalanobis(v1: cv2.typing.MatLike, v2: cv2.typing.MatLike, icovar: cv2.typing.MatLike) -> float: ... -@typing.overload -def Mahalanobis(v1: UMat, v2: UMat, icovar: UMat) -> float: ... -@typing.overload -def PCABackProject( - data: cv2.typing.MatLike, - mean: cv2.typing.MatLike, - eigenvectors: cv2.typing.MatLike, - result: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def PCABackProject(data: UMat, mean: UMat, eigenvectors: UMat, result: UMat | None = ...) -> UMat: ... -@typing.overload -def PCACompute( - data: cv2.typing.MatLike, - mean: cv2.typing.MatLike, - eigenvectors: cv2.typing.MatLike | None = ..., - maxComponents: int = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def PCACompute( - data: UMat, - mean: UMat, - eigenvectors: UMat | None = ..., - maxComponents: int = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def PCACompute( - data: cv2.typing.MatLike, - mean: cv2.typing.MatLike, - retainedVariance: float, - eigenvectors: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def PCACompute( - data: UMat, - mean: UMat, - retainedVariance: float, - eigenvectors: UMat | None = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def PCACompute2( - data: cv2.typing.MatLike, - mean: cv2.typing.MatLike, - eigenvectors: cv2.typing.MatLike | None = ..., - eigenvalues: cv2.typing.MatLike | None = ..., - maxComponents: int = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def PCACompute2( - data: UMat, - mean: UMat, - eigenvectors: UMat | None = ..., - eigenvalues: UMat | None = ..., - maxComponents: int = ..., -) -> tuple[ - UMat, - UMat, - UMat, -]: ... -@typing.overload -def PCACompute2( - data: cv2.typing.MatLike, - mean: cv2.typing.MatLike, - retainedVariance: float, - eigenvectors: cv2.typing.MatLike | None = ..., - eigenvalues: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def PCACompute2( - data: UMat, - mean: UMat, - retainedVariance: float, - eigenvectors: UMat | None = ..., - eigenvalues: UMat | None = ..., -) -> tuple[ - UMat, - UMat, - UMat, -]: ... -@typing.overload -def PCAProject( - data: cv2.typing.MatLike, - mean: cv2.typing.MatLike, - eigenvectors: cv2.typing.MatLike, - result: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def PCAProject(data: UMat, mean: UMat, eigenvectors: UMat, result: UMat | None = ...) -> UMat: ... -@typing.overload -def PSNR(src1: cv2.typing.MatLike, src2: cv2.typing.MatLike, R: float = ...) -> float: ... -@typing.overload -def PSNR(src1: UMat, src2: UMat, R: float = ...) -> float: ... -@typing.overload -def RQDecomp3x3( - src: cv2.typing.MatLike, - mtxR: cv2.typing.MatLike | None = ..., - mtxQ: cv2.typing.MatLike | None = ..., - Qx: cv2.typing.MatLike | None = ..., - Qy: cv2.typing.MatLike | None = ..., - Qz: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.Vec3d, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def RQDecomp3x3( - src: UMat, - mtxR: UMat | None = ..., - mtxQ: UMat | None = ..., - Qx: UMat | None = ..., - Qy: UMat | None = ..., - Qz: UMat | None = ..., -) -> tuple[ - cv2.typing.Vec3d, - UMat, - UMat, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def Rodrigues( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - jacobian: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def Rodrigues(src: UMat, dst: UMat | None = ..., jacobian: UMat | None = ...) -> tuple[UMat, UMat]: ... -@typing.overload -def SVBackSubst( - w: cv2.typing.MatLike, - u: cv2.typing.MatLike, - vt: cv2.typing.MatLike, - rhs: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def SVBackSubst(w: UMat, u: UMat, vt: UMat, rhs: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def SVDecomp( - src: cv2.typing.MatLike, - w: cv2.typing.MatLike | None = ..., - u: cv2.typing.MatLike | None = ..., - vt: cv2.typing.MatLike | None = ..., - flags: int = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def SVDecomp( - src: UMat, - w: UMat | None = ..., - u: UMat | None = ..., - vt: UMat | None = ..., - flags: int = ..., -) -> tuple[ - UMat, - UMat, - UMat, -]: ... -@typing.overload -def Scharr( - src: cv2.typing.MatLike, - ddepth: int, - dx: int, - dy: int, - dst: cv2.typing.MatLike | None = ..., - scale: float = ..., - delta: float = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def Scharr( - src: UMat, - ddepth: int, - dx: int, - dy: int, - dst: UMat | None = ..., - scale: float = ..., - delta: float = ..., - borderType: int = ..., -) -> UMat: ... -@typing.overload -def Sobel( - src: cv2.typing.MatLike, - ddepth: int, - dx: int, - dy: int, - dst: cv2.typing.MatLike | None = ..., - ksize: int = ..., - scale: float = ..., - delta: float = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def Sobel( - src: UMat, - ddepth: int, - dx: int, - dy: int, - dst: UMat | None = ..., - ksize: int = ..., - scale: float = ..., - delta: float = ..., - borderType: int = ..., -) -> UMat: ... -@typing.overload -def absdiff( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def absdiff(src1: UMat, src2: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def accumulate( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike, - mask: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def accumulate(src: UMat, dst: UMat, mask: UMat | None = ...) -> UMat: ... -@typing.overload -def accumulateProduct( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike, - mask: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def accumulateProduct(src1: UMat, src2: UMat, dst: UMat, mask: UMat | None = ...) -> UMat: ... -@typing.overload -def accumulateSquare( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike, - mask: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def accumulateSquare(src: UMat, dst: UMat, mask: UMat | None = ...) -> UMat: ... -@typing.overload -def accumulateWeighted( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike, - alpha: float, - mask: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def accumulateWeighted(src: UMat, dst: UMat, alpha: float, mask: UMat | None = ...) -> UMat: ... -@typing.overload -def adaptiveThreshold( - src: cv2.typing.MatLike, - maxValue: float, - adaptiveMethod: int, - thresholdType: int, - blockSize: int, - C: float, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def adaptiveThreshold( - src: UMat, - maxValue: float, - adaptiveMethod: int, - thresholdType: int, - blockSize: int, - C: float, - dst: UMat | None = ..., -) -> UMat: ... -@typing.overload -def add( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - mask: cv2.typing.MatLike | None = ..., - dtype: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def add(src1: UMat, src2: UMat, dst: UMat | None = ..., mask: UMat | None = ..., dtype: int = ...) -> UMat: ... -def addText( - img: cv2.typing.MatLike, - text: str, - org: cv2.typing.Point, - nameFont: str, - pointSize: int = ..., - color: cv2.typing.Scalar = ..., - weight: int = ..., - style: int = ..., - spacing: int = ..., -) -> None: ... -@typing.overload -def addWeighted( - src1: cv2.typing.MatLike, - alpha: float, - src2: cv2.typing.MatLike, - beta: float, - gamma: float, - dst: cv2.typing.MatLike | None = ..., - dtype: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def addWeighted( - src1: UMat, - alpha: float, - src2: UMat, - beta: float, - gamma: float, - dst: UMat | None = ..., - dtype: int = ..., -) -> UMat: ... -@typing.overload -def applyColorMap( - src: cv2.typing.MatLike, - colormap: int, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def applyColorMap(src: UMat, colormap: int, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def applyColorMap( - src: cv2.typing.MatLike, - userColor: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def applyColorMap(src: UMat, userColor: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def approxPolyDP( - curve: cv2.typing.MatLike, - epsilon: float, - closed: bool, - approxCurve: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def approxPolyDP(curve: UMat, epsilon: float, closed: bool, approxCurve: UMat | None = ...) -> UMat: ... -@typing.overload -def arcLength(curve: cv2.typing.MatLike, closed: bool) -> float: ... -@typing.overload -def arcLength(curve: UMat, closed: bool) -> float: ... -@typing.overload -def arrowedLine( - img: cv2.typing.MatLike, - pt1: cv2.typing.Point, - pt2: cv2.typing.Point, - color: cv2.typing.Scalar, - thickness: int = ..., - line_type: int = ..., - shift: int = ..., - tipLength: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def arrowedLine( - img: UMat, - pt1: cv2.typing.Point, - pt2: cv2.typing.Point, - color: cv2.typing.Scalar, - thickness: int = ..., - line_type: int = ..., - shift: int = ..., - tipLength: float = ..., -) -> UMat: ... -@typing.overload -def batchDistance( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - dtype: int, - dist: cv2.typing.MatLike | None = ..., - nidx: cv2.typing.MatLike | None = ..., - normType: int = ..., - K: int = ..., - mask: cv2.typing.MatLike | None = ..., - update: int = ..., - crosscheck: bool = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def batchDistance( - src1: UMat, - src2: UMat, - dtype: int, - dist: UMat | None = ..., - nidx: UMat | None = ..., - normType: int = ..., - K: int = ..., - mask: UMat | None = ..., - update: int = ..., - crosscheck: bool = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def bilateralFilter( - src: cv2.typing.MatLike, - d: int, - sigmaColor: float, - sigmaSpace: float, - dst: cv2.typing.MatLike | None = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def bilateralFilter( - src: UMat, - d: int, - sigmaColor: float, - sigmaSpace: float, - dst: UMat | None = ..., - borderType: int = ..., -) -> UMat: ... -@typing.overload -def bitwise_and( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - mask: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def bitwise_and(src1: UMat, src2: UMat, dst: UMat | None = ..., mask: UMat | None = ...) -> UMat: ... -@typing.overload -def bitwise_not( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - mask: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def bitwise_not(src: UMat, dst: UMat | None = ..., mask: UMat | None = ...) -> UMat: ... -@typing.overload -def bitwise_or( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - mask: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def bitwise_or(src1: UMat, src2: UMat, dst: UMat | None = ..., mask: UMat | None = ...) -> UMat: ... -@typing.overload -def bitwise_xor( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - mask: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def bitwise_xor(src1: UMat, src2: UMat, dst: UMat | None = ..., mask: UMat | None = ...) -> UMat: ... -@typing.overload -def blendLinear( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - weights1: cv2.typing.MatLike, - weights2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def blendLinear(src1: UMat, src2: UMat, weights1: UMat, weights2: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def blur( - src: cv2.typing.MatLike, - ksize: cv2.typing.Size, - dst: cv2.typing.MatLike | None = ..., - anchor: cv2.typing.Point = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def blur( - src: UMat, - ksize: cv2.typing.Size, - dst: UMat | None = ..., - anchor: cv2.typing.Point = ..., - borderType: int = ..., -) -> UMat: ... -def borderInterpolate(p: int, len: int, borderType: int) -> int: ... -@typing.overload -def boundingRect(array: cv2.typing.MatLike) -> cv2.typing.Rect: ... -@typing.overload -def boundingRect(array: UMat) -> cv2.typing.Rect: ... -@typing.overload -def boxFilter( - src: cv2.typing.MatLike, - ddepth: int, - ksize: cv2.typing.Size, - dst: cv2.typing.MatLike | None = ..., - anchor: cv2.typing.Point = ..., - normalize: bool = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def boxFilter( - src: UMat, - ddepth: int, - ksize: cv2.typing.Size, - dst: UMat | None = ..., - anchor: cv2.typing.Point = ..., - normalize: bool = ..., - borderType: int = ..., -) -> UMat: ... -@typing.overload -def boxPoints(box: cv2.typing.RotatedRect, points: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def boxPoints(box: cv2.typing.RotatedRect, points: UMat | None = ...) -> UMat: ... -@typing.overload -def buildOpticalFlowPyramid( - img: cv2.typing.MatLike, - winSize: cv2.typing.Size, - maxLevel: int, - pyramid: typing.Sequence[cv2.typing.MatLike] | None = ..., - withDerivatives: bool = ..., - pyrBorder: int = ..., - derivBorder: int = ..., - tryReuseInputImage: bool = ..., -) -> tuple[ - int, - typing.Sequence[cv2.typing.MatLike], -]: ... -@typing.overload -def buildOpticalFlowPyramid( - img: UMat, - winSize: cv2.typing.Size, - maxLevel: int, - pyramid: typing.Sequence[UMat] | None = ..., - withDerivatives: bool = ..., - pyrBorder: int = ..., - derivBorder: int = ..., - tryReuseInputImage: bool = ..., -) -> tuple[ - int, - typing.Sequence[UMat], -]: ... -@typing.overload -def calcBackProject( - images: typing.Sequence[cv2.typing.MatLike], - channels: typing.Sequence[int], - hist: cv2.typing.MatLike, - ranges: typing.Sequence[float], - scale: float, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def calcBackProject( - images: typing.Sequence[UMat], - channels: typing.Sequence[int], - hist: UMat, - ranges: typing.Sequence[float], - scale: float, - dst: UMat | None = ..., -) -> UMat: ... -@typing.overload -def calcCovarMatrix( - samples: cv2.typing.MatLike, - mean: cv2.typing.MatLike, - flags: int, - covar: cv2.typing.MatLike | None = ..., - ctype: int = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def calcCovarMatrix( - samples: UMat, - mean: UMat, - flags: int, - covar: UMat | None = ..., - ctype: int = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def calcHist( - images: typing.Sequence[cv2.typing.MatLike], - channels: typing.Sequence[int], - mask: cv2.typing.MatLike | None, - histSize: typing.Sequence[int], - ranges: typing.Sequence[float], - hist: cv2.typing.MatLike | None = ..., - accumulate: bool = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def calcHist( - images: typing.Sequence[UMat], - channels: typing.Sequence[int], - mask: UMat | None, - histSize: typing.Sequence[int], - ranges: typing.Sequence[float], - hist: UMat | None = ..., - accumulate: bool = ..., -) -> UMat: ... -@typing.overload -def calcOpticalFlowFarneback( - prev: cv2.typing.MatLike, - next: cv2.typing.MatLike, - flow: cv2.typing.MatLike, - pyr_scale: float, - levels: int, - winsize: int, - iterations: int, - poly_n: int, - poly_sigma: float, - flags: int, -) -> cv2.typing.MatLike: ... -@typing.overload -def calcOpticalFlowFarneback( - prev: UMat, - next: UMat, - flow: UMat, - pyr_scale: float, - levels: int, - winsize: int, - iterations: int, - poly_n: int, - poly_sigma: float, - flags: int, -) -> UMat: ... -@typing.overload -def calcOpticalFlowPyrLK( - prevImg: cv2.typing.MatLike, - nextImg: cv2.typing.MatLike, - prevPts: cv2.typing.MatLike, - nextPts: cv2.typing.MatLike, - status: cv2.typing.MatLike | None = ..., - err: cv2.typing.MatLike | None = ..., - winSize: cv2.typing.Size = ..., - maxLevel: int = ..., - criteria: cv2.typing.TermCriteria = ..., - flags: int = ..., - minEigThreshold: float = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def calcOpticalFlowPyrLK( - prevImg: UMat, - nextImg: UMat, - prevPts: UMat, - nextPts: UMat, - status: UMat | None = ..., - err: UMat | None = ..., - winSize: cv2.typing.Size = ..., - maxLevel: int = ..., - criteria: cv2.typing.TermCriteria = ..., - flags: int = ..., - minEigThreshold: float = ..., -) -> tuple[ - UMat, - UMat, - UMat, -]: ... -@typing.overload -def calibrateCamera( - objectPoints: typing.Sequence[cv2.typing.MatLike], - imagePoints: typing.Sequence[cv2.typing.MatLike], - imageSize: cv2.typing.Size, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - rvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., - tvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - cv2.typing.MatLike, - cv2.typing.MatLike, - typing.Sequence[cv2.typing.MatLike], - typing.Sequence[cv2.typing.MatLike], -]: ... -@typing.overload -def calibrateCamera( - objectPoints: typing.Sequence[UMat], - imagePoints: typing.Sequence[UMat], - imageSize: cv2.typing.Size, - cameraMatrix: UMat, - distCoeffs: UMat, - rvecs: typing.Sequence[UMat] | None = ..., - tvecs: typing.Sequence[UMat] | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - UMat, - UMat, - typing.Sequence[UMat], - typing.Sequence[UMat], -]: ... -@typing.overload -def calibrateCameraExtended( - objectPoints: typing.Sequence[cv2.typing.MatLike], - imagePoints: typing.Sequence[cv2.typing.MatLike], - imageSize: cv2.typing.Size, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - rvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., - tvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., - stdDeviationsIntrinsics: cv2.typing.MatLike | None = ..., - stdDeviationsExtrinsics: cv2.typing.MatLike | None = ..., - perViewErrors: cv2.typing.MatLike | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - cv2.typing.MatLike, - cv2.typing.MatLike, - typing.Sequence[cv2.typing.MatLike], - typing.Sequence[cv2.typing.MatLike], - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def calibrateCameraExtended( - objectPoints: typing.Sequence[UMat], - imagePoints: typing.Sequence[UMat], - imageSize: cv2.typing.Size, - cameraMatrix: UMat, - distCoeffs: UMat, - rvecs: typing.Sequence[UMat] | None = ..., - tvecs: typing.Sequence[UMat] | None = ..., - stdDeviationsIntrinsics: UMat | None = ..., - stdDeviationsExtrinsics: UMat | None = ..., - perViewErrors: UMat | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - UMat, - UMat, - typing.Sequence[UMat], - typing.Sequence[UMat], - UMat, - UMat, - UMat, -]: ... -@typing.overload -def calibrateCameraRO( - objectPoints: typing.Sequence[cv2.typing.MatLike], - imagePoints: typing.Sequence[cv2.typing.MatLike], - imageSize: cv2.typing.Size, - iFixedPoint: int, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - rvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., - tvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., - newObjPoints: cv2.typing.MatLike | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - cv2.typing.MatLike, - cv2.typing.MatLike, - typing.Sequence[cv2.typing.MatLike], - typing.Sequence[cv2.typing.MatLike], - cv2.typing.MatLike, -]: ... -@typing.overload -def calibrateCameraRO( - objectPoints: typing.Sequence[UMat], - imagePoints: typing.Sequence[UMat], - imageSize: cv2.typing.Size, - iFixedPoint: int, - cameraMatrix: UMat, - distCoeffs: UMat, - rvecs: typing.Sequence[UMat] | None = ..., - tvecs: typing.Sequence[UMat] | None = ..., - newObjPoints: UMat | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - UMat, - UMat, - typing.Sequence[UMat], - typing.Sequence[UMat], - UMat, -]: ... -@typing.overload -def calibrateCameraROExtended( - objectPoints: typing.Sequence[cv2.typing.MatLike], - imagePoints: typing.Sequence[cv2.typing.MatLike], - imageSize: cv2.typing.Size, - iFixedPoint: int, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - rvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., - tvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., - newObjPoints: cv2.typing.MatLike | None = ..., - stdDeviationsIntrinsics: cv2.typing.MatLike | None = ..., - stdDeviationsExtrinsics: cv2.typing.MatLike | None = ..., - stdDeviationsObjPoints: cv2.typing.MatLike | None = ..., - perViewErrors: cv2.typing.MatLike | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - cv2.typing.MatLike, - cv2.typing.MatLike, - typing.Sequence[cv2.typing.MatLike], - typing.Sequence[cv2.typing.MatLike], - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def calibrateCameraROExtended( - objectPoints: typing.Sequence[UMat], - imagePoints: typing.Sequence[UMat], - imageSize: cv2.typing.Size, - iFixedPoint: int, - cameraMatrix: UMat, - distCoeffs: UMat, - rvecs: typing.Sequence[UMat] | None = ..., - tvecs: typing.Sequence[UMat] | None = ..., - newObjPoints: UMat | None = ..., - stdDeviationsIntrinsics: UMat | None = ..., - stdDeviationsExtrinsics: UMat | None = ..., - stdDeviationsObjPoints: UMat | None = ..., - perViewErrors: UMat | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - UMat, - UMat, - typing.Sequence[UMat], - typing.Sequence[UMat], - UMat, - UMat, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def calibrateHandEye( - R_gripper2base: typing.Sequence[cv2.typing.MatLike], - t_gripper2base: typing.Sequence[cv2.typing.MatLike], - R_target2cam: typing.Sequence[cv2.typing.MatLike], - t_target2cam: typing.Sequence[cv2.typing.MatLike], - R_cam2gripper: cv2.typing.MatLike | None = ..., - t_cam2gripper: cv2.typing.MatLike | None = ..., - method: HandEyeCalibrationMethod = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def calibrateHandEye( - R_gripper2base: typing.Sequence[UMat], - t_gripper2base: typing.Sequence[UMat], - R_target2cam: typing.Sequence[UMat], - t_target2cam: typing.Sequence[UMat], - R_cam2gripper: UMat | None = ..., - t_cam2gripper: UMat | None = ..., - method: HandEyeCalibrationMethod = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def calibrateRobotWorldHandEye( - R_world2cam: typing.Sequence[cv2.typing.MatLike], - t_world2cam: typing.Sequence[cv2.typing.MatLike], - R_base2gripper: typing.Sequence[cv2.typing.MatLike], - t_base2gripper: typing.Sequence[cv2.typing.MatLike], - R_base2world: cv2.typing.MatLike | None = ..., - t_base2world: cv2.typing.MatLike | None = ..., - R_gripper2cam: cv2.typing.MatLike | None = ..., - t_gripper2cam: cv2.typing.MatLike | None = ..., - method: RobotWorldHandEyeCalibrationMethod = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def calibrateRobotWorldHandEye( - R_world2cam: typing.Sequence[UMat], - t_world2cam: typing.Sequence[UMat], - R_base2gripper: typing.Sequence[UMat], - t_base2gripper: typing.Sequence[UMat], - R_base2world: UMat | None = ..., - t_base2world: UMat | None = ..., - R_gripper2cam: UMat | None = ..., - t_gripper2cam: UMat | None = ..., - method: RobotWorldHandEyeCalibrationMethod = ..., -) -> tuple[ - UMat, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def calibrationMatrixValues( - cameraMatrix: cv2.typing.MatLike, - imageSize: cv2.typing.Size, - apertureWidth: float, - apertureHeight: float, -) -> tuple[ - float, - float, - float, - cv2.typing.Point2d, - float, -]: ... -@typing.overload -def calibrationMatrixValues( - cameraMatrix: UMat, - imageSize: cv2.typing.Size, - apertureWidth: float, - apertureHeight: float, -) -> tuple[ - float, - float, - float, - cv2.typing.Point2d, - float, -]: ... -@typing.overload -def cartToPolar( - x: cv2.typing.MatLike, - y: cv2.typing.MatLike, - magnitude: cv2.typing.MatLike | None = ..., - angle: cv2.typing.MatLike | None = ..., - angleInDegrees: bool = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def cartToPolar( - x: UMat, - y: UMat, - magnitude: UMat | None = ..., - angle: UMat | None = ..., - angleInDegrees: bool = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def checkChessboard(img: cv2.typing.MatLike, size: cv2.typing.Size) -> bool: ... -@typing.overload -def checkChessboard(img: UMat, size: cv2.typing.Size) -> bool: ... -def checkHardwareSupport(feature: int) -> bool: ... -@typing.overload -def checkRange( - a: cv2.typing.MatLike, - quiet: bool = ..., - minVal: float = ..., - maxVal: float = ..., -) -> tuple[ - bool, - cv2.typing.Point, -]: ... -@typing.overload -def checkRange( - a: UMat, - quiet: bool = ..., - minVal: float = ..., - maxVal: float = ..., -) -> tuple[ - bool, - cv2.typing.Point, -]: ... -@typing.overload -def circle( - img: cv2.typing.MatLike, - center: cv2.typing.Point, - radius: int, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - shift: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def circle( - img: UMat, - center: cv2.typing.Point, - radius: int, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - shift: int = ..., -) -> UMat: ... -def clipLine( - imgRect: cv2.typing.Rect, - pt1: cv2.typing.Point, - pt2: cv2.typing.Point, -) -> tuple[ - bool, - cv2.typing.Point, - cv2.typing.Point, -]: ... -@typing.overload -def colorChange( - src: cv2.typing.MatLike, - mask: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - red_mul: float = ..., - green_mul: float = ..., - blue_mul: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def colorChange( - src: UMat, - mask: UMat, - dst: UMat | None = ..., - red_mul: float = ..., - green_mul: float = ..., - blue_mul: float = ..., -) -> UMat: ... -@typing.overload -def compare( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - cmpop: int, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def compare(src1: UMat, src2: UMat, cmpop: int, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def compareHist(H1: cv2.typing.MatLike, H2: cv2.typing.MatLike, method: int) -> float: ... -@typing.overload -def compareHist(H1: UMat, H2: UMat, method: int) -> float: ... -@typing.overload -def completeSymm(m: cv2.typing.MatLike, lowerToUpper: bool = ...) -> cv2.typing.MatLike: ... -@typing.overload -def completeSymm(m: UMat, lowerToUpper: bool = ...) -> UMat: ... -@typing.overload -def composeRT( - rvec1: cv2.typing.MatLike, - tvec1: cv2.typing.MatLike, - rvec2: cv2.typing.MatLike, - tvec2: cv2.typing.MatLike, - rvec3: cv2.typing.MatLike | None = ..., - tvec3: cv2.typing.MatLike | None = ..., - dr3dr1: cv2.typing.MatLike | None = ..., - dr3dt1: cv2.typing.MatLike | None = ..., - dr3dr2: cv2.typing.MatLike | None = ..., - dr3dt2: cv2.typing.MatLike | None = ..., - dt3dr1: cv2.typing.MatLike | None = ..., - dt3dt1: cv2.typing.MatLike | None = ..., - dt3dr2: cv2.typing.MatLike | None = ..., - dt3dt2: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def composeRT( - rvec1: UMat, - tvec1: UMat, - rvec2: UMat, - tvec2: UMat, - rvec3: UMat | None = ..., - tvec3: UMat | None = ..., - dr3dr1: UMat | None = ..., - dr3dt1: UMat | None = ..., - dr3dr2: UMat | None = ..., - dr3dt2: UMat | None = ..., - dt3dr1: UMat | None = ..., - dt3dt1: UMat | None = ..., - dt3dr2: UMat | None = ..., - dt3dt2: UMat | None = ..., -) -> tuple[ - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def computeCorrespondEpilines( - points: cv2.typing.MatLike, - whichImage: int, - F: cv2.typing.MatLike, - lines: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def computeCorrespondEpilines(points: UMat, whichImage: int, F: UMat, lines: UMat | None = ...) -> UMat: ... -@typing.overload -def computeECC( - templateImage: cv2.typing.MatLike, - inputImage: cv2.typing.MatLike, - inputMask: cv2.typing.MatLike | None = ..., -) -> float: ... -@typing.overload -def computeECC(templateImage: UMat, inputImage: UMat, inputMask: UMat | None = ...) -> float: ... -@typing.overload -def connectedComponents( - image: cv2.typing.MatLike, - labels: cv2.typing.MatLike | None = ..., - connectivity: int = ..., - ltype: int = ..., -) -> tuple[ - int, - cv2.typing.MatLike, -]: ... -@typing.overload -def connectedComponents( - image: UMat, - labels: UMat | None = ..., - connectivity: int = ..., - ltype: int = ..., -) -> tuple[ - int, - UMat, -]: ... -@typing.overload -def connectedComponentsWithAlgorithm( - image: cv2.typing.MatLike, - connectivity: int, - ltype: int, - ccltype: int, - labels: cv2.typing.MatLike | None = ..., -) -> tuple[ - int, - cv2.typing.MatLike, -]: ... -@typing.overload -def connectedComponentsWithAlgorithm( - image: UMat, - connectivity: int, - ltype: int, - ccltype: int, - labels: UMat | None = ..., -) -> tuple[ - int, - UMat, -]: ... -@typing.overload -def connectedComponentsWithStats( - image: cv2.typing.MatLike, - labels: cv2.typing.MatLike | None = ..., - stats: cv2.typing.MatLike | None = ..., - centroids: cv2.typing.MatLike | None = ..., - connectivity: int = ..., - ltype: int = ..., -) -> tuple[ - int, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def connectedComponentsWithStats( - image: UMat, - labels: UMat | None = ..., - stats: UMat | None = ..., - centroids: UMat | None = ..., - connectivity: int = ..., - ltype: int = ..., -) -> tuple[ - int, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def connectedComponentsWithStatsWithAlgorithm( - image: cv2.typing.MatLike, - connectivity: int, - ltype: int, - ccltype: int, - labels: cv2.typing.MatLike | None = ..., - stats: cv2.typing.MatLike | None = ..., - centroids: cv2.typing.MatLike | None = ..., -) -> tuple[ - int, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def connectedComponentsWithStatsWithAlgorithm( - image: UMat, - connectivity: int, - ltype: int, - ccltype: int, - labels: UMat | None = ..., - stats: UMat | None = ..., - centroids: UMat | None = ..., -) -> tuple[ - int, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def contourArea(contour: cv2.typing.MatLike, oriented: bool = ...) -> float: ... -@typing.overload -def contourArea(contour: UMat, oriented: bool = ...) -> float: ... -@typing.overload -def convertFp16(src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def convertFp16(src: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def convertMaps( - map1: cv2.typing.MatLike, - map2: cv2.typing.MatLike, - dstmap1type: int, - dstmap1: cv2.typing.MatLike | None = ..., - dstmap2: cv2.typing.MatLike | None = ..., - nninterpolation: bool = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def convertMaps( - map1: UMat, - map2: UMat, - dstmap1type: int, - dstmap1: UMat | None = ..., - dstmap2: UMat | None = ..., - nninterpolation: bool = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def convertPointsFromHomogeneous( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def convertPointsFromHomogeneous(src: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def convertPointsToHomogeneous(src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def convertPointsToHomogeneous(src: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def convertScaleAbs( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - alpha: float = ..., - beta: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def convertScaleAbs(src: UMat, dst: UMat | None = ..., alpha: float = ..., beta: float = ...) -> UMat: ... -@typing.overload -def convexHull( - points: cv2.typing.MatLike, - hull: cv2.typing.MatLike | None = ..., - clockwise: bool = ..., - returnPoints: bool = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def convexHull(points: UMat, hull: UMat | None = ..., clockwise: bool = ..., returnPoints: bool = ...) -> UMat: ... -@typing.overload -def convexityDefects( - contour: cv2.typing.MatLike, - convexhull: cv2.typing.MatLike, - convexityDefects: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def convexityDefects(contour: UMat, convexhull: UMat, convexityDefects: UMat | None = ...) -> UMat: ... -@typing.overload -def copyMakeBorder( - src: cv2.typing.MatLike, - top: int, - bottom: int, - left: int, - right: int, - borderType: int, - dst: cv2.typing.MatLike | None = ..., - value: cv2.typing.Scalar = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def copyMakeBorder( - src: UMat, - top: int, - bottom: int, - left: int, - right: int, - borderType: int, - dst: UMat | None = ..., - value: cv2.typing.Scalar = ..., -) -> UMat: ... -@typing.overload -def copyTo( - src: cv2.typing.MatLike, - mask: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def copyTo(src: UMat, mask: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def cornerEigenValsAndVecs( - src: cv2.typing.MatLike, - blockSize: int, - ksize: int, - dst: cv2.typing.MatLike | None = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def cornerEigenValsAndVecs( - src: UMat, - blockSize: int, - ksize: int, - dst: UMat | None = ..., - borderType: int = ..., -) -> UMat: ... -@typing.overload -def cornerHarris( - src: cv2.typing.MatLike, - blockSize: int, - ksize: int, - k: float, - dst: cv2.typing.MatLike | None = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def cornerHarris( - src: UMat, - blockSize: int, - ksize: int, - k: float, - dst: UMat | None = ..., - borderType: int = ..., -) -> UMat: ... -@typing.overload -def cornerMinEigenVal( - src: cv2.typing.MatLike, - blockSize: int, - dst: cv2.typing.MatLike | None = ..., - ksize: int = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def cornerMinEigenVal( - src: UMat, - blockSize: int, - dst: UMat | None = ..., - ksize: int = ..., - borderType: int = ..., -) -> UMat: ... -@typing.overload -def cornerSubPix( - image: cv2.typing.MatLike, - corners: cv2.typing.MatLike, - winSize: cv2.typing.Size, - zeroZone: cv2.typing.Size, - criteria: cv2.typing.TermCriteria, -) -> cv2.typing.MatLike: ... -@typing.overload -def cornerSubPix( - image: UMat, - corners: UMat, - winSize: cv2.typing.Size, - zeroZone: cv2.typing.Size, - criteria: cv2.typing.TermCriteria, -) -> UMat: ... -@typing.overload -def correctMatches( - F: cv2.typing.MatLike, - points1: cv2.typing.MatLike, - points2: cv2.typing.MatLike, - newPoints1: cv2.typing.MatLike | None = ..., - newPoints2: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def correctMatches( - F: UMat, - points1: UMat, - points2: UMat, - newPoints1: UMat | None = ..., - newPoints2: UMat | None = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def countNonZero(src: cv2.typing.MatLike) -> int: ... -@typing.overload -def countNonZero(src: UMat) -> int: ... -def createAlignMTB(max_bits: int = ..., exclude_range: int = ..., cut: bool = ...) -> AlignMTB: ... -def createBackgroundSubtractorKNN( - history: int = ..., - dist2Threshold: float = ..., - detectShadows: bool = ..., -) -> BackgroundSubtractorKNN: ... -def createBackgroundSubtractorMOG2( - history: int = ..., - varThreshold: float = ..., - detectShadows: bool = ..., -) -> BackgroundSubtractorMOG2: ... -def createCLAHE(clipLimit: float = ..., tileGridSize: cv2.typing.Size = ...) -> CLAHE: ... -def createCalibrateDebevec(samples: int = ..., lambda_: float = ..., random: bool = ...) -> CalibrateDebevec: ... -def createCalibrateRobertson(max_iter: int = ..., threshold: float = ...) -> CalibrateRobertson: ... -def createGeneralizedHoughBallard() -> GeneralizedHoughBallard: ... -def createGeneralizedHoughGuil() -> GeneralizedHoughGuil: ... -@typing.overload -def createHanningWindow( - winSize: cv2.typing.Size, - type: int, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def createHanningWindow(winSize: cv2.typing.Size, type: int, dst: UMat | None = ...) -> UMat: ... -def createLineSegmentDetector( - refine: int = ..., - scale: float = ..., - sigma_scale: float = ..., - quant: float = ..., - ang_th: float = ..., - log_eps: float = ..., - density_th: float = ..., - n_bins: int = ..., -) -> LineSegmentDetector: ... -def createMergeDebevec() -> MergeDebevec: ... -def createMergeMertens( - contrast_weight: float = ..., - saturation_weight: float = ..., - exposure_weight: float = ..., -) -> MergeMertens: ... -def createMergeRobertson() -> MergeRobertson: ... -def createTonemap(gamma: float = ...) -> Tonemap: ... -def createTonemapDrago(gamma: float = ..., saturation: float = ..., bias: float = ...) -> TonemapDrago: ... -def createTonemapMantiuk(gamma: float = ..., scale: float = ..., saturation: float = ...) -> TonemapMantiuk: ... -def createTonemapReinhard( - gamma: float = ..., - intensity: float = ..., - light_adapt: float = ..., - color_adapt: float = ..., -) -> TonemapReinhard: ... -def cubeRoot(val: float) -> float: ... -@typing.overload -def cvtColor( - src: cv2.typing.MatLike, - code: int, - dst: cv2.typing.MatLike | None = ..., - dstCn: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def cvtColor(src: UMat, code: int, dst: UMat | None = ..., dstCn: int = ...) -> UMat: ... -@typing.overload -def cvtColorTwoPlane( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - code: int, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def cvtColorTwoPlane(src1: UMat, src2: UMat, code: int, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def dct(src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ..., flags: int = ...) -> cv2.typing.MatLike: ... -@typing.overload -def dct(src: UMat, dst: UMat | None = ..., flags: int = ...) -> UMat: ... -@typing.overload -def decolor( - src: cv2.typing.MatLike, - grayscale: cv2.typing.MatLike | None = ..., - color_boost: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def decolor(src: UMat, grayscale: UMat | None = ..., color_boost: UMat | None = ...) -> tuple[UMat, UMat]: ... -@typing.overload -def decomposeEssentialMat( - E: cv2.typing.MatLike, - R1: cv2.typing.MatLike | None = ..., - R2: cv2.typing.MatLike | None = ..., - t: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def decomposeEssentialMat( - E: UMat, - R1: UMat | None = ..., - R2: UMat | None = ..., - t: UMat | None = ..., -) -> tuple[ - UMat, - UMat, - UMat, -]: ... -@typing.overload -def decomposeHomographyMat( - H: cv2.typing.MatLike, - K: cv2.typing.MatLike, - rotations: typing.Sequence[cv2.typing.MatLike] | None = ..., - translations: typing.Sequence[cv2.typing.MatLike] | None = ..., - normals: typing.Sequence[cv2.typing.MatLike] | None = ..., -) -> tuple[ - int, - typing.Sequence[cv2.typing.MatLike], - typing.Sequence[cv2.typing.MatLike], - typing.Sequence[cv2.typing.MatLike], -]: ... -@typing.overload -def decomposeHomographyMat( - H: UMat, - K: UMat, - rotations: typing.Sequence[UMat] | None = ..., - translations: typing.Sequence[UMat] | None = ..., - normals: typing.Sequence[UMat] | None = ..., -) -> tuple[ - int, - typing.Sequence[UMat], - typing.Sequence[UMat], - typing.Sequence[UMat], -]: ... -@typing.overload -def decomposeProjectionMatrix( - projMatrix: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike | None = ..., - rotMatrix: cv2.typing.MatLike | None = ..., - transVect: cv2.typing.MatLike | None = ..., - rotMatrixX: cv2.typing.MatLike | None = ..., - rotMatrixY: cv2.typing.MatLike | None = ..., - rotMatrixZ: cv2.typing.MatLike | None = ..., - eulerAngles: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def decomposeProjectionMatrix( - projMatrix: UMat, - cameraMatrix: UMat | None = ..., - rotMatrix: UMat | None = ..., - transVect: UMat | None = ..., - rotMatrixX: UMat | None = ..., - rotMatrixY: UMat | None = ..., - rotMatrixZ: UMat | None = ..., - eulerAngles: UMat | None = ..., -) -> tuple[ - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def demosaicing( - src: cv2.typing.MatLike, - code: int, - dst: cv2.typing.MatLike | None = ..., - dstCn: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def demosaicing(src: UMat, code: int, dst: UMat | None = ..., dstCn: int = ...) -> UMat: ... -def denoise_TVL1( - observations: typing.Sequence[cv2.typing.MatLike], - result: cv2.typing.MatLike, - lambda_: float = ..., - niters: int = ..., -) -> None: ... -def destroyAllWindows() -> None: ... -def destroyWindow(winname: str) -> None: ... -@typing.overload -def detailEnhance( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - sigma_s: float = ..., - sigma_r: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def detailEnhance(src: UMat, dst: UMat | None = ..., sigma_s: float = ..., sigma_r: float = ...) -> UMat: ... -@typing.overload -def determinant(mtx: cv2.typing.MatLike) -> float: ... -@typing.overload -def determinant(mtx: UMat) -> float: ... -@typing.overload -def dft( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - flags: int = ..., - nonzeroRows: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def dft(src: UMat, dst: UMat | None = ..., flags: int = ..., nonzeroRows: int = ...) -> UMat: ... -@typing.overload -def dilate( - src: cv2.typing.MatLike, - kernel: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - anchor: cv2.typing.Point = ..., - iterations: int = ..., - borderType: int = ..., - borderValue: cv2.typing.Scalar = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def dilate( - src: UMat, - kernel: UMat, - dst: UMat | None = ..., - anchor: cv2.typing.Point = ..., - iterations: int = ..., - borderType: int = ..., - borderValue: cv2.typing.Scalar = ..., -) -> UMat: ... -def displayOverlay(winname: str, text: str, delayms: int = ...) -> None: ... -def displayStatusBar(winname: str, text: str, delayms: int = ...) -> None: ... -@typing.overload -def distanceTransform( - src: cv2.typing.MatLike, - distanceType: int, - maskSize: int, - dst: cv2.typing.MatLike | None = ..., - dstType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def distanceTransform( - src: UMat, - distanceType: int, - maskSize: int, - dst: UMat | None = ..., - dstType: int = ..., -) -> UMat: ... -@typing.overload -def distanceTransformWithLabels( - src: cv2.typing.MatLike, - distanceType: int, - maskSize: int, - dst: cv2.typing.MatLike | None = ..., - labels: cv2.typing.MatLike | None = ..., - labelType: int = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def distanceTransformWithLabels( - src: UMat, - distanceType: int, - maskSize: int, - dst: UMat | None = ..., - labels: UMat | None = ..., - labelType: int = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def divSpectrums( - a: cv2.typing.MatLike, - b: cv2.typing.MatLike, - flags: int, - c: cv2.typing.MatLike | None = ..., - conjB: bool = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def divSpectrums(a: UMat, b: UMat, flags: int, c: UMat | None = ..., conjB: bool = ...) -> UMat: ... -@typing.overload -def divide( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - scale: float = ..., - dtype: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def divide(src1: UMat, src2: UMat, dst: UMat | None = ..., scale: float = ..., dtype: int = ...) -> UMat: ... -@typing.overload -def divide( - scale: float, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - dtype: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def divide(scale: float, src2: UMat, dst: UMat | None = ..., dtype: int = ...) -> UMat: ... -@typing.overload -def drawChessboardCorners( - image: cv2.typing.MatLike, - patternSize: cv2.typing.Size, - corners: cv2.typing.MatLike, - patternWasFound: bool, -) -> cv2.typing.MatLike: ... -@typing.overload -def drawChessboardCorners(image: UMat, patternSize: cv2.typing.Size, corners: UMat, patternWasFound: bool) -> UMat: ... -@typing.overload -def drawContours( - image: cv2.typing.MatLike, - contours: typing.Sequence[cv2.typing.MatLike], - contourIdx: int, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - hierarchy: cv2.typing.MatLike | None = ..., - maxLevel: int = ..., - offset: cv2.typing.Point = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def drawContours( - image: UMat, - contours: typing.Sequence[UMat], - contourIdx: int, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - hierarchy: UMat | None = ..., - maxLevel: int = ..., - offset: cv2.typing.Point = ..., -) -> UMat: ... -@typing.overload -def drawFrameAxes( - image: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - rvec: cv2.typing.MatLike, - tvec: cv2.typing.MatLike, - length: float, - thickness: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def drawFrameAxes( - image: UMat, - cameraMatrix: UMat, - distCoeffs: UMat, - rvec: UMat, - tvec: UMat, - length: float, - thickness: int = ..., -) -> UMat: ... -@typing.overload -def drawKeypoints( - image: cv2.typing.MatLike, - keypoints: typing.Sequence[KeyPoint], - outImage: cv2.typing.MatLike, - color: cv2.typing.Scalar = ..., - flags: DrawMatchesFlags = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def drawKeypoints( - image: UMat, - keypoints: typing.Sequence[KeyPoint], - outImage: UMat, - color: cv2.typing.Scalar = ..., - flags: DrawMatchesFlags = ..., -) -> UMat: ... -@typing.overload -def drawMarker( - img: cv2.typing.MatLike, - position: cv2.typing.Point, - color: cv2.typing.Scalar, - markerType: int = ..., - markerSize: int = ..., - thickness: int = ..., - line_type: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def drawMarker( - img: UMat, - position: cv2.typing.Point, - color: cv2.typing.Scalar, - markerType: int = ..., - markerSize: int = ..., - thickness: int = ..., - line_type: int = ..., -) -> UMat: ... -@typing.overload -def drawMatches( - img1: cv2.typing.MatLike, - keypoints1: typing.Sequence[KeyPoint], - img2: cv2.typing.MatLike, - keypoints2: typing.Sequence[KeyPoint], - matches1to2: typing.Sequence[DMatch], - outImg: cv2.typing.MatLike, - matchColor: cv2.typing.Scalar = ..., - singlePointColor: cv2.typing.Scalar = ..., - matchesMask: typing.Sequence[str] = ..., - flags: DrawMatchesFlags = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def drawMatches( - img1: UMat, - keypoints1: typing.Sequence[KeyPoint], - img2: UMat, - keypoints2: typing.Sequence[KeyPoint], - matches1to2: typing.Sequence[DMatch], - outImg: UMat, - matchColor: cv2.typing.Scalar = ..., - singlePointColor: cv2.typing.Scalar = ..., - matchesMask: typing.Sequence[str] = ..., - flags: DrawMatchesFlags = ..., -) -> UMat: ... -@typing.overload -def drawMatches( - img1: cv2.typing.MatLike, - keypoints1: typing.Sequence[KeyPoint], - img2: cv2.typing.MatLike, - keypoints2: typing.Sequence[KeyPoint], - matches1to2: typing.Sequence[DMatch], - outImg: cv2.typing.MatLike, - matchesThickness: int, - matchColor: cv2.typing.Scalar = ..., - singlePointColor: cv2.typing.Scalar = ..., - matchesMask: typing.Sequence[str] = ..., - flags: DrawMatchesFlags = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def drawMatches( - img1: UMat, - keypoints1: typing.Sequence[KeyPoint], - img2: UMat, - keypoints2: typing.Sequence[KeyPoint], - matches1to2: typing.Sequence[DMatch], - outImg: UMat, - matchesThickness: int, - matchColor: cv2.typing.Scalar = ..., - singlePointColor: cv2.typing.Scalar = ..., - matchesMask: typing.Sequence[str] = ..., - flags: DrawMatchesFlags = ..., -) -> UMat: ... -@typing.overload -def drawMatchesKnn( - img1: cv2.typing.MatLike, - keypoints1: typing.Sequence[KeyPoint], - img2: cv2.typing.MatLike, - keypoints2: typing.Sequence[KeyPoint], - matches1to2: typing.Sequence[typing.Sequence[DMatch]], - outImg: cv2.typing.MatLike, - matchColor: cv2.typing.Scalar = ..., - singlePointColor: cv2.typing.Scalar = ..., - matchesMask: typing.Sequence[typing.Sequence[str]] = ..., - flags: DrawMatchesFlags = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def drawMatchesKnn( - img1: UMat, - keypoints1: typing.Sequence[KeyPoint], - img2: UMat, - keypoints2: typing.Sequence[KeyPoint], - matches1to2: typing.Sequence[typing.Sequence[DMatch]], - outImg: UMat, - matchColor: cv2.typing.Scalar = ..., - singlePointColor: cv2.typing.Scalar = ..., - matchesMask: typing.Sequence[typing.Sequence[str]] = ..., - flags: DrawMatchesFlags = ..., -) -> UMat: ... -@typing.overload -def edgePreservingFilter( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - flags: int = ..., - sigma_s: float = ..., - sigma_r: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def edgePreservingFilter( - src: UMat, - dst: UMat | None = ..., - flags: int = ..., - sigma_s: float = ..., - sigma_r: float = ..., -) -> UMat: ... -@typing.overload -def eigen( - src: cv2.typing.MatLike, - eigenvalues: cv2.typing.MatLike | None = ..., - eigenvectors: cv2.typing.MatLike | None = ..., -) -> tuple[ - bool, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def eigen(src: UMat, eigenvalues: UMat | None = ..., eigenvectors: UMat | None = ...) -> tuple[bool, UMat, UMat]: ... -@typing.overload -def eigenNonSymmetric( - src: cv2.typing.MatLike, - eigenvalues: cv2.typing.MatLike | None = ..., - eigenvectors: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def eigenNonSymmetric( - src: UMat, - eigenvalues: UMat | None = ..., - eigenvectors: UMat | None = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def ellipse( - img: cv2.typing.MatLike, - center: cv2.typing.Point, - axes: cv2.typing.Size, - angle: float, - startAngle: float, - endAngle: float, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - shift: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def ellipse( - img: UMat, - center: cv2.typing.Point, - axes: cv2.typing.Size, - angle: float, - startAngle: float, - endAngle: float, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - shift: int = ..., -) -> UMat: ... -@typing.overload -def ellipse( - img: cv2.typing.MatLike, - box: cv2.typing.RotatedRect, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def ellipse( - img: UMat, - box: cv2.typing.RotatedRect, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., -) -> UMat: ... -def ellipse2Poly( - center: cv2.typing.Point, - axes: cv2.typing.Size, - angle: int, - arcStart: int, - arcEnd: int, - delta: int, -) -> typing.Sequence[cv2.typing.Point]: ... -def empty_array_desc() -> GArrayDesc: ... -def empty_gopaque_desc() -> GOpaqueDesc: ... -def empty_scalar_desc() -> GScalarDesc: ... -@typing.overload -def equalizeHist(src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def equalizeHist(src: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def erode( - src: cv2.typing.MatLike, - kernel: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - anchor: cv2.typing.Point = ..., - iterations: int = ..., - borderType: int = ..., - borderValue: cv2.typing.Scalar = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def erode( - src: UMat, - kernel: UMat, - dst: UMat | None = ..., - anchor: cv2.typing.Point = ..., - iterations: int = ..., - borderType: int = ..., - borderValue: cv2.typing.Scalar = ..., -) -> UMat: ... -@typing.overload -def estimateAffine2D( - from_: cv2.typing.MatLike, - to: cv2.typing.MatLike, - inliers: cv2.typing.MatLike | None = ..., - method: int = ..., - ransacReprojThreshold: float = ..., - maxIters: int = ..., - confidence: float = ..., - refineIters: int = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def estimateAffine2D( - from_: UMat, - to: UMat, - inliers: UMat | None = ..., - method: int = ..., - ransacReprojThreshold: float = ..., - maxIters: int = ..., - confidence: float = ..., - refineIters: int = ..., -) -> tuple[ - cv2.typing.MatLike, - UMat, -]: ... -@typing.overload -def estimateAffine2D( - pts1: cv2.typing.MatLike, - pts2: cv2.typing.MatLike, - params: UsacParams, - inliers: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def estimateAffine2D( - pts1: UMat, - pts2: UMat, - params: UsacParams, - inliers: UMat | None = ..., -) -> tuple[ - cv2.typing.MatLike, - UMat, -]: ... -@typing.overload -def estimateAffine3D( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike, - out: cv2.typing.MatLike | None = ..., - inliers: cv2.typing.MatLike | None = ..., - ransacThreshold: float = ..., - confidence: float = ..., -) -> tuple[ - int, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def estimateAffine3D( - src: UMat, - dst: UMat, - out: UMat | None = ..., - inliers: UMat | None = ..., - ransacThreshold: float = ..., - confidence: float = ..., -) -> tuple[ - int, - UMat, - UMat, -]: ... -@typing.overload -def estimateAffine3D( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike, - force_rotation: bool = ..., -) -> tuple[ - cv2.typing.MatLike, - float, -]: ... -@typing.overload -def estimateAffine3D(src: UMat, dst: UMat, force_rotation: bool = ...) -> tuple[cv2.typing.MatLike, float]: ... -@typing.overload -def estimateAffinePartial2D( - from_: cv2.typing.MatLike, - to: cv2.typing.MatLike, - inliers: cv2.typing.MatLike | None = ..., - method: int = ..., - ransacReprojThreshold: float = ..., - maxIters: int = ..., - confidence: float = ..., - refineIters: int = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def estimateAffinePartial2D( - from_: UMat, - to: UMat, - inliers: UMat | None = ..., - method: int = ..., - ransacReprojThreshold: float = ..., - maxIters: int = ..., - confidence: float = ..., - refineIters: int = ..., -) -> tuple[ - cv2.typing.MatLike, - UMat, -]: ... -@typing.overload -def estimateChessboardSharpness( - image: cv2.typing.MatLike, - patternSize: cv2.typing.Size, - corners: cv2.typing.MatLike, - rise_distance: float = ..., - vertical: bool = ..., - sharpness: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.Scalar, - cv2.typing.MatLike, -]: ... -@typing.overload -def estimateChessboardSharpness( - image: UMat, - patternSize: cv2.typing.Size, - corners: UMat, - rise_distance: float = ..., - vertical: bool = ..., - sharpness: UMat | None = ..., -) -> tuple[ - cv2.typing.Scalar, - UMat, -]: ... -@typing.overload -def estimateTranslation3D( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike, - out: cv2.typing.MatLike | None = ..., - inliers: cv2.typing.MatLike | None = ..., - ransacThreshold: float = ..., - confidence: float = ..., -) -> tuple[ - int, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def estimateTranslation3D( - src: UMat, - dst: UMat, - out: UMat | None = ..., - inliers: UMat | None = ..., - ransacThreshold: float = ..., - confidence: float = ..., -) -> tuple[ - int, - UMat, - UMat, -]: ... -@typing.overload -def exp(src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def exp(src: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def extractChannel(src: cv2.typing.MatLike, coi: int, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def extractChannel(src: UMat, coi: int, dst: UMat | None = ...) -> UMat: ... -def fastAtan2(y: float, x: float) -> float: ... -@typing.overload -def fastNlMeansDenoising( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - h: float = ..., - templateWindowSize: int = ..., - searchWindowSize: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def fastNlMeansDenoising( - src: UMat, - dst: UMat | None = ..., - h: float = ..., - templateWindowSize: int = ..., - searchWindowSize: int = ..., -) -> UMat: ... -@typing.overload -def fastNlMeansDenoising( - src: cv2.typing.MatLike, - h: typing.Sequence[float], - dst: cv2.typing.MatLike | None = ..., - templateWindowSize: int = ..., - searchWindowSize: int = ..., - normType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def fastNlMeansDenoising( - src: UMat, - h: typing.Sequence[float], - dst: UMat | None = ..., - templateWindowSize: int = ..., - searchWindowSize: int = ..., - normType: int = ..., -) -> UMat: ... -@typing.overload -def fastNlMeansDenoisingColored( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - h: float = ..., - hColor: float = ..., - templateWindowSize: int = ..., - searchWindowSize: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def fastNlMeansDenoisingColored( - src: UMat, - dst: UMat | None = ..., - h: float = ..., - hColor: float = ..., - templateWindowSize: int = ..., - searchWindowSize: int = ..., -) -> UMat: ... -@typing.overload -def fastNlMeansDenoisingColoredMulti( - srcImgs: typing.Sequence[cv2.typing.MatLike], - imgToDenoiseIndex: int, - temporalWindowSize: int, - dst: cv2.typing.MatLike | None = ..., - h: float = ..., - hColor: float = ..., - templateWindowSize: int = ..., - searchWindowSize: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def fastNlMeansDenoisingColoredMulti( - srcImgs: typing.Sequence[UMat], - imgToDenoiseIndex: int, - temporalWindowSize: int, - dst: UMat | None = ..., - h: float = ..., - hColor: float = ..., - templateWindowSize: int = ..., - searchWindowSize: int = ..., -) -> UMat: ... -@typing.overload -def fastNlMeansDenoisingMulti( - srcImgs: typing.Sequence[cv2.typing.MatLike], - imgToDenoiseIndex: int, - temporalWindowSize: int, - dst: cv2.typing.MatLike | None = ..., - h: float = ..., - templateWindowSize: int = ..., - searchWindowSize: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def fastNlMeansDenoisingMulti( - srcImgs: typing.Sequence[UMat], - imgToDenoiseIndex: int, - temporalWindowSize: int, - dst: UMat | None = ..., - h: float = ..., - templateWindowSize: int = ..., - searchWindowSize: int = ..., -) -> UMat: ... -@typing.overload -def fastNlMeansDenoisingMulti( - srcImgs: typing.Sequence[cv2.typing.MatLike], - imgToDenoiseIndex: int, - temporalWindowSize: int, - h: typing.Sequence[float], - dst: cv2.typing.MatLike | None = ..., - templateWindowSize: int = ..., - searchWindowSize: int = ..., - normType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def fastNlMeansDenoisingMulti( - srcImgs: typing.Sequence[UMat], - imgToDenoiseIndex: int, - temporalWindowSize: int, - h: typing.Sequence[float], - dst: UMat | None = ..., - templateWindowSize: int = ..., - searchWindowSize: int = ..., - normType: int = ..., -) -> UMat: ... -@typing.overload -def fillConvexPoly( - img: cv2.typing.MatLike, - points: cv2.typing.MatLike, - color: cv2.typing.Scalar, - lineType: int = ..., - shift: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def fillConvexPoly( - img: UMat, - points: UMat, - color: cv2.typing.Scalar, - lineType: int = ..., - shift: int = ..., -) -> UMat: ... -@typing.overload -def fillPoly( - img: cv2.typing.MatLike, - pts: typing.Sequence[cv2.typing.MatLike], - color: cv2.typing.Scalar, - lineType: int = ..., - shift: int = ..., - offset: cv2.typing.Point = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def fillPoly( - img: UMat, - pts: typing.Sequence[UMat], - color: cv2.typing.Scalar, - lineType: int = ..., - shift: int = ..., - offset: cv2.typing.Point = ..., -) -> UMat: ... -@typing.overload -def filter2D( - src: cv2.typing.MatLike, - ddepth: int, - kernel: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - anchor: cv2.typing.Point = ..., - delta: float = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def filter2D( - src: UMat, - ddepth: int, - kernel: UMat, - dst: UMat | None = ..., - anchor: cv2.typing.Point = ..., - delta: float = ..., - borderType: int = ..., -) -> UMat: ... -@typing.overload -def filterHomographyDecompByVisibleRefpoints( - rotations: typing.Sequence[cv2.typing.MatLike], - normals: typing.Sequence[cv2.typing.MatLike], - beforePoints: cv2.typing.MatLike, - afterPoints: cv2.typing.MatLike, - possibleSolutions: cv2.typing.MatLike | None = ..., - pointsMask: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def filterHomographyDecompByVisibleRefpoints( - rotations: typing.Sequence[UMat], - normals: typing.Sequence[UMat], - beforePoints: UMat, - afterPoints: UMat, - possibleSolutions: UMat | None = ..., - pointsMask: UMat | None = ..., -) -> UMat: ... -@typing.overload -def filterSpeckles( - img: cv2.typing.MatLike, - newVal: float, - maxSpeckleSize: int, - maxDiff: float, - buf: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def filterSpeckles( - img: UMat, - newVal: float, - maxSpeckleSize: int, - maxDiff: float, - buf: UMat | None = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def find4QuadCornerSubpix( - img: cv2.typing.MatLike, - corners: cv2.typing.MatLike, - region_size: cv2.typing.Size, -) -> tuple[ - bool, - cv2.typing.MatLike, -]: ... -@typing.overload -def find4QuadCornerSubpix(img: UMat, corners: UMat, region_size: cv2.typing.Size) -> tuple[bool, UMat]: ... -@typing.overload -def findChessboardCorners( - image: cv2.typing.MatLike, - patternSize: cv2.typing.Size, - corners: cv2.typing.MatLike | None = ..., - flags: int = ..., -) -> tuple[ - bool, - cv2.typing.MatLike, -]: ... -@typing.overload -def findChessboardCorners( - image: UMat, - patternSize: cv2.typing.Size, - corners: UMat | None = ..., - flags: int = ..., -) -> tuple[ - bool, - UMat, -]: ... -@typing.overload -def findChessboardCornersSB( - image: cv2.typing.MatLike, - patternSize: cv2.typing.Size, - corners: cv2.typing.MatLike | None = ..., - flags: int = ..., -) -> tuple[ - bool, - cv2.typing.MatLike, -]: ... -@typing.overload -def findChessboardCornersSB( - image: UMat, - patternSize: cv2.typing.Size, - corners: UMat | None = ..., - flags: int = ..., -) -> tuple[ - bool, - UMat, -]: ... -@typing.overload -def findChessboardCornersSBWithMeta( - image: cv2.typing.MatLike, - patternSize: cv2.typing.Size, - flags: int, - corners: cv2.typing.MatLike | None = ..., - meta: cv2.typing.MatLike | None = ..., -) -> tuple[ - bool, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def findChessboardCornersSBWithMeta( - image: UMat, - patternSize: cv2.typing.Size, - flags: int, - corners: UMat | None = ..., - meta: UMat | None = ..., -) -> tuple[ - bool, - UMat, - UMat, -]: ... -@typing.overload -def findCirclesGrid( - image: cv2.typing.MatLike, - patternSize: cv2.typing.Size, - flags: int, - blobDetector: cv2.typing.FeatureDetector, - parameters: CirclesGridFinderParameters, - centers: cv2.typing.MatLike | None = ..., -) -> tuple[ - bool, - cv2.typing.MatLike, -]: ... -@typing.overload -def findCirclesGrid( - image: UMat, - patternSize: cv2.typing.Size, - flags: int, - blobDetector: cv2.typing.FeatureDetector, - parameters: CirclesGridFinderParameters, - centers: UMat | None = ..., -) -> tuple[ - bool, - UMat, -]: ... -@typing.overload -def findCirclesGrid( - image: cv2.typing.MatLike, - patternSize: cv2.typing.Size, - centers: cv2.typing.MatLike | None = ..., - flags: int = ..., - blobDetector: cv2.typing.FeatureDetector = ..., -) -> tuple[ - bool, - cv2.typing.MatLike, -]: ... -@typing.overload -def findCirclesGrid( - image: UMat, - patternSize: cv2.typing.Size, - centers: UMat | None = ..., - flags: int = ..., - blobDetector: cv2.typing.FeatureDetector = ..., -) -> tuple[ - bool, - UMat, -]: ... -@typing.overload -def findContours( - image: cv2.typing.MatLike, - mode: int, - method: int, - contours: typing.Sequence[cv2.typing.MatLike] | None = ..., - hierarchy: cv2.typing.MatLike | None = ..., - offset: cv2.typing.Point = ..., -) -> tuple[ - typing.Sequence[cv2.typing.MatLike], - cv2.typing.MatLike, -]: ... -@typing.overload -def findContours( - image: UMat, - mode: int, - method: int, - contours: typing.Sequence[UMat] | None = ..., - hierarchy: UMat | None = ..., - offset: cv2.typing.Point = ..., -) -> tuple[ - typing.Sequence[UMat], - UMat, -]: ... -@typing.overload -def findEssentialMat( - points1: cv2.typing.MatLike, - points2: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - method: int = ..., - prob: float = ..., - threshold: float = ..., - maxIters: int = ..., - mask: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def findEssentialMat( - points1: UMat, - points2: UMat, - cameraMatrix: UMat, - method: int = ..., - prob: float = ..., - threshold: float = ..., - maxIters: int = ..., - mask: UMat | None = ..., -) -> tuple[ - cv2.typing.MatLike, - UMat, -]: ... -@typing.overload -def findEssentialMat( - points1: cv2.typing.MatLike, - points2: cv2.typing.MatLike, - focal: float = ..., - pp: cv2.typing.Point2d = ..., - method: int = ..., - prob: float = ..., - threshold: float = ..., - maxIters: int = ..., - mask: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def findEssentialMat( - points1: UMat, - points2: UMat, - focal: float = ..., - pp: cv2.typing.Point2d = ..., - method: int = ..., - prob: float = ..., - threshold: float = ..., - maxIters: int = ..., - mask: UMat | None = ..., -) -> tuple[ - cv2.typing.MatLike, - UMat, -]: ... -@typing.overload -def findEssentialMat( - points1: cv2.typing.MatLike, - points2: cv2.typing.MatLike, - cameraMatrix1: cv2.typing.MatLike, - distCoeffs1: cv2.typing.MatLike, - cameraMatrix2: cv2.typing.MatLike, - distCoeffs2: cv2.typing.MatLike, - method: int = ..., - prob: float = ..., - threshold: float = ..., - mask: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def findEssentialMat( - points1: UMat, - points2: UMat, - cameraMatrix1: UMat, - distCoeffs1: UMat, - cameraMatrix2: UMat, - distCoeffs2: UMat, - method: int = ..., - prob: float = ..., - threshold: float = ..., - mask: UMat | None = ..., -) -> tuple[ - cv2.typing.MatLike, - UMat, -]: ... -@typing.overload -def findEssentialMat( - points1: cv2.typing.MatLike, - points2: cv2.typing.MatLike, - cameraMatrix1: cv2.typing.MatLike, - cameraMatrix2: cv2.typing.MatLike, - dist_coeff1: cv2.typing.MatLike, - dist_coeff2: cv2.typing.MatLike, - params: UsacParams, - mask: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def findEssentialMat( - points1: UMat, - points2: UMat, - cameraMatrix1: UMat, - cameraMatrix2: UMat, - dist_coeff1: UMat, - dist_coeff2: UMat, - params: UsacParams, - mask: UMat | None = ..., -) -> tuple[ - cv2.typing.MatLike, - UMat, -]: ... -@typing.overload -def findFundamentalMat( - points1: cv2.typing.MatLike, - points2: cv2.typing.MatLike, - method: int, - ransacReprojThreshold: float, - confidence: float, - maxIters: int, - mask: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def findFundamentalMat( - points1: UMat, - points2: UMat, - method: int, - ransacReprojThreshold: float, - confidence: float, - maxIters: int, - mask: UMat | None = ..., -) -> tuple[ - cv2.typing.MatLike, - UMat, -]: ... -@typing.overload -def findFundamentalMat( - points1: cv2.typing.MatLike, - points2: cv2.typing.MatLike, - method: int = ..., - ransacReprojThreshold: float = ..., - confidence: float = ..., - mask: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def findFundamentalMat( - points1: UMat, - points2: UMat, - method: int = ..., - ransacReprojThreshold: float = ..., - confidence: float = ..., - mask: UMat | None = ..., -) -> tuple[ - cv2.typing.MatLike, - UMat, -]: ... -@typing.overload -def findFundamentalMat( - points1: cv2.typing.MatLike, - points2: cv2.typing.MatLike, - params: UsacParams, - mask: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def findFundamentalMat( - points1: UMat, - points2: UMat, - params: UsacParams, - mask: UMat | None = ..., -) -> tuple[ - cv2.typing.MatLike, - UMat, -]: ... -@typing.overload -def findHomography( - srcPoints: cv2.typing.MatLike, - dstPoints: cv2.typing.MatLike, - method: int = ..., - ransacReprojThreshold: float = ..., - mask: cv2.typing.MatLike | None = ..., - maxIters: int = ..., - confidence: float = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def findHomography( - srcPoints: UMat, - dstPoints: UMat, - method: int = ..., - ransacReprojThreshold: float = ..., - mask: UMat | None = ..., - maxIters: int = ..., - confidence: float = ..., -) -> tuple[ - cv2.typing.MatLike, - UMat, -]: ... -@typing.overload -def findHomography( - srcPoints: cv2.typing.MatLike, - dstPoints: cv2.typing.MatLike, - params: UsacParams, - mask: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def findHomography( - srcPoints: UMat, - dstPoints: UMat, - params: UsacParams, - mask: UMat | None = ..., -) -> tuple[ - cv2.typing.MatLike, - UMat, -]: ... -@typing.overload -def findNonZero(src: cv2.typing.MatLike, idx: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def findNonZero(src: UMat, idx: UMat | None = ...) -> UMat: ... -@typing.overload -def findTransformECC( - templateImage: cv2.typing.MatLike, - inputImage: cv2.typing.MatLike, - warpMatrix: cv2.typing.MatLike, - motionType: int, - criteria: cv2.typing.TermCriteria, - inputMask: cv2.typing.MatLike, - gaussFiltSize: int, -) -> tuple[ - float, - cv2.typing.MatLike, -]: ... -@typing.overload -def findTransformECC( - templateImage: UMat, - inputImage: UMat, - warpMatrix: UMat, - motionType: int, - criteria: cv2.typing.TermCriteria, - inputMask: UMat, - gaussFiltSize: int, -) -> tuple[ - float, - UMat, -]: ... -@typing.overload -def findTransformECC( - templateImage: cv2.typing.MatLike, - inputImage: cv2.typing.MatLike, - warpMatrix: cv2.typing.MatLike, - motionType: int = ..., - criteria: cv2.typing.TermCriteria = ..., - inputMask: cv2.typing.MatLike | None = ..., -) -> tuple[ - float, - cv2.typing.MatLike, -]: ... -@typing.overload -def findTransformECC( - templateImage: UMat, - inputImage: UMat, - warpMatrix: UMat, - motionType: int = ..., - criteria: cv2.typing.TermCriteria = ..., - inputMask: UMat | None = ..., -) -> tuple[ - float, - UMat, -]: ... -@typing.overload -def fitEllipse(points: cv2.typing.MatLike) -> cv2.typing.RotatedRect: ... -@typing.overload -def fitEllipse(points: UMat) -> cv2.typing.RotatedRect: ... -@typing.overload -def fitEllipseAMS(points: cv2.typing.MatLike) -> cv2.typing.RotatedRect: ... -@typing.overload -def fitEllipseAMS(points: UMat) -> cv2.typing.RotatedRect: ... -@typing.overload -def fitEllipseDirect(points: cv2.typing.MatLike) -> cv2.typing.RotatedRect: ... -@typing.overload -def fitEllipseDirect(points: UMat) -> cv2.typing.RotatedRect: ... -@typing.overload -def fitLine( - points: cv2.typing.MatLike, - distType: int, - param: float, - reps: float, - aeps: float, - line: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def fitLine(points: UMat, distType: int, param: float, reps: float, aeps: float, line: UMat | None = ...) -> UMat: ... -@typing.overload -def flip(src: cv2.typing.MatLike, flipCode: int, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def flip(src: UMat, flipCode: int, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def flipND(src: cv2.typing.MatLike, axis: int, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def flipND(src: UMat, axis: int, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def floodFill( - image: cv2.typing.MatLike, - mask: cv2.typing.MatLike, - seedPoint: cv2.typing.Point, - newVal: cv2.typing.Scalar, - loDiff: cv2.typing.Scalar = ..., - upDiff: cv2.typing.Scalar = ..., - flags: int = ..., -) -> tuple[ - int, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.Rect, -]: ... -@typing.overload -def floodFill( - image: UMat, - mask: UMat, - seedPoint: cv2.typing.Point, - newVal: cv2.typing.Scalar, - loDiff: cv2.typing.Scalar = ..., - upDiff: cv2.typing.Scalar = ..., - flags: int = ..., -) -> tuple[ - int, - UMat, - UMat, - cv2.typing.Rect, -]: ... -@typing.overload -def gemm( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - alpha: float, - src3: cv2.typing.MatLike, - beta: float, - dst: cv2.typing.MatLike | None = ..., - flags: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def gemm( - src1: UMat, - src2: UMat, - alpha: float, - src3: UMat, - beta: float, - dst: UMat | None = ..., - flags: int = ..., -) -> UMat: ... -@typing.overload -def getAffineTransform(src: cv2.typing.MatLike, dst: cv2.typing.MatLike) -> cv2.typing.MatLike: ... -@typing.overload -def getAffineTransform(src: UMat, dst: UMat) -> cv2.typing.MatLike: ... -def getBuildInformation() -> str: ... -def getCPUFeaturesLine() -> str: ... -def getCPUTickCount() -> int: ... -@typing.overload -def getDefaultNewCameraMatrix( - cameraMatrix: cv2.typing.MatLike, - imgsize: cv2.typing.Size = ..., - centerPrincipalPoint: bool = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def getDefaultNewCameraMatrix( - cameraMatrix: UMat, - imgsize: cv2.typing.Size = ..., - centerPrincipalPoint: bool = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def getDerivKernels( - dx: int, - dy: int, - ksize: int, - kx: cv2.typing.MatLike | None = ..., - ky: cv2.typing.MatLike | None = ..., - normalize: bool = ..., - ktype: int = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def getDerivKernels( - dx: int, - dy: int, - ksize: int, - kx: UMat | None = ..., - ky: UMat | None = ..., - normalize: bool = ..., - ktype: int = ..., -) -> tuple[ - UMat, - UMat, -]: ... -def getFontScaleFromHeight(fontFace: int, pixelHeight: int, thickness: int = ...) -> float: ... -def getGaborKernel( - ksize: cv2.typing.Size, - sigma: float, - theta: float, - lambd: float, - gamma: float, - psi: float = ..., - ktype: int = ..., -) -> cv2.typing.MatLike: ... -def getGaussianKernel(ksize: int, sigma: float, ktype: int = ...) -> cv2.typing.MatLike: ... -def getHardwareFeatureName(feature: int) -> str: ... -def getLogLevel() -> int: ... -def getNumThreads() -> int: ... -def getNumberOfCPUs() -> int: ... -def getOptimalDFTSize(vecsize: int) -> int: ... -@typing.overload -def getOptimalNewCameraMatrix( - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - imageSize: cv2.typing.Size, - alpha: float, - newImgSize: cv2.typing.Size = ..., - centerPrincipalPoint: bool = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.Rect, -]: ... -@typing.overload -def getOptimalNewCameraMatrix( - cameraMatrix: UMat, - distCoeffs: UMat, - imageSize: cv2.typing.Size, - alpha: float, - newImgSize: cv2.typing.Size = ..., - centerPrincipalPoint: bool = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.Rect, -]: ... -@typing.overload -def getPerspectiveTransform( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike, - solveMethod: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def getPerspectiveTransform(src: UMat, dst: UMat, solveMethod: int = ...) -> cv2.typing.MatLike: ... -@typing.overload -def getRectSubPix( - image: cv2.typing.MatLike, - patchSize: cv2.typing.Size, - center: cv2.typing.Point2f, - patch: cv2.typing.MatLike | None = ..., - patchType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def getRectSubPix( - image: UMat, - patchSize: cv2.typing.Size, - center: cv2.typing.Point2f, - patch: UMat | None = ..., - patchType: int = ..., -) -> UMat: ... -def getRotationMatrix2D(center: cv2.typing.Point2f, angle: float, scale: float) -> cv2.typing.MatLike: ... -def getStructuringElement(shape: int, ksize: cv2.typing.Size, anchor: cv2.typing.Point = ...) -> cv2.typing.MatLike: ... -def getTextSize(text: str, fontFace: int, fontScale: float, thickness: int) -> tuple[cv2.typing.Size, int]: ... -def getThreadNum() -> int: ... -def getTickCount() -> int: ... -def getTickFrequency() -> float: ... -def getTrackbarPos(trackbarname: str, winname: str) -> int: ... -def getValidDisparityROI( - roi1: cv2.typing.Rect, - roi2: cv2.typing.Rect, - minDisparity: int, - numberOfDisparities: int, - blockSize: int, -) -> cv2.typing.Rect: ... -def getVersionMajor() -> int: ... -def getVersionMinor() -> int: ... -def getVersionRevision() -> int: ... -def getVersionString() -> str: ... -def getWindowImageRect(winname: str) -> cv2.typing.Rect: ... -def getWindowProperty(winname: str, prop_id: int) -> float: ... -@typing.overload -def goodFeaturesToTrack( - image: cv2.typing.MatLike, - maxCorners: int, - qualityLevel: float, - minDistance: float, - corners: cv2.typing.MatLike | None = ..., - mask: cv2.typing.MatLike | None = ..., - blockSize: int = ..., - useHarrisDetector: bool = ..., - k: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def goodFeaturesToTrack( - image: UMat, - maxCorners: int, - qualityLevel: float, - minDistance: float, - corners: UMat | None = ..., - mask: UMat | None = ..., - blockSize: int = ..., - useHarrisDetector: bool = ..., - k: float = ..., -) -> UMat: ... -@typing.overload -def goodFeaturesToTrack( - image: cv2.typing.MatLike, - maxCorners: int, - qualityLevel: float, - minDistance: float, - mask: cv2.typing.MatLike, - blockSize: int, - gradientSize: int, - corners: cv2.typing.MatLike | None = ..., - useHarrisDetector: bool = ..., - k: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def goodFeaturesToTrack( - image: UMat, - maxCorners: int, - qualityLevel: float, - minDistance: float, - mask: UMat, - blockSize: int, - gradientSize: int, - corners: UMat | None = ..., - useHarrisDetector: bool = ..., - k: float = ..., -) -> UMat: ... -@typing.overload -def goodFeaturesToTrackWithQuality( - image: cv2.typing.MatLike, - maxCorners: int, - qualityLevel: float, - minDistance: float, - mask: cv2.typing.MatLike, - corners: cv2.typing.MatLike | None = ..., - cornersQuality: cv2.typing.MatLike | None = ..., - blockSize: int = ..., - gradientSize: int = ..., - useHarrisDetector: bool = ..., - k: float = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def goodFeaturesToTrackWithQuality( - image: UMat, - maxCorners: int, - qualityLevel: float, - minDistance: float, - mask: UMat, - corners: UMat | None = ..., - cornersQuality: UMat | None = ..., - blockSize: int = ..., - gradientSize: int = ..., - useHarrisDetector: bool = ..., - k: float = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def grabCut( - img: cv2.typing.MatLike, - mask: cv2.typing.MatLike, - rect: cv2.typing.Rect, - bgdModel: cv2.typing.MatLike, - fgdModel: cv2.typing.MatLike, - iterCount: int, - mode: int = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def grabCut( - img: UMat, - mask: UMat, - rect: cv2.typing.Rect, - bgdModel: UMat, - fgdModel: UMat, - iterCount: int, - mode: int = ..., -) -> tuple[ - UMat, - UMat, - UMat, -]: ... -def groupRectangles( - rectList: typing.Sequence[cv2.typing.Rect], - groupThreshold: int, - eps: float = ..., -) -> tuple[ - typing.Sequence[cv2.typing.Rect], - typing.Sequence[int], -]: ... -@typing.overload -def hasNonZero(src: cv2.typing.MatLike) -> bool: ... -@typing.overload -def hasNonZero(src: UMat) -> bool: ... -def haveImageReader(filename: str) -> bool: ... -def haveImageWriter(filename: str) -> bool: ... -def haveOpenVX() -> bool: ... -@typing.overload -def hconcat(src: typing.Sequence[cv2.typing.MatLike], dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def hconcat(src: typing.Sequence[UMat], dst: UMat | None = ...) -> UMat: ... -@typing.overload -def idct(src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ..., flags: int = ...) -> cv2.typing.MatLike: ... -@typing.overload -def idct(src: UMat, dst: UMat | None = ..., flags: int = ...) -> UMat: ... -@typing.overload -def idft( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - flags: int = ..., - nonzeroRows: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def idft(src: UMat, dst: UMat | None = ..., flags: int = ..., nonzeroRows: int = ...) -> UMat: ... -@typing.overload -def illuminationChange( - src: cv2.typing.MatLike, - mask: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - alpha: float = ..., - beta: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def illuminationChange( - src: UMat, - mask: UMat, - dst: UMat | None = ..., - alpha: float = ..., - beta: float = ..., -) -> UMat: ... -def imcount(filename: str, flags: int = ...) -> int: ... -@typing.overload -def imdecode(buf: cv2.typing.MatLike, flags: int) -> cv2.typing.MatLike: ... -@typing.overload -def imdecode(buf: UMat, flags: int) -> cv2.typing.MatLike: ... -@typing.overload -def imdecodemulti( - buf: cv2.typing.MatLike, - flags: int, - mats: typing.Sequence[cv2.typing.MatLike] | None = ..., -) -> tuple[ - bool, - typing.Sequence[cv2.typing.MatLike], -]: ... -@typing.overload -def imdecodemulti( - buf: UMat, - flags: int, - mats: typing.Sequence[cv2.typing.MatLike] | None = ..., -) -> tuple[ - bool, - typing.Sequence[cv2.typing.MatLike], -]: ... -@typing.overload -def imencode( - ext: str, - img: cv2.typing.MatLike, - params: typing.Sequence[int] = ..., -) -> tuple[ - bool, - numpy.ndarray[ - typing.Any, - numpy.dtype[numpy.uint8], - ], -]: ... -@typing.overload -def imencode( - ext: str, - img: UMat, - params: typing.Sequence[int] = ..., -) -> tuple[ - bool, - numpy.ndarray[ - typing.Any, - numpy.dtype[numpy.uint8], - ], -]: ... -def imread(filename: str, flags: int = ...) -> cv2.typing.MatLike: ... -@typing.overload -def imreadmulti( - filename: str, - mats: typing.Sequence[cv2.typing.MatLike] | None = ..., - flags: int = ..., -) -> tuple[ - bool, - typing.Sequence[cv2.typing.MatLike], -]: ... -@typing.overload -def imreadmulti( - filename: str, - start: int, - count: int, - mats: typing.Sequence[cv2.typing.MatLike] | None = ..., - flags: int = ..., -) -> tuple[ - bool, - typing.Sequence[cv2.typing.MatLike], -]: ... -@typing.overload -def imshow(winname: str, mat: cv2.typing.MatLike) -> None: ... -@typing.overload -def imshow(winname: str, mat: cv2.cuda.GpuMat) -> None: ... -@typing.overload -def imshow(winname: str, mat: UMat) -> None: ... -@typing.overload -def imwrite(filename: str, img: cv2.typing.MatLike, params: typing.Sequence[int] = ...) -> bool: ... -@typing.overload -def imwrite(filename: str, img: UMat, params: typing.Sequence[int] = ...) -> bool: ... -@typing.overload -def imwritemulti( - filename: str, - img: typing.Sequence[cv2.typing.MatLike], - params: typing.Sequence[int] = ..., -) -> bool: ... -@typing.overload -def imwritemulti(filename: str, img: typing.Sequence[UMat], params: typing.Sequence[int] = ...) -> bool: ... -@typing.overload -def inRange( - src: cv2.typing.MatLike, - lowerb: cv2.typing.MatLike, - upperb: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def inRange(src: UMat, lowerb: UMat, upperb: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def initCameraMatrix2D( - objectPoints: typing.Sequence[cv2.typing.MatLike], - imagePoints: typing.Sequence[cv2.typing.MatLike], - imageSize: cv2.typing.Size, - aspectRatio: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def initCameraMatrix2D( - objectPoints: typing.Sequence[UMat], - imagePoints: typing.Sequence[UMat], - imageSize: cv2.typing.Size, - aspectRatio: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def initInverseRectificationMap( - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - R: cv2.typing.MatLike, - newCameraMatrix: cv2.typing.MatLike, - size: cv2.typing.Size, - m1type: int, - map1: cv2.typing.MatLike | None = ..., - map2: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def initInverseRectificationMap( - cameraMatrix: UMat, - distCoeffs: UMat, - R: UMat, - newCameraMatrix: UMat, - size: cv2.typing.Size, - m1type: int, - map1: UMat | None = ..., - map2: UMat | None = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def initUndistortRectifyMap( - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - R: cv2.typing.MatLike, - newCameraMatrix: cv2.typing.MatLike, - size: cv2.typing.Size, - m1type: int, - map1: cv2.typing.MatLike | None = ..., - map2: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def initUndistortRectifyMap( - cameraMatrix: UMat, - distCoeffs: UMat, - R: UMat, - newCameraMatrix: UMat, - size: cv2.typing.Size, - m1type: int, - map1: UMat | None = ..., - map2: UMat | None = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def inpaint( - src: cv2.typing.MatLike, - inpaintMask: cv2.typing.MatLike, - inpaintRadius: float, - flags: int, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def inpaint(src: UMat, inpaintMask: UMat, inpaintRadius: float, flags: int, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def insertChannel(src: cv2.typing.MatLike, dst: cv2.typing.MatLike, coi: int) -> cv2.typing.MatLike: ... -@typing.overload -def insertChannel(src: UMat, dst: UMat, coi: int) -> UMat: ... -@typing.overload -def integral( - src: cv2.typing.MatLike, - sum: cv2.typing.MatLike | None = ..., - sdepth: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def integral(src: UMat, sum: UMat | None = ..., sdepth: int = ...) -> UMat: ... -@typing.overload -def integral2( - src: cv2.typing.MatLike, - sum: cv2.typing.MatLike | None = ..., - sqsum: cv2.typing.MatLike | None = ..., - sdepth: int = ..., - sqdepth: int = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def integral2( - src: UMat, - sum: UMat | None = ..., - sqsum: UMat | None = ..., - sdepth: int = ..., - sqdepth: int = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def integral3( - src: cv2.typing.MatLike, - sum: cv2.typing.MatLike | None = ..., - sqsum: cv2.typing.MatLike | None = ..., - tilted: cv2.typing.MatLike | None = ..., - sdepth: int = ..., - sqdepth: int = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def integral3( - src: UMat, - sum: UMat | None = ..., - sqsum: UMat | None = ..., - tilted: UMat | None = ..., - sdepth: int = ..., - sqdepth: int = ..., -) -> tuple[ - UMat, - UMat, - UMat, -]: ... -@typing.overload -def intersectConvexConvex( - p1: cv2.typing.MatLike, - p2: cv2.typing.MatLike, - p12: cv2.typing.MatLike | None = ..., - handleNested: bool = ..., -) -> tuple[ - float, - cv2.typing.MatLike, -]: ... -@typing.overload -def intersectConvexConvex( - p1: UMat, - p2: UMat, - p12: UMat | None = ..., - handleNested: bool = ..., -) -> tuple[ - float, - UMat, -]: ... -@typing.overload -def invert( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - flags: int = ..., -) -> tuple[ - float, - cv2.typing.MatLike, -]: ... -@typing.overload -def invert(src: UMat, dst: UMat | None = ..., flags: int = ...) -> tuple[float, UMat]: ... -@typing.overload -def invertAffineTransform(M: cv2.typing.MatLike, iM: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def invertAffineTransform(M: UMat, iM: UMat | None = ...) -> UMat: ... -@typing.overload -def isContourConvex(contour: cv2.typing.MatLike) -> bool: ... -@typing.overload -def isContourConvex(contour: UMat) -> bool: ... -@typing.overload -def kmeans( - data: cv2.typing.MatLike, - K: int, - bestLabels: cv2.typing.MatLike, - criteria: cv2.typing.TermCriteria, - attempts: int, - flags: int, - centers: cv2.typing.MatLike | None = ..., -) -> tuple[ - float, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def kmeans( - data: UMat, - K: int, - bestLabels: UMat, - criteria: cv2.typing.TermCriteria, - attempts: int, - flags: int, - centers: UMat | None = ..., -) -> tuple[ - float, - UMat, - UMat, -]: ... -@typing.overload -def line( - img: cv2.typing.MatLike, - pt1: cv2.typing.Point, - pt2: cv2.typing.Point, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - shift: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def line( - img: UMat, - pt1: cv2.typing.Point, - pt2: cv2.typing.Point, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - shift: int = ..., -) -> UMat: ... -@typing.overload -def linearPolar( - src: cv2.typing.MatLike, - center: cv2.typing.Point2f, - maxRadius: float, - flags: int, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def linearPolar( - src: UMat, - center: cv2.typing.Point2f, - maxRadius: float, - flags: int, - dst: UMat | None = ..., -) -> UMat: ... -@typing.overload -def log(src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def log(src: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def logPolar( - src: cv2.typing.MatLike, - center: cv2.typing.Point2f, - M: float, - flags: int, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def logPolar(src: UMat, center: cv2.typing.Point2f, M: float, flags: int, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def magnitude( - x: cv2.typing.MatLike, - y: cv2.typing.MatLike, - magnitude: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def magnitude(x: UMat, y: UMat, magnitude: UMat | None = ...) -> UMat: ... -@typing.overload -def matMulDeriv( - A: cv2.typing.MatLike, - B: cv2.typing.MatLike, - dABdA: cv2.typing.MatLike | None = ..., - dABdB: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def matMulDeriv(A: UMat, B: UMat, dABdA: UMat | None = ..., dABdB: UMat | None = ...) -> tuple[UMat, UMat]: ... -@typing.overload -def matchShapes(contour1: cv2.typing.MatLike, contour2: cv2.typing.MatLike, method: int, parameter: float) -> float: ... -@typing.overload -def matchShapes(contour1: UMat, contour2: UMat, method: int, parameter: float) -> float: ... -@typing.overload -def matchTemplate( - image: cv2.typing.MatLike, - templ: cv2.typing.MatLike, - method: int, - result: cv2.typing.MatLike | None = ..., - mask: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def matchTemplate( - image: UMat, - templ: UMat, - method: int, - result: UMat | None = ..., - mask: UMat | None = ..., -) -> UMat: ... -@typing.overload -def max( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def max(src1: UMat, src2: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def mean(src: cv2.typing.MatLike, mask: cv2.typing.MatLike | None = ...) -> cv2.typing.Scalar: ... -@typing.overload -def mean(src: UMat, mask: UMat | None = ...) -> cv2.typing.Scalar: ... -@typing.overload -def meanShift( - probImage: cv2.typing.MatLike, - window: cv2.typing.Rect, - criteria: cv2.typing.TermCriteria, -) -> tuple[ - int, - cv2.typing.Rect, -]: ... -@typing.overload -def meanShift( - probImage: UMat, - window: cv2.typing.Rect, - criteria: cv2.typing.TermCriteria, -) -> tuple[ - int, - cv2.typing.Rect, -]: ... -@typing.overload -def meanStdDev( - src: cv2.typing.MatLike, - mean: cv2.typing.MatLike | None = ..., - stddev: cv2.typing.MatLike | None = ..., - mask: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def meanStdDev( - src: UMat, - mean: UMat | None = ..., - stddev: UMat | None = ..., - mask: UMat | None = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def medianBlur(src: cv2.typing.MatLike, ksize: int, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def medianBlur(src: UMat, ksize: int, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def merge(mv: typing.Sequence[cv2.typing.MatLike], dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def merge(mv: typing.Sequence[UMat], dst: UMat | None = ...) -> UMat: ... -@typing.overload -def min( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def min(src1: UMat, src2: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def minAreaRect(points: cv2.typing.MatLike) -> cv2.typing.RotatedRect: ... -@typing.overload -def minAreaRect(points: UMat) -> cv2.typing.RotatedRect: ... -@typing.overload -def minEnclosingCircle(points: cv2.typing.MatLike) -> tuple[cv2.typing.Point2f, float]: ... -@typing.overload -def minEnclosingCircle(points: UMat) -> tuple[cv2.typing.Point2f, float]: ... -@typing.overload -def minEnclosingTriangle( - points: cv2.typing.MatLike, - triangle: cv2.typing.MatLike | None = ..., -) -> tuple[ - float, - cv2.typing.MatLike, -]: ... -@typing.overload -def minEnclosingTriangle(points: UMat, triangle: UMat | None = ...) -> tuple[float, UMat]: ... -@typing.overload -def minMaxLoc( - src: cv2.typing.MatLike, - mask: cv2.typing.MatLike | None = ..., -) -> tuple[ - float, - float, - cv2.typing.Point, - cv2.typing.Point, -]: ... -@typing.overload -def minMaxLoc(src: UMat, mask: UMat | None = ...) -> tuple[float, float, cv2.typing.Point, cv2.typing.Point]: ... -@typing.overload -def mixChannels( - src: typing.Sequence[cv2.typing.MatLike], - dst: typing.Sequence[cv2.typing.MatLike], - fromTo: typing.Sequence[int], -) -> typing.Sequence[cv2.typing.MatLike]: ... -@typing.overload -def mixChannels( - src: typing.Sequence[UMat], - dst: typing.Sequence[UMat], - fromTo: typing.Sequence[int], -) -> typing.Sequence[UMat]: ... -@typing.overload -def moments(array: cv2.typing.MatLike, binaryImage: bool = ...) -> cv2.typing.Moments: ... -@typing.overload -def moments(array: UMat, binaryImage: bool = ...) -> cv2.typing.Moments: ... -@typing.overload -def morphologyEx( - src: cv2.typing.MatLike, - op: int, - kernel: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - anchor: cv2.typing.Point = ..., - iterations: int = ..., - borderType: int = ..., - borderValue: cv2.typing.Scalar = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def morphologyEx( - src: UMat, - op: int, - kernel: UMat, - dst: UMat | None = ..., - anchor: cv2.typing.Point = ..., - iterations: int = ..., - borderType: int = ..., - borderValue: cv2.typing.Scalar = ..., -) -> UMat: ... -def moveWindow(winname: str, x: int, y: int) -> None: ... -@typing.overload -def mulSpectrums( - a: cv2.typing.MatLike, - b: cv2.typing.MatLike, - flags: int, - c: cv2.typing.MatLike | None = ..., - conjB: bool = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def mulSpectrums(a: UMat, b: UMat, flags: int, c: UMat | None = ..., conjB: bool = ...) -> UMat: ... -@typing.overload -def mulTransposed( - src: cv2.typing.MatLike, - aTa: bool, - dst: cv2.typing.MatLike | None = ..., - delta: cv2.typing.MatLike | None = ..., - scale: float = ..., - dtype: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def mulTransposed( - src: UMat, - aTa: bool, - dst: UMat | None = ..., - delta: UMat | None = ..., - scale: float = ..., - dtype: int = ..., -) -> UMat: ... -@typing.overload -def multiply( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - scale: float = ..., - dtype: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def multiply(src1: UMat, src2: UMat, dst: UMat | None = ..., scale: float = ..., dtype: int = ...) -> UMat: ... -def namedWindow(winname: str, flags: int = ...) -> None: ... -@typing.overload -def norm(src1: cv2.typing.MatLike, normType: int = ..., mask: cv2.typing.MatLike | None = ...) -> float: ... -@typing.overload -def norm(src1: UMat, normType: int = ..., mask: UMat | None = ...) -> float: ... -@typing.overload -def norm( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - normType: int = ..., - mask: cv2.typing.MatLike | None = ..., -) -> float: ... -@typing.overload -def norm(src1: UMat, src2: UMat, normType: int = ..., mask: UMat | None = ...) -> float: ... -@typing.overload -def normalize( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike, - alpha: float = ..., - beta: float = ..., - norm_type: int = ..., - dtype: int = ..., - mask: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def normalize( - src: UMat, - dst: UMat, - alpha: float = ..., - beta: float = ..., - norm_type: int = ..., - dtype: int = ..., - mask: UMat | None = ..., -) -> UMat: ... -@typing.overload -def patchNaNs(a: cv2.typing.MatLike, val: float = ...) -> cv2.typing.MatLike: ... -@typing.overload -def patchNaNs(a: UMat, val: float = ...) -> UMat: ... -@typing.overload -def pencilSketch( - src: cv2.typing.MatLike, - dst1: cv2.typing.MatLike | None = ..., - dst2: cv2.typing.MatLike | None = ..., - sigma_s: float = ..., - sigma_r: float = ..., - shade_factor: float = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def pencilSketch( - src: UMat, - dst1: UMat | None = ..., - dst2: UMat | None = ..., - sigma_s: float = ..., - sigma_r: float = ..., - shade_factor: float = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def perspectiveTransform( - src: cv2.typing.MatLike, - m: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def perspectiveTransform(src: UMat, m: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def phase( - x: cv2.typing.MatLike, - y: cv2.typing.MatLike, - angle: cv2.typing.MatLike | None = ..., - angleInDegrees: bool = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def phase(x: UMat, y: UMat, angle: UMat | None = ..., angleInDegrees: bool = ...) -> UMat: ... -@typing.overload -def phaseCorrelate( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - window: cv2.typing.MatLike | None = ..., -) -> tuple[ - cv2.typing.Point2d, - float, -]: ... -@typing.overload -def phaseCorrelate(src1: UMat, src2: UMat, window: UMat | None = ...) -> tuple[cv2.typing.Point2d, float]: ... -@typing.overload -def pointPolygonTest(contour: cv2.typing.MatLike, pt: cv2.typing.Point2f, measureDist: bool) -> float: ... -@typing.overload -def pointPolygonTest(contour: UMat, pt: cv2.typing.Point2f, measureDist: bool) -> float: ... -@typing.overload -def polarToCart( - magnitude: cv2.typing.MatLike, - angle: cv2.typing.MatLike, - x: cv2.typing.MatLike | None = ..., - y: cv2.typing.MatLike | None = ..., - angleInDegrees: bool = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def polarToCart( - magnitude: UMat, - angle: UMat, - x: UMat | None = ..., - y: UMat | None = ..., - angleInDegrees: bool = ..., -) -> tuple[ - UMat, - UMat, -]: ... -def pollKey() -> int: ... -@typing.overload -def polylines( - img: cv2.typing.MatLike, - pts: typing.Sequence[cv2.typing.MatLike], - isClosed: bool, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - shift: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def polylines( - img: UMat, - pts: typing.Sequence[UMat], - isClosed: bool, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - shift: int = ..., -) -> UMat: ... -@typing.overload -def pow(src: cv2.typing.MatLike, power: float, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def pow(src: UMat, power: float, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def preCornerDetect( - src: cv2.typing.MatLike, - ksize: int, - dst: cv2.typing.MatLike | None = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def preCornerDetect(src: UMat, ksize: int, dst: UMat | None = ..., borderType: int = ...) -> UMat: ... -@typing.overload -def projectPoints( - objectPoints: cv2.typing.MatLike, - rvec: cv2.typing.MatLike, - tvec: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - imagePoints: cv2.typing.MatLike | None = ..., - jacobian: cv2.typing.MatLike | None = ..., - aspectRatio: float = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def projectPoints( - objectPoints: UMat, - rvec: UMat, - tvec: UMat, - cameraMatrix: UMat, - distCoeffs: UMat, - imagePoints: UMat | None = ..., - jacobian: UMat | None = ..., - aspectRatio: float = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def putText( - img: cv2.typing.MatLike, - text: str, - org: cv2.typing.Point, - fontFace: int, - fontScale: float, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - bottomLeftOrigin: bool = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def putText( - img: UMat, - text: str, - org: cv2.typing.Point, - fontFace: int, - fontScale: float, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - bottomLeftOrigin: bool = ..., -) -> UMat: ... -@typing.overload -def pyrDown( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - dstsize: cv2.typing.Size = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def pyrDown(src: UMat, dst: UMat | None = ..., dstsize: cv2.typing.Size = ..., borderType: int = ...) -> UMat: ... -@typing.overload -def pyrMeanShiftFiltering( - src: cv2.typing.MatLike, - sp: float, - sr: float, - dst: cv2.typing.MatLike | None = ..., - maxLevel: int = ..., - termcrit: cv2.typing.TermCriteria = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def pyrMeanShiftFiltering( - src: UMat, - sp: float, - sr: float, - dst: UMat | None = ..., - maxLevel: int = ..., - termcrit: cv2.typing.TermCriteria = ..., -) -> UMat: ... -@typing.overload -def pyrUp( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - dstsize: cv2.typing.Size = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def pyrUp(src: UMat, dst: UMat | None = ..., dstsize: cv2.typing.Size = ..., borderType: int = ...) -> UMat: ... -@typing.overload -def randShuffle(dst: cv2.typing.MatLike, iterFactor: float = ...) -> cv2.typing.MatLike: ... -@typing.overload -def randShuffle(dst: UMat, iterFactor: float = ...) -> UMat: ... -@typing.overload -def randn(dst: cv2.typing.MatLike, mean: cv2.typing.MatLike, stddev: cv2.typing.MatLike) -> cv2.typing.MatLike: ... -@typing.overload -def randn(dst: UMat, mean: UMat, stddev: UMat) -> UMat: ... -@typing.overload -def randu(dst: cv2.typing.MatLike, low: cv2.typing.MatLike, high: cv2.typing.MatLike) -> cv2.typing.MatLike: ... -@typing.overload -def randu(dst: UMat, low: UMat, high: UMat) -> UMat: ... -def readOpticalFlow(path: str) -> cv2.typing.MatLike: ... -@typing.overload -def recoverPose( - points1: cv2.typing.MatLike, - points2: cv2.typing.MatLike, - cameraMatrix1: cv2.typing.MatLike, - distCoeffs1: cv2.typing.MatLike, - cameraMatrix2: cv2.typing.MatLike, - distCoeffs2: cv2.typing.MatLike, - E: cv2.typing.MatLike | None = ..., - R: cv2.typing.MatLike | None = ..., - t: cv2.typing.MatLike | None = ..., - method: int = ..., - prob: float = ..., - threshold: float = ..., - mask: cv2.typing.MatLike | None = ..., -) -> tuple[ - int, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def recoverPose( - points1: UMat, - points2: UMat, - cameraMatrix1: UMat, - distCoeffs1: UMat, - cameraMatrix2: UMat, - distCoeffs2: UMat, - E: UMat | None = ..., - R: UMat | None = ..., - t: UMat | None = ..., - method: int = ..., - prob: float = ..., - threshold: float = ..., - mask: UMat | None = ..., -) -> tuple[ - int, - UMat, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def recoverPose( - E: cv2.typing.MatLike, - points1: cv2.typing.MatLike, - points2: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - R: cv2.typing.MatLike | None = ..., - t: cv2.typing.MatLike | None = ..., - mask: cv2.typing.MatLike | None = ..., -) -> tuple[ - int, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def recoverPose( - E: UMat, - points1: UMat, - points2: UMat, - cameraMatrix: UMat, - R: UMat | None = ..., - t: UMat | None = ..., - mask: UMat | None = ..., -) -> tuple[ - int, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def recoverPose( - E: cv2.typing.MatLike, - points1: cv2.typing.MatLike, - points2: cv2.typing.MatLike, - R: cv2.typing.MatLike | None = ..., - t: cv2.typing.MatLike | None = ..., - focal: float = ..., - pp: cv2.typing.Point2d = ..., - mask: cv2.typing.MatLike | None = ..., -) -> tuple[ - int, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def recoverPose( - E: UMat, - points1: UMat, - points2: UMat, - R: UMat | None = ..., - t: UMat | None = ..., - focal: float = ..., - pp: cv2.typing.Point2d = ..., - mask: UMat | None = ..., -) -> tuple[ - int, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def recoverPose( - E: cv2.typing.MatLike, - points1: cv2.typing.MatLike, - points2: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distanceThresh: float, - R: cv2.typing.MatLike | None = ..., - t: cv2.typing.MatLike | None = ..., - mask: cv2.typing.MatLike | None = ..., - triangulatedPoints: cv2.typing.MatLike | None = ..., -) -> tuple[ - int, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def recoverPose( - E: UMat, - points1: UMat, - points2: UMat, - cameraMatrix: UMat, - distanceThresh: float, - R: UMat | None = ..., - t: UMat | None = ..., - mask: UMat | None = ..., - triangulatedPoints: UMat | None = ..., -) -> tuple[ - int, - UMat, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def rectangle( - img: cv2.typing.MatLike, - pt1: cv2.typing.Point, - pt2: cv2.typing.Point, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - shift: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def rectangle( - img: UMat, - pt1: cv2.typing.Point, - pt2: cv2.typing.Point, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - shift: int = ..., -) -> UMat: ... -@typing.overload -def rectangle( - img: cv2.typing.MatLike, - rec: cv2.typing.Rect, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - shift: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def rectangle( - img: UMat, - rec: cv2.typing.Rect, - color: cv2.typing.Scalar, - thickness: int = ..., - lineType: int = ..., - shift: int = ..., -) -> UMat: ... -def rectangleIntersectionArea(a: cv2.typing.Rect2d, b: cv2.typing.Rect2d) -> float: ... -@typing.overload -def rectify3Collinear( - cameraMatrix1: cv2.typing.MatLike, - distCoeffs1: cv2.typing.MatLike, - cameraMatrix2: cv2.typing.MatLike, - distCoeffs2: cv2.typing.MatLike, - cameraMatrix3: cv2.typing.MatLike, - distCoeffs3: cv2.typing.MatLike, - imgpt1: typing.Sequence[cv2.typing.MatLike], - imgpt3: typing.Sequence[cv2.typing.MatLike], - imageSize: cv2.typing.Size, - R12: cv2.typing.MatLike, - T12: cv2.typing.MatLike, - R13: cv2.typing.MatLike, - T13: cv2.typing.MatLike, - alpha: float, - newImgSize: cv2.typing.Size, - flags: int, - R1: cv2.typing.MatLike | None = ..., - R2: cv2.typing.MatLike | None = ..., - R3: cv2.typing.MatLike | None = ..., - P1: cv2.typing.MatLike | None = ..., - P2: cv2.typing.MatLike | None = ..., - P3: cv2.typing.MatLike | None = ..., - Q: cv2.typing.MatLike | None = ..., -) -> tuple[ - float, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.Rect, - cv2.typing.Rect, -]: ... -@typing.overload -def rectify3Collinear( - cameraMatrix1: UMat, - distCoeffs1: UMat, - cameraMatrix2: UMat, - distCoeffs2: UMat, - cameraMatrix3: UMat, - distCoeffs3: UMat, - imgpt1: typing.Sequence[UMat], - imgpt3: typing.Sequence[UMat], - imageSize: cv2.typing.Size, - R12: UMat, - T12: UMat, - R13: UMat, - T13: UMat, - alpha: float, - newImgSize: cv2.typing.Size, - flags: int, - R1: UMat | None = ..., - R2: UMat | None = ..., - R3: UMat | None = ..., - P1: UMat | None = ..., - P2: UMat | None = ..., - P3: UMat | None = ..., - Q: UMat | None = ..., -) -> tuple[ - float, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - cv2.typing.Rect, - cv2.typing.Rect, -]: ... -@typing.overload -def reduce( - src: cv2.typing.MatLike, - dim: int, - rtype: int, - dst: cv2.typing.MatLike | None = ..., - dtype: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def reduce(src: UMat, dim: int, rtype: int, dst: UMat | None = ..., dtype: int = ...) -> UMat: ... -@typing.overload -def reduceArgMax( - src: cv2.typing.MatLike, - axis: int, - dst: cv2.typing.MatLike | None = ..., - lastIndex: bool = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def reduceArgMax(src: UMat, axis: int, dst: UMat | None = ..., lastIndex: bool = ...) -> UMat: ... -@typing.overload -def reduceArgMin( - src: cv2.typing.MatLike, - axis: int, - dst: cv2.typing.MatLike | None = ..., - lastIndex: bool = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def reduceArgMin(src: UMat, axis: int, dst: UMat | None = ..., lastIndex: bool = ...) -> UMat: ... -@typing.overload -def remap( - src: cv2.typing.MatLike, - map1: cv2.typing.MatLike, - map2: cv2.typing.MatLike, - interpolation: int, - dst: cv2.typing.MatLike | None = ..., - borderMode: int = ..., - borderValue: cv2.typing.Scalar = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def remap( - src: UMat, - map1: UMat, - map2: UMat, - interpolation: int, - dst: UMat | None = ..., - borderMode: int = ..., - borderValue: cv2.typing.Scalar = ..., -) -> UMat: ... -@typing.overload -def repeat(src: cv2.typing.MatLike, ny: int, nx: int, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def repeat(src: UMat, ny: int, nx: int, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def reprojectImageTo3D( - disparity: cv2.typing.MatLike, - Q: cv2.typing.MatLike, - _3dImage: cv2.typing.MatLike | None = ..., - handleMissingValues: bool = ..., - ddepth: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def reprojectImageTo3D( - disparity: UMat, - Q: UMat, - _3dImage: UMat | None = ..., - handleMissingValues: bool = ..., - ddepth: int = ..., -) -> UMat: ... -@typing.overload -def resize( - src: cv2.typing.MatLike, - dsize: cv2.typing.Size | None, - dst: cv2.typing.MatLike | None = ..., - fx: float = ..., - fy: float = ..., - interpolation: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def resize( - src: UMat, - dsize: cv2.typing.Size | None, - dst: UMat | None = ..., - fx: float = ..., - fy: float = ..., - interpolation: int = ..., -) -> UMat: ... -@typing.overload -def resizeWindow(winname: str, width: int, height: int) -> None: ... -@typing.overload -def resizeWindow(winname: str, size: cv2.typing.Size) -> None: ... -@typing.overload -def rotate(src: cv2.typing.MatLike, rotateCode: int, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def rotate(src: UMat, rotateCode: int, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def rotatedRectangleIntersection( - rect1: cv2.typing.RotatedRect, - rect2: cv2.typing.RotatedRect, - intersectingRegion: cv2.typing.MatLike | None = ..., -) -> tuple[ - int, - cv2.typing.MatLike, -]: ... -@typing.overload -def rotatedRectangleIntersection( - rect1: cv2.typing.RotatedRect, - rect2: cv2.typing.RotatedRect, - intersectingRegion: UMat | None = ..., -) -> tuple[ - int, - UMat, -]: ... -@typing.overload -def sampsonDistance(pt1: cv2.typing.MatLike, pt2: cv2.typing.MatLike, F: cv2.typing.MatLike) -> float: ... -@typing.overload -def sampsonDistance(pt1: UMat, pt2: UMat, F: UMat) -> float: ... -@typing.overload -def scaleAdd( - src1: cv2.typing.MatLike, - alpha: float, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def scaleAdd(src1: UMat, alpha: float, src2: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def seamlessClone( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike, - mask: cv2.typing.MatLike, - p: cv2.typing.Point, - flags: int, - blend: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def seamlessClone( - src: UMat, - dst: UMat, - mask: UMat, - p: cv2.typing.Point, - flags: int, - blend: UMat | None = ..., -) -> UMat: ... -@typing.overload -def selectROI( - windowName: str, - img: cv2.typing.MatLike, - showCrosshair: bool = ..., - fromCenter: bool = ..., - printNotice: bool = ..., -) -> cv2.typing.Rect: ... -@typing.overload -def selectROI( - windowName: str, - img: UMat, - showCrosshair: bool = ..., - fromCenter: bool = ..., - printNotice: bool = ..., -) -> cv2.typing.Rect: ... -@typing.overload -def selectROI( - img: cv2.typing.MatLike, - showCrosshair: bool = ..., - fromCenter: bool = ..., - printNotice: bool = ..., -) -> cv2.typing.Rect: ... -@typing.overload -def selectROI( - img: UMat, - showCrosshair: bool = ..., - fromCenter: bool = ..., - printNotice: bool = ..., -) -> cv2.typing.Rect: ... -@typing.overload -def selectROIs( - windowName: str, - img: cv2.typing.MatLike, - showCrosshair: bool = ..., - fromCenter: bool = ..., - printNotice: bool = ..., -) -> typing.Sequence[cv2.typing.Rect]: ... -@typing.overload -def selectROIs( - windowName: str, - img: UMat, - showCrosshair: bool = ..., - fromCenter: bool = ..., - printNotice: bool = ..., -) -> typing.Sequence[cv2.typing.Rect]: ... -@typing.overload -def sepFilter2D( - src: cv2.typing.MatLike, - ddepth: int, - kernelX: cv2.typing.MatLike, - kernelY: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - anchor: cv2.typing.Point = ..., - delta: float = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def sepFilter2D( - src: UMat, - ddepth: int, - kernelX: UMat, - kernelY: UMat, - dst: UMat | None = ..., - anchor: cv2.typing.Point = ..., - delta: float = ..., - borderType: int = ..., -) -> UMat: ... -@typing.overload -def setIdentity(mtx: cv2.typing.MatLike, s: cv2.typing.Scalar = ...) -> cv2.typing.MatLike: ... -@typing.overload -def setIdentity(mtx: UMat, s: cv2.typing.Scalar = ...) -> UMat: ... -def setLogLevel(level: int) -> int: ... -def setNumThreads(nthreads: int) -> None: ... -def setRNGSeed(seed: int) -> None: ... -def setTrackbarMax(trackbarname: str, winname: str, maxval: int) -> None: ... -def setTrackbarMin(trackbarname: str, winname: str, minval: int) -> None: ... -def setTrackbarPos(trackbarname: str, winname: str, pos: int) -> None: ... -def setUseOpenVX(flag: bool) -> None: ... -def setUseOptimized(onoff: bool) -> None: ... -def setWindowProperty(winname: str, prop_id: int, prop_value: float) -> None: ... -def setWindowTitle(winname: str, title: str) -> None: ... -@typing.overload -def solve( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - flags: int = ..., -) -> tuple[ - bool, - cv2.typing.MatLike, -]: ... -@typing.overload -def solve(src1: UMat, src2: UMat, dst: UMat | None = ..., flags: int = ...) -> tuple[bool, UMat]: ... -@typing.overload -def solveCubic( - coeffs: cv2.typing.MatLike, - roots: cv2.typing.MatLike | None = ..., -) -> tuple[ - int, - cv2.typing.MatLike, -]: ... -@typing.overload -def solveCubic(coeffs: UMat, roots: UMat | None = ...) -> tuple[int, UMat]: ... -@typing.overload -def solveLP( - Func: cv2.typing.MatLike, - Constr: cv2.typing.MatLike, - constr_eps: float, - z: cv2.typing.MatLike | None = ..., -) -> tuple[ - int, - cv2.typing.MatLike, -]: ... -@typing.overload -def solveLP(Func: UMat, Constr: UMat, constr_eps: float, z: UMat | None = ...) -> tuple[int, UMat]: ... -@typing.overload -def solveLP( - Func: cv2.typing.MatLike, - Constr: cv2.typing.MatLike, - z: cv2.typing.MatLike | None = ..., -) -> tuple[ - int, - cv2.typing.MatLike, -]: ... -@typing.overload -def solveLP(Func: UMat, Constr: UMat, z: UMat | None = ...) -> tuple[int, UMat]: ... -@typing.overload -def solveP3P( - objectPoints: cv2.typing.MatLike, - imagePoints: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - flags: int, - rvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., - tvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., -) -> tuple[ - int, - typing.Sequence[cv2.typing.MatLike], - typing.Sequence[cv2.typing.MatLike], -]: ... -@typing.overload -def solveP3P( - objectPoints: UMat, - imagePoints: UMat, - cameraMatrix: UMat, - distCoeffs: UMat, - flags: int, - rvecs: typing.Sequence[UMat] | None = ..., - tvecs: typing.Sequence[UMat] | None = ..., -) -> tuple[ - int, - typing.Sequence[UMat], - typing.Sequence[UMat], -]: ... -@typing.overload -def solvePnP( - objectPoints: cv2.typing.MatLike, - imagePoints: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - rvec: cv2.typing.MatLike | None = ..., - tvec: cv2.typing.MatLike | None = ..., - useExtrinsicGuess: bool = ..., - flags: int = ..., -) -> tuple[ - bool, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def solvePnP( - objectPoints: UMat, - imagePoints: UMat, - cameraMatrix: UMat, - distCoeffs: UMat, - rvec: UMat | None = ..., - tvec: UMat | None = ..., - useExtrinsicGuess: bool = ..., - flags: int = ..., -) -> tuple[ - bool, - UMat, - UMat, -]: ... -@typing.overload -def solvePnPGeneric( - objectPoints: cv2.typing.MatLike, - imagePoints: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - rvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., - tvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., - useExtrinsicGuess: bool = ..., - flags: SolvePnPMethod = ..., - rvec: cv2.typing.MatLike | None = ..., - tvec: cv2.typing.MatLike | None = ..., - reprojectionError: cv2.typing.MatLike | None = ..., -) -> tuple[ - int, - typing.Sequence[cv2.typing.MatLike], - typing.Sequence[cv2.typing.MatLike], - cv2.typing.MatLike, -]: ... -@typing.overload -def solvePnPGeneric( - objectPoints: UMat, - imagePoints: UMat, - cameraMatrix: UMat, - distCoeffs: UMat, - rvecs: typing.Sequence[UMat] | None = ..., - tvecs: typing.Sequence[UMat] | None = ..., - useExtrinsicGuess: bool = ..., - flags: SolvePnPMethod = ..., - rvec: UMat | None = ..., - tvec: UMat | None = ..., - reprojectionError: UMat | None = ..., -) -> tuple[ - int, - typing.Sequence[UMat], - typing.Sequence[UMat], - UMat, -]: ... -@typing.overload -def solvePnPRansac( - objectPoints: cv2.typing.MatLike, - imagePoints: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - rvec: cv2.typing.MatLike | None = ..., - tvec: cv2.typing.MatLike | None = ..., - useExtrinsicGuess: bool = ..., - iterationsCount: int = ..., - reprojectionError: float = ..., - confidence: float = ..., - inliers: cv2.typing.MatLike | None = ..., - flags: int = ..., -) -> tuple[ - bool, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def solvePnPRansac( - objectPoints: UMat, - imagePoints: UMat, - cameraMatrix: UMat, - distCoeffs: UMat, - rvec: UMat | None = ..., - tvec: UMat | None = ..., - useExtrinsicGuess: bool = ..., - iterationsCount: int = ..., - reprojectionError: float = ..., - confidence: float = ..., - inliers: UMat | None = ..., - flags: int = ..., -) -> tuple[ - bool, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def solvePnPRansac( - objectPoints: cv2.typing.MatLike, - imagePoints: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - rvec: cv2.typing.MatLike | None = ..., - tvec: cv2.typing.MatLike | None = ..., - inliers: cv2.typing.MatLike | None = ..., - params: UsacParams = ..., -) -> tuple[ - bool, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def solvePnPRansac( - objectPoints: UMat, - imagePoints: UMat, - cameraMatrix: UMat, - distCoeffs: UMat, - rvec: UMat | None = ..., - tvec: UMat | None = ..., - inliers: UMat | None = ..., - params: UsacParams = ..., -) -> tuple[ - bool, - UMat, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def solvePnPRefineLM( - objectPoints: cv2.typing.MatLike, - imagePoints: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - rvec: cv2.typing.MatLike, - tvec: cv2.typing.MatLike, - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def solvePnPRefineLM( - objectPoints: UMat, - imagePoints: UMat, - cameraMatrix: UMat, - distCoeffs: UMat, - rvec: UMat, - tvec: UMat, - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def solvePnPRefineVVS( - objectPoints: cv2.typing.MatLike, - imagePoints: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - rvec: cv2.typing.MatLike, - tvec: cv2.typing.MatLike, - criteria: cv2.typing.TermCriteria = ..., - VVSlambda: float = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def solvePnPRefineVVS( - objectPoints: UMat, - imagePoints: UMat, - cameraMatrix: UMat, - distCoeffs: UMat, - rvec: UMat, - tvec: UMat, - criteria: cv2.typing.TermCriteria = ..., - VVSlambda: float = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def solvePoly( - coeffs: cv2.typing.MatLike, - roots: cv2.typing.MatLike | None = ..., - maxIters: int = ..., -) -> tuple[ - float, - cv2.typing.MatLike, -]: ... -@typing.overload -def solvePoly(coeffs: UMat, roots: UMat | None = ..., maxIters: int = ...) -> tuple[float, UMat]: ... -@typing.overload -def sort(src: cv2.typing.MatLike, flags: int, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def sort(src: UMat, flags: int, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def sortIdx(src: cv2.typing.MatLike, flags: int, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def sortIdx(src: UMat, flags: int, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def spatialGradient( - src: cv2.typing.MatLike, - dx: cv2.typing.MatLike | None = ..., - dy: cv2.typing.MatLike | None = ..., - ksize: int = ..., - borderType: int = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def spatialGradient( - src: UMat, - dx: UMat | None = ..., - dy: UMat | None = ..., - ksize: int = ..., - borderType: int = ..., -) -> tuple[ - UMat, - UMat, -]: ... -@typing.overload -def split( - m: cv2.typing.MatLike, - mv: typing.Sequence[cv2.typing.MatLike] | None = ..., -) -> typing.Sequence[cv2.typing.MatLike]: ... -@typing.overload -def split(m: UMat, mv: typing.Sequence[UMat] | None = ...) -> typing.Sequence[UMat]: ... -@typing.overload -def sqrBoxFilter( - src: cv2.typing.MatLike, - ddepth: int, - ksize: cv2.typing.Size, - dst: cv2.typing.MatLike | None = ..., - anchor: cv2.typing.Point = ..., - normalize: bool = ..., - borderType: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def sqrBoxFilter( - src: UMat, - ddepth: int, - ksize: cv2.typing.Size, - dst: UMat | None = ..., - anchor: cv2.typing.Point = ..., - normalize: bool = ..., - borderType: int = ..., -) -> UMat: ... -@typing.overload -def sqrt(src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def sqrt(src: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def stackBlur( - src: cv2.typing.MatLike, - ksize: cv2.typing.Size, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def stackBlur(src: UMat, ksize: cv2.typing.Size, dst: UMat | None = ...) -> UMat: ... -def startWindowThread() -> int: ... -@typing.overload -def stereoCalibrate( - objectPoints: typing.Sequence[cv2.typing.MatLike], - imagePoints1: typing.Sequence[cv2.typing.MatLike], - imagePoints2: typing.Sequence[cv2.typing.MatLike], - cameraMatrix1: cv2.typing.MatLike, - distCoeffs1: cv2.typing.MatLike, - cameraMatrix2: cv2.typing.MatLike, - distCoeffs2: cv2.typing.MatLike, - imageSize: cv2.typing.Size, - R: cv2.typing.MatLike | None = ..., - T: cv2.typing.MatLike | None = ..., - E: cv2.typing.MatLike | None = ..., - F: cv2.typing.MatLike | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def stereoCalibrate( - objectPoints: typing.Sequence[UMat], - imagePoints1: typing.Sequence[UMat], - imagePoints2: typing.Sequence[UMat], - cameraMatrix1: UMat, - distCoeffs1: UMat, - cameraMatrix2: UMat, - distCoeffs2: UMat, - imageSize: cv2.typing.Size, - R: UMat | None = ..., - T: UMat | None = ..., - E: UMat | None = ..., - F: UMat | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def stereoCalibrate( - objectPoints: typing.Sequence[cv2.typing.MatLike], - imagePoints1: typing.Sequence[cv2.typing.MatLike], - imagePoints2: typing.Sequence[cv2.typing.MatLike], - cameraMatrix1: cv2.typing.MatLike, - distCoeffs1: cv2.typing.MatLike, - cameraMatrix2: cv2.typing.MatLike, - distCoeffs2: cv2.typing.MatLike, - imageSize: cv2.typing.Size, - R: cv2.typing.MatLike, - T: cv2.typing.MatLike, - E: cv2.typing.MatLike | None = ..., - F: cv2.typing.MatLike | None = ..., - perViewErrors: cv2.typing.MatLike | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def stereoCalibrate( - objectPoints: typing.Sequence[UMat], - imagePoints1: typing.Sequence[UMat], - imagePoints2: typing.Sequence[UMat], - cameraMatrix1: UMat, - distCoeffs1: UMat, - cameraMatrix2: UMat, - distCoeffs2: UMat, - imageSize: cv2.typing.Size, - R: UMat, - T: UMat, - E: UMat | None = ..., - F: UMat | None = ..., - perViewErrors: UMat | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, -]: ... -@typing.overload -def stereoCalibrateExtended( - objectPoints: typing.Sequence[cv2.typing.MatLike], - imagePoints1: typing.Sequence[cv2.typing.MatLike], - imagePoints2: typing.Sequence[cv2.typing.MatLike], - cameraMatrix1: cv2.typing.MatLike, - distCoeffs1: cv2.typing.MatLike, - cameraMatrix2: cv2.typing.MatLike, - distCoeffs2: cv2.typing.MatLike, - imageSize: cv2.typing.Size, - R: cv2.typing.MatLike, - T: cv2.typing.MatLike, - E: cv2.typing.MatLike | None = ..., - F: cv2.typing.MatLike | None = ..., - rvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., - tvecs: typing.Sequence[cv2.typing.MatLike] | None = ..., - perViewErrors: cv2.typing.MatLike | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - typing.Sequence[cv2.typing.MatLike], - typing.Sequence[cv2.typing.MatLike], - cv2.typing.MatLike, -]: ... -@typing.overload -def stereoCalibrateExtended( - objectPoints: typing.Sequence[UMat], - imagePoints1: typing.Sequence[UMat], - imagePoints2: typing.Sequence[UMat], - cameraMatrix1: UMat, - distCoeffs1: UMat, - cameraMatrix2: UMat, - distCoeffs2: UMat, - imageSize: cv2.typing.Size, - R: UMat, - T: UMat, - E: UMat | None = ..., - F: UMat | None = ..., - rvecs: typing.Sequence[UMat] | None = ..., - tvecs: typing.Sequence[UMat] | None = ..., - perViewErrors: UMat | None = ..., - flags: int = ..., - criteria: cv2.typing.TermCriteria = ..., -) -> tuple[ - float, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - UMat, - typing.Sequence[UMat], - typing.Sequence[UMat], - UMat, -]: ... -@typing.overload -def stereoRectify( - cameraMatrix1: cv2.typing.MatLike, - distCoeffs1: cv2.typing.MatLike, - cameraMatrix2: cv2.typing.MatLike, - distCoeffs2: cv2.typing.MatLike, - imageSize: cv2.typing.Size, - R: cv2.typing.MatLike, - T: cv2.typing.MatLike, - R1: cv2.typing.MatLike | None = ..., - R2: cv2.typing.MatLike | None = ..., - P1: cv2.typing.MatLike | None = ..., - P2: cv2.typing.MatLike | None = ..., - Q: cv2.typing.MatLike | None = ..., - flags: int = ..., - alpha: float = ..., - newImageSize: cv2.typing.Size = ..., -) -> tuple[ - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.MatLike, - cv2.typing.Rect, - cv2.typing.Rect, -]: ... -@typing.overload -def stereoRectify( - cameraMatrix1: UMat, - distCoeffs1: UMat, - cameraMatrix2: UMat, - distCoeffs2: UMat, - imageSize: cv2.typing.Size, - R: UMat, - T: UMat, - R1: UMat | None = ..., - R2: UMat | None = ..., - P1: UMat | None = ..., - P2: UMat | None = ..., - Q: UMat | None = ..., - flags: int = ..., - alpha: float = ..., - newImageSize: cv2.typing.Size = ..., -) -> tuple[ - UMat, - UMat, - UMat, - UMat, - UMat, - cv2.typing.Rect, - cv2.typing.Rect, -]: ... -@typing.overload -def stereoRectifyUncalibrated( - points1: cv2.typing.MatLike, - points2: cv2.typing.MatLike, - F: cv2.typing.MatLike, - imgSize: cv2.typing.Size, - H1: cv2.typing.MatLike | None = ..., - H2: cv2.typing.MatLike | None = ..., - threshold: float = ..., -) -> tuple[ - bool, - cv2.typing.MatLike, - cv2.typing.MatLike, -]: ... -@typing.overload -def stereoRectifyUncalibrated( - points1: UMat, - points2: UMat, - F: UMat, - imgSize: cv2.typing.Size, - H1: UMat | None = ..., - H2: UMat | None = ..., - threshold: float = ..., -) -> tuple[ - bool, - UMat, - UMat, -]: ... -@typing.overload -def stylization( - src: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - sigma_s: float = ..., - sigma_r: float = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def stylization(src: UMat, dst: UMat | None = ..., sigma_s: float = ..., sigma_r: float = ...) -> UMat: ... -@typing.overload -def subtract( - src1: cv2.typing.MatLike, - src2: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - mask: cv2.typing.MatLike | None = ..., - dtype: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def subtract(src1: UMat, src2: UMat, dst: UMat | None = ..., mask: UMat | None = ..., dtype: int = ...) -> UMat: ... -@typing.overload -def sumElems(src: cv2.typing.MatLike) -> cv2.typing.Scalar: ... -@typing.overload -def sumElems(src: UMat) -> cv2.typing.Scalar: ... -@typing.overload -def textureFlattening( - src: cv2.typing.MatLike, - mask: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - low_threshold: float = ..., - high_threshold: float = ..., - kernel_size: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def textureFlattening( - src: UMat, - mask: UMat, - dst: UMat | None = ..., - low_threshold: float = ..., - high_threshold: float = ..., - kernel_size: int = ..., -) -> UMat: ... -@typing.overload -def threshold( - src: cv2.typing.MatLike, - thresh: float, - maxval: float, - type: int, - dst: cv2.typing.MatLike | None = ..., -) -> tuple[ - float, - cv2.typing.MatLike, -]: ... -@typing.overload -def threshold(src: UMat, thresh: float, maxval: float, type: int, dst: UMat | None = ...) -> tuple[float, UMat]: ... -@typing.overload -def trace(mtx: cv2.typing.MatLike) -> cv2.typing.Scalar: ... -@typing.overload -def trace(mtx: UMat) -> cv2.typing.Scalar: ... -@typing.overload -def transform( - src: cv2.typing.MatLike, - m: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def transform(src: UMat, m: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def transpose(src: cv2.typing.MatLike, dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def transpose(src: UMat, dst: UMat | None = ...) -> UMat: ... -@typing.overload -def transposeND( - src: cv2.typing.MatLike, - order: typing.Sequence[int], - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def transposeND(src: UMat, order: typing.Sequence[int], dst: UMat | None = ...) -> UMat: ... -@typing.overload -def triangulatePoints( - projMatr1: cv2.typing.MatLike, - projMatr2: cv2.typing.MatLike, - projPoints1: cv2.typing.MatLike, - projPoints2: cv2.typing.MatLike, - points4D: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def triangulatePoints( - projMatr1: UMat, - projMatr2: UMat, - projPoints1: UMat, - projPoints2: UMat, - points4D: UMat | None = ..., -) -> UMat: ... -@typing.overload -def undistort( - src: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - newCameraMatrix: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def undistort( - src: UMat, - cameraMatrix: UMat, - distCoeffs: UMat, - dst: UMat | None = ..., - newCameraMatrix: UMat | None = ..., -) -> UMat: ... -@typing.overload -def undistortImagePoints( - src: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - arg1: cv2.typing.TermCriteria = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def undistortImagePoints( - src: UMat, - cameraMatrix: UMat, - distCoeffs: UMat, - dst: UMat | None = ..., - arg1: cv2.typing.TermCriteria = ..., -) -> UMat: ... -@typing.overload -def undistortPoints( - src: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - dst: cv2.typing.MatLike | None = ..., - R: cv2.typing.MatLike | None = ..., - P: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def undistortPoints( - src: UMat, - cameraMatrix: UMat, - distCoeffs: UMat, - dst: UMat | None = ..., - R: UMat | None = ..., - P: UMat | None = ..., -) -> UMat: ... -@typing.overload -def undistortPointsIter( - src: cv2.typing.MatLike, - cameraMatrix: cv2.typing.MatLike, - distCoeffs: cv2.typing.MatLike, - R: cv2.typing.MatLike, - P: cv2.typing.MatLike, - criteria: cv2.typing.TermCriteria, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def undistortPointsIter( - src: UMat, - cameraMatrix: UMat, - distCoeffs: UMat, - R: UMat, - P: UMat, - criteria: cv2.typing.TermCriteria, - dst: UMat | None = ..., -) -> UMat: ... -def useOpenVX() -> bool: ... -def useOptimized() -> bool: ... -@typing.overload -def validateDisparity( - disparity: cv2.typing.MatLike, - cost: cv2.typing.MatLike, - minDisparity: int, - numberOfDisparities: int, - disp12MaxDisp: int = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def validateDisparity( - disparity: UMat, - cost: UMat, - minDisparity: int, - numberOfDisparities: int, - disp12MaxDisp: int = ..., -) -> UMat: ... -@typing.overload -def vconcat(src: typing.Sequence[cv2.typing.MatLike], dst: cv2.typing.MatLike | None = ...) -> cv2.typing.MatLike: ... -@typing.overload -def vconcat(src: typing.Sequence[UMat], dst: UMat | None = ...) -> UMat: ... -def waitKey(delay: int = ...) -> int: ... -def waitKeyEx(delay: int = ...) -> int: ... -@typing.overload -def warpAffine( - src: cv2.typing.MatLike, - M: cv2.typing.MatLike, - dsize: cv2.typing.Size, - dst: cv2.typing.MatLike | None = ..., - flags: int = ..., - borderMode: int = ..., - borderValue: cv2.typing.Scalar = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def warpAffine( - src: UMat, - M: UMat, - dsize: cv2.typing.Size, - dst: UMat | None = ..., - flags: int = ..., - borderMode: int = ..., - borderValue: cv2.typing.Scalar = ..., -) -> UMat: ... -@typing.overload -def warpPerspective( - src: cv2.typing.MatLike, - M: cv2.typing.MatLike, - dsize: cv2.typing.Size, - dst: cv2.typing.MatLike | None = ..., - flags: int = ..., - borderMode: int = ..., - borderValue: cv2.typing.Scalar = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def warpPerspective( - src: UMat, - M: UMat, - dsize: cv2.typing.Size, - dst: UMat | None = ..., - flags: int = ..., - borderMode: int = ..., - borderValue: cv2.typing.Scalar = ..., -) -> UMat: ... -@typing.overload -def warpPolar( - src: cv2.typing.MatLike, - dsize: cv2.typing.Size, - center: cv2.typing.Point2f, - maxRadius: float, - flags: int, - dst: cv2.typing.MatLike | None = ..., -) -> cv2.typing.MatLike: ... -@typing.overload -def warpPolar( - src: UMat, - dsize: cv2.typing.Size, - center: cv2.typing.Point2f, - maxRadius: float, - flags: int, - dst: UMat | None = ..., -) -> UMat: ... -@typing.overload -def watershed(image: cv2.typing.MatLike, markers: cv2.typing.MatLike) -> cv2.typing.MatLike: ... -@typing.overload -def watershed(image: UMat, markers: UMat) -> UMat: ... -@typing.overload -def writeOpticalFlow(path: str, flow: cv2.typing.MatLike) -> bool: ... -@typing.overload -def writeOpticalFlow(path: str, flow: UMat) -> bool: ... diff --git a/typings/cv2/mat_wrapper/__init__.pyi b/typings/cv2/mat_wrapper/__init__.pyi deleted file mode 100644 index 0bfcd316..00000000 --- a/typings/cv2/mat_wrapper/__init__.pyi +++ /dev/null @@ -1,14 +0,0 @@ -from typing import TypeAlias - -import numpy as np -from _typeshed import Unused - -__all__: list[str] = [] -_NDArray: TypeAlias = np.ndarray[float, np.dtype[np.generic]] - -class Mat(_NDArray): - wrap_channels: bool | None - - def __new__(cls, arr: _NDArray, wrap_channels: bool = ..., **kwargs: Unused) -> _NDArray: ... - def __init__(self, arr: _NDArray, wrap_channels: bool = ...) -> None: ... - def __array_finalize__(self, obj: _NDArray | None) -> None: ... diff --git a/typings/scipy/__init__.pyi b/typings/scipy/__init__.pyi index a684150f..8e85a093 100644 --- a/typings/scipy/__init__.pyi +++ b/typings/scipy/__init__.pyi @@ -27,6 +27,8 @@ from scipy._lib._testutils import PytestTester from scipy.version import version as __version__ __all__ = [ + "LowLevelCallable", + "__version__", "cluster", "constants", "datasets", @@ -40,15 +42,13 @@ __all__ = [ "ndimage", "odr", "optimize", + "show_config", "signal", "sparse", "spatial", "special", "stats", - "LowLevelCallable", "test", - "show_config", - "__version__", ] test: PytestTester diff --git a/typings/scipy/fft/__init__.pyi b/typings/scipy/fft/__init__.pyi index ea8b8f3d..d27753fe 100644 --- a/typings/scipy/fft/__init__.pyi +++ b/typings/scipy/fft/__init__.pyi @@ -28,45 +28,45 @@ from scipy.fft._pocketfft.helper import get_workers, set_workers from scipy.fft._realtransforms import dct, dctn, dst, dstn, idct, idctn, idst, idstn __all__ = [ + "dct", + "dctn", + "dst", + "dstn", "fft", - "ifft", "fft2", - "ifft2", + "fftfreq", "fftn", - "ifftn", - "rfft", - "irfft", - "rfft2", - "irfft2", - "rfftn", - "irfftn", + "fftshift", + "fht", + "fhtoffset", + "get_workers", "hfft", - "ihfft", "hfft2", - "ihfft2", "hfftn", - "ihfftn", - "fftfreq", - "rfftfreq", - "fftshift", - "ifftshift", - "next_fast_len", - "dct", "idct", - "dst", - "idst", - "dctn", "idctn", - "dstn", + "idst", "idstn", - "fht", + "ifft", + "ifft2", + "ifftn", + "ifftshift", "ifht", - "fhtoffset", + "ihfft", + "ihfft2", + "ihfftn", + "irfft", + "irfft2", + "irfftn", + "next_fast_len", + "register_backend", + "rfft", + "rfft2", + "rfftfreq", + "rfftn", "set_backend", - "skip_backend", "set_global_backend", - "register_backend", - "get_workers", "set_workers", + "skip_backend", ] test: PytestTester diff --git a/typings/scipy/fft/_realtransforms.pyi b/typings/scipy/fft/_realtransforms.pyi index bbddf386..fd68b022 100644 --- a/typings/scipy/fft/_realtransforms.pyi +++ b/typings/scipy/fft/_realtransforms.pyi @@ -2,8 +2,7 @@ from _typeshed import Incomplete from numpy import float64, generic from numpy.typing import NDArray -__all__ = ["dct", "idct", "dst", "idst", "dctn", "idctn", "dstn", "idstn"] - +__all__ = ["dct", "dctn", "dst", "dstn", "idct", "idctn", "idst", "idstn"] def dctn( x, @@ -17,7 +16,6 @@ def dctn( orthogonalize=None, ): ... - def idctn( x, type=2, @@ -29,7 +27,6 @@ def idctn( orthogonalize=None, ): ... - def dstn( x, type=2, @@ -41,7 +38,6 @@ def dstn( orthogonalize=None, ): ... - def idstn( x, type=2, @@ -53,7 +49,6 @@ def idstn( orthogonalize=None, ): ... - def dct( x: NDArray[generic], type: int = 2, @@ -65,7 +60,6 @@ def dct( orthogonalize: Incomplete | None = None, ) -> NDArray[float64]: ... - def idct( x, type=2, @@ -77,7 +71,6 @@ def idct( orthogonalize=None, ): ... - def dst( x, type=2, @@ -89,7 +82,6 @@ def dst( orthogonalize=None, ): ... - def idst( x, type=2,