Skip to content

Commit

Permalink
[pre-commit.ci] pre-commit autoupdate (#323)
Browse files Browse the repository at this point in the history
* [pre-commit.ci] pre-commit autoupdate

updates:
- [github.com/psf/black: 22.12.0 → 23.1.0](psf/black@22.12.0...23.1.0)

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
  • Loading branch information
pre-commit-ci[bot] authored Feb 6, 2023
1 parent 92ec35f commit 1aeb6e9
Show file tree
Hide file tree
Showing 4 changed files with 1 addition and 5 deletions.
2 changes: 1 addition & 1 deletion .pre-commit-config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ repos:
additional_dependencies:
- black==22.6.0 # Matches hook
- repo: https://github.com/psf/black
rev: 22.12.0
rev: 23.1.0
hooks:
- id: black
args: [--line-length, "100"]
Expand Down
2 changes: 0 additions & 2 deletions solo/backbones/poolformer/poolformer.py
Original file line number Diff line number Diff line change
Expand Up @@ -182,7 +182,6 @@ def __init__(
use_layer_scale=True,
layer_scale_init_value=1e-5,
):

super().__init__()

self.norm1 = norm_layer(dim)
Expand Down Expand Up @@ -299,7 +298,6 @@ def __init__(
pretrained=None,
**kwargs,
):

super().__init__()

if not fork_feat:
Expand Down
1 change: 0 additions & 1 deletion solo/data/dali_dataloader.py
Original file line number Diff line number Diff line change
Expand Up @@ -718,7 +718,6 @@ def __init__(
dali_device: str = "gpu",
encode_indexes_into_labels: bool = False,
):

"""DataModule for pretrain data using Nvidia DALI.
Args:
Expand Down
1 change: 0 additions & 1 deletion solo/utils/kmeans.py
Original file line number Diff line number Diff line change
Expand Up @@ -98,7 +98,6 @@ def cluster_memory(
dist.broadcast(centroids, 0)

for n_iter in range(self.kmeans_iters + 1):

# E step
dot_products = torch.mm(local_memory_embeddings[j], centroids.t())
_, local_assignments = dot_products.max(dim=1)
Expand Down

0 comments on commit 1aeb6e9

Please sign in to comment.