-
Notifications
You must be signed in to change notification settings - Fork 861
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Environment installation problem #481
Comments
I got the same error problem. Looking for a fix |
I guess all the issues arise from the OpenMM Lab (MMCV and MMseg) dependencies. I handle a lot of installation because I am an applied Deep Learning engineer and spend a lot of time coding in deep learning. And I am stuck for 9 hours now (yup) trying to go through numerous permutations of PyTorch, MMCV, and MMsegmentation. The easiest way (which I think will work) is to install CUDA toolkit 11.7 and then run the conda install from the project. However, I believe hardly anyone will have global CUDA 11.7 in 2024. I have CUDA 12.x globally enabled and cannot risk messing up the system with multiple global CUDA versions. |
Just want to mention that I also have this problem. It seems that there is a conda conflict with Full conda error
I can replicate this error with this minimal environment file: name: dinov2
channels:
- defaults
- pytorch
- nvidia
- xformers
- conda-forge
dependencies:
- pytorch::pytorch=2.0.0
- xformers::xformers=0.0.18 Conda error for the minimal environment file.
The main README.md says that it is only tested on these two versions, so, maybe these versions have changed underneath this repo? |
I found a solution that seems to work. I initially had CUDA 12.4 installed globally on my system. This project requires CUDA 11.7, however, I was unwilling the uninstall the current version and a bit hesitant to install two CUDA versions. Still, I went ahead and installed CUDA 11.7 through local runfile steps. The image below shows the settings. After installation and adding the paths to bashrc, everything seems to work. |
It's a conflict between But I tried it anyway, just to make sure. It didn't change anything for me.# min.yaml
name: dinov2
channels:
- nvidia
- pytorch
- xformers
- conda-forge
dependencies:
- pytorch::pytorch=2.0.0
- pytorch::pytorch-cuda=11.7.*
- xformers::xformers=0.0.18 FROM nvidia/cuda:11.7.1-devel-ubuntu22.04
RUN apt-get update && apt-get install -y curl gcc
RUN cd /root && curl -o ./install.sh -L micro.mamba.pm/install.sh && chmod +x ./install.sh && bash ./install.sh
ADD min.yaml /root
So, I'm going to try using |
@Multihuntr |
I encountered the same issue on 5 RTX 4090 GPUs. I commented out the xformer part in the config file to set up the environment, and then installed the requirements using pip. Training and testing can be performed successfully, but both can only run on a single GPU, and multi-GPU training and testing are not working. I haven’t figured out the reason yet. |
I want to install the environment on 8*4090 machine fellow the instruction in readme.md
Could not solve for environment specs The following packages are incompatible ├─ pytorch-cuda 11.7.0* is requested and can be installed; ├─ pytorch 2.0.0* is installable with the potential options │ ├─ pytorch 2.0.0 would require │ │ └─ pytorch-mutex 1.0 cpu, which can be installed; │ ├─ pytorch 2.0.0 would require │ │ └─ pytorch-mutex 1.0 cuda, which conflicts with any installable versions previously reported; │ └─ pytorch 2.0.0 would require │ ├─ pytorch-cuda >=11.8,<11.9 , which conflicts with any installable versions previously reported; │ └─ pytorch-mutex 1.0 cuda, which conflicts with any installable versions previously reported; ├─ torchvision 0.15.0* is installable with the potential options │ ├─ torchvision 0.15.0 would require │ │ └─ pytorch-mutex 1.0 cpu, which can be installed; │ ├─ torchvision 0.15.0 would require │ │ ├─ pytorch-cuda 11.7.* , which can be installed; │ │ └─ pytorch-mutex 1.0 cuda, which conflicts with any installable versions previously reported; │ └─ torchvision 0.15.0 would require │ └─ pytorch-cuda 11.8.* , which conflicts with any installable versions previously reported; └─ xformers 0.0.18* is not installable because there are no viable options ├─ xformers 0.0.18 would require │ └─ pytorch 1.12.1.* but there are no viable options │ ├─ pytorch 1.12.1 conflicts with any installable versions previously reported; │ ├─ pytorch [1.12.1|1.13.1] would require │ │ └─ pytorch-mutex 1.0 cpu, which can be installed; │ └─ pytorch [1.12.1|1.13.1] would require │ └─ pytorch-mutex 1.0 cuda, which conflicts with any installable versions previously reported; └─ xformers 0.0.18 would require └─ pytorch 1.13.1.* but there are no viable options ├─ pytorch [1.12.1|1.13.1], which cannot be installed (as previously explained); ├─ pytorch [1.12.1|1.13.1], which cannot be installed (as previously explained); ├─ pytorch 1.13.1 conflicts with any installable versions previously reported; └─ pytorch 1.13.1 would require ├─ pytorch-cuda >=11.6,<11.7 , which conflicts with any installable versions previously reported; └─ pytorch-mutex 1.0 cuda, which conflicts with any installable versions previously reported.
The text was updated successfully, but these errors were encountered: