|
## Installation |
|
|
|
Our [Colab Notebook](https://colab.research.google.com/drive/16jcaJoc6bCFAQ96jDe2HwtXj7BMD_-m5) |
|
has step-by-step instructions that install detectron2. |
|
The [Dockerfile](docker) |
|
also installs detectron2 with a few simple commands. |
|
|
|
### Requirements |
|
- Linux or macOS with Python β₯ 3.6 |
|
- PyTorch β₯ 1.4 |
|
- [torchvision](https://github.com/pytorch/vision/) that matches the PyTorch installation. |
|
You can install them together at [pytorch.org](https://pytorch.org) to make sure of this. |
|
- OpenCV, optional, needed by demo and visualization |
|
- pycocotools: `pip install cython; pip install -U 'git+https://github.com/cocodataset/cocoapi.git#subdirectory=PythonAPI'` |
|
|
|
|
|
### Build Detectron2 from Source |
|
|
|
gcc & g++ β₯ 5 are required. [ninja](https://ninja-build.org/) is recommended for faster build. |
|
After having them, run: |
|
``` |
|
python -m pip install 'git+https://github.com/facebookresearch/detectron2.git' |
|
# (add --user if you don't have permission) |
|
|
|
# Or, to install it from a local clone: |
|
git clone https://github.com/facebookresearch/detectron2.git |
|
python -m pip install -e detectron2 |
|
|
|
# Or if you are on macOS |
|
# CC=clang CXX=clang++ python -m pip install -e . |
|
``` |
|
|
|
To __rebuild__ detectron2 that's built from a local clone, use `rm -rf build/ **/*.so` to clean the |
|
old build first. You often need to rebuild detectron2 after reinstalling PyTorch. |
|
|
|
### Install Pre-Built Detectron2 (Linux only) |
|
``` |
|
# for CUDA 10.1: |
|
python -m pip install detectron2 -f https://dl.fbaipublicfiles.com/detectron2/wheels/cu101/index.html |
|
``` |
|
You can replace cu101 with "cu{100,92}" or "cpu". |
|
|
|
Note that: |
|
1. Such installation has to be used with certain version of official PyTorch release. |
|
See [releases](https://github.com/facebookresearch/detectron2/releases) for requirements. |
|
It will not work with a different version of PyTorch or a non-official build of PyTorch. |
|
2. Such installation is out-of-date w.r.t. master branch of detectron2. It may not be |
|
compatible with the master branch of a research project that uses detectron2 (e.g. those in |
|
[projects](projects) or [meshrcnn](https://github.com/facebookresearch/meshrcnn/)). |
|
|
|
### Common Installation Issues |
|
|
|
If you met issues using the pre-built detectron2, please uninstall it and try building it from source. |
|
|
|
Click each issue for its solutions: |
|
|
|
<details> |
|
<summary> |
|
Undefined torch/aten/caffe2 symbols, or segmentation fault immediately when running the library. |
|
</summary> |
|
<br/> |
|
|
|
This usually happens when detectron2 or torchvision is not |
|
compiled with the version of PyTorch you're running. |
|
|
|
Pre-built torchvision or detectron2 has to work with the corresponding official release of pytorch. |
|
If the error comes from a pre-built torchvision, uninstall torchvision and pytorch and reinstall them |
|
following [pytorch.org](http://pytorch.org). So the versions will match. |
|
|
|
If the error comes from a pre-built detectron2, check [release notes](https://github.com/facebookresearch/detectron2/releases) |
|
to see the corresponding pytorch version required for each pre-built detectron2. |
|
|
|
If the error comes from detectron2 or torchvision that you built manually from source, |
|
remove files you built (`build/`, `**/*.so`) and rebuild it so it can pick up the version of pytorch currently in your environment. |
|
|
|
If you cannot resolve this problem, please include the output of `gdb -ex "r" -ex "bt" -ex "quit" --args python -m detectron2.utils.collect_env` |
|
in your issue. |
|
</details> |
|
|
|
<details> |
|
<summary> |
|
Undefined C++ symbols (e.g. `GLIBCXX`) or C++ symbols not found. |
|
</summary> |
|
<br/> |
|
Usually it's because the library is compiled with a newer C++ compiler but run with an old C++ runtime. |
|
|
|
This often happens with old anaconda. |
|
Try `conda update libgcc`. Then rebuild detectron2. |
|
|
|
The fundamental solution is to run the code with proper C++ runtime. |
|
One way is to use `LD_PRELOAD=/path/to/libstdc++.so`. |
|
|
|
</details> |
|
|
|
<details> |
|
<summary> |
|
"Not compiled with GPU support" or "Detectron2 CUDA Compiler: not available". |
|
</summary> |
|
<br/> |
|
CUDA is not found when building detectron2. |
|
You should make sure |
|
|
|
``` |
|
python -c 'import torch; from torch.utils.cpp_extension import CUDA_HOME; print(torch.cuda.is_available(), CUDA_HOME)' |
|
``` |
|
|
|
print valid outputs at the time you build detectron2. |
|
|
|
Most models can run inference (but not training) without GPU support. To use CPUs, set `MODEL.DEVICE='cpu'` in the config. |
|
</details> |
|
|
|
<details> |
|
<summary> |
|
"invalid device function" or "no kernel image is available for execution". |
|
</summary> |
|
<br/> |
|
Two possibilities: |
|
|
|
* You build detectron2 with one version of CUDA but run it with a different version. |
|
|
|
To check whether it is the case, |
|
use `python -m detectron2.utils.collect_env` to find out inconsistent CUDA versions. |
|
In the output of this command, you should expect "Detectron2 CUDA Compiler", "CUDA_HOME", "PyTorch built with - CUDA" |
|
to contain cuda libraries of the same version. |
|
|
|
When they are inconsistent, |
|
you need to either install a different build of PyTorch (or build by yourself) |
|
to match your local CUDA installation, or install a different version of CUDA to match PyTorch. |
|
|
|
* Detectron2 or PyTorch/torchvision is not built for the correct GPU architecture (compute compatibility). |
|
|
|
The GPU architecture for PyTorch/detectron2/torchvision is available in the "architecture flags" in |
|
`python -m detectron2.utils.collect_env`. |
|
|
|
The GPU architecture flags of detectron2/torchvision by default matches the GPU model detected |
|
during compilation. This means the compiled code may not work on a different GPU model. |
|
To overwrite the GPU architecture for detectron2/torchvision, use `TORCH_CUDA_ARCH_LIST` environment variable during compilation. |
|
|
|
For example, `export TORCH_CUDA_ARCH_LIST=6.0,7.0` makes it compile for both P100s and V100s. |
|
Visit [developer.nvidia.com/cuda-gpus](https://developer.nvidia.com/cuda-gpus) to find out |
|
the correct compute compatibility number for your device. |
|
|
|
</details> |
|
|
|
<details> |
|
<summary> |
|
Undefined CUDA symbols; cannot open libcudart.so; other nvcc failures. |
|
</summary> |
|
<br/> |
|
The version of NVCC you use to build detectron2 or torchvision does |
|
not match the version of CUDA you are running with. |
|
This often happens when using anaconda's CUDA runtime. |
|
|
|
Use `python -m detectron2.utils.collect_env` to find out inconsistent CUDA versions. |
|
In the output of this command, you should expect "Detectron2 CUDA Compiler", "CUDA_HOME", "PyTorch built with - CUDA" |
|
to contain cuda libraries of the same version. |
|
|
|
When they are inconsistent, |
|
you need to either install a different build of PyTorch (or build by yourself) |
|
to match your local CUDA installation, or install a different version of CUDA to match PyTorch. |
|
</details> |
|
|
|
|
|
<details> |
|
<summary> |
|
"ImportError: cannot import name '_C'". |
|
</summary> |
|
<br/> |
|
Please build and install detectron2 following the instructions above. |
|
|
|
If you are running code from detectron2's root directory, `cd` to a different one. |
|
Otherwise you may not import the code that you installed. |
|
</details> |
|
|
|
<details> |
|
<summary> |
|
ONNX conversion segfault after some "TraceWarning". |
|
</summary> |
|
<br/> |
|
The ONNX package is compiled with too old compiler. |
|
|
|
Please build and install ONNX from its source code using a compiler |
|
whose version is closer to what's used by PyTorch (available in `torch.__config__.show()`). |
|
</details> |
|
|