- Notifications
You must be signed in to change notification settings - Fork26k
Development Tips
If you want to have no-op incremental rebuilds (which are fast), seeMake no-op build fast below.
If you don't need CUDA, build using USE_CUDA=0: the build is significantly faster. There are also a lot of other build flags that help get rid of components that you might not work on. Below is an opinionated build command that gets rid of a lot of different options that don't get used very often.
USE_KINETO=0 BUILD_CAFFE2=0 USE_DISTRIBUTED=0 USE_NCCL=0 BUILD_TEST=0 USE_XNNPACK=0 USE_FBGEMM=0 USE_QNNPACK=0 USE_MKLDNN=0 USE_MIOPEN=0 USE_NNPACK=0 BUILD_CAFFE2_OPS=0 USE_TENSORPIPE=0 python setup.py developSeeBuild only what you need for a list of useful build flags.
When developing PyTorch, instead of branching off of
master, you can branch off ofviable/strict.viable/strictis a branch that lags behind master and guarantees that all PyTorch tests are passing on the branch. Basing your work off ofviable/strictgives you confidence that any test failures are actually your code's fault.# Creating a new feature branch off of viable/strictgit checkout viable/strictgit checkout -b my_new_feature# Rebasing your work to appear on top of viable/strict, assuming upstream points to pytorch/pytorch.# (Some people develop with origin pointing to pytorch/pytorch)git pull --rebase upstream viable/strict
python setup.py develop will build everything by default, but sometimes you areonly interested in a specific component.
- Working on a test binary? Run
(cd build && ninja bin/test_binary_name)torebuild only that test binary (without rerunning cmake). (Replaceninjawithmakeif you don't have ninja installed). - Don't need Caffe2? Pass
BUILD_CAFFE2=0to disable Caffe2 build.
On the initial build, you can also speed things up with the environmentvariablesDEBUG,USE_DISTRIBUTED,USE_MKLDNN,USE_CUDA,BUILD_TEST,USE_FBGEMM,USE_NNPACK andUSE_QNNPACK.
DEBUG=1will enable debug builds (-g -O0)REL_WITH_DEB_INFO=1will enable debug symbols with optimizations (-g -O3)USE_DISTRIBUTED=0will disable distributed (c10d, gloo, mpi, etc.) build.USE_MKLDNN=0will disable using MKL-DNN.USE_CUDA=0will disable compiling CUDA (in case you are developing on something not CUDA related), to save compile time.BUILD_TEST=0will disable building C++ test binaries.USE_FBGEMM=0will disable using FBGEMM (quantized 8-bit server operators).USE_NNPACK=0will disable compiling with NNPACK.USE_QNNPACK=0will disable QNNPACK build (quantized 8-bit operators).USE_XNNPACK=0will disable compiling with XNNPACK.
For example:
DEBUG=1 USE_DISTRIBUTED=0 USE_MKLDNN=0 USE_CUDA=0 BUILD_TEST=0 USE_FBGEMM=0 USE_NNPACK=0 USE_QNNPACK=0 USE_XNNPACK=0 python setup.py develop
For subsequent builds (i.e., whenbuild/CMakeCache.txt exists), the buildoptions passed for the first time will persist; please runccmake build/, runcmake-gui build/, or directly editbuild/CMakeCache.txt to adapt buildoptions.
When installing withpython setup.py develop (in contrast topython setup.py install) Python runtime will usethe current local source-tree when importingtorch package. (This is done by creating.egg-link file insite-packages folder)This way you do not need to repeatedly install after modifying Python files (.py).However, you would need to reinstall if you modify Python interface (.pyi,.pyi.in) ornon-Python files (.cpp,.cc,.cu,.h, ...).
One way to avoid runningpython setup.py develop every time one makes a change to C++/CUDA/ObjectiveC files on Linux/Mac, is to create a symbolic link frombuild folder totorch/lib, for example, by issuing following:bash pushd torch/lib; sh -c "ln -sf ../../build/lib/libtorch_cpu.* ."; popdAfterwards rebuilding a library (for example to rebuildlibtorch_cpu.so issueninja torch_cpu frombuild folder), would be sufficient to make change visible intorch package.
If you are working on the C++ code, there are a few important things that youwill want to keep in mind:
- How to rebuild only the code you are working on.
- How to make rebuilds in the absence of changes go faster.
When usingpython setup.py develop, PyTorch will generateacompile_commands.json file that can be used by many editorsto provide command completion and error highlighting for PyTorch'sC++ code. You need topip install ninja to generate accurateinformation for the code intorch/csrc. More information at:
By default, cmake will use its Makefile generator to generate your buildsystem. You can get faster builds if you install the ninja build systemwithpip install ninja. If PyTorch was already built, you will needto runpython setup.py clean once after installing ninja for builds tosucceed.
Even when dependencies are tracked with file modification, there are manysituations where files get rebuilt when a previous compilation was exactly thesame. Using ccache in a situation like this is a real time-saver.
Before building pytorch, install ccache from your package manager of choice:
conda install ccache -c conda-forgesudo apt install ccachesudo yum install ccachebrew install ccache
You may also find the default cache size in ccache is too small to be useful.The cache sizes can be increased from the command line:
# config: cache dir is ~/.ccache, conf file ~/.ccache/ccache.conf# max size of cacheccache -M 25Gi# -M 0 for unlimited# unlimited number of filesccache -F 0
To check this is working, do two clean builds of pytorch in a row. The secondbuild should be substantially and noticeably faster than the first build. Ifthis doesn't seem to be the case, check theCMAKE_<LANG>_COMPILER_LAUNCHERrules inbuild/CMakeCache.txt, where<LANG> isC,CXX andCUDA.Each of these 3 variables should contain ccache, e.g.
//CXX compiler launcherCMAKE_CXX_COMPILER_LAUNCHER:STRING=/usr/bin/ccacheIf not, you can define these variables on the command line before invokingsetup.py.
export CMAKE_C_COMPILER_LAUNCHER=ccacheexport CMAKE_CXX_COMPILER_LAUNCHER=ccacheexport CMAKE_CUDA_COMPILER_LAUNCHER=ccachepython setup.py develop
If you are editing a single file and rebuilding in a tight loop, the time spentlinking will dominate. The system linker available in most Linux distributions(GNUld) is quite slow. Use a faster linker, likelld.
People on Mac, followthis guide instead.
The easiest way to uselld this is download thelatest LLVM binaries and run:
ln -s /path/to/downloaded/ld.lld /usr/local/bin/ld
Sometimes there's no way of getting around rebuilding lots of files, for exampleeditingnative_functions.yaml usually means 1000+ files being rebuilt. Ifyou're using CMake newer than 3.16, you can enable pre-compiled headers bysettingUSE_PRECOMPILED_HEADERS=1 either on first setup, or in theCMakeCache.txt file.
USE_PRECOMPILED_HEADERS=1 python setup.py develop
This adds a build step where the compiler takes<ATen/ATen.h> and essentiallydumps it's internal AST to a file so the compiler can avoid repeating itself forevery.cpp file.
One caveat is that when enabled, this header gets included in every file by default.Which may change what code is legal, for example:
- internal functions can never alias existing names in
<ATen/ATen.h> - names in
<ATen/ATen.h>will work even if you don't explicitly include it.
If re-building without modifying any files results in several CUDA files beingre-compiled, you may be running into annvcc bug where header dependencies arenot converted to absolute paths before reporting it to the build system. Thismakesninja think one of the header files has been deleted, so it runs thebuild again.
A compiler-wrapper to fix this is provided intools/nvcc_fix_deps.py. You can usethis as a compiler launcher, similar toccache
export CMAKE_CUDA_COMPILER_LAUNCHER="python;`pwd`/tools/nvcc_fix_deps.py;ccache"python setup.py develop
We have very extensive tests in thetest/cpp/api folder. Thetests are a great way to see how certain components are intended to be used.When compiling PyTorch from source, the test runner binary will be written tobuild/bin/test_api. The tests use theGoogleTestframework, which you can read up about to learn how to configure the test runner. Whensubmitting a new feature, we care very much that you write appropriate tests.Please follow the lead of the other tests to see how to write a new test case.
If you are debugging pytorch inside GDB, you might be interested inpytorch-gdb. This script introduces somepytorch-specific commands which you can use from the GDB prompt. Inparticular,torch-tensor-repr prints a human-readable repr of an at::Tensorobject. Example of usage:
$ gdb pythonGNU gdb (GDB) 9.2[...](gdb) # insert a breakpoint when we call .neg()(gdb) break at::Tensor::negFunction "at::Tensor::neg" not defined.Make breakpoint pending on future shared library load? (y or [n]) yBreakpoint 1 (at::Tensor::neg) pending.(gdb) run[...]>>> import torch>>> t = torch.tensor([1, 2, 3, 4], dtype=torch.float64)>>> ttensor([1., 2., 3., 4.], dtype=torch.float64)>>> t.neg()Thread 1 "python" hit Breakpoint 1, at::Tensor::neg (this=0x7ffb118a9c88) at aten/src/ATen/core/TensorBody.h:32953295 inline at::Tensor Tensor::neg() const {(gdb) # the default repr of 'this' is not very useful(gdb) p this$1 = (const at::Tensor * const) 0x7ffb118a9c88(gdb) p *this$2 = {impl_ = {target_ = 0x55629b5cd330}}(gdb) torch-tensor-repr *thisPython-level repr of *this:tensor([1., 2., 3., 4.], dtype=torch.float64)GDB tries to automatically loadpytorch-gdb thanks to the.gdbinit at the root of the pytorch repo. However, auto-loadings is disabled by default, because of security reasons:
$ gdbwarning: File"/path/to/pytorch/.gdbinit" auto-loading has been declined by your`auto-load safe-path' set to "$debugdir:$datadir/auto-load".To enable execution of this file add add-auto-load-safe-path /path/to/pytorch/.gdbinitline to your configuration file "/home/YOUR-USERNAME/.gdbinit".To completely disable this security protection add set auto-load safe-path /line to your configuration file "/home/YOUR-USERNAME/.gdbinit".For more information about this security protection see the"Auto-loading safe path" section in the GDB manual. E.g., run from the shell: info "(gdb)Auto-loading safe path"(gdb)
As gdb itself suggests, the best way to enable auto-loading ofpytorch-gdbis to add the following line to your~/.gdbinit (i.e., the.gdbinit filewhich is in your home directory,not/path/to/pytorch/.gdbinit):
add-auto-load-safe-path /path/to/pytorch/.gdbinit
SetTORCH_SHOW_CPP_STACKTRACES=1 to get the C++ stacktrace when an error occurs in Python.
If you are working on the CUDA code, here are some useful CUDA debugging tips:
CUDA_DEVICE_DEBUG=1will enable CUDA device function debug symbols (-g -G).This will be particularly helpful in debugging device code. However, it willslow down the build process for about 50% (compared to onlyDEBUG=1), so use wisely.cuda-gdbandcuda-memcheckare your best CUDA debugging friends. Unlikegdb,cuda-gdbcan display actual values in a CUDA tensor (rather than all zeros).- CUDA supports a lot of C++11/14 features such as,
std::numeric_limits,std::nextafter,std::tupleetc. in device code. Many of such features are possible because of the--expt-relaxed-constexprnvcc flag. There is a knownissuethat ROCm errors out on device code, which uses such stl functions. - A good performance metric for a CUDA kernel is theEffective Memory Bandwidth.It is useful for you to measure this metric whenever you are writing/optimizing a CUDAkernel. Following script shows how we can measure the effective bandwidth of CUDA
uniform_kernel.importtorchfromtorch.utils.benchmarkimportTimersize=128*512nrep=100nbytes_read_write=4# this is number of bytes read + written by a kernel. Change this to fit your kernel.foriinrange(10):a=torch.empty(size).cuda().uniform_()torch.cuda.synchronize()out=a.uniform_()torch.cuda.synchronize()t=Timer(stmt="a.uniform_()",globals=globals())res=t.blocked_autorange()timec=res.medianprint("uniform, size, elements",size,"forward",timec,"bandwidth (GB/s)",size*(nbytes_read_write)*1e-9/timec)size*=2
See more cuda development tipshere
For building from source on Windows, consultour documentation on it.
Occasionally, you will write a patch which works on Linux, but fails CI on Windows.There are a few aspects in which MSVC (the Windows compiler toolchain we use) is stricterthan Linux, which are worth keeping in mind when fixing these problems.
Symbols are NOT exported by default on Windows; instead, you have to explicitlymark a symbol as exported/imported in a header file with
__declspec(dllexport)/__declspec(dllimport). We have codified this pattern into a set of macroswhich follow the convention*_API, e.g.,TORCH_APIinside Caffe2, Aten and Torch.(Every separate shared library needs a unique macro name, because symbol visibilityis on a per shared library basis. See c10/macros/Macros.h for more details.)The upshot is if you see an "unresolved external" error in your Windows build, thisis probably because you forgot to mark a function with
*_API. However, there isone important counterexample to this principle: if you want atemplated functionto be instantiated at the call site, do NOT mark it with*_API(if you do mark it,you'll have to explicitly instantiate all of the specializations used by the callsites.)If you link against a library, this does not make its dependencies transitivelyvisible. You must explicitly specify a link dependency against every library whosesymbols you use. (This is different from Linux where in most environments,transitive dependencies can be used to fulfill unresolved symbols.)
If you have a Windows box (we have a few on EC2 which you can request access to) andyou want to run the build, the easiest way is to just run
.ci/pytorch/win-build.sh.If you need to rebuild, runREBUILD=1 .ci/pytorch/win-build.sh(this will avoidblowing away your Conda environment.)
Even if you don't know anything about MSVC, you can use cmake to build simple programs onWindows; this can be helpful if you want to learn more about some peculiar linking behaviorby reproducing it on a small example. Here's a simple example cmake file that definestwo dynamic libraries, one linking with the other:
project(myproject CXX)set(CMAKE_CXX_STANDARD 14)add_library(foo SHARED foo.cpp)add_library(bar SHARED bar.cpp)# NB: don't forget to __declspec(dllexport) at least one symbol from foo,# otherwise foo.lib will not be created.target_link_libraries(barPUBLIC foo)
You can build it with:
mkdir buildcd buildcmake ..cmake --build.
The PyTorch codebase sometimes likes to use exciting C++ features, andthese exciting features lead to exciting bugs in Windows compilers.To add insult to injury, the error messages will often not tell youwhich line of code actually induced the erroring template instantiation.
We've found the most effective way to debug these problems is tocarefully read over diffs, keeping in mind known bugs in MSVC/NVCC.Here are a few well known pitfalls and workarounds:
This is not actually a bug per se, but in general, code generated by MSVCis more sensitive to memory errors; you may have written some codethat does a use-after-free or stack overflows; on Linux the codemight work, but on Windows your program will crash. ASAN may notcatch all of these problems: stay vigilant to the possibility thatyour crash is due to a real memory problem.
(NVCC)
c10::optionaldoes not work when used from device code. Don't useit from kernels. Upstream issue:https://github.com/akrzemi1/Optional/issues/58and our local issue #10329.constexprgenerally works less well on MSVC.- The idiom
static_assert(f() == f())to test iffis constexprdoes not work; you'll get "error C2131: expression did not evaluateto a constant". Don't use these asserts on Windows.(Example:c10/util/intrusive_ptr.h)
- The idiom
(NVCC) Code you access inside a
static_assertwill eagerly beevaluated as if it were device code, and so you might get an errorthat the code is "not accessible".
classA {static A singleton_;staticconstexprinline A*singleton() {return &singleton_; }};static_assert(std::is_same(A*,decltype(A::singleton()))::value,"hmm");
The compiler will run out of heap space if you attempt to compile files thatare too large. Splitting such files into separate files helps.(Example:
THTensorMath,THTensorMoreMath,THTensorEvenMoreMath.)MSVC's preprocessor (but not the standard compiler) has a bugwhere it incorrectly tokenizes raw string literals, ending when it sees a
".This causes preprocessor tokens inside the literal like an#endifto be incorrectlytreated as preprocessor directives. Seehttps://godbolt.org/z/eVTIJq as an example.Either MSVC or the Windows headers have a PURE macro defined and will replaceany occurrences of the PURE token in code with an empty string. This is whywe have AliasAnalysisKind::PURE_FUNCTION and not AliasAnalysisKind::PURE.The same is likely true for other identifiers that we just didn't try to use yet.
CUDA, MSVC, and PyTorch versions are interdependent; please install matching versions from this table:
| CUDA version | Newest supported VS version | PyTorch version |
|---|---|---|
| 10.1 | Visual Studio 2019 (16.X) (_MSC_VER < 1930) | 1.3.0 ~ 1.7.0 |
| 10.2 | Visual Studio 2019 (16.X) (_MSC_VER < 1930) | 1.5.0 ~ 1.7.0 |
| 11.0 | Visual Studio 2019 (16.X) (_MSC_VER < 1930) | 1.7.0 |
Note: There's acompilation issue in several Visual Studio 2019 versions since 16.7.1, so please make sure your Visual Studio 2019 version is not in 16.7.1 ~ 16.7.5
ASAN is veryuseful for debugging memory errors in C++. We run it in CI, but here's how toget the same thing to run on your local machine.
First, install LLVM 8. The easiest way is to getprebuiltbinaries and extract them tofolder (later called$LLVM_ROOT).
Then set up the appropriate scripts. You can put this in your.bashrc:
LLVM_ROOT=<wherever your llvm install is>PYTORCH_ROOT=<wherever your pytorch checkout is>LIBASAN_RT="$LLVM_ROOT/lib/clang/8.0.0/lib/linux/libclang_rt.asan-x86_64.so"build_with_asan(){ LD_PRELOAD=${LIBASAN_RT} \ CC="$LLVM_ROOT/bin/clang" \ CXX="$LLVM_ROOT/bin/clang++" \ LDSHARED="clang --shared" \ LDFLAGS="-stdlib=libstdc++" \ CFLAGS="-fsanitize=address -fno-sanitize-recover=all -shared-libasan -pthread" \ CXX_FLAGS="-pthread" \ USE_CUDA=0 USE_OPENMP=0 BUILD_CAFFE2_OPS=0 USE_DISTRIBUTED=0 DEBUG=1 \ python setup.py develop}run_with_asan(){ LD_PRELOAD=${LIBASAN_RT}$@}# you can look at build-asan.sh to find the latest options the CI usesexport ASAN_OPTIONS=detect_leaks=0:symbolize=1:strict_init_order=trueexport UBSAN_OPTIONS=print_stacktrace=1:suppressions=$PYTORCH_ROOT/ubsan.suppexport ASAN_SYMBOLIZER_PATH=$LLVM_ROOT/bin/llvm-symbolizer
Then you can use the scripts like:
suo-devfair ~/pytorch ❯ build_with_asansuo-devfair ~/pytorch ❯ run_with_asan python test/test_jit.pyThe scripts above specify theclang andclang++ binaries directly, whichbypassesccache. Here's how to getccache to work:
- Make sure the ccache symlinks for
clangandclang++are set up (seeCONTRIBUTING.md) - Make sure
$LLVM_ROOT/binis available on your$PATH. - Change the
CCandCXXvariables inbuild_with_asan()to pointdirectly toclangandclang++.
The “standard” workflow for ASAN assumes you have a standalone binary:
- Recompile your binary with
-fsanitize=address. - Run the binary, and ASAN will report whatever errors it find.
Unfortunately, PyTorch is a distributed as a shared library that is loaded bya third-party executable (Python). It’s too much of a hassle to recompile allof Python every time we want to use ASAN. Luckily, the ASAN folks have aworkaround for cases like this:
- Recompile your library with
-fsanitize=address -shared-libasan. Theextra-shared-libasantells the compiler to ask for the shared ASANruntime library. - Use
LD_PRELOADto tell the dynamic linker to load the ASAN runtimelibrary before anything else.
More information can be foundhere.
We needLD_PRELOAD because there is a cmake check that ensures that asimple program builds and runs. If we are building with ASAN as a sharedlibrary, we need toLD_PRELOAD the runtime library, otherwise there willdynamic linker errors and the check will fail.
We don’t actually need either of these if we fix the cmake checks.
Python leaks a lot of memory. Possibly we could configure a suppression file,but we haven’t gotten around to it.
I would love to contribute to PyTorch!