Note
relion is distributed under a GPLv2 license, i.e. it is completely free, open-source software for both academia and industry.
Installation¶
The sections below explain how to download and install relion on your computer.
Note that relion depends on and uses several external programs and libraries.
- C++ compiler:
RELION 5.0 requires a C++ compiler that fully supports the C++14 standard. For GCC, this means version 5.0 or later. Note that GCC 4.8, which comes with RedHat Enterprise Linux / Cent OS 7.x, is too old. You can obtain newer GCC via devtoolset or use free Intel compiler that comes with oneAPI toolkit (see below).
- MPI:
Your system will need MPI runtime (most flavours will do). If you don’t have an MPI installation already on your system, we recommend installing OpenMPI.
- CUDA, HIP/ROCm, SYCL or oneAPI intel compilers:
If you have GPUs from nvidia, AMD or Intel, you can accelerate many jobs considerably.
By default, relion will build with GPU-acceleration support, for which you’ll need cuda. Download it from NVIDIA website. Note that CUDA toolkits support only a limited range of C compilers. Also note that a newer CUDA toolkit requires a newer GPU driver. Carefully read the release note and make sure you have a compatible set of GPU driver, C compiler and CUDA toolkit.
- If you want to compile with HIP/ROCm, you will need
- If you want to compile with SYCL, you will need
Intel oneAPI Base Toolkit and HPC Toolkit (All components recommended; this is also recommended if you want to build the CPU acceleration path, see below)
Codeplay oneAPI for NVIDIA GPU (optional)
Codeplay oneAPI for AMD GPU (optional)
- CTFFIND-4.1:
CTF estimation is not part of relion. Instead, relion provides a wrapper to Alexis Rohou and Niko Grigorieff’s ctffind 4 [RG15]. Please obtain CTFFIND 4.1.x from their Web site. Note that CTFFIND 5.x is not supported.
- Ghostscript:
RELION uses Ghostscript to generate PDF files.
- FLTK (only for GUI):
RELION uses FLTK as a GUI tool kit. This will be installed automatically (see below).
- X Window system libraries (only for GUI):
RELION needs basic X11 libraries together with Xft for the GUI. Most Linux distributions have packages called
libxft-dev
orlibXft-devel
andlibX11-devel
. Note that you need developer packages if you build your own FLTK.- FFT libraries:
RELION needs an FFT library. The default is FFTW. This will be installed automatically (see below). Depending on your CPU, Intel MKL FFT or AMD optimised FFTW might run faster. See below how to use them.
- libtiff:
RELION needs libtiff version >= 4.0. Most Linux distributions have packages called
libtiff-dev
orlibtiff-devel
. Note that you need a developer package.- libpng:
RELION needs libpng. Most Linux distributions have packages called
libpng-dev
orlibpng-devel
. Note that you need a developer package.- pbzip2, xz, zstd:
RELION needs these commands in the
PATH
to read MRC movies compressed by bzip2, xz or ZStandard, respectively. Note that RELION usespbzip2
, notbzip2
. Most Linux distributions provide packages for these utilities.- UCSF MotionCor2 (optional):
relion implements its own (CPU-only) implementation of the UCSF motioncor2 algorithm for whole-frame micrograph movie-alignment [ZPA+17]. If you want, you can still use the (GPU-accelerated) UCSF program. You can download it from David Agard’s page and follow his installation instructions. Note that using the UCSF program does not make full advantage of the opportunities provided in Bayesian polishing.
- ResMap (optional):
Local-resolution estimation may be performed inside relion’s own postprocessing program. Alternatively, one can also use Alp Kucukelbir’s resmap [KST14]. Download it from Alp’s ResMap website and follow his installation instructions.
In practice, most of these dependencies can be installed by system’s package manager if you have the root priviledge.
In Debian or Ubuntu:
sudo apt install cmake git build-essential mpi-default-bin mpi-default-dev libfftw3-dev libtiff-dev libpng-dev ghostscript libxft-dev
In RHEL, Cent OS, Scientific Linux:
sudo yum install cmake git gcc gcc-c++ openmpi-devel fftw-devel libtiff-devel libpng-devel ghostscript libXft-devel libX11-devel
Download RELION¶
We store the public release versions of relion on GitHub, a site that provides code-development with version control and issue tracking through the use of git
.
We will not describe the use of git in general, as you will not need more than very basic features.
Below we outline the few commands needed on a UNIX-system, please refer to general git descriptions and tutorials to suit your system.
To get the code, you clone or download the repository.
We recommend cloning, because it allows you very easily update the code when new versions are released.
To do so, use the shell command-line:
git clone https://github.com/3dem/relion.git
This will create a local Git repository. All subsequent git-commands should be run inside this directory.
The master branch (default) contains the stable release of relion-4.0. By performing:
git checkout ver5.0
you can access the latest (developmental) updates for RELION 5.0x.
The code will be intermittently updated to amend issues. To incorporate these changes, use the command-line:
git pull
inside you local repository (the source-code directory downloaded). If you have changed the code in some way, this will force you to commit a local merge. You are free to do so, but we will assume you have not changed the code. Refer to external instructions regarding git and merging so-called conflicts if you have changed the code an need to keep those changes.
Setup a conda environment¶
To add support for Python modules (e.g. Blush, ModelAngelo and DynaMight) you will have to setup a Python environment with dependencies. We recommend installing via Miniforge to avoid inadvertently installing packages from the restrictively licensed “default” conda repository.
Once you have conda setup, you can install all the RELION Python dependencies into a new environment by running:
conda env create -f environment.yml
Also code in this environment will be updated intermittently. You can incorporate the latest changes by running:
conda env update -f environment.yml
Warning
You should NOT activate this relion-5.0
conda environment when compiling and using RELION;
RELION activates it automatically only when necessary.
Otherwise, system-wide installation of compilers/libraries/MPI runtime might get mixed up with those provided by conda, leading to compilation failures or runtime errors.
The same applies to other software packages that provide their own libraries/MPI runtime, such as CCPEM, CCP4, EMAN2, DIALS, PHENIX.
The cmake
command should automatically detect the relion-5.0
conda environment created above.
If it does not, you can specify -DPYTHON_EXE_PATH=path/to/your/conda/python
.
Additionally, if you intend to make use of automatically downloaded pretrained model weights (used in e.g. Blush, ModelAngelo and class_ranker), it is recommended to explicitly set the destination directory (TORCH_HOME
) by including the flag -DTORCH_HOME_PATH=path/to/torch/home
.
You have to create this directory before running cmake
.
Otherwise, it will be downloaded to the default location (usually ~/.cache/torch
).
At the moment, the model weights for Blush are stored on MRC-LMB’s FTP server. If your network blocks FTP, please follow instructions here.
Compilation¶
relion has an installation procedure which relies on cmake
.
You will need to have this program installed, but most UNIX-systems have this by default.
You will need to make a build-directory in which the code will be compiled.
This can be placed inside the repository:
cd relion
mkdir build
cd build
You then invoke cmake
inside the build-directoy, but point to the source-directoy to configure the installation.
This will not install relion, just configure the build:
cmake ..
The output will notify you of what was detected and what type of build will be installed. Because relion is rich in terms of the possible configurations, it is important to check this output. For instance:
The path to the MPI library.
GPU-capability will only be included if a CUDA SDK is detected. If not, the program will install, but without support for GPUs.
The path to the Python interpreter.
If FFTW is not detected, instructions are included to download and install it in a local directory known to the relion installation.
As above, regarding FLTK (required for GUI). If a GUI is not desired, this can be escaped as explained in the following section.
The MPI library must be the one you intend to use relion with.
Compiling relion with one version of MPI and running the resulting binary with mpirun
from another version can cause crash.
Note that some software packages (e.g. CCPEM, crYOLO, EMAN2) come with their own MPI runtime.
Sourcing/activating their environment might update PATH
and LD_LIBRARY_PATH
environmental variables and put their MPI runtime into the highest priority.
The MPI C++ compiler (mpicxx
) and CUDA compiler (nvcc
) internally calls a C++ compiler.
This must match the compiler cmake
picked up.
Otherwise, the compilation might fail at the linking step.
Following the completion of cmake-configuration without errors, make
is used to install the program:
make -j N
, where N
is the number of processes to use during installation.
Using a higher number simply means that it will compile faster.
Take note of any warnings or errors reported.
relion will be installed in the build
directory’s sub-directory called bin
.
To make the installation system-wide, see below.
Wherever you install relion, make sure your PATH
environmental variable points to the directory containing relion binaries.
Launching relion with a path like /path/to/relion
is not the right way;
this starts the right GUI, but the GUI might invoke other versions of relion in the PATH
.
General configuration¶
CMake allows configuration of many aspects of the installation, some of which are outlined here. Note that by default, relion is configured to build with CUA acceleration on NVidia GPUs. Instructions for building with CPU, HIP/Rocm (AMD) SYCL (Intel et al) acceleration are given in the next section below.
Most options can be set by adding options to the cmake
configuration.
Under the below subheadings, some example replacement commands are given to substitute the original configuration command.
It is also recommended to clean or purge your build-directory between builds, since CMake caches some of previous configurations:
cd build
rm -fr *
And of course, any of the below options can be combined.
- Omitting the GUI:
cmake -DGUI=OFF ..
(default is ON)With this option, GUI programs (e.g.
relion
,relion_manualpick
,relion_display
) are not be built and FLTK becomes unnecessary.- Using single-precision on the CPU:
cmake -DDoublePrec_CPU=OFF ..
(default is ON)This will reduce (CPU but not GPU) memory consumption to about half. This is useful when memory hungry tasks such as motion correction and Polishing run out of memory. This is safe in most cases but please use the default double precision build if CtfRefine produces NaNs.
- Using double-precision on the GPU:
cmake -DDoublePrec_GPU=ON ..
(default is OFF)This will slow down GPU-execution considerably, while this does NOT improve the resolution. Thus, this option is not recommended.
- Compiling NVIDIA GPU codes for your architecture:
cmake -DCUDA_ARCH=52 ..
(default is 35, meaning compute capability 3.5, which is the lowest supported by relion)CUDA-capable NVIDIA devices have a so-called compute capability, which code can be compiled against for optimal performance. The compute capability of your card can be looked up at the table in NVIDIA website. WARNING: If you use a wrong number, compilation might succeed but the resulting binary can fail at the runtime.
- Forcing build and use of local FFTW:
cmake -DFORCE_OWN_FFTW=ON ..
This will download, verify and install FFTW during the installation process.
- Forcing build and use of AMD optimized FFTW:
cmake -DFORCE_OWN_FFTW=ON -DAMDFFTW=ON ..
This will download, verify and install AMD optimized version of FFTW during the installation process. This is recommended for AMD CPUs (e.g. Ryzen, Threadripper, EPYC).
- Forcing build and use of Intel MKL FFT:
cmake -DMKLFFT=ON ..
This will use FFT library from Intel MKL. In contrast to the FFTW options above, this will not download MKL automatically. You have to install MKL and set relevants paths (usually by sourcing the
mkl_vars.sh
script).- Forcing build and use of local FLTK:
cmake -DFORCE_OWN_FLTK=ON ..
This will download, verify and install FLTK during the installation process. If any of these are not detected during configuration, this will happen automatically anyway, and you should not have to specify the below options manually.
- Specify location of libtiff:
cmake -DTIFF_INCLUDE_DIR=/path/to/include -DTIFF_LIBRARY=/path/to/libtiff.so.5
This option is to use libtiff installed in non-standard location.
- Specifying an installation location:
To allow relion a system-wide installation use:
cmake -DCMAKE_INSTALL_PREFIX=/path/to/install/dir/ .. make -j N make install
Warning
Do not specify the build
directory itself as CMAKE_INSTALL_PREFIX
.
This does not work!
If you are happy with binaries in the build directory, leave CMAKE_INSTALL_PREFIX
as default and omit the make install
step.
Configuration with CPU acceleration¶
- Enable accelerated CPU code path:
cmake -DALTCPU=ON
Note that this is mutually exclusive with GPU acceleration (
-DCUDA=ON
). Intel Classic compilers are recommended for this option (see below).- Use Intel compilers:
There are two Intel compilers: Intel Classic compiler and Intel oneAPI DPC++/C++ compiler. They often generate faster binaries for Intel CPUs, especially when combined with the accelerated CPU code path above. As of 2024 April, the classic compiler generate faster binaries than the DPC++/C++ compiler.
Both compilers used to be available free of chage as part of Intel oneAPI HPC toolkit. Unfortunately, the classic compiler was removed in the 2024 release and only the DPC++/C++ compiler is currently distributed. (Apparently older versions seem to be available via YUM/APT, but we do not know how long they remain.) We recommend you to use the classic compiler if you have older oneAPI toolkit installers at hand.
To use Intel Classic compiler, run below after sourcing the initialization script (setvars.sh):
mkdir build-cpu cd build-cpu cmake .. -DMKLFFT=ON \ -DCMAKE_C_COMPILER=icc -DCMAKE_CXX_COMPILER=icpc -DMPI_C_COMPILER=mpiicc -DMPI_CXX_COMPILER=mpiicpc \ -DCMAKE_C_FLAGS="-O3 -ip -g -xCOMMON-AVX512 -restrict " \ -DCMAKE_CXX_FLAGS="-O3 -ip -g -xCOMMON-AVX512 -restrict " make -j 24
This generates binaries optimized with AVX512 instructions. If your CPU supports only up to AVX256, use
-xCORE-AVX2
instead of-xCOMMON-AVX512
.If you do not have Intel Classic compiler, use the DPC++/C++ compiler from the latest oneAPI release. Its performance is being improved. The
cmake
line should be:cmake .. -DMKLFFT=ON \ -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx -DMPI_C_COMPILER=mpiicx -DMPI_CXX_COMPILER=mpiicpx \ -DCMAKE_C_FLAGS="-O3 -qopenmp-simd -xCORE-AVX512 -qopt-zmm-usage=high -qoverride-limits " \ -DCMAKE_CXX_FLAGS="-O3 -qopenmp-simd -xCORE-AVX512 -qopt-zmm-usage=high -qoverride-limits "
If your CPU supports only up to AVX256, use
-xCORE-AVX2
instead of-xCORE-AVX512
.If you don’t want to use Intel MPI, change
DMPI_C_COMPILER
andDMPI_CXX_COMPILER
variables accordingly. For example, to use OpenMPI with Intel Classic compiler, specifympicc
andmpicxx
after setting environmental variablesOMPI_CC=icc
andOMPI_CXX=icpc
. Ifcmake
still picks up Intel MPI, specifyMPI_HOME
. See OpenMPI FAQ and FindMPI manual for details.
Configuration with HIP/ROCm acceleration for AMD GPUs¶
- Enable the accelerated HIP/ROCm code path with:
cmake -DHIP=ON
Note that this is mutually exclusive with other accelerated code paths (e.g. CUDA, ALTCPU and SYCL). On our system, we build with HIP/ROCm acceleration to use AMD GPUs with the following commands:
export LD_LIBRARY_PATH=/opt/rocm/lib:$LD_LIBRARY_PATH
export PATH=/opt/rocm/:$PATH
export ROCM_PATH=/opt/rocm/
mkdir build-amd
cd build-amd
cmake -DCMAKE_BUILD_TYPE=Release -DHIP=ON -DHIP_ARCH="gfx90a,gfx908" -DFORCE_OWN_FFTW=ON -DAMDFFTW=on ..
make -j 24
If you get problems finding omp.h
, make sure you have openmp-extras-devel
installed on your system too.
Configuration with SYCL acceleration (Intel GPUs)¶
- Enable accelerated the SYCL code path with:
cmake -DSYCL=ON
Note that this is mutually exclusive with other accelerated code paths (e.g. CUDA, ALTCPU and HIP/ROCm). Technically speaking, you can build SYCL for AMD and NVIDIA GPUs to make a single binary that runs on NVIDIA, AMD and Intel GPUs, but this is highly experimental and not tested well.
For now, this way of building RELION is explained here:.
Set-up queue job submission¶
The GUI allows the user to submit jobs to a job queueing system with a single click. For this to work, a template job submission script needs to be provided for the queueing system at hand (e.g. TORQUE, PBS, SGE). In this script a set of strings (variables) in the template script is replaced by the values given in the GUI. The following table contains all defined variables:
String |
Variable |
Meaning |
---|---|---|
|
string |
The standard output log file RELION GUI displays. |
|
string |
The standard error log file RELION GUI displays. |
|
string |
relion command + arguments |
|
string |
Name of the queue to submit job to |
|
integer |
The number of MPI processes to use |
|
integer |
The number of threads to use on each MPI process |
|
integer |
The number of MPI processes times the number of threads |
|
integer |
The minimum number of cores on each node (use this to fill entire nodes) |
|
integer |
The total number of nodes to be requested |
|
string |
Installation-specific, see below |
|
string |
Installation-specific, see below |
The XXXcommandXXX
variable needs a special care.
For non-MPI commands (e.g. relion_refine
) not only the variable but the whole line is replaced.
Thus, mpirun XXXcommandXXX
will be mpirun relion_refine_mpi
for an MPI job but relion_refine
for a non-MPI job.
Also note that some jobs consist of multiple lines of commands.
See CCPEM threads (1 and 2) for typical pitfalls.
There are two environment variables that control the use of the entry of the ‘Minimum number of dedicated cores per node’ on the Running tabs of the GUI: RELION_MINIMUM_DEDICATED
sets its default value (1 if not set); RELION_ALLOW_CHANGE_MINIMUM_DEDICATED
sets whether the user will be able to change this entry. At LMB, we set the default to 24 and do not allow users to change it. In this way, we enforce that our hyper-threaded 12-core nodes get filled up entirely by individual relion jobs.
By default, the XXXextra1XXX
, XXXextra2XXX
, … variables are not used.
They provide additional flexibility for queueing systems that require additional variables.
They may be activated by first setting RELION_QSUB_EXTRA_COUNT
to the number of fields you need (e.g.
2) and then setting the RELION_QSUB_EXTRA1
, RELION_QSUB_EXTRA2
, … environment variables, respectively. This will result in extra input fields in the GUI, with the label text being equal to the value of the environment variable.
Likewise, their default values (upon starting the GUI) can be set through environment variables RELION_QSUB_EXTRA1_DEFAULT
, RELION_QSUB_EXTRA2_DEFAULT
, etc and their help messages can be set through environmental variables RELION_QSUB_EXTRA1_HELP
, RELION_QSUB_EXTRA2_HELP
and so on.
But note that (unlike all other entries in the GUI) the extra values are not remembered from one run to the other.
The template job submission script may be saved in any location.
By default, the one used at the LMB is present as gui/qsub.csh
in the relion tar-ball.
Upon installation this file is copied to the bin directory.
It is convenient for the user if he does not have to select this file each time he opens the relion GUI in a new directory.
Therefore, one may set the environment variable RELION_QSUB_TEMPLATE
to point to the location of the script for the system at hand.
This value will be pre-set as default in the GUI.
(Note the user still has the liberty to define and use his own template!)
Note
If somehow the job queue submission cannot be set up, relion may still be run in parallel and on a job queueing system. The GUI comprises a Print command button that prints the entire relion command, including all arguments, to the screen. Pasting of this command to a job queue submission script, and manual submission of this script may then be used to submit the parallel job to a queueing system.
Edit the environment set-up¶
For relion, we source the following C-shell setup in our .cshrc
file.
You’ll need to change all the paths for your own system, and translate the script in case you use a bash shell (which uses export instead of setenv etc).
#!/bin/csh -f
# Setup openMPI if not already done so
if ("" == "`echo $path | grep /public/EM/OpenMPI/openmpi/bin`") then
set path=(/public/EM/OpenMPI/openmpi/bin $path)
endif
if ("1" == "$?LD_LIBRARY_PATH") then
if ("$LD_LIBRARY_PATH" !~ */public/EM/OpenMPI/openmpi/lib*) then
setenv LD_LIBRARY_PATH /public/EM/OpenMPI/openmpi/lib:$LD_LIBRARY_PATH
endif
else
setenv LD_LIBRARY_PATH /public/EM/OpenMPI/openmpi/lib
endif
# Setup |RELION| if not already done so
if ("" == "`echo $path | grep /public/EM/RELION/relion/bin`") then
set path=(/public/EM/RELION/relion/bin $path)
endif
if ("1" == "$?LD_LIBRARY_PATH") then
if ("$LD_LIBRARY_PATH" !~ */public/EM/RELION/relion/lib*) then
setenv LD_LIBRARY_PATH /public/EM/RELION/relion/lib:$LD_LIBRARY_PATH
endif
else
setenv LD_LIBRARY_PATH /public/EM/RELION/relion/lib
endif
# CUDA for RELION
setenv PATH /public/EM/CUDA/Cuda11.4/bin:$PATH
setenv LD_LIBRARY_PATH /public/EM/CUDA/Cuda11.4/lib64:$LD_LIBRARY_PATH
setenv CUDA_HOME /public/EM/CUDA/Cuda11.4
# Where is qsub template script stored
setenv RELION_QSUB_TEMPLATE /public/EM/RELION/relion-devel/bin/qsub.csh
# Default PDF viewer
setenv RELION_PDFVIEWER_EXECUTABLE evince
# Default MOTIONCOR2 executable
setenv RELION_MOTIONCOR2_EXECUTABLE /public/EM/MOTIONCOR2/bin/MotionCor2_1.0.4
# Default CTFFIND-4.1+ executable
setenv RELION_CTFFIND_EXECUTABLE /public/EM/ctffind/ctffind.exe
# Default ResMap executable
setenv RELION_RESMAP_EXECUTABLE /public/EM/ResMap/ResMap-1.1.4-linux64
# Enforce cluster jobs to occupy entire nodes with 24 hyperthreads
setenv RELION_MINIMUM_DEDICATED 24
# Do not allow the user to change the enforcement of entire nodes
setenv RELION_ALLOW_CHANGE_MINIMUM_DEDICATED 0
# Ask for confirmation if users try to submit local jobs with more than 12 MPI nodes
setenv RELION_WARNING_LOCAL_MPI 12
# Other useful variables
# RELION_MPI_RUN: The mpi runtime ('mpirun' by default)
# RELION_QSUB_NRMPI: The default for 'Number of MPI procs'
# RELION_MPI_MAX: The maximum number of MPI processes available from the GUI
# RELION_QSUB_NRTHREADS: The default for 'Number of threads'
# RELION_THREAD_MAX: The maximum number of threads per MPI process available from the GUI
# RELION_QUEUE_USE: The default for 'Submit to queue?'. "yes" or "no".
# RELION_QUEUE_NAME: The default for 'Queue Name"
# RELION_QSUB_COMMAND: The default for 'Queue submit command'
# RELION_MINIMUM_DEDICATED: The default for 'Minimum dedicated cores per node'
# RELION_ALLOW_CHANGE_MINIMUM_DEDICATED: Whether to allow a user to change the 'Minimum dedicated cores per node' field in the GUI
# RELION_SHELL: A shell used to launch CTFFIND in CtfFind jobs ('csh' by default; only available from 3.1)
# RELION_SCRATCH_DIR: The default scratch directory in the GUI
# RELION_STACK_BUFFER: The buffer size used for MRC(S) file I/O, potentially useful on GPFS or Lustre file system. See https://github.com/3dem/relion/pull/783 for details.