BiOinfOrmatics ThreAded Benchmark tooLsuitE (BOOTABLE) This toolsuite currently consists of the following tools
- BBMap
- Bowtie2
- BWA
- Diamond
- Velvet
- IDBA
- GROMACS
- Tensowflow
- SPAdes
- Clustal Omega
- MAFFT
- SINA
more are coming soon ...
All the tools and datasets included in this bundle are carfeully chosen to cover a range of different application cases and make strong use of multithreading in order to get some benchmarks of the underlying hardware and about the scalability of the system. In the following is explained how to install BOOTABLE, how to run it and what kind of options you have. Information about results are stated in the Results section.
If you have any questions or remarks just write me a mail: maximilian.hanussek@uni-tuebingen.de
The docker version has been tested in the meantime, there was a small bug concerning the installation of MAFFT. The Dockerfile has already been updated. The rebuild image will follow as soon as possible.
The docker version has also been updated and now incorporates the newly added tools. A fully build container is available on DockerHub. Currently the Docker image is untested, test will be done in the upcoming week.
BOOTABLE has some new tools. We added the tools BBMap, BWA, MAFFT and SINA. We ahve also added some new datasets in order to use SINA. Currently the updated tools are only available using the bare metal installation. We are working on the update of the other versions (Docker, Singularity, QCOW2 Image, Ansible). GPU resouces are still on the agenda.
We apologize for the longer lasting maintenance of our infrastructure but now everything is restored and you should be able access the chosen datasets implemented in BOOTABLE. For the next months we also plan an expanded version of BOOTABLE including benchmark tests using GPU resources.
Due to a restructuring of our cloud and storage infrastructure the automated file download from our S3 storage system will not work. As soon as we have it back online it will be listed in the news section.
BOOTABLE has been puplished in the FGCS LIFE 2019 special issue. So if you want to learn more about the technical details of BOOTABLE please have a look at the paper and please cite it if it has been useful in your own work.
We added the possibility to install BOOTABLE via Ansible.
- You will find more about how to use it in the Using Ansible section under the installation possibilities.
We added a generic approach to make it possible to test every tool not already inlcuded in BOOTABLE.
- You can use all built in plot variants
- You can use the scaling mode to test your tool with regards to scalability
- You can use all already available datasets within BOOTABLE
- Use a simple textfile approach to specify your tool name the command to run it and what size of dataset you are using.
More description can be found in the Start Benchmarks section.
We changed some things regarding the report generator and the executing benchmark script.
- New scaling mode is now available in the
run_benchmark.sh
script (flag -s) more in the Start Benchmarks section - New scaling plots are available in the report generator tool (parameter: scaling) more in the Results section
- New shorter name of the report generator tool. Before:
threadedBioBenchsuiteStatsGenerator.R
Now:BOOTABLE_report_generator.R
This current version is only tested for CentOS 7. Ubuntu support will be added soon.
If you want to install the tools and scripts BOOTABLE use please be sure to have following packages installed via yum (CentOS):
- epel-release
- Development Tools (group install)
- nano
- curl
- wget
- vim
- htop
- time
- git
- nmon
- zlib-devel.x86_64
- python-pip (in Version 9.0.3)
- cmake3
- tbb-devel.x86_64
- argtable-devel
- R
- inxi
- uitl-linux
- hwloc
- hwloc-devel
- gmp-devel
- mpfr-devel
- libmpc-devel
yum update -y yum install epel-release -y yum group install "Development Tools" -y yum install nano curl wget vim htop nmon time git zlib-devel.x86_64 python-pip inxi cmake3 tbb-devel.x86_64 argtable-devel util-linux hwloc hwloc-devel gmp-devel mpfr-devel libmpc-devel R -y
Further the system needs access to the internet to load the specific datasets and some R packages.
You can also install the required R packages in beforehand which are:
- grid
- gridExtra
- RColorBrewer
- stringr
If you can not wait to test your system or tool and want to read through the whole tutorial here just the commands to start on a bare metal system.
- Clone the github repo
git clone https://github.com/MaximilianHanussek/BOOTABLE.git
- Change into the root directory of the cloned repo
cd BOOTABLE/
- Install BOOTABLE
sh install_bootable.sh
- Check if installation is correct
sh install_check.sh
- Run your first benchmark
sh run_benchmarks.sh
- After the benchmarking has finished create a report if wanted
Rscript --vanilla BOOTABLE_report_generator.R
There are 6 different kinds of installation possibilities.
- Bare metal installation
- Using a .qcow2 image with everything already preinstalled
- Using Docker and the provided Dockerfile to install everything into a Docker Container
- Using Singularity and the provided Singularity file
- Transform the Docker container from option 3. into a Singularity container
- Use Ansible and the provided Ansible script to install the tools on a CentOS based system from scratch
In order to compile and install the tools directly on a server without a virtualization technique
in between, clone this github repo and run the install_bootable.sh
script like in the following.
Clone the github repo
git clone https://github.com/MaximilianHanussek/BOOTABLE.git
Change into the root directory of the cloned repo
cd BOOTABLE/
Run the installation script and all the tools will be compiled and installed in the BOOTABLE directory except some GROMACS tools which will be installed under /usr/local/ at the moment and tensorflow which will be installed via pip
sh install_bootable.sh
During the installation you will see a percentage bar which will fill up after they tasks are finished. The compilation of the gcc compiler will take the most time so do not wonder if it can take up to an hour. You can find all the logs from the installation process in the log
directory.
Afterwards please run the install_ckeck.sh
tool from the BOOTABLE root directory in order to check everything is installed
and works as it is intended.
sh install_check.sh
If everything is correct all the output on the screen will appear in green. If not, the problematic tools/files will be marked in red.
If everything is green you can start with running your first benchmark.
This instructions tell you how to use a .qcow2 image with everything preinstalled. It is further assumed that you know how to install an image in an virtual environment or have a virtual environment already running. Make sure you have at least 50GB of disk space.
The image is deposited in an S3 bucket with the following URL https://s3.denbi.uni-tuebingen.de/max/benchmark_image.qcow2
.
You can just download it into your current directory via wget
for example:
wget https://s3.denbi.uni-tuebingen.de/max/benchmark_image.qcow2
Just start the image in any virtuel environment you like, for example with virt-manager
under CentOS.
You can login with the following credentials:
- Username: root
- Password: rootdenbi
After the login please change the password of the user centos
via the passwd command to any you like.
Afterwards logout and login to the virtual machine as user centos
with the password you have entered before.
Change into the home directory (/home/centos/
), where you will find all tools already installed. Start the benchmarks by running the following command:
sh run_benchmarks
This will start the calculations of the different tools first with the maximal number of CPU cores, second the half of the maximal number of CPU cores and at the end with one CPU core. Each number of CPU cores is running three times to get the mean value. As a comparison value, on a 28 core machine this will take some days (~80 hours).
After the benchmarks have finished run the following Rscript to generate a brief report in .pdf format for each number of benchmarked CPUs:
Rscript --vanilla report_generator.R
If you want to re-run the benchmarks use the -c flag in order to delete the created output files and benchmark results from the run before:
sh run_benchmarks -c
This instructions assume that docker is already installed and running. If not you should find most of the information on the Docker website
The first possibility is to pull the already configured BOOTABLE Docker image from docker hub with the following docker pull command.
docker pull maximilianhanussek/bootable
After the image has been downloaded you can start an interactive Docker container with the following command
docker run --rm -it maximilianhanussek/bootable
After you are in the container change into the home directory (root)
cd /root/
In the home directory you will find all the tools already installed. Before you start the benchmark you can run the install_check.sh script to be sure everything is fine
sh install_check.sh
In order to start the benchmark just run
sh run_benchmarks.sh
or enter some additional parameters via the provided flags.
The second option is to build the Docker container by yourself with the provided Dockerfile
. You can find it in the github repo in the docker
directory.
To build the Docker container by yourself clone this github repo, change into the docker
directory and run the following command:
docker build --tag bootable .
The build will take a couple of minutes (30-60) as the first step is to update the underlying operating sytem /(centOS7) to the latest version and installing all the required packages. Further some large datasets need to be downloaded and this depends on your network conectivity. Afterwards most of the tools have to be compiled and installed.
After the Docker image has been build you can start the benchmark with the same command like if you had pulled it from Docker Hub:
docker run --rm -it maximilianhanussek/bootable
cd /root/
sh run_benchmarks.sh
This instructions assume that singularity is already installed and working. If not you should find most of the information on the Singularity website
The first possibility is to pull the already configured and build container from singularity hub with the following singularity pull command.
singularity pull shub://MaximilianHanussek/BOOTABLE
After the image has been downloaded you can start an interactive Singularity container with the following command
sudo singularity shell bootable.simg
If you use a newer Singularity Version the image will come in the .sif
format and has to be run as the following
sudo singularity shell BOOTABLE_latest.sif
After you are in the container change to the root directory
cd /
In this directory you will find all the tools already installed. Before you start the benchmark you can run the install_check.sh script to be sure everything is fine
sh install_check.sh
In order to start the benchmark just run
sh run_benchmarks.sh
or enter some additional parameters via the provided flags.
The second option is to pull the already configured and build BOOTABLE Docker container from docker hub with the following singularity pull command.
singularity pull docker://maximilianhanussek/bootable
After the image has been downloaded into your current working directory and converted into a singularity container you can start an interactive Singularity container with the following command
sudo singularity shell MaximilianHanussek-BOOTABLE-master-latest.simg
After you are in the container, change to the root home directory /root/
cd /
In this directory you will find all the tools already installed. In order to start the benchmark just run
sh run_benchmarks.sh
or enter some additional parameters via the provided flags.
The third option is to build the Singularity container by yourself with the provided Singularity recipe
. You can find it in the github repo in the singularity
directory.
To build the Docker container by yourself clone this github repo, change into the singularity
directory and run the following command:
sudo singularity build bootable.simg singularity_bootable_recipe
The build will take a couple of minutes (30-60) as the first step is to update the underlying operating sytem /(centOS7) to the latest version and installing all the required packages. Further some large datasets need to be downloaded and this depends on your network conectivity. Afterwards most of the tools have to be compiled and installed.
After the Singularity image has been build you can start the benchmark with the same command like if you had pulled it from Singularity Hub:
sudo singularity shell bootable.simg
cd /
sh run_benchmarks.sh
This guide assumes that you have Ansible already installed and are familiar with it. The playbook has been tested with Ansible version 2.7.2 and is made for Redhat/Centos operating systems. In prinicpal the installation should work the same for other operating systems like Ubuntu but has not been tested yet. Especially the packages to install maybe have to be renamed.
A short warning in beforehand, the playbook will at first update all packages to latest available version, afterwards reboot the machine, install the epel package and the Development tolls package group. If you do not want to update all package versions please uncomment the resepctive tasks.
The first step to use the playbook is to clone or download the BOOTABLE github repository. You will find everything regarding ansible in the directory ansible
outgoing from the BOOTABLE repository root directory. Currently everything will be installed under the directory /home/centos/
if you want to install BOOTABLE in an other path please adjust it via the site.yml
file. Further you might need root/sudo access for the GROMACS tool to install some things in /usr/local/
, additional R packages and tensorflow via pip. We will also fix this in an upcoming version that no specific rights are needed at all.
The configuration of the hosts is done via the inventory
file which includes a template which you have to fill out.
;[BOOTABLE] bootable_machine ansible_host= ansible_user= ansible_ssh_private_key_file=
- REMOTE_IP: Here you have to specifiy the IP of the machine where you want to install BOOTABLE via the ansible playbook
- REMOTE_USER: Here you have to specifiy the user name of the remote host. Which could be root or any other user available on the remote machine
- ansible_ssh_private_key_file: The access to the machine is granted over an ssh key which should be already placed on the remote machine
You should also be able to install BOOTABLE via the ansible playbook on the same machine where you have downloaded it but than you could use the install_bootable.sh
directly described above in the Bare Metal section.
The site.yml
file specififies which roles and hosts should be used, which are currently all that are part of the inventory file. Further you will find the variable root_path
, that specifies the installation path prefix. Please change it to any path you want to. BOOTABLE will then be installed in this directory.
In the roles directory you will find one single role, BOOTABLE
. In this directory you will find the three important directories and their corresponding yaml files (main.yml
).
- defaults: In the file of the defaults directory is specified which packages need to be installed during the installation
- files: The directory contains the executable files required to run the BOOTABLE benchmarks and create the reports.
- tasks: The file of the tasks directory is the main part of the BOOTABLE installation process. It contains all instructions how to download, configure and install the offered tools and other things required.
In order to run the playbook after you have configured the inventory
file is the following. Change into the ansible
directory where you can see the inventory
and site.yml
file and execute the follwoing command:
ansible-playbook site.yml
After that ansible will start and BOOTABLE will be installed in an automated way. Ansible will also inform you if something went wrong. Please be patient during the GCC compilation task ( Compile gcc compiler version 7.3.0) as this can take a while (up to hours depending on the hardware).
Afterwards you can use BOOTABLE like in the following chapter.
In order to start a benchmark just stay in the BOOTABLE root directory and run the run_benchmarks.sh
script.
If you provide no flags the default parameters will be chosen like stated in the following.
-
c: Clean option, if you want to run an other new benchmark, the old results will be cleaned up and a backup will be placed in in the directory backed_up_benchmark_results if you confirm that. Per default the cleanup will not be done and the old date will be overwritten.
-
d: Dataset option, if you want to change the default dataset just set the flag
-d
and one of the corresponding keywords (large
,medium
,small
). The default parameters are the one of the medium group.
Parameter | small | medium | large |
---|---|---|---|
Dataset | ERR016155.filt.fastq/SRR741411.filt.fa | ERR016155.filt.fastq/ERR015528.filt.fa | ERR251006.filt.fastq |
Reference dataset | DRR001012.fa | DRR001025.fa | GRCh38_full_analysis_set_plus_decoy_hla.fa |
ClustalOmega | wgs.ANCA.1_200.fsa | wgs.ANCA.1_400.fsa | wgs.ANCA.1_500.fsa |
Tensorflow steps | 1000 | 2500 | 5000 |
GROMACS steps | 10000 | 30000 | 50000 |
-
p: Number of cores you want to use to run the benchmark. You can choose any arbitrary integer number or one of the three keywords
full
,half
andone
.- full: All cores which are accesible (Default)
- half: Half of the cores that are accesible
- one: Only one core (very long runtime)
-
r: Number of replicates you want to use. The more replicates the better is the chance to get a trustworthy result and regulate outliers. You can choose any integer value you want. The default value is 3.
-
s: The scaling mode automatically performs a benchmark of the specified tool(s) with all available cores, half of the available cores, quarter of the available cores and a single core. If you choose the scaling mode you do not have top specify the -p flag. Further choose on of the following keywords for the
-s
flag (default is medium)- large: Refers to to the large dataset option
- medium: Refers to the medium dataset option
- small: Refers to the small dataset option
-
t: Toolgroup you want to use. You can choose between
all
which will use all tools,genomics
which will use only genomic tools (Bowtie2, Velvet, IDBA, SPAdes),ml
which will only use Tensorflow andquant
which will only use GROMACS. The default option isall
. You can also choose just single tools. Currently these arebowtie2-build
,velvet
,idba
,tensorflow
,gromacs
,SPAdes
,clustalomega
, which are also the keywords that this flag exceppts.
Some execution examples:
sh run_benchmarks.sh
This would be the same if we would run the command:
sh run_benchmarks.sh -d medium -p full -r 3 -t all
If you want to run a very long benchmark with maximal CPU usage you could use:
sh run_benchmarks.sh -d large -p full -r 5 -t all
If you have already executed a benchmark and want to run a new one use the -c
flag:
sh run_benchmarks.sh -c -d medium -p full -r 3 -t all
If you just want to run some tensorflow benchmarks use for example:
sh run_benchmarks.sh -d large -p full -r 3 -t ml
If you want to test the scaling behaviour of a tool use the scaling mode with the -s
flag here with three replicates, medium size dataset and all tools that belong to the genomics group:
sh run_benchmarks.sh -r 3 -s medium -t genomics
If you want to use a tool that is not part of BOOTABLE you can include this one by using the -o
flag and specifying the path to the Toolfile which is explained more in detail in the section Generic tools wrapper:
sh run_benchmarks.sh -c -d small -o /absolute/path/to/toolfile.btbl -p full -r 1
You can also skip flags where you just want to use the default values.
After the benchmarks are finished you can create a Summary file in pdf format running the following command from BOOTABLE root directory (If you run it for the first time please run it as root or with sudo as some R packages needs to be installed):
Rscript --vanilla BOOTABLE_report_generator.R
If you used the scaling mode for the benchmarks we suggest to plot also the scaling plots with the following parameter
Rscript --vanilla BOOTABLE_report_generator.R scaling
The output file will be named scaling_plot_DATE_TIME.pdf
.
You will find the pdf file(s) in the BOOTABLE root directory.
For the case you want to benchmark a tool or other tools that are not part of BOOTABLE you can also do this and benefit of all already implemented BOOTABLE features, like automated scaling benchmarks, meaningful reports and plots or pre-selected datasets.
You just need to write a short textfile named as you like it and saved where you have access to it in the following manner (example for the bowtie2 index builder tool):
Toolname:bowtie2_build Dataset:Medium Command:bowtie2/bowtie2-2.3.4.2/bowtie2-build --threads $cores --seed 42 $reference benchmark_output/bowtie2_build/benchmark
For the lines Parameters Toolname
and Dataset
you can choose what ever you want it is just for you and has no direct impact on the calculations.
The Command
parameter is more important. Here you need to specify exactly how your tool has to be executed with some following guidelines:
-
If your tool has a parameter to choose the number of used CPU cores/thread we suggest to use the
$cores
variable as above, so this will be handled by BOOTABLE and the input parameter. You can also hard code it directly in your command but then the scaling mode will not work as expected. -
For the datasets you can do the same. You can use one of our already selected and available ones if they fit. In the example above we use the
$reference
variable and you can use any dataset available through BOOTABLE with the already known parameterssmall
,middle
orlarge
. But you can also specify any dataset or input data you want to use by giving the full path. Here a list of the datasets you will get with the corresponding variable: -
$reference: with -d flag small -> DRR001012.fa
-
$reference: with -d flag medium -> DRR001025.fa
-
$reference: with -d flag large -> GRCh38_full_analysis_set_plus_decoy_hla.fa
-
$dataset: with -d flag small -> ERR016155.filt.fastq
-
$dataset: with -d flag medium -> ERR016155.filt.fastq
-
$dataset: with -d flag large -> ERR251006.filt.fastq
-
If your tool produces any ouptut you can also let handle this by BOOTABLE. You only need to specify the output folder to point to the directory BOOTABLE/benchmark_output/
Toolname
as specified in the example. Please do not overwrite already existing directories of other tools already existing in thebenchmark_output
directory. -
And of course you need to install your tool and all required dependencies yourself and check everything is working fine.
BOOTABLE will produce different result files. Mainly the most interesting file is the benchmark_summary_\*.txt
file which you will find in the BOOTABLE root directory. The content looks like the following:
Replica_1 Bowtie2_build with 28 cores on dataset DRR001012 Mi 7. Nov 09:55:58 UTC 2018 real 211.35 user 3097.55 sys 14.13 Replica_1 Bowtie2_align with 28 cores on dataset ERR016155 Mi 7. Nov 09:59:29 UTC 2018 real 6.66 user 66.12 sys 7.77 Replica_1 Velveth with 28 cores on dataset ERR016155 Mi 7. Nov 09:59:36 UTC 2018 real 10.77 user 105.95 sys 6.74 Replica_1 Velvetg with 28 cores on dataset ERR016155 Mi 7. Nov 09:59:47 UTC 2018 real 64.57 user 155.62 sys 1.30 Replica_1 IDBA with 28 cores on dataset ERR016155 Mi 7. Nov 10:00:51 UTC 2018 real 110.99 user 2277.08 sys 10.19 Replica_1 Tensorflow with 28 cores on dataset cifar10 with 1000 Mi 7. Nov 10:02:42 UTC 2018 real 148.45 user 1934.74 sys 310.51 Replica_1 GROMACS with 28 cores on dataset adh_cubic calculating 10000 steps with CPU pinning enabled Mi 7. Nov 10:05:12 UTC 2018 real 79.11 user 1707.00 sys 465.29 Replica_1 SPAdes with 28 cores on dataset ERR016155 Mi 7. Nov 10:06:31 UTC 2018 real 440.31 user 4642.39 sys 784.05
Every block gives you information about which replica (Replica_1) of which tool with how many cores and on which dataset or stepsize took how long. The different runtimes are stated in seconds.
- real: The real walltime the tool used from start to end
- user: The used CPU time taking the number of cores into account
- sys: The amount of CPU time spent in the kernel for system calls
Further you will find the file benchmark_summary_\*.pdf
which gives you a more compact and statistical overview about the executed benchmarks. You will find this file only if you have run the BOOTABLE_report_generator.R
script as stated above.
If you also used the scaling parameter you will find the scaling_plots with the number of cores on the x-axis and the walltime on the y-axis. Both axes are on logarithmic scale to identify linear behaviour.
Also in the .pdf file all numbers are stated in seconds.
Futher you can find a detailed list, tool by tool, in the results directory. There you can also find the stdout output (benchmark_<tool>_output_<cores>.txt
) of the single tools and recapture the steps they have done.
Further you will find a summary of the underlying hardware/system where the benchmark has been executed, which flags of BOOTABLE has been used and how Bowtie2 and GROMACS has been compiled in the file bootable_system_info.txt
.
Further you find lots of interesting graphs and metadata collected and ilustrated by the tool nmon chart in the nmon_stats directory.
In order to uninstall BOOTABLE just run the following commands from a directory one level higher than the BOOTABLE directory:
sudo rm -rf BOOTABLE/ sudo rm -rf /usr/local/gromacs/ sudo pip uninstall tensorflow==1.4.0