_____________
< mspipeline1 >
-------------
\
___......__ _ \
_.-' ~-_ _.=a~~-_
--=====-.-.-_----------~ .--. _ -.__.-~ ( ___===>
'''--...__ ( \ \\\ { ) _.-~
=_ ~_ \\-~~~//~~~~-=-~
|-=-~_ \\ \\
|_/ =. ) ~}
|} ||
// ||
_// {{
'='~' \\_ =
~~'
If you want to use fragpipe using the command line interface, then this is the tool for you.
This pipeline takes 1) a list of .d files and 2) a list of fasta-amino acid files and outputs sane protein calls with abundances. It uses philosopher database and fragpipe to do the job. The snakemake pipeline maintains a nice output file tree.
Because it makes sure that all outputs are updated when you change input-parameters. It also yells at you if something fails, and hopefully makes it a bit easier to find the error.
- Prerequisites:
- Preferably a HPC system, or a beefy local workstation.
- An conda package manager on that system. (We recommend miniforge)
-
Clone this repo on the HPC/workstation where you want to work.
git clone https://github.com/cmkobel/mspipeline1.git && cd mspipeline1
-
If you don't already have an environment with snakemake and mamba installed, use the following command to install a "snakemake" environment with the bundled environment file:
conda env create -f environment.yaml -n mspipeline1
This environment can then be activated by typing
conda activate mspipeline1
-
If needed, tweak the profiles/slurm/ configuration so that it matches your execution environment. There is a profile for local execution without a job management system (profiles/local/) as well as a few profiles for different HPC environments like PBS and SLURM.
The file config_template.yaml contains all the parameters needed to run this pipeline. You should change the parameters to reflect your sample batch.
Because nesvilab do not make their executables immediately publicly available, you need to tell the pipeline where to find them on your system. Update addresses for the keys philosopher_executable
, msfragger_jar
, ionquant_jar
and fragpipe_executable
which can be downloaded here, here, here and here, respectively.
Currently the pipeline is only tested on the input of .d-files (agilent/bruker): Create an item in batch_parameters where you define key d_base
which is the base directory where all .d-files reside. Define key database_glob
which is a path (or glob) to the fasta-amino acid files that you want to include in the target protein database.
Define items under the samples
key which link sample names to the .d-files.
Lastly, set the batch
key to point at the batch that you want to run.
Finally, run the pipeline in your command line with:
$ snakemake --profile profiles/slurm/
Below is visualization of the workflow graph:
This pipeline might involve an R-markdown performing trivial QC. Also, a test data set that accelerates the development cycle. 🚴♀️