Skip to content

Commit

Permalink
added fair directive to all processes
Browse files Browse the repository at this point in the history
  • Loading branch information
mattheww95 committed Jun 5, 2024
1 parent df48c71 commit bb409b9
Show file tree
Hide file tree
Showing 56 changed files with 56 additions and 0 deletions.
1 change: 1 addition & 0 deletions modules/local/abricate.nf
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ process ABRICATE {
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"


fair true
input:
tuple val(meta), path(assembly)

Expand Down
1 change: 1 addition & 0 deletions modules/local/bakta_annotate.nf
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ process BAKTA_ANNOTATE {
label 'process_high'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(fasta)
path db
Expand Down
1 change: 1 addition & 0 deletions modules/local/bandage_image.nf
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ process BANDAGE_IMAGE {
label "process_low"
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(gfa)

Expand Down
1 change: 1 addition & 0 deletions modules/local/bin_kraken2.nf
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,7 @@ process BIN_KRAKEN2{
cache 'deep' // ! Deep caching is required to not bungle up the later metadata updates on resumes
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(contigs), path(kraken_report), path(kraken_output)
val taxonomic_level
Expand Down
1 change: 1 addition & 0 deletions modules/local/check_ont.nf
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,7 @@ process CHECK_ONT{
// TODO perhaps reads should just be dedupped by header...
// TODO Awk would be faster...

fair true
input:
tuple val(meta), path(reads)

Expand Down
1 change: 1 addition & 0 deletions modules/local/checkm_lineagewf.nf
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ process CHECKM_LINEAGEWF {
label 'process_high'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(fasta)

Expand Down
1 change: 1 addition & 0 deletions modules/local/chopper_trim.nf
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ process CHOPPER_TRIM{
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"


fair true
input:
tuple val(meta), path(reads)

Expand Down
1 change: 1 addition & 0 deletions modules/local/combine_data.nf
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ process COMBINE_DATA{
label "process_low"
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(fastq_1), path(fastq_2), path(long_reads), path(assembly)

Expand Down
1 change: 1 addition & 0 deletions modules/local/ectyper.nf
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ process ECTYPER{
label 'process_medium'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(fasta)

Expand Down
1 change: 1 addition & 0 deletions modules/local/fastp_trim.nf
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ process FASTP_TRIM{
label "process_medium" // fastp uses very little memory in reality, but for duplicate analysis it is better to give it more memory
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(reads)

Expand Down
1 change: 1 addition & 0 deletions modules/local/fastqscan.nf
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@ process FASTQSCAN {
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"


fair true
input:
tuple val(meta), path(reads)

Expand Down
1 change: 1 addition & 0 deletions modules/local/flye_assemble.nf
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,7 @@ process FLYE_ASSEMBLE{
memory { task.memory * task.attempt}
// TODO check if --debug flag should be added to flye to actually turns off the debug logging?

fair true
input:
tuple val(meta), path(reads)
val mode
Expand Down
1 change: 1 addition & 0 deletions modules/local/gzip_files.nf
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ process GZIP_FILES {
label 'process_low'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(file_in)

Expand Down
1 change: 1 addition & 0 deletions modules/local/kat_hist.nf
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,7 @@ process KAT_HIST{
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"


fair true
input:
tuple val(meta), path(reads)
val long_reads_p
Expand Down
1 change: 1 addition & 0 deletions modules/local/kleborate.nf
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ process KLEBORATE {
label 'process_medium'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(fastas)

Expand Down
1 change: 1 addition & 0 deletions modules/local/kraken.nf
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ process KRAKEN {
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"


fair true
input:
tuple val(meta), path(contigs)
path db
Expand Down
1 change: 1 addition & 0 deletions modules/local/lissero.nf
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,7 @@ process LISSERO {
// TODO add in log message saying what went wrong with the sample
errorStrategy 'ignore' // TODO set a proper strategy once the issues with the mash parsing script are solved e.g. the ambiguous top hits

fair true
input:
tuple val(meta), path(fasta)

Expand Down
1 change: 1 addition & 0 deletions modules/local/locidex_extract.nf
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,7 @@ process LOCIDEX_EXTRACT {
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"


fair true
input:
tuple val(meta), path(fasta), path(db)

Expand Down
1 change: 1 addition & 0 deletions modules/local/locidex_report.nf
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ process LOCIDEX_REPORT {
label "process_low"
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(seq_store)

Expand Down
1 change: 1 addition & 0 deletions modules/local/locidex_search.nf
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ process LOCIDEX_SEARCH {
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"


fair true
input:
tuple val(meta), path(fasta), path(db)

Expand Down
1 change: 1 addition & 0 deletions modules/local/mash_estimate.nf
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ process MASH_ESTIMATE{
tag "${prefix}"
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(reads)
val long_reads_p
Expand Down
1 change: 1 addition & 0 deletions modules/local/mash_paste.nf
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ process MASH_PASTE{
label 'process_low'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(sketches) // getting all output sketches

Expand Down
1 change: 1 addition & 0 deletions modules/local/mash_screen.nf
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,7 @@ process MASH_SCREEN {
label 'process_medium'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(reads)
path sequences_sketch
Expand Down
1 change: 1 addition & 0 deletions modules/local/mash_sketch.nf
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ process MASH_SKETCH{
label 'process_low'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(fasta), val(comment)

Expand Down
1 change: 1 addition & 0 deletions modules/local/medaka_polish.nf
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ process MEDAKA_POLISH{
beforeScript 'rm -rf medaka' // Same reasoning as above
errorStrategy { sleep(Math.pow(2, task.attempt) * 200 as long); return 'retry' } // May be having issues with medaka model copying

fair true
input:
tuple val(meta), path(reads), path(assembly)
val model
Expand Down
1 change: 1 addition & 0 deletions modules/local/minimap2_index.nf
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ process MINIMAP2_INDEX{
label 'process_low'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(contigs)

Expand Down
1 change: 1 addition & 0 deletions modules/local/minimap2_map.nf
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ process MINIMAP2_MAP {
label 'process_medium'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(reads), path(index), path(contigs)
val paf_out
Expand Down
1 change: 1 addition & 0 deletions modules/local/mlst.nf
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ process MLST {
label "process_low"
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(fasta)

Expand Down
1 change: 1 addition & 0 deletions modules/local/mob_recon.nf
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ process MOBSUITE_RECON {

container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(fasta)

Expand Down
1 change: 1 addition & 0 deletions modules/local/parse_fastp.nf
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ process PARSE_FASTP{
tag "$meta.id"
label "process_single"

fair true
input:
tuple val(meta), val(json)

Expand Down
1 change: 1 addition & 0 deletions modules/local/parse_kat.nf
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ process PARSE_KAT {
label "process_single"
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), val(json)

Expand Down
1 change: 1 addition & 0 deletions modules/local/parse_kraken.nf
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ process PARSE_KRAKEN {
label "process_low"
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(kraken_report)

Expand Down
1 change: 1 addition & 0 deletions modules/local/parse_mash.nf
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ process PARSE_MASH{
label "process_low"
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(mash_screen)
val run_mode
Expand Down
1 change: 1 addition & 0 deletions modules/local/pilon_polish.nf
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,7 @@ process PILON_POLISH {


// NOTE some inputs not used, they are just there to maintain a common interface with other polishing programs
fair true
input:
tuple val(meta), path(reads), path(contigs), path(bam), path(bai) // may need to copy the bai into the work dir

Expand Down
1 change: 1 addition & 0 deletions modules/local/pilon_polisher.nf
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ process PILON_ITER {
memory {task.memory * task.attempt}
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(reads), path(contigs)

Expand Down
1 change: 1 addition & 0 deletions modules/local/quast_assembly.nf
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ process QUAST {
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"


fair true
input:
tuple val(meta), path(contigs), path(trimmed_reads)

Expand Down
1 change: 1 addition & 0 deletions modules/local/racon_polish.nf
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ process RACON_POLISH {
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"
afterScript "rm ${input_reads}"

fair true
input:
tuple val(meta), path(reads), path(sam), path(contigs)

Expand Down
1 change: 1 addition & 0 deletions modules/local/read_summary.nf
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ process READ_SCAN{
tag "${meta.id}"
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(reads), path(l_reads)

Expand Down
1 change: 1 addition & 0 deletions modules/local/remove_contaminants.nf
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ process REMOVE_CONTAMINANTS {
label 'process_medium'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(reads)
path contaminant_fa
Expand Down
1 change: 1 addition & 0 deletions modules/local/report.nf
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,7 @@ process REPORT{
tag "Report Generation"
label "process_single"

fair true
input:
val test_in

Expand Down
1 change: 1 addition & 0 deletions modules/local/report_aggregate.nf
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ process REPORT_AGGREGATE{
label 'process_medium'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
path summary_report

Expand Down
1 change: 1 addition & 0 deletions modules/local/sam_to_bam.nf
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ process SAM_TO_BAM{
label 'process_low'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(sam)

Expand Down
1 change: 1 addition & 0 deletions modules/local/samplesheet_check.nf
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ process SAMPLESHEET_CHECK {
'https://depot.galaxyproject.org/singularity/python:3.8.3' :
'quay.io/biocontainers/python:3.8.3' }"

fair true
input:
path samplesheet

Expand Down
1 change: 1 addition & 0 deletions modules/local/select_pointfinder.nf
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@ process IDENTIFY_POINTDB {
tag "$meta.id"
label "process_single"

fair true
input:
tuple val(meta), val(species)

Expand Down
1 change: 1 addition & 0 deletions modules/local/seqkit_filter.nf
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ process SEQKIT_FILTER {
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"


fair true
input:
tuple val(meta), path(contigs), path(reads)
val min_length
Expand Down
1 change: 1 addition & 0 deletions modules/local/seqkit_stats.nf
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ process SEQKIT_STATS {
label 'process_single'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(contigs), path(trimmed_reads)

Expand Down
1 change: 1 addition & 0 deletions modules/local/seqtk_fasta_fastq.nf
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ process SEQTK_FASTA_FASTQ{
label 'process_low'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(fasta)

Expand Down
1 change: 1 addition & 0 deletions modules/local/seqtk_sample.nf
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ process SEQTK_SAMPLE{
label "process_low"
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(reads), val(sample_fraction)

Expand Down
1 change: 1 addition & 0 deletions modules/local/seqtk_size.nf
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,7 @@ process SEQTK_SIZE{
label "process_low"
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(reads)

Expand Down
1 change: 1 addition & 0 deletions modules/local/shigatyper.nf
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ process SHIGATYPER{
label 'process_low'
container "${workflow.containerEngine == 'singularity' || workflow.containerEngine == 'apptainer' ? task.ext.parameters.get('singularity') : task.ext.parameters.get('docker')}"

fair true
input:
tuple val(meta), path(reads) // in mkkondo it will only work with assemblies

Expand Down
Loading

0 comments on commit bb409b9

Please sign in to comment.