When I have a large number of parallel tasks at the same time, the following error will be reported:
About 1000 parallel samples,Taking one sample as an example, run the command as follows:
qiime tools import --type 'SampleData[PairedEndSequencesWithQuality]' --input-path Process/A5_1/A5_1_list.tsv --output-path Process/A5_1/A5_1_demux.qza --input-format PairedEndFastqManifestPhred33V2
qiime cutadapt trim-paired --i-demultiplexed-sequences Process/A5_1/A5_1_demux.qza --p-minimum-length 100 --p-overlap 17 --p-front-f CCTAGGTGNTTAWGCAG --p-front-r GATGACHAACCTAATCC --o-trimmed-sequences Process/A5_1/A5_1_demux-trimmed.qza --verbose
qiime vsearch merge-pairs --i-demultiplexed-seqs Process/A5_1/A5_1_demux-trimmed.qza --o-merged-sequences Process/A5_1/A5_1_merged.qza --o-unmerged-sequences Process/A5_1/A5_1_unmerged.qza
qiime quality-filter q-score --i-demux Process/A5_1/A5_1_merged.qza --o-filtered-sequences Process/A5_1/A5_1_filtered.qza --o-filter-stats Process/A5_1/A5_1_filter_stats.qza --p-min-quality 20 --p-quality-window 100 --p-max-ambiguous 5
qiime tools export --input-path Process/A5_1/A5_1_filtered.qza --output-path Process/A5_1/A5_1_demux-filtered-export
qiime demux summarize --i-data Process/A5_1/A5_1_filtered.qza --o-visualization Process/A5_1/A5_1_Clean-data-length.qzv
/public1/Softwares/vsearch-2.3.4/bin/vsearch --uchime_ref Process/A5_1/A5_1_demux-filtered-export/A5_1_L001_R1_001.fasta --db /mnt/dgfs/database/BLAST_db/RDP_GOLD/rdp_gold.fa --nonchimeras Process/A5_1/A5_1_demux-filtered-export/A5_1_no_chimera.fasta --threads 4 > Process/A5_1/A5_1_demux-filtered-export/A5_1_no_chimera.fasta.log 2>&1
During the whole process, the following error messages may be triggered when running one of the steps (each step may be possible):
Traceback (most recent call last):
File "/opt/conda/envs/qiime2-amplicon-2024.10/lib/python3.10/site-packages/q2cli/builtin/tools.py", line 852, in import
artifact = qiime2.sdk.Artifact.import_data(
File "/opt/conda/envs/qiime2-amplicon-2024.10/lib/python3.10/site-packages/qiime2/sdk/result.py", line 345, in import_data
provenance_capture = archive.ImportProvenanceCapture(format, md5sums)
File "/opt/conda/envs/qiime2-amplicon-2024.10/lib/python3.10/site-packages/qiime2/core/archive/provenance.py", line 525, in init
super().init()
File "/opt/conda/envs/qiime2-amplicon-2024.10/lib/python3.10/site-packages/qiime2/core/archive/provenance.py", line 296, in init
self._build_paths()
File "/opt/conda/envs/qiime2-amplicon-2024.10/lib/python3.10/site-packages/qiime2/core/archive/provenance.py", line 303, in _build_paths
self.path = qiime2.core.path.ProvenancePath()
File "/opt/conda/envs/qiime2-amplicon-2024.10/lib/python3.10/site-packages/qiime2/core/path.py", line 138, in new
cache = get_cache()
File "/opt/conda/envs/qiime2-amplicon-2024.10/lib/python3.10/site-packages/qiime2/core/cache.py", line 113, in get_cache
_CACHE.temp_cache = Cache()
File "/opt/conda/envs/qiime2-amplicon-2024.10/lib/python3.10/site-packages/qiime2/core/cache.py", line 417, in init
self.__init(path=path, process_pool_lifespan=process_pool_lifespan)
File "/opt/conda/envs/qiime2-amplicon-2024.10/lib/python3.10/site-packages/qiime2/core/cache.py", line 457, in __init
with self.lock:
File "/opt/conda/envs/qiime2-amplicon-2024.10/lib/python3.10/site-packages/qiime2/core/cache.py", line 304, in exit
self.flufl_lock.unlock()
File "/opt/conda/envs/qiime2-amplicon-2024.10/lib/python3.10/site-packages/flufl/lock/_lockfile.py", line 417, in unlock
raise NotLockedError('Already unlocked')
flufl.lock._lockfile.NotLockedError: Already unlocked
This problem is prone to occur when running large sample sizes, and it may not occur in all samples. When running around 1000 parallel samples, this error may occur in a small subset of samples, and sometimes the exception may not be displayed after re running.