Thanks @BenKaehler and @SoilRotifer for your help with this. I just have a couple more downloads that are still not working. The first is under 100,000 sequences and I used the following script:
qiime rescript get-ncbi-data *
** --p-query ‘txid189478[ORGN] OR txid147549[ORGN] OR txid205932[ORGN] OR txid147537[ORGN] OR txid451866[ORGN] OR txid129384[ORGN] OR txid136265[ORGN] OR txid2283618[ORGN] OR txid112252[ORGN] AND (LSU[TITLE] OR 28S[TITLE] OR large ribosomal subunit[TITLE] NOT uncultured[TITLE] NOT unidentified[TITLE] NOT unclassified[TITLE] NOT unverified[TITLE])’ *
** --o-sequences ncbi-LSU-seqs-fungi4.qza **
** --o-taxonomy ncbi-LSU-taxonomy-fungi4.qza **
** --verbose**
when I try to run this i get the following error message:
Traceback (most recent call last):
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/q2cli/commands.py”, line 329, in call
results = action(**arguments)
File “”, line 2, in get_ncbi_data
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/qiime2/sdk/action.py”, line 245, in bound_callable
output_types, provenance)
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/qiime2/sdk/action.py”, line 390, in callable_executor
output_views = self._callable(**view_args)
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/rescript/ncbi.py”, line 89, in get_ncbi_data
query, logging_level, n_jobs, request_lock, _entrez_delay)
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/rescript/ncbi.py”, line 365, in get_nuc_for_query
for chunk in range(0, expected_num_records, 5000))
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/joblib/parallel.py”, line 1044, in call
while self.dispatch_one_batch(iterator):
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/joblib/parallel.py”, line 859, in dispatch_one_batch
self._dispatch(tasks)
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/joblib/parallel.py”, line 777, in _dispatch
job = self._backend.apply_async(batch, callback=cb)
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/joblib/_parallel_backends.py”, line 208, in apply_async
result = ImmediateResult(func)
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/joblib/_parallel_backends.py”, line 572, in init
self.results = batch()
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/joblib/parallel.py”, line 263, in call
for func, args, kwargs in self.items]
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/joblib/parallel.py”, line 263, in
for func, args, kwargs in self.items]
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/rescript/ncbi.py”, line 342, in _get_query_chunk
raise RuntimeError(‘Download did not finish. Reason unknown.’)
RuntimeError: Download did not finish. Reason unknown.
Plugin error from rescript:
Download did not finish. Reason unknown.
See above for debug info.
The second download is larger at 123,000 sequences and I used the following script:
qiime rescript get-ncbi-data *
** --p-query ‘txid6960[ORGN] AND (LSU[TITLE] OR 28S[TITLE] OR large ribosomal subunit[TITLE] NOT uncultured[TITLE] NOT unidentified[TITLE] NOT unclassified[TITLE] NOT unverified[TITLE])’ *
** --p-n-jobs 5 **
** --o-sequences ncbi-LSU-seqs-metazoa1.qza **
** --o-taxonomy ncbi-LSU-taxonomy-metazoa1.qza **
** --verbose**
For this one it looks like I get a slightly different error message:
Traceback (most recent call last):
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/q2cli/commands.py”, line 329, in call
results = action(**arguments)
File “”, line 2, in get_ncbi_data
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/qiime2/sdk/action.py”, line 245, in bound_callable
output_types, provenance)
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/qiime2/sdk/action.py”, line 390, in callable_executor
output_views = self._callable(**view_args)
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/rescript/ncbi.py”, line 89, in get_ncbi_data
query, logging_level, n_jobs, request_lock, _entrez_delay)
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/rescript/ncbi.py”, line 365, in get_nuc_for_query
for chunk in range(0, expected_num_records, 5000))
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/joblib/parallel.py”, line 1054, in call
self.retrieve()
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/joblib/parallel.py”, line 933, in retrieve
self._output.extend(job.get(timeout=self.timeout))
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/site-packages/joblib/_parallel_backends.py”, line 542, in wrap_future_result
return future.result(timeout=timeout)
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/concurrent/futures/_base.py”, line 432, in result
return self.__get_result()
File “/mnt/scratch/nodelete/sbiat4/qiime_conda/envs/rescript/lib/python3.6/concurrent/futures/_base.py”, line 384, in __get_result
raise self._exception
RuntimeError: Download did not finish. Reason unknown.
Plugin error from rescript:
Download did not finish. Reason unknown.
Any thoughts on this would be greatly appreciated, I am so close to having all the database downloaded!
Anya