@@ -184,15 +184,16 @@ def fetch_uniprot_data(organism_id):
184
184
def fetch_uniprot_reference_proteome_data () -> list :
185
185
"""Single URL request for Uniprot proteome data."""
186
186
file_path = Path (RAW_DATA_DIR ) / f"{ PROTEOMES_FILENAME } .{ UNIPROT_DESIRED_FORMAT } "
187
- all_proteomes_query = "%28*%29"
188
- # filtered_proteomes_query = (
189
- # "((superkingdom:Bacteria)+OR+(superkingdom:Archaea))+AND+((proteome_type:1)+OR+(proteome_type:2))"
190
- # )
187
+ # all_proteomes_query = "%28*%29"
188
+ filtered_proteomes_query = (
189
+ "((superkingdom:Bacteria)+OR+(superkingdom:Archaea))+AND+((proteome_type:1)+OR+(proteome_type:2))"
190
+ )
191
+ import pdb ; pdb .set_trace ()
191
192
192
193
url = construct_query_url (
193
194
UNIPROT_REFERENCE_PROTEOMES_URL ,
194
195
UNIPROT_DESIRED_FORMAT ,
195
- all_proteomes_query ,
196
+ filtered_proteomes_query ,
196
197
UNIPROT_REFERENCE_PROTEOMES_FIELDS ,
197
198
UNIPROT_SIZE ,
198
199
)
@@ -311,7 +312,7 @@ def run_uniprot_api_parallel(
311
312
# If show_status is True, use process_map to display a progress bar
312
313
if show_status :
313
314
process_map (
314
- fetch_func , taxa_id_common_with_proteomes_list , max_workers = workers
315
+ fetch_func , taxa_id_common_with_proteomes_list , max_workers = workers , chunksize = CHUNK_SIZE_PER_WORKER
315
316
)
316
317
else :
317
318
# Set up a pool of worker processes without a progress bar
0 commit comments