{\rtf1\ansi\ansicpg1252\cocoartf2639 \cocoatextscaling0\cocoaplatform0{\fonttbl\f0\fswiss\fcharset0 Helvetica;} {\colortbl;\red255\green255\blue255;} {\*\expandedcolortbl;;} \margl1440\margr1440\vieww17880\viewh8720\viewkind0 \pard\tx720\tx1440\tx2160\tx2880\tx3600\tx4320\tx5040\tx5760\tx6480\tx7200\tx7920\tx8640\pardirnatural\partightenfactor0 \f0\fs24 \cf0 Enable Galaxy reports authentification \ Checking /export...\ Disable Galaxy Interactive Environments. Start with --privileged to enable IE's.\ Unlinking stale socket /var/run/supervisor.sock\ Starting postgres\ postgresql: started\ Checking if database is up and running\ Database connected\ Starting cron\ cron: started\ Starting ProFTP\ proftpd: started\ Starting Galaxy reports webapp\ reports: started\ Starting nodejs\ galaxy:galaxy_nodejs_proxy: started\ Starting condor\ condor: started\ Starting slurmctld\ Starting slurmd\ Creating admin user admin with key fakekey and password password if not existing\ ==> /home/galaxy/logs/handler0.log <==\ galaxy.workflow.scheduling_manager INFO 2022-08-08 23:23:57,141 [pN:handler0,p:254,tN:MainThread] No workflow schedulers plugin config file defined, using default scheduler.\ galaxy.web_stack.handlers INFO 2022-08-08 23:23:57,141 [pN:handler0,p:254,tN:MainThread] WorkflowSchedulingManager: No job handler assignment method is set, defaulting to 'HANDLER_ASSIGNMENT_METHODS.DB_SKIP_LOCKED', set the `assign_with` attribute on to override the default\ galaxy.workflow.scheduling_manager DEBUG 2022-08-08 23:23:57,142 [pN:handler0,p:254,tN:MainThread] Starting workflow schedulers\ galaxy.jobs.handler INFO 2022-08-08 23:23:57,471 [pN:handler0,p:254,tN:WorkflowRequestMonitor.monitor_thread] Handler job grabber initialized with 'HANDLER_ASSIGNMENT_METHODS.DB_SKIP_LOCKED' assignment method for handler 'handler0', tag(s): handlers, None\ galaxy.queue_worker INFO 2022-08-08 23:23:57,495 [pN:handler0,p:254,tN:MainThread] Binding and starting galaxy control worker for handler0\ galaxy.queue_worker INFO 2022-08-08 23:23:57,563 [pN:handler0,p:254,tN:MainThread] Queuing async task rebuild_toolbox_search_index for handler0.\ galaxy.app INFO 2022-08-08 23:23:58,152 [pN:handler0,p:254,tN:MainThread] Galaxy app startup finished (14503.039 ms)\ galaxy.web_stack INFO 2022-08-08 23:23:58,153 [pN:handler0,p:254,tN:MainThread] Galaxy server instance 'handler0' is running\ galaxy.queue_worker INFO 2022-08-08 23:23:59,066 [pN:handler0,p:254,tN:Thread-1] Instance 'handler0' received 'rebuild_toolbox_search_index' task, executing now.\ galaxy.queue_worker DEBUG 2022-08-08 23:23:59,066 [pN:handler0,p:254,tN:Thread-1] App is not a webapp, not building a search index\ \ ==> /home/galaxy/logs/handler1.log <==\ galaxy.web_stack.handlers INFO 2022-08-08 23:23:57,127 [pN:handler1,p:255,tN:MainThread] WorkflowSchedulingManager: No job handler assignment method is set, defaulting to 'HANDLER_ASSIGNMENT_METHODS.DB_SKIP_LOCKED', set the `assign_with` attribute on to override the default\ galaxy.workflow.scheduling_manager DEBUG 2022-08-08 23:23:57,127 [pN:handler1,p:255,tN:MainThread] Starting workflow schedulers\ galaxy.jobs.handler INFO 2022-08-08 23:23:57,393 [pN:handler1,p:255,tN:WorkflowRequestMonitor.monitor_thread] Handler job grabber initialized with 'HANDLER_ASSIGNMENT_METHODS.DB_SKIP_LOCKED' assignment method for handler 'handler1', tag(s): handlers, None\ galaxy.queue_worker INFO 2022-08-08 23:23:57,413 [pN:handler1,p:255,tN:MainThread] Binding and starting galaxy control worker for handler1\ galaxy.queue_worker INFO 2022-08-08 23:23:57,599 [pN:handler1,p:255,tN:MainThread] Queuing async task rebuild_toolbox_search_index for handler1.\ galaxy.model.database_heartbeat DEBUG 2022-08-08 23:23:57,798 [pN:handler1,p:255,tN:database_heartbeart_handler1.thread] handler1 is config watcher\ galaxy.app INFO 2022-08-08 23:23:58,203 [pN:handler1,p:255,tN:MainThread] Galaxy app startup finished (14548.751 ms)\ galaxy.web_stack INFO 2022-08-08 23:23:58,203 [pN:handler1,p:255,tN:MainThread] Galaxy server instance 'handler1' is running\ galaxy.queue_worker INFO 2022-08-08 23:23:59,121 [pN:handler1,p:255,tN:Thread-1] Instance 'handler1' received 'rebuild_toolbox_search_index' task, executing now.\ galaxy.queue_worker DEBUG 2022-08-08 23:23:59,121 [pN:handler1,p:255,tN:Thread-1] App is not a webapp, not building a search index\ \ ==> /home/galaxy/logs/reports.log <==\ galaxy.web.framework.base DEBUG 2022-08-08 23:24:02,523 [pN:main,p:1799,tN:MainThread] Enabling 'history' controller, class: History\ galaxy.web.framework.base DEBUG 2022-08-08 23:24:02,525 [pN:main,p:1799,tN:MainThread] Enabling 'users' controller, class: Users\ galaxy.web.framework.base DEBUG 2022-08-08 23:24:02,526 [pN:main,p:1799,tN:MainThread] Enabling 'jobs' controller, class: Jobs\ galaxy.webapps.util DEBUG 2022-08-08 23:24:02,532 [pN:main,p:1799,tN:MainThread] Enabling 'paste.httpexceptions' middleware\ galaxy.webapps.util DEBUG 2022-08-08 23:24:02,537 [pN:main,p:1799,tN:MainThread] Enabling 'RecursiveMiddleware' middleware\ galaxy.webapps.util DEBUG 2022-08-08 23:24:02,549 [pN:main,p:1799,tN:MainThread] Enabling 'ErrorMiddleware' middleware\ galaxy.webapps.util DEBUG 2022-08-08 23:24:02,551 [pN:main,p:1799,tN:MainThread] Enabling 'TransLogger' middleware\ galaxy.webapps.util DEBUG 2022-08-08 23:24:02,553 [pN:main,p:1799,tN:MainThread] Enabling 'XForwardedHostMiddleware' middleware\ Starting server in PID 1799.\ serving on http://127.0.0.1:9001\ \ ==> /home/galaxy/logs/slurmctld.log <==\ [2022-08-08T23:24:12.703] No trigger state file (/tmp/slurm/trigger_state.old) to recover\ [2022-08-08T23:24:12.703] _preserve_plugins: backup_controller not specified\ [2022-08-08T23:24:12.703] Reinitializing job accounting state\ [2022-08-08T23:24:12.704] cons_res: select_p_reconfigure\ [2022-08-08T23:24:12.704] cons_res: select_p_node_init\ [2022-08-08T23:24:12.704] cons_res: preparing for 1 partitions\ [2022-08-08T23:24:12.705] Running as primary controller\ [2022-08-08T23:24:12.709] No parameter for mcs plugin, default values set\ [2022-08-08T23:24:12.709] mcs: MCSParameters = (null). ondemand set.\ [2022-08-08T23:24:15.730] SchedulerParameters=default_queue_depth=100,max_rpc_cnt=0,max_sched_time=2,partition_job_depth=0,sched_max_job_start=0,sched_min_interval=2\ \ ==> /home/galaxy/logs/slurmd.log <==\ [2022-08-06T00:41:30.214] slurmd version 17.11.2 started\ [2022-08-06T00:41:30.217] slurmd started on Sat, 06 Aug 2022 00:41:30 +0000\ [2022-08-06T00:41:30.221] CPUs=4 Boards=1 Sockets=1 Cores=4 Threads=1 Memory=7851 TmpDisk=59819 Uptime=1007 CPUSpecList=(null) FeaturesAvail=(null) FeaturesActive=(null)\ [2022-08-08T23:24:13.098] error: Domain socket directory /var/spool/slurmd: No such file or directory\ [2022-08-08T23:24:13.116] Node reconfigured socket/core boundaries SocketsPerBoard=4:1(hw) CoresPerSocket=1:4(hw)\ [2022-08-08T23:24:13.117] Message aggregation disabled\ [2022-08-08T23:24:13.122] CPU frequency setting not configured for this node\ [2022-08-08T23:24:13.129] slurmd version 17.11.2 started\ [2022-08-08T23:24:13.134] slurmd started on Mon, 08 Aug 2022 23:24:13 +0000\ [2022-08-08T23:24:13.138] CPUs=4 Boards=1 Sockets=1 Cores=4 Threads=1 Memory=7851 TmpDisk=59819 Uptime=1124 CPUSpecList=(null) FeaturesAvail=(null) FeaturesActive=(null)\ \ ==> /home/galaxy/logs/uwsgi.log <==\ Mon Aug 8 23:24:13 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:14 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:15 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:16 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:17 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:18 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:19 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:20 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:21 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:22 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:23 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:24 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ 172.17.0.1 - - [08/Aug/2022:23:24:24 +0000] "GET / HTTP/1.1" 200 - "-" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2275|app: 0|req: 1/2] 172.17.0.1 () \{58 vars in 1360 bytes\} [Mon Aug 8 23:24:24 2022] GET / => generated 95099 bytes in 692 msecs (HTTP/1.1 200) 8 headers in 393 bytes (1 switches on core 0)\ Mon Aug 8 23:24:25 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ 172.17.0.1 - - [08/Aug/2022:23:24:25 +0000] "GET /api/webhooks HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2275|app: 0|req: 2/3] 172.17.0.1 () \{58 vars in 1312 bytes\} [Mon Aug 8 23:24:25 2022] GET /api/webhooks => generated 7467 bytes in 30 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 1)\ 172.17.0.1 - - [08/Aug/2022:23:24:25 +0000] "GET /api/webhooks HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2275|app: 0|req: 3/4] 172.17.0.1 () \{58 vars in 1312 bytes\} [Mon Aug 8 23:24:25 2022] GET /api/webhooks => generated 7467 bytes in 44 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 2)\ 172.17.0.1 - - [08/Aug/2022:23:24:25 +0000] "GET /api/webhooks HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2275|app: 0|req: 4/5] 172.17.0.1 () \{58 vars in 1312 bytes\} [Mon Aug 8 23:24:25 2022] GET /api/webhooks => generated 7467 bytes in 31 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 3)\ 172.17.0.1 - - [08/Aug/2022:23:24:25 +0000] "GET /welcome HTTP/1.1" 302 281 "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2279|app: 0|req: 3/6] 172.17.0.1 () \{58 vars in 1392 bytes\} [Mon Aug 8 23:24:25 2022] GET /welcome => generated 281 bytes in 184 msecs (HTTP/1.1 302) 4 headers in 173 bytes (1 switches on core 3)\ 172.17.0.1 - - [08/Aug/2022:23:24:25 +0000] "GET /history/current_history_json HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2279|app: 0|req: 3/7] 172.17.0.1 () \{58 vars in 1344 bytes\} [Mon Aug 8 23:24:25 2022] GET /history/current_history_json => generated 548 bytes in 339 msecs (HTTP/1.1 200) 7 headers in 207 bytes (1 switches on core 2)\ 172.17.0.1 - - [08/Aug/2022:23:24:26 +0000] "GET /api/users/current HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2279|app: 0|req: 5/8] 172.17.0.1 () \{58 vars in 1322 bytes\} [Mon Aug 8 23:24:26 2022] GET /api/users/current => generated 82 bytes in 78 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 0)\ 172.17.0.1 - - [08/Aug/2022:23:24:26 +0000] "GET /api/histories/f597429621d6eb2b/contents?limit=500&offset=0&order=hid&v=dev&q=deleted&q=purged&q=visible&qv=False&qv=False&qv=True HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2275|app: 0|req: 5/9] 172.17.0.1 () \{58 vars in 1545 bytes\} [Mon Aug 8 23:24:26 2022] GET /api/histories/f597429621d6eb2b/contents?limit=500&offset=0&order=hid&v=dev&q=deleted&q=purged&q=visible&qv=False&qv=False&qv=True => generated 2 bytes in 158 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 0)\ Mon Aug 8 23:24:26 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ 172.17.0.1 - - [08/Aug/2022:23:24:26 +0000] "GET /history/set_as_current?id=f597429621d6eb2b HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2279|app: 0|req: 5/10] 172.17.0.1 () \{56 vars in 1319 bytes\} [Mon Aug 8 23:24:26 2022] GET /history/set_as_current?id=f597429621d6eb2b => generated 548 bytes in 173 msecs (HTTP/1.1 200) 7 headers in 207 bytes (1 switches on core 1)\ 172.17.0.1 - - [08/Aug/2022:23:24:26 +0000] "GET /api/histories/f597429621d6eb2b?keys=size%2Cnon_ready_jobs%2Ccontents_active%2Chid_counter HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2275|app: 0|req: 7/11] 172.17.0.1 () \{58 vars in 1465 bytes\} [Mon Aug 8 23:24:26 2022] GET /api/histories/f597429621d6eb2b?keys=size%2Cnon_ready_jobs%2Ccontents_active%2Chid_counter => generated 112 bytes in 178 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 1)\ 172.17.0.1 - - [08/Aug/2022:23:24:26 +0000] "GET /api/histories/f597429621d6eb2b?view=betawebclient HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2275|app: 0|req: 7/12] 172.17.0.1 () \{56 vars in 1333 bytes\} [Mon Aug 8 23:24:26 2022] GET /api/histories/f597429621d6eb2b?view=betawebclient => generated 577 bytes in 408 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 2)\ Mon Aug 8 23:24:27 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:28 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:29 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:30 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ 172.17.0.1 - - [08/Aug/2022:23:24:30 +0000] "GET /api/genomes HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ 172.17.0.1 - - [08/Aug/2022:23:24:30 +0000] "GET /api/histories?view=summary HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2279|app: 0|req: 8/13] 172.17.0.1 () \{56 vars in 1258 bytes\} [Mon Aug 8 23:24:30 2022] GET /api/genomes => generated 55436 bytes in 85 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 2)\ [pid: 2275|app: 0|req: 9/14] 172.17.0.1 () \{56 vars in 1287 bytes\} [Mon Aug 8 23:24:30 2022] GET /api/histories?view=summary => generated 255 bytes in 91 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 3)\ 172.17.0.1 - - [08/Aug/2022:23:24:30 +0000] "GET /api/users/current HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2279|app: 0|req: 8/15] 172.17.0.1 () \{56 vars in 1270 bytes\} [Mon Aug 8 23:24:30 2022] GET /api/users/current => generated 82 bytes in 103 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 3)\ 172.17.0.1 - - [08/Aug/2022:23:24:30 +0000] "GET /history/current_history_json HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2275|app: 0|req: 9/16] 172.17.0.1 () \{56 vars in 1292 bytes\} [Mon Aug 8 23:24:30 2022] GET /history/current_history_json => generated 548 bytes in 148 msecs (HTTP/1.1 200) 7 headers in 207 bytes (1 switches on core 0)\ 172.17.0.1 - - [08/Aug/2022:23:24:30 +0000] "GET /api/datatypes?extension_only=False HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2279|app: 0|req: 8/17] 172.17.0.1 () \{56 vars in 1303 bytes\} [Mon Aug 8 23:24:30 2022] GET /api/datatypes?extension_only=False => generated 63194 bytes in 215 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 0)\ Mon Aug 8 23:24:31 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:32 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:33 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:34 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:35 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:36 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:37 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:38 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:39 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:40 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ galaxy.tools INFO 2022-08-08 23:24:41,130 [pN:main.web.2,p:2279,w:2,m:0,tN:uWSGIWorker2Core1] Validated and populated state for tool request (104.331 ms)\ galaxy.tools.actions.upload DEBUG 2022-08-08 23:24:41,132 [pN:main.web.2,p:2279,w:2,m:0,tN:uWSGIWorker2Core1] Persisted uploads (0.640 ms)\ galaxy.tools.actions.upload DEBUG 2022-08-08 23:24:41,364 [pN:main.web.2,p:2279,w:2,m:0,tN:uWSGIWorker2Core1] Checked uploads (231.282 ms)\ Mon Aug 8 23:24:41 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ galaxy.tools.actions.upload_common INFO 2022-08-08 23:24:41,436 [pN:main.web.2,p:2279,w:2,m:0,tN:uWSGIWorker2Core1] tool upload1 created job id 101\ \ ==> /home/galaxy/logs/handler1.log <==\ galaxy.jobs.handler DEBUG 2022-08-08 23:24:41,653 [pN:handler1,p:255,tN:JobHandlerQueue.monitor_thread] Grabbed Job(s): 101\ galaxy.jobs.mapper DEBUG 2022-08-08 23:24:41,827 [pN:handler1,p:255,tN:JobHandlerQueue.monitor_thread] (101) Mapped job to destination id: slurm_cluster\ galaxy.jobs.handler DEBUG 2022-08-08 23:24:41,909 [pN:handler1,p:255,tN:JobHandlerQueue.monitor_thread] (101) Dispatching to slurm runner\ galaxy.jobs DEBUG 2022-08-08 23:24:42,032 [pN:handler1,p:255,tN:JobHandlerQueue.monitor_thread] (101) Persisting job destination (destination id: slurm_cluster)\ galaxy.jobs DEBUG 2022-08-08 23:24:42,123 [pN:handler1,p:255,tN:JobHandlerQueue.monitor_thread] (101) Working directory for job is: /export/galaxy-central/database/job_working_directory/000/101\ galaxy.jobs.runners DEBUG 2022-08-08 23:24:42,165 [pN:handler1,p:255,tN:JobHandlerQueue.monitor_thread] Job [101] queued (255.762 ms)\ galaxy.jobs.handler INFO 2022-08-08 23:24:42,183 [pN:handler1,p:255,tN:JobHandlerQueue.monitor_thread] (101) Job dispatched\ \ ==> /home/galaxy/logs/uwsgi.log <==\ galaxy.jobs.manager DEBUG 2022-08-08 23:24:41,505 [pN:main.web.2,p:2279,w:2,m:0,tN:uWSGIWorker2Core1] (Job[id=101,tool_id=upload1]) Configured job handler for tool 'upload1' is: handlers\ galaxy.web_stack.handlers INFO 2022-08-08 23:24:41,570 [pN:main.web.2,p:2279,w:2,m:0,tN:uWSGIWorker2Core1] Flushed transaction for Job[id=101,tool_id=upload1] (63.807 ms)\ galaxy.web_stack.handlers INFO 2022-08-08 23:24:41,570 [pN:main.web.2,p:2279,w:2,m:0,tN:uWSGIWorker2Core1] (Job[id=101,tool_id=upload1]) Handler 'handlers' assigned using 'HANDLER_ASSIGNMENT_METHODS.DB_SKIP_LOCKED' assignment method\ galaxy.tools.actions.upload DEBUG 2022-08-08 23:24:41,571 [pN:main.web.2,p:2279,w:2,m:0,tN:uWSGIWorker2Core1] Created upload job (206.387 ms)\ galaxy.tools.execute DEBUG 2022-08-08 23:24:41,571 [pN:main.web.2,p:2279,w:2,m:0,tN:uWSGIWorker2Core1] Tool upload1 created job 101 (439.709 ms)\ galaxy.jobs.manager DEBUG 2022-08-08 23:24:41,635 [pN:main.web.2,p:2279,w:2,m:0,tN:uWSGIWorker2Core1] (Job[id=101,tool_id=upload1]) Configured job handler for tool 'upload1' is: handlers\ galaxy.web_stack.handlers INFO 2022-08-08 23:24:41,636 [pN:main.web.2,p:2279,w:2,m:0,tN:uWSGIWorker2Core1] (Job[id=101,tool_id=upload1]) Handler 'handlers' assigned using 'HANDLER_ASSIGNMENT_METHODS.DB_SKIP_LOCKED' assignment method\ galaxy.tools.execute DEBUG 2022-08-08 23:24:41,662 [pN:main.web.2,p:2279,w:2,m:0,tN:uWSGIWorker2Core1] Executed 1 job(s) for tool upload1 request (531.310 ms)\ 172.17.0.1 - - [08/Aug/2022:23:24:40 +0000] "POST /api/tools HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2279|app: 0|req: 9/18] 172.17.0.1 () \{66 vars in 1533 bytes\} [Mon Aug 8 23:24:40 2022] POST /api/tools => generated 1055 bytes in 760 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 1)\ 172.17.0.1 - - [08/Aug/2022:23:24:41 +0000] "GET /api/histories/f597429621d6eb2b/contents?order=hid&v=dev&q=update_time-ge&q=deleted&q=purged&qv=2022-08-08T23%3A24%3A26.199Z&qv=False&qv=False HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2275|app: 0|req: 10/19] 172.17.0.1 () \{58 vars in 1569 bytes\} [Mon Aug 8 23:24:41 2022] GET /api/histories/f597429621d6eb2b/contents?order=hid&v=dev&q=update_time-ge&q=deleted&q=purged&qv=2022-08-08T23%3A24%3A26.199Z&qv=False&qv=False => generated 490 bytes in 152 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 1)\ Mon Aug 8 23:24:42 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ \ ==> /home/galaxy/logs/handler1.log <==\ galaxy.jobs DEBUG 2022-08-08 23:24:42,690 [pN:handler1,p:255,tN:SlurmRunner.work_thread-0] Job wrapper for Job [101] prepared (439.562 ms)\ galaxy.jobs.command_factory INFO 2022-08-08 23:24:42,751 [pN:handler1,p:255,tN:SlurmRunner.work_thread-0] Built script [/export/galaxy-central/database/job_working_directory/000/101/tool_script.sh] for tool command [python '/galaxy-central/tools/data_source/upload.py' '/galaxy-central' '/export/galaxy-central/database/job_working_directory/000/101/registry.xml' '/export/galaxy-central/database/job_working_directory/000/101/upload_params.json' '101:/export/galaxy-central/database/job_working_directory/000/101/working/dataset_101_files:/export/galaxy-central/database/files/000/dataset_101.dat']\ galaxy.tool_util.deps DEBUG 2022-08-08 23:24:42,876 [pN:handler1,p:255,tN:SlurmRunner.work_thread-0] Using dependency bcftools version 1.5 of type conda\ galaxy.jobs.runners DEBUG 2022-08-08 23:24:42,884 [pN:handler1,p:255,tN:SlurmRunner.work_thread-0] (101) command is: mkdir -p working outputs configs\ if [ -d _working ]; then\ rm -rf working/ outputs/ configs/; cp -R _working working; cp -R _outputs outputs; cp -R _configs configs\ else\ cp -R working _working; cp -R outputs _outputs; cp -R configs _configs\ fi\ cd working; /bin/bash /export/galaxy-central/database/job_working_directory/000/101/tool_script.sh > ../outputs/tool_stdout 2> ../outputs/tool_stderr; return_code=$?; cd '/export/galaxy-central/database/job_working_directory/000/101'; \ [ "$GALAXY_VIRTUAL_ENV" = "None" ] && GALAXY_VIRTUAL_ENV="$_GALAXY_VIRTUAL_ENV"; _galaxy_setup_environment True\ export PATH=$PATH:'/export/tool_deps/_conda/envs/__bcftools@1.5/bin' ; python "metadata/set.py"; sh -c "exit $return_code"\ galaxy.jobs.runners.drmaa DEBUG 2022-08-08 23:24:42,966 [pN:handler1,p:255,tN:SlurmRunner.work_thread-0] (101) submitting file /export/galaxy-central/database/job_working_directory/000/101/galaxy_101.sh\ galaxy.jobs.runners.drmaa DEBUG 2022-08-08 23:24:42,966 [pN:handler1,p:255,tN:SlurmRunner.work_thread-0] (101) native specification is: --ntasks=1 --share\ galaxy.jobs.runners.drmaa INFO 2022-08-08 23:24:43,017 [pN:handler1,p:255,tN:SlurmRunner.work_thread-0] (101) queued as 2\ galaxy.jobs.runners.drmaa DEBUG 2022-08-08 23:24:43,190 [pN:handler1,p:255,tN:SlurmRunner.monitor_thread] (101/2) state change: job is queued and active\ \ ==> /home/galaxy/logs/slurmctld.log <==\ [2022-08-08T23:24:43.013] _slurm_rpc_submit_batch_job: JobId=2 InitPrio=4294901759 usec=6911\ [2022-08-08T23:24:43.015] error: Could not open job state file /tmp/slurm/job_state: No such file or directory\ [2022-08-08T23:24:43.015] error: NOTE: Trying backup state save file. Jobs may be lost!\ [2022-08-08T23:24:43.015] No job state file (/tmp/slurm/job_state.old) found\ \ ==> /home/galaxy/logs/uwsgi.log <==\ Mon Aug 8 23:24:43 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ \ ==> /home/galaxy/logs/handler1.log <==\ galaxy.jobs.runners.drmaa DEBUG 2022-08-08 23:24:44,226 [pN:handler1,p:255,tN:SlurmRunner.monitor_thread] (101/2) state change: job finished, but failed\ galaxy.jobs.runners.slurm WARNING 2022-08-08 23:24:44,366 [pN:handler1,p:255,tN:SlurmRunner.monitor_thread] (101/2) Job failed due to unknown reasons, job state in SLURM was: FAILED\ \ ==> /home/galaxy/logs/slurmctld.log <==\ [2022-08-08T23:24:43.966] sched: Allocate JobID=2 NodeList=23895fdc3425 #CPUs=1 Partition=debug\ [2022-08-08T23:24:44.105] _job_complete: JobID=2 State=0x1 NodeCnt=1 WEXITSTATUS 1\ [2022-08-08T23:24:44.106] _job_complete: JobID=2 State=0x8005 NodeCnt=1 done\ \ ==> /home/galaxy/logs/slurmd.log <==\ [2022-08-08T23:24:43.988] _run_prolog: run job script took usec=427\ [2022-08-08T23:24:43.988] _run_prolog: prolog with lock for job 2 ran for 0 seconds\ [2022-08-08T23:24:43.988] Launching batch job 2 for UID 1450\ [2022-08-08T23:24:44.076] [2.batch] error: setgroups: Operation not permitted\ [2022-08-08T23:24:44.095] [2.batch] sending REQUEST_COMPLETE_BATCH_SCRIPT, error:0 status 256\ [2022-08-08T23:24:44.110] [2.batch] done with job\ \ ==> /home/galaxy/logs/uwsgi.log <==\ Mon Aug 8 23:24:44 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ \ ==> /home/galaxy/logs/handler1.log <==\ galaxy.tools.error_reports DEBUG 2022-08-08 23:24:44,760 [pN:handler1,p:255,tN:SlurmRunner.work_thread-1] Bug report plugin generated response None\ galaxy.jobs.runners DEBUG 2022-08-08 23:24:44,805 [pN:handler1,p:255,tN:SlurmRunner.work_thread-1] (101/2) Unable to cleanup /export/galaxy-central/database/job_working_directory/000/101/galaxy_101.sh: [Errno 2] No such file or directory: '/export/galaxy-central/database/job_working_directory/000/101/galaxy_101.sh'\ galaxy.jobs.runners DEBUG 2022-08-08 23:24:44,819 [pN:handler1,p:255,tN:SlurmRunner.work_thread-1] (101/2) Unable to cleanup /export/galaxy-central/database/job_working_directory/000/101/galaxy_101.o: [Errno 2] No such file or directory: '/export/galaxy-central/database/job_working_directory/000/101/galaxy_101.o'\ galaxy.jobs.runners DEBUG 2022-08-08 23:24:44,836 [pN:handler1,p:255,tN:SlurmRunner.work_thread-1] (101/2) Unable to cleanup /export/galaxy-central/database/job_working_directory/000/101/galaxy_101.e: [Errno 2] No such file or directory: '/export/galaxy-central/database/job_working_directory/000/101/galaxy_101.e'\ galaxy.jobs.runners DEBUG 2022-08-08 23:24:44,852 [pN:handler1,p:255,tN:SlurmRunner.work_thread-1] (101/2) Unable to cleanup /export/galaxy-central/database/job_working_directory/000/101/galaxy_101.ec: [Errno 2] No such file or directory: '/export/galaxy-central/database/job_working_directory/000/101/galaxy_101.ec'\ \ ==> /home/galaxy/logs/uwsgi.log <==\ Mon Aug 8 23:24:45 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ 172.17.0.1 - - [08/Aug/2022:23:24:45 +0000] "GET /api/histories/f597429621d6eb2b/contents?order=hid&v=dev&q=update_time-ge&q=deleted&q=purged&qv=2022-08-08T23%3A24%3A41.000Z&qv=False&qv=False HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2275|app: 0|req: 11/20] 172.17.0.1 () \{58 vars in 1569 bytes\} [Mon Aug 8 23:24:45 2022] GET /api/histories/f597429621d6eb2b/contents?order=hid&v=dev&q=update_time-ge&q=deleted&q=purged&qv=2022-08-08T23%3A24%3A41.000Z&qv=False&qv=False => generated 489 bytes in 120 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 2)\ 172.17.0.1 - - [08/Aug/2022:23:24:46 +0000] "GET /api/histories/f597429621d6eb2b?keys=size%2Cnon_ready_jobs%2Ccontents_active%2Chid_counter HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2279|app: 0|req: 10/21] 172.17.0.1 () \{58 vars in 1465 bytes\} [Mon Aug 8 23:24:46 2022] GET /api/histories/f597429621d6eb2b?keys=size%2Cnon_ready_jobs%2Ccontents_active%2Chid_counter => generated 112 bytes in 132 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 2)\ Mon Aug 8 23:24:46 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:47 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ 172.17.0.1 - - [08/Aug/2022:23:24:47 +0000] "GET /api/histories/f597429621d6eb2b/contents/datasets/6f91353f3eb0fa4a HTTP/1.1" 200 - "http://localhost:8080/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2275|app: 0|req: 12/22] 172.17.0.1 () \{58 vars in 1418 bytes\} [Mon Aug 8 23:24:47 2022] GET /api/histories/f597429621d6eb2b/contents/datasets/6f91353f3eb0fa4a => generated 1645 bytes in 266 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 3)\ Mon Aug 8 23:24:48 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:49 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:50 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:51 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:52 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:53 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:54 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:55 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:56 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:57 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:58 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:24:59 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:00 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:01 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ 172.17.0.1 - - [08/Aug/2022:23:25:01 +0000] "GET /api/datasets/6f91353f3eb0fa4a HTTP/1.1" 200 - "http://localhost:8080/datasets/error?dataset_id=6f91353f3eb0fa4a&__identifer=y39qzuksufb" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2279|app: 0|req: 11/23] 172.17.0.1 () \{60 vars in 1422 bytes\} [Mon Aug 8 23:25:01 2022] GET /api/datasets/6f91353f3eb0fa4a => generated 1645 bytes in 250 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 3)\ 172.17.0.1 - - [08/Aug/2022:23:25:01 +0000] "GET /api/jobs/6f91353f3eb0fa4a?full=True HTTP/1.1" 200 - "http://localhost:8080/datasets/error" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2275|app: 0|req: 13/24] 172.17.0.1 () \{60 vars in 1381 bytes\} [Mon Aug 8 23:25:01 2022] GET /api/jobs/6f91353f3eb0fa4a?full=True => generated 1141 bytes in 283 msecs (HTTP/1.1 200) 3 headers in 139 bytes (1 switches on core 0)\ 172.17.0.1 - - [08/Aug/2022:23:25:02 +0000] "GET /api/jobs/6f91353f3eb0fa4a/common_problems HTTP/1.1" 403 - "http://localhost:8080/datasets/error" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36"\ [pid: 2279|app: 0|req: 12/25] 172.17.0.1 () \{60 vars in 1394 bytes\} [Mon Aug 8 23:25:02 2022] GET /api/jobs/6f91353f3eb0fa4a/common_problems => generated 79 bytes in 26 msecs (HTTP/1.1 403) 2 headers in 95 bytes (1 switches on core 0)\ Mon Aug 8 23:25:02 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:03 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:04 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:05 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:06 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:07 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:08 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:09 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:10 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:11 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:12 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:13 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:14 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:15 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:16 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:17 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:18 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:19 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:20 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:21 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:22 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:23 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:24 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:25 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:26 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:27 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:28 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:29 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:30 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:31 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:32 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:33 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:34 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***\ Mon Aug 8 23:25:35 2022 - *** uWSGI listen queue of socket "127.0.0.1:4001" (fd: 6) full !!! (28770288/64) ***}