cclcentral.cse.nd.edu catalog server

glados.crc.nd.edu

return to catalog view

time_application 344768446
time_polling 45369605746
time_internal 6683566
time_status_msgs 14549647
time_receive 75015016
time_send 14769814
disk_largest 0
disk_total 0
memory_largest 0
memory_total 0
cores_largest 0
cores_total 0
workers_largest 0
workers_total 0
workers_idled_out 38
workers_removed 42
workers_idle 0
workers_connected 0
workers 0
capacity_weighted 299
tasks_left 134695
name glados.crc.nd.edu
lastheardfrom 1728604765
address 10.32.93.156
gpus_largest 0
gpus_smallest 0
gpus_total 0
gpus_inuse 0
disk_smallest 0
memory_smallest 0
cores_smallest 0
workers_smallest 0
workers_inuse 0
master_load 0
bytes_sent 870774773
time_workers_execute_exhaustion 37290266004
time_send_good 0
time_when_started 1728573318672581
tasks_exhausted_attempts 52
tasks_cancelled 0
tasks_failed 1
tasks_dispatched 73
tasks_submitted 154
tasks_with_results 0
tasks_waiting 152
workers_lost 0
workers_fast_aborted 0
workers_released 0
workers_able 0
workers_init 0
network_interfaces [{"family":"AF_INET","host":"10.32.93.156","interface":"eno12399np0"}]
master_preferred_connection by_ip
priority 0
port 1026
version 7.1.0 FINAL
owner apiccine
working_dir /tmpscratch/users/apiccine/FullProduction/FullR2/UL16APV/Round1/Batch1/LHE_step/v2
starttime 1728573318
project lobster_apiccine_EFT_LHE_GLADOSandT3_UL16APV_v2
type wq_master
workers_busy 0
tasks_on_workers 0
tasks_running 0
tasks_done 54
tasks_complete 54
time_receive_good 1257
time_workers_execute 81121265307
time_workers_execute_good 30718977514
bytes_received 77572
capacity_tasks 284
capacity_cores 1278
capacity_memory 4136744
capacity_disk 4811528
capacity_instantaneous 280
cores_inuse 0
memory_inuse 0
disk_inuse 0
categories [{"max_allocation_count":152,"first_allocation_count":0,"max_disk":"~4557","max_memory":"~5382","max_cores":">26","tasks_cancelled":0,"tasks_failed":53,"tasks_done":1,"tasks_dispatched":0,"tasks_running":0,"tasks_waiting":152,"category":"lhe_ttgamma"}]
tasks_total_cores 4104
tasks_total_memory 818064
tasks_total_disk 692664
workers_joined 42