displays hosts and their static and dynamic resources
By default, returns the following information about all hosts: host name, host status, job state statistics, and job slot limits.
bhosts displays output for condensed host groups and compute units. These host groups and compute units are defined by CONDENSE in the HostGroup and ComputeUnit sections of lsb.hosts. Condensed host groups and compute units are displayed as a single entry with the name as defined by GROUP_NAME or NAME in lsb.hosts.
When LSF adds more resources to a running resizable job, bhosts displays the added resources. When LSF removes resources from a running resizable job, bhosts displays the updated resources.
The -l and -X options display uncondensed output.
The -s option displays information about the numeric shared resources and their associated hosts.
With MultiCluster, displays the information about hosts available to the local cluster. Use -e to view information about exported hosts.
Dynamic Cluster only. Shows information about all hosts, including Dynamic Cluster virtual machine hosts configured with the jobvm resource. Default output includes only standard LSF hosts and Dynamic Cluster hosts configured with the dchost resource.
Displays host topology information for CPU and memory affinity scheduling.
In LSF Advanced Edition, includes the cluster name for execution cluster hosts and host groups in output.
MultiCluster only. Displays information about resources that have been exported to another cluster.
Displays host information in a (long) multi-line format. In addition to the default fields, displays information about the CPU factor, the current load, and the load thresholds. Also displays the value of slots for each host. slots is the greatest number of unused slots on a host.
bhosts -l also displays information about the dispatch windows.
When PowerPolicy is enabled (in lsb.threshold), bhosts -l also displays host power states. Final power states are on or suspend. Intermediate power states are restarting, resuming, and suspending. The final power state under admin control is closed_Power. The final power state under policy control is ok_Power. If the host status becomes unknown (power operation due to failure), the power state is shown as a dash (“-”).
If you specified an administrator comment with the -C option of the host control commands hclose or hopen, -l displays the comment text.
Displays host information in wide format. Fields are displayed without truncation.
For condensed host groups and compute units , the -w option displays the overall status and the number of hosts with the ok, unavail, unreach, and busy status in the following format:
host_group_status num_ok/num_unavail/num_unreach/num_busy
where
host_group_status is the overall status of the host group or compute unit. If a single host in the group or unit is ok, the overall status is also ok.
num_ok, num_unavail, num_unreach, and num_busy are the number of hosts that are ok, unavail, unreach, and busy, respectively.
For example, if there are five ok, two unavail, one unreach, and three busy hosts in a condensed host group hg1, its status is displayed as the following:
hg1 ok 5/2/1/3
If any hosts in the host group or compute unit are closed, the status for the host group is displayed as closed, with no status for the other states:
hg1 closed
Display hosts whose job exit rate has exceeded the threshold configured by EXIT_RATE in lsb.hosts for longer than JOB_EXIT_RATE_DURATION configured in lsb.params, and are still high. By default, these hosts are closed the next time LSF checks host exceptions and invokes eadmin.
Use with the -l option to show detailed information about host exceptions.
If no hosts exceed the job exit rate, bhosts -x displays:
There is no exceptional host found
Displays uncondensed output for host groups and compute units.
Only displays information about hosts that satisfy the resource requirement expression. For more information about resource requirements, see Administering IBM Platform LSF. The size of the resource requirement string is limited to 512 bytes.
LSF supports ordering of resource requirements on all load indices, including external load indices, either static or dynamic.
In LSF Advanced Edition, includes the cluster name for execution cluster hosts in output. The output displayed is sorted by cluster and then by host name.
Specify shared numeric resources only. Displays information about the specified resources. Returns the following information: the resource names, the total and reserved amounts, and the resource locations.
bhosts -s only shows consumable resources.
When LOCAL_TO is configured for a license feature in lsf.licensescheduler, bhosts -s shows different resource information depending on the cluster locality of the features. For example:
bhosts -s
RESOURCE TOTAL RESERVED LOCATION
hspice 36.0 0.0 host1
bhosts -s
RESOURCE TOTAL RESERVED LOCATION
hspice 76.0 0.0 host2
Only displays information about the specified hosts. Do not use quotes when specifying multiple hosts.
For host groups and compute units, the names of the member hosts are displayed instead of the name of the host group or compute unit. Do not use quotes when specifying multiple host groups or compute units.
MultiCluster only. Displays information about hosts in the specified cluster.
Prints command usage to stderr and exits.
Prints LSF release version to stderr and exits.
Displays the following fields:
The name of the host. If a host has batch jobs running and the host is removed from the configuration, the host name is displayed as lost_and_found.
For condensed host groups, this is the name of host group.
With MultiCluster, not shown for fully exported hosts.
The current status of the host and the sbatchd daemon. Batch jobs can only be dispatched to hosts with an ok status. The possible values for host status are as follows:
The host is available to accept batch jobs.
For condensed host groups, if a single host in the host group is ok, the overall status is also shown as ok.
If any host in the host group or compute unit is not ok, bhosts displays the first host status it encounters as the overall status for the condensed host group. Use bhosts -X to see the status of individual hosts in the host group or compute unit.
The host is down, or LIM and sbatchd on the host are unreachable.
LIM on the host is running but sbatchd is unreachable.
The host is not allowed to accept any remote batch jobs. There are several reasons for the host to be closed (see Host-Based -l Options).
This host is a member of a compute unit running an exclusive compute unit job.
With MultiCluster, not shown for fully exported hosts.
The maximum number of job slots that the host can process on a per user basis. If a dash (-) is displayed, there is no limit.
For condensed host groups or compute units, this is the total number of job slots that all hosts in the group or unit can process on a per user basis.
The host does not allocate more than JL/U job slots for one user at the same time. These job slots are used by running jobs, as well as by suspended or pending jobs that have slots reserved for them.
For preemptive scheduling, the accounting is different. These job slots are used by running jobs and by pending jobs that have slots reserved for them (see the description of PREEMPTIVE in lsb.queues(5) and JL/U in lsb.hosts(5)).
The maximum number of job slots available. If a dash (-) is displayed, there is no limit.
For condensed host groups and compute units, this is the total maximum number of job slots available in all hosts in the host group or compute unit.
These job slots are used by running jobs, as well as by suspended or pending jobs that have slots reserved for them.
If preemptive scheduling is used, suspended jobs are not counted (see the description of PREEMPTIVE in lsb.queues(5) and MXJ in lsb.hosts(5)).
A host does not always have to allocate this many job slots if there are waiting jobs; the host must also satisfy its configured load conditions to accept more jobs.
The number of job slots used by jobs dispatched to the host. This includes running, suspended, and chunk jobs.
For condensed host groups and compute units, this is the total number of job slots used by jobs dispatched to any host in the host group or compute unit.
The number of job slots used by jobs running on the host.
For condensed host groups and compute units, this is the total number of job slots used by jobs running on any host in the host group or compute unit.
The number of job slots used by system suspended jobs on the host.
For condensed host groups and compute units, this is the total number of job slots used by system suspended jobs on any host in the host group or compute unit.
The number of job slots used by user suspended jobs on the host. Jobs can be suspended by the user or by the LSF administrator.
For condensed host groups and compute units, this is the total number of job slots used by user suspended jobs on any host in the host group or compute unit.
The number of job slots used by pending jobs that have jobs slots reserved on the host.
For condensed host groups and compute units, this is the total number of job slots used by pending jobs that have job slots reserved on any host in the host group or compute unit.
In addition to the above fields, the -l option also displays the following:
The scheduling and suspending thresholds for the host. If a threshold is not defined, the threshold from the queue definition applies. If both the host and the queue define a threshold for a load index, the most restrictive threshold is used.
The migration threshold is the time that a job dispatched to this host can remain suspended by the system before LSF attempts to migrate the job to another host.
If the host’s operating system supports checkpoint copy, this is indicated here. With checkpoint copy, the operating system automatically copies all open files to the checkpoint directory when a process is checkpointed. Checkpoint copy is currently supported only on Cray systems.
The long format shown by the -l option gives the possible reasons for a host to be closed. If PowerPolicy is enabled in lsb.threshold, it will show the power state:
The host is closed by the LSF administrator or root (see badmin(8)) using badmin hclose. No job can be dispatched to the host, but jobs that are running on the host are not affected.
The host is overloaded. At least one load index exceeds the configured threshold (see lsb.hosts(5)). Indices that exceed their threshold are identified by an asterisk (*). No job can be dispatched to the host, but jobs that are running on the host are not affected.
This host is a member of a compute unit running an exclusive compute unit job (bsub -R "cu[excl]").
For EGO-enabled SLA scheduling, host is closed because it has not been allocated by EGO to run LSF jobs. Hosts allocated from EGO display status ok.
The host is running an exclusive job (bsub -x).
The maximum number of job slots on the host has been reached. No job can be dispatched to the host, but jobs that are running on the host are not affected.
LIM on the host is unreachable, but sbatchd is running.
The host is locked by the LSF administrator or root (see lsadmin(8)) using lsadmin limlock. Running jobs on the host are suspended by LSF (SSUSP). Use lsadmin limunlock to unlock LIM on the local host.
The host is closed by a dispatch window defined in the configuration file lsb.hosts(5). No job can be dispatched to the host, but jobs that are running on the host are not affected.
Displays the CPU normalization factor of the host (see lshosts(1)).
Displays the dispatch windows for each host. Dispatch windows are the time windows during the week when batch jobs can be run on each host. Jobs already started are not affected by the dispatch windows. When the dispatch windows close, jobs are not suspended. Jobs already running continue to run, but no new jobs are started until the windows reopen. The default for the dispatch window is no restriction or always open (that is, twenty-four hours a day and seven days a week). For the dispatch window specification, see the description for the DISPATCH_WINDOWS keyword under the -l option in bqueues(1).
Displays the total and reserved host load.
You specify reserved resources by using bsub -R. These resources are reserved by jobs running on the host.
The total load has different meanings depending on whether the load index is increasing or decreasing.
For increasing load indices, such as run queue lengths, CPU utilization, paging activity, logins, and disk I/O, the total load is the consumed plus the reserved amount. The total load is calculated as the sum of the current load and the reserved load. The current load is the load seen by lsload(1).
For decreasing load indices, such as available memory, idle time, available swap space, and available space in tmp, the total load is the available amount. The total load is the difference between the current load and the reserved load. This difference is the available resource as seen by lsload(1).
Displays the scheduling threshold loadSched and the suspending threshold loadStop. Also displays the migration threshold if defined and the checkpoint support if the host supports checkpointing.
The format for the thresholds is the same as for batch job queues (see bqueues(1)) and lsb.queues(5)). For an explanation of the thresholds and load indices, see the description for the "QUEUE SCHEDULING PARAMETERS" keyword under the -l option in bqueues(1).
Displays the configured threshold of EXIT_RATE for the host and its current load value for host exceptions.
If the LSF administrator specified an administrator comment with the -C option of the badmin host control commands hclose or hopen, the comment text is displayed.
Displays network resource information for IBM Parallel Edition (PE) jobs submitted with the bsub -network option, or to a queue (defined in lsb.queues) or an application profile (defined in lsb.applications) with the NETWORK_REQ parameter defined.
bhosts -l
...
PE NETWORK INFORMATION
NetworkID Status rsv_windows/total_windows
1111111 ok 4/64
2222222 closed_Dedicated 4/64
...
NetworkID is the physical network ID returned by PE.
ok - normal status
closed_Full - all network windows are reserved
closed_Dedicated - a dedicated PE job is running on the network (usage=dedicated specified in the network resource requirement string)
unavail - network information is not available
The host is configured in lsb.hosts to accept jobs for CPU and memory affinity scheduling. If AFFINITY is configured as Y, the keyword all is displayed. If a CPU list is specified under the AFFINITY column, the configured CPU list for affinity scheduling is displayed.
The -s option displays the following: the amounts used for scheduling, the amounts reserved, and the associated hosts for the resources. Only resources (shared or host-based) with numeric values are displayed. See lim, and lsf.cluster on how to configure shared resources.
The following fields are displayed:
The name of the resource.
The total amount free of a resource used for scheduling.
The amount reserved by jobs. You specify the reserved resource using bsub -R.
The hosts that are associated with the resource.
The -aff option displays host topology information for CPU and memory affinity scheduling. Only the topology nodes containing CPUs in the CPULIST defined in lsb.hosts are displayed.
The following fields are displayed:
If the host is configured in lsb.hosts to accept jobs for CPU and memory affinity scheduling, and the host supports affinity scheduling, AFFINITY: Enabled is displayed. If the host is configured in lsb.hosts to accept jobs for CPU and memory affinity scheduling, but the host does not support affinity scheduling, AFFINITY: Disabled (not supported) is displayed. If the host is LIM is not available or sbatchd is unreachable, AFFINITY: UNKNOWN is displayed.
Maximum available memory on the host. If memory availability cannot be determined, a dash (-) is displayed for the host. If the -l option is specified with the -aff option, the host name is not displayed.
For hosts that do not support affinity scheduling, a dash (-) is displayed for host memory and no host topology is displayed.
Requested and total NUMA node memory. It is possible for requested memory for the NUMA node to be greater than the maximum available memory displayed.
A socket is a collection of cores with a direct pipe to memory. Each socket contains 1 or more cores. This does not necessarily refer to a physical socket, but rather to the memory architecture of the machine.
A core is a single entity capable of performing computations.
A node contains sockets, a socket contains cores, and a core can contain threads if the core is enabled for multithreading.
If no NUMA nodes are present, then the NUMA layer in the output is not shown. Other relevant items such as host, socket, core and thread are still shown.
If the host is not available, only the host name is displayed. A dash (-) is shown where available host memory would normally be displayed.
bhosts -l -aff hostA
HOST hostA
STATUS CPUF JL/U MAX NJOBS RUN SSUSP USUSP RSV DISPATCH_WINDOW
ok 60.00 - 8 0 0 0 0 0 -
CURRENT LOAD USED FOR SCHEDULING:
r15s r1m r15m ut pg io ls it tmp swp mem slots
Total 0.0 0.0 0.0 30% 0.0 193 25 0 8605M 5.8G 13.2G 8
Reserved 0.0 0.0 0.0 0% 0.0 0 0 0 0M 0M 0M -
LOAD THRESHOLD USED FOR SCHEDULING:
r15s r1m r15m ut pg io ls it tmp swp mem
loadSched - - - - - - - - - - -
loadStop - - - - - - - - - - -
CONFIGURED AFFINITY CPU LIST: all
AFFINITY: Enabled
Host[15.7G]
NUMA[0: 100M / 15.7G]
Socket0
core0(0)
Socket1
core0(1)
Socket2
core0(2)
Socket3
core0(3)
Socket4
core0(4)
Socket5
core0(5)
Socket6
core0(6)
Socket7
core0(7)
...
Host[1.4G] hostB
NUMA[0: 1.4G / 1.4G] (*0 *1)
...
A job requesting 2 cores, or 2 sockets, or 2 CPUs will run. Requesting 2 cores from the same NUMA node will also run. However, a job requesting 2 cores from the same socket will remain pending.
Reads lsb.hosts.
lsb.hosts, bqueues, lshosts, badmin, lsadmin