Commit b094a36f authored by Sebastian Sanchez's avatar Sebastian Sanchez Committed by Doug Ledford

IB/hfi1: Refine user process affinity algorithm

When performing process affinity recommendations for MPI ranks, the current
algorithm doesn't take into account multiple HFI units. Also, real
cores and HT cores are not distinguished from one another. Therefore,
all HT cores are recommended to be assigned first within the local NUMA
node before recommending the assignments of cores in other NUMA nodes.
It's ideal to assign all real cores across all NUMA nodes first, then all
HT 1 cores, then all HT 2 cores, and so on to balance CPU workload. CPU
cores in other NUMA nodes could be running interrupt handlers, and this is
not taken into account.

To balance the CPU workload for user processes, the following
recommendation algorithm is used:

 For each user process that is opening a context on HFI Y:
  a) If all cores are assigned to user processes, start assignments all
	 over from the first core
  b) Assign real cores first, then HT cores (First set of HT cores on
	 all physical cores, then second set of HT cores, and, so on) in the
	 following order:

	 1. Same NUMA node as HFI Y and not running an IRQ handler
	 2. Same NUMA node as HFI Y and running an IRQ handler
	 3. Different NUMA node to HFI Y and not running an IRQ handler
	 4. Different NUMA node to HFI Y and running an IRQ handler
  c) Mark core as assigned in the global affinity structure. As user
	 processes are done, remove core assignments from global affinity
	 structure.

This implementation allows an arbitrary number of HT cores and provides
support for multiple HFIs.

This is being included in the kernel rather than user space due to the
fact that user space has no way of knowing the CPU recommendations for
contexts running as part of other jobs.
Reviewed-by: default avatarIra Weiny <ira.weiny@intel.com>
Reviewed-by: default avatarMitko Haralanov <mitko.haralanov@intel.com>
Reviewed-by: default avatarDennis Dalessandro <dennis.dalessandro@intel.com>
Signed-off-by: default avatarSebastian Sanchez <sebastian.sanchez@intel.com>
Signed-off-by: default avatarDoug Ledford <dledford@redhat.com>
parent d6373019
This diff is collapsed.
...@@ -73,7 +73,6 @@ struct cpu_mask_set { ...@@ -73,7 +73,6 @@ struct cpu_mask_set {
struct hfi1_affinity { struct hfi1_affinity {
struct cpu_mask_set def_intr; struct cpu_mask_set def_intr;
struct cpu_mask_set rcv_intr; struct cpu_mask_set rcv_intr;
struct cpu_mask_set proc;
struct cpumask real_cpu_mask; struct cpumask real_cpu_mask;
/* spin lock to protect affinity struct */ /* spin lock to protect affinity struct */
spinlock_t lock; spinlock_t lock;
...@@ -99,9 +98,9 @@ void hfi1_put_irq_affinity(struct hfi1_devdata *, struct hfi1_msix_entry *); ...@@ -99,9 +98,9 @@ void hfi1_put_irq_affinity(struct hfi1_devdata *, struct hfi1_msix_entry *);
* Determine a CPU affinity for a user process, if the process does not * Determine a CPU affinity for a user process, if the process does not
* have an affinity set yet. * have an affinity set yet.
*/ */
int hfi1_get_proc_affinity(struct hfi1_devdata *, int); int hfi1_get_proc_affinity(int);
/* Release a CPU used by a user process. */ /* Release a CPU used by a user process. */
void hfi1_put_proc_affinity(struct hfi1_devdata *, int); void hfi1_put_proc_affinity(int);
struct hfi1_affinity_node { struct hfi1_affinity_node {
int node; int node;
...@@ -115,6 +114,9 @@ struct hfi1_affinity_node_list { ...@@ -115,6 +114,9 @@ struct hfi1_affinity_node_list {
struct list_head list; struct list_head list;
struct cpumask real_cpu_mask; struct cpumask real_cpu_mask;
struct cpu_mask_set proc; struct cpu_mask_set proc;
int num_core_siblings;
int num_online_nodes;
int num_online_cpus;
/* protect affinity node list */ /* protect affinity node list */
spinlock_t lock; spinlock_t lock;
}; };
......
...@@ -715,7 +715,7 @@ static int hfi1_file_close(struct inode *inode, struct file *fp) ...@@ -715,7 +715,7 @@ static int hfi1_file_close(struct inode *inode, struct file *fp)
hfi1_user_sdma_free_queues(fdata); hfi1_user_sdma_free_queues(fdata);
/* release the cpu */ /* release the cpu */
hfi1_put_proc_affinity(dd, fdata->rec_cpu_num); hfi1_put_proc_affinity(fdata->rec_cpu_num);
/* /*
* Clear any left over, unhandled events so the next process that * Clear any left over, unhandled events so the next process that
...@@ -815,9 +815,10 @@ static int assign_ctxt(struct file *fp, struct hfi1_user_info *uinfo) ...@@ -815,9 +815,10 @@ static int assign_ctxt(struct file *fp, struct hfi1_user_info *uinfo)
ret = find_shared_ctxt(fp, uinfo); ret = find_shared_ctxt(fp, uinfo);
if (ret < 0) if (ret < 0)
goto done_unlock; goto done_unlock;
if (ret) if (ret) {
fd->rec_cpu_num = hfi1_get_proc_affinity( fd->rec_cpu_num =
fd->uctxt->dd, fd->uctxt->numa_id); hfi1_get_proc_affinity(fd->uctxt->numa_id);
}
} }
/* /*
...@@ -929,7 +930,11 @@ static int allocate_ctxt(struct file *fp, struct hfi1_devdata *dd, ...@@ -929,7 +930,11 @@ static int allocate_ctxt(struct file *fp, struct hfi1_devdata *dd,
if (ctxt == dd->num_rcv_contexts) if (ctxt == dd->num_rcv_contexts)
return -EBUSY; return -EBUSY;
fd->rec_cpu_num = hfi1_get_proc_affinity(dd, -1); /*
* If we don't have a NUMA node requested, preference is towards
* device NUMA node.
*/
fd->rec_cpu_num = hfi1_get_proc_affinity(dd->node);
if (fd->rec_cpu_num != -1) if (fd->rec_cpu_num != -1)
numa = cpu_to_node(fd->rec_cpu_num); numa = cpu_to_node(fd->rec_cpu_num);
else else
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment