diff mbox series

[RFC] mm/mempolicy: Weighted interleave auto-tuning

Message ID 20241210215439.94819-1-joshua.hahnjy@gmail.com
State New
Headers show
Series [RFC] mm/mempolicy: Weighted interleave auto-tuning | expand

Commit Message

Joshua Hahn Dec. 10, 2024, 9:54 p.m. UTC
On machines with multiple memory nodes, interleaving page allocations
across nodes allows for better utilization of each node's bandwidth.
Previous work by Gregory Price [1] introduced weighted interleave, which
allowed for pages to be allocated across NUMA nodes according to
user-set ratios.

Ideally, these weights should be proportional to their bandwidth, so
that under bandwidth pressure, each node uses its maximal efficient
bandwidth and prevents latency from increasing exponentially.

At the same time, we want these weights to be as small as possible.
Having ratios that involve large co-prime numbers like 7639:1345:7 leads
to awkward and inefficient allocations, since the node with weight 7
will remain mostly unused (and despite being proportional to bandwidth,
will not aid in relieving the pressure present in the other two nodes).

This patch introduces an auto-configuration for the interleave weights
that aims to balance the two goals of setting node weights to be
proportional to their bandwidths and keeping the weight values low.
This balance is controlled by a value max_node_weight, which defines the
maximum weight a single node can take.

Large max_node_weights generally lead to increased weight-bandwidth
proportionality, but can lead to underutilized nodes (think worst-case
scenario, which is 1:max_node_weight). Lower max_node_weights reduce the
effects of underutilized nodes, but may lead to improperly loaded
distributions.

This knob is exposed as a sysfs interface with a default value of 32.
Weights are re-calculated once at boottime and then every time the knob
is changed by the user, or when the ACPI table is updated.

[1] https://lore.kernel.org/linux-mm/20240202170238.90004-1-gregory.price@memverge.com/

Signed-off-by: Joshua Hahn <joshua.hahnjy@gmail.com>
Signed-off-by: Gregory Price <gourry@gourry.net>
Co-Developed-by: Gregory Price <gourry@gourry.net>
---
 ...fs-kernel-mm-mempolicy-weighted-interleave |  24 +++
 drivers/acpi/numa/hmat.c                      |   1 +
 drivers/base/node.c                           |   7 +
 include/linux/mempolicy.h                     |   4 +
 mm/mempolicy.c                                | 195 ++++++++++++++++--
 5 files changed, 211 insertions(+), 20 deletions(-)

Comments

Hyeonggon Yoo Dec. 13, 2024, 6:19 a.m. UTC | #1
On 2024-12-11 06:54 AM, Joshua Hahn wrote:
> On machines with multiple memory nodes, interleaving page allocations
> across nodes allows for better utilization of each node's bandwidth.
> Previous work by Gregory Price [1] introduced weighted interleave, which
> allowed for pages to be allocated across NUMA nodes according to
> user-set ratios.
> 
> Ideally, these weights should be proportional to their bandwidth, so
> that under bandwidth pressure, each node uses its maximal efficient
> bandwidth and prevents latency from increasing exponentially.
> 
> At the same time, we want these weights to be as small as possible.
> Having ratios that involve large co-prime numbers like 7639:1345:7 leads
> to awkward and inefficient allocations, since the node with weight 7
> will remain mostly unused (and despite being proportional to bandwidth,
> will not aid in relieving the pressure present in the other two nodes).
> 
> This patch introduces an auto-configuration for the interleave weights
> that aims to balance the two goals of setting node weights to be
> proportional to their bandwidths and keeping the weight values low.
> This balance is controlled by a value max_node_weight, which defines the
> maximum weight a single node can take.

Hi Joshua,

I am wondering how this is going to work for host memory + CXL memory 
interleaving. I guess by "the ACPI table" you mean the ACPI HMAT or CXL 
CDAT, both of which does not provide the bandwidth of host memory.
If this feature does not consider the bandwidth of host memory, manual 
configuration will be inevitable anyway.

> Large max_node_weights generally lead to increased weight-bandwidth
> proportionality, but can lead to underutilized nodes (think worst-case
> scenario, which is 1:max_node_weight). Lower max_node_weights reduce the
> effects of underutilized nodes, but may lead to improperly loaded
> distributions.
> 
> This knob is exposed as a sysfs interface with a default value of 32.
> Weights are re-calculated once at boottime and then every time the knob
> is changed by the user, or when the ACPI table is updated.
>> [1] 
https://lore.kernel.org/linux-mm/20240202170238.90004-1-gregory.price@memverge.com/
> 
> Signed-off-by: Joshua Hahn <joshua.hahnjy@gmail.com>
> Signed-off-by: Gregory Price <gourry@gourry.net>
> Co-Developed-by: Gregory Price <gourry@gourry.net>
> ---
>   ...fs-kernel-mm-mempolicy-weighted-interleave |  24 +++
>   drivers/acpi/numa/hmat.c                      |   1 +
>   drivers/base/node.c                           |   7 +
>   include/linux/mempolicy.h                     |   4 +
>   mm/mempolicy.c                                | 195 ++++++++++++++++--
>   5 files changed, 211 insertions(+), 20 deletions(-)
> 
> diff --git a/Documentation/ABI/testing/sysfs-kernel-mm-mempolicy-weighted-interleave b/Documentation/ABI/testing/sysfs-kernel-mm-mempolicy-weighted-interleave
> index 0b7972de04e9..2ef9a87ce878 100644
> --- a/Documentation/ABI/testing/sysfs-kernel-mm-mempolicy-weighted-interleave
> +++ b/Documentation/ABI/testing/sysfs-kernel-mm-mempolicy-weighted-interleave
> @@ -23,3 +23,27 @@ Description:	Weight configuration interface for nodeN
>   		Writing an empty string or `0` will reset the weight to the
>   		system default. The system default may be set by the kernel
>   		or drivers at boot or during hotplug events.
> +
> +What:		/sys/kernel/mm/mempolicy/weighted_interleave/max_node_weight
> +Date:		December 2024
> +Contact:	Linux memory management mailing list <linux-mm@kvack.org>
> +Description:	Weight limiting / scaling interface
> +
> +		The maximum interleave weight for a memory node. When it is
> +		updated, any previous changes to interleave weights (i.e. via
> +		the nodeN sysfs interfaces) are ignored, and new weights are
> +		calculated using ACPI-reported bandwidths and scaled.
> +

At first this paragraph sounded like "previously stored weights are 
discarded after setting max_node_weight", but I think you mean
"User can override the default values, but defaults values are 
calculated regardless of the values set by the user". Right?

> +		It is possible for weights to be greater than max_node_weight if
> +		the nodeN interfaces are directly modified to be greater.
> +
> +		Minimum weight: 1
> +		Default value: 32
> +		Maximum weight: 255
> +
> +		Writing an empty string will set the value to be the default
> +		(32). Writing a value outside the valid range  will return
> +		EINVAL and will not re-trigger a weight scaling.
> +
> +		Setting max_node_weight to 1 is equivalent to unweighted
> +		interleave.
> diff --git a/mm/mempolicy.c b/mm/mempolicy.c
> index ee32a10e992c..f789280acdcb 100644
> --- a/mm/mempolicy.c
> +++ b/mm/mempolicy.c
> @@ -109,6 +109,7 @@
>   #include <linux/mmu_notifier.h>
>   #include <linux/printk.h>
>   #include <linux/swapops.h>
> +#include <linux/gcd.h>
>   
>   #include <asm/tlbflush.h>
>   #include <asm/tlb.h>
> @@ -153,24 +154,116 @@ static unsigned int mempolicy_behavior;

[...snip...]

> +int mempolicy_set_node_perf(unsigned int node, struct access_coordinate *coords)
> +{
> +	unsigned long *old_bw, *new_bw;
> +	unsigned long bw_val;
> +	u8 *old_iw, *new_iw;
> +
> +	/*
> +	 * Bandwidths above this limit causes rounding errors when reducing
> +	 * weights. This value is ~16 exabytes, which is unreasonable anyways.
> +	 */
> +	bw_val = min(coords->read_bandwidth, coords->write_bandwidth);
> +	if (bw_val > (U64_MAX / 10))
> +		return -EINVAL;
> +
> +	new_bw = kcalloc(nr_node_ids, sizeof(unsigned long), GFP_KERNEL);
> +	if (!new_bw)
> +		return -ENOMEM;
> +
> +	new_iw = kzalloc(nr_node_ids, GFP_KERNEL);

I think kcalloc(nr_node_ids, sizeof(u8), GFP_KERNEL); will be more readable.

> +	if (!new_iw) {
> +		kfree(new_bw);
> +		return -ENOMEM;
> +	}
> +
> +	mutex_lock(&default_iwt_lock);
> +	old_bw = node_bw_table;
> +	old_iw = rcu_dereference_protected(default_iw_table,
> +					   lockdep_is_held(&default_iwt_lock));
> +
> +	if (old_bw)
> +		memcpy(new_bw, old_bw, nr_node_ids*sizeof(unsigned long));
> +	new_bw[node] = bw_val;
> +	node_bw_table = new_bw;
> +
> +	reduce_interleave_weights(new_bw, new_iw);
> +	rcu_assign_pointer(default_iw_table, new_iw);
> +
> +	mutex_unlock(&default_iwt_lock);
> +	synchronize_rcu();
> +	kfree(old_bw);
> +	kfree(old_iw);
> +	return 0;
> +}
> +
>   /**
>    * numa_nearest_node - Find nearest node by state
>    * @node: Node id to start the search
> @@ -2001,7 +2094,7 @@ static unsigned int weighted_interleave_nid(struct mempolicy *pol, pgoff_t ilx)
>   {
>   	nodemask_t nodemask;
>   	unsigned int target, nr_nodes;
> -	u8 *table;
> +	u8 *table, *defaults;
>   	unsigned int weight_total = 0;
>   	u8 weight;
>   	int nid;
> @@ -2012,11 +2105,12 @@ static unsigned int weighted_interleave_nid(struct mempolicy *pol, pgoff_t ilx)
>   
>   	rcu_read_lock();
>   	table = rcu_dereference(iw_table);
> +	defaults = rcu_dereference(iw_table);

Probably you intended rcu_dereference(default_iw_table)?

>   	/* calculate the total weight */
>   	for_each_node_mask(nid, nodemask) {
>   		/* detect system default usage */
> -		weight = table ? table[nid] : 1;
> -		weight = weight ? weight : 1;
> +		weight = table ? table[nid] : 0;
> +		weight = weight ? weight : (defaults ? defaults[nid] : 1);
>   		weight_total += weight;
>   	}
>   
> @@ -2025,8 +2119,8 @@ static unsigned int weighted_interleave_nid(struct mempolicy *pol, pgoff_t ilx)
>   	nid = first_node(nodemask);
>   	while (target) {
>   		/* detect system default usage */
> -		weight = table ? table[nid] : 1;
> -		weight = weight ? weight : 1;
> +		weight = table ? table[nid] : 0;
> +		weight = weight ? weight : (defaults ? defaults[nid] : 1);
>   		if (target < weight)
>   			break;
>   		target -= weight;
> @@ -2409,7 +2503,7 @@ static unsigned long alloc_pages_bulk_array_weighted_interleave(gfp_t gfp,
>   	unsigned long nr_allocated = 0;
>   	unsigned long rounds;
>   	unsigned long node_pages, delta;
> -	u8 *table, *weights, weight;
> +	u8 *weights, weight;
>   	unsigned int weight_total = 0;
>   	unsigned long rem_pages = nr_pages;
>   	nodemask_t nodes;
> @@ -2458,16 +2552,8 @@ static unsigned long alloc_pages_bulk_array_weighted_interleave(gfp_t gfp,
>   	if (!weights)
>   		return total_allocated;
>   
> -	rcu_read_lock();
> -	table = rcu_dereference(iw_table);
> -	if (table)
> -		memcpy(weights, table, nr_node_ids);
> -	rcu_read_unlock();
> -
> -	/* calculate total, detect system default usage */
>   	for_each_node_mask(node, nodes) {
> -		if (!weights[node])
> -			weights[node] = 1;
> +		weights[node] = get_il_weight(node);
>   		weight_total += weights[node];
>   	}
>   
> @@ -3396,6 +3482,7 @@ static ssize_t node_store(struct kobject *kobj, struct kobj_attribute *attr,
>   }
>   
>   static struct iw_node_attr **node_attrs;
> +static struct kobj_attribute *max_nw_attr;

Where is max_nw_attr initialized?

Best,
Hyeonggon

>   static void sysfs_wi_node_release(struct iw_node_attr *node_attr,
>   				  struct kobject *parent)
> @@ -3413,6 +3500,10 @@ static void sysfs_wi_release(struct kobject *wi_kobj)
>   
>   	for (i = 0; i < nr_node_ids; i++)
>   		sysfs_wi_node_release(node_attrs[i], wi_kobj);
> +
> +	sysfs_remove_file(wi_kobj, &max_nw_attr->attr);
> +	kfree(max_nw_attr->attr.name);
> +	kfree(max_nw_attr);
>   	kobject_put(wi_kobj);
>   }
Gregory Price Dec. 13, 2024, 4:28 p.m. UTC | #2
On Fri, Dec 13, 2024 at 03:19:20PM +0900, Hyeonggon Yoo wrote:
> On 2024-12-11 06:54 AM, Joshua Hahn wrote:
> > This patch introduces an auto-configuration for the interleave weights
> > that aims to balance the two goals of setting node weights to be
> > proportional to their bandwidths and keeping the weight values low.
> > This balance is controlled by a value max_node_weight, which defines the
> > maximum weight a single node can take.
> 
> Hi Joshua,
> 
> I am wondering how this is going to work for host memory + CXL memory
> interleaving. I guess by "the ACPI table" you mean the ACPI HMAT or CXL
> CDAT, both of which does not provide the bandwidth of host memory.

Then your BIOS vendor needs to fix their ACPI table generation, because
HMAT can absolutely contain that information.

[078h 0120   2]               Structure Type : 0001 [System Locality Latency and Bandwidth Information]
[07Ah 0122   2]                     Reserved : 0000
[07Ch 0124   4]                       Length : 00000030
[080h 0128   1]        Flags (decoded below) : 00
                            Memory Hierarchy : 0
                   Use Minimum Transfer Size : 0
                    Non-sequential Transfers : 0
[081h 0129   1]                    Data Type : 00
[082h 0130   1]        Minimum Transfer Size : 00
[083h 0131   1]                    Reserved1 : 00
[084h 0132   4] Initiator Proximity Domains # : 00000001
[088h 0136   4]   Target Proximity Domains # : 00000002
[08Ch 0140   4]                    Reserved2 : 00000000
[090h 0144   8]              Entry Base Unit : 00000000000003E8
[098h 0152   4] Initiator Proximity Domain List : 00000000
[09Ch 0156   4] Target Proximity Domain List : 00000000
[0A0h 0160   4] Target Proximity Domain List : 00000001
[0A4h 0164   2]                        Entry : 006E
[0A6h 0166   2]                        Entry : 01FE

[0A8h 0168   2]               Structure Type : 0001 [System Locality Latency and Bandwidth Information]
[0AAh 0170   2]                     Reserved : 0000
[0ACh 0172   4]                       Length : 00000030
[0B0h 0176   1]        Flags (decoded below) : 00
                            Memory Hierarchy : 0
                   Use Minimum Transfer Size : 0
                    Non-sequential Transfers : 0
[0B1h 0177   1]                    Data Type : 03
[0B2h 0178   1]        Minimum Transfer Size : 00
[0B3h 0179   1]                    Reserved1 : 00
[0B4h 0180   4] Initiator Proximity Domains # : 00000001
[0B8h 0184   4]   Target Proximity Domains # : 00000002
[0BCh 0188   4]                    Reserved2 : 00000000
[0C0h 0192   8]              Entry Base Unit : 0000000000000064
[0C8h 0200   4] Initiator Proximity Domain List : 00000000
[0CCh 0204   4] Target Proximity Domain List : 00000000
[0D0h 0208   4] Target Proximity Domain List : 00000001
[0D4h 0212   2]                        Entry : 1200
[0D6h 0214   2]                        Entry : 0064

Obviously if information is missing, then manual is the only way forward.

> > +		The maximum interleave weight for a memory node. When it is
> > +		updated, any previous changes to interleave weights (i.e. via
> > +		the nodeN sysfs interfaces) are ignored, and new weights are
> > +		calculated using ACPI-reported bandwidths and scaled.
> > +
> 
> At first this paragraph sounded like "previously stored weights are
> discarded after setting max_node_weight", but I think you mean
> "User can override the default values, but defaults values are calculated
> regardless of the values set by the user". Right?
> 

Agree that these comments need clarification, we'll workshop it.

~Gregory
Hyeonggon Yoo Dec. 16, 2024, 7:53 a.m. UTC | #3
On 2024-12-14 4:57 AM, Joshua Hahn wrote:
> On Fri, 13 Dec 2024 15:19:20 +0900 Hyeonggon Yoo <hyeonggon.yoo@sk.com> wrote:
> 
>> On 2024-12-11 06:54 AM, Joshua Hahn wrote:
>>> This patch introduces an auto-configuration for the interleave weights
>>> that aims to balance the two goals of setting node weights to be
>>> proportional to their bandwidths and keeping the weight values low.
>>> This balance is controlled by a value max_node_weight, which defines the
>>> maximum weight a single node can take.
>>
>> Hi Joshua,
>>
>> I am wondering how this is going to work for host memory + CXL memory
>> interleaving. I guess by "the ACPI table" you mean the ACPI HMAT or CXL
>> CDAT, both of which does not provide the bandwidth of host memory.
>> If this feature does not consider the bandwidth of host memory, manual
>> configuration will be inevitable anyway.
> 
> Hi Hyeonggon,
> 
> Thank you for reviewing my patch!

No problem :)

> As Gregory showed in his reply, I think it would be possible to get host bandwidth information
> using the ACPI HMAT.

You're right. I was wrong. I checked on our server, and its bandwidth 
information was valid for both local memory and CXL memory. Sorry for
the noise.

> [-----8<-----]
> 
>>> +What:		/sys/kernel/mm/mempolicy/weighted_interleave/max_node_weight
>>> +Date:		December 2024
>>> +Contact:	Linux memory management mailing list <linux-mm@kvack.org>
>>> +Description:	Weight limiting / scaling interface
>>> +
>>> +		The maximum interleave weight for a memory node. When it is
>>> +		updated, any previous changes to interleave weights (i.e. via
>>> +		the nodeN sysfs interfaces) are ignored, and new weights are
>>> +		calculated using ACPI-reported bandwidths and scaled.
>>> +
>>
>> At first this paragraph sounded like "previously stored weights are
>> discarded after setting max_node_weight", but I think you mean
>> "User can override the default values, but defaults values are
>> calculated regardless of the values set by the user". Right?
> 
> In the implementation, the first way you interpreted is the correct
> description. That is, if a user manually changes a ndoe weight,
> then updates the max_node_weight, the previous manual change will
> be overwritten by the newly scaled values.
 >
> Does this behavior make sense?

Ok. then current implementation overwrites the node weights
previously set by the user.

I think it makes sense to re-scale all nodes and overwrite manually set 
node weights, because it's what the knob is meant to do, and the user 
still can override the weight by setting node weight after updating
max_node_weight.

By the way, could you please explain which part of this patch implements
this rule? IIUC it does not invalidate iw_table after updating these
default values, which makes get_il_weight() to use manully set node 
weights even after updating max_node_weight. (Or probably I just
misunderstood the code?)

> Have a great day!

Have a great day too :)

--
Hyeonggon
Huang, Ying Dec. 21, 2024, 5:57 a.m. UTC | #4
Hi, Joshua,

Joshua Hahn <joshua.hahnjy@gmail.com> writes:

> On Fri, 13 Dec 2024 15:19:20 +0900 Hyeonggon Yoo <hyeonggon.yoo@sk.com> wrote:
>
>> On 2024-12-11 06:54 AM, Joshua Hahn wrote:

[snip]

>
> [-----8<-----]
>
>> > +What:		/sys/kernel/mm/mempolicy/weighted_interleave/max_node_weight

I don't think that we need a new knob for this.  Just use a reasonable
default value, for example, 32 or 16.  If it turns out that a knob will
be really helpful, we can add it at that time.  For now, I don't think
the requirements are clear.  And, this is a new ABI and we need to
maintain it almost for ever.  We must be careful about new knob.

>> > +Date:		December 2024
>> > +Contact:	Linux memory management mailing list <linux-mm@kvack.org>
>> > +Description:	Weight limiting / scaling interface
>> > +
>> > +		The maximum interleave weight for a memory node. When it is
>> > +		updated, any previous changes to interleave weights (i.e. via
>> > +		the nodeN sysfs interfaces) are ignored, and new weights are
>> > +		calculated using ACPI-reported bandwidths and scaled.
>> > +
>> 
>> At first this paragraph sounded like "previously stored weights are 
>> discarded after setting max_node_weight", but I think you mean
>> "User can override the default values, but defaults values are 
>> calculated regardless of the values set by the user". Right?
>
> In the implementation, the first way you interpreted is the correct
> description. That is, if a user manually changes a ndoe weight,
> then updates the max_node_weight, the previous manual change will
> be overwritten by the newly scaled values.
>
> Does this behavior make sense? Perhaps it makes sense to ignore
> user-changed values when doing the re-scaling if a user decides to
> change the max_node_weight value themselves. 
>
> Regardless of what implementation makes sense, I can re-write the
> description so that there is no ambiguity when it comes to the
> expected behavior of the code. Thank you for pointing this out!

I don't think that it's a good idea to override the user supplied
configuration values.  User configurations always have higher priority
than system default configurations.  IIUC, this is the general rule of
Linux kernel user space interface.

---
Best Regards,
Huang, Ying
Gregory Price Dec. 21, 2024, 2:58 p.m. UTC | #5
On Sat, Dec 21, 2024 at 01:57:58PM +0800, Huang, Ying wrote:
> Hi, Joshua,
> 
> Joshua Hahn <joshua.hahnjy@gmail.com> writes:
> 
> > On Fri, 13 Dec 2024 15:19:20 +0900 Hyeonggon Yoo <hyeonggon.yoo@sk.com> wrote:
> >
> >> On 2024-12-11 06:54 AM, Joshua Hahn wrote:
> 
> [snip]
> 
> >
> > [-----8<-----]
> >
> >> > +What:		/sys/kernel/mm/mempolicy/weighted_interleave/max_node_weight
> 
> I don't think that we need a new knob for this.  Just use a reasonable
> default value, for example, 32 or 16.  If it turns out that a knob will
> be really helpful, we can add it at that time.  For now, I don't think
> the requirements are clear.  And, this is a new ABI and we need to
> maintain it almost for ever.  We must be careful about new knob.
> 

This is fair.  We spent a good amount of time modeling the best
effective maximum weight and basically came to the conclusion that 32
has a good balance of minimizing error and being somewhat aggressive.

Ripping out the sysfs is easy enough.

> >
> > Regardless of what implementation makes sense, I can re-write the
> > description so that there is no ambiguity when it comes to the
> > expected behavior of the code. Thank you for pointing this out!
> 
> I don't think that it's a good idea to override the user supplied
> configuration values.  User configurations always have higher priority
> than system default configurations.  IIUC, this is the general rule of
> Linux kernel user space interface.
> 

We discussed this and decided it was confusing no matter what we did.

If new data comes in (CDAT data from a hotplug event), then the weights
are now wrong for the new global state - regardless of whether the user
set a weight manually or not.  This also allowed us to simplify the
implementation a bit.

But if generally we need to preserve user settings, then I think the
best we can do to provide a sane system is ignore the user setting when
re-weighting on a hotplug event.

e.g. user has not set a value

default_values [5,2,-] <- 1 node not set, expected to be hotplugged
user_values    [-,-,-] <- user has not set values
effective      [5,2,-]

hotplug event
default_values [2,1,1] - reweight has occurred
user_values    [-,-,-]
effective      [2,1,1]

e.g. user has set a value

default_values [5,2,-] <- 1 node not set, expected to be hotplugged
user_values    [4,-,-] <- user has only set one value
effective      [4,2,-]

hotplug event
default_values [2,1,1] - reweight has occurred
user_values    [4,-,-]
effective      [4,1,1]


So default values get updated, but user values get left alone.

If that's sane we'll fix it up.

> ---
> Best Regards,
> Huang, Ying
Huang, Ying Dec. 22, 2024, 8:29 a.m. UTC | #6
Gregory Price <gourry@gourry.net> writes:

> On Sat, Dec 21, 2024 at 01:57:58PM +0800, Huang, Ying wrote:
>> Hi, Joshua,
>> 
>> Joshua Hahn <joshua.hahnjy@gmail.com> writes:
>> 
>> > On Fri, 13 Dec 2024 15:19:20 +0900 Hyeonggon Yoo <hyeonggon.yoo@sk.com> wrote:
>> >
>> >> On 2024-12-11 06:54 AM, Joshua Hahn wrote:
>>

[snip]

>
>> >
>> > Regardless of what implementation makes sense, I can re-write the
>> > description so that there is no ambiguity when it comes to the
>> > expected behavior of the code. Thank you for pointing this out!
>> 
>> I don't think that it's a good idea to override the user supplied
>> configuration values.  User configurations always have higher priority
>> than system default configurations.  IIUC, this is the general rule of
>> Linux kernel user space interface.
>> 
>
> We discussed this and decided it was confusing no matter what we did.
>
> If new data comes in (CDAT data from a hotplug event), then the weights
> are now wrong for the new global state - regardless of whether the user
> set a weight manually or not.  This also allowed us to simplify the
> implementation a bit.
>
> But if generally we need to preserve user settings, then I think the
> best we can do to provide a sane system is ignore the user setting when
> re-weighting on a hotplug event.
>
> e.g. user has not set a value
>
> default_values [5,2,-] <- 1 node not set, expected to be hotplugged
> user_values    [-,-,-] <- user has not set values
> effective      [5,2,-]
>
> hotplug event
> default_values [2,1,1] - reweight has occurred
> user_values    [-,-,-]
> effective      [2,1,1]
>
> e.g. user has set a value
>
> default_values [5,2,-] <- 1 node not set, expected to be hotplugged
> user_values    [4,-,-] <- user has only set one value
> effective      [4,2,-]
>
> hotplug event
> default_values [2,1,1] - reweight has occurred
> user_values    [4,-,-]
> effective      [4,1,1]

Another choice is that if the user set a value, he/she set all values
effectively.  Even if he/she doesn't set the other values, he/she thinks
that the other values are good, and more importantly, the ratio is good.
If so,

default_values [5,2,-] <- 1 node not set, expected to be hotplugged
user_values    [4,2,0] <- user has only set one value, not populated nodes have value 0
effective      [4,2,0]

hotplug event
default_values [2,1,1] - reweight has occurred
user_values    [4,2,0]
effective      [4,2,0]

In this way, 0 becomes a valid value too.

What do you think about this?

> So default values get updated, but user values get left alone.
>
> If that's sane we'll fix it up.

---
Best Regards,
Huang, Ying
Gregory Price Dec. 22, 2024, 4:54 p.m. UTC | #7
On Sun, Dec 22, 2024 at 04:29:30PM +0800, Huang, Ying wrote:
> Gregory Price <gourry@gourry.net> writes:
> 
> > On Sat, Dec 21, 2024 at 01:57:58PM +0800, Huang, Ying wrote:
> 
> Another choice is that if the user set a value, he/she set all values
> effectively.  Even if he/she doesn't set the other values, he/she thinks
> that the other values are good, and more importantly, the ratio is good.

This is probably the actual way to go.  

> If so,
> 
> default_values [5,2,-] <- 1 node not set, expected to be hotplugged
> user_values    [4,2,0] <- user has only set one value, not populated nodes have value 0
> effective      [4,2,0]
>
> hotplug event
> default_values [2,1,1] - reweight has occurred
> user_values    [4,2,0]
> effective      [4,2,0]
> 
> In this way, 0 becomes a valid value too.
> 
> What do you think about this?
> 

We decided when implementing weights that 0 was a special value that
reverts to the system default:

  Writing an empty string or `0` will reset the weight to the
  system default. The system default may be set by the kernel
  or drivers at boot or during hotplug events.

I'm ok pulling the default weights in collectively once the first one is
written, but 0 is an invalid value which causes issues.

We went through that when we initially implemented the feature w/ task-local
weights and why the help function overrides it to 1 if it's ever seen.

We'll revert back to our initial implementation w/ default_iw_table and
iw_table - where iw_table contains user-defined weights.  Writing a 0 to
iw_table[N] will allow get_il_weight() to retrieve default_iw_table[N]
as the docs imply it should.

~Gregory
diff mbox series

Patch

diff --git a/Documentation/ABI/testing/sysfs-kernel-mm-mempolicy-weighted-interleave b/Documentation/ABI/testing/sysfs-kernel-mm-mempolicy-weighted-interleave
index 0b7972de04e9..2ef9a87ce878 100644
--- a/Documentation/ABI/testing/sysfs-kernel-mm-mempolicy-weighted-interleave
+++ b/Documentation/ABI/testing/sysfs-kernel-mm-mempolicy-weighted-interleave
@@ -23,3 +23,27 @@  Description:	Weight configuration interface for nodeN
 		Writing an empty string or `0` will reset the weight to the
 		system default. The system default may be set by the kernel
 		or drivers at boot or during hotplug events.
+
+What:		/sys/kernel/mm/mempolicy/weighted_interleave/max_node_weight
+Date:		December 2024
+Contact:	Linux memory management mailing list <linux-mm@kvack.org>
+Description:	Weight limiting / scaling interface
+
+		The maximum interleave weight for a memory node. When it is
+		updated, any previous changes to interleave weights (i.e. via
+		the nodeN sysfs interfaces) are ignored, and new weights are
+		calculated using ACPI-reported bandwidths and scaled.
+
+		It is possible for weights to be greater than max_node_weight if
+		the nodeN interfaces are directly modified to be greater.
+
+		Minimum weight: 1
+		Default value: 32
+		Maximum weight: 255
+
+		Writing an empty string will set the value to be the default
+		(32). Writing a value outside the valid range  will return
+		EINVAL and will not re-trigger a weight scaling.
+
+		Setting max_node_weight to 1 is equivalent to unweighted
+		interleave.
diff --git a/drivers/acpi/numa/hmat.c b/drivers/acpi/numa/hmat.c
index a2f9e7a4b479..83f3858a773f 100644
--- a/drivers/acpi/numa/hmat.c
+++ b/drivers/acpi/numa/hmat.c
@@ -20,6 +20,7 @@ 
 #include <linux/list_sort.h>
 #include <linux/memregion.h>
 #include <linux/memory.h>
+#include <linux/mempolicy.h>
 #include <linux/mutex.h>
 #include <linux/node.h>
 #include <linux/sysfs.h>
diff --git a/drivers/base/node.c b/drivers/base/node.c
index eb72580288e6..d45216386c03 100644
--- a/drivers/base/node.c
+++ b/drivers/base/node.c
@@ -7,6 +7,7 @@ 
 #include <linux/init.h>
 #include <linux/mm.h>
 #include <linux/memory.h>
+#include <linux/mempolicy.h>
 #include <linux/vmstat.h>
 #include <linux/notifier.h>
 #include <linux/node.h>
@@ -214,6 +215,12 @@  void node_set_perf_attrs(unsigned int nid, struct access_coordinate *coord,
 			break;
 		}
 	}
+
+	/* When setting CPU access coordinates, update mempolicy */
+	if (access == ACCESS_COORDINATE_CPU) {
+		if (mempolicy_set_node_perf(nid, coord))
+			pr_info("failed to set node%d mempolicy attrs\n", nid);
+	}
 }
 EXPORT_SYMBOL_GPL(node_set_perf_attrs);
 
diff --git a/include/linux/mempolicy.h b/include/linux/mempolicy.h
index 931b118336f4..d564e9e893ea 100644
--- a/include/linux/mempolicy.h
+++ b/include/linux/mempolicy.h
@@ -11,6 +11,7 @@ 
 #include <linux/slab.h>
 #include <linux/rbtree.h>
 #include <linux/spinlock.h>
+#include <linux/node.h>
 #include <linux/nodemask.h>
 #include <linux/pagemap.h>
 #include <uapi/linux/mempolicy.h>
@@ -177,6 +178,9 @@  static inline bool mpol_is_preferred_many(struct mempolicy *pol)
 
 extern bool apply_policy_zone(struct mempolicy *policy, enum zone_type zone);
 
+extern int mempolicy_set_node_perf(unsigned int node,
+				   struct access_coordinate *coords);
+
 #else
 
 struct mempolicy {};
diff --git a/mm/mempolicy.c b/mm/mempolicy.c
index ee32a10e992c..f789280acdcb 100644
--- a/mm/mempolicy.c
+++ b/mm/mempolicy.c
@@ -109,6 +109,7 @@ 
 #include <linux/mmu_notifier.h>
 #include <linux/printk.h>
 #include <linux/swapops.h>
+#include <linux/gcd.h>
 
 #include <asm/tlbflush.h>
 #include <asm/tlb.h>
@@ -153,24 +154,116 @@  static unsigned int mempolicy_behavior;
  *
  * iw_table is RCU protected
  */
+static unsigned long *node_bw_table;
+static u8 __rcu *default_iw_table;
+static DEFINE_MUTEX(default_iwt_lock);
+
 static u8 __rcu *iw_table;
 static DEFINE_MUTEX(iw_table_lock);
 
+static int max_node_weight = 32;
+
 static u8 get_il_weight(int node)
 {
-	u8 *table;
+	u8 *table, *defaults;
 	u8 weight;
 
 	rcu_read_lock();
+	defaults = rcu_dereference(default_iw_table);
 	table = rcu_dereference(iw_table);
-	/* if no iw_table, use system default */
-	weight = table ? table[node] : 1;
-	/* if value in iw_table is 0, use system default */
-	weight = weight ? weight : 1;
+	/* if no iw_table, use system default - if no default, use 1 */
+	weight = table ? table[node] : 0;
+	weight = weight ? weight : (defaults ? defaults[node] : 1);
 	rcu_read_unlock();
 	return weight;
 }
 
+/*
+ * Convert ACPI-reported bandwidths into weighted interleave weights for
+ * informed page allocation.
+ * Call with default_iwt_lock held
+ */
+static void reduce_interleave_weights(unsigned long *bw, u8 *new_iw)
+{
+	uint64_t ttl_bw = 0, ttl_iw = 0, scaling_factor = 1;
+	unsigned int iw_gcd = 1, i = 0;
+
+	/* Recalculate the bandwidth distribution given the new info */
+	for (i = 0; i < nr_node_ids; i++)
+		ttl_bw += bw[i];
+
+	/* If node is not set or has < 1% of total bw, use minimum value of 1 */
+	for (i = 0; i < nr_node_ids; i++) {
+		if (bw[i]) {
+			scaling_factor = 100 * bw[i];
+			new_iw[i] = max(scaling_factor / ttl_bw, 1);
+		} else {
+			new_iw[i] = 1;
+		}
+		ttl_iw += new_iw[i];
+	}
+
+	/*
+	 * Scale each node's share of the total bandwidth from percentages
+	 * to whole numbers in the range [1, max_node_weight]
+	 */
+	for (i = 0; i < nr_node_ids; i++) {
+		scaling_factor = max_node_weight * new_iw[i];
+		new_iw[i] = max(scaling_factor / ttl_iw, 1);
+		if (unlikely(i == 0))
+			iw_gcd = new_iw[0];
+		iw_gcd = gcd(iw_gcd, new_iw[i]);
+	}
+
+	/* 1:2 is strictly better than 16:32. Reduce by the weights' GCD. */
+	for (i = 0; i < nr_node_ids; i++)
+		new_iw[i] /= iw_gcd;
+}
+
+int mempolicy_set_node_perf(unsigned int node, struct access_coordinate *coords)
+{
+	unsigned long *old_bw, *new_bw;
+	unsigned long bw_val;
+	u8 *old_iw, *new_iw;
+
+	/*
+	 * Bandwidths above this limit causes rounding errors when reducing
+	 * weights. This value is ~16 exabytes, which is unreasonable anyways.
+	 */
+	bw_val = min(coords->read_bandwidth, coords->write_bandwidth);
+	if (bw_val > (U64_MAX / 10))
+		return -EINVAL;
+
+	new_bw = kcalloc(nr_node_ids, sizeof(unsigned long), GFP_KERNEL);
+	if (!new_bw)
+		return -ENOMEM;
+
+	new_iw = kzalloc(nr_node_ids, GFP_KERNEL);
+	if (!new_iw) {
+		kfree(new_bw);
+		return -ENOMEM;
+	}
+
+	mutex_lock(&default_iwt_lock);
+	old_bw = node_bw_table;
+	old_iw = rcu_dereference_protected(default_iw_table,
+					   lockdep_is_held(&default_iwt_lock));
+
+	if (old_bw)
+		memcpy(new_bw, old_bw, nr_node_ids*sizeof(unsigned long));
+	new_bw[node] = bw_val;
+	node_bw_table = new_bw;
+
+	reduce_interleave_weights(new_bw, new_iw);
+	rcu_assign_pointer(default_iw_table, new_iw);
+
+	mutex_unlock(&default_iwt_lock);
+	synchronize_rcu();
+	kfree(old_bw);
+	kfree(old_iw);
+	return 0;
+}
+
 /**
  * numa_nearest_node - Find nearest node by state
  * @node: Node id to start the search
@@ -2001,7 +2094,7 @@  static unsigned int weighted_interleave_nid(struct mempolicy *pol, pgoff_t ilx)
 {
 	nodemask_t nodemask;
 	unsigned int target, nr_nodes;
-	u8 *table;
+	u8 *table, *defaults;
 	unsigned int weight_total = 0;
 	u8 weight;
 	int nid;
@@ -2012,11 +2105,12 @@  static unsigned int weighted_interleave_nid(struct mempolicy *pol, pgoff_t ilx)
 
 	rcu_read_lock();
 	table = rcu_dereference(iw_table);
+	defaults = rcu_dereference(iw_table);
 	/* calculate the total weight */
 	for_each_node_mask(nid, nodemask) {
 		/* detect system default usage */
-		weight = table ? table[nid] : 1;
-		weight = weight ? weight : 1;
+		weight = table ? table[nid] : 0;
+		weight = weight ? weight : (defaults ? defaults[nid] : 1);
 		weight_total += weight;
 	}
 
@@ -2025,8 +2119,8 @@  static unsigned int weighted_interleave_nid(struct mempolicy *pol, pgoff_t ilx)
 	nid = first_node(nodemask);
 	while (target) {
 		/* detect system default usage */
-		weight = table ? table[nid] : 1;
-		weight = weight ? weight : 1;
+		weight = table ? table[nid] : 0;
+		weight = weight ? weight : (defaults ? defaults[nid] : 1);
 		if (target < weight)
 			break;
 		target -= weight;
@@ -2409,7 +2503,7 @@  static unsigned long alloc_pages_bulk_array_weighted_interleave(gfp_t gfp,
 	unsigned long nr_allocated = 0;
 	unsigned long rounds;
 	unsigned long node_pages, delta;
-	u8 *table, *weights, weight;
+	u8 *weights, weight;
 	unsigned int weight_total = 0;
 	unsigned long rem_pages = nr_pages;
 	nodemask_t nodes;
@@ -2458,16 +2552,8 @@  static unsigned long alloc_pages_bulk_array_weighted_interleave(gfp_t gfp,
 	if (!weights)
 		return total_allocated;
 
-	rcu_read_lock();
-	table = rcu_dereference(iw_table);
-	if (table)
-		memcpy(weights, table, nr_node_ids);
-	rcu_read_unlock();
-
-	/* calculate total, detect system default usage */
 	for_each_node_mask(node, nodes) {
-		if (!weights[node])
-			weights[node] = 1;
+		weights[node] = get_il_weight(node);
 		weight_total += weights[node];
 	}
 
@@ -3396,6 +3482,7 @@  static ssize_t node_store(struct kobject *kobj, struct kobj_attribute *attr,
 }
 
 static struct iw_node_attr **node_attrs;
+static struct kobj_attribute *max_nw_attr;
 
 static void sysfs_wi_node_release(struct iw_node_attr *node_attr,
 				  struct kobject *parent)
@@ -3413,6 +3500,10 @@  static void sysfs_wi_release(struct kobject *wi_kobj)
 
 	for (i = 0; i < nr_node_ids; i++)
 		sysfs_wi_node_release(node_attrs[i], wi_kobj);
+
+	sysfs_remove_file(wi_kobj, &max_nw_attr->attr);
+	kfree(max_nw_attr->attr.name);
+	kfree(max_nw_attr);
 	kobject_put(wi_kobj);
 }
 
@@ -3454,6 +3545,63 @@  static int add_weight_node(int nid, struct kobject *wi_kobj)
 	return 0;
 }
 
+static ssize_t max_nw_show(struct kobject *kobj, struct kobj_attribute *attr,
+			char *buf)
+{
+	return sysfs_emit(buf, "%d\n", max_node_weight);
+}
+
+static ssize_t max_nw_store(struct kobject *kobj,
+		struct kobj_attribute *attr, const char *buf, size_t count)
+{
+	unsigned long *bw;
+	u8 *old_iw, *new_iw;
+	u8 max_weight;
+
+	if (count == 0 || sysfs_streq(buf, ""))
+		max_weight = 32;
+	else if (kstrtou8(buf, 0, &max_weight) || max_weight == 0)
+		return -EINVAL;
+
+	new_iw = kzalloc(nr_node_ids, GFP_KERNEL);
+	if (!new_iw)
+		return -ENOMEM;
+
+	mutex_lock(&default_iwt_lock);
+	bw = node_bw_table;
+
+	if (!bw) {
+		mutex_unlock(&default_iwt_lock);
+		kfree(new_iw);
+		return -ENODEV;
+	}
+
+	max_node_weight = max_weight;
+	old_iw = rcu_dereference_protected(default_iw_table,
+					   lockdep_is_held(&default_iwt_lock));
+
+	reduce_interleave_weights(bw, new_iw);
+	rcu_assign_pointer(default_iw_table, new_iw);
+	mutex_unlock(&default_iwt_lock);
+
+	synchronize_rcu();
+	kfree(old_iw);
+
+	return count;
+}
+
+static struct kobj_attribute wi_attr =
+	__ATTR(max_node_weight, 0664, max_nw_show, max_nw_store);
+
+static struct attribute *wi_default_attrs[] = {
+	&wi_attr.attr,
+	NULL
+};
+
+static const struct attribute_group wi_attr_group = {
+	.attrs = wi_default_attrs,
+};
+
 static int add_weighted_interleave_group(struct kobject *root_kobj)
 {
 	struct kobject *wi_kobj;
@@ -3470,6 +3618,13 @@  static int add_weighted_interleave_group(struct kobject *root_kobj)
 		return err;
 	}
 
+	err = sysfs_create_group(wi_kobj, &wi_attr_group);
+	if (err) {
+		pr_err("failed to add sysfs [max_node_weight]\n");
+		kobject_put(wi_kobj);
+		return err;
+	}
+
 	for_each_node_state(nid, N_POSSIBLE) {
 		err = add_weight_node(nid, wi_kobj);
 		if (err) {