Commit ace30fb2 authored by SeongJae Park's avatar SeongJae Park Committed by Andrew Morton

mm/damon/core: use pseudo-moving sum for nr_accesses_bp

Let nr_accesses_bp be calculated as a pseudo-moving sum that updated for
every sampling interval, using damon_moving_sum().  This is assumed to be
useful for cases that the aggregation interval is set quite huge, but the
monivoting results need to be collected earlier than next aggregation
interval is passed.

Link: https://lkml.kernel.org/r/20230915025251.72816-7-sj@kernel.orgSigned-off-by: default avatarSeongJae Park <sj@kernel.org>
Cc: Brendan Higgins <brendanhiggins@google.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
parent 80333828
...@@ -40,7 +40,8 @@ struct damon_addr_range { ...@@ -40,7 +40,8 @@ struct damon_addr_range {
* @ar: The address range of the region. * @ar: The address range of the region.
* @sampling_addr: Address of the sample for the next access check. * @sampling_addr: Address of the sample for the next access check.
* @nr_accesses: Access frequency of this region. * @nr_accesses: Access frequency of this region.
* @nr_accesses_bp: @nr_accesses in basis point (0.01%). * @nr_accesses_bp: @nr_accesses in basis point (0.01%) that updated for
* each sampling interval.
* @list: List head for siblings. * @list: List head for siblings.
* @age: Age of this region. * @age: Age of this region.
* *
...@@ -51,7 +52,11 @@ struct damon_addr_range { ...@@ -51,7 +52,11 @@ struct damon_addr_range {
* damon_update_region_access_rate(). * damon_update_region_access_rate().
* *
* @nr_accesses_bp is another representation of @nr_accesses in basis point * @nr_accesses_bp is another representation of @nr_accesses in basis point
* (1 in 10,000) that updated every aggregation interval. * (1 in 10,000) that updated for every &damon_attrs->sample_interval in a
* manner similar to moving sum. By the algorithm, this value becomes
* @nr_accesses * 10000 for every &struct damon_attrs->aggr_interval. This can
* be used when the aggregation interval is too huge and therefore cannot wait
* for it before getting the access monitoring results.
* *
* @age is initially zero, increased for each aggregation interval, and reset * @age is initially zero, increased for each aggregation interval, and reset
* to zero again if the access frequency is significantly changed. If two * to zero again if the access frequency is significantly changed. If two
...@@ -629,7 +634,8 @@ int damon_set_regions(struct damon_target *t, struct damon_addr_range *ranges, ...@@ -629,7 +634,8 @@ int damon_set_regions(struct damon_target *t, struct damon_addr_range *ranges,
unsigned int nr_ranges); unsigned int nr_ranges);
unsigned int damon_moving_sum(unsigned int mvsum, unsigned int nomvsum, unsigned int damon_moving_sum(unsigned int mvsum, unsigned int nomvsum,
unsigned int len_window, unsigned int new_value); unsigned int len_window, unsigned int new_value);
void damon_update_region_access_rate(struct damon_region *r, bool accessed); void damon_update_region_access_rate(struct damon_region *r, bool accessed,
struct damon_attrs *attrs);
struct damos_filter *damos_new_filter(enum damos_filter_type type, struct damos_filter *damos_new_filter(enum damos_filter_type type,
bool matching); bool matching);
......
...@@ -1599,14 +1599,28 @@ unsigned int damon_moving_sum(unsigned int mvsum, unsigned int nomvsum, ...@@ -1599,14 +1599,28 @@ unsigned int damon_moving_sum(unsigned int mvsum, unsigned int nomvsum,
* damon_update_region_access_rate() - Update the access rate of a region. * damon_update_region_access_rate() - Update the access rate of a region.
* @r: The DAMON region to update for its access check result. * @r: The DAMON region to update for its access check result.
* @accessed: Whether the region has accessed during last sampling interval. * @accessed: Whether the region has accessed during last sampling interval.
* @attrs: The damon_attrs of the DAMON context.
* *
* Update the access rate of a region with the region's last sampling interval * Update the access rate of a region with the region's last sampling interval
* access check result. * access check result.
* *
* Usually this will be called by &damon_operations->check_accesses callback. * Usually this will be called by &damon_operations->check_accesses callback.
*/ */
void damon_update_region_access_rate(struct damon_region *r, bool accessed) void damon_update_region_access_rate(struct damon_region *r, bool accessed,
struct damon_attrs *attrs)
{ {
unsigned int len_window = 1;
/*
* sample_interval can be zero, but cannot be larger than
* aggr_interval, owing to validation of damon_set_attrs().
*/
if (attrs->sample_interval)
len_window = attrs->aggr_interval / attrs->sample_interval;
r->nr_accesses_bp = damon_moving_sum(r->nr_accesses_bp,
r->last_nr_accesses * 10000, len_window,
accessed ? 10000 : 0);
if (accessed) if (accessed)
r->nr_accesses++; r->nr_accesses++;
} }
......
...@@ -148,7 +148,8 @@ static bool damon_pa_young(unsigned long paddr, unsigned long *folio_sz) ...@@ -148,7 +148,8 @@ static bool damon_pa_young(unsigned long paddr, unsigned long *folio_sz)
return accessed; return accessed;
} }
static void __damon_pa_check_access(struct damon_region *r) static void __damon_pa_check_access(struct damon_region *r,
struct damon_attrs *attrs)
{ {
static unsigned long last_addr; static unsigned long last_addr;
static unsigned long last_folio_sz = PAGE_SIZE; static unsigned long last_folio_sz = PAGE_SIZE;
...@@ -157,12 +158,12 @@ static void __damon_pa_check_access(struct damon_region *r) ...@@ -157,12 +158,12 @@ static void __damon_pa_check_access(struct damon_region *r)
/* If the region is in the last checked page, reuse the result */ /* If the region is in the last checked page, reuse the result */
if (ALIGN_DOWN(last_addr, last_folio_sz) == if (ALIGN_DOWN(last_addr, last_folio_sz) ==
ALIGN_DOWN(r->sampling_addr, last_folio_sz)) { ALIGN_DOWN(r->sampling_addr, last_folio_sz)) {
damon_update_region_access_rate(r, last_accessed); damon_update_region_access_rate(r, last_accessed, attrs);
return; return;
} }
last_accessed = damon_pa_young(r->sampling_addr, &last_folio_sz); last_accessed = damon_pa_young(r->sampling_addr, &last_folio_sz);
damon_update_region_access_rate(r, last_accessed); damon_update_region_access_rate(r, last_accessed, attrs);
last_addr = r->sampling_addr; last_addr = r->sampling_addr;
} }
...@@ -175,7 +176,7 @@ static unsigned int damon_pa_check_accesses(struct damon_ctx *ctx) ...@@ -175,7 +176,7 @@ static unsigned int damon_pa_check_accesses(struct damon_ctx *ctx)
damon_for_each_target(t, ctx) { damon_for_each_target(t, ctx) {
damon_for_each_region(r, t) { damon_for_each_region(r, t) {
__damon_pa_check_access(r); __damon_pa_check_access(r, &ctx->attrs);
max_nr_accesses = max(r->nr_accesses, max_nr_accesses); max_nr_accesses = max(r->nr_accesses, max_nr_accesses);
} }
} }
......
...@@ -558,26 +558,27 @@ static bool damon_va_young(struct mm_struct *mm, unsigned long addr, ...@@ -558,26 +558,27 @@ static bool damon_va_young(struct mm_struct *mm, unsigned long addr,
* r the region to be checked * r the region to be checked
*/ */
static void __damon_va_check_access(struct mm_struct *mm, static void __damon_va_check_access(struct mm_struct *mm,
struct damon_region *r, bool same_target) struct damon_region *r, bool same_target,
struct damon_attrs *attrs)
{ {
static unsigned long last_addr; static unsigned long last_addr;
static unsigned long last_folio_sz = PAGE_SIZE; static unsigned long last_folio_sz = PAGE_SIZE;
static bool last_accessed; static bool last_accessed;
if (!mm) { if (!mm) {
damon_update_region_access_rate(r, false); damon_update_region_access_rate(r, false, attrs);
return; return;
} }
/* If the region is in the last checked page, reuse the result */ /* If the region is in the last checked page, reuse the result */
if (same_target && (ALIGN_DOWN(last_addr, last_folio_sz) == if (same_target && (ALIGN_DOWN(last_addr, last_folio_sz) ==
ALIGN_DOWN(r->sampling_addr, last_folio_sz))) { ALIGN_DOWN(r->sampling_addr, last_folio_sz))) {
damon_update_region_access_rate(r, last_accessed); damon_update_region_access_rate(r, last_accessed, attrs);
return; return;
} }
last_accessed = damon_va_young(mm, r->sampling_addr, &last_folio_sz); last_accessed = damon_va_young(mm, r->sampling_addr, &last_folio_sz);
damon_update_region_access_rate(r, last_accessed); damon_update_region_access_rate(r, last_accessed, attrs);
last_addr = r->sampling_addr; last_addr = r->sampling_addr;
} }
...@@ -594,7 +595,8 @@ static unsigned int damon_va_check_accesses(struct damon_ctx *ctx) ...@@ -594,7 +595,8 @@ static unsigned int damon_va_check_accesses(struct damon_ctx *ctx)
mm = damon_get_mm(t); mm = damon_get_mm(t);
same_target = false; same_target = false;
damon_for_each_region(r, t) { damon_for_each_region(r, t) {
__damon_va_check_access(mm, r, same_target); __damon_va_check_access(mm, r, same_target,
&ctx->attrs);
max_nr_accesses = max(r->nr_accesses, max_nr_accesses); max_nr_accesses = max(r->nr_accesses, max_nr_accesses);
same_target = true; same_target = true;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment