Contributors: 14
Author |
Tokens |
Token Proportion |
Commits |
Commit Proportion |
Johannes Weiner |
239 |
55.71% |
5 |
16.13% |
Roman Gushchin |
75 |
17.48% |
8 |
25.81% |
Maarten Lankhorst |
33 |
7.69% |
2 |
6.45% |
Shakeel Butt |
26 |
6.06% |
4 |
12.90% |
David Finkel |
22 |
5.13% |
1 |
3.23% |
Jakub Kiciński |
14 |
3.26% |
1 |
3.23% |
Eric Dumazet |
7 |
1.63% |
2 |
6.45% |
Glauber de Oliveira Costa |
6 |
1.40% |
2 |
6.45% |
David S. Miller |
2 |
0.47% |
1 |
3.23% |
Kamezawa Hiroyuki |
1 |
0.23% |
1 |
3.23% |
Vladimir Davydov |
1 |
0.23% |
1 |
3.23% |
Balbir Singh |
1 |
0.23% |
1 |
3.23% |
Greg Kroah-Hartman |
1 |
0.23% |
1 |
3.23% |
Christophe Jaillet |
1 |
0.23% |
1 |
3.23% |
Total |
429 |
|
31 |
|
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef _LINUX_PAGE_COUNTER_H
#define _LINUX_PAGE_COUNTER_H
#include <linux/atomic.h>
#include <linux/cache.h>
#include <linux/limits.h>
#include <asm/page.h>
struct page_counter {
/*
* Make sure 'usage' does not share cacheline with any other field in
* v2. The memcg->memory.usage is a hot member of struct mem_cgroup.
*/
atomic_long_t usage;
unsigned long failcnt; /* v1-only field */
CACHELINE_PADDING(_pad1_);
/* effective memory.min and memory.min usage tracking */
unsigned long emin;
atomic_long_t min_usage;
atomic_long_t children_min_usage;
/* effective memory.low and memory.low usage tracking */
unsigned long elow;
atomic_long_t low_usage;
atomic_long_t children_low_usage;
unsigned long watermark;
/* Latest cg2 reset watermark */
unsigned long local_watermark;
/* Keep all the read most fields in a separete cacheline. */
CACHELINE_PADDING(_pad2_);
bool protection_support;
bool track_failcnt;
unsigned long min;
unsigned long low;
unsigned long high;
unsigned long max;
struct page_counter *parent;
} ____cacheline_internodealigned_in_smp;
#if BITS_PER_LONG == 32
#define PAGE_COUNTER_MAX LONG_MAX
#else
#define PAGE_COUNTER_MAX (LONG_MAX / PAGE_SIZE)
#endif
/*
* Protection is supported only for the first counter (with id 0).
*/
static inline void page_counter_init(struct page_counter *counter,
struct page_counter *parent,
bool protection_support)
{
counter->usage = (atomic_long_t)ATOMIC_LONG_INIT(0);
counter->max = PAGE_COUNTER_MAX;
counter->parent = parent;
counter->protection_support = protection_support;
counter->track_failcnt = false;
}
static inline unsigned long page_counter_read(struct page_counter *counter)
{
return atomic_long_read(&counter->usage);
}
void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages);
void page_counter_charge(struct page_counter *counter, unsigned long nr_pages);
bool page_counter_try_charge(struct page_counter *counter,
unsigned long nr_pages,
struct page_counter **fail);
void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages);
void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages);
void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages);
static inline void page_counter_set_high(struct page_counter *counter,
unsigned long nr_pages)
{
WRITE_ONCE(counter->high, nr_pages);
}
int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages);
int page_counter_memparse(const char *buf, const char *max,
unsigned long *nr_pages);
static inline void page_counter_reset_watermark(struct page_counter *counter)
{
unsigned long usage = page_counter_read(counter);
/*
* Update local_watermark first, so it's always <= watermark
* (modulo CPU/compiler re-ordering)
*/
counter->local_watermark = usage;
counter->watermark = usage;
}
#if IS_ENABLED(CONFIG_MEMCG) || IS_ENABLED(CONFIG_CGROUP_DMEM)
void page_counter_calculate_protection(struct page_counter *root,
struct page_counter *counter,
bool recursive_protection);
#else
static inline void page_counter_calculate_protection(struct page_counter *root,
struct page_counter *counter,
bool recursive_protection) {}
#endif
#endif /* _LINUX_PAGE_COUNTER_H */