Contributors: 8
Author |
Tokens |
Token Proportion |
Commits |
Commit Proportion |
Ingo Molnar |
148 |
69.81% |
6 |
30.00% |
Linus Torvalds (pre-git) |
23 |
10.85% |
5 |
25.00% |
Mel Gorman |
18 |
8.49% |
3 |
15.00% |
Rik Van Riel |
10 |
4.72% |
2 |
10.00% |
Jann Horn |
6 |
2.83% |
1 |
5.00% |
Kefeng Wang |
4 |
1.89% |
1 |
5.00% |
David S. Miller |
2 |
0.94% |
1 |
5.00% |
Greg Kroah-Hartman |
1 |
0.47% |
1 |
5.00% |
Total |
212 |
|
20 |
|
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef _LINUX_SCHED_NUMA_BALANCING_H
#define _LINUX_SCHED_NUMA_BALANCING_H
/*
* This is the interface between the scheduler and the MM that
* implements memory access pattern based NUMA-balancing:
*/
#include <linux/sched.h>
#define TNF_MIGRATED 0x01
#define TNF_NO_GROUP 0x02
#define TNF_SHARED 0x04
#define TNF_FAULT_LOCAL 0x08
#define TNF_MIGRATE_FAIL 0x10
enum numa_vmaskip_reason {
NUMAB_SKIP_UNSUITABLE,
NUMAB_SKIP_SHARED_RO,
NUMAB_SKIP_INACCESSIBLE,
NUMAB_SKIP_SCAN_DELAY,
NUMAB_SKIP_PID_INACTIVE,
NUMAB_SKIP_IGNORE_PID,
NUMAB_SKIP_SEQ_COMPLETED,
};
#ifdef CONFIG_NUMA_BALANCING
extern void task_numa_fault(int last_node, int node, int pages, int flags);
extern pid_t task_numa_group_id(struct task_struct *p);
extern void set_numabalancing_state(bool enabled);
extern void task_numa_free(struct task_struct *p, bool final);
bool should_numa_migrate_memory(struct task_struct *p, struct folio *folio,
int src_nid, int dst_cpu);
#else
static inline void task_numa_fault(int last_node, int node, int pages,
int flags)
{
}
static inline pid_t task_numa_group_id(struct task_struct *p)
{
return 0;
}
static inline void set_numabalancing_state(bool enabled)
{
}
static inline void task_numa_free(struct task_struct *p, bool final)
{
}
static inline bool should_numa_migrate_memory(struct task_struct *p,
struct folio *folio, int src_nid, int dst_cpu)
{
return true;
}
#endif
#endif /* _LINUX_SCHED_NUMA_BALANCING_H */