Contributors: 16
Author Tokens Token Proportion Commits Commit Proportion
Lorenzo Stoakes 415 64.54% 4 16.67%
David Hildenbrand 161 25.04% 1 4.17%
Linus Torvalds 11 1.71% 2 8.33%
Michel Lespinasse 8 1.24% 1 4.17%
Suren Baghdasaryan 7 1.09% 3 12.50%
David Howells 7 1.09% 1 4.17%
Vincent Guittot 6 0.93% 1 4.17%
Ingo Molnar 5 0.78% 1 4.17%
Colin Cross 5 0.78% 1 4.17%
Linus Torvalds (pre-git) 4 0.62% 2 8.33%
Andrew Morton 4 0.62% 1 4.17%
Joel A Fernandes 3 0.47% 1 4.17%
Yu Zhao 3 0.47% 1 4.17%
Peter Xu 2 0.31% 2 8.33%
Heinrich Schuchardt 1 0.16% 1 4.17%
Mel Gorman 1 0.16% 1 4.17%
Total 643 24


// SPDX-License-Identifier: GPL-2.0-or-later

/*
 * Functions for initialisaing, allocating, freeing and duplicating VMAs. Shared
 * between CONFIG_MMU and non-CONFIG_MMU kernel configurations.
 */

#include "vma_internal.h"
#include "vma.h"

/* SLAB cache for vm_area_struct structures */
static struct kmem_cache *vm_area_cachep;

void __init vma_state_init(void)
{
	struct kmem_cache_args args = {
		.use_freeptr_offset = true,
		.freeptr_offset = offsetof(struct vm_area_struct, vm_freeptr),
	};

	vm_area_cachep = kmem_cache_create("vm_area_struct",
			sizeof(struct vm_area_struct), &args,
			SLAB_HWCACHE_ALIGN|SLAB_PANIC|SLAB_TYPESAFE_BY_RCU|
			SLAB_ACCOUNT);
}

struct vm_area_struct *vm_area_alloc(struct mm_struct *mm)
{
	struct vm_area_struct *vma;

	vma = kmem_cache_alloc(vm_area_cachep, GFP_KERNEL);
	if (!vma)
		return NULL;

	vma_init(vma, mm);

	return vma;
}

static void vm_area_init_from(const struct vm_area_struct *src,
			      struct vm_area_struct *dest)
{
	dest->vm_mm = src->vm_mm;
	dest->vm_ops = src->vm_ops;
	dest->vm_start = src->vm_start;
	dest->vm_end = src->vm_end;
	dest->anon_vma = src->anon_vma;
	dest->vm_pgoff = src->vm_pgoff;
	dest->vm_file = src->vm_file;
	dest->vm_private_data = src->vm_private_data;
	vm_flags_init(dest, src->vm_flags);
	memcpy(&dest->vm_page_prot, &src->vm_page_prot,
	       sizeof(dest->vm_page_prot));
	/*
	 * src->shared.rb may be modified concurrently when called from
	 * dup_mmap(), but the clone will reinitialize it.
	 */
	data_race(memcpy(&dest->shared, &src->shared, sizeof(dest->shared)));
	memcpy(&dest->vm_userfaultfd_ctx, &src->vm_userfaultfd_ctx,
	       sizeof(dest->vm_userfaultfd_ctx));
#ifdef CONFIG_ANON_VMA_NAME
	dest->anon_name = src->anon_name;
#endif
#ifdef CONFIG_SWAP
	memcpy(&dest->swap_readahead_info, &src->swap_readahead_info,
	       sizeof(dest->swap_readahead_info));
#endif
#ifndef CONFIG_MMU
	dest->vm_region = src->vm_region;
#endif
#ifdef CONFIG_NUMA
	dest->vm_policy = src->vm_policy;
#endif
#ifdef __HAVE_PFNMAP_TRACKING
	dest->pfnmap_track_ctx = NULL;
#endif
}

#ifdef __HAVE_PFNMAP_TRACKING
static inline int vma_pfnmap_track_ctx_dup(struct vm_area_struct *orig,
		struct vm_area_struct *new)
{
	struct pfnmap_track_ctx *ctx = orig->pfnmap_track_ctx;

	if (likely(!ctx))
		return 0;

	/*
	 * We don't expect to ever hit this. If ever required, we would have
	 * to duplicate the tracking.
	 */
	if (unlikely(kref_read(&ctx->kref) >= REFCOUNT_MAX))
		return -ENOMEM;
	kref_get(&ctx->kref);
	new->pfnmap_track_ctx = ctx;
	return 0;
}

static inline void vma_pfnmap_track_ctx_release(struct vm_area_struct *vma)
{
	struct pfnmap_track_ctx *ctx = vma->pfnmap_track_ctx;

	if (likely(!ctx))
		return;

	kref_put(&ctx->kref, pfnmap_track_ctx_release);
	vma->pfnmap_track_ctx = NULL;
}
#else
static inline int vma_pfnmap_track_ctx_dup(struct vm_area_struct *orig,
		struct vm_area_struct *new)
{
	return 0;
}
static inline void vma_pfnmap_track_ctx_release(struct vm_area_struct *vma)
{
}
#endif

struct vm_area_struct *vm_area_dup(struct vm_area_struct *orig)
{
	struct vm_area_struct *new = kmem_cache_alloc(vm_area_cachep, GFP_KERNEL);

	if (!new)
		return NULL;

	ASSERT_EXCLUSIVE_WRITER(orig->vm_flags);
	ASSERT_EXCLUSIVE_WRITER(orig->vm_file);
	vm_area_init_from(orig, new);

	if (vma_pfnmap_track_ctx_dup(orig, new)) {
		kmem_cache_free(vm_area_cachep, new);
		return NULL;
	}
	vma_lock_init(new, true);
	INIT_LIST_HEAD(&new->anon_vma_chain);
	vma_numab_state_init(new);
	dup_anon_vma_name(orig, new);

	return new;
}

void vm_area_free(struct vm_area_struct *vma)
{
	/* The vma should be detached while being destroyed. */
	vma_assert_detached(vma);
	vma_numab_state_free(vma);
	free_anon_vma_name(vma);
	vma_pfnmap_track_ctx_release(vma);
	kmem_cache_free(vm_area_cachep, vma);
}