rmap.h 4.03 KB
Newer Older
Linus Torvalds's avatar
Linus Torvalds committed
1
2
3
4
5
6
7
8
9
10
#ifndef _LINUX_RMAP_H
#define _LINUX_RMAP_H
/*
 * Declarations for Reverse Mapping functions in mm/rmap.c
 */

#include <linux/list.h>
#include <linux/slab.h>
#include <linux/mm.h>
#include <linux/spinlock.h>
11
#include <linux/memcontrol.h>
Linus Torvalds's avatar
Linus Torvalds committed
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28

/*
 * The anon_vma heads a list of private "related" vmas, to scan if
 * an anonymous page pointing to this anon_vma needs to be unmapped:
 * the vmas on the list will be related by forking, or by splitting.
 *
 * Since vmas come and go as they are split and merged (particularly
 * in mprotect), the mapping field of an anonymous page cannot point
 * directly to a vma: instead it points to an anon_vma, on whose list
 * the related vmas can be easily linked or unlinked.
 *
 * After unlinking the last vma on the list, we must garbage collect
 * the anon_vma object itself: we're guaranteed no page can be
 * pointing to this anon_vma once its vma list is empty.
 */
struct anon_vma {
	spinlock_t lock;	/* Serialize access to vma list */
29
30
31
32
33
34
35
36
	/*
	 * NOTE: the LSB of the head.next is set by
	 * mm_take_all_locks() _after_ taking the above lock. So the
	 * head must only be read/written after taking the above lock
	 * to be sure to see a valid next pointer. The LSB bit itself
	 * is serialized by a system wide lock only visible to
	 * mm_take_all_locks() (mm_all_locks_mutex).
	 */
Linus Torvalds's avatar
Linus Torvalds committed
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
	struct list_head head;	/* List of private "related" vmas */
};

#ifdef CONFIG_MMU

static inline void anon_vma_lock(struct vm_area_struct *vma)
{
	struct anon_vma *anon_vma = vma->anon_vma;
	if (anon_vma)
		spin_lock(&anon_vma->lock);
}

static inline void anon_vma_unlock(struct vm_area_struct *vma)
{
	struct anon_vma *anon_vma = vma->anon_vma;
	if (anon_vma)
		spin_unlock(&anon_vma->lock);
}

/*
 * anon_vma helper functions.
 */
void anon_vma_init(void);	/* create anon_vma_cachep */
int  anon_vma_prepare(struct vm_area_struct *);
void __anon_vma_merge(struct vm_area_struct *, struct vm_area_struct *);
void anon_vma_unlink(struct vm_area_struct *);
void anon_vma_link(struct vm_area_struct *);
void __anon_vma_link(struct vm_area_struct *);

/*
 * rmap interfaces called when adding or removing pte of page
 */
void page_add_anon_rmap(struct page *, struct vm_area_struct *, unsigned long);
Nick Piggin's avatar
Nick Piggin committed
70
void page_add_new_anon_rmap(struct page *, struct vm_area_struct *, unsigned long);
Linus Torvalds's avatar
Linus Torvalds committed
71
void page_add_file_rmap(struct page *);
72
void page_remove_rmap(struct page *);
Linus Torvalds's avatar
Linus Torvalds committed
73

Nick Piggin's avatar
Nick Piggin committed
74
75
76
77
#ifdef CONFIG_DEBUG_VM
void page_dup_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address);
#else
static inline void page_dup_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address)
Linus Torvalds's avatar
Linus Torvalds committed
78
79
80
{
	atomic_inc(&page->_mapcount);
}
Nick Piggin's avatar
Nick Piggin committed
81
#endif
Linus Torvalds's avatar
Linus Torvalds committed
82
83
84
85

/*
 * Called from mm/vmscan.c to handle paging out
 */
86
87
int page_referenced(struct page *, int is_locked,
			struct mem_cgroup *cnt, unsigned long *vm_flags);
88
int try_to_unmap(struct page *, int ignore_refs);
Linus Torvalds's avatar
Linus Torvalds committed
89

90
91
92
/*
 * Called from mm/filemap_xip.c to unmap empty zero page
 */
93
pte_t *page_check_address(struct page *, struct mm_struct *,
Nick Piggin's avatar
Nick Piggin committed
94
				unsigned long, spinlock_t **, int);
95

Linus Torvalds's avatar
Linus Torvalds committed
96
97
98
99
100
/*
 * Used by swapoff to help locate where page is expected in vma.
 */
unsigned long page_address_in_vma(struct page *, struct vm_area_struct *);

101
102
103
104
105
106
107
108
/*
 * Cleans the PTEs of shared mappings.
 * (and since clean PTEs should also be readonly, write protects them too)
 *
 * returns the number of cleaned PTEs.
 */
int page_mkclean(struct page *);

109
110
111
112
113
114
/*
 * called in munlock()/munmap() path to check for other vmas holding
 * the page mlocked.
 */
int try_to_munlock(struct page *);

Linus Torvalds's avatar
Linus Torvalds committed
115
116
117
118
119
120
#else	/* !CONFIG_MMU */

#define anon_vma_init()		do {} while (0)
#define anon_vma_prepare(vma)	(0)
#define anon_vma_link(vma)	do {} while (0)

121
122
123
124
125
126
127
128
static inline int page_referenced(struct page *page, int is_locked,
				  struct mem_cgroup *cnt,
				  unsigned long *vm_flags)
{
	*vm_flags = 0;
	return TestClearPageReferenced(page);
}

129
#define try_to_unmap(page, refs) SWAP_FAIL
Linus Torvalds's avatar
Linus Torvalds committed
130

131
132
133
134
135
136
static inline int page_mkclean(struct page *page)
{
	return 0;
}


Linus Torvalds's avatar
Linus Torvalds committed
137
138
139
140
141
142
143
144
#endif	/* CONFIG_MMU */

/*
 * Return values of try_to_unmap
 */
#define SWAP_SUCCESS	0
#define SWAP_AGAIN	1
#define SWAP_FAIL	2
145
#define SWAP_MLOCK	3
Linus Torvalds's avatar
Linus Torvalds committed
146
147

#endif	/* _LINUX_RMAP_H */