1 #ifndef _LINUX_KHUGEPAGED_H
2 #define _LINUX_KHUGEPAGED_H
4 #include <linux/sched.h> /* MMF_VM_HUGEPAGE */
6 #ifdef CONFIG_TRANSPARENT_HUGEPAGE
7 extern int __khugepaged_enter(struct mm_struct *mm);
8 extern void __khugepaged_exit(struct mm_struct *mm);
9 extern int khugepaged_enter_vma_merge(struct vm_area_struct *vma,
10 unsigned long vm_flags);
12 #define khugepaged_enabled() \
13 (transparent_hugepage_flags & \
14 ((1<<TRANSPARENT_HUGEPAGE_FLAG) | \
15 (1<<TRANSPARENT_HUGEPAGE_REQ_MADV_FLAG)))
16 #define khugepaged_always() \
17 (transparent_hugepage_flags & \
18 (1<<TRANSPARENT_HUGEPAGE_FLAG))
19 #define khugepaged_req_madv() \
20 (transparent_hugepage_flags & \
21 (1<<TRANSPARENT_HUGEPAGE_REQ_MADV_FLAG))
22 #define khugepaged_defrag() \
23 (transparent_hugepage_flags & \
24 (1<<TRANSPARENT_HUGEPAGE_DEFRAG_KHUGEPAGED_FLAG))
26 static inline int khugepaged_fork(struct mm_struct *mm, struct mm_struct *oldmm)
28 if (test_bit(MMF_VM_HUGEPAGE, &oldmm->flags))
29 return __khugepaged_enter(mm);
33 static inline void khugepaged_exit(struct mm_struct *mm)
35 if (test_bit(MMF_VM_HUGEPAGE, &mm->flags))
36 __khugepaged_exit(mm);
39 static inline int khugepaged_enter(struct vm_area_struct *vma,
40 unsigned long vm_flags)
42 if (!test_bit(MMF_VM_HUGEPAGE, &vma->vm_mm->flags))
43 if ((khugepaged_always() ||
44 (khugepaged_req_madv() && (vm_flags & VM_HUGEPAGE))) &&
45 !(vm_flags & VM_NOHUGEPAGE))
46 if (__khugepaged_enter(vma->vm_mm))
50 #else /* CONFIG_TRANSPARENT_HUGEPAGE */
51 static inline int khugepaged_fork(struct mm_struct *mm, struct mm_struct *oldmm)
55 static inline void khugepaged_exit(struct mm_struct *mm)
58 static inline int khugepaged_enter(struct vm_area_struct *vma,
59 unsigned long vm_flags)
63 static inline int khugepaged_enter_vma_merge(struct vm_area_struct *vma,
64 unsigned long vm_flags)
68 #endif /* CONFIG_TRANSPARENT_HUGEPAGE */
70 #endif /* _LINUX_KHUGEPAGED_H */