]> Git Repo - J-linux.git/blob - arch/arm/mm/mmap.c
Merge tag 'vfs-6.13-rc7.fixes' of git://git.kernel.org/pub/scm/linux/kernel/git/vfs/vfs
[J-linux.git] / arch / arm / mm / mmap.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  *  linux/arch/arm/mm/mmap.c
4  */
5 #include <linux/fs.h>
6 #include <linux/mm.h>
7 #include <linux/mman.h>
8 #include <linux/shm.h>
9 #include <linux/sched/signal.h>
10 #include <linux/sched/mm.h>
11 #include <linux/io.h>
12 #include <linux/personality.h>
13 #include <linux/random.h>
14 #include <asm/cachetype.h>
15
16 #define COLOUR_ALIGN(addr,pgoff)                \
17         ((((addr)+SHMLBA-1)&~(SHMLBA-1)) +      \
18          (((pgoff)<<PAGE_SHIFT) & (SHMLBA-1)))
19
20 /*
21  * We need to ensure that shared mappings are correctly aligned to
22  * avoid aliasing issues with VIPT caches.  We need to ensure that
23  * a specific page of an object is always mapped at a multiple of
24  * SHMLBA bytes.
25  *
26  * We unconditionally provide this function for all cases, however
27  * in the VIVT case, we optimise out the alignment rules.
28  */
29 unsigned long
30 arch_get_unmapped_area(struct file *filp, unsigned long addr,
31                 unsigned long len, unsigned long pgoff,
32                 unsigned long flags, vm_flags_t vm_flags)
33 {
34         struct mm_struct *mm = current->mm;
35         struct vm_area_struct *vma;
36         int do_align = 0;
37         int aliasing = cache_is_vipt_aliasing();
38         struct vm_unmapped_area_info info = {};
39
40         /*
41          * We only need to do colour alignment if either the I or D
42          * caches alias.
43          */
44         if (aliasing)
45                 do_align = filp || (flags & MAP_SHARED);
46
47         /*
48          * We enforce the MAP_FIXED case.
49          */
50         if (flags & MAP_FIXED) {
51                 if (aliasing && flags & MAP_SHARED &&
52                     (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))
53                         return -EINVAL;
54                 return addr;
55         }
56
57         if (len > TASK_SIZE)
58                 return -ENOMEM;
59
60         if (addr) {
61                 if (do_align)
62                         addr = COLOUR_ALIGN(addr, pgoff);
63                 else
64                         addr = PAGE_ALIGN(addr);
65
66                 vma = find_vma(mm, addr);
67                 if (TASK_SIZE - len >= addr &&
68                     (!vma || addr + len <= vm_start_gap(vma)))
69                         return addr;
70         }
71
72         info.length = len;
73         info.low_limit = mm->mmap_base;
74         info.high_limit = TASK_SIZE;
75         info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0;
76         info.align_offset = pgoff << PAGE_SHIFT;
77         return vm_unmapped_area(&info);
78 }
79
80 unsigned long
81 arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0,
82                         const unsigned long len, const unsigned long pgoff,
83                         const unsigned long flags, vm_flags_t vm_flags)
84 {
85         struct vm_area_struct *vma;
86         struct mm_struct *mm = current->mm;
87         unsigned long addr = addr0;
88         int do_align = 0;
89         int aliasing = cache_is_vipt_aliasing();
90         struct vm_unmapped_area_info info = {};
91
92         /*
93          * We only need to do colour alignment if either the I or D
94          * caches alias.
95          */
96         if (aliasing)
97                 do_align = filp || (flags & MAP_SHARED);
98
99         /* requested length too big for entire address space */
100         if (len > TASK_SIZE)
101                 return -ENOMEM;
102
103         if (flags & MAP_FIXED) {
104                 if (aliasing && flags & MAP_SHARED &&
105                     (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))
106                         return -EINVAL;
107                 return addr;
108         }
109
110         /* requesting a specific address */
111         if (addr) {
112                 if (do_align)
113                         addr = COLOUR_ALIGN(addr, pgoff);
114                 else
115                         addr = PAGE_ALIGN(addr);
116                 vma = find_vma(mm, addr);
117                 if (TASK_SIZE - len >= addr &&
118                                 (!vma || addr + len <= vm_start_gap(vma)))
119                         return addr;
120         }
121
122         info.flags = VM_UNMAPPED_AREA_TOPDOWN;
123         info.length = len;
124         info.low_limit = FIRST_USER_ADDRESS;
125         info.high_limit = mm->mmap_base;
126         info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0;
127         info.align_offset = pgoff << PAGE_SHIFT;
128         addr = vm_unmapped_area(&info);
129
130         /*
131          * A failed mmap() very likely causes application failure,
132          * so fall back to the bottom-up function here. This scenario
133          * can happen with large stack limits and large mmap()
134          * allocations.
135          */
136         if (addr & ~PAGE_MASK) {
137                 VM_BUG_ON(addr != -ENOMEM);
138                 info.flags = 0;
139                 info.low_limit = mm->mmap_base;
140                 info.high_limit = TASK_SIZE;
141                 addr = vm_unmapped_area(&info);
142         }
143
144         return addr;
145 }
146
147 /*
148  * You really shouldn't be using read() or write() on /dev/mem.  This
149  * might go away in the future.
150  */
151 int valid_phys_addr_range(phys_addr_t addr, size_t size)
152 {
153         if (addr < PHYS_OFFSET)
154                 return 0;
155         if (addr + size > __pa(high_memory - 1) + 1)
156                 return 0;
157
158         return 1;
159 }
160
161 /*
162  * Do not allow /dev/mem mappings beyond the supported physical range.
163  */
164 int valid_mmap_phys_addr_range(unsigned long pfn, size_t size)
165 {
166         return (pfn + (size >> PAGE_SHIFT)) <= (1 + (PHYS_MASK >> PAGE_SHIFT));
167 }
This page took 0.035613 seconds and 4 git commands to generate.