]> Git Repo - linux.git/blob - drivers/firmware/efi/unaccepted_memory.c
ACPI: CPPC: Adjust debug messages in amd_set_max_freq_ratio() to warn
[linux.git] / drivers / firmware / efi / unaccepted_memory.c
1 // SPDX-License-Identifier: GPL-2.0-only
2
3 #include <linux/efi.h>
4 #include <linux/memblock.h>
5 #include <linux/spinlock.h>
6 #include <linux/crash_dump.h>
7 #include <linux/nmi.h>
8 #include <asm/unaccepted_memory.h>
9
10 /* Protects unaccepted memory bitmap and accepting_list */
11 static DEFINE_SPINLOCK(unaccepted_memory_lock);
12
13 struct accept_range {
14         struct list_head list;
15         unsigned long start;
16         unsigned long end;
17 };
18
19 static LIST_HEAD(accepting_list);
20
21 /*
22  * accept_memory() -- Consult bitmap and accept the memory if needed.
23  *
24  * Only memory that is explicitly marked as unaccepted in the bitmap requires
25  * an action. All the remaining memory is implicitly accepted and doesn't need
26  * acceptance.
27  *
28  * No need to accept:
29  *  - anything if the system has no unaccepted table;
30  *  - memory that is below phys_base;
31  *  - memory that is above the memory that addressable by the bitmap;
32  */
33 void accept_memory(phys_addr_t start, phys_addr_t end)
34 {
35         struct efi_unaccepted_memory *unaccepted;
36         unsigned long range_start, range_end;
37         struct accept_range range, *entry;
38         unsigned long flags;
39         u64 unit_size;
40
41         unaccepted = efi_get_unaccepted_table();
42         if (!unaccepted)
43                 return;
44
45         unit_size = unaccepted->unit_size;
46
47         /*
48          * Only care for the part of the range that is represented
49          * in the bitmap.
50          */
51         if (start < unaccepted->phys_base)
52                 start = unaccepted->phys_base;
53         if (end < unaccepted->phys_base)
54                 return;
55
56         /* Translate to offsets from the beginning of the bitmap */
57         start -= unaccepted->phys_base;
58         end -= unaccepted->phys_base;
59
60         /*
61          * load_unaligned_zeropad() can lead to unwanted loads across page
62          * boundaries. The unwanted loads are typically harmless. But, they
63          * might be made to totally unrelated or even unmapped memory.
64          * load_unaligned_zeropad() relies on exception fixup (#PF, #GP and now
65          * #VE) to recover from these unwanted loads.
66          *
67          * But, this approach does not work for unaccepted memory. For TDX, a
68          * load from unaccepted memory will not lead to a recoverable exception
69          * within the guest. The guest will exit to the VMM where the only
70          * recourse is to terminate the guest.
71          *
72          * There are two parts to fix this issue and comprehensively avoid
73          * access to unaccepted memory. Together these ensure that an extra
74          * "guard" page is accepted in addition to the memory that needs to be
75          * used:
76          *
77          * 1. Implicitly extend the range_contains_unaccepted_memory(start, end)
78          *    checks up to end+unit_size if 'end' is aligned on a unit_size
79          *    boundary.
80          *
81          * 2. Implicitly extend accept_memory(start, end) to end+unit_size if
82          *    'end' is aligned on a unit_size boundary. (immediately following
83          *    this comment)
84          */
85         if (!(end % unit_size))
86                 end += unit_size;
87
88         /* Make sure not to overrun the bitmap */
89         if (end > unaccepted->size * unit_size * BITS_PER_BYTE)
90                 end = unaccepted->size * unit_size * BITS_PER_BYTE;
91
92         range.start = start / unit_size;
93         range.end = DIV_ROUND_UP(end, unit_size);
94 retry:
95         spin_lock_irqsave(&unaccepted_memory_lock, flags);
96
97         /*
98          * Check if anybody works on accepting the same range of the memory.
99          *
100          * The check is done with unit_size granularity. It is crucial to catch
101          * all accept requests to the same unit_size block, even if they don't
102          * overlap on physical address level.
103          */
104         list_for_each_entry(entry, &accepting_list, list) {
105                 if (entry->end <= range.start)
106                         continue;
107                 if (entry->start >= range.end)
108                         continue;
109
110                 /*
111                  * Somebody else accepting the range. Or at least part of it.
112                  *
113                  * Drop the lock and retry until it is complete.
114                  */
115                 spin_unlock_irqrestore(&unaccepted_memory_lock, flags);
116                 goto retry;
117         }
118
119         /*
120          * Register that the range is about to be accepted.
121          * Make sure nobody else will accept it.
122          */
123         list_add(&range.list, &accepting_list);
124
125         range_start = range.start;
126         for_each_set_bitrange_from(range_start, range_end, unaccepted->bitmap,
127                                    range.end) {
128                 unsigned long phys_start, phys_end;
129                 unsigned long len = range_end - range_start;
130
131                 phys_start = range_start * unit_size + unaccepted->phys_base;
132                 phys_end = range_end * unit_size + unaccepted->phys_base;
133
134                 /*
135                  * Keep interrupts disabled until the accept operation is
136                  * complete in order to prevent deadlocks.
137                  *
138                  * Enabling interrupts before calling arch_accept_memory()
139                  * creates an opportunity for an interrupt handler to request
140                  * acceptance for the same memory. The handler will continuously
141                  * spin with interrupts disabled, preventing other task from
142                  * making progress with the acceptance process.
143                  */
144                 spin_unlock(&unaccepted_memory_lock);
145
146                 arch_accept_memory(phys_start, phys_end);
147
148                 spin_lock(&unaccepted_memory_lock);
149                 bitmap_clear(unaccepted->bitmap, range_start, len);
150         }
151
152         list_del(&range.list);
153
154         touch_softlockup_watchdog();
155
156         spin_unlock_irqrestore(&unaccepted_memory_lock, flags);
157 }
158
159 bool range_contains_unaccepted_memory(phys_addr_t start, phys_addr_t end)
160 {
161         struct efi_unaccepted_memory *unaccepted;
162         unsigned long flags;
163         bool ret = false;
164         u64 unit_size;
165
166         unaccepted = efi_get_unaccepted_table();
167         if (!unaccepted)
168                 return false;
169
170         unit_size = unaccepted->unit_size;
171
172         /*
173          * Only care for the part of the range that is represented
174          * in the bitmap.
175          */
176         if (start < unaccepted->phys_base)
177                 start = unaccepted->phys_base;
178         if (end < unaccepted->phys_base)
179                 return false;
180
181         /* Translate to offsets from the beginning of the bitmap */
182         start -= unaccepted->phys_base;
183         end -= unaccepted->phys_base;
184
185         /*
186          * Also consider the unaccepted state of the *next* page. See fix #1 in
187          * the comment on load_unaligned_zeropad() in accept_memory().
188          */
189         if (!(end % unit_size))
190                 end += unit_size;
191
192         /* Make sure not to overrun the bitmap */
193         if (end > unaccepted->size * unit_size * BITS_PER_BYTE)
194                 end = unaccepted->size * unit_size * BITS_PER_BYTE;
195
196         spin_lock_irqsave(&unaccepted_memory_lock, flags);
197         while (start < end) {
198                 if (test_bit(start / unit_size, unaccepted->bitmap)) {
199                         ret = true;
200                         break;
201                 }
202
203                 start += unit_size;
204         }
205         spin_unlock_irqrestore(&unaccepted_memory_lock, flags);
206
207         return ret;
208 }
209
210 #ifdef CONFIG_PROC_VMCORE
211 static bool unaccepted_memory_vmcore_pfn_is_ram(struct vmcore_cb *cb,
212                                                 unsigned long pfn)
213 {
214         return !pfn_is_unaccepted_memory(pfn);
215 }
216
217 static struct vmcore_cb vmcore_cb = {
218         .pfn_is_ram = unaccepted_memory_vmcore_pfn_is_ram,
219 };
220
221 static int __init unaccepted_memory_init_kdump(void)
222 {
223         register_vmcore_cb(&vmcore_cb);
224         return 0;
225 }
226 core_initcall(unaccepted_memory_init_kdump);
227 #endif /* CONFIG_PROC_VMCORE */
This page took 0.042746 seconds and 4 git commands to generate.