]>
Commit | Line | Data |
---|---|---|
b0c9778b TH |
1 | /* |
2 | * mm/percpu-km.c - kernel memory based chunk allocation | |
3 | * | |
4 | * Copyright (C) 2010 SUSE Linux Products GmbH | |
5 | * Copyright (C) 2010 Tejun Heo <[email protected]> | |
6 | * | |
7 | * This file is released under the GPLv2. | |
8 | * | |
9 | * Chunks are allocated as a contiguous kernel memory using gfp | |
10 | * allocation. This is to be used on nommu architectures. | |
11 | * | |
12 | * To use percpu-km, | |
13 | * | |
14 | * - define CONFIG_NEED_PER_CPU_KM from the arch Kconfig. | |
15 | * | |
16 | * - CONFIG_NEED_PER_CPU_PAGE_FIRST_CHUNK must not be defined. It's | |
17 | * not compatible with PER_CPU_KM. EMBED_FIRST_CHUNK should work | |
18 | * fine. | |
19 | * | |
20 | * - NUMA is not supported. When setting up the first chunk, | |
21 | * @cpu_distance_fn should be NULL or report all CPUs to be nearer | |
22 | * than or at LOCAL_DISTANCE. | |
23 | * | |
24 | * - It's best if the chunk size is power of two multiple of | |
25 | * PAGE_SIZE. Because each chunk is allocated as a contiguous | |
26 | * kernel memory block using alloc_pages(), memory will be wasted if | |
27 | * chunk size is not aligned. percpu-km code will whine about it. | |
28 | */ | |
29 | ||
bbddff05 | 30 | #if defined(CONFIG_SMP) && defined(CONFIG_NEED_PER_CPU_PAGE_FIRST_CHUNK) |
b0c9778b TH |
31 | #error "contiguous percpu allocation is incompatible with paged first chunk" |
32 | #endif | |
33 | ||
34 | #include <linux/log2.h> | |
35 | ||
a93ace48 TH |
36 | static int pcpu_populate_chunk(struct pcpu_chunk *chunk, |
37 | int page_start, int page_end) | |
b0c9778b | 38 | { |
b0c9778b TH |
39 | return 0; |
40 | } | |
41 | ||
a93ace48 TH |
42 | static void pcpu_depopulate_chunk(struct pcpu_chunk *chunk, |
43 | int page_start, int page_end) | |
b0c9778b TH |
44 | { |
45 | /* nada */ | |
46 | } | |
47 | ||
48 | static struct pcpu_chunk *pcpu_create_chunk(void) | |
49 | { | |
50 | const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; | |
51 | struct pcpu_chunk *chunk; | |
52 | struct page *pages; | |
53 | int i; | |
54 | ||
55 | chunk = pcpu_alloc_chunk(); | |
56 | if (!chunk) | |
57 | return NULL; | |
58 | ||
59 | pages = alloc_pages(GFP_KERNEL, order_base_2(nr_pages)); | |
60 | if (!pages) { | |
61 | pcpu_free_chunk(chunk); | |
62 | return NULL; | |
63 | } | |
64 | ||
65 | for (i = 0; i < nr_pages; i++) | |
66 | pcpu_set_page_chunk(nth_page(pages, i), chunk); | |
67 | ||
68 | chunk->data = pages; | |
69 | chunk->base_addr = page_address(pages) - pcpu_group_offsets[0]; | |
70 | return chunk; | |
71 | } | |
72 | ||
73 | static void pcpu_destroy_chunk(struct pcpu_chunk *chunk) | |
74 | { | |
75 | const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; | |
76 | ||
77 | if (chunk && chunk->data) | |
78 | __free_pages(chunk->data, order_base_2(nr_pages)); | |
79 | pcpu_free_chunk(chunk); | |
80 | } | |
81 | ||
82 | static struct page *pcpu_addr_to_page(void *addr) | |
83 | { | |
84 | return virt_to_page(addr); | |
85 | } | |
86 | ||
87 | static int __init pcpu_verify_alloc_info(const struct pcpu_alloc_info *ai) | |
88 | { | |
89 | size_t nr_pages, alloc_pages; | |
90 | ||
91 | /* all units must be in a single group */ | |
92 | if (ai->nr_groups != 1) { | |
93 | printk(KERN_CRIT "percpu: can't handle more than one groups\n"); | |
94 | return -EINVAL; | |
95 | } | |
96 | ||
97 | nr_pages = (ai->groups[0].nr_units * ai->unit_size) >> PAGE_SHIFT; | |
98 | alloc_pages = roundup_pow_of_two(nr_pages); | |
99 | ||
100 | if (alloc_pages > nr_pages) | |
101 | printk(KERN_WARNING "percpu: wasting %zu pages per chunk\n", | |
102 | alloc_pages - nr_pages); | |
103 | ||
104 | return 0; | |
105 | } |