]> Git Repo - linux.git/blob - drivers/md/dm-vdo/flush.c
Merge tag 'kbuild-v6.9' of git://git.kernel.org/pub/scm/linux/kernel/git/masahiroy...
[linux.git] / drivers / md / dm-vdo / flush.c
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright 2023 Red Hat
4  */
5
6 #include "flush.h"
7
8 #include <linux/mempool.h>
9 #include <linux/spinlock.h>
10
11 #include "logger.h"
12 #include "memory-alloc.h"
13 #include "permassert.h"
14
15 #include "admin-state.h"
16 #include "completion.h"
17 #include "io-submitter.h"
18 #include "logical-zone.h"
19 #include "slab-depot.h"
20 #include "types.h"
21 #include "vdo.h"
22
23 struct flusher {
24         struct vdo_completion completion;
25         /* The vdo to which this flusher belongs */
26         struct vdo *vdo;
27         /* The administrative state of the flusher */
28         struct admin_state state;
29         /* The current flush generation of the vdo */
30         sequence_number_t flush_generation;
31         /* The first unacknowledged flush generation */
32         sequence_number_t first_unacknowledged_generation;
33         /* The queue of flush requests waiting to notify other threads */
34         struct vdo_wait_queue notifiers;
35         /* The queue of flush requests waiting for VIOs to complete */
36         struct vdo_wait_queue pending_flushes;
37         /* The flush generation for which notifications are being sent */
38         sequence_number_t notify_generation;
39         /* The logical zone to notify next */
40         struct logical_zone *logical_zone_to_notify;
41         /* The ID of the thread on which flush requests should be made */
42         thread_id_t thread_id;
43         /* The pool of flush requests */
44         mempool_t *flush_pool;
45         /* Bios waiting for a flush request to become available */
46         struct bio_list waiting_flush_bios;
47         /* The lock to protect the previous fields */
48         spinlock_t lock;
49         /* The rotor for selecting the bio queue for submitting flush bios */
50         zone_count_t bio_queue_rotor;
51         /* The number of flushes submitted to the current bio queue */
52         int flush_count;
53 };
54
55 /**
56  * assert_on_flusher_thread() - Check that we are on the flusher thread.
57  * @flusher: The flusher.
58  * @caller: The function which is asserting.
59  */
60 static inline void assert_on_flusher_thread(struct flusher *flusher, const char *caller)
61 {
62         VDO_ASSERT_LOG_ONLY((vdo_get_callback_thread_id() == flusher->thread_id),
63                             "%s() called from flusher thread", caller);
64 }
65
66 /**
67  * as_flusher() - Convert a generic vdo_completion to a flusher.
68  * @completion: The completion to convert.
69  *
70  * Return: The completion as a flusher.
71  */
72 static struct flusher *as_flusher(struct vdo_completion *completion)
73 {
74         vdo_assert_completion_type(completion, VDO_FLUSH_NOTIFICATION_COMPLETION);
75         return container_of(completion, struct flusher, completion);
76 }
77
78 /**
79  * completion_as_vdo_flush() - Convert a generic vdo_completion to a vdo_flush.
80  * @completion: The completion to convert.
81  *
82  * Return: The completion as a vdo_flush.
83  */
84 static inline struct vdo_flush *completion_as_vdo_flush(struct vdo_completion *completion)
85 {
86         vdo_assert_completion_type(completion, VDO_FLUSH_COMPLETION);
87         return container_of(completion, struct vdo_flush, completion);
88 }
89
90 /**
91  * vdo_waiter_as_flush() - Convert a vdo_flush's generic wait queue entry back to the vdo_flush.
92  * @waiter: The wait queue entry to convert.
93  *
94  * Return: The wait queue entry as a vdo_flush.
95  */
96 static struct vdo_flush *vdo_waiter_as_flush(struct vdo_waiter *waiter)
97 {
98         return container_of(waiter, struct vdo_flush, waiter);
99 }
100
101 static void *allocate_flush(gfp_t gfp_mask, void *pool_data)
102 {
103         struct vdo_flush *flush = NULL;
104
105         if ((gfp_mask & GFP_NOWAIT) == GFP_NOWAIT) {
106                 flush = vdo_allocate_memory_nowait(sizeof(struct vdo_flush), __func__);
107         } else {
108                 int result = vdo_allocate(1, struct vdo_flush, __func__, &flush);
109
110                 if (result != VDO_SUCCESS)
111                         vdo_log_error_strerror(result, "failed to allocate spare flush");
112         }
113
114         if (flush != NULL) {
115                 struct flusher *flusher = pool_data;
116
117                 vdo_initialize_completion(&flush->completion, flusher->vdo,
118                                           VDO_FLUSH_COMPLETION);
119         }
120
121         return flush;
122 }
123
124 static void free_flush(void *element, void *pool_data __always_unused)
125 {
126         vdo_free(element);
127 }
128
129 /**
130  * vdo_make_flusher() - Make a flusher for a vdo.
131  * @vdo: The vdo which owns the flusher.
132  *
133  * Return: VDO_SUCCESS or an error.
134  */
135 int vdo_make_flusher(struct vdo *vdo)
136 {
137         int result = vdo_allocate(1, struct flusher, __func__, &vdo->flusher);
138
139         if (result != VDO_SUCCESS)
140                 return result;
141
142         vdo->flusher->vdo = vdo;
143         vdo->flusher->thread_id = vdo->thread_config.packer_thread;
144         vdo_set_admin_state_code(&vdo->flusher->state, VDO_ADMIN_STATE_NORMAL_OPERATION);
145         vdo_initialize_completion(&vdo->flusher->completion, vdo,
146                                   VDO_FLUSH_NOTIFICATION_COMPLETION);
147
148         spin_lock_init(&vdo->flusher->lock);
149         bio_list_init(&vdo->flusher->waiting_flush_bios);
150         vdo->flusher->flush_pool = mempool_create(1, allocate_flush, free_flush,
151                                                   vdo->flusher);
152         return ((vdo->flusher->flush_pool == NULL) ? -ENOMEM : VDO_SUCCESS);
153 }
154
155 /**
156  * vdo_free_flusher() - Free a flusher.
157  * @flusher: The flusher to free.
158  */
159 void vdo_free_flusher(struct flusher *flusher)
160 {
161         if (flusher == NULL)
162                 return;
163
164         if (flusher->flush_pool != NULL)
165                 mempool_destroy(vdo_forget(flusher->flush_pool));
166         vdo_free(flusher);
167 }
168
169 /**
170  * vdo_get_flusher_thread_id() - Get the ID of the thread on which flusher functions should be
171  *                               called.
172  * @flusher: The flusher to query.
173  *
174  * Return: The ID of the thread which handles the flusher.
175  */
176 thread_id_t vdo_get_flusher_thread_id(struct flusher *flusher)
177 {
178         return flusher->thread_id;
179 }
180
181 static void notify_flush(struct flusher *flusher);
182 static void vdo_complete_flush(struct vdo_flush *flush);
183
184 /**
185  * finish_notification() - Finish the notification process.
186  * @completion: The flusher completion.
187  *
188  * Finishes the notification process by checking if any flushes have completed and then starting
189  * the notification of the next flush request if one came in while the current notification was in
190  * progress. This callback is registered in flush_packer_callback().
191  */
192 static void finish_notification(struct vdo_completion *completion)
193 {
194         struct flusher *flusher = as_flusher(completion);
195
196         assert_on_flusher_thread(flusher, __func__);
197
198         vdo_waitq_enqueue_waiter(&flusher->pending_flushes,
199                                  vdo_waitq_dequeue_waiter(&flusher->notifiers));
200         vdo_complete_flushes(flusher);
201         if (vdo_waitq_has_waiters(&flusher->notifiers))
202                 notify_flush(flusher);
203 }
204
205 /**
206  * flush_packer_callback() - Flush the packer.
207  * @completion: The flusher completion.
208  *
209  * Flushes the packer now that all of the logical and physical zones have been notified of the new
210  * flush request. This callback is registered in increment_generation().
211  */
212 static void flush_packer_callback(struct vdo_completion *completion)
213 {
214         struct flusher *flusher = as_flusher(completion);
215
216         vdo_increment_packer_flush_generation(flusher->vdo->packer);
217         vdo_launch_completion_callback(completion, finish_notification,
218                                        flusher->thread_id);
219 }
220
221 /**
222  * increment_generation() - Increment the flush generation in a logical zone.
223  * @completion: The flusher as a completion.
224  *
225  * If there are more logical zones, go on to the next one, otherwise, prepare the physical zones.
226  * This callback is registered both in notify_flush() and in itself.
227  */
228 static void increment_generation(struct vdo_completion *completion)
229 {
230         struct flusher *flusher = as_flusher(completion);
231         struct logical_zone *zone = flusher->logical_zone_to_notify;
232
233         vdo_increment_logical_zone_flush_generation(zone, flusher->notify_generation);
234         if (zone->next == NULL) {
235                 vdo_launch_completion_callback(completion, flush_packer_callback,
236                                                flusher->thread_id);
237                 return;
238         }
239
240         flusher->logical_zone_to_notify = zone->next;
241         vdo_launch_completion_callback(completion, increment_generation,
242                                        flusher->logical_zone_to_notify->thread_id);
243 }
244
245 /**
246  * notify_flush() - Launch a flush notification.
247  * @flusher: The flusher doing the notification.
248  */
249 static void notify_flush(struct flusher *flusher)
250 {
251         struct vdo_flush *flush =
252                 vdo_waiter_as_flush(vdo_waitq_get_first_waiter(&flusher->notifiers));
253
254         flusher->notify_generation = flush->flush_generation;
255         flusher->logical_zone_to_notify = &flusher->vdo->logical_zones->zones[0];
256         flusher->completion.requeue = true;
257         vdo_launch_completion_callback(&flusher->completion, increment_generation,
258                                        flusher->logical_zone_to_notify->thread_id);
259 }
260
261 /**
262  * flush_vdo() - Start processing a flush request.
263  * @completion: A flush request (as a vdo_completion)
264  *
265  * This callback is registered in launch_flush().
266  */
267 static void flush_vdo(struct vdo_completion *completion)
268 {
269         struct vdo_flush *flush = completion_as_vdo_flush(completion);
270         struct flusher *flusher = completion->vdo->flusher;
271         bool may_notify;
272         int result;
273
274         assert_on_flusher_thread(flusher, __func__);
275         result = VDO_ASSERT(vdo_is_state_normal(&flusher->state),
276                             "flusher is in normal operation");
277         if (result != VDO_SUCCESS) {
278                 vdo_enter_read_only_mode(flusher->vdo, result);
279                 vdo_complete_flush(flush);
280                 return;
281         }
282
283         flush->flush_generation = flusher->flush_generation++;
284         may_notify = !vdo_waitq_has_waiters(&flusher->notifiers);
285         vdo_waitq_enqueue_waiter(&flusher->notifiers, &flush->waiter);
286         if (may_notify)
287                 notify_flush(flusher);
288 }
289
290 /**
291  * check_for_drain_complete() - Check whether the flusher has drained.
292  * @flusher: The flusher.
293  */
294 static void check_for_drain_complete(struct flusher *flusher)
295 {
296         bool drained;
297
298         if (!vdo_is_state_draining(&flusher->state) ||
299             vdo_waitq_has_waiters(&flusher->pending_flushes))
300                 return;
301
302         spin_lock(&flusher->lock);
303         drained = bio_list_empty(&flusher->waiting_flush_bios);
304         spin_unlock(&flusher->lock);
305
306         if (drained)
307                 vdo_finish_draining(&flusher->state);
308 }
309
310 /**
311  * vdo_complete_flushes() - Attempt to complete any flushes which might have finished.
312  * @flusher: The flusher.
313  */
314 void vdo_complete_flushes(struct flusher *flusher)
315 {
316         sequence_number_t oldest_active_generation = U64_MAX;
317         struct logical_zone *zone;
318
319         assert_on_flusher_thread(flusher, __func__);
320
321         for (zone = &flusher->vdo->logical_zones->zones[0]; zone != NULL; zone = zone->next)
322                 oldest_active_generation =
323                         min(oldest_active_generation,
324                             READ_ONCE(zone->oldest_active_generation));
325
326         while (vdo_waitq_has_waiters(&flusher->pending_flushes)) {
327                 struct vdo_flush *flush =
328                         vdo_waiter_as_flush(vdo_waitq_get_first_waiter(&flusher->pending_flushes));
329
330                 if (flush->flush_generation >= oldest_active_generation)
331                         return;
332
333                 VDO_ASSERT_LOG_ONLY((flush->flush_generation ==
334                                      flusher->first_unacknowledged_generation),
335                                     "acknowledged next expected flush, %llu, was: %llu",
336                                     (unsigned long long) flusher->first_unacknowledged_generation,
337                                     (unsigned long long) flush->flush_generation);
338                 vdo_waitq_dequeue_waiter(&flusher->pending_flushes);
339                 vdo_complete_flush(flush);
340                 flusher->first_unacknowledged_generation++;
341         }
342
343         check_for_drain_complete(flusher);
344 }
345
346 /**
347  * vdo_dump_flusher() - Dump the flusher, in a thread-unsafe fashion.
348  * @flusher: The flusher.
349  */
350 void vdo_dump_flusher(const struct flusher *flusher)
351 {
352         vdo_log_info("struct flusher");
353         vdo_log_info("  flush_generation=%llu first_unacknowledged_generation=%llu",
354                      (unsigned long long) flusher->flush_generation,
355                      (unsigned long long) flusher->first_unacknowledged_generation);
356         vdo_log_info("  notifiers queue is %s; pending_flushes queue is %s",
357                      (vdo_waitq_has_waiters(&flusher->notifiers) ? "not empty" : "empty"),
358                      (vdo_waitq_has_waiters(&flusher->pending_flushes) ? "not empty" : "empty"));
359 }
360
361 /**
362  * initialize_flush() - Initialize a vdo_flush structure.
363  * @flush: The flush to initialize.
364  * @vdo: The vdo being flushed.
365  *
366  * Initializes a vdo_flush structure, transferring all the bios in the flusher's waiting_flush_bios
367  * list to it. The caller MUST already hold the lock.
368  */
369 static void initialize_flush(struct vdo_flush *flush, struct vdo *vdo)
370 {
371         bio_list_init(&flush->bios);
372         bio_list_merge(&flush->bios, &vdo->flusher->waiting_flush_bios);
373         bio_list_init(&vdo->flusher->waiting_flush_bios);
374 }
375
376 static void launch_flush(struct vdo_flush *flush)
377 {
378         struct vdo_completion *completion = &flush->completion;
379
380         vdo_prepare_completion(completion, flush_vdo, flush_vdo,
381                                completion->vdo->thread_config.packer_thread, NULL);
382         vdo_enqueue_completion(completion, VDO_DEFAULT_Q_FLUSH_PRIORITY);
383 }
384
385 /**
386  * vdo_launch_flush() - Function called to start processing a flush request.
387  * @vdo: The vdo.
388  * @bio: The bio containing an empty flush request.
389  *
390  * This is called when we receive an empty flush bio from the block layer, and before acknowledging
391  * a non-empty bio with the FUA flag set.
392  */
393 void vdo_launch_flush(struct vdo *vdo, struct bio *bio)
394 {
395         /*
396          * Try to allocate a vdo_flush to represent the flush request. If the allocation fails,
397          * we'll deal with it later.
398          */
399         struct vdo_flush *flush = mempool_alloc(vdo->flusher->flush_pool, GFP_NOWAIT);
400         struct flusher *flusher = vdo->flusher;
401         const struct admin_state_code *code = vdo_get_admin_state_code(&flusher->state);
402
403         VDO_ASSERT_LOG_ONLY(!code->quiescent, "Flushing not allowed in state %s",
404                             code->name);
405
406         spin_lock(&flusher->lock);
407
408         /* We have a new bio to start. Add it to the list. */
409         bio_list_add(&flusher->waiting_flush_bios, bio);
410
411         if (flush == NULL) {
412                 spin_unlock(&flusher->lock);
413                 return;
414         }
415
416         /* We have flushes to start. Capture them in the vdo_flush structure. */
417         initialize_flush(flush, vdo);
418         spin_unlock(&flusher->lock);
419
420         /* Finish launching the flushes. */
421         launch_flush(flush);
422 }
423
424 /**
425  * release_flush() - Release a vdo_flush structure that has completed its work.
426  * @flush: The completed flush structure to re-use or free.
427  *
428  * If there are any pending flush requests whose vdo_flush allocation failed, they will be launched
429  * by immediately re-using the released vdo_flush. If there is no spare vdo_flush, the released
430  * structure will become the spare. Otherwise, the vdo_flush will be freed.
431  */
432 static void release_flush(struct vdo_flush *flush)
433 {
434         bool relaunch_flush;
435         struct flusher *flusher = flush->completion.vdo->flusher;
436
437         spin_lock(&flusher->lock);
438         if (bio_list_empty(&flusher->waiting_flush_bios)) {
439                 relaunch_flush = false;
440         } else {
441                 /* We have flushes to start. Capture them in a flush request. */
442                 initialize_flush(flush, flusher->vdo);
443                 relaunch_flush = true;
444         }
445         spin_unlock(&flusher->lock);
446
447         if (relaunch_flush) {
448                 /* Finish launching the flushes. */
449                 launch_flush(flush);
450                 return;
451         }
452
453         mempool_free(flush, flusher->flush_pool);
454 }
455
456 /**
457  * vdo_complete_flush_callback() - Function called to complete and free a flush request, registered
458  *                                 in vdo_complete_flush().
459  * @completion: The flush request.
460  */
461 static void vdo_complete_flush_callback(struct vdo_completion *completion)
462 {
463         struct vdo_flush *flush = completion_as_vdo_flush(completion);
464         struct vdo *vdo = completion->vdo;
465         struct bio *bio;
466
467         while ((bio = bio_list_pop(&flush->bios)) != NULL) {
468                 /*
469                  * We're not acknowledging this bio now, but we'll never touch it again, so this is
470                  * the last chance to account for it.
471                  */
472                 vdo_count_bios(&vdo->stats.bios_acknowledged, bio);
473
474                 /* Update the device, and send it on down... */
475                 bio_set_dev(bio, vdo_get_backing_device(vdo));
476                 atomic64_inc(&vdo->stats.flush_out);
477                 submit_bio_noacct(bio);
478         }
479
480
481         /*
482          * Release the flush structure, freeing it, re-using it as the spare, or using it to launch
483          * any flushes that had to wait when allocations failed.
484          */
485         release_flush(flush);
486 }
487
488 /**
489  * select_bio_queue() - Select the bio queue on which to finish a flush request.
490  * @flusher: The flusher finishing the request.
491  */
492 static thread_id_t select_bio_queue(struct flusher *flusher)
493 {
494         struct vdo *vdo = flusher->vdo;
495         zone_count_t bio_threads = flusher->vdo->thread_config.bio_thread_count;
496         int interval;
497
498         if (bio_threads == 1)
499                 return vdo->thread_config.bio_threads[0];
500
501         interval = vdo->device_config->thread_counts.bio_rotation_interval;
502         if (flusher->flush_count == interval) {
503                 flusher->flush_count = 1;
504                 flusher->bio_queue_rotor = ((flusher->bio_queue_rotor + 1) % bio_threads);
505         } else {
506                 flusher->flush_count++;
507         }
508
509         return vdo->thread_config.bio_threads[flusher->bio_queue_rotor];
510 }
511
512 /**
513  * vdo_complete_flush() - Complete and free a vdo flush request.
514  * @flush: The flush request.
515  */
516 static void vdo_complete_flush(struct vdo_flush *flush)
517 {
518         struct vdo_completion *completion = &flush->completion;
519
520         vdo_prepare_completion(completion, vdo_complete_flush_callback,
521                                vdo_complete_flush_callback,
522                                select_bio_queue(completion->vdo->flusher), NULL);
523         vdo_enqueue_completion(completion, BIO_Q_FLUSH_PRIORITY);
524 }
525
526 /**
527  * initiate_drain() - Initiate a drain.
528  *
529  * Implements vdo_admin_initiator_fn.
530  */
531 static void initiate_drain(struct admin_state *state)
532 {
533         check_for_drain_complete(container_of(state, struct flusher, state));
534 }
535
536 /**
537  * vdo_drain_flusher() - Drain the flusher.
538  * @flusher: The flusher to drain.
539  * @completion: The completion to finish when the flusher has drained.
540  *
541  * Drains the flusher by preventing any more VIOs from entering the flusher and then flushing. The
542  * flusher will be left in the suspended state.
543  */
544 void vdo_drain_flusher(struct flusher *flusher, struct vdo_completion *completion)
545 {
546         assert_on_flusher_thread(flusher, __func__);
547         vdo_start_draining(&flusher->state, VDO_ADMIN_STATE_SUSPENDING, completion,
548                            initiate_drain);
549 }
550
551 /**
552  * vdo_resume_flusher() - Resume a flusher which has been suspended.
553  * @flusher: The flusher to resume.
554  * @parent: The completion to finish when the flusher has resumed.
555  */
556 void vdo_resume_flusher(struct flusher *flusher, struct vdo_completion *parent)
557 {
558         assert_on_flusher_thread(flusher, __func__);
559         vdo_continue_completion(parent, vdo_resume_if_quiescent(&flusher->state));
560 }
This page took 0.055361 seconds and 4 git commands to generate.