123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139 |
- /*
- * Copyright (C) 2011-2017 Red Hat, Inc.
- *
- * This file is released under the GPL.
- */
- #ifndef DM_BIO_PRISON_H
- #define DM_BIO_PRISON_H
- #include "persistent-data/dm-block-manager.h" /* FIXME: for dm_block_t */
- #include "dm-thin-metadata.h" /* FIXME: for dm_thin_id */
- #include <linux/bio.h>
- #include <linux/rbtree.h>
- /*----------------------------------------------------------------*/
- /*
- * Sometimes we can't deal with a bio straight away. We put them in prison
- * where they can't cause any mischief. Bios are put in a cell identified
- * by a key, multiple bios can be in the same cell. When the cell is
- * subsequently unlocked the bios become available.
- */
- struct dm_bio_prison;
- /*
- * Keys define a range of blocks within either a virtual or physical
- * device.
- */
- struct dm_cell_key {
- int virtual;
- dm_thin_id dev;
- dm_block_t block_begin, block_end;
- };
- /*
- * Treat this as opaque, only in header so callers can manage allocation
- * themselves.
- */
- struct dm_bio_prison_cell {
- struct list_head user_list; /* for client use */
- struct rb_node node;
- struct dm_cell_key key;
- struct bio *holder;
- struct bio_list bios;
- };
- struct dm_bio_prison *dm_bio_prison_create(void);
- void dm_bio_prison_destroy(struct dm_bio_prison *prison);
- /*
- * These two functions just wrap a mempool. This is a transitory step:
- * Eventually all bio prison clients should manage their own cell memory.
- *
- * Like mempool_alloc(), dm_bio_prison_alloc_cell() can only fail if called
- * in interrupt context or passed GFP_NOWAIT.
- */
- struct dm_bio_prison_cell *dm_bio_prison_alloc_cell(struct dm_bio_prison *prison,
- gfp_t gfp);
- void dm_bio_prison_free_cell(struct dm_bio_prison *prison,
- struct dm_bio_prison_cell *cell);
- /*
- * Creates, or retrieves a cell that overlaps the given key.
- *
- * Returns 1 if pre-existing cell returned, zero if new cell created using
- * @cell_prealloc.
- */
- int dm_get_cell(struct dm_bio_prison *prison,
- struct dm_cell_key *key,
- struct dm_bio_prison_cell *cell_prealloc,
- struct dm_bio_prison_cell **cell_result);
- /*
- * An atomic op that combines retrieving or creating a cell, and adding a
- * bio to it.
- *
- * Returns 1 if the cell was already held, 0 if @inmate is the new holder.
- */
- int dm_bio_detain(struct dm_bio_prison *prison,
- struct dm_cell_key *key,
- struct bio *inmate,
- struct dm_bio_prison_cell *cell_prealloc,
- struct dm_bio_prison_cell **cell_result);
- void dm_cell_release(struct dm_bio_prison *prison,
- struct dm_bio_prison_cell *cell,
- struct bio_list *bios);
- void dm_cell_release_no_holder(struct dm_bio_prison *prison,
- struct dm_bio_prison_cell *cell,
- struct bio_list *inmates);
- void dm_cell_error(struct dm_bio_prison *prison,
- struct dm_bio_prison_cell *cell, blk_status_t error);
- /*
- * Visits the cell and then releases. Guarantees no new inmates are
- * inserted between the visit and release.
- */
- void dm_cell_visit_release(struct dm_bio_prison *prison,
- void (*visit_fn)(void *, struct dm_bio_prison_cell *),
- void *context, struct dm_bio_prison_cell *cell);
- /*
- * Rather than always releasing the prisoners in a cell, the client may
- * want to promote one of them to be the new holder. There is a race here
- * though between releasing an empty cell, and other threads adding new
- * inmates. So this function makes the decision with its lock held.
- *
- * This function can have two outcomes:
- * i) An inmate is promoted to be the holder of the cell (return value of 0).
- * ii) The cell has no inmate for promotion and is released (return value of 1).
- */
- int dm_cell_promote_or_release(struct dm_bio_prison *prison,
- struct dm_bio_prison_cell *cell);
- /*----------------------------------------------------------------*/
- /*
- * We use the deferred set to keep track of pending reads to shared blocks.
- * We do this to ensure the new mapping caused by a write isn't performed
- * until these prior reads have completed. Otherwise the insertion of the
- * new mapping could free the old block that the read bios are mapped to.
- */
- struct dm_deferred_set;
- struct dm_deferred_entry;
- struct dm_deferred_set *dm_deferred_set_create(void);
- void dm_deferred_set_destroy(struct dm_deferred_set *ds);
- struct dm_deferred_entry *dm_deferred_entry_inc(struct dm_deferred_set *ds);
- void dm_deferred_entry_dec(struct dm_deferred_entry *entry, struct list_head *head);
- int dm_deferred_set_add_work(struct dm_deferred_set *ds, struct list_head *work);
- /*----------------------------------------------------------------*/
- #endif
|