#ifndef _BCACHE_REQUEST_H_ #define _BCACHE_REQUEST_H_ #include struct search { /* Stack frame for bio_complete */ struct closure cl; struct bcache_device *d; struct task_struct *task; struct bbio bio; struct bio *orig_bio; struct bio *cache_miss; unsigned cache_bio_sectors; unsigned recoverable:1; unsigned unaligned_bvec:1; unsigned write:1; unsigned writeback:1; /* IO error returned to s->bio */ short error; unsigned long start_time; /* Anything past op->keys won't get zeroed in do_bio_hook */ struct btree_op op; }; void bch_cache_read_endio(struct bio *, int); int bch_get_congested(struct cache_set *); void bch_insert_data(struct closure *cl); void bch_btree_insert_async(struct closure *); void bch_cache_read_endio(struct bio *, int); void bch_open_buckets_free(struct cache_set *); int bch_open_buckets_alloc(struct cache_set *); void bch_cached_dev_request_init(struct cached_dev *dc); void bch_flash_dev_request_init(struct bcache_device *d); extern struct kmem_cache *bch_search_cache, *bch_passthrough_cache; struct bch_cgroup { #ifdef CONFIG_CGROUP_BCACHE struct cgroup_subsys_state css; #endif /* * We subtract one from the index into bch_cache_modes[], so that * default == -1; this makes it so the rest match up with d->cache_mode, * and we use d->cache_mode if cgrp->cache_mode < 0 */ short cache_mode; bool verify; struct cache_stat_collector stats; }; struct bch_cgroup *bch_bio_to_cgroup(struct bio *bio); #endif /* _BCACHE_REQUEST_H_ */