| #undef TRACE_SYSTEM |
| #define TRACE_SYSTEM bcache |
| |
| #if !defined(_TRACE_BCACHE_H) || defined(TRACE_HEADER_MULTI_READ) |
| #define _TRACE_BCACHE_H |
| |
| #include <linux/tracepoint.h> |
| |
| struct search; |
| |
| DECLARE_EVENT_CLASS(bcache_request, |
| TP_PROTO(struct search *s, struct bio *bio), |
| TP_ARGS(s, bio), |
| |
| TP_STRUCT__entry( |
| __field(dev_t, dev ) |
| __field(unsigned int, orig_major ) |
| __field(unsigned int, orig_minor ) |
| __field(sector_t, sector ) |
| __field(dev_t, orig_sector ) |
| __field(unsigned int, nr_sector ) |
| __array(char, rwbs, 6 ) |
| ), |
| |
| TP_fast_assign( |
| __entry->dev = bio->bi_bdev->bd_dev; |
| __entry->orig_major = s->d->disk->major; |
| __entry->orig_minor = s->d->disk->first_minor; |
| __entry->sector = bio->bi_sector; |
| __entry->orig_sector = bio->bi_sector - 16; |
| __entry->nr_sector = bio->bi_size >> 9; |
| blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_size); |
| ), |
| |
| TP_printk("%d,%d %s %llu + %u (from %d,%d @ %llu)", |
| MAJOR(__entry->dev), MINOR(__entry->dev), |
| __entry->rwbs, (unsigned long long)__entry->sector, |
| __entry->nr_sector, __entry->orig_major, __entry->orig_minor, |
| (unsigned long long)__entry->orig_sector) |
| ); |
| |
| DECLARE_EVENT_CLASS(bkey, |
| TP_PROTO(struct bkey *k), |
| TP_ARGS(k), |
| |
| TP_STRUCT__entry( |
| __field(u32, size ) |
| __field(u32, inode ) |
| __field(u64, offset ) |
| __field(bool, dirty ) |
| ), |
| |
| TP_fast_assign( |
| __entry->inode = KEY_INODE(k); |
| __entry->offset = KEY_OFFSET(k); |
| __entry->size = KEY_SIZE(k); |
| __entry->dirty = KEY_DIRTY(k); |
| ), |
| |
| TP_printk("%u:%llu len %u dirty %u", __entry->inode, |
| __entry->offset, __entry->size, __entry->dirty) |
| ); |
| |
| DECLARE_EVENT_CLASS(btree_node, |
| TP_PROTO(struct btree *b), |
| TP_ARGS(b), |
| |
| TP_STRUCT__entry( |
| __field(size_t, bucket ) |
| ), |
| |
| TP_fast_assign( |
| __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); |
| ), |
| |
| TP_printk("bucket %zu", __entry->bucket) |
| ); |
| |
| /* request.c */ |
| |
| DEFINE_EVENT(bcache_request, bcache_request_start, |
| TP_PROTO(struct search *s, struct bio *bio), |
| TP_ARGS(s, bio) |
| ); |
| |
| DEFINE_EVENT(bcache_request, bcache_request_end, |
| TP_PROTO(struct search *s, struct bio *bio), |
| TP_ARGS(s, bio) |
| ); |
| |
| DECLARE_EVENT_CLASS(bcache_bio, |
| TP_PROTO(struct bio *bio), |
| TP_ARGS(bio), |
| |
| TP_STRUCT__entry( |
| __field(dev_t, dev ) |
| __field(sector_t, sector ) |
| __field(unsigned int, nr_sector ) |
| __array(char, rwbs, 6 ) |
| ), |
| |
| TP_fast_assign( |
| __entry->dev = bio->bi_bdev->bd_dev; |
| __entry->sector = bio->bi_sector; |
| __entry->nr_sector = bio->bi_size >> 9; |
| blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_size); |
| ), |
| |
| TP_printk("%d,%d %s %llu + %u", |
| MAJOR(__entry->dev), MINOR(__entry->dev), __entry->rwbs, |
| (unsigned long long)__entry->sector, __entry->nr_sector) |
| ); |
| |
| DEFINE_EVENT(bcache_bio, bcache_bypass_sequential, |
| TP_PROTO(struct bio *bio), |
| TP_ARGS(bio) |
| ); |
| |
| DEFINE_EVENT(bcache_bio, bcache_bypass_congested, |
| TP_PROTO(struct bio *bio), |
| TP_ARGS(bio) |
| ); |
| |
| TRACE_EVENT(bcache_read, |
| TP_PROTO(struct bio *bio, bool hit, bool bypass), |
| TP_ARGS(bio, hit, bypass), |
| |
| TP_STRUCT__entry( |
| __field(dev_t, dev ) |
| __field(sector_t, sector ) |
| __field(unsigned int, nr_sector ) |
| __array(char, rwbs, 6 ) |
| __field(bool, cache_hit ) |
| __field(bool, bypass ) |
| ), |
| |
| TP_fast_assign( |
| __entry->dev = bio->bi_bdev->bd_dev; |
| __entry->sector = bio->bi_sector; |
| __entry->nr_sector = bio->bi_size >> 9; |
| blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_size); |
| __entry->cache_hit = hit; |
| __entry->bypass = bypass; |
| ), |
| |
| TP_printk("%d,%d %s %llu + %u hit %u bypass %u", |
| MAJOR(__entry->dev), MINOR(__entry->dev), |
| __entry->rwbs, (unsigned long long)__entry->sector, |
| __entry->nr_sector, __entry->cache_hit, __entry->bypass) |
| ); |
| |
| TRACE_EVENT(bcache_write, |
| TP_PROTO(struct bio *bio, bool writeback, bool bypass), |
| TP_ARGS(bio, writeback, bypass), |
| |
| TP_STRUCT__entry( |
| __field(dev_t, dev ) |
| __field(sector_t, sector ) |
| __field(unsigned int, nr_sector ) |
| __array(char, rwbs, 6 ) |
| __field(bool, writeback ) |
| __field(bool, bypass ) |
| ), |
| |
| TP_fast_assign( |
| __entry->dev = bio->bi_bdev->bd_dev; |
| __entry->sector = bio->bi_sector; |
| __entry->nr_sector = bio->bi_size >> 9; |
| blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_size); |
| __entry->writeback = writeback; |
| __entry->bypass = bypass; |
| ), |
| |
| TP_printk("%d,%d %s %llu + %u hit %u bypass %u", |
| MAJOR(__entry->dev), MINOR(__entry->dev), |
| __entry->rwbs, (unsigned long long)__entry->sector, |
| __entry->nr_sector, __entry->writeback, __entry->bypass) |
| ); |
| |
| DEFINE_EVENT(bcache_bio, bcache_read_retry, |
| TP_PROTO(struct bio *bio), |
| TP_ARGS(bio) |
| ); |
| |
| DEFINE_EVENT(bkey, bcache_cache_insert, |
| TP_PROTO(struct bkey *k), |
| TP_ARGS(k) |
| ); |
| |
| /* Journal */ |
| |
| DECLARE_EVENT_CLASS(cache_set, |
| TP_PROTO(struct cache_set *c), |
| TP_ARGS(c), |
| |
| TP_STRUCT__entry( |
| __array(char, uuid, 16 ) |
| ), |
| |
| TP_fast_assign( |
| memcpy(__entry->uuid, c->sb.set_uuid, 16); |
| ), |
| |
| TP_printk("%pU", __entry->uuid) |
| ); |
| |
| DEFINE_EVENT(bkey, bcache_journal_replay_key, |
| TP_PROTO(struct bkey *k), |
| TP_ARGS(k) |
| ); |
| |
| DEFINE_EVENT(cache_set, bcache_journal_full, |
| TP_PROTO(struct cache_set *c), |
| TP_ARGS(c) |
| ); |
| |
| DEFINE_EVENT(cache_set, bcache_journal_entry_full, |
| TP_PROTO(struct cache_set *c), |
| TP_ARGS(c) |
| ); |
| |
| DEFINE_EVENT(bcache_bio, bcache_journal_write, |
| TP_PROTO(struct bio *bio), |
| TP_ARGS(bio) |
| ); |
| |
| /* Btree */ |
| |
| DEFINE_EVENT(cache_set, bcache_btree_cache_cannibalize, |
| TP_PROTO(struct cache_set *c), |
| TP_ARGS(c) |
| ); |
| |
| DEFINE_EVENT(btree_node, bcache_btree_read, |
| TP_PROTO(struct btree *b), |
| TP_ARGS(b) |
| ); |
| |
| TRACE_EVENT(bcache_btree_write, |
| TP_PROTO(struct btree *b), |
| TP_ARGS(b), |
| |
| TP_STRUCT__entry( |
| __field(size_t, bucket ) |
| __field(unsigned, block ) |
| __field(unsigned, keys ) |
| ), |
| |
| TP_fast_assign( |
| __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); |
| __entry->block = b->written; |
| __entry->keys = b->sets[b->nsets].data->keys; |
| ), |
| |
| TP_printk("bucket %zu", __entry->bucket) |
| ); |
| |
| DEFINE_EVENT(btree_node, bcache_btree_node_alloc, |
| TP_PROTO(struct btree *b), |
| TP_ARGS(b) |
| ); |
| |
| DEFINE_EVENT(btree_node, bcache_btree_node_alloc_fail, |
| TP_PROTO(struct btree *b), |
| TP_ARGS(b) |
| ); |
| |
| DEFINE_EVENT(btree_node, bcache_btree_node_free, |
| TP_PROTO(struct btree *b), |
| TP_ARGS(b) |
| ); |
| |
| TRACE_EVENT(bcache_btree_gc_coalesce, |
| TP_PROTO(unsigned nodes), |
| TP_ARGS(nodes), |
| |
| TP_STRUCT__entry( |
| __field(unsigned, nodes ) |
| ), |
| |
| TP_fast_assign( |
| __entry->nodes = nodes; |
| ), |
| |
| TP_printk("coalesced %u nodes", __entry->nodes) |
| ); |
| |
| DEFINE_EVENT(cache_set, bcache_gc_start, |
| TP_PROTO(struct cache_set *c), |
| TP_ARGS(c) |
| ); |
| |
| DEFINE_EVENT(cache_set, bcache_gc_end, |
| TP_PROTO(struct cache_set *c), |
| TP_ARGS(c) |
| ); |
| |
| DEFINE_EVENT(bkey, bcache_gc_copy, |
| TP_PROTO(struct bkey *k), |
| TP_ARGS(k) |
| ); |
| |
| DEFINE_EVENT(bkey, bcache_gc_copy_collision, |
| TP_PROTO(struct bkey *k), |
| TP_ARGS(k) |
| ); |
| |
| TRACE_EVENT(bcache_btree_insert_key, |
| TP_PROTO(struct btree *b, struct bkey *k, unsigned op, unsigned status), |
| TP_ARGS(b, k, op, status), |
| |
| TP_STRUCT__entry( |
| __field(u64, btree_node ) |
| __field(u32, btree_level ) |
| __field(u32, inode ) |
| __field(u64, offset ) |
| __field(u32, size ) |
| __field(u8, dirty ) |
| __field(u8, op ) |
| __field(u8, status ) |
| ), |
| |
| TP_fast_assign( |
| __entry->btree_node = PTR_BUCKET_NR(b->c, &b->key, 0); |
| __entry->btree_level = b->level; |
| __entry->inode = KEY_INODE(k); |
| __entry->offset = KEY_OFFSET(k); |
| __entry->size = KEY_SIZE(k); |
| __entry->dirty = KEY_DIRTY(k); |
| __entry->op = op; |
| __entry->status = status; |
| ), |
| |
| TP_printk("%u for %u at %llu(%u): %u:%llu len %u dirty %u", |
| __entry->status, __entry->op, |
| __entry->btree_node, __entry->btree_level, |
| __entry->inode, __entry->offset, |
| __entry->size, __entry->dirty) |
| ); |
| |
| DECLARE_EVENT_CLASS(btree_split, |
| TP_PROTO(struct btree *b, unsigned keys), |
| TP_ARGS(b, keys), |
| |
| TP_STRUCT__entry( |
| __field(size_t, bucket ) |
| __field(unsigned, keys ) |
| ), |
| |
| TP_fast_assign( |
| __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); |
| __entry->keys = keys; |
| ), |
| |
| TP_printk("bucket %zu keys %u", __entry->bucket, __entry->keys) |
| ); |
| |
| DEFINE_EVENT(btree_split, bcache_btree_node_split, |
| TP_PROTO(struct btree *b, unsigned keys), |
| TP_ARGS(b, keys) |
| ); |
| |
| DEFINE_EVENT(btree_split, bcache_btree_node_compact, |
| TP_PROTO(struct btree *b, unsigned keys), |
| TP_ARGS(b, keys) |
| ); |
| |
| DEFINE_EVENT(btree_node, bcache_btree_set_root, |
| TP_PROTO(struct btree *b), |
| TP_ARGS(b) |
| ); |
| |
| /* Allocator */ |
| |
| TRACE_EVENT(bcache_alloc_invalidate, |
| TP_PROTO(struct cache *ca), |
| TP_ARGS(ca), |
| |
| TP_STRUCT__entry( |
| __field(unsigned, free ) |
| __field(unsigned, free_inc ) |
| __field(unsigned, free_inc_size ) |
| __field(unsigned, unused ) |
| ), |
| |
| TP_fast_assign( |
| __entry->free = fifo_used(&ca->free); |
| __entry->free_inc = fifo_used(&ca->free_inc); |
| __entry->free_inc_size = ca->free_inc.size; |
| __entry->unused = fifo_used(&ca->unused); |
| ), |
| |
| TP_printk("free %u free_inc %u/%u unused %u", __entry->free, |
| __entry->free_inc, __entry->free_inc_size, __entry->unused) |
| ); |
| |
| TRACE_EVENT(bcache_alloc_fail, |
| TP_PROTO(struct cache *ca), |
| TP_ARGS(ca), |
| |
| TP_STRUCT__entry( |
| __field(unsigned, free ) |
| __field(unsigned, free_inc ) |
| __field(unsigned, unused ) |
| __field(unsigned, blocked ) |
| ), |
| |
| TP_fast_assign( |
| __entry->free = fifo_used(&ca->free); |
| __entry->free_inc = fifo_used(&ca->free_inc); |
| __entry->unused = fifo_used(&ca->unused); |
| __entry->blocked = atomic_read(&ca->set->prio_blocked); |
| ), |
| |
| TP_printk("free %u free_inc %u unused %u blocked %u", __entry->free, |
| __entry->free_inc, __entry->unused, __entry->blocked) |
| ); |
| |
| /* Background writeback */ |
| |
| DEFINE_EVENT(bkey, bcache_writeback, |
| TP_PROTO(struct bkey *k), |
| TP_ARGS(k) |
| ); |
| |
| DEFINE_EVENT(bkey, bcache_writeback_collision, |
| TP_PROTO(struct bkey *k), |
| TP_ARGS(k) |
| ); |
| |
| #endif /* _TRACE_BCACHE_H */ |
| |
| /* This part must be outside protection */ |
| #include <trace/define_trace.h> |