mirror of
https://github.com/torvalds/linux.git
synced 2026-04-18 06:44:00 -04:00
Merge tag 'bcachefs-2023-11-29' of https://evilpiepirate.org/git/bcachefs
Pull more bcachefs bugfixes from Kent Overstreet: - bcache & bcachefs were broken with CFI enabled; patch for closures to fix type punning - mark erasure coding as extra-experimental; there are incompatible disk space accounting changes coming for erasure coding, and I'm still seeing checksum errors in some tests - several fixes for durability-related issues (durability is a device specific setting where we can tell bcachefs that data on a given device should be counted as replicated x times) - a fix for a rare livelock when a btree node merge then updates a parent node that is almost full - fix a race in the device removal path, where dropping a pointer in a btree node to a device would be clobbered by an in flight btree write updating the btree node key on completion - fix one SRCU lock hold time warning in the btree gc code - ther's still a bunch more of these to fix - fix a rare race where we'd start copygc before initializing the "are we rw" percpu refcount; copygc would think we were already ro and die immediately * tag 'bcachefs-2023-11-29' of https://evilpiepirate.org/git/bcachefs: (23 commits) bcachefs: Extra kthread_should_stop() calls for copygc bcachefs: Convert gc_alloc_start() to for_each_btree_key2() bcachefs: Fix race between btree writes and metadata drop bcachefs: move journal seq assertion bcachefs: -EROFS doesn't count as move_extent_start_fail bcachefs: trace_move_extent_start_fail() now includes errcode bcachefs: Fix split_race livelock bcachefs: Fix bucket data type for stripe buckets bcachefs: Add missing validation for jset_entry_data_usage bcachefs: Fix zstd compress workspace size bcachefs: bpos is misaligned on big endian bcachefs: Fix ec + durability calculation bcachefs: Data update path won't accidentaly grow replicas bcachefs: deallocate_extra_replicas() bcachefs: Proper refcounting for journal_keys bcachefs: preserve device path as device name bcachefs: Fix an endianness conversion bcachefs: Start gc, copygc, rebalance threads after initing writes ref bcachefs: Don't stop copygc thread on device resize bcachefs: Make sure bch2_move_ratelimit() also waits for move_ops ...
This commit is contained in:
@@ -327,9 +327,9 @@ static void __write_super(struct cache_sb *sb, struct cache_sb_disk *out,
|
||||
submit_bio(bio);
|
||||
}
|
||||
|
||||
static void bch_write_bdev_super_unlock(struct closure *cl)
|
||||
static CLOSURE_CALLBACK(bch_write_bdev_super_unlock)
|
||||
{
|
||||
struct cached_dev *dc = container_of(cl, struct cached_dev, sb_write);
|
||||
closure_type(dc, struct cached_dev, sb_write);
|
||||
|
||||
up(&dc->sb_write_mutex);
|
||||
}
|
||||
@@ -363,9 +363,9 @@ static void write_super_endio(struct bio *bio)
|
||||
closure_put(&ca->set->sb_write);
|
||||
}
|
||||
|
||||
static void bcache_write_super_unlock(struct closure *cl)
|
||||
static CLOSURE_CALLBACK(bcache_write_super_unlock)
|
||||
{
|
||||
struct cache_set *c = container_of(cl, struct cache_set, sb_write);
|
||||
closure_type(c, struct cache_set, sb_write);
|
||||
|
||||
up(&c->sb_write_mutex);
|
||||
}
|
||||
@@ -407,9 +407,9 @@ static void uuid_endio(struct bio *bio)
|
||||
closure_put(cl);
|
||||
}
|
||||
|
||||
static void uuid_io_unlock(struct closure *cl)
|
||||
static CLOSURE_CALLBACK(uuid_io_unlock)
|
||||
{
|
||||
struct cache_set *c = container_of(cl, struct cache_set, uuid_write);
|
||||
closure_type(c, struct cache_set, uuid_write);
|
||||
|
||||
up(&c->uuid_write_mutex);
|
||||
}
|
||||
@@ -1344,9 +1344,9 @@ void bch_cached_dev_release(struct kobject *kobj)
|
||||
module_put(THIS_MODULE);
|
||||
}
|
||||
|
||||
static void cached_dev_free(struct closure *cl)
|
||||
static CLOSURE_CALLBACK(cached_dev_free)
|
||||
{
|
||||
struct cached_dev *dc = container_of(cl, struct cached_dev, disk.cl);
|
||||
closure_type(dc, struct cached_dev, disk.cl);
|
||||
|
||||
if (test_and_clear_bit(BCACHE_DEV_WB_RUNNING, &dc->disk.flags))
|
||||
cancel_writeback_rate_update_dwork(dc);
|
||||
@@ -1378,9 +1378,9 @@ static void cached_dev_free(struct closure *cl)
|
||||
kobject_put(&dc->disk.kobj);
|
||||
}
|
||||
|
||||
static void cached_dev_flush(struct closure *cl)
|
||||
static CLOSURE_CALLBACK(cached_dev_flush)
|
||||
{
|
||||
struct cached_dev *dc = container_of(cl, struct cached_dev, disk.cl);
|
||||
closure_type(dc, struct cached_dev, disk.cl);
|
||||
struct bcache_device *d = &dc->disk;
|
||||
|
||||
mutex_lock(&bch_register_lock);
|
||||
@@ -1499,9 +1499,9 @@ void bch_flash_dev_release(struct kobject *kobj)
|
||||
kfree(d);
|
||||
}
|
||||
|
||||
static void flash_dev_free(struct closure *cl)
|
||||
static CLOSURE_CALLBACK(flash_dev_free)
|
||||
{
|
||||
struct bcache_device *d = container_of(cl, struct bcache_device, cl);
|
||||
closure_type(d, struct bcache_device, cl);
|
||||
|
||||
mutex_lock(&bch_register_lock);
|
||||
atomic_long_sub(bcache_dev_sectors_dirty(d),
|
||||
@@ -1512,9 +1512,9 @@ static void flash_dev_free(struct closure *cl)
|
||||
kobject_put(&d->kobj);
|
||||
}
|
||||
|
||||
static void flash_dev_flush(struct closure *cl)
|
||||
static CLOSURE_CALLBACK(flash_dev_flush)
|
||||
{
|
||||
struct bcache_device *d = container_of(cl, struct bcache_device, cl);
|
||||
closure_type(d, struct bcache_device, cl);
|
||||
|
||||
mutex_lock(&bch_register_lock);
|
||||
bcache_device_unlink(d);
|
||||
@@ -1670,9 +1670,9 @@ void bch_cache_set_release(struct kobject *kobj)
|
||||
module_put(THIS_MODULE);
|
||||
}
|
||||
|
||||
static void cache_set_free(struct closure *cl)
|
||||
static CLOSURE_CALLBACK(cache_set_free)
|
||||
{
|
||||
struct cache_set *c = container_of(cl, struct cache_set, cl);
|
||||
closure_type(c, struct cache_set, cl);
|
||||
struct cache *ca;
|
||||
|
||||
debugfs_remove(c->debug);
|
||||
@@ -1711,9 +1711,9 @@ static void cache_set_free(struct closure *cl)
|
||||
kobject_put(&c->kobj);
|
||||
}
|
||||
|
||||
static void cache_set_flush(struct closure *cl)
|
||||
static CLOSURE_CALLBACK(cache_set_flush)
|
||||
{
|
||||
struct cache_set *c = container_of(cl, struct cache_set, caching);
|
||||
closure_type(c, struct cache_set, caching);
|
||||
struct cache *ca = c->cache;
|
||||
struct btree *b;
|
||||
|
||||
@@ -1808,9 +1808,9 @@ static void conditional_stop_bcache_device(struct cache_set *c,
|
||||
}
|
||||
}
|
||||
|
||||
static void __cache_set_unregister(struct closure *cl)
|
||||
static CLOSURE_CALLBACK(__cache_set_unregister)
|
||||
{
|
||||
struct cache_set *c = container_of(cl, struct cache_set, caching);
|
||||
closure_type(c, struct cache_set, caching);
|
||||
struct cached_dev *dc;
|
||||
struct bcache_device *d;
|
||||
size_t i;
|
||||
|
||||
Reference in New Issue
Block a user