int truncate_bdev_range(struct block_device *bdev, fmode_t mode,
loff_t lstart, loff_t lend)
{
- struct block_device *claimed_bdev = NULL;
- int err;
-
/*
* If we don't hold exclusive handle for the device, upgrade to it
* while we discard the buffer cache to avoid discarding buffers
* under live filesystem.
*/
if (!(mode & FMODE_EXCL)) {
- claimed_bdev = bdev->bd_contains;
- err = bd_prepare_to_claim(bdev, claimed_bdev,
- truncate_bdev_range);
+ int err = bd_prepare_to_claim(bdev, truncate_bdev_range);
if (err)
return err;
}
+
truncate_inode_pages_range(bdev->bd_inode->i_mapping, lstart, lend);
- if (claimed_bdev)
- bd_abort_claiming(bdev, claimed_bdev, truncate_bdev_range);
+ if (!(mode & FMODE_EXCL))
+ bd_abort_claiming(bdev, truncate_bdev_range);
return 0;
}
EXPORT_SYMBOL(truncate_bdev_range);
spin_lock_init(&bdev->bd_size_lock);
bdev->bd_disk = disk;
bdev->bd_partno = partno;
- bdev->bd_contains = NULL;
bdev->bd_super = NULL;
bdev->bd_inode = inode;
bdev->bd_part_count = 0;
/**
* bd_prepare_to_claim - claim a block device
* @bdev: block device of interest
- * @whole: the whole device containing @bdev, may equal @bdev
* @holder: holder trying to claim @bdev
*
* Claim @bdev. This function fails if @bdev is already claimed by another
* RETURNS:
* 0 if @bdev can be claimed, -EBUSY otherwise.
*/
-int bd_prepare_to_claim(struct block_device *bdev, struct block_device *whole,
- void *holder)
+int bd_prepare_to_claim(struct block_device *bdev, void *holder)
{
+ struct block_device *whole = bdev_whole(bdev);
+
+ if (WARN_ON_ONCE(!holder))
+ return -EINVAL;
retry:
spin_lock(&bdev_lock);
/* if someone else claimed, fail */
/**
* bd_finish_claiming - finish claiming of a block device
* @bdev: block device of interest
- * @whole: whole block device
* @holder: holder that has claimed @bdev
*
* Finish exclusive open of a block device. Mark the device as exlusively
* open by the holder and wake up all waiters for exclusive open to finish.
*/
-static void bd_finish_claiming(struct block_device *bdev,
- struct block_device *whole, void *holder)
+static void bd_finish_claiming(struct block_device *bdev, void *holder)
{
+ struct block_device *whole = bdev_whole(bdev);
+
spin_lock(&bdev_lock);
BUG_ON(!bd_may_claim(bdev, whole, holder));
/*
* also used when exclusive open is not actually desired and we just needed
* to block other exclusive openers for a while.
*/
-void bd_abort_claiming(struct block_device *bdev, struct block_device *whole,
- void *holder)
+void bd_abort_claiming(struct block_device *bdev, void *holder)
{
spin_lock(&bdev_lock);
- bd_clear_claiming(whole, holder);
+ bd_clear_claiming(bdev_whole(bdev), holder);
spin_unlock(&bdev_lock);
}
EXPORT_SYMBOL(bd_abort_claiming);
int ret;
if (!bdev->bd_openers) {
- bdev->bd_contains = bdev;
-
- if (!bdev->bd_partno) {
+ if (!bdev_is_partition(bdev)) {
ret = -ENXIO;
bdev->bd_part = disk_get_part(disk, 0);
if (!bdev->bd_part)
whole->bd_part_count++;
mutex_unlock(&whole->bd_mutex);
- bdev->bd_contains = whole;
bdev->bd_part = disk_get_part(disk, bdev->bd_partno);
if (!(disk->flags & GENHD_FL_UP) ||
!bdev->bd_part || !bdev->bd_part->nr_sects) {
if (bdev->bd_bdi == &noop_backing_dev_info)
bdev->bd_bdi = bdi_get(disk->queue->backing_dev_info);
} else {
- if (bdev->bd_contains == bdev) {
+ if (!bdev_is_partition(bdev)) {
ret = 0;
if (bdev->bd_disk->fops->open)
ret = bdev->bd_disk->fops->open(bdev, mode);
out_clear:
disk_put_part(bdev->bd_part);
bdev->bd_part = NULL;
- bdev->bd_contains = NULL;
return ret;
}
*/
struct block_device *blkdev_get_by_dev(dev_t dev, fmode_t mode, void *holder)
{
- struct block_device *claiming;
bool unblock_events = true;
struct block_device *bdev;
struct gendisk *disk;
disk = bdev->bd_disk;
if (mode & FMODE_EXCL) {
- WARN_ON_ONCE(!holder);
-
- ret = -ENOMEM;
- claiming = bdget_disk(disk, 0);
- if (!claiming)
- goto put_blkdev;
- ret = bd_prepare_to_claim(bdev, claiming, holder);
+ ret = bd_prepare_to_claim(bdev, holder);
if (ret)
- goto put_claiming;
+ goto put_blkdev;
}
disk_block_events(disk);
if (ret)
goto abort_claiming;
if (mode & FMODE_EXCL) {
- bd_finish_claiming(bdev, claiming, holder);
+ bd_finish_claiming(bdev, holder);
/*
* Block event polling for write claims if requested. Any write
if (unblock_events)
disk_unblock_events(disk);
- if (mode & FMODE_EXCL)
- bdput(claiming);
return bdev;
abort_claiming:
if (mode & FMODE_EXCL)
- bd_abort_claiming(bdev, claiming, holder);
+ bd_abort_claiming(bdev, holder);
mutex_unlock(&bdev->bd_mutex);
disk_unblock_events(disk);
-put_claiming:
- if (mode & FMODE_EXCL)
- bdput(claiming);
put_blkdev:
blkdev_put_no_open(bdev);
if (ret == -ERESTARTSYS)
disk_put_part(bdev->bd_part);
bdev->bd_part = NULL;
if (bdev_is_partition(bdev))
- victim = bdev->bd_contains;
- bdev->bd_contains = NULL;
+ victim = bdev_whole(bdev);
} else {
if (!bdev_is_partition(bdev) && disk->fops->release)
disk->fops->release(disk, mode);
mutex_lock(&bdev->bd_mutex);
if (mode & FMODE_EXCL) {
+ struct block_device *whole = bdev_whole(bdev);
bool bdev_free;
/*
spin_lock(&bdev_lock);
WARN_ON_ONCE(--bdev->bd_holders < 0);
- WARN_ON_ONCE(--bdev->bd_contains->bd_holders < 0);
+ WARN_ON_ONCE(--whole->bd_holders < 0);
- /* bd_contains might point to self, check in a separate step */
if ((bdev_free = !bdev->bd_holders))
bdev->bd_holder = NULL;
- if (!bdev->bd_contains->bd_holders)
- bdev->bd_contains->bd_holder = NULL;
+ if (!whole->bd_holders)
+ whole->bd_holder = NULL;
spin_unlock(&bdev_lock);