io_uring: end waiting before task cancel attempts
[linux-2.6-microblaze.git] / fs / block_dev.c
index b350ed3..9e56ee1 100644 (file)
@@ -32,6 +32,7 @@
 #include <linux/cleancache.h>
 #include <linux/task_io_accounting_ops.h>
 #include <linux/falloc.h>
+#include <linux/part_stat.h>
 #include <linux/uaccess.h>
 #include <linux/suspend.h>
 #include "internal.h"
@@ -110,24 +111,20 @@ EXPORT_SYMBOL(invalidate_bdev);
 int truncate_bdev_range(struct block_device *bdev, fmode_t mode,
                        loff_t lstart, loff_t lend)
 {
-       struct block_device *claimed_bdev = NULL;
-       int err;
-
        /*
         * If we don't hold exclusive handle for the device, upgrade to it
         * while we discard the buffer cache to avoid discarding buffers
         * under live filesystem.
         */
        if (!(mode & FMODE_EXCL)) {
-               claimed_bdev = bdev->bd_contains;
-               err = bd_prepare_to_claim(bdev, claimed_bdev,
-                                         truncate_bdev_range);
+               int err = bd_prepare_to_claim(bdev, truncate_bdev_range);
                if (err)
                        return err;
        }
+
        truncate_inode_pages_range(bdev->bd_inode->i_mapping, lstart, lend);
-       if (claimed_bdev)
-               bd_abort_claiming(bdev, claimed_bdev, truncate_bdev_range);
+       if (!(mode & FMODE_EXCL))
+               bd_abort_claiming(bdev, truncate_bdev_range);
        return 0;
 }
 EXPORT_SYMBOL(truncate_bdev_range);
@@ -785,23 +782,19 @@ static struct inode *bdev_alloc_inode(struct super_block *sb)
 
 static void bdev_free_inode(struct inode *inode)
 {
+       struct block_device *bdev = I_BDEV(inode);
+
+       free_percpu(bdev->bd_stats);
+       kfree(bdev->bd_meta_info);
+
        kmem_cache_free(bdev_cachep, BDEV_I(inode));
 }
 
-static void init_once(void *foo)
+static void init_once(void *data)
 {
-       struct bdev_inode *ei = (struct bdev_inode *) foo;
-       struct block_device *bdev = &ei->bdev;
+       struct bdev_inode *ei = data;
 
-       memset(bdev, 0, sizeof(*bdev));
-       mutex_init(&bdev->bd_mutex);
-#ifdef CONFIG_SYSFS
-       INIT_LIST_HEAD(&bdev->bd_holder_disks);
-#endif
-       bdev->bd_bdi = &noop_backing_dev_info;
        inode_init_once(&ei->vfs_inode);
-       /* Initialize mutex for freeze. */
-       mutex_init(&bdev->bd_fsfreeze_mutex);
 }
 
 static void bdev_evict_inode(struct inode *inode)
@@ -877,13 +870,22 @@ struct block_device *bdev_alloc(struct gendisk *disk, u8 partno)
        mapping_set_gfp_mask(&inode->i_data, GFP_USER);
 
        bdev = I_BDEV(inode);
+       memset(bdev, 0, sizeof(*bdev));
+       mutex_init(&bdev->bd_mutex);
+       mutex_init(&bdev->bd_fsfreeze_mutex);
        spin_lock_init(&bdev->bd_size_lock);
        bdev->bd_disk = disk;
        bdev->bd_partno = partno;
-       bdev->bd_contains = NULL;
-       bdev->bd_super = NULL;
        bdev->bd_inode = inode;
-       bdev->bd_part_count = 0;
+       bdev->bd_bdi = &noop_backing_dev_info;
+#ifdef CONFIG_SYSFS
+       INIT_LIST_HEAD(&bdev->bd_holder_disks);
+#endif
+       bdev->bd_stats = alloc_percpu(struct disk_stats);
+       if (!bdev->bd_stats) {
+               iput(inode);
+               return NULL;
+       }
        return bdev;
 }
 
@@ -920,11 +922,6 @@ struct block_device *bdgrab(struct block_device *bdev)
 }
 EXPORT_SYMBOL(bdgrab);
 
-struct block_device *bdget_part(struct hd_struct *part)
-{
-       return bdget(part_devt(part));
-}
-
 long nr_blockdev_pages(void)
 {
        struct inode *inode;
@@ -979,7 +976,6 @@ static bool bd_may_claim(struct block_device *bdev, struct block_device *whole,
 /**
  * bd_prepare_to_claim - claim a block device
  * @bdev: block device of interest
- * @whole: the whole device containing @bdev, may equal @bdev
  * @holder: holder trying to claim @bdev
  *
  * Claim @bdev.  This function fails if @bdev is already claimed by another
@@ -989,9 +985,12 @@ static bool bd_may_claim(struct block_device *bdev, struct block_device *whole,
  * RETURNS:
  * 0 if @bdev can be claimed, -EBUSY otherwise.
  */
-int bd_prepare_to_claim(struct block_device *bdev, struct block_device *whole,
-               void *holder)
+int bd_prepare_to_claim(struct block_device *bdev, void *holder)
 {
+       struct block_device *whole = bdev_whole(bdev);
+
+       if (WARN_ON_ONCE(!holder))
+               return -EINVAL;
 retry:
        spin_lock(&bdev_lock);
        /* if someone else claimed, fail */
@@ -1031,15 +1030,15 @@ static void bd_clear_claiming(struct block_device *whole, void *holder)
 /**
  * bd_finish_claiming - finish claiming of a block device
  * @bdev: block device of interest
- * @whole: whole block device
  * @holder: holder that has claimed @bdev
  *
  * Finish exclusive open of a block device. Mark the device as exlusively
  * open by the holder and wake up all waiters for exclusive open to finish.
  */
-static void bd_finish_claiming(struct block_device *bdev,
-               struct block_device *whole, void *holder)
+static void bd_finish_claiming(struct block_device *bdev, void *holder)
 {
+       struct block_device *whole = bdev_whole(bdev);
+
        spin_lock(&bdev_lock);
        BUG_ON(!bd_may_claim(bdev, whole, holder));
        /*
@@ -1064,11 +1063,10 @@ static void bd_finish_claiming(struct block_device *bdev,
  * also used when exclusive open is not actually desired and we just needed
  * to block other exclusive openers for a while.
  */
-void bd_abort_claiming(struct block_device *bdev, struct block_device *whole,
-                      void *holder)
+void bd_abort_claiming(struct block_device *bdev, void *holder)
 {
        spin_lock(&bdev_lock);
-       bd_clear_claiming(whole, holder);
+       bd_clear_claiming(bdev_whole(bdev), holder);
        spin_unlock(&bdev_lock);
 }
 EXPORT_SYMBOL(bd_abort_claiming);
@@ -1139,7 +1137,7 @@ int bd_link_disk_holder(struct block_device *bdev, struct gendisk *disk)
        WARN_ON_ONCE(!bdev->bd_holder);
 
        /* FIXME: remove the following once add_disk() handles errors */
-       if (WARN_ON(!disk->slave_dir || !bdev->bd_part->holder_dir))
+       if (WARN_ON(!disk->slave_dir || !bdev->bd_holder_dir))
                goto out_unlock;
 
        holder = bd_find_holder_disk(bdev, disk);
@@ -1162,14 +1160,14 @@ int bd_link_disk_holder(struct block_device *bdev, struct gendisk *disk)
        if (ret)
                goto out_free;
 
-       ret = add_symlink(bdev->bd_part->holder_dir, &disk_to_dev(disk)->kobj);
+       ret = add_symlink(bdev->bd_holder_dir, &disk_to_dev(disk)->kobj);
        if (ret)
                goto out_del;
        /*
         * bdev could be deleted beneath us which would implicitly destroy
         * the holder directory.  Hold on to it.
         */
-       kobject_get(bdev->bd_part->holder_dir);
+       kobject_get(bdev->bd_holder_dir);
 
        list_add(&holder->list, &bdev->bd_holder_disks);
        goto out_unlock;
@@ -1204,9 +1202,8 @@ void bd_unlink_disk_holder(struct block_device *bdev, struct gendisk *disk)
 
        if (!WARN_ON_ONCE(holder == NULL) && !--holder->refcnt) {
                del_symlink(disk->slave_dir, bdev_kobj(bdev));
-               del_symlink(bdev->bd_part->holder_dir,
-                           &disk_to_dev(disk)->kobj);
-               kobject_put(bdev->bd_part->holder_dir);
+               del_symlink(bdev->bd_holder_dir, &disk_to_dev(disk)->kobj);
+               kobject_put(bdev->bd_holder_dir);
                list_del_init(&holder->list);
                kfree(holder);
        }
@@ -1216,70 +1213,6 @@ void bd_unlink_disk_holder(struct block_device *bdev, struct gendisk *disk)
 EXPORT_SYMBOL_GPL(bd_unlink_disk_holder);
 #endif
 
-/**
- * check_disk_size_change - checks for disk size change and adjusts bdev size.
- * @disk: struct gendisk to check
- * @bdev: struct bdev to adjust.
- * @verbose: if %true log a message about a size change if there is any
- *
- * This routine checks to see if the bdev size does not match the disk size
- * and adjusts it if it differs. When shrinking the bdev size, its all caches
- * are freed.
- */
-static void check_disk_size_change(struct gendisk *disk,
-               struct block_device *bdev, bool verbose)
-{
-       loff_t disk_size, bdev_size;
-
-       spin_lock(&bdev->bd_size_lock);
-       disk_size = (loff_t)get_capacity(disk) << 9;
-       bdev_size = i_size_read(bdev->bd_inode);
-       if (disk_size != bdev_size) {
-               if (verbose) {
-                       printk(KERN_INFO
-                              "%s: detected capacity change from %lld to %lld\n",
-                              disk->disk_name, bdev_size, disk_size);
-               }
-               i_size_write(bdev->bd_inode, disk_size);
-       }
-       spin_unlock(&bdev->bd_size_lock);
-}
-
-/**
- * revalidate_disk_size - checks for disk size change and adjusts bdev size.
- * @disk: struct gendisk to check
- * @verbose: if %true log a message about a size change if there is any
- *
- * This routine checks to see if the bdev size does not match the disk size
- * and adjusts it if it differs. When shrinking the bdev size, its all caches
- * are freed.
- */
-void revalidate_disk_size(struct gendisk *disk, bool verbose)
-{
-       struct block_device *bdev;
-
-       /*
-        * Hidden disks don't have associated bdev so there's no point in
-        * revalidating them.
-        */
-       if (disk->flags & GENHD_FL_HIDDEN)
-               return;
-
-       bdev = bdget_disk(disk, 0);
-       if (bdev) {
-               check_disk_size_change(disk, bdev, verbose);
-               bdput(bdev);
-       }
-}
-
-void bd_set_nr_sectors(struct block_device *bdev, sector_t sectors)
-{
-       spin_lock(&bdev->bd_size_lock);
-       i_size_write(bdev->bd_inode, (loff_t)sectors << SECTOR_SHIFT);
-       spin_unlock(&bdev->bd_size_lock);
-}
-EXPORT_SYMBOL(bd_set_nr_sectors);
-
 static void __blkdev_put(struct block_device *bdev, fmode_t mode, int for_part);
 
 int bdev_disk_changed(struct block_device *bdev, bool invalidate)
@@ -1313,8 +1246,6 @@ rescan:
                        disk->fops->revalidate_disk(disk);
        }
 
-       check_disk_size_change(disk, bdev, !invalidate);
-
        if (get_capacity(disk)) {
                ret = blk_add_partitions(disk, bdev);
                if (ret == -EAGAIN)
@@ -1344,25 +1275,16 @@ EXPORT_SYMBOL_GPL(bdev_disk_changed);
 static int __blkdev_get(struct block_device *bdev, fmode_t mode)
 {
        struct gendisk *disk = bdev->bd_disk;
-       int ret;
+       int ret = 0;
 
        if (!bdev->bd_openers) {
-               bdev->bd_contains = bdev;
-
-               if (!bdev->bd_partno) {
-                       ret = -ENXIO;
-                       bdev->bd_part = disk_get_part(disk, 0);
-                       if (!bdev->bd_part)
-                               goto out_clear;
-
+               if (!bdev_is_partition(bdev)) {
                        ret = 0;
                        if (disk->fops->open)
                                ret = disk->fops->open(bdev, mode);
 
-                       if (!ret) {
-                               bd_set_nr_sectors(bdev, get_capacity(disk));
+                       if (!ret)
                                set_init_blocksize(bdev);
-                       }
 
                        /*
                         * If the device is invalidated, rescan partition
@@ -1375,38 +1297,33 @@ static int __blkdev_get(struct block_device *bdev, fmode_t mode)
                                bdev_disk_changed(bdev, ret == -ENOMEDIUM);
 
                        if (ret)
-                               goto out_clear;
+                               return ret;
                } else {
-                       struct block_device *whole = bdget_disk(disk, 0);
+                       struct block_device *whole = bdgrab(disk->part0);
 
                        mutex_lock_nested(&whole->bd_mutex, 1);
                        ret = __blkdev_get(whole, mode);
                        if (ret) {
                                mutex_unlock(&whole->bd_mutex);
                                bdput(whole);
-                               goto out_clear;
+                               return ret;
                        }
                        whole->bd_part_count++;
                        mutex_unlock(&whole->bd_mutex);
 
-                       bdev->bd_contains = whole;
-                       bdev->bd_part = disk_get_part(disk, bdev->bd_partno);
                        if (!(disk->flags & GENHD_FL_UP) ||
-                           !bdev->bd_part || !bdev->bd_part->nr_sects) {
+                           !bdev_nr_sectors(bdev)) {
                                __blkdev_put(whole, mode, 1);
                                bdput(whole);
-                               ret = -ENXIO;
-                               goto out_clear;
+                               return -ENXIO;
                        }
-                       bd_set_nr_sectors(bdev, bdev->bd_part->nr_sects);
                        set_init_blocksize(bdev);
                }
 
                if (bdev->bd_bdi == &noop_backing_dev_info)
                        bdev->bd_bdi = bdi_get(disk->queue->backing_dev_info);
        } else {
-               if (bdev->bd_contains == bdev) {
-                       ret = 0;
+               if (!bdev_is_partition(bdev)) {
                        if (bdev->bd_disk->fops->open)
                                ret = bdev->bd_disk->fops->open(bdev, mode);
                        /* the same as first opener case, read comment there */
@@ -1419,12 +1336,6 @@ static int __blkdev_get(struct block_device *bdev, fmode_t mode)
        }
        bdev->bd_openers++;
        return 0;
-
- out_clear:
-       disk_put_part(bdev->bd_part);
-       bdev->bd_part = NULL;
-       bdev->bd_contains = NULL;
-       return ret;
 }
 
 struct block_device *blkdev_get_no_open(dev_t dev)
@@ -1492,7 +1403,6 @@ void blkdev_put_no_open(struct block_device *bdev)
  */
 struct block_device *blkdev_get_by_dev(dev_t dev, fmode_t mode, void *holder)
 {
-       struct block_device *claiming;
        bool unblock_events = true;
        struct block_device *bdev;
        struct gendisk *disk;
@@ -1515,15 +1425,9 @@ retry:
        disk = bdev->bd_disk;
 
        if (mode & FMODE_EXCL) {
-               WARN_ON_ONCE(!holder);
-       
-               ret = -ENOMEM;
-               claiming = bdget_disk(disk, 0);
-               if (!claiming)
-                       goto put_blkdev;
-               ret = bd_prepare_to_claim(bdev, claiming, holder);
+               ret = bd_prepare_to_claim(bdev, holder);
                if (ret)
-                       goto put_claiming;
+                       goto put_blkdev;
        }
 
        disk_block_events(disk);
@@ -1533,7 +1437,7 @@ retry:
        if (ret)
                goto abort_claiming;
        if (mode & FMODE_EXCL) {
-               bd_finish_claiming(bdev, claiming, holder);
+               bd_finish_claiming(bdev, holder);
 
                /*
                 * Block event polling for write claims if requested.  Any write
@@ -1552,18 +1456,13 @@ retry:
 
        if (unblock_events)
                disk_unblock_events(disk);
-       if (mode & FMODE_EXCL)
-               bdput(claiming);
        return bdev;
 
 abort_claiming:
        if (mode & FMODE_EXCL)
-               bd_abort_claiming(bdev, claiming, holder);
+               bd_abort_claiming(bdev, holder);
        mutex_unlock(&bdev->bd_mutex);
        disk_unblock_events(disk);
-put_claiming:
-       if (mode & FMODE_EXCL)
-               bdput(claiming);
 put_blkdev:
        blkdev_put_no_open(bdev);
        if (ret == -ERESTARTSYS)
@@ -1663,19 +1562,12 @@ static void __blkdev_put(struct block_device *bdev, fmode_t mode, int for_part)
                sync_blockdev(bdev);
                kill_bdev(bdev);
                bdev_write_inode(bdev);
-
-               if (!bdev_is_partition(bdev) && disk->fops->release)
-                       disk->fops->release(disk, mode);
-
-               disk_put_part(bdev->bd_part);
-               bdev->bd_part = NULL;
                if (bdev_is_partition(bdev))
-                       victim = bdev->bd_contains;
-               bdev->bd_contains = NULL;
-       } else {
-               if (!bdev_is_partition(bdev) && disk->fops->release)
-                       disk->fops->release(disk, mode);
+                       victim = bdev_whole(bdev);
        }
+
+       if (!bdev_is_partition(bdev) && disk->fops->release)
+               disk->fops->release(disk, mode);
        mutex_unlock(&bdev->bd_mutex);
        if (victim) {
                __blkdev_put(victim, mode, 1);
@@ -1690,6 +1582,7 @@ void blkdev_put(struct block_device *bdev, fmode_t mode)
        mutex_lock(&bdev->bd_mutex);
 
        if (mode & FMODE_EXCL) {
+               struct block_device *whole = bdev_whole(bdev);
                bool bdev_free;
 
                /*
@@ -1700,13 +1593,12 @@ void blkdev_put(struct block_device *bdev, fmode_t mode)
                spin_lock(&bdev_lock);
 
                WARN_ON_ONCE(--bdev->bd_holders < 0);
-               WARN_ON_ONCE(--bdev->bd_contains->bd_holders < 0);
+               WARN_ON_ONCE(--whole->bd_holders < 0);
 
-               /* bd_contains might point to self, check in a separate step */
                if ((bdev_free = !bdev->bd_holders))
                        bdev->bd_holder = NULL;
-               if (!bdev->bd_contains->bd_holders)
-                       bdev->bd_contains->bd_holder = NULL;
+               if (!whole->bd_holders)
+                       whole->bd_holder = NULL;
 
                spin_unlock(&bdev_lock);