Merge branches 'pm-core', 'pm-sleep' and 'powercap'
[linux-2.6-microblaze.git] / drivers / base / power / runtime.c
index d4059e6..676dc72 100644 (file)
@@ -263,7 +263,7 @@ static int rpm_check_suspend_allowed(struct device *dev)
                retval = -EINVAL;
        else if (dev->power.disable_depth > 0)
                retval = -EACCES;
-       else if (atomic_read(&dev->power.usage_count) > 0)
+       else if (atomic_read(&dev->power.usage_count))
                retval = -EAGAIN;
        else if (!dev->power.ignore_children &&
                        atomic_read(&dev->power.child_count))
@@ -1039,13 +1039,33 @@ int pm_schedule_suspend(struct device *dev, unsigned int delay)
 }
 EXPORT_SYMBOL_GPL(pm_schedule_suspend);
 
+static int rpm_drop_usage_count(struct device *dev)
+{
+       int ret;
+
+       ret = atomic_sub_return(1, &dev->power.usage_count);
+       if (ret >= 0)
+               return ret;
+
+       /*
+        * Because rpm_resume() does not check the usage counter, it will resume
+        * the device even if the usage counter is 0 or negative, so it is
+        * sufficient to increment the usage counter here to reverse the change
+        * made above.
+        */
+       atomic_inc(&dev->power.usage_count);
+       dev_warn(dev, "Runtime PM usage count underflow!\n");
+       return -EINVAL;
+}
+
 /**
  * __pm_runtime_idle - Entry point for runtime idle operations.
  * @dev: Device to send idle notification for.
  * @rpmflags: Flag bits.
  *
  * If the RPM_GET_PUT flag is set, decrement the device's usage count and
- * return immediately if it is larger than zero.  Then carry out an idle
+ * return immediately if it is larger than zero (if it becomes negative, log a
+ * warning, increment it, and return an error).  Then carry out an idle
  * notification, either synchronous or asynchronous.
  *
  * This routine may be called in atomic context if the RPM_ASYNC flag is set,
@@ -1057,7 +1077,10 @@ int __pm_runtime_idle(struct device *dev, int rpmflags)
        int retval;
 
        if (rpmflags & RPM_GET_PUT) {
-               if (!atomic_dec_and_test(&dev->power.usage_count)) {
+               retval = rpm_drop_usage_count(dev);
+               if (retval < 0) {
+                       return retval;
+               } else if (retval > 0) {
                        trace_rpm_usage_rcuidle(dev, rpmflags);
                        return 0;
                }
@@ -1079,7 +1102,8 @@ EXPORT_SYMBOL_GPL(__pm_runtime_idle);
  * @rpmflags: Flag bits.
  *
  * If the RPM_GET_PUT flag is set, decrement the device's usage count and
- * return immediately if it is larger than zero.  Then carry out a suspend,
+ * return immediately if it is larger than zero (if it becomes negative, log a
+ * warning, increment it, and return an error).  Then carry out a suspend,
  * either synchronous or asynchronous.
  *
  * This routine may be called in atomic context if the RPM_ASYNC flag is set,
@@ -1091,7 +1115,10 @@ int __pm_runtime_suspend(struct device *dev, int rpmflags)
        int retval;
 
        if (rpmflags & RPM_GET_PUT) {
-               if (!atomic_dec_and_test(&dev->power.usage_count)) {
+               retval = rpm_drop_usage_count(dev);
+               if (retval < 0) {
+                       return retval;
+               } else if (retval > 0) {
                        trace_rpm_usage_rcuidle(dev, rpmflags);
                        return 0;
                }
@@ -1210,12 +1237,13 @@ int __pm_runtime_set_status(struct device *dev, unsigned int status)
 {
        struct device *parent = dev->parent;
        bool notify_parent = false;
+       unsigned long flags;
        int error = 0;
 
        if (status != RPM_ACTIVE && status != RPM_SUSPENDED)
                return -EINVAL;
 
-       spin_lock_irq(&dev->power.lock);
+       spin_lock_irqsave(&dev->power.lock, flags);
 
        /*
         * Prevent PM-runtime from being enabled for the device or return an
@@ -1226,7 +1254,7 @@ int __pm_runtime_set_status(struct device *dev, unsigned int status)
        else
                error = -EAGAIN;
 
-       spin_unlock_irq(&dev->power.lock);
+       spin_unlock_irqrestore(&dev->power.lock, flags);
 
        if (error)
                return error;
@@ -1247,7 +1275,7 @@ int __pm_runtime_set_status(struct device *dev, unsigned int status)
                device_links_read_unlock(idx);
        }
 
-       spin_lock_irq(&dev->power.lock);
+       spin_lock_irqsave(&dev->power.lock, flags);
 
        if (dev->power.runtime_status == status || !parent)
                goto out_set;
@@ -1288,7 +1316,7 @@ int __pm_runtime_set_status(struct device *dev, unsigned int status)
                dev->power.runtime_error = 0;
 
  out:
-       spin_unlock_irq(&dev->power.lock);
+       spin_unlock_irqrestore(&dev->power.lock, flags);
 
        if (notify_parent)
                pm_request_idle(parent);
@@ -1527,14 +1555,17 @@ EXPORT_SYMBOL_GPL(pm_runtime_forbid);
  */
 void pm_runtime_allow(struct device *dev)
 {
+       int ret;
+
        spin_lock_irq(&dev->power.lock);
        if (dev->power.runtime_auto)
                goto out;
 
        dev->power.runtime_auto = true;
-       if (atomic_dec_and_test(&dev->power.usage_count))
+       ret = rpm_drop_usage_count(dev);
+       if (ret == 0)
                rpm_idle(dev, RPM_AUTO | RPM_ASYNC);
-       else
+       else if (ret > 0)
                trace_rpm_usage_rcuidle(dev, RPM_AUTO | RPM_ASYNC);
 
  out: