{
        struct zram *zram = dev_to_zram(dev);
 
-       return sprintf(buf, "%u\n", zram->stats.pages_zero);
+       return sprintf(buf, "%u\n", atomic_read(&zram->stats.pages_zero));
 }
 
 static ssize_t orig_data_size_show(struct device *dev,
        struct zram *zram = dev_to_zram(dev);
 
        return sprintf(buf, "%llu\n",
-               (u64)(zram->stats.pages_stored) << PAGE_SHIFT);
+               (u64)(atomic_read(&zram->stats.pages_stored)) << PAGE_SHIFT);
 }
 
 static ssize_t compr_data_size_show(struct device *dev,
                 */
                if (zram_test_flag(meta, index, ZRAM_ZERO)) {
                        zram_clear_flag(meta, index, ZRAM_ZERO);
-                       zram->stats.pages_zero--;
+                       atomic_dec(&zram->stats.pages_zero);
                }
                return;
        }
 
        if (unlikely(size > max_zpage_size))
-               zram->stats.bad_compress--;
+               atomic_dec(&zram->stats.bad_compress);
 
        zs_free(meta->mem_pool, handle);
 
        if (size <= PAGE_SIZE / 2)
-               zram->stats.good_compress--;
+               atomic_dec(&zram->stats.good_compress);
 
        atomic64_sub(meta->table[index].size, &zram->stats.compr_size);
-       zram->stats.pages_stored--;
+       atomic_dec(&zram->stats.pages_stored);
 
        meta->table[index].handle = 0;
        meta->table[index].size = 0;
                /* Free memory associated with this sector now. */
                zram_free_page(zram, index);
 
-               zram->stats.pages_zero++;
+               atomic_inc(&zram->stats.pages_zero);
                zram_set_flag(meta, index, ZRAM_ZERO);
                ret = 0;
                goto out;
        }
 
        if (unlikely(clen > max_zpage_size)) {
-               zram->stats.bad_compress++;
+               atomic_inc(&zram->stats.bad_compress);
                clen = PAGE_SIZE;
                src = NULL;
                if (is_partial_io(bvec))
 
        /* Update stats */
        atomic64_add(clen, &zram->stats.compr_size);
-       zram->stats.pages_stored++;
+       atomic_inc(&zram->stats.pages_stored);
        if (clen <= PAGE_SIZE / 2)
-               zram->stats.good_compress++;
+               atomic_inc(&zram->stats.good_compress);
 
 out:
        if (is_partial_io(bvec))
 
        u8 flags;
 } __aligned(4);
 
-/*
- * All 64bit fields should only be manipulated by 64bit atomic accessors.
- * All modifications to 32bit counter should be protected by zram->lock.
- */
 struct zram_stats {
        atomic64_t compr_size;  /* compressed size of pages stored */
        atomic64_t num_reads;   /* failed + successful */
        atomic64_t failed_writes;       /* can happen when memory is too low */
        atomic64_t invalid_io;  /* non-page-aligned I/O requests */
        atomic64_t notify_free; /* no. of swap slot free notifications */
-       u32 pages_zero;         /* no. of zero filled pages */
-       u32 pages_stored;       /* no. of pages currently stored */
-       u32 good_compress;      /* % of pages with compression ratio<=50% */
-       u32 bad_compress;       /* % of pages with compression ratio>=75% */
+       atomic_t pages_zero;            /* no. of zero filled pages */
+       atomic_t pages_stored;  /* no. of pages currently stored */
+       atomic_t good_compress; /* % of pages with compression ratio<=50% */
+       atomic_t bad_compress;  /* % of pages with compression ratio>=75% */
 };
 
 struct zram_meta {
 struct zram {
        struct zram_meta *meta;
        struct rw_semaphore lock; /* protect compression buffers, table,
-                                  * 32bit stat counters against concurrent
-                                  * notifications, reads and writes */
+                                  * reads and writes
+                                  */
 
        struct work_struct free_work;  /* handle pending free request */
        struct zram_slot_free *slot_free_rq; /* list head of free request */