projects
/
linux-2.6-microblaze.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
net: disable netpoll on fresh napis
[linux-2.6-microblaze.git]
/
mm
/
rmap.c
diff --git
a/mm/rmap.c
b/mm/rmap.c
index
6cce9ef
..
83cc459
100644
(file)
--- a/
mm/rmap.c
+++ b/
mm/rmap.c
@@
-672,7
+672,7
@@
static bool should_defer_flush(struct mm_struct *mm, enum ttu_flags flags)
*/
void flush_tlb_batched_pending(struct mm_struct *mm)
{
*/
void flush_tlb_batched_pending(struct mm_struct *mm)
{
- if (
mm->tlb_flush_batched
) {
+ if (
data_race(mm->tlb_flush_batched)
) {
flush_tlb_mm(mm);
/*
flush_tlb_mm(mm);
/*
@@
-1130,7
+1130,7
@@
void do_page_add_anon_rmap(struct page *page,
}
if (first) {
}
if (first) {
- int nr = compound ?
hpage
_nr_pages(page) : 1;
+ int nr = compound ?
thp
_nr_pages(page) : 1;
/*
* We use the irq-unsafe __{inc|mod}_zone_page_stat because
* these counters are not modified in interrupt context, and
/*
* We use the irq-unsafe __{inc|mod}_zone_page_stat because
* these counters are not modified in interrupt context, and
@@
-1169,7
+1169,7
@@
void do_page_add_anon_rmap(struct page *page,
void page_add_new_anon_rmap(struct page *page,
struct vm_area_struct *vma, unsigned long address, bool compound)
{
void page_add_new_anon_rmap(struct page *page,
struct vm_area_struct *vma, unsigned long address, bool compound)
{
- int nr = compound ?
hpage
_nr_pages(page) : 1;
+ int nr = compound ?
thp
_nr_pages(page) : 1;
VM_BUG_ON_VMA(address < vma->vm_start || address >= vma->vm_end, vma);
__SetPageSwapBacked(page);
VM_BUG_ON_VMA(address < vma->vm_start || address >= vma->vm_end, vma);
__SetPageSwapBacked(page);
@@
-1860,7
+1860,7
@@
static void rmap_walk_anon(struct page *page, struct rmap_walk_control *rwc,
return;
pgoff_start = page_to_pgoff(page);
return;
pgoff_start = page_to_pgoff(page);
- pgoff_end = pgoff_start +
hpage
_nr_pages(page) - 1;
+ pgoff_end = pgoff_start +
thp
_nr_pages(page) - 1;
anon_vma_interval_tree_foreach(avc, &anon_vma->rb_root,
pgoff_start, pgoff_end) {
struct vm_area_struct *vma = avc->vma;
anon_vma_interval_tree_foreach(avc, &anon_vma->rb_root,
pgoff_start, pgoff_end) {
struct vm_area_struct *vma = avc->vma;
@@
-1913,7
+1913,7
@@
static void rmap_walk_file(struct page *page, struct rmap_walk_control *rwc,
return;
pgoff_start = page_to_pgoff(page);
return;
pgoff_start = page_to_pgoff(page);
- pgoff_end = pgoff_start +
hpage
_nr_pages(page) - 1;
+ pgoff_end = pgoff_start +
thp
_nr_pages(page) - 1;
if (!locked)
i_mmap_lock_read(mapping);
vma_interval_tree_foreach(vma, &mapping->i_mmap,
if (!locked)
i_mmap_lock_read(mapping);
vma_interval_tree_foreach(vma, &mapping->i_mmap,