From: Christoph Lameter Date: Fri, 30 Jun 2006 08:55:41 +0000 (-0700) Subject: [PATCH] zoned vm counters: conversion of nr_bounce to per zone counter X-Git-Tag: v2.6.18-rc1~303 X-Git-Url: https://err.no/cgi-bin/gitweb.cgi?a=commitdiff_plain;h=d2c5e30c9a1420902262aa923794d2ae4e0bc391;p=linux-2.6 [PATCH] zoned vm counters: conversion of nr_bounce to per zone counter Conversion of nr_bounce to a per zone counter nr_bounce is only used for proc output. So it could be left as an event counter. However, the event counters may not be accurate and nr_bounce is categorizing types of pages in a zone. So we really need this to also be a per zone counter. [akpm@osdl.org: bugfix] Signed-off-by: Christoph Lameter Signed-off-by: Andrew Morton Signed-off-by: Linus Torvalds --- diff --git a/drivers/base/node.c b/drivers/base/node.c index b73d869c9d..772eadac57 100644 --- a/drivers/base/node.c +++ b/drivers/base/node.c @@ -65,6 +65,7 @@ static ssize_t node_read_meminfo(struct sys_device * dev, char * buf) "Node %d AnonPages: %8lu kB\n" "Node %d PageTables: %8lu kB\n" "Node %d NFS Unstable: %8lu kB\n" + "Node %d Bounce: %8lu kB\n" "Node %d Slab: %8lu kB\n", nid, K(i.totalram), nid, K(i.freeram), @@ -82,6 +83,7 @@ static ssize_t node_read_meminfo(struct sys_device * dev, char * buf) nid, K(node_page_state(nid, NR_ANON_PAGES)), nid, K(node_page_state(nid, NR_PAGETABLE)), nid, K(node_page_state(nid, NR_UNSTABLE_NFS)), + nid, K(node_page_state(nid, NR_BOUNCE)), nid, K(node_page_state(nid, NR_SLAB))); n += hugetlb_report_node_meminfo(nid, buf + n); return n; diff --git a/fs/proc/proc_misc.c b/fs/proc/proc_misc.c index eae66b41a1..6aa2aa779a 100644 --- a/fs/proc/proc_misc.c +++ b/fs/proc/proc_misc.c @@ -171,6 +171,7 @@ static int meminfo_read_proc(char *page, char **start, off_t off, "Slab: %8lu kB\n" "PageTables: %8lu kB\n" "NFS Unstable: %8lu kB\n" + "Bounce: %8lu kB\n" "CommitLimit: %8lu kB\n" "Committed_AS: %8lu kB\n" "VmallocTotal: %8lu kB\n" @@ -196,6 +197,7 @@ static int meminfo_read_proc(char *page, char **start, off_t off, K(global_page_state(NR_SLAB)), K(global_page_state(NR_PAGETABLE)), K(global_page_state(NR_UNSTABLE_NFS)), + K(global_page_state(NR_BOUNCE)), K(allowed), K(committed), (unsigned long)VMALLOC_TOTAL >> 10, diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h index e9d80697f5..2dbeec1d28 100644 --- a/include/linux/mmzone.h +++ b/include/linux/mmzone.h @@ -56,6 +56,7 @@ enum zone_stat_item { NR_FILE_DIRTY, NR_WRITEBACK, NR_UNSTABLE_NFS, /* NFS unstable pages */ + NR_BOUNCE, NR_VM_ZONE_STAT_ITEMS }; struct per_cpu_pages { diff --git a/include/linux/vmstat.h b/include/linux/vmstat.h index 9de2a41c88..5b5b96afc3 100644 --- a/include/linux/vmstat.h +++ b/include/linux/vmstat.h @@ -67,7 +67,6 @@ struct page_state { unsigned long allocstall; /* direct reclaim calls */ unsigned long pgrotated; /* pages rotated to tail of the LRU */ - unsigned long nr_bounce; /* pages for bounce buffers */ }; extern void get_full_page_state(struct page_state *ret); diff --git a/mm/highmem.c b/mm/highmem.c index 9b274fdf9d..9b2a5403c4 100644 --- a/mm/highmem.c +++ b/mm/highmem.c @@ -315,8 +315,8 @@ static void bounce_end_io(struct bio *bio, mempool_t *pool, int err) if (bvec->bv_page == org_vec->bv_page) continue; - mempool_free(bvec->bv_page, pool); - dec_page_state(nr_bounce); + dec_zone_page_state(bvec->bv_page, NR_BOUNCE); + mempool_free(bvec->bv_page, pool); } bio_endio(bio_orig, bio_orig->bi_size, err); @@ -397,7 +397,7 @@ static void __blk_queue_bounce(request_queue_t *q, struct bio **bio_orig, to->bv_page = mempool_alloc(pool, q->bounce_gfp); to->bv_len = from->bv_len; to->bv_offset = from->bv_offset; - inc_page_state(nr_bounce); + inc_zone_page_state(to->bv_page, NR_BOUNCE); if (rw == WRITE) { char *vto, *vfrom; diff --git a/mm/vmstat.c b/mm/vmstat.c index 9dc270aed5..25e5ca7c17 100644 --- a/mm/vmstat.c +++ b/mm/vmstat.c @@ -381,6 +381,7 @@ static char *vmstat_text[] = { "nr_dirty", "nr_writeback", "nr_unstable", + "nr_bounce", /* Event counters */ "pgpgin", @@ -428,7 +429,6 @@ static char *vmstat_text[] = { "allocstall", "pgrotated", - "nr_bounce", }; /*