@@ -211,20 +211,28 @@ extern unsigned long zone_reclaimable_pages(struct zone *zone);
static inline unsigned long node_page_state(int node,
enum zone_stat_item item)
{
- struct zone *zones = NODE_DATA(node)->node_zones;
+ int i;
+ unsigned long page_state = 0;
+
+ for_each_mem_region_in_nid(i, node) {
+ mr_data_t *mrdat = &(NODE_DATA(node)->mem_regions[i]);
+ struct zone *zones = mrdat->zones;
+
+ page_state =
- return
#ifdef CONFIG_ZONE_DMA
- zone_page_state(&zones[ZONE_DMA], item) +
+ zone_page_state(&zones[ZONE_DMA], item) +
#endif
#ifdef CONFIG_ZONE_DMA32
- zone_page_state(&zones[ZONE_DMA32], item) +
+ zone_page_state(&zones[ZONE_DMA32], item) +
#endif
#ifdef CONFIG_HIGHMEM
- zone_page_state(&zones[ZONE_HIGHMEM], item) +
+ zone_page_state(&zones[ZONE_HIGHMEM], item) +
#endif
- zone_page_state(&zones[ZONE_NORMAL], item) +
- zone_page_state(&zones[ZONE_MOVABLE], item);
+ zone_page_state(&zones[ZONE_NORMAL], item) +
+ zone_page_state(&zones[ZONE_MOVABLE], item);
+ }
+ return page_state;
}
extern void zone_statistics(struct zone *, struct zone *, gfp_t gfp);
@@ -191,17 +191,21 @@ void set_pgdat_percpu_threshold(pg_data_t *pgdat,
struct zone *zone;
int cpu;
int threshold;
- int i;
+ int i, p;
for (i = 0; i < pgdat->nr_zones; i++) {
- zone = &pgdat->node_zones[i];
- if (!zone->percpu_drift_mark)
- continue;
-
- threshold = (*calculate_pressure)(zone);
- for_each_possible_cpu(cpu)
- per_cpu_ptr(zone->pageset, cpu)->stat_threshold
- = threshold;
+ for_each_mem_region_in_nid(p, pgdat->node_id) {
+ mem_region_t *mem_region = &pgdat->mem_regions[p];
+ struct zone *zone = mem_region->zones + i;
+
+ if (!zone->percpu_drift_mark)
+ continue;
+
+ threshold = (*calculate_pressure)(zone);
+ for_each_possible_cpu(cpu)
+ per_cpu_ptr(zone->pageset, cpu)->stat_threshold
+ = threshold;
+ }
}
}
@@ -642,19 +646,23 @@ static void frag_stop(struct seq_file *m, void *arg)
/* Walk all the zones in a node and print using a callback */
static void walk_zones_in_node(struct seq_file *m, pg_data_t *pgdat,
- void (*print)(struct seq_file *m, pg_data_t *, struct zone *))
+ void (*print)(struct seq_file *m, pg_data_t *,
+ mem_region_t *, struct zone *))
{
- struct zone *zone;
- struct zone *node_zones = pgdat->node_zones;
unsigned long flags;
-
- for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) {
- if (!populated_zone(zone))
- continue;
-
- spin_lock_irqsave(&zone->lock, flags);
- print(m, pgdat, zone);
- spin_unlock_irqrestore(&zone->lock, flags);
+ int i, j;
+
+ for (i = 0; i < MAX_NR_ZONES; ++i) {
+ for_each_mem_region_in_nid(j, pgdat->node_id) {
+ mem_region_t *mem_region = &pgdat->mem_regions[j];
+ struct zone *zone = mem_region->zones + i;
+ if (!populated_zone(zone))
+ continue;
+
+ spin_lock_irqsave(&zone->lock, flags);
+ print(m, pgdat, mem_region, zone);
+ spin_unlock_irqrestore(&zone->lock, flags);
+ }
}
}
#endif
Change the way vmstats are collected. Since the zones are now present inside regions, scan through all the regions to obtain zone specific statistics. Signed-off-by: Ankita Garg <ankita@in.ibm.com> --- include/linux/vmstat.h | 22 +++++++++++++++------- mm/vmstat.c | 48 ++++++++++++++++++++++++++++-------------------- 2 files changed, 43 insertions(+), 27 deletions(-)