@@ -471,6 +471,26 @@ Memory Area, or VMA) there is a series of lines such as the following::
KernelPageSize: 4 kB
MMUPageSize: 4 kB
Locked: 0 kB
+ AnonCont4K: 0 kB
+ AnonCont8K: 0 kB
+ AnonCont16K: 0 kB
+ AnonCont32K: 0 kB
+ AnonCont64K: 0 kB
+ AnonCont128K: 0 kB
+ AnonCont256K: 0 kB
+ AnonCont512K: 0 kB
+ AnonCont1M: 0 kB
+ AnonCont2M: 0 kB
+ FileCont4K: 348 kB
+ FileCont8K: 0 kB
+ FileCont16K: 32 kB
+ FileCont32K: 0 kB
+ FileCont64K: 512 kB
+ FileCont128K: 0 kB
+ FileCont256K: 0 kB
+ FileCont512K: 0 kB
+ FileCont1M: 0 kB
+ FileCont2M: 0 kB
THPeligible: 0
VmFlags: rd ex mr mw me dw
@@ -524,6 +544,12 @@ replaced by copy-on-write) part of the underlying shmem object out on swap.
does not take into account swapped out page of underlying shmem objects.
"Locked" indicates whether the mapping is locked in memory or not.
+"AnonContXXX" and "FileContXXX" indicate physically contiguous runs of memory,
+binned into power-of-2 sizes starting with the page size and ending with the
+pmd size. Therefore the exact set of keys will vary by platform. It only
+includes pte-mapped memory and reports on anonymous and file-backed memory
+separately.
+
"THPeligible" indicates whether the mapping is eligible for allocating THP
pages as well as the THP is PMD mappable or not - 1 if true, 0 otherwise.
It just shows the current status.
@@ -397,6 +397,49 @@ const struct file_operations proc_pid_maps_operations = {
#define PSS_SHIFT 12
#ifdef CONFIG_PROC_PAGE_MONITOR
+
+#define CONT_ORDER_MAX (PMD_SHIFT-PAGE_SHIFT)
+#define CONT_LABEL_FIELD_SIZE 8
+#define CONT_LABEL_BUF_SIZE 32
+
+static char *cont_label(int order, char buf[CONT_LABEL_BUF_SIZE])
+{
+ unsigned long size = ((1UL << order) * PAGE_SIZE) >> 10;
+ char suffix = 'K';
+ int count;
+
+ if (size >= SZ_1K) {
+ size >>= 10;
+ suffix = 'M';
+ }
+
+ if (size >= SZ_1K) {
+ size >>= 10;
+ suffix = 'G';
+ }
+
+ count = snprintf(buf, CONT_LABEL_BUF_SIZE, "%lu%c:", size, suffix);
+
+ /*
+ * If the string is less than the field size, pad it with spaces so that
+ * the values line up in smaps.
+ */
+ if (count < CONT_LABEL_FIELD_SIZE) {
+ memset(&buf[count], ' ', CONT_LABEL_FIELD_SIZE - count);
+ buf[CONT_LABEL_FIELD_SIZE] = '\0';
+ }
+
+ return buf;
+}
+
+struct cont_accumulator {
+ bool anon;
+ unsigned long folio_start_pfn;
+ unsigned long folio_end_pfn;
+ unsigned long next_pfn;
+ unsigned long nrpages;
+};
+
struct mem_size_stats {
unsigned long resident;
unsigned long shared_clean;
@@ -419,8 +462,60 @@ struct mem_size_stats {
u64 pss_dirty;
u64 pss_locked;
u64 swap_pss;
+ unsigned long anon_cont[CONT_ORDER_MAX + 1];
+ unsigned long file_cont[CONT_ORDER_MAX + 1];
+ struct cont_accumulator cacc;
};
+static void cacc_init(struct mem_size_stats *mss)
+{
+ struct cont_accumulator *cacc = &mss->cacc;
+
+ cacc->next_pfn = -1;
+ cacc->nrpages = 0;
+}
+
+static void cacc_drain(struct mem_size_stats *mss)
+{
+ struct cont_accumulator *cacc = &mss->cacc;
+ unsigned long *cont = cacc->anon ? mss->anon_cont : mss->file_cont;
+ unsigned long order;
+ unsigned long nrpages;
+
+ while (cacc->nrpages > 0) {
+ order = ilog2(cacc->nrpages);
+ nrpages = 1UL << order;
+ cacc->nrpages -= nrpages;
+ cont[order] += nrpages * PAGE_SIZE;
+ }
+}
+
+static void cacc_accumulate(struct mem_size_stats *mss, struct page *page)
+{
+ struct cont_accumulator *cacc = &mss->cacc;
+ unsigned long pfn = page_to_pfn(page);
+ bool anon = PageAnon(page);
+ struct folio *folio;
+ unsigned long start_pfn;
+
+ if (cacc->next_pfn == pfn && cacc->anon == anon &&
+ pfn >= cacc->folio_start_pfn && pfn < cacc->folio_end_pfn) {
+ cacc->next_pfn++;
+ cacc->nrpages++;
+ } else {
+ cacc_drain(mss);
+
+ folio = page_folio(page);
+ start_pfn = page_to_pfn(&folio->page);
+
+ cacc->anon = anon;
+ cacc->folio_start_pfn = start_pfn;
+ cacc->folio_end_pfn = start_pfn + folio_nr_pages(folio);
+ cacc->next_pfn = pfn + 1;
+ cacc->nrpages = 1;
+ }
+}
+
static void smaps_page_accumulate(struct mem_size_stats *mss,
struct page *page, unsigned long size, unsigned long pss,
bool dirty, bool locked, bool private)
@@ -473,6 +568,10 @@ static void smaps_account(struct mem_size_stats *mss, struct page *page,
if (young || page_is_young(page) || PageReferenced(page))
mss->referenced += size;
+ /* Accumulate physically contiguous map size information. */
+ if (!compound)
+ cacc_accumulate(mss, page);
+
/*
* Then accumulate quantities that may depend on sharing, or that may
* differ page-by-page.
@@ -622,6 +721,7 @@ static int smaps_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end,
struct mm_walk *walk)
{
struct vm_area_struct *vma = walk->vma;
+ struct mem_size_stats *mss = walk->private;
pte_t *pte;
spinlock_t *ptl;
@@ -632,6 +732,7 @@ static int smaps_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end,
goto out;
}
+ cacc_init(mss);
pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
if (!pte) {
walk->action = ACTION_AGAIN;
@@ -640,6 +741,7 @@ static int smaps_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end,
for (; addr != end; pte++, addr += PAGE_SIZE)
smaps_pte_entry(pte, addr, walk);
pte_unmap_unlock(pte - 1, ptl);
+ cacc_drain(mss);
out:
cond_resched();
return 0;
@@ -816,6 +918,9 @@ static void smap_gather_stats(struct vm_area_struct *vma,
static void __show_smap(struct seq_file *m, const struct mem_size_stats *mss,
bool rollup_mode)
{
+ int i;
+ char label[CONT_LABEL_BUF_SIZE];
+
SEQ_PUT_DEC("Rss: ", mss->resident);
SEQ_PUT_DEC(" kB\nPss: ", mss->pss >> PSS_SHIFT);
SEQ_PUT_DEC(" kB\nPss_Dirty: ", mss->pss_dirty >> PSS_SHIFT);
@@ -849,6 +954,16 @@ static void __show_smap(struct seq_file *m, const struct mem_size_stats *mss,
mss->swap_pss >> PSS_SHIFT);
SEQ_PUT_DEC(" kB\nLocked: ",
mss->pss_locked >> PSS_SHIFT);
+ for (i = 0; i <= CONT_ORDER_MAX; i++) {
+ seq_printf(m, " kB\nAnonCont%s%8lu",
+ cont_label(i, label),
+ mss->anon_cont[i] >> 10);
+ }
+ for (i = 0; i <= CONT_ORDER_MAX; i++) {
+ seq_printf(m, " kB\nFileCont%s%8lu",
+ cont_label(i, label),
+ mss->file_cont[i] >> 10);
+ }
seq_puts(m, " kB\n");
}