Loading include/trace/events/cma.h +39 −1 Original line number Diff line number Diff line Loading @@ -7,7 +7,7 @@ #include <linux/types.h> #include <linux/tracepoint.h> TRACE_EVENT(cma_alloc, DECLARE_EVENT_CLASS(cma_alloc_class, TP_PROTO(unsigned long pfn, const struct page *page, unsigned int count, unsigned int align), Loading Loading @@ -60,6 +60,44 @@ TRACE_EVENT(cma_release, __entry->count) ); TRACE_EVENT(cma_alloc_start, TP_PROTO(unsigned int count, unsigned int align), TP_ARGS(count, align), TP_STRUCT__entry( __field(unsigned int, count) __field(unsigned int, align) ), TP_fast_assign( __entry->count = count; __entry->align = align; ), TP_printk("count=%u align=%u", __entry->count, __entry->align) ); DEFINE_EVENT(cma_alloc_class, cma_alloc, TP_PROTO(unsigned long pfn, const struct page *page, unsigned int count, unsigned int align), TP_ARGS(pfn, page, count, align) ); DEFINE_EVENT(cma_alloc_class, cma_alloc_busy_retry, TP_PROTO(unsigned long pfn, const struct page *page, unsigned int count, unsigned int align), TP_ARGS(pfn, page, count, align) ); #endif /* _TRACE_CMA_H */ /* This part must be outside protection */ Loading include/trace/events/migrate.h +21 −0 Original line number Diff line number Diff line Loading @@ -96,6 +96,27 @@ TRACE_EVENT(mm_numa_migrate_ratelimit, __entry->dst_nid, __entry->nr_pages) ); TRACE_EVENT(mm_migrate_pages_start, TP_PROTO(enum migrate_mode mode, int reason), TP_ARGS(mode, reason), TP_STRUCT__entry( __field(enum migrate_mode, mode) __field(int, reason) ), TP_fast_assign( __entry->mode = mode; __entry->reason = reason; ), TP_printk("mode=%s reason=%s", __print_symbolic(__entry->mode, MIGRATE_MODE), __print_symbolic(__entry->reason, MIGRATE_REASON)) ); #endif /* _TRACE_MIGRATE_H */ /* This part must be outside protection */ Loading mm/cma.c +4 −0 Original line number Diff line number Diff line Loading @@ -380,6 +380,8 @@ struct page *cma_alloc(struct cma *cma, size_t count, unsigned int align) if (!count) return NULL; trace_cma_alloc_start(count, align); mask = cma_bitmap_aligned_mask(cma, align); offset = cma_bitmap_aligned_offset(cma, align); bitmap_maxno = cma_bitmap_maxno(cma); Loading Loading @@ -420,6 +422,8 @@ struct page *cma_alloc(struct cma *cma, size_t count, unsigned int align) pr_debug("%s(): memory range at %p is busy, retrying\n", __func__, pfn_to_page(pfn)); trace_cma_alloc_busy_retry(pfn, pfn_to_page(pfn), count, align); /* try again with a bit different memory target */ start = bitmap_no + mask + 1; } Loading mm/migrate.c +2 −0 Original line number Diff line number Diff line Loading @@ -1319,6 +1319,8 @@ int migrate_pages(struct list_head *from, new_page_t get_new_page, int swapwrite = current->flags & PF_SWAPWRITE; int rc; trace_mm_migrate_pages_start(mode, reason); if (!swapwrite) current->flags |= PF_SWAPWRITE; Loading Loading
include/trace/events/cma.h +39 −1 Original line number Diff line number Diff line Loading @@ -7,7 +7,7 @@ #include <linux/types.h> #include <linux/tracepoint.h> TRACE_EVENT(cma_alloc, DECLARE_EVENT_CLASS(cma_alloc_class, TP_PROTO(unsigned long pfn, const struct page *page, unsigned int count, unsigned int align), Loading Loading @@ -60,6 +60,44 @@ TRACE_EVENT(cma_release, __entry->count) ); TRACE_EVENT(cma_alloc_start, TP_PROTO(unsigned int count, unsigned int align), TP_ARGS(count, align), TP_STRUCT__entry( __field(unsigned int, count) __field(unsigned int, align) ), TP_fast_assign( __entry->count = count; __entry->align = align; ), TP_printk("count=%u align=%u", __entry->count, __entry->align) ); DEFINE_EVENT(cma_alloc_class, cma_alloc, TP_PROTO(unsigned long pfn, const struct page *page, unsigned int count, unsigned int align), TP_ARGS(pfn, page, count, align) ); DEFINE_EVENT(cma_alloc_class, cma_alloc_busy_retry, TP_PROTO(unsigned long pfn, const struct page *page, unsigned int count, unsigned int align), TP_ARGS(pfn, page, count, align) ); #endif /* _TRACE_CMA_H */ /* This part must be outside protection */ Loading
include/trace/events/migrate.h +21 −0 Original line number Diff line number Diff line Loading @@ -96,6 +96,27 @@ TRACE_EVENT(mm_numa_migrate_ratelimit, __entry->dst_nid, __entry->nr_pages) ); TRACE_EVENT(mm_migrate_pages_start, TP_PROTO(enum migrate_mode mode, int reason), TP_ARGS(mode, reason), TP_STRUCT__entry( __field(enum migrate_mode, mode) __field(int, reason) ), TP_fast_assign( __entry->mode = mode; __entry->reason = reason; ), TP_printk("mode=%s reason=%s", __print_symbolic(__entry->mode, MIGRATE_MODE), __print_symbolic(__entry->reason, MIGRATE_REASON)) ); #endif /* _TRACE_MIGRATE_H */ /* This part must be outside protection */ Loading
mm/cma.c +4 −0 Original line number Diff line number Diff line Loading @@ -380,6 +380,8 @@ struct page *cma_alloc(struct cma *cma, size_t count, unsigned int align) if (!count) return NULL; trace_cma_alloc_start(count, align); mask = cma_bitmap_aligned_mask(cma, align); offset = cma_bitmap_aligned_offset(cma, align); bitmap_maxno = cma_bitmap_maxno(cma); Loading Loading @@ -420,6 +422,8 @@ struct page *cma_alloc(struct cma *cma, size_t count, unsigned int align) pr_debug("%s(): memory range at %p is busy, retrying\n", __func__, pfn_to_page(pfn)); trace_cma_alloc_busy_retry(pfn, pfn_to_page(pfn), count, align); /* try again with a bit different memory target */ start = bitmap_no + mask + 1; } Loading
mm/migrate.c +2 −0 Original line number Diff line number Diff line Loading @@ -1319,6 +1319,8 @@ int migrate_pages(struct list_head *from, new_page_t get_new_page, int swapwrite = current->flags & PF_SWAPWRITE; int rc; trace_mm_migrate_pages_start(mode, reason); if (!swapwrite) current->flags |= PF_SWAPWRITE; Loading