diff options
| author | Stefan Monnier | 2012-09-24 10:38:10 -0400 |
|---|---|---|
| committer | Stefan Monnier | 2012-09-24 10:38:10 -0400 |
| commit | 3d80c99f3817bf5eccd6acc6a79498a4fde979a4 (patch) | |
| tree | 5377692a9d9b96157a42b8ae693a8f7d18a8bc85 /src | |
| parent | 0970d85fef9830e3b8e5cbfbdc04dbf00cc4c027 (diff) | |
| download | emacs-3d80c99f3817bf5eccd6acc6a79498a4fde979a4.tar.gz emacs-3d80c99f3817bf5eccd6acc6a79498a4fde979a4.zip | |
Rewrite sampler to use Elisp hash-tables.
* src/profiler.c: Remove filtering functionality.
(is_in_trace, Qgc): Remove vars.
(make_log, record_backtrace, Fsample_profiler_log):
Rewrite, using Elisp hash-tables.
(approximate_median, evict_lower_half): New functions.
(cpu_log): Rename from sample_log.
(cpu_gc_count): New var.
(Fsample_profiler_reset, Fmemory_profiler_reset): Remove.
(sigprof_handler): Add count to cpu_gc_count during GC, detected via
backtrace_list.
(block_sigprof, unblock_sigprof): Remove.
(gc_probe, mark_profiler): Remove functions.
(syms_of_profiler): Staticpro cpu_log and memory_log.
* lisp/profiler.el (profiler-sample-interval): Move before first use.
Change default to 1ms.
(profiler-entry=, profiler-backtrace-reverse, profiler-log-fixup-slot)
(profiler-calltree-elapsed<, profiler-calltree-elapsed>): Remove functions.
(profiler-entry-format): Don't use type-of.
(profiler-slot, profiler-log): Remove structs.
(profiler-log-timestamp, profiler-log-type, profiler-log-diff-p):
Redefine for new log representation.
(profiler-log-diff, profiler-log-fixup, profiler-calltree-build-1):
Rewrite for new log representation.
(profiler-calltree): Remove `elapsed' fields.
(profiler-calltree-count<, profiler-report-make-entry-part):
Remove gc special case.
(profiler-calltree-find): Use equal.
(profiler-calltree-walk): Remove `args'; rely on closures instead.
(profiler-calltree-compute-percentages-1): Remove; inlined.
(profiler-calltree-compute-percentages): Simplify.
(profiler-report-log, profiler-report-reversed)
(profiler-report-order): Use defvar-local.
(profiler-report-line-format): Remove `elapsed', do a bit of CSE.
(profiler-report-mode-map): Remove up/down bindings.
(profiler-report-make-buffer-name): Simplify by CSE.
(profiler-report-mode): Remove redundant code.
(profiler-report-expand-entry, profiler-report-collapse-entry):
Use inhibit-read-only.
(profiler-report-render-calltree-1): Simplify by CSE.
(profiler-reset): Rewrite for new subroutines.
(profiler--report-cpu): Rename from sample-profiler-report.
(profiler--report-memory): Rename from memory-profiler-report.
* src/alloc.c (Fgarbage_collect): Record itself in backtrace_list.
Don't set is_in_trace any more. Don't call mark_profiler.
Only call gc_probe for the memory profiler.
(syms_of_alloc): Define Qautomatic_gc.
* src/lisp.h (SXHASH_COMBINE): Move back to...
* src/fns.c (SXHASH_COMBINE): ...here.
* src/xdisp.c (Qautomatic_redisplay): New constant.
(redisplay_internal): Record itself in backtrace_list.
(syms_of_xdisp): Define Qautomatic_redisplay.
* .dir-locals.el (indent-tabs-mode): Remove personal preference.
Diffstat (limited to 'src')
| -rw-r--r-- | src/ChangeLog | 28 | ||||
| -rw-r--r-- | src/alloc.c | 26 | ||||
| -rw-r--r-- | src/fns.c | 7 | ||||
| -rw-r--r-- | src/lisp.h | 11 | ||||
| -rw-r--r-- | src/profiler.c | 1030 | ||||
| -rw-r--r-- | src/xdisp.c | 18 |
6 files changed, 229 insertions, 891 deletions
diff --git a/src/ChangeLog b/src/ChangeLog index feb9c6219fb..1b90ae8b976 100644 --- a/src/ChangeLog +++ b/src/ChangeLog | |||
| @@ -1,3 +1,31 @@ | |||
| 1 | 2012-09-24 Stefan Monnier <monnier@iro.umontreal.ca> | ||
| 2 | |||
| 3 | * xdisp.c (Qautomatic_redisplay): New constant. | ||
| 4 | (redisplay_internal): Record itself in backtrace_list. | ||
| 5 | (syms_of_xdisp): Define Qautomatic_redisplay. | ||
| 6 | |||
| 7 | * profiler.c: Remove filtering functionality. | ||
| 8 | (is_in_trace, Qgc): Remove vars. | ||
| 9 | (make_log, record_backtrace, Fsample_profiler_log): | ||
| 10 | Rewrite, using Elisp hash-tables. | ||
| 11 | (approximate_median, evict_lower_half): New functions. | ||
| 12 | (cpu_log): Rename from sample_log. | ||
| 13 | (cpu_gc_count): New var. | ||
| 14 | (Fsample_profiler_reset, Fmemory_profiler_reset): Remove. | ||
| 15 | (sigprof_handler): Add count to cpu_gc_count during GC, detected via | ||
| 16 | backtrace_list. | ||
| 17 | (block_sigprof, unblock_sigprof): Remove. | ||
| 18 | (gc_probe, mark_profiler): Remove functions. | ||
| 19 | (syms_of_profiler): Staticpro cpu_log and memory_log. | ||
| 20 | |||
| 21 | * lisp.h (SXHASH_COMBINE): Move back to... | ||
| 22 | * fns.c (SXHASH_COMBINE): ...here. | ||
| 23 | |||
| 24 | * alloc.c (Fgarbage_collect): Record itself in backtrace_list. | ||
| 25 | Don't set is_in_trace any more. Don't call mark_profiler. | ||
| 26 | Only call gc_probe for the memory profiler. | ||
| 27 | (syms_of_alloc): Define Qautomatic_gc. | ||
| 28 | |||
| 1 | 2012-09-15 Tomohiro Matsuyama <tomo@cx4a.org> | 29 | 2012-09-15 Tomohiro Matsuyama <tomo@cx4a.org> |
| 2 | 30 | ||
| 3 | * alloc.c (emacs_blocked_malloc): Remove redundant MALLOC_PROBE. | 31 | * alloc.c (emacs_blocked_malloc): Remove redundant MALLOC_PROBE. |
diff --git a/src/alloc.c b/src/alloc.c index 36adb49f835..2fc93f825d1 100644 --- a/src/alloc.c +++ b/src/alloc.c | |||
| @@ -264,6 +264,7 @@ static Lisp_Object Qintervals; | |||
| 264 | static Lisp_Object Qbuffers; | 264 | static Lisp_Object Qbuffers; |
| 265 | static Lisp_Object Qstring_bytes, Qvector_slots, Qheap; | 265 | static Lisp_Object Qstring_bytes, Qvector_slots, Qheap; |
| 266 | static Lisp_Object Qgc_cons_threshold; | 266 | static Lisp_Object Qgc_cons_threshold; |
| 267 | Lisp_Object Qautomatic_gc; | ||
| 267 | Lisp_Object Qchar_table_extra_slots; | 268 | Lisp_Object Qchar_table_extra_slots; |
| 268 | 269 | ||
| 269 | /* Hook run after GC has finished. */ | 270 | /* Hook run after GC has finished. */ |
| @@ -5421,6 +5422,7 @@ See Info node `(elisp)Garbage Collection'. */) | |||
| 5421 | EMACS_TIME start; | 5422 | EMACS_TIME start; |
| 5422 | Lisp_Object retval = Qnil; | 5423 | Lisp_Object retval = Qnil; |
| 5423 | size_t tot_before = 0; | 5424 | size_t tot_before = 0; |
| 5425 | struct backtrace backtrace; | ||
| 5424 | 5426 | ||
| 5425 | if (abort_on_gc) | 5427 | if (abort_on_gc) |
| 5426 | abort (); | 5428 | abort (); |
| @@ -5430,6 +5432,14 @@ See Info node `(elisp)Garbage Collection'. */) | |||
| 5430 | if (pure_bytes_used_before_overflow) | 5432 | if (pure_bytes_used_before_overflow) |
| 5431 | return Qnil; | 5433 | return Qnil; |
| 5432 | 5434 | ||
| 5435 | /* Record this function, so it appears on the profiler's backtraces. */ | ||
| 5436 | backtrace.next = backtrace_list; | ||
| 5437 | backtrace.function = &Qautomatic_gc; | ||
| 5438 | backtrace.args = &Qautomatic_gc; | ||
| 5439 | backtrace.nargs = 0; | ||
| 5440 | backtrace.debug_on_exit = 0; | ||
| 5441 | backtrace_list = &backtrace; | ||
| 5442 | |||
| 5433 | check_cons_list (); | 5443 | check_cons_list (); |
| 5434 | 5444 | ||
| 5435 | /* Don't keep undo information around forever. | 5445 | /* Don't keep undo information around forever. |
| @@ -5486,7 +5496,6 @@ See Info node `(elisp)Garbage Collection'. */) | |||
| 5486 | shrink_regexp_cache (); | 5496 | shrink_regexp_cache (); |
| 5487 | 5497 | ||
| 5488 | gc_in_progress = 1; | 5498 | gc_in_progress = 1; |
| 5489 | is_in_trace = 1; | ||
| 5490 | 5499 | ||
| 5491 | /* Mark all the special slots that serve as the roots of accessibility. */ | 5500 | /* Mark all the special slots that serve as the roots of accessibility. */ |
| 5492 | 5501 | ||
| @@ -5538,8 +5547,6 @@ See Info node `(elisp)Garbage Collection'. */) | |||
| 5538 | mark_backtrace (); | 5547 | mark_backtrace (); |
| 5539 | #endif | 5548 | #endif |
| 5540 | 5549 | ||
| 5541 | mark_profiler (); | ||
| 5542 | |||
| 5543 | #ifdef HAVE_WINDOW_SYSTEM | 5550 | #ifdef HAVE_WINDOW_SYSTEM |
| 5544 | mark_fringe_data (); | 5551 | mark_fringe_data (); |
| 5545 | #endif | 5552 | #endif |
| @@ -5607,7 +5614,6 @@ See Info node `(elisp)Garbage Collection'. */) | |||
| 5607 | check_cons_list (); | 5614 | check_cons_list (); |
| 5608 | 5615 | ||
| 5609 | gc_in_progress = 0; | 5616 | gc_in_progress = 0; |
| 5610 | is_in_trace = 0; | ||
| 5611 | 5617 | ||
| 5612 | consing_since_gc = 0; | 5618 | consing_since_gc = 0; |
| 5613 | if (gc_cons_threshold < GC_DEFAULT_THRESHOLD / 10) | 5619 | if (gc_cons_threshold < GC_DEFAULT_THRESHOLD / 10) |
| @@ -5720,24 +5726,19 @@ See Info node `(elisp)Garbage Collection'. */) | |||
| 5720 | gcs_done++; | 5726 | gcs_done++; |
| 5721 | 5727 | ||
| 5722 | /* Collect profiling data. */ | 5728 | /* Collect profiling data. */ |
| 5723 | if (sample_profiler_running || memory_profiler_running) | 5729 | if (memory_profiler_running) |
| 5724 | { | 5730 | { |
| 5725 | size_t swept = 0; | 5731 | size_t swept = 0; |
| 5726 | size_t elapsed = 0; | ||
| 5727 | if (memory_profiler_running) | 5732 | if (memory_profiler_running) |
| 5728 | { | 5733 | { |
| 5729 | size_t tot_after = total_bytes_of_live_objects (); | 5734 | size_t tot_after = total_bytes_of_live_objects (); |
| 5730 | if (tot_before > tot_after) | 5735 | if (tot_before > tot_after) |
| 5731 | swept = tot_before - tot_after; | 5736 | swept = tot_before - tot_after; |
| 5732 | } | 5737 | } |
| 5733 | if (sample_profiler_running) | 5738 | malloc_probe (swept); |
| 5734 | { | ||
| 5735 | EMACS_TIME since_start = sub_emacs_time (current_emacs_time (), start); | ||
| 5736 | elapsed = EMACS_TIME_TO_DOUBLE (since_start) * 1000; | ||
| 5737 | } | ||
| 5738 | gc_probe (swept, elapsed); | ||
| 5739 | } | 5739 | } |
| 5740 | 5740 | ||
| 5741 | backtrace_list = backtrace.next; | ||
| 5741 | return retval; | 5742 | return retval; |
| 5742 | } | 5743 | } |
| 5743 | 5744 | ||
| @@ -6867,6 +6868,7 @@ do hash-consing of the objects allocated to pure space. */); | |||
| 6867 | DEFSYM (Qstring_bytes, "string-bytes"); | 6868 | DEFSYM (Qstring_bytes, "string-bytes"); |
| 6868 | DEFSYM (Qvector_slots, "vector-slots"); | 6869 | DEFSYM (Qvector_slots, "vector-slots"); |
| 6869 | DEFSYM (Qheap, "heap"); | 6870 | DEFSYM (Qheap, "heap"); |
| 6871 | DEFSYM (Qautomatic_gc, "Automatic GC"); | ||
| 6870 | 6872 | ||
| 6871 | DEFSYM (Qgc_cons_threshold, "gc-cons-threshold"); | 6873 | DEFSYM (Qgc_cons_threshold, "gc-cons-threshold"); |
| 6872 | DEFSYM (Qchar_table_extra_slots, "char-table-extra-slots"); | 6874 | DEFSYM (Qchar_table_extra_slots, "char-table-extra-slots"); |
| @@ -4096,6 +4096,13 @@ sweep_weak_hash_tables (void) | |||
| 4096 | 4096 | ||
| 4097 | #define SXHASH_MAX_LEN 7 | 4097 | #define SXHASH_MAX_LEN 7 |
| 4098 | 4098 | ||
| 4099 | /* Combine two integers X and Y for hashing. The result might not fit | ||
| 4100 | into a Lisp integer. */ | ||
| 4101 | |||
| 4102 | #define SXHASH_COMBINE(X, Y) \ | ||
| 4103 | ((((EMACS_UINT) (X) << 4) + ((EMACS_UINT) (X) >> (BITS_PER_EMACS_INT - 4))) \ | ||
| 4104 | + (EMACS_UINT) (Y)) | ||
| 4105 | |||
| 4099 | /* Hash X, returning a value that fits into a Lisp integer. */ | 4106 | /* Hash X, returning a value that fits into a Lisp integer. */ |
| 4100 | #define SXHASH_REDUCE(X) \ | 4107 | #define SXHASH_REDUCE(X) \ |
| 4101 | ((((X) ^ (X) >> (BITS_PER_EMACS_INT - FIXNUM_BITS))) & INTMASK) | 4108 | ((((X) ^ (X) >> (BITS_PER_EMACS_INT - FIXNUM_BITS))) & INTMASK) |
diff --git a/src/lisp.h b/src/lisp.h index 894b18c838c..09a812829a3 100644 --- a/src/lisp.h +++ b/src/lisp.h | |||
| @@ -2679,11 +2679,6 @@ extern void init_syntax_once (void); | |||
| 2679 | extern void syms_of_syntax (void); | 2679 | extern void syms_of_syntax (void); |
| 2680 | 2680 | ||
| 2681 | /* Defined in fns.c */ | 2681 | /* Defined in fns.c */ |
| 2682 | /* Combine two integers X and Y for hashing. The result might not fit | ||
| 2683 | into a Lisp integer. */ | ||
| 2684 | #define SXHASH_COMBINE(X, Y) \ | ||
| 2685 | ((((EMACS_UINT) (X) << 4) + ((EMACS_UINT) (X) >> (BITS_PER_EMACS_INT - 4))) \ | ||
| 2686 | + (EMACS_UINT) (Y)) | ||
| 2687 | extern Lisp_Object QCrehash_size, QCrehash_threshold; | 2682 | extern Lisp_Object QCrehash_size, QCrehash_threshold; |
| 2688 | enum { NEXT_ALMOST_PRIME_LIMIT = 11 }; | 2683 | enum { NEXT_ALMOST_PRIME_LIMIT = 11 }; |
| 2689 | EXFUN (Fidentity, 1) ATTRIBUTE_CONST; | 2684 | EXFUN (Fidentity, 1) ATTRIBUTE_CONST; |
| @@ -2921,6 +2916,7 @@ build_string (const char *str) | |||
| 2921 | 2916 | ||
| 2922 | extern Lisp_Object pure_cons (Lisp_Object, Lisp_Object); | 2917 | extern Lisp_Object pure_cons (Lisp_Object, Lisp_Object); |
| 2923 | extern void make_byte_code (struct Lisp_Vector *); | 2918 | extern void make_byte_code (struct Lisp_Vector *); |
| 2919 | extern Lisp_Object Qautomatic_gc; | ||
| 2924 | extern Lisp_Object Qchar_table_extra_slots; | 2920 | extern Lisp_Object Qchar_table_extra_slots; |
| 2925 | extern struct Lisp_Vector *allocate_vector (EMACS_INT); | 2921 | extern struct Lisp_Vector *allocate_vector (EMACS_INT); |
| 2926 | extern struct Lisp_Vector *allocate_pseudovector (int memlen, int lisplen, int tag); | 2922 | extern struct Lisp_Vector *allocate_pseudovector (int memlen, int lisplen, int tag); |
| @@ -3532,19 +3528,14 @@ void syms_of_dbusbind (void); | |||
| 3532 | /* Defined in profiler.c */ | 3528 | /* Defined in profiler.c */ |
| 3533 | extern bool sample_profiler_running; | 3529 | extern bool sample_profiler_running; |
| 3534 | extern bool memory_profiler_running; | 3530 | extern bool memory_profiler_running; |
| 3535 | extern bool is_in_trace; | ||
| 3536 | extern Lisp_Object Qgc; | ||
| 3537 | extern void malloc_probe (size_t); | 3531 | extern void malloc_probe (size_t); |
| 3538 | extern void gc_probe (size_t, size_t); | 3532 | extern void gc_probe (size_t, size_t); |
| 3539 | #define ENTER_TRACE (is_in_trace = 1) | ||
| 3540 | #define LEAVE_TRACE (is_in_trace = 0) | ||
| 3541 | #define MALLOC_PROBE(size) \ | 3533 | #define MALLOC_PROBE(size) \ |
| 3542 | do { \ | 3534 | do { \ |
| 3543 | if (memory_profiler_running) \ | 3535 | if (memory_profiler_running) \ |
| 3544 | malloc_probe (size); \ | 3536 | malloc_probe (size); \ |
| 3545 | } while (0) | 3537 | } while (0) |
| 3546 | 3538 | ||
| 3547 | extern void mark_profiler (void); | ||
| 3548 | extern void syms_of_profiler (void); | 3539 | extern void syms_of_profiler (void); |
| 3549 | 3540 | ||
| 3550 | #ifdef DOS_NT | 3541 | #ifdef DOS_NT |
diff --git a/src/profiler.c b/src/profiler.c index 0ef20a9a70c..5eaaaf3330f 100644 --- a/src/profiler.c +++ b/src/profiler.c | |||
| @@ -33,702 +33,103 @@ bool sample_profiler_running; | |||
| 33 | 33 | ||
| 34 | bool memory_profiler_running; | 34 | bool memory_profiler_running; |
| 35 | 35 | ||
| 36 | /* True during tracing. */ | ||
| 37 | |||
| 38 | bool is_in_trace; | ||
| 39 | |||
| 40 | /* Tag for GC entry. */ | ||
| 41 | |||
| 42 | Lisp_Object Qgc; | ||
| 43 | |||
| 44 | static void sigprof_handler (int, siginfo_t *, void *); | 36 | static void sigprof_handler (int, siginfo_t *, void *); |
| 45 | static void block_sigprof (void); | ||
| 46 | static void unblock_sigprof (void); | ||
| 47 | |||
| 48 | |||
| 49 | /* Pattern matching. */ | ||
| 50 | |||
| 51 | enum pattern_type | ||
| 52 | { | ||
| 53 | pattern_exact, /* foo */ | ||
| 54 | pattern_body_exact, /* *foo* */ | ||
| 55 | pattern_pre_any, /* *foo */ | ||
| 56 | pattern_post_any, /* foo* */ | ||
| 57 | pattern_body_any /* foo*bar */ | ||
| 58 | }; | ||
| 59 | |||
| 60 | struct pattern | ||
| 61 | { | ||
| 62 | enum pattern_type type; | ||
| 63 | char *exact; | ||
| 64 | char *extra; | ||
| 65 | int exact_length; | ||
| 66 | int extra_length; | ||
| 67 | }; | ||
| 68 | |||
| 69 | static struct pattern * | ||
| 70 | parse_pattern (const char *pattern) | ||
| 71 | { | ||
| 72 | int length = strlen (pattern); | ||
| 73 | enum pattern_type type; | ||
| 74 | char *exact; | ||
| 75 | char *extra = 0; | ||
| 76 | struct pattern *pat = | ||
| 77 | (struct pattern *) xmalloc (sizeof (struct pattern)); | ||
| 78 | |||
| 79 | if (length > 1 | ||
| 80 | && *pattern == '*' | ||
| 81 | && pattern[length - 1] == '*') | ||
| 82 | { | ||
| 83 | type = pattern_body_exact; | ||
| 84 | exact = xstrdup (pattern + 1); | ||
| 85 | exact[length - 2] = 0; | ||
| 86 | } | ||
| 87 | else if (*pattern == '*') | ||
| 88 | { | ||
| 89 | type = pattern_pre_any; | ||
| 90 | exact = xstrdup (pattern + 1); | ||
| 91 | } | ||
| 92 | else if (pattern[length - 1] == '*') | ||
| 93 | { | ||
| 94 | type = pattern_post_any; | ||
| 95 | exact = xstrdup (pattern); | ||
| 96 | exact[length - 1] = 0; | ||
| 97 | } | ||
| 98 | else if (strchr (pattern, '*')) | ||
| 99 | { | ||
| 100 | type = pattern_body_any; | ||
| 101 | exact = xstrdup (pattern); | ||
| 102 | extra = strchr (exact, '*'); | ||
| 103 | *extra++ = 0; | ||
| 104 | } | ||
| 105 | else | ||
| 106 | { | ||
| 107 | type = pattern_exact; | ||
| 108 | exact = xstrdup (pattern); | ||
| 109 | } | ||
| 110 | |||
| 111 | pat->type = type; | ||
| 112 | pat->exact = exact; | ||
| 113 | pat->extra = extra; | ||
| 114 | pat->exact_length = strlen (exact); | ||
| 115 | pat->extra_length = extra ? strlen (extra) : 0; | ||
| 116 | |||
| 117 | return pat; | ||
| 118 | } | ||
| 119 | |||
| 120 | static void | ||
| 121 | free_pattern (struct pattern *pattern) | ||
| 122 | { | ||
| 123 | xfree (pattern->exact); | ||
| 124 | xfree (pattern); | ||
| 125 | } | ||
| 126 | |||
| 127 | static int | ||
| 128 | pattern_match_1 (enum pattern_type type, | ||
| 129 | const char *exact, | ||
| 130 | int exact_length, | ||
| 131 | const char *string, | ||
| 132 | int length) | ||
| 133 | { | ||
| 134 | if (exact_length > length) | ||
| 135 | return 0; | ||
| 136 | switch (type) | ||
| 137 | { | ||
| 138 | case pattern_exact: | ||
| 139 | return exact_length == length && !strncmp (exact, string, length); | ||
| 140 | case pattern_body_exact: | ||
| 141 | return strstr (string, exact) != 0; | ||
| 142 | case pattern_pre_any: | ||
| 143 | return !strncmp (exact, string + (length - exact_length), exact_length); | ||
| 144 | case pattern_post_any: | ||
| 145 | return !strncmp (exact, string, exact_length); | ||
| 146 | case pattern_body_any: | ||
| 147 | return 0; | ||
| 148 | } | ||
| 149 | } | ||
| 150 | |||
| 151 | static int | ||
| 152 | pattern_match (struct pattern *pattern, const char *string) | ||
| 153 | { | ||
| 154 | int length = strlen (string); | ||
| 155 | switch (pattern->type) | ||
| 156 | { | ||
| 157 | case pattern_body_any: | ||
| 158 | if (pattern->exact_length + pattern->extra_length > length) | ||
| 159 | return 0; | ||
| 160 | return pattern_match_1 (pattern_post_any, | ||
| 161 | pattern->exact, | ||
| 162 | pattern->exact_length, | ||
| 163 | string, length) | ||
| 164 | && pattern_match_1 (pattern_pre_any, | ||
| 165 | pattern->extra, | ||
| 166 | pattern->extra_length, | ||
| 167 | string, length); | ||
| 168 | default: | ||
| 169 | return pattern_match_1 (pattern->type, | ||
| 170 | pattern->exact, | ||
| 171 | pattern->exact_length, | ||
| 172 | string, length); | ||
| 173 | } | ||
| 174 | } | ||
| 175 | |||
| 176 | #if 0 | ||
| 177 | static int | ||
| 178 | match (const char *pattern, const char *string) | ||
| 179 | { | ||
| 180 | int res; | ||
| 181 | struct pattern *pat = parse_pattern (pattern); | ||
| 182 | res = pattern_match (pat, string); | ||
| 183 | free_pattern (pat); | ||
| 184 | return res; | ||
| 185 | } | ||
| 186 | |||
| 187 | static void | ||
| 188 | should_match (const char *pattern, const char *string) | ||
| 189 | { | ||
| 190 | putchar (match (pattern, string) ? '.' : 'F'); | ||
| 191 | } | ||
| 192 | |||
| 193 | static void | ||
| 194 | should_not_match (const char *pattern, const char *string) | ||
| 195 | { | ||
| 196 | putchar (match (pattern, string) ? 'F' : '.'); | ||
| 197 | } | ||
| 198 | |||
| 199 | static void | ||
| 200 | pattern_match_tests (void) | ||
| 201 | { | ||
| 202 | should_match ("", ""); | ||
| 203 | should_not_match ("", "a"); | ||
| 204 | should_match ("a", "a"); | ||
| 205 | should_not_match ("a", "ab"); | ||
| 206 | should_not_match ("ab", "a"); | ||
| 207 | should_match ("*a*", "a"); | ||
| 208 | should_match ("*a*", "ab"); | ||
| 209 | should_match ("*a*", "ba"); | ||
| 210 | should_match ("*a*", "bac"); | ||
| 211 | should_not_match ("*a*", ""); | ||
| 212 | should_not_match ("*a*", "b"); | ||
| 213 | should_match ("*", ""); | ||
| 214 | should_match ("*", "a"); | ||
| 215 | should_match ("a*", "a"); | ||
| 216 | should_match ("a*", "ab"); | ||
| 217 | should_not_match ("a*", ""); | ||
| 218 | should_not_match ("a*", "ba"); | ||
| 219 | should_match ("*a", "a"); | ||
| 220 | should_match ("*a", "ba"); | ||
| 221 | should_not_match ("*a", ""); | ||
| 222 | should_not_match ("*a", "ab"); | ||
| 223 | should_match ("a*b", "ab"); | ||
| 224 | should_match ("a*b", "acb"); | ||
| 225 | should_match ("a*b", "aab"); | ||
| 226 | should_match ("a*b", "abb"); | ||
| 227 | should_not_match ("a*b", ""); | ||
| 228 | should_not_match ("a*b", ""); | ||
| 229 | should_not_match ("a*b", "abc"); | ||
| 230 | puts (""); | ||
| 231 | } | ||
| 232 | #endif | ||
| 233 | |||
| 234 | |||
| 235 | /* Filters. */ | ||
| 236 | |||
| 237 | static struct pattern *filter_pattern; | ||
| 238 | |||
| 239 | /* Set the current filter pattern. If PATTERN is null, unset the | ||
| 240 | current filter pattern instead. */ | ||
| 241 | |||
| 242 | static void | ||
| 243 | set_filter_pattern (const char *pattern) | ||
| 244 | { | ||
| 245 | if (sample_profiler_running) | ||
| 246 | block_sigprof (); | ||
| 247 | |||
| 248 | if (filter_pattern) | ||
| 249 | { | ||
| 250 | free_pattern (filter_pattern); | ||
| 251 | filter_pattern = 0; | ||
| 252 | } | ||
| 253 | if (pattern) | ||
| 254 | filter_pattern = parse_pattern (pattern); | ||
| 255 | |||
| 256 | if (sample_profiler_running) | ||
| 257 | unblock_sigprof (); | ||
| 258 | } | ||
| 259 | |||
| 260 | /* Return true if the current filter pattern is matched with FUNCTION. | ||
| 261 | FUNCTION should be a symbol or a subroutine, otherwise return | ||
| 262 | false. */ | ||
| 263 | |||
| 264 | static int | ||
| 265 | apply_filter_1 (Lisp_Object function) | ||
| 266 | { | ||
| 267 | const char *name; | ||
| 268 | |||
| 269 | if (!filter_pattern) | ||
| 270 | return 1; | ||
| 271 | |||
| 272 | if (SYMBOLP (function)) | ||
| 273 | name = SDATA (SYMBOL_NAME (function)); | ||
| 274 | else if (SUBRP (function)) | ||
| 275 | name = XSUBR (function)->symbol_name; | ||
| 276 | else | ||
| 277 | return 0; | ||
| 278 | |||
| 279 | return pattern_match (filter_pattern, name); | ||
| 280 | } | ||
| 281 | |||
| 282 | /* Return true if the current filter pattern is matched with at least | ||
| 283 | one entry in BACKLIST. */ | ||
| 284 | |||
| 285 | static int | ||
| 286 | apply_filter (struct backtrace *backlist) | ||
| 287 | { | ||
| 288 | while (backlist) | ||
| 289 | { | ||
| 290 | if (apply_filter_1 (*backlist->function)) | ||
| 291 | return 1; | ||
| 292 | backlist = backlist->next; | ||
| 293 | } | ||
| 294 | return 0; | ||
| 295 | } | ||
| 296 | |||
| 297 | DEFUN ("profiler-set-filter-pattern", | ||
| 298 | Fprofiler_set_filter_pattern, Sprofiler_set_filter_pattern, | ||
| 299 | 1, 1, "sPattern: ", | ||
| 300 | doc: /* Set the current filter pattern. PATTERN can contain | ||
| 301 | one or two wildcards (*) as follows: | ||
| 302 | |||
| 303 | - foo | ||
| 304 | - *foo | ||
| 305 | - foo* | ||
| 306 | - *foo* | ||
| 307 | - foo*bar | ||
| 308 | |||
| 309 | If PATTERN is nil or an empty string, then unset the current filter | ||
| 310 | pattern. */) | ||
| 311 | (Lisp_Object pattern) | ||
| 312 | { | ||
| 313 | if (NILP (pattern) | ||
| 314 | || (STRINGP (pattern) && !SREF (pattern, 0))) | ||
| 315 | { | ||
| 316 | set_filter_pattern (0); | ||
| 317 | message ("Profiler filter pattern unset"); | ||
| 318 | return Qt; | ||
| 319 | } | ||
| 320 | else if (!STRINGP (pattern)) | ||
| 321 | error ("Invalid type of profiler filter pattern"); | ||
| 322 | |||
| 323 | set_filter_pattern (SDATA (pattern)); | ||
| 324 | |||
| 325 | return Qt; | ||
| 326 | } | ||
| 327 | 37 | ||
| 328 | 38 | ||
| 329 | /* Backtraces. */ | 39 | /* Logs. */ |
| 330 | 40 | ||
| 41 | typedef struct Lisp_Hash_Table log_t; | ||
| 331 | 42 | ||
| 332 | static Lisp_Object | 43 | static Lisp_Object |
| 333 | make_backtrace (int size) | 44 | make_log (int heap_size, int max_stack_depth) |
| 334 | { | 45 | { |
| 335 | return Fmake_vector (make_number (size), Qnil); | 46 | /* We use a standard Elisp hash-table object, but we use it in |
| 47 | a special way. This is OK as long as the object is not exposed | ||
| 48 | to Elisp, i.e. until it is returned by *-profiler-log, after which | ||
| 49 | it can't be used any more. */ | ||
| 50 | Lisp_Object log = make_hash_table (Qequal, make_number (heap_size), | ||
| 51 | make_float (DEFAULT_REHASH_SIZE), | ||
| 52 | make_float (DEFAULT_REHASH_THRESHOLD), | ||
| 53 | Qnil, Qnil, Qnil); | ||
| 54 | struct Lisp_Hash_Table *h = XHASH_TABLE (log); | ||
| 55 | |||
| 56 | /* What is special about our hash-tables is that the keys are pre-filled | ||
| 57 | with the vectors we'll put in them. */ | ||
| 58 | int i = ASIZE (h->key_and_value) / 2; | ||
| 59 | while (0 < i) | ||
| 60 | set_hash_key_slot (h, --i, | ||
| 61 | Fmake_vector (make_number (max_stack_depth), Qnil)); | ||
| 62 | return log; | ||
| 336 | } | 63 | } |
| 337 | 64 | ||
| 338 | static EMACS_UINT | 65 | /* Evict the least used half of the hash_table. |
| 339 | backtrace_hash (Lisp_Object backtrace) | ||
| 340 | { | ||
| 341 | int i; | ||
| 342 | EMACS_UINT hash = 0; | ||
| 343 | for (i = 0; i < ASIZE (backtrace); i++) | ||
| 344 | /* FIXME */ | ||
| 345 | hash = SXHASH_COMBINE (XUINT (AREF (backtrace, i)), hash); | ||
| 346 | return hash; | ||
| 347 | } | ||
| 348 | 66 | ||
| 349 | static int | 67 | When the table is full, we have to evict someone. |
| 350 | backtrace_equal (Lisp_Object a, Lisp_Object b) | 68 | The easiest and most efficient is to evict the value we're about to add |
| 351 | { | 69 | (i.e. once the table is full, stop sampling). |
| 352 | int i, j; | ||
| 353 | 70 | ||
| 354 | for (i = 0, j = 0;; i++, j++) | 71 | We could also pick the element with the lowest count and evict it, |
| 355 | { | 72 | but finding it is O(N) and for that amount of work we get very |
| 356 | Lisp_Object x = i < ASIZE (a) ? AREF (a, i) : Qnil; | 73 | little in return: for the next sample, this latest sample will have |
| 357 | Lisp_Object y = j < ASIZE (b) ? AREF (b, j) : Qnil; | 74 | count==1 and will hence be a prime candidate for eviction :-( |
| 358 | if (NILP (x) && NILP (y)) | ||
| 359 | break; | ||
| 360 | else if (!EQ (x, y)) | ||
| 361 | return 0; | ||
| 362 | } | ||
| 363 | 75 | ||
| 364 | return 1; | 76 | So instead, we take O(N) time to eliminate more or less half of the |
| 365 | } | 77 | entries (the half with the lowest counts). So we get an amortized |
| 78 | cost of O(1) and we get O(N) time for a new entry to grow larger | ||
| 79 | than the other least counts before a new round of eviction. */ | ||
| 366 | 80 | ||
| 367 | static Lisp_Object | 81 | static EMACS_INT approximate_median (log_t *log, |
| 368 | backtrace_object_1 (Lisp_Object backtrace, int i) | 82 | ptrdiff_t start, ptrdiff_t size) |
| 369 | { | 83 | { |
| 370 | if (i >= ASIZE (backtrace) || NILP (AREF (backtrace, i))) | 84 | eassert (size > 0); |
| 371 | return Qnil; | 85 | if (size < 2) |
| 86 | return XINT (HASH_VALUE (log, start)); | ||
| 87 | if (size < 3) | ||
| 88 | /* Not an actual median, but better for our application than | ||
| 89 | choosing either of the two numbers. */ | ||
| 90 | return ((XINT (HASH_VALUE (log, start)) | ||
| 91 | + XINT (HASH_VALUE (log, start + 1))) | ||
| 92 | / 2); | ||
| 372 | else | 93 | else |
| 373 | return Fcons (AREF (backtrace, i), backtrace_object_1 (backtrace, i + 1)); | ||
| 374 | } | ||
| 375 | |||
| 376 | /* Convert BACKTRACE to a list. */ | ||
| 377 | |||
| 378 | static Lisp_Object | ||
| 379 | backtrace_object (Lisp_Object backtrace) | ||
| 380 | { | ||
| 381 | backtrace_object_1 (backtrace, 0); | ||
| 382 | } | ||
| 383 | |||
| 384 | |||
| 385 | /* Slots. */ | ||
| 386 | |||
| 387 | /* Slot data structure. */ | ||
| 388 | |||
| 389 | struct slot | ||
| 390 | { | ||
| 391 | /* Point to next free slot or next hash table link. */ | ||
| 392 | struct slot *next; | ||
| 393 | /* Point to previous hash table link. */ | ||
| 394 | struct slot *prev; | ||
| 395 | /* Backtrace object with fixed size. */ | ||
| 396 | Lisp_Object backtrace; | ||
| 397 | /* How many times a profiler sees the slot, or how much resouce | ||
| 398 | allocated during profiling. */ | ||
| 399 | size_t count; | ||
| 400 | /* How long the slot takes to execute. */ | ||
| 401 | size_t elapsed; | ||
| 402 | /* True in used. */ | ||
| 403 | unsigned char used : 1; | ||
| 404 | }; | ||
| 405 | |||
| 406 | static void | ||
| 407 | mark_slot (struct slot *slot) | ||
| 408 | { | ||
| 409 | mark_object (slot->backtrace); | ||
| 410 | } | ||
| 411 | |||
| 412 | /* Convert SLOT to a list. */ | ||
| 413 | |||
| 414 | static Lisp_Object | ||
| 415 | slot_object (struct slot *slot) | ||
| 416 | { | ||
| 417 | return list3 (backtrace_object (slot->backtrace), | ||
| 418 | make_number (slot->count), | ||
| 419 | make_number (slot->elapsed)); | ||
| 420 | } | ||
| 421 | |||
| 422 | |||
| 423 | |||
| 424 | /* Slot heaps. */ | ||
| 425 | |||
| 426 | struct slot_heap | ||
| 427 | { | ||
| 428 | /* Number of slots allocated to the heap. */ | ||
| 429 | unsigned int size; | ||
| 430 | /* Actual data area. */ | ||
| 431 | struct slot *data; | ||
| 432 | /* Free list. */ | ||
| 433 | struct slot *free_list; | ||
| 434 | }; | ||
| 435 | |||
| 436 | static void | ||
| 437 | clear_slot_heap (struct slot_heap *heap) | ||
| 438 | { | ||
| 439 | int i; | ||
| 440 | struct slot *data; | ||
| 441 | struct slot *free_list; | ||
| 442 | |||
| 443 | data = heap->data; | ||
| 444 | |||
| 445 | /* Mark all slots unsused. */ | ||
| 446 | for (i = 0; i < heap->size; i++) | ||
| 447 | data[i].used = 0; | ||
| 448 | |||
| 449 | /* Rebuild a free list. */ | ||
| 450 | free_list = heap->free_list = heap->data; | ||
| 451 | for (i = 1; i < heap->size; i++) | ||
| 452 | { | 94 | { |
| 453 | free_list->next = &data[i]; | 95 | ptrdiff_t newsize = size / 3; |
| 454 | free_list = free_list->next; | 96 | ptrdiff_t start2 = start + newsize; |
| 97 | EMACS_INT i1 = approximate_median (log, start, newsize); | ||
| 98 | EMACS_INT i2 = approximate_median (log, start2, newsize); | ||
| 99 | EMACS_INT i3 = approximate_median (log, start2 + newsize, | ||
| 100 | size - 2 * newsize); | ||
| 101 | return (i1 < i2 | ||
| 102 | ? (i2 < i3 ? i2 : (i1 < i3 ? i3 : i1)) | ||
| 103 | : (i1 < i3 ? i1 : (i2 < i3 ? i3 : i2))); | ||
| 455 | } | 104 | } |
| 456 | free_list->next = 0; | ||
| 457 | } | 105 | } |
| 458 | 106 | ||
| 459 | /* Make a slot heap with SIZE. MAX_STACK_DEPTH is a fixed size of | 107 | static void evict_lower_half (log_t *log) |
| 460 | allocated slots. */ | ||
| 461 | |||
| 462 | static struct slot_heap * | ||
| 463 | make_slot_heap (unsigned int size, int max_stack_depth) | ||
| 464 | { | 108 | { |
| 465 | int i; | 109 | ptrdiff_t size = ASIZE (log->key_and_value) / 2; |
| 466 | struct slot_heap *heap; | 110 | EMACS_INT median = approximate_median (log, 0, size); |
| 467 | struct slot *data; | 111 | ptrdiff_t i; |
| 468 | 112 | ||
| 469 | data = (struct slot *) xmalloc (sizeof (struct slot) * size); | ||
| 470 | for (i = 0; i < size; i++) | 113 | for (i = 0; i < size; i++) |
| 471 | data[i].backtrace = make_backtrace (max_stack_depth); | 114 | /* Evict not only values smaller but also values equal to the median, |
| 472 | 115 | so as to make sure we evict something no matter what. */ | |
| 473 | heap = (struct slot_heap *) xmalloc (sizeof (struct slot_heap)); | 116 | if (XINT (HASH_VALUE (log, i)) <= median) |
| 474 | heap->size = size; | 117 | { |
| 475 | heap->data = data; | 118 | Lisp_Object key = HASH_KEY (log, i); |
| 476 | clear_slot_heap (heap); | 119 | { /* FIXME: we could make this more efficient. */ |
| 477 | 120 | Lisp_Object tmp; | |
| 478 | return heap; | 121 | XSET_HASH_TABLE (tmp, log); /* FIXME: Use make_lisp_ptr. */ |
| 479 | } | 122 | Fremhash (key, tmp); |
| 480 | 123 | } | |
| 481 | static void | 124 | eassert (EQ (log->next_free, make_number (i))); |
| 482 | free_slot_heap (struct slot_heap *heap) | 125 | { |
| 483 | { | 126 | int j; |
| 484 | int i; | 127 | eassert (VECTORP (key)); |
| 485 | struct slot *data = heap->data; | 128 | for (j = 0; j < ASIZE (key); j++) |
| 486 | for (i = 0; i < heap->size; i++) | 129 | ASET (key, i, Qnil); |
| 487 | data[i].backtrace = Qnil; | 130 | } |
| 488 | xfree (data); | 131 | set_hash_key_slot (log, i, key); |
| 489 | xfree (heap); | 132 | } |
| 490 | } | ||
| 491 | |||
| 492 | static void | ||
| 493 | mark_slot_heap (struct slot_heap *heap) | ||
| 494 | { | ||
| 495 | int i; | ||
| 496 | for (i = 0; i < heap->size; i++) | ||
| 497 | mark_slot (&heap->data[i]); | ||
| 498 | } | ||
| 499 | |||
| 500 | /* Allocate one slot from HEAP. Return 0 if no free slot in HEAP. */ | ||
| 501 | |||
| 502 | static struct slot * | ||
| 503 | allocate_slot (struct slot_heap *heap) | ||
| 504 | { | ||
| 505 | struct slot *slot; | ||
| 506 | if (!heap->free_list) | ||
| 507 | return 0; | ||
| 508 | slot = heap->free_list; | ||
| 509 | slot->count = 0; | ||
| 510 | slot->elapsed = 0; | ||
| 511 | slot->used = 1; | ||
| 512 | heap->free_list = heap->free_list->next; | ||
| 513 | return slot; | ||
| 514 | } | ||
| 515 | |||
| 516 | static void | ||
| 517 | free_slot (struct slot_heap *heap, struct slot *slot) | ||
| 518 | { | ||
| 519 | eassert (slot->used); | ||
| 520 | slot->used = 0; | ||
| 521 | slot->next = heap->free_list; | ||
| 522 | heap->free_list = slot; | ||
| 523 | } | ||
| 524 | |||
| 525 | /* Return a minimal slot from HEAP. "Minimal" means that such a slot | ||
| 526 | is meaningless for profiling. */ | ||
| 527 | |||
| 528 | static struct slot * | ||
| 529 | min_slot (struct slot_heap *heap) | ||
| 530 | { | ||
| 531 | int i; | ||
| 532 | struct slot *min = 0; | ||
| 533 | for (i = 0; i < heap->size; i++) | ||
| 534 | { | ||
| 535 | struct slot *slot = &heap->data[i]; | ||
| 536 | if (!min || (slot->used && slot->count < min->count)) | ||
| 537 | min = slot; | ||
| 538 | } | ||
| 539 | return min; | ||
| 540 | } | ||
| 541 | |||
| 542 | |||
| 543 | /* Slot hash tables. */ | ||
| 544 | |||
| 545 | struct slot_table | ||
| 546 | { | ||
| 547 | /* Number of slot buckets. */ | ||
| 548 | unsigned int size; | ||
| 549 | /* Buckets data area. */ | ||
| 550 | struct slot **data; | ||
| 551 | }; | ||
| 552 | |||
| 553 | static void | ||
| 554 | clear_slot_table (struct slot_table *table) | ||
| 555 | { | ||
| 556 | int i; | ||
| 557 | for (i = 0; i < table->size; i++) | ||
| 558 | table->data[i] = 0; | ||
| 559 | } | ||
| 560 | |||
| 561 | static struct slot_table * | ||
| 562 | make_slot_table (int size) | ||
| 563 | { | ||
| 564 | struct slot_table *table | ||
| 565 | = (struct slot_table *) xmalloc (sizeof (struct slot_table)); | ||
| 566 | table->size = size; | ||
| 567 | table->data = (struct slot **) xmalloc (sizeof (struct slot *) * size); | ||
| 568 | clear_slot_table (table); | ||
| 569 | return table; | ||
| 570 | } | ||
| 571 | |||
| 572 | static void | ||
| 573 | free_slot_table (struct slot_table *table) | ||
| 574 | { | ||
| 575 | xfree (table->data); | ||
| 576 | xfree (table); | ||
| 577 | } | ||
| 578 | |||
| 579 | static void | ||
| 580 | remove_slot (struct slot_table *table, struct slot *slot) | ||
| 581 | { | ||
| 582 | if (slot->prev) | ||
| 583 | slot->prev->next = slot->next; | ||
| 584 | else | ||
| 585 | { | ||
| 586 | EMACS_UINT hash = backtrace_hash (slot->backtrace); | ||
| 587 | table->data[hash % table->size] = slot->next; | ||
| 588 | } | ||
| 589 | if (slot->next) | ||
| 590 | slot->next->prev = slot->prev; | ||
| 591 | } | ||
| 592 | |||
| 593 | |||
| 594 | /* Logs. */ | ||
| 595 | |||
| 596 | struct log | ||
| 597 | { | ||
| 598 | /* Type of log in symbol. `sample' or `memory'. */ | ||
| 599 | Lisp_Object type; | ||
| 600 | /* Backtrace for working. */ | ||
| 601 | Lisp_Object backtrace; | ||
| 602 | struct slot_heap *slot_heap; | ||
| 603 | struct slot_table *slot_table; | ||
| 604 | size_t others_count; | ||
| 605 | size_t others_elapsed; | ||
| 606 | }; | ||
| 607 | |||
| 608 | static struct log * | ||
| 609 | make_log (const char *type, int heap_size, int max_stack_depth) | ||
| 610 | { | ||
| 611 | struct log *log = | ||
| 612 | (struct log *) xmalloc (sizeof (struct log)); | ||
| 613 | log->type = intern (type); | ||
| 614 | log->backtrace = make_backtrace (max_stack_depth); | ||
| 615 | log->slot_heap = make_slot_heap (heap_size, max_stack_depth); | ||
| 616 | /* Number of buckets of hash table will be 10% of HEAP_SIZE. */ | ||
| 617 | log->slot_table = make_slot_table (max (256, heap_size) / 10); | ||
| 618 | log->others_count = 0; | ||
| 619 | log->others_elapsed = 0; | ||
| 620 | return log; | ||
| 621 | } | ||
| 622 | |||
| 623 | static void | ||
| 624 | free_log (struct log *log) | ||
| 625 | { | ||
| 626 | log->backtrace = Qnil; | ||
| 627 | free_slot_heap (log->slot_heap); | ||
| 628 | free_slot_table (log->slot_table); | ||
| 629 | } | ||
| 630 | |||
| 631 | static void | ||
| 632 | mark_log (struct log *log) | ||
| 633 | { | ||
| 634 | mark_object (log->type); | ||
| 635 | mark_object (log->backtrace); | ||
| 636 | mark_slot_heap (log->slot_heap); | ||
| 637 | } | ||
| 638 | |||
| 639 | static void | ||
| 640 | clear_log (struct log *log) | ||
| 641 | { | ||
| 642 | clear_slot_heap (log->slot_heap); | ||
| 643 | clear_slot_table (log->slot_table); | ||
| 644 | log->others_count = 0; | ||
| 645 | log->others_elapsed = 0; | ||
| 646 | } | ||
| 647 | |||
| 648 | /* Evint SLOT from LOG and accumulate the slot counts into others | ||
| 649 | counts. */ | ||
| 650 | |||
| 651 | static void | ||
| 652 | evict_slot (struct log *log, struct slot *slot) | ||
| 653 | { | ||
| 654 | log->others_count += slot->count; | ||
| 655 | log->others_elapsed += slot->elapsed; | ||
| 656 | remove_slot (log->slot_table, slot); | ||
| 657 | free_slot (log->slot_heap, slot); | ||
| 658 | } | ||
| 659 | |||
| 660 | /* Evict a minimal slot from LOG. */ | ||
| 661 | |||
| 662 | static void | ||
| 663 | evict_min_slot (struct log *log) | ||
| 664 | { | ||
| 665 | struct slot *min = min_slot (log->slot_heap); | ||
| 666 | if (min) | ||
| 667 | evict_slot (log, min); | ||
| 668 | } | ||
| 669 | |||
| 670 | /* Allocate a new slot for BACKTRACE from LOG. The returen value must | ||
| 671 | be a valid pointer to the slot. */ | ||
| 672 | |||
| 673 | static struct slot * | ||
| 674 | new_slot (struct log *log, Lisp_Object backtrace) | ||
| 675 | { | ||
| 676 | int i; | ||
| 677 | struct slot *slot = allocate_slot (log->slot_heap); | ||
| 678 | |||
| 679 | /* If failed to allocate a slot, free some slots to make a room in | ||
| 680 | heap. */ | ||
| 681 | if (!slot) | ||
| 682 | { | ||
| 683 | evict_min_slot (log); | ||
| 684 | slot = allocate_slot (log->slot_heap); | ||
| 685 | /* Must be allocated. */ | ||
| 686 | eassert (slot); | ||
| 687 | } | ||
| 688 | |||
| 689 | slot->prev = 0; | ||
| 690 | slot->next = 0; | ||
| 691 | |||
| 692 | /* Assign BACKTRACE to the slot. */ | ||
| 693 | for (i = 0; i < ASIZE (backtrace); i++) | ||
| 694 | ASET (slot->backtrace, i, AREF (backtrace, i)); | ||
| 695 | |||
| 696 | return slot; | ||
| 697 | } | ||
| 698 | |||
| 699 | /* Make sure that a slot for BACKTRACE is in LOG and return the | ||
| 700 | slot. The return value must be a valid pointer to the slot. */ | ||
| 701 | |||
| 702 | static struct slot * | ||
| 703 | ensure_slot (struct log *log, Lisp_Object backtrace) | ||
| 704 | { | ||
| 705 | EMACS_UINT hash = backtrace_hash (backtrace); | ||
| 706 | int index = hash % log->slot_table->size; | ||
| 707 | struct slot *slot = log->slot_table->data[index]; | ||
| 708 | struct slot *prev = slot; | ||
| 709 | |||
| 710 | /* Looking up in hash table bucket. */ | ||
| 711 | while (slot) | ||
| 712 | { | ||
| 713 | if (backtrace_equal (backtrace, slot->backtrace)) | ||
| 714 | goto found; | ||
| 715 | prev = slot; | ||
| 716 | slot = slot->next; | ||
| 717 | } | ||
| 718 | |||
| 719 | /* If not found, allocate a new slot for BACKTRACE from LOG and link | ||
| 720 | it with bucket chain. */ | ||
| 721 | slot = new_slot (log, backtrace); | ||
| 722 | if (prev) | ||
| 723 | { | ||
| 724 | slot->prev = prev; | ||
| 725 | prev->next = slot; | ||
| 726 | } | ||
| 727 | else | ||
| 728 | log->slot_table->data[index] = slot; | ||
| 729 | |||
| 730 | found: | ||
| 731 | return slot; | ||
| 732 | } | 133 | } |
| 733 | 134 | ||
| 734 | /* Record the current backtrace in LOG. BASE is a special name for | 135 | /* Record the current backtrace in LOG. BASE is a special name for |
| @@ -738,83 +139,60 @@ ensure_slot (struct log *log, Lisp_Object backtrace) | |||
| 738 | took. */ | 139 | took. */ |
| 739 | 140 | ||
| 740 | static void | 141 | static void |
| 741 | record_backtrace_under (struct log *log, Lisp_Object base, | 142 | record_backtrace (log_t *log, size_t count) |
| 742 | size_t count, size_t elapsed) | ||
| 743 | { | 143 | { |
| 744 | int i = 0; | ||
| 745 | Lisp_Object backtrace = log->backtrace; | ||
| 746 | struct backtrace *backlist = backtrace_list; | 144 | struct backtrace *backlist = backtrace_list; |
| 145 | Lisp_Object backtrace; | ||
| 146 | ptrdiff_t index, i = 0; | ||
| 147 | ptrdiff_t asize; | ||
| 747 | 148 | ||
| 748 | /* First of all, apply filter on the bactkrace. */ | 149 | if (!INTEGERP (log->next_free)) |
| 749 | if (!apply_filter (backlist)) return; | 150 | evict_lower_half (log); |
| 151 | index = XINT (log->next_free); | ||
| 750 | 152 | ||
| 751 | /* Record BASE if necessary. */ | 153 | /* Get a "working memory" vector. */ |
| 752 | if (!NILP (base) && ASIZE (backtrace) > 0) | 154 | backtrace = HASH_KEY (log, index); |
| 753 | ASET (backtrace, i++, base); | 155 | asize = ASIZE (backtrace); |
| 754 | 156 | ||
| 755 | /* Copy the backtrace contents into working memory. */ | 157 | /* Copy the backtrace contents into working memory. */ |
| 756 | for (; i < ASIZE (backtrace) && backlist; backlist = backlist->next) | 158 | for (; i < asize && backlist; i++, backlist = backlist->next) |
| 757 | { | 159 | ASET (backtrace, i, *backlist->function); |
| 758 | Lisp_Object function = *backlist->function; | ||
| 759 | if (FUNCTIONP (function)) | ||
| 760 | ASET (backtrace, i++, function); | ||
| 761 | } | ||
| 762 | /* Make sure that unused space of working memory is filled with | ||
| 763 | nil. */ | ||
| 764 | for (; i < ASIZE (backtrace); i++) | ||
| 765 | ASET (backtrace, i, Qnil); | ||
| 766 | |||
| 767 | /* If the backtrace is not empty, */ | ||
| 768 | if (!NILP (AREF (backtrace, 0))) | ||
| 769 | { | ||
| 770 | /* then record counts. */ | ||
| 771 | struct slot *slot = ensure_slot (log, backtrace); | ||
| 772 | slot->count += count; | ||
| 773 | slot->elapsed += elapsed; | ||
| 774 | } | ||
| 775 | } | ||
| 776 | |||
| 777 | static void | ||
| 778 | record_backtrace (struct log *log, size_t count, size_t elapsed) | ||
| 779 | { | ||
| 780 | record_backtrace_under (log, Qnil, count, elapsed); | ||
| 781 | } | ||
| 782 | |||
| 783 | /* Convert LOG to a list. */ | ||
| 784 | 160 | ||
| 785 | static Lisp_Object | 161 | /* Make sure that unused space of working memory is filled with nil. */ |
| 786 | log_object (struct log *log) | 162 | for (; i < asize; i++) |
| 787 | { | 163 | ASET (backtrace, i, Qnil); |
| 788 | int i; | ||
| 789 | Lisp_Object slots = Qnil; | ||
| 790 | |||
| 791 | if (log->others_count != 0 || log->others_elapsed != 0) | ||
| 792 | { | ||
| 793 | /* Add others slot. */ | ||
| 794 | Lisp_Object others_slot | ||
| 795 | = list3 (list1 (Qt), | ||
| 796 | make_number (log->others_count), | ||
| 797 | make_number (log->others_elapsed)); | ||
| 798 | slots = list1 (others_slot); | ||
| 799 | } | ||
| 800 | |||
| 801 | for (i = 0; i < log->slot_heap->size; i++) | ||
| 802 | { | ||
| 803 | struct slot *s = &log->slot_heap->data[i]; | ||
| 804 | if (s->used) | ||
| 805 | { | ||
| 806 | Lisp_Object slot = slot_object (s); | ||
| 807 | slots = Fcons (slot, slots); | ||
| 808 | } | ||
| 809 | } | ||
| 810 | 164 | ||
| 811 | return list4 (log->type, Qnil, Fcurrent_time (), slots); | 165 | { /* We basically do a `gethash+puthash' here, except that we have to be |
| 166 | careful to avoid memory allocation since we're in a signal | ||
| 167 | handler, and we optimize the code to try and avoid computing the | ||
| 168 | hash+lookup twice. See fns.c:Fputhash for reference. */ | ||
| 169 | EMACS_UINT hash; | ||
| 170 | ptrdiff_t j = hash_lookup (log, backtrace, &hash); | ||
| 171 | if (j >= 0) | ||
| 172 | set_hash_value_slot (log, j, | ||
| 173 | make_number (count + XINT (HASH_VALUE (log, j)))); | ||
| 174 | else | ||
| 175 | { /* BEWARE! hash_put in general can allocate memory. | ||
| 176 | But currently it only does that if log->next_free is nil. */ | ||
| 177 | int j; | ||
| 178 | eassert (!NILP (log->next_free)); | ||
| 179 | j = hash_put (log, backtrace, make_number (count), hash); | ||
| 180 | /* Let's make sure we've put `backtrace' right where it | ||
| 181 | already was to start with. */ | ||
| 182 | eassert (index == j); | ||
| 183 | |||
| 184 | /* FIXME: If the hash-table is almost full, we should set | ||
| 185 | some global flag so that some Elisp code can offload its | ||
| 186 | data elsewhere, so as to avoid the eviction code. */ | ||
| 187 | } | ||
| 188 | } | ||
| 812 | } | 189 | } |
| 813 | |||
| 814 | 190 | ||
| 815 | /* Sample profiler. */ | 191 | /* Sample profiler. */ |
| 816 | 192 | ||
| 817 | static struct log *sample_log; | 193 | static Lisp_Object cpu_log; |
| 194 | /* Separate counter for the time spent in the GC. */ | ||
| 195 | static EMACS_INT cpu_gc_count; | ||
| 818 | 196 | ||
| 819 | /* The current sample interval in millisecond. */ | 197 | /* The current sample interval in millisecond. */ |
| 820 | 198 | ||
| @@ -833,10 +211,12 @@ take samples each SAMPLE-INTERVAL in millisecond. See also | |||
| 833 | if (sample_profiler_running) | 211 | if (sample_profiler_running) |
| 834 | error ("Sample profiler is already running"); | 212 | error ("Sample profiler is already running"); |
| 835 | 213 | ||
| 836 | if (!sample_log) | 214 | if (NILP (cpu_log)) |
| 837 | sample_log = make_log ("sample", | 215 | { |
| 838 | profiler_slot_heap_size, | 216 | cpu_gc_count = 0; |
| 839 | profiler_max_stack_depth); | 217 | cpu_log = make_log (profiler_slot_heap_size, |
| 218 | profiler_max_stack_depth); | ||
| 219 | } | ||
| 840 | 220 | ||
| 841 | current_sample_interval = XINT (sample_interval); | 221 | current_sample_interval = XINT (sample_interval); |
| 842 | 222 | ||
| @@ -869,27 +249,6 @@ DEFUN ("sample-profiler-stop", Fsample_profiler_stop, Ssample_profiler_stop, | |||
| 869 | return Qt; | 249 | return Qt; |
| 870 | } | 250 | } |
| 871 | 251 | ||
| 872 | DEFUN ("sample-profiler-reset", Fsample_profiler_reset, Ssample_profiler_reset, | ||
| 873 | 0, 0, 0, | ||
| 874 | doc: /* Clear sample profiler log. */) | ||
| 875 | (void) | ||
| 876 | { | ||
| 877 | if (sample_log) | ||
| 878 | { | ||
| 879 | if (sample_profiler_running) | ||
| 880 | { | ||
| 881 | block_sigprof (); | ||
| 882 | clear_log (sample_log); | ||
| 883 | unblock_sigprof (); | ||
| 884 | } | ||
| 885 | else | ||
| 886 | { | ||
| 887 | free_log (sample_log); | ||
| 888 | sample_log = 0; | ||
| 889 | } | ||
| 890 | } | ||
| 891 | } | ||
| 892 | |||
| 893 | DEFUN ("sample-profiler-running-p", | 252 | DEFUN ("sample-profiler-running-p", |
| 894 | Fsample_profiler_running_p, Ssample_profiler_running_p, | 253 | Fsample_profiler_running_p, Ssample_profiler_running_p, |
| 895 | 0, 0, 0, | 254 | 0, 0, 0, |
| @@ -907,28 +266,24 @@ DEFUN ("sample-profiler-log", | |||
| 907 | log is collected and SLOTS is a list of slots. */) | 266 | log is collected and SLOTS is a list of slots. */) |
| 908 | (void) | 267 | (void) |
| 909 | { | 268 | { |
| 910 | int i; | 269 | Lisp_Object result = cpu_log; |
| 911 | Lisp_Object result = Qnil; | 270 | /* Here we're making the log visible to Elisp , so it's not safe any |
| 912 | 271 | more for our use afterwards since we can't rely on its special | |
| 913 | if (sample_log) | 272 | pre-allocated keys anymore. So we have to allocate a new one. */ |
| 914 | { | 273 | cpu_log = (sample_profiler_running |
| 915 | if (sample_profiler_running) | 274 | ? make_log (profiler_slot_heap_size, profiler_max_stack_depth) |
| 916 | { | 275 | : Qnil); |
| 917 | block_sigprof (); | 276 | Fputhash (Fmake_vector (make_number (1), Qautomatic_gc), |
| 918 | result = log_object (sample_log); | 277 | make_number (cpu_gc_count), |
| 919 | unblock_sigprof (); | 278 | result); |
| 920 | } | 279 | cpu_gc_count = 0; |
| 921 | else | ||
| 922 | result = log_object (sample_log); | ||
| 923 | } | ||
| 924 | |||
| 925 | return result; | 280 | return result; |
| 926 | } | 281 | } |
| 927 | 282 | ||
| 928 | 283 | ||
| 929 | /* Memory profiler. */ | 284 | /* Memory profiler. */ |
| 930 | 285 | ||
| 931 | static struct log *memory_log; | 286 | static Lisp_Object memory_log; |
| 932 | 287 | ||
| 933 | DEFUN ("memory-profiler-start", Fmemory_profiler_start, Smemory_profiler_start, | 288 | DEFUN ("memory-profiler-start", Fmemory_profiler_start, Smemory_profiler_start, |
| 934 | 0, 0, 0, | 289 | 0, 0, 0, |
| @@ -939,9 +294,8 @@ DEFUN ("memory-profiler-start", Fmemory_profiler_start, Smemory_profiler_start, | |||
| 939 | if (memory_profiler_running) | 294 | if (memory_profiler_running) |
| 940 | error ("Memory profiler is already running"); | 295 | error ("Memory profiler is already running"); |
| 941 | 296 | ||
| 942 | if (!memory_log) | 297 | if (NILP (memory_log)) |
| 943 | memory_log = make_log ("memory", | 298 | memory_log = make_log (profiler_slot_heap_size, |
| 944 | profiler_slot_heap_size, | ||
| 945 | profiler_max_stack_depth); | 299 | profiler_max_stack_depth); |
| 946 | 300 | ||
| 947 | memory_profiler_running = 1; | 301 | memory_profiler_running = 1; |
| @@ -962,24 +316,6 @@ DEFUN ("memory-profiler-stop", | |||
| 962 | return Qt; | 316 | return Qt; |
| 963 | } | 317 | } |
| 964 | 318 | ||
| 965 | DEFUN ("memory-profiler-reset", | ||
| 966 | Fmemory_profiler_reset, Smemory_profiler_reset, | ||
| 967 | 0, 0, 0, | ||
| 968 | doc: /* Clear memory profiler log. */) | ||
| 969 | (void) | ||
| 970 | { | ||
| 971 | if (memory_log) | ||
| 972 | { | ||
| 973 | if (memory_profiler_running) | ||
| 974 | clear_log (memory_log); | ||
| 975 | else | ||
| 976 | { | ||
| 977 | free_log (memory_log); | ||
| 978 | memory_log = 0; | ||
| 979 | } | ||
| 980 | } | ||
| 981 | } | ||
| 982 | |||
| 983 | DEFUN ("memory-profiler-running-p", | 319 | DEFUN ("memory-profiler-running-p", |
| 984 | Fmemory_profiler_running_p, Smemory_profiler_running_p, | 320 | Fmemory_profiler_running_p, Smemory_profiler_running_p, |
| 985 | 0, 0, 0, | 321 | 0, 0, 0, |
| @@ -997,11 +333,13 @@ DEFUN ("memory-profiler-log", | |||
| 997 | log is collected and SLOTS is a list of slots. */) | 333 | log is collected and SLOTS is a list of slots. */) |
| 998 | (void) | 334 | (void) |
| 999 | { | 335 | { |
| 1000 | Lisp_Object result = Qnil; | 336 | Lisp_Object result = memory_log; |
| 1001 | 337 | /* Here we're making the log visible to Elisp , so it's not safe any | |
| 1002 | if (memory_log) | 338 | more for our use afterwards since we can't rely on its special |
| 1003 | result = log_object (memory_log); | 339 | pre-allocated keys anymore. So we have to allocate a new one. */ |
| 1004 | 340 | memory_log = (memory_profiler_running | |
| 341 | ? make_log (profiler_slot_heap_size, profiler_max_stack_depth) | ||
| 342 | : Qnil); | ||
| 1005 | return result; | 343 | return result; |
| 1006 | } | 344 | } |
| 1007 | 345 | ||
| @@ -1013,73 +351,31 @@ log is collected and SLOTS is a list of slots. */) | |||
| 1013 | static void | 351 | static void |
| 1014 | sigprof_handler (int signal, siginfo_t *info, void *ctx) | 352 | sigprof_handler (int signal, siginfo_t *info, void *ctx) |
| 1015 | { | 353 | { |
| 1016 | if (!is_in_trace && sample_log) | 354 | eassert (HASH_TABLE_P (cpu_log)); |
| 1017 | record_backtrace (sample_log, 1, current_sample_interval); | 355 | if (backtrace_list && EQ (*backtrace_list->function, Qautomatic_gc)) |
| 1018 | } | 356 | /* Special case the time-count inside GC because the hash-table |
| 1019 | 357 | code is not prepared to be used while the GC is running. | |
| 1020 | static void | 358 | More specifically it uses ASIZE at many places where it does |
| 1021 | block_sigprof (void) | 359 | not expect the ARRAY_MARK_FLAG to be set. We could try and |
| 1022 | { | 360 | harden the hash-table code, but it doesn't seem worth the |
| 1023 | sigset_t sigset; | 361 | effort. */ |
| 1024 | sigemptyset (&sigset); | 362 | cpu_gc_count += current_sample_interval; |
| 1025 | sigaddset (&sigset, SIGPROF); | 363 | else |
| 1026 | sigprocmask (SIG_BLOCK, &sigset, 0); | 364 | record_backtrace (XHASH_TABLE (cpu_log), current_sample_interval); |
| 1027 | } | ||
| 1028 | |||
| 1029 | static void | ||
| 1030 | unblock_sigprof (void) | ||
| 1031 | { | ||
| 1032 | sigset_t sigset; | ||
| 1033 | sigemptyset (&sigset); | ||
| 1034 | sigaddset (&sigset, SIGPROF); | ||
| 1035 | sigprocmask (SIG_UNBLOCK, &sigset, 0); | ||
| 1036 | } | 365 | } |
| 1037 | 366 | ||
| 1038 | /* Record that the current backtrace allocated SIZE bytes. */ | 367 | /* Record that the current backtrace allocated SIZE bytes. */ |
| 1039 | 368 | /* FIXME: Inline it everywhere! */ | |
| 1040 | void | 369 | void |
| 1041 | malloc_probe (size_t size) | 370 | malloc_probe (size_t size) |
| 1042 | { | 371 | { |
| 1043 | if (memory_log) | 372 | if (HASH_TABLE_P (memory_log)) |
| 1044 | record_backtrace (memory_log, size, 0); | 373 | record_backtrace (XHASH_TABLE (memory_log), size); |
| 1045 | } | ||
| 1046 | |||
| 1047 | /* Record that GC happened in the current backtrace. */ | ||
| 1048 | |||
| 1049 | void | ||
| 1050 | gc_probe (size_t size, size_t elapsed) | ||
| 1051 | { | ||
| 1052 | if (sample_log) | ||
| 1053 | record_backtrace_under (sample_log, Qgc, 1, elapsed); | ||
| 1054 | if (memory_log) | ||
| 1055 | record_backtrace_under (memory_log, Qgc, size, elapsed); | ||
| 1056 | } | ||
| 1057 | |||
| 1058 | |||
| 1059 | |||
| 1060 | void | ||
| 1061 | mark_profiler (void) | ||
| 1062 | { | ||
| 1063 | if (sample_log) | ||
| 1064 | { | ||
| 1065 | if (sample_profiler_running) | ||
| 1066 | { | ||
| 1067 | block_sigprof (); | ||
| 1068 | mark_log (sample_log); | ||
| 1069 | unblock_sigprof (); | ||
| 1070 | } | ||
| 1071 | else | ||
| 1072 | mark_log (sample_log); | ||
| 1073 | } | ||
| 1074 | if (memory_log) | ||
| 1075 | mark_log (memory_log); | ||
| 1076 | } | 374 | } |
| 1077 | 375 | ||
| 1078 | void | 376 | void |
| 1079 | syms_of_profiler (void) | 377 | syms_of_profiler (void) |
| 1080 | { | 378 | { |
| 1081 | DEFSYM (Qgc, "gc"); | ||
| 1082 | |||
| 1083 | DEFVAR_INT ("profiler-max-stack-depth", profiler_max_stack_depth, | 379 | DEFVAR_INT ("profiler-max-stack-depth", profiler_max_stack_depth, |
| 1084 | doc: /* FIXME */); | 380 | doc: /* FIXME */); |
| 1085 | profiler_max_stack_depth = 16; | 381 | profiler_max_stack_depth = 16; |
| @@ -1087,17 +383,19 @@ syms_of_profiler (void) | |||
| 1087 | doc: /* FIXME */); | 383 | doc: /* FIXME */); |
| 1088 | profiler_slot_heap_size = 10000; | 384 | profiler_slot_heap_size = 10000; |
| 1089 | 385 | ||
| 1090 | defsubr (&Sprofiler_set_filter_pattern); | 386 | cpu_log = memory_log = Qnil; |
| 387 | staticpro (&cpu_log); | ||
| 388 | staticpro (&memory_log); | ||
| 1091 | 389 | ||
| 390 | /* FIXME: Rename things to start with "profiler-", to use "cpu" instead of | ||
| 391 | "sample", and to make them sound like they're internal or something. */ | ||
| 1092 | defsubr (&Ssample_profiler_start); | 392 | defsubr (&Ssample_profiler_start); |
| 1093 | defsubr (&Ssample_profiler_stop); | 393 | defsubr (&Ssample_profiler_stop); |
| 1094 | defsubr (&Ssample_profiler_reset); | ||
| 1095 | defsubr (&Ssample_profiler_running_p); | 394 | defsubr (&Ssample_profiler_running_p); |
| 1096 | defsubr (&Ssample_profiler_log); | 395 | defsubr (&Ssample_profiler_log); |
| 1097 | 396 | ||
| 1098 | defsubr (&Smemory_profiler_start); | 397 | defsubr (&Smemory_profiler_start); |
| 1099 | defsubr (&Smemory_profiler_stop); | 398 | defsubr (&Smemory_profiler_stop); |
| 1100 | defsubr (&Smemory_profiler_reset); | ||
| 1101 | defsubr (&Smemory_profiler_running_p); | 399 | defsubr (&Smemory_profiler_running_p); |
| 1102 | defsubr (&Smemory_profiler_log); | 400 | defsubr (&Smemory_profiler_log); |
| 1103 | } | 401 | } |
diff --git a/src/xdisp.c b/src/xdisp.c index f5edb4b16f8..ccfa251fd1c 100644 --- a/src/xdisp.c +++ b/src/xdisp.c | |||
| @@ -333,10 +333,10 @@ static Lisp_Object Qinhibit_eval_during_redisplay; | |||
| 333 | static Lisp_Object Qbuffer_position, Qposition, Qobject; | 333 | static Lisp_Object Qbuffer_position, Qposition, Qobject; |
| 334 | static Lisp_Object Qright_to_left, Qleft_to_right; | 334 | static Lisp_Object Qright_to_left, Qleft_to_right; |
| 335 | 335 | ||
| 336 | /* Cursor shapes */ | 336 | /* Cursor shapes. */ |
| 337 | Lisp_Object Qbar, Qhbar, Qbox, Qhollow; | 337 | Lisp_Object Qbar, Qhbar, Qbox, Qhollow; |
| 338 | 338 | ||
| 339 | /* Pointer shapes */ | 339 | /* Pointer shapes. */ |
| 340 | static Lisp_Object Qarrow, Qhand; | 340 | static Lisp_Object Qarrow, Qhand; |
| 341 | Lisp_Object Qtext; | 341 | Lisp_Object Qtext; |
| 342 | 342 | ||
| @@ -347,6 +347,7 @@ static Lisp_Object Qfontification_functions; | |||
| 347 | 347 | ||
| 348 | static Lisp_Object Qwrap_prefix; | 348 | static Lisp_Object Qwrap_prefix; |
| 349 | static Lisp_Object Qline_prefix; | 349 | static Lisp_Object Qline_prefix; |
| 350 | static Lisp_Object Qautomatic_redisplay; | ||
| 350 | 351 | ||
| 351 | /* Non-nil means don't actually do any redisplay. */ | 352 | /* Non-nil means don't actually do any redisplay. */ |
| 352 | 353 | ||
| @@ -12931,12 +12932,13 @@ redisplay_internal (void) | |||
| 12931 | struct frame *sf; | 12932 | struct frame *sf; |
| 12932 | int polling_stopped_here = 0; | 12933 | int polling_stopped_here = 0; |
| 12933 | Lisp_Object old_frame = selected_frame; | 12934 | Lisp_Object old_frame = selected_frame; |
| 12935 | struct backtrace backtrace; | ||
| 12934 | 12936 | ||
| 12935 | /* Non-zero means redisplay has to consider all windows on all | 12937 | /* Non-zero means redisplay has to consider all windows on all |
| 12936 | frames. Zero means, only selected_window is considered. */ | 12938 | frames. Zero means, only selected_window is considered. */ |
| 12937 | int consider_all_windows_p; | 12939 | int consider_all_windows_p; |
| 12938 | 12940 | ||
| 12939 | /* Non-zero means redisplay has to redisplay the miniwindow */ | 12941 | /* Non-zero means redisplay has to redisplay the miniwindow. */ |
| 12940 | int update_miniwindow_p = 0; | 12942 | int update_miniwindow_p = 0; |
| 12941 | 12943 | ||
| 12942 | TRACE ((stderr, "redisplay_internal %d\n", redisplaying_p)); | 12944 | TRACE ((stderr, "redisplay_internal %d\n", redisplaying_p)); |
| @@ -12974,6 +12976,14 @@ redisplay_internal (void) | |||
| 12974 | ++redisplaying_p; | 12976 | ++redisplaying_p; |
| 12975 | specbind (Qinhibit_free_realized_faces, Qnil); | 12977 | specbind (Qinhibit_free_realized_faces, Qnil); |
| 12976 | 12978 | ||
| 12979 | /* Record this function, so it appears on the profiler's backtraces. */ | ||
| 12980 | backtrace.next = backtrace_list; | ||
| 12981 | backtrace.function = &Qautomatic_redisplay; | ||
| 12982 | backtrace.args = &Qautomatic_redisplay; | ||
| 12983 | backtrace.nargs = 0; | ||
| 12984 | backtrace.debug_on_exit = 0; | ||
| 12985 | backtrace_list = &backtrace; | ||
| 12986 | |||
| 12977 | { | 12987 | { |
| 12978 | Lisp_Object tail, frame; | 12988 | Lisp_Object tail, frame; |
| 12979 | 12989 | ||
| @@ -13671,6 +13681,7 @@ redisplay_internal (void) | |||
| 13671 | #endif /* HAVE_WINDOW_SYSTEM */ | 13681 | #endif /* HAVE_WINDOW_SYSTEM */ |
| 13672 | 13682 | ||
| 13673 | end_of_redisplay: | 13683 | end_of_redisplay: |
| 13684 | backtrace_list = backtrace.next; | ||
| 13674 | unbind_to (count, Qnil); | 13685 | unbind_to (count, Qnil); |
| 13675 | RESUME_POLLING; | 13686 | RESUME_POLLING; |
| 13676 | } | 13687 | } |
| @@ -28696,6 +28707,7 @@ syms_of_xdisp (void) | |||
| 28696 | staticpro (&Vmessage_stack); | 28707 | staticpro (&Vmessage_stack); |
| 28697 | 28708 | ||
| 28698 | DEFSYM (Qinhibit_redisplay, "inhibit-redisplay"); | 28709 | DEFSYM (Qinhibit_redisplay, "inhibit-redisplay"); |
| 28710 | DEFSYM (Qautomatic_redisplay, "Automatic Redisplay"); | ||
| 28699 | 28711 | ||
| 28700 | message_dolog_marker1 = Fmake_marker (); | 28712 | message_dolog_marker1 = Fmake_marker (); |
| 28701 | staticpro (&message_dolog_marker1); | 28713 | staticpro (&message_dolog_marker1); |