Commit | Line | Data |
---|---|---|
0efc778b | 1 | /* Profiler implementation. |
c2d7786e TM |
2 | |
3 | Copyright (C) 2012 Free Software Foundation, Inc. | |
4 | ||
5 | This file is part of GNU Emacs. | |
6 | ||
7 | GNU Emacs is free software: you can redistribute it and/or modify | |
8 | it under the terms of the GNU General Public License as published by | |
9 | the Free Software Foundation, either version 3 of the License, or | |
10 | (at your option) any later version. | |
11 | ||
12 | GNU Emacs is distributed in the hope that it will be useful, | |
13 | but WITHOUT ANY WARRANTY; without even the implied warranty of | |
14 | MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
15 | GNU General Public License for more details. | |
16 | ||
17 | You should have received a copy of the GNU General Public License | |
18 | along with GNU Emacs. If not, see <http://www.gnu.org/licenses/>. */ | |
19 | ||
20 | #include <config.h> | |
c2d7786e | 21 | #include "lisp.h" |
704d3f45 | 22 | #include "syssignal.h" |
d89460ed PE |
23 | #include "systime.h" |
24 | ||
25 | /* Return A + B, but return the maximum fixnum if the result would overflow. | |
26 | Assume A and B are nonnegative and in fixnum range. */ | |
27 | ||
28 | static EMACS_INT | |
29 | saturated_add (EMACS_INT a, EMACS_INT b) | |
30 | { | |
31 | return min (a + b, MOST_POSITIVE_FIXNUM); | |
32 | } | |
c2d7786e | 33 | |
3d80c99f | 34 | /* Logs. */ |
c2d7786e | 35 | |
3d80c99f | 36 | typedef struct Lisp_Hash_Table log_t; |
c2d7786e TM |
37 | |
38 | static Lisp_Object | |
3d80c99f SM |
39 | make_log (int heap_size, int max_stack_depth) |
40 | { | |
41 | /* We use a standard Elisp hash-table object, but we use it in | |
42 | a special way. This is OK as long as the object is not exposed | |
43 | to Elisp, i.e. until it is returned by *-profiler-log, after which | |
44 | it can't be used any more. */ | |
45 | Lisp_Object log = make_hash_table (Qequal, make_number (heap_size), | |
46 | make_float (DEFAULT_REHASH_SIZE), | |
47 | make_float (DEFAULT_REHASH_THRESHOLD), | |
48 | Qnil, Qnil, Qnil); | |
49 | struct Lisp_Hash_Table *h = XHASH_TABLE (log); | |
50 | ||
51 | /* What is special about our hash-tables is that the keys are pre-filled | |
52 | with the vectors we'll put in them. */ | |
53 | int i = ASIZE (h->key_and_value) / 2; | |
54 | while (0 < i) | |
55 | set_hash_key_slot (h, --i, | |
56 | Fmake_vector (make_number (max_stack_depth), Qnil)); | |
57 | return log; | |
c2d7786e TM |
58 | } |
59 | ||
3d80c99f | 60 | /* Evict the least used half of the hash_table. |
c2d7786e | 61 | |
3d80c99f SM |
62 | When the table is full, we have to evict someone. |
63 | The easiest and most efficient is to evict the value we're about to add | |
64 | (i.e. once the table is full, stop sampling). | |
c2d7786e | 65 | |
3d80c99f SM |
66 | We could also pick the element with the lowest count and evict it, |
67 | but finding it is O(N) and for that amount of work we get very | |
68 | little in return: for the next sample, this latest sample will have | |
69 | count==1 and will hence be a prime candidate for eviction :-( | |
c2d7786e | 70 | |
3d80c99f SM |
71 | So instead, we take O(N) time to eliminate more or less half of the |
72 | entries (the half with the lowest counts). So we get an amortized | |
73 | cost of O(1) and we get O(N) time for a new entry to grow larger | |
74 | than the other least counts before a new round of eviction. */ | |
c2d7786e | 75 | |
3d80c99f SM |
76 | static EMACS_INT approximate_median (log_t *log, |
77 | ptrdiff_t start, ptrdiff_t size) | |
c2d7786e | 78 | { |
3d80c99f SM |
79 | eassert (size > 0); |
80 | if (size < 2) | |
81 | return XINT (HASH_VALUE (log, start)); | |
82 | if (size < 3) | |
83 | /* Not an actual median, but better for our application than | |
84 | choosing either of the two numbers. */ | |
85 | return ((XINT (HASH_VALUE (log, start)) | |
86 | + XINT (HASH_VALUE (log, start + 1))) | |
87 | / 2); | |
c2d7786e | 88 | else |
c2d7786e | 89 | { |
3d80c99f SM |
90 | ptrdiff_t newsize = size / 3; |
91 | ptrdiff_t start2 = start + newsize; | |
92 | EMACS_INT i1 = approximate_median (log, start, newsize); | |
93 | EMACS_INT i2 = approximate_median (log, start2, newsize); | |
94 | EMACS_INT i3 = approximate_median (log, start2 + newsize, | |
95 | size - 2 * newsize); | |
96 | return (i1 < i2 | |
97 | ? (i2 < i3 ? i2 : (i1 < i3 ? i3 : i1)) | |
98 | : (i1 < i3 ? i1 : (i2 < i3 ? i3 : i2))); | |
c2d7786e | 99 | } |
c2d7786e TM |
100 | } |
101 | ||
3d80c99f | 102 | static void evict_lower_half (log_t *log) |
c2d7786e | 103 | { |
3d80c99f SM |
104 | ptrdiff_t size = ASIZE (log->key_and_value) / 2; |
105 | EMACS_INT median = approximate_median (log, 0, size); | |
106 | ptrdiff_t i; | |
c2d7786e | 107 | |
c2d7786e | 108 | for (i = 0; i < size; i++) |
3d80c99f SM |
109 | /* Evict not only values smaller but also values equal to the median, |
110 | so as to make sure we evict something no matter what. */ | |
111 | if (XINT (HASH_VALUE (log, i)) <= median) | |
112 | { | |
113 | Lisp_Object key = HASH_KEY (log, i); | |
114 | { /* FIXME: we could make this more efficient. */ | |
115 | Lisp_Object tmp; | |
116 | XSET_HASH_TABLE (tmp, log); /* FIXME: Use make_lisp_ptr. */ | |
117 | Fremhash (key, tmp); | |
118 | } | |
119 | eassert (EQ (log->next_free, make_number (i))); | |
120 | { | |
121 | int j; | |
122 | eassert (VECTORP (key)); | |
123 | for (j = 0; j < ASIZE (key); j++) | |
ad942b63 | 124 | ASET (key, j, Qnil); |
3d80c99f SM |
125 | } |
126 | set_hash_key_slot (log, i, key); | |
127 | } | |
c2d7786e TM |
128 | } |
129 | ||
d89460ed | 130 | /* Record the current backtrace in LOG. COUNT is the weight of this |
b3ecad33 PE |
131 | current backtrace: interrupt counts for CPU, and the allocation |
132 | size for memory. */ | |
0efc778b | 133 | |
c2d7786e | 134 | static void |
d89460ed | 135 | record_backtrace (log_t *log, EMACS_INT count) |
c2d7786e | 136 | { |
c2d7786e | 137 | struct backtrace *backlist = backtrace_list; |
3d80c99f SM |
138 | Lisp_Object backtrace; |
139 | ptrdiff_t index, i = 0; | |
140 | ptrdiff_t asize; | |
c2d7786e | 141 | |
3d80c99f | 142 | if (!INTEGERP (log->next_free)) |
611b7507 JB |
143 | /* FIXME: transfer the evicted counts to a special entry rather |
144 | than dropping them on the floor. */ | |
3d80c99f SM |
145 | evict_lower_half (log); |
146 | index = XINT (log->next_free); | |
c2d7786e | 147 | |
3d80c99f SM |
148 | /* Get a "working memory" vector. */ |
149 | backtrace = HASH_KEY (log, index); | |
150 | asize = ASIZE (backtrace); | |
12b3895d | 151 | |
0efc778b | 152 | /* Copy the backtrace contents into working memory. */ |
3d80c99f | 153 | for (; i < asize && backlist; i++, backlist = backlist->next) |
611b7507 | 154 | /* FIXME: For closures we should ignore the environment. */ |
e7c1b6ef | 155 | ASET (backtrace, i, backlist->function); |
0efc778b | 156 | |
3d80c99f SM |
157 | /* Make sure that unused space of working memory is filled with nil. */ |
158 | for (; i < asize; i++) | |
159 | ASET (backtrace, i, Qnil); | |
c2d7786e | 160 | |
3d80c99f SM |
161 | { /* We basically do a `gethash+puthash' here, except that we have to be |
162 | careful to avoid memory allocation since we're in a signal | |
163 | handler, and we optimize the code to try and avoid computing the | |
164 | hash+lookup twice. See fns.c:Fputhash for reference. */ | |
165 | EMACS_UINT hash; | |
166 | ptrdiff_t j = hash_lookup (log, backtrace, &hash); | |
167 | if (j >= 0) | |
d89460ed PE |
168 | { |
169 | EMACS_INT old_val = XINT (HASH_VALUE (log, j)); | |
170 | EMACS_INT new_val = saturated_add (old_val, count); | |
171 | set_hash_value_slot (log, j, make_number (new_val)); | |
172 | } | |
3d80c99f SM |
173 | else |
174 | { /* BEWARE! hash_put in general can allocate memory. | |
175 | But currently it only does that if log->next_free is nil. */ | |
176 | int j; | |
177 | eassert (!NILP (log->next_free)); | |
178 | j = hash_put (log, backtrace, make_number (count), hash); | |
179 | /* Let's make sure we've put `backtrace' right where it | |
180 | already was to start with. */ | |
181 | eassert (index == j); | |
182 | ||
183 | /* FIXME: If the hash-table is almost full, we should set | |
184 | some global flag so that some Elisp code can offload its | |
611b7507 JB |
185 | data elsewhere, so as to avoid the eviction code. |
186 | There are 2 ways to do that, AFAICT: | |
187 | - Set a flag checked in QUIT, such that QUIT can then call | |
188 | Fprofiler_cpu_log and stash the full log for later use. | |
189 | - Set a flag check in post-gc-hook, so that Elisp code can call | |
190 | profiler-cpu-log. That gives us more flexibility since that | |
191 | Elisp code can then do all kinds of fun stuff like write | |
192 | the log to disk. Or turn it right away into a call tree. | |
193 | Of course, using Elisp is generally preferable, but it may | |
194 | take longer until we get a chance to run the Elisp code, so | |
195 | there's more risk that the table will get full before we | |
196 | get there. */ | |
3d80c99f SM |
197 | } |
198 | } | |
c2d7786e | 199 | } |
c2d7786e | 200 | \f |
c22bac2c | 201 | /* Sampling profiler. */ |
c2d7786e | 202 | |
d89460ed PE |
203 | #ifdef PROFILER_CPU_SUPPORT |
204 | ||
205 | /* The profiler timer and whether it was properly initialized, if | |
206 | POSIX timers are available. */ | |
2b794d69 | 207 | #ifdef HAVE_ITIMERSPEC |
d89460ed PE |
208 | static timer_t profiler_timer; |
209 | static bool profiler_timer_ok; | |
210 | #endif | |
ad942b63 | 211 | |
d89460ed PE |
212 | /* Status of sampling profiler. */ |
213 | static enum profiler_cpu_running | |
214 | { NOT_RUNNING, TIMER_SETTIME_RUNNING, SETITIMER_RUNNING } | |
215 | profiler_cpu_running; | |
6521894d | 216 | |
d89460ed | 217 | /* Hash-table log of CPU profiler. */ |
3d80c99f | 218 | static Lisp_Object cpu_log; |
d89460ed | 219 | |
3d80c99f SM |
220 | /* Separate counter for the time spent in the GC. */ |
221 | static EMACS_INT cpu_gc_count; | |
0efc778b | 222 | |
b3ecad33 | 223 | /* The current sampling interval in nanoseconds. */ |
c22bac2c | 224 | static EMACS_INT current_sampling_interval; |
c2d7786e | 225 | |
c22bac2c | 226 | /* Signal handler for sampling profiler. */ |
6521894d SM |
227 | |
228 | static void | |
d89460ed | 229 | handle_profiler_signal (int signal) |
6521894d | 230 | { |
e7c1b6ef | 231 | if (backtrace_list && EQ (backtrace_list->function, Qautomatic_gc)) |
6521894d SM |
232 | /* Special case the time-count inside GC because the hash-table |
233 | code is not prepared to be used while the GC is running. | |
234 | More specifically it uses ASIZE at many places where it does | |
235 | not expect the ARRAY_MARK_FLAG to be set. We could try and | |
236 | harden the hash-table code, but it doesn't seem worth the | |
237 | effort. */ | |
b3ecad33 | 238 | cpu_gc_count = saturated_add (cpu_gc_count, 1); |
6521894d | 239 | else |
d89460ed | 240 | { |
a3c5c0c5 | 241 | Lisp_Object oquit; |
b3ecad33 | 242 | EMACS_INT count = 1; |
2b794d69 | 243 | #ifdef HAVE_ITIMERSPEC |
b3ecad33 PE |
244 | if (profiler_timer_ok) |
245 | { | |
246 | int overruns = timer_getoverrun (profiler_timer); | |
247 | eassert (0 <= overruns); | |
248 | count += overruns; | |
249 | } | |
250 | #endif | |
a3c5c0c5 PE |
251 | /* record_backtrace uses hash functions that call Fequal, which |
252 | uses QUIT, which can call malloc, which can cause disaster in | |
253 | a signal handler. So inhibit QUIT. */ | |
254 | oquit = Vinhibit_quit; | |
255 | Vinhibit_quit = Qt; | |
256 | ||
d89460ed | 257 | eassert (HASH_TABLE_P (cpu_log)); |
b3ecad33 | 258 | record_backtrace (XHASH_TABLE (cpu_log), count); |
a3c5c0c5 PE |
259 | |
260 | Vinhibit_quit = oquit; | |
d89460ed | 261 | } |
6521894d SM |
262 | } |
263 | ||
704d3f45 | 264 | static void |
d89460ed PE |
265 | deliver_profiler_signal (int signal) |
266 | { | |
267 | deliver_process_signal (signal, handle_profiler_signal); | |
268 | } | |
269 | ||
270 | static enum profiler_cpu_running | |
c22bac2c | 271 | setup_cpu_timer (Lisp_Object sampling_interval) |
704d3f45 | 272 | { |
d89460ed PE |
273 | struct sigaction action; |
274 | struct itimerval timer; | |
275 | struct timespec interval; | |
b3ecad33 | 276 | int billion = 1000000000; |
d89460ed | 277 | |
c22bac2c | 278 | if (! RANGED_INTEGERP (1, sampling_interval, |
b3ecad33 PE |
279 | (TYPE_MAXIMUM (time_t) < EMACS_INT_MAX / billion |
280 | ? ((EMACS_INT) TYPE_MAXIMUM (time_t) * billion | |
281 | + (billion - 1)) | |
d89460ed PE |
282 | : EMACS_INT_MAX))) |
283 | return NOT_RUNNING; | |
284 | ||
c22bac2c | 285 | current_sampling_interval = XINT (sampling_interval); |
b3ecad33 PE |
286 | interval = make_emacs_time (current_sampling_interval / billion, |
287 | current_sampling_interval % billion); | |
d89460ed PE |
288 | emacs_sigaction_init (&action, deliver_profiler_signal); |
289 | sigaction (SIGPROF, &action, 0); | |
290 | ||
2b794d69 | 291 | #ifdef HAVE_ITIMERSPEC |
d89460ed PE |
292 | if (! profiler_timer_ok) |
293 | { | |
294 | /* System clocks to try, in decreasing order of desirability. */ | |
295 | static clockid_t const system_clock[] = { | |
296 | #ifdef CLOCK_THREAD_CPUTIME_ID | |
297 | CLOCK_THREAD_CPUTIME_ID, | |
298 | #endif | |
299 | #ifdef CLOCK_PROCESS_CPUTIME_ID | |
300 | CLOCK_PROCESS_CPUTIME_ID, | |
301 | #endif | |
302 | #ifdef CLOCK_MONOTONIC | |
303 | CLOCK_MONOTONIC, | |
304 | #endif | |
305 | CLOCK_REALTIME | |
306 | }; | |
307 | int i; | |
308 | struct sigevent sigev; | |
309 | sigev.sigev_value.sival_ptr = &profiler_timer; | |
310 | sigev.sigev_signo = SIGPROF; | |
311 | sigev.sigev_notify = SIGEV_SIGNAL; | |
312 | ||
313 | for (i = 0; i < sizeof system_clock / sizeof *system_clock; i++) | |
314 | if (timer_create (system_clock[i], &sigev, &profiler_timer) == 0) | |
315 | { | |
316 | profiler_timer_ok = 1; | |
317 | break; | |
318 | } | |
319 | } | |
320 | ||
321 | if (profiler_timer_ok) | |
322 | { | |
323 | struct itimerspec ispec; | |
324 | ispec.it_value = ispec.it_interval = interval; | |
2b794d69 PE |
325 | if (timer_settime (profiler_timer, 0, &ispec, 0) == 0) |
326 | return TIMER_SETTIME_RUNNING; | |
d89460ed PE |
327 | } |
328 | #endif | |
329 | ||
2b794d69 | 330 | #ifdef HAVE_SETITIMER |
d89460ed | 331 | timer.it_value = timer.it_interval = make_timeval (interval); |
2b794d69 PE |
332 | if (setitimer (ITIMER_PROF, &timer, 0) == 0) |
333 | return SETITIMER_RUNNING; | |
334 | #endif | |
335 | ||
336 | return NOT_RUNNING; | |
704d3f45 TM |
337 | } |
338 | ||
6521894d | 339 | DEFUN ("profiler-cpu-start", Fprofiler_cpu_start, Sprofiler_cpu_start, |
c2d7786e | 340 | 1, 1, 0, |
6521894d | 341 | doc: /* Start or restart the cpu profiler. |
b3ecad33 | 342 | It takes call-stack samples each SAMPLING-INTERVAL nanoseconds, approximately. |
6521894d | 343 | See also `profiler-log-size' and `profiler-max-stack-depth'. */) |
c22bac2c | 344 | (Lisp_Object sampling_interval) |
c2d7786e | 345 | { |
6521894d | 346 | if (profiler_cpu_running) |
c22bac2c | 347 | error ("CPU profiler is already running"); |
c2d7786e | 348 | |
3d80c99f SM |
349 | if (NILP (cpu_log)) |
350 | { | |
351 | cpu_gc_count = 0; | |
6521894d | 352 | cpu_log = make_log (profiler_log_size, |
3d80c99f SM |
353 | profiler_max_stack_depth); |
354 | } | |
c2d7786e | 355 | |
c22bac2c | 356 | profiler_cpu_running = setup_cpu_timer (sampling_interval); |
d89460ed | 357 | if (! profiler_cpu_running) |
c22bac2c | 358 | error ("Invalid sampling interval"); |
c2d7786e TM |
359 | |
360 | return Qt; | |
361 | } | |
362 | ||
6521894d | 363 | DEFUN ("profiler-cpu-stop", Fprofiler_cpu_stop, Sprofiler_cpu_stop, |
c2d7786e | 364 | 0, 0, 0, |
234148bf SM |
365 | doc: /* Stop the cpu profiler. The profiler log is not affected. |
366 | Return non-nil if the profiler was running. */) | |
c2d7786e TM |
367 | (void) |
368 | { | |
d89460ed PE |
369 | switch (profiler_cpu_running) |
370 | { | |
371 | case NOT_RUNNING: | |
372 | return Qnil; | |
373 | ||
2b794d69 | 374 | #ifdef HAVE_ITIMERSPEC |
d89460ed PE |
375 | case TIMER_SETTIME_RUNNING: |
376 | { | |
377 | struct itimerspec disable; | |
378 | memset (&disable, 0, sizeof disable); | |
379 | timer_settime (profiler_timer, 0, &disable, 0); | |
380 | } | |
381 | break; | |
84f72efd | 382 | #endif |
c2d7786e | 383 | |
2b794d69 | 384 | #ifdef HAVE_SETITIMER |
d89460ed PE |
385 | case SETITIMER_RUNNING: |
386 | { | |
387 | struct itimerval disable; | |
388 | memset (&disable, 0, sizeof disable); | |
389 | setitimer (ITIMER_PROF, &disable, 0); | |
390 | } | |
391 | break; | |
2b794d69 | 392 | #endif |
d89460ed | 393 | } |
c2d7786e | 394 | |
d89460ed PE |
395 | signal (SIGPROF, SIG_IGN); |
396 | profiler_cpu_running = NOT_RUNNING; | |
c2d7786e TM |
397 | return Qt; |
398 | } | |
399 | ||
6521894d SM |
400 | DEFUN ("profiler-cpu-running-p", |
401 | Fprofiler_cpu_running_p, Sprofiler_cpu_running_p, | |
c2d7786e | 402 | 0, 0, 0, |
6521894d | 403 | doc: /* Return non-nil iff cpu profiler is running. */) |
c2d7786e TM |
404 | (void) |
405 | { | |
6521894d | 406 | return profiler_cpu_running ? Qt : Qnil; |
c2d7786e TM |
407 | } |
408 | ||
6521894d | 409 | DEFUN ("profiler-cpu-log", Fprofiler_cpu_log, Sprofiler_cpu_log, |
c2d7786e | 410 | 0, 0, 0, |
6521894d SM |
411 | doc: /* Return the current cpu profiler log. |
412 | The log is a hash-table mapping backtraces to counters which represent | |
413 | the amount of time spent at those points. Every backtrace is a vector | |
414 | of functions, where the last few elements may be nil. | |
415 | Before returning, a new log is allocated for future samples. */) | |
c2d7786e TM |
416 | (void) |
417 | { | |
3d80c99f | 418 | Lisp_Object result = cpu_log; |
d89460ed | 419 | /* Here we're making the log visible to Elisp, so it's not safe any |
3d80c99f SM |
420 | more for our use afterwards since we can't rely on its special |
421 | pre-allocated keys anymore. So we have to allocate a new one. */ | |
6521894d SM |
422 | cpu_log = (profiler_cpu_running |
423 | ? make_log (profiler_log_size, profiler_max_stack_depth) | |
3d80c99f SM |
424 | : Qnil); |
425 | Fputhash (Fmake_vector (make_number (1), Qautomatic_gc), | |
426 | make_number (cpu_gc_count), | |
427 | result); | |
428 | cpu_gc_count = 0; | |
c2d7786e TM |
429 | return result; |
430 | } | |
d89460ed | 431 | #endif /* PROFILER_CPU_SUPPORT */ |
c2d7786e | 432 | \f |
0efc778b | 433 | /* Memory profiler. */ |
c2d7786e | 434 | |
6521894d SM |
435 | /* True if memory profiler is running. */ |
436 | bool profiler_memory_running; | |
437 | ||
3d80c99f | 438 | static Lisp_Object memory_log; |
c2d7786e | 439 | |
6521894d | 440 | DEFUN ("profiler-memory-start", Fprofiler_memory_start, Sprofiler_memory_start, |
c2d7786e | 441 | 0, 0, 0, |
6521894d SM |
442 | doc: /* Start/restart the memory profiler. |
443 | The memory profiler will take samples of the call-stack whenever a new | |
444 | allocation takes place. Note that most small allocations only trigger | |
445 | the profiler occasionally. | |
446 | See also `profiler-log-size' and `profiler-max-stack-depth'. */) | |
c2d7786e TM |
447 | (void) |
448 | { | |
6521894d | 449 | if (profiler_memory_running) |
c2d7786e TM |
450 | error ("Memory profiler is already running"); |
451 | ||
3d80c99f | 452 | if (NILP (memory_log)) |
6521894d | 453 | memory_log = make_log (profiler_log_size, |
c2d7786e TM |
454 | profiler_max_stack_depth); |
455 | ||
234148bf | 456 | profiler_memory_running = true; |
c2d7786e TM |
457 | |
458 | return Qt; | |
459 | } | |
460 | ||
6521894d SM |
461 | DEFUN ("profiler-memory-stop", |
462 | Fprofiler_memory_stop, Sprofiler_memory_stop, | |
c2d7786e | 463 | 0, 0, 0, |
234148bf SM |
464 | doc: /* Stop the memory profiler. The profiler log is not affected. |
465 | Return non-nil if the profiler was running. */) | |
c2d7786e TM |
466 | (void) |
467 | { | |
6521894d | 468 | if (!profiler_memory_running) |
234148bf SM |
469 | return Qnil; |
470 | profiler_memory_running = false; | |
c2d7786e TM |
471 | return Qt; |
472 | } | |
473 | ||
6521894d SM |
474 | DEFUN ("profiler-memory-running-p", |
475 | Fprofiler_memory_running_p, Sprofiler_memory_running_p, | |
c2d7786e | 476 | 0, 0, 0, |
6521894d | 477 | doc: /* Return non-nil if memory profiler is running. */) |
c2d7786e TM |
478 | (void) |
479 | { | |
6521894d | 480 | return profiler_memory_running ? Qt : Qnil; |
c2d7786e TM |
481 | } |
482 | ||
6521894d SM |
483 | DEFUN ("profiler-memory-log", |
484 | Fprofiler_memory_log, Sprofiler_memory_log, | |
c2d7786e | 485 | 0, 0, 0, |
6521894d SM |
486 | doc: /* Return the current memory profiler log. |
487 | The log is a hash-table mapping backtraces to counters which represent | |
488 | the amount of memory allocated at those points. Every backtrace is a vector | |
489 | of functions, where the last few elements may be nil. | |
490 | Before returning, a new log is allocated for future samples. */) | |
c2d7786e TM |
491 | (void) |
492 | { | |
3d80c99f SM |
493 | Lisp_Object result = memory_log; |
494 | /* Here we're making the log visible to Elisp , so it's not safe any | |
495 | more for our use afterwards since we can't rely on its special | |
496 | pre-allocated keys anymore. So we have to allocate a new one. */ | |
6521894d SM |
497 | memory_log = (profiler_memory_running |
498 | ? make_log (profiler_log_size, profiler_max_stack_depth) | |
3d80c99f | 499 | : Qnil); |
c2d7786e TM |
500 | return result; |
501 | } | |
502 | ||
503 | \f | |
0efc778b | 504 | /* Signals and probes. */ |
c2d7786e | 505 | |
0efc778b | 506 | /* Record that the current backtrace allocated SIZE bytes. */ |
c2d7786e TM |
507 | void |
508 | malloc_probe (size_t size) | |
509 | { | |
ad942b63 | 510 | eassert (HASH_TABLE_P (memory_log)); |
d89460ed | 511 | record_backtrace (XHASH_TABLE (memory_log), min (size, MOST_POSITIVE_FIXNUM)); |
c2d7786e TM |
512 | } |
513 | ||
514 | void | |
515 | syms_of_profiler (void) | |
516 | { | |
517 | DEFVAR_INT ("profiler-max-stack-depth", profiler_max_stack_depth, | |
6521894d | 518 | doc: /* Number of elements from the call-stack recorded in the log. */); |
c2d7786e | 519 | profiler_max_stack_depth = 16; |
6521894d SM |
520 | DEFVAR_INT ("profiler-log-size", profiler_log_size, |
521 | doc: /* Number of distinct call-stacks that can be recorded in a profiler log. | |
522 | If the log gets full, some of the least-seen call-stacks will be evicted | |
523 | to make room for new entries. */); | |
524 | profiler_log_size = 10000; | |
c2d7786e | 525 | |
ad942b63 | 526 | #ifdef PROFILER_CPU_SUPPORT |
d89460ed | 527 | profiler_cpu_running = NOT_RUNNING; |
ad942b63 SM |
528 | cpu_log = Qnil; |
529 | staticpro (&cpu_log); | |
6521894d SM |
530 | defsubr (&Sprofiler_cpu_start); |
531 | defsubr (&Sprofiler_cpu_stop); | |
532 | defsubr (&Sprofiler_cpu_running_p); | |
533 | defsubr (&Sprofiler_cpu_log); | |
ad942b63 | 534 | #endif |
234148bf | 535 | profiler_memory_running = false; |
ad942b63 SM |
536 | memory_log = Qnil; |
537 | staticpro (&memory_log); | |
6521894d SM |
538 | defsubr (&Sprofiler_memory_start); |
539 | defsubr (&Sprofiler_memory_stop); | |
540 | defsubr (&Sprofiler_memory_running_p); | |
541 | defsubr (&Sprofiler_memory_log); | |
c2d7786e | 542 | } |