Remove the opt.lg_prof_tcmax option.

Remove the opt.lg_prof_tcmax option and hard-code a cache size of 1024.
This setting is something that users just shouldn't have to worry about.
If lock contention actually ends up being a problem, the simple solution
available to the user is to reduce sampling frequency.
This commit is contained in:
Jason Evans
2012-02-13 18:04:26 -08:00
parent e7a1058aaa
commit 0b526ff94d
6 changed files with 8 additions and 50 deletions

View File

@@ -938,8 +938,6 @@ malloc_conf = "xmalloc:true";]]></programlisting>
option for probabilistic sampling control. See the <link
linkend="opt.prof_accum"><mallctl>opt.prof_accum</mallctl></link>
option for control of cumulative sample reporting. See the <link
linkend="opt.lg_prof_tcmax"><mallctl>opt.lg_prof_tcmax</mallctl></link>
option for control of per thread backtrace caching. See the <link
linkend="opt.lg_prof_interval"><mallctl>opt.lg_prof_interval</mallctl></link>
option for information on interval-triggered profile dumping, and the
<link linkend="opt.prof_gdump"><mallctl>opt.prof_gdump</mallctl></link>
@@ -1017,28 +1015,8 @@ malloc_conf = "xmalloc:true";]]></programlisting>
dumps enabled/disabled. If this option is enabled, every unique
backtrace must be stored for the duration of execution. Depending on
the application, this can impose a large memory overhead, and the
cumulative counts are not always of interest. See the
<link
linkend="opt.lg_prof_tcmax"><mallctl>opt.lg_prof_tcmax</mallctl></link>
option for control of per thread backtrace caching, which has important
interactions. This option is enabled by default.</para></listitem>
</varlistentry>
<varlistentry id="opt.lg_prof_tcmax">
<term>
<mallctl>opt.lg_prof_tcmax</mallctl>
(<type>ssize_t</type>)
<literal>r-</literal>
[<option>--enable-prof</option>]
</term>
<listitem><para>Maximum per thread backtrace cache (log base 2) used
for heap profiling. A backtrace can only be discarded if the
<link linkend="opt.prof_accum"><mallctl>opt.prof_accum</mallctl></link>
option is disabled, and no thread caches currently refer to the
backtrace. Therefore, a backtrace cache limit should be imposed if the
intention is to limit how much memory is used by backtraces. By
default, no limit is imposed (encoded as -1).
</para></listitem>
cumulative counts are not always of interest. This option is enabled
by default.</para></listitem>
</varlistentry>
<varlistentry id="opt.lg_prof_interval">