Whamcloud - gitweb
LUDOC-504 nodemap: servers must be in a trusted+admin group
[doc/manual.git] / LustreTuning.xml
index 5b89770..9460749 100644 (file)
@@ -1,7 +1,7 @@
 <?xml version='1.0' encoding='utf-8'?>
 <chapter xmlns="http://docbook.org/ns/docbook"
-xmlns:xl="http://www.w3.org/1999/xlink" version="5.0" xml:lang="en-US"
-xml:id="lustretuning">
+ xmlns:xl="http://www.w3.org/1999/xlink" version="5.0" xml:lang="en-US"
+ xml:id="lustretuning">
   <title xml:id="lustretuning.title">Tuning a Lustre File System</title>
   <para>This chapter contains information about tuning a Lustre file system for
   better performance.</para>
@@ -10,7 +10,7 @@ xml:id="lustretuning">
     parameters. These parameters are contained in the 
     <literal>/etc/modprobe.d/lustre.conf</literal> file.</para>
   </note>
-  <section xml:id="dbdoclet.50438272_55226">
+  <section xml:id="tuning_optimizing_service_threads">
     <title>
     <indexterm>
       <primary>tuning</primary>
@@ -97,7 +97,7 @@ lctl {get,set}_param {service}.thread_{min,max,started}
       <para>
       This works in a similar fashion to 
       binding of threads on MDS. MDS thread tuning is covered in 
-      <xref linkend="dbdoclet.mdsbinding" />.</para>
+      <xref linkend="mdsbinding" />.</para>
       <itemizedlist>
         <listitem>
           <para>
@@ -113,9 +113,9 @@ lctl {get,set}_param {service}.thread_{min,max,started}
         </listitem>
       </itemizedlist>
       <para>For further details, see 
-      <xref linkend="dbdoclet.50438271_87260" />.</para>
+      <xref linkend="tuning_setting_thread_count" />.</para>
     </section>
-    <section xml:id="dbdoclet.mdstuning">
+    <section xml:id="mdstuning">
       <title>
       <indexterm>
         <primary>tuning</primary>
@@ -136,7 +136,7 @@ lctl {get,set}_param {service}.thread_{min,max,started}
 </screen>
       </para>
       <para>For details, see 
-      <xref linkend="dbdoclet.50438271_87260" />.</para>
+      <xref linkend="tuning_setting_thread_count" />.</para>
       <para>The number of MDS service threads started depends on system size
       and the load on the server, and has a default maximum of 64. The
       maximum potential number of threads (<literal>MDS_MAX_THREADS</literal>)
@@ -149,8 +149,8 @@ lctl {get,set}_param {service}.thread_{min,max,started}
        service immediately and disables automatic thread creation behavior.
        </para>
       </note>
-      <para condition='l23'>Lustre software release 2.3 introduced new
-      parameters to provide more control to administrators.</para>
+      <para>Parameters are available to provide administrators control
+        over the number of service threads.</para>
       <itemizedlist>
         <listitem>
           <para>
@@ -158,29 +158,22 @@ lctl {get,set}_param {service}.thread_{min,max,started}
           in providing the read page service. The read page service handles
           file close and readdir operations.</para>
         </listitem>
-        <listitem>
-          <para>
-          <literal>mds_attr_num_threads</literal> controls the number of threads
-          in providing the setattr service to clients running Lustre software
-          release 1.8.</para>
-        </listitem>
       </itemizedlist>
     </section>
   </section>
-  <section xml:id="dbdoclet.mdsbinding" condition='l23'>
+  <section xml:id="mdsbinding">
     <title>
     <indexterm>
       <primary>tuning</primary>
       <secondary>MDS binding</secondary>
     </indexterm>Binding MDS Service Thread to CPU Partitions</title>
-    <para>With the introduction of Node Affinity (
-    <xref linkend="nodeaffdef" />) in Lustre software release 2.3, MDS threads
-    can be bound to particular CPU partitions (CPTs) to improve CPU cache
-    usage and memory locality.  Default values for CPT counts and CPU core
+    <para>With the Node Affinity (<xref linkend="nodeaffdef" />) feature,
+    MDS threads can be bound to particular CPU partitions (CPTs) to improve CPU
+    cache usage and memory locality.  Default values for CPT counts and CPU core
     bindings are selected automatically to provide good overall performance for
     a given CPU count. However, an administrator can deviate from these setting
     if they choose.  For details on specifying the mapping of CPU cores to
-    CPTs see <xref linkend="dbdoclet.libcfstuning"/>.
+    CPTs see <xref linkend="libcfstuning"/>.
     </para>
     <itemizedlist>
       <listitem>
@@ -202,14 +195,8 @@ lctl {get,set}_param {service}.thread_{min,max,started}
         to 
         <literal>CPT4</literal>.</para>
       </listitem>
-      <listitem>
-        <para>
-        <literal>mds_attr_num_cpts=[EXPRESSION]</literal> binds the setattr
-        service threads to CPTs defined by 
-        <literal>EXPRESSION</literal>.</para>
-      </listitem>
     </itemizedlist>
-       <para>Parameters must be set before module load in the file 
+    <para>Parameters must be set before module load in the file 
     <literal>/etc/modprobe.d/lustre.conf</literal>. For example:
     <example><title>lustre.conf</title>
     <screen>options lnet networks=tcp0(eth0)
@@ -217,7 +204,7 @@ options mdt mds_num_cpts=[0]</screen>
     </example>
     </para>
   </section>
-  <section xml:id="dbdoclet.50438272_73839">
+  <section xml:id="tuning_lnet_params">
     <title>
     <indexterm>
       <primary>LNet</primary>
@@ -268,18 +255,17 @@ options ksocklnd enable_irq_affinity=0
       <para>By default, this parameter is off. As always, you should test the
       performance to compare the impact of changing this parameter.</para>
     </section>
-    <section condition='l23'>
+    <section>
       <title>
       <indexterm>
         <primary>tuning</primary>
         <secondary>Network interface binding</secondary>
       </indexterm>Binding Network Interface Against CPU Partitions</title>
-      <para>Lustre software release 2.3 and beyond provide enhanced network
-      interface control. The enhancement means that an administrator can bind
-      an interface to one or more CPU partitions. Bindings are specified as
-      options to the LNet modules. For more information on specifying module
-      options, see 
-      <xref linkend="dbdoclet.50438293_15350" /></para>
+      <para>Lustre allows enhanced network interface control. This means that
+      an administrator can bind an interface to one or more CPU partitions.
+      Bindings are specified as options to the LNet modules. For more
+      information on specifying module options, see 
+      <xref linkend="tuning_lnet_mod_params" /></para>
       <para>For example, 
       <literal>o2ib0(ib0)[0,1]</literal> will ensure that all messages for 
       <literal>o2ib0</literal> will be handled by LND threads executing on 
@@ -324,9 +310,9 @@ ksocklnd credits=256
       <screen>
 ko2iblnd credits=256
 </screen>
-      <note condition="l23">
-        <para>In Lustre software release 2.3 and beyond, LNet may revalidate
-        the NI credits, so the administrator's request may not persist.</para>
+      <note>
+        <para>LNet may revalidate the NI credits, so the administrator's
+       request may not persist.</para>
       </note>
     </section>
     <section>
@@ -375,10 +361,9 @@ ko2iblnd credits=256
       <screen>
 lnet large_router_buffers=8192
 </screen>
-      <note condition="l23">
-        <para>In Lustre software release 2.3 and beyond, LNet may revalidate
-        the router buffer setting, so the administrator's request may not
-        persist.</para>
+      <note>
+        <para>LNet may revalidate the router buffer setting, so the
+       administrator's request may not persist.</para>
       </note>
     </section>
     <section>
@@ -399,9 +384,8 @@ lnet large_router_buffers=8192
       events across all CPTs. This may balance load better across the CPU but
       can incur a cross CPU overhead.</para>
       <para>The current policy can be changed by an administrator with 
-      <literal>echo 
-      <replaceable>value</replaceable>&gt;
-      /proc/sys/lnet/portal_rotor</literal>. There are four options for 
+      <literal>lctl set_param portal_rotor=value</literal>. 
+      There are four options for 
       <literal>
         <replaceable>value</replaceable>
       </literal>:</para>
@@ -480,7 +464,7 @@ lnet large_router_buffers=8192
           interface. The default setting is 1. (For more information about the
           LNet routes parameter, see 
           <xref xmlns:xlink="http://www.w3.org/1999/xlink"
-          linkend="dbdoclet.50438216_71227" /></para>
+          linkend="lnet_module_routes" /></para>
           <para>A router is considered down if any of its NIDs are down. For
           example, router X has three NIDs: 
           <literal>Xnid1</literal>, 
@@ -525,16 +509,16 @@ lnet large_router_buffers=8192
       be MAX.</para>
     </section>
   </section>
-  <section xml:id="dbdoclet.libcfstuning" condition='l23'>
+  <section xml:id="libcfstuning">
     <title>
     <indexterm>
       <primary>tuning</primary>
       <secondary>libcfs</secondary>
     </indexterm>libcfs Tuning</title>
-    <para>Lustre software release 2.3 introduced binding service threads via
-    CPU Partition Tables (CPTs). This allows the system administrator to
-    fine-tune on which CPU cores the Lustre service threads are run, for both
-    OSS and MDS services, as well as on the client.
+    <para>Lustre allows binding service threads via CPU Partition Tables
+      (CPTs). This allows the system administrator to fine-tune on which CPU
+      cores the Lustre service threads are run, for both OSS and MDS services,
+      as well as on the client.
     </para>
     <para>CPTs are useful to reserve some cores on the OSS or MDS nodes for
     system functions such as system monitoring, HA heartbeat, or similar
@@ -606,7 +590,7 @@ cpu_partition_table=
       </para>
     </section>
   </section>
-  <section xml:id="dbdoclet.lndtuning">
+  <section xml:id="lndtuning">
     <title>
     <indexterm>
       <primary>tuning</primary>
@@ -619,12 +603,11 @@ cpu_partition_table=
     <literal>nscheds</literal> parameter. This adjusts the number of threads for
     each partition, not the overall number of threads on the LND.</para>
     <note>
-      <para>Lustre software release 2.3 has greatly decreased the default
-      number of threads for 
+      <para>The default number of threads for 
       <literal>ko2iblnd</literal> and 
-      <literal>ksocklnd</literal> on high-core count machines. The current
-      default values are automatically set and are chosen to work well across a
-      number of typical scenarios.</para>
+      <literal>ksocklnd</literal> are automatically set and are chosen to
+      work well across a number of typical scenarios, for systems with both
+      high and low core counts.</para>
     </note>
     <section>
        <title>ko2iblnd Tuning</title>
@@ -729,7 +712,7 @@ cpu_partition_table=
                </entry>
                <entry>
                  <para>Introduced in 2.10. Number of connections to each peer. Messages
-          are sent round-robin over the connection pool.  Provides signifiant
+          are sent round-robin over the connection pool.  Provides significant
           improvement with OmniPath.</para>
                </entry>
              </row>
@@ -1050,7 +1033,7 @@ cpu_partition_table=
        </informaltable>
     </section>
   </section>
-  <section xml:id="dbdoclet.nrstuning" condition='l24'>
+  <section xml:id="nrstuning">
     <title>
     <indexterm>
       <primary>tuning</primary>
@@ -1585,7 +1568,7 @@ ost.OSS.ost_io.nrs_orr_supported=reg_supported:reads_and_writes
         </listitem>
       </itemizedlist>
     </section>
-    <section xml:id="dbdoclet.tbftuning" condition='l26'>
+    <section xml:id="tbftuning" condition='l26'>
       <title>
       <indexterm>
         <primary>tuning</primary>
@@ -1599,8 +1582,8 @@ ost.OSS.ost_io.nrs_orr_supported=reg_supported:reads_and_writes
         <title>The internal structure of TBF policy</title>
         <mediaobject>
           <imageobject>
-            <imagedata scalefit="1" width="100%"
-            fileref="figures/TBF_policy.svg" />
+            <imagedata scalefit="1" width="50%"
+            fileref="figures/TBF_policy.png" />
           </imageobject>
           <textobject>
             <phrase>The internal structure of TBF policy</phrase>
@@ -1634,15 +1617,16 @@ ost.OSS.ost_io.nrs_orr_supported=reg_supported:reads_and_writes
        <screen>lctl set_param ost.OSS.ost_io.nrs_policies="tbf &lt;<replaceable>policy</replaceable>&gt;"
        </screen>
        <para>For now, the RPCs can be classified into the different types
-       according to their NID, JOBID and OPCode. (UID/GID will be supported
-       soon.) When enabling TBF policy, you can specify one of the types, or
-       just use "tbf" to enable all of them to do a fine-grained RPC requests
-       classification.</para>
+       according to their NID, JOBID, OPCode and UID/GID. When enabling TBF
+       policy, you can specify one of the types, or just use "tbf" to enable
+       all of them to do a fine-grained RPC requests classification.</para>
        <para>Example:</para>
        <screen>$ lctl set_param ost.OSS.ost_io.nrs_policies="tbf"
 $ lctl set_param ost.OSS.ost_io.nrs_policies="tbf nid"
 $ lctl set_param ost.OSS.ost_io.nrs_policies="tbf jobid"
-$ lctl set_param ost.OSS.ost_io.nrs_policies="tbf opcode"</screen>
+$ lctl set_param ost.OSS.ost_io.nrs_policies="tbf opcode"
+$ lctl set_param ost.OSS.ost_io.nrs_policies="tbf uid"
+$ lctl set_param ost.OSS.ost_io.nrs_policies="tbf gid"</screen>
       </section>
       <section remap="h4">
        <title>Start a TBF rule</title>
@@ -1652,10 +1636,13 @@ $ lctl set_param ost.OSS.ost_io.nrs_policies="tbf opcode"</screen>
        <screen>lctl set_param x.x.x.nrs_tbf_rule=
 "[reg|hp] start <replaceable>rule_name</replaceable> <replaceable>arguments</replaceable>..."
        </screen>
-       <para>'<replaceable>rule_name</replaceable>' is a string of the TBF
-       policy rule's name and '<replaceable>arguments</replaceable>' is a
-       string to specify the detailed rule according to the different types.
-       </para>
+        <para>'<replaceable>rule_name</replaceable>' is a string up to 15
+          characters which identifies the TBF policy rule's name. Alphanumeric
+          characters and underscores are accepted (e.g: "test_rule_A1").
+        </para>
+        <para>'<replaceable>arguments</replaceable>' is a string to specify the
+          detailed rule according to the different types.
+        </para>
        <itemizedlist>
        <para>Next, the different types of TBF policies will be described.</para>
          <listitem>
@@ -1703,10 +1690,10 @@ $ lctl set_param ost.OSS.ost_io.nrs_tbf_rule=\
            <para><emphasis role="bold">JobID based TBF policy</emphasis></para>
             <para>For the JobID, please see
             <xref xmlns:xlink="http://www.w3.org/1999/xlink"
-            linkend="dbdoclet.jobstats" /> for more details.</para>
+            linkend="jobstats" /> for more details.</para>
            <para>Command:</para>
             <screen>lctl set_param x.x.x.nrs_tbf_rule=
-"[reg|hp] start <replaceable>name</replaceable> jobid={<replaceable>jobid_list</replaceable>} rate=<replaceable>rate</replaceable>"
+"[reg|hp] start <replaceable>rule_name</replaceable> jobid={<replaceable>jobid_list</replaceable>} rate=<replaceable>rate</replaceable>"
            </screen>
            <para>Wildcard is supported in
            {<replaceable>jobid_list</replaceable>}.</para>
@@ -1730,7 +1717,7 @@ $ lctl set_param ost.OSS.ost_io.nrs_tbf_rule=\
            <para><emphasis role="bold">Opcode based TBF policy</emphasis></para>
            <para>Command:</para>
             <screen>$ lctl set_param x.x.x.nrs_tbf_rule=
-"[reg|hp] start <replaceable>name</replaceable> opcode={<replaceable>opcode_list</replaceable>} rate=<replaceable>rate</replaceable>"
+"[reg|hp] start <replaceable>rule_name</replaceable> opcode={<replaceable>opcode_list</replaceable>} rate=<replaceable>rate</replaceable>"
            </screen>
             <para>Example:</para>
            <screen>$ lctl set_param ost.OSS.ost_io.nrs_tbf_rule=\
@@ -1745,11 +1732,33 @@ $ lctl set_param ost.OSS.ost_io.nrs_tbf_rule=\
 "reg start iozone_user1 opcode={ost_read} rate=100"</screen>
          </listitem>
          <listitem>
+      <para><emphasis role="bold">UID/GID based TBF policy</emphasis></para>
+           <para>Command:</para>
+           <screen>$ lctl set_param ost.OSS.*.nrs_tbf_rule=\
+"[reg][hp] start <replaceable>rule_name</replaceable> uid={<replaceable>uid</replaceable>} rate=<replaceable>rate</replaceable>"
+$ lctl set_param ost.OSS.*.nrs_tbf_rule=\
+"[reg][hp] start <replaceable>rule_name</replaceable> gid={<replaceable>gid</replaceable>} rate=<replaceable>rate</replaceable>"</screen>
+           <para>Exapmle:</para>
+           <para>Limit the rate of RPC requests of the uid 500</para>
+           <screen>$ lctl set_param ost.OSS.*.nrs_tbf_rule=\
+"start tbf_name uid={500} rate=100"</screen>
+           <para>Limit the rate of RPC requests of the gid 500</para>
+           <screen>$ lctl set_param ost.OSS.*.nrs_tbf_rule=\
+"start tbf_name gid={500} rate=100"</screen>
+           <para>Also, you can use the following rule to control all reqs
+           to mds:</para>
+           <para>Start the tbf uid QoS on MDS:</para>
+           <screen>$ lctl set_param mds.MDS.*.nrs_policies="tbf uid"</screen>
+           <para>Limit the rate of RPC requests of the uid 500</para>
+           <screen>$ lctl set_param mds.MDS.*.nrs_tbf_rule=\
+"start tbf_name uid={500} rate=100"</screen>
+         </listitem>
+         <listitem>
            <para><emphasis role="bold">Policy combination</emphasis></para>
-           <para>To support rules with complex expressions of NID/JOBID/OPCode
-           conditions, TBF classifier is extented to classify RPC in a more
-           fine-grained way. This feature supports logical conditional
-           conjunction and disjunction operations among different types.
+           <para>To support TBF rules with complex expressions of conditions,
+           TBF classifier is extented to classify RPC in a more fine-grained
+           way. This feature supports logical conditional conjunction and
+           disjunction operations among different types.
            In the rule:
            "&amp;" represents the conditional conjunction and
            "," represents the conditional disjunction.</para>
@@ -1780,6 +1789,11 @@ default * 10000, ref 0
 CPT 1:
 comp_rule opcode={ost_write}&amp;jobid={dd.0},nid={192.168.1.[1-128]@tcp 0@lo} 100, ref 0
 default * 10000, ref 0</screen>
+           <para>Example:</para>
+           <screen>$ lctl set_param ost.OSS.*.nrs_tbf_rule=\
+"start tbf_name uid={500}&amp;gid={500} rate=100"</screen>
+           <para>In this example, those RPC requests whose uid is 500 and
+           gid is 500 will be processed at the rate of 100 req/sec.</para>
          </listitem>
        </itemizedlist>
       </section>
@@ -1902,7 +1916,7 @@ default * 10000, ref 0</screen>
        </itemizedlist>
       </section>
     </section>
-    <section xml:id="dbdoclet.delaytuning" condition='l2A'>
+    <section xml:id="delaytuning" condition='l2A'>
       <title>
       <indexterm>
         <primary>tuning</primary>
@@ -2048,7 +2062,7 @@ ost.OSS.ost_io.nrs_delay_pct=hp_delay_pct:5
       </itemizedlist>
     </section>
   </section>
-  <section xml:id="dbdoclet.50438272_25884">
+  <section xml:id="tuning_lockless_IO">
     <title>
     <indexterm>
       <primary>tuning</primary>
@@ -2084,9 +2098,7 @@ ldlm.namespaces.filter-<replaceable>fsname</replaceable>-*.
         <para>
           <emphasis role="bold">Client-side:</emphasis>
         </para>
-        <screen>
-/proc/fs/lustre/llite/lustre-*
-</screen>
+        <screen>llite.<replaceable>fsname</replaceable>-*</screen>
         <para>
         <literal>contention_seconds</literal>- 
         <literal>llite</literal> inode remembers its contended state for the
@@ -2097,8 +2109,8 @@ ldlm.namespaces.filter-<replaceable>fsname</replaceable>-*.
           <emphasis role="bold">Client-side statistics:</emphasis>
         </para>
         <para>The 
-        <literal>/proc/fs/lustre/llite/lustre-*/stats</literal> file has new
-        rows for lockless I/O statistics.</para>
+        <literal>llite.<replaceable>fsname</replaceable>-*.stats</literal>
+        parameter has several entries for lockless I/O statistics.</para>
         <para>
         <literal>lockless_read_bytes</literal> and 
         <literal>lockless_write_bytes</literal>- To count the total bytes read
@@ -2119,7 +2131,7 @@ ldlm.namespaces.filter-<replaceable>fsname</replaceable>-*.
         Server-Side Advice and Hinting
       </title>
       <section><title>Overview</title>
-      <para>Use the <literal>lfs ladvise</literal> command give file access
+      <para>Use the <literal>lfs ladvise</literal> command to give file access
       advices or hints to servers.</para>
       <screen>lfs ladvise [--advice|-a ADVICE ] [--background|-b]
 [--start|-s START[kMGT]]
@@ -2154,6 +2166,10 @@ ldlm.namespaces.filter-<replaceable>fsname</replaceable>-*.
                 cache</para>
                 <para><literal>dontneed</literal> to cleanup data cache on
                 server</para>
+                <para><literal>lockahead</literal> Request an LDLM extent lock
+                of the given mode on the given byte range </para>
+                <para><literal>noexpand</literal> Disable extent lock expansion
+                behavior for I/O to this file descriptor</para>
               </entry>
             </row>
             <row>
@@ -2199,6 +2215,16 @@ ldlm.namespaces.filter-<replaceable>fsname</replaceable>-*.
                   <literal>-e</literal> option.</para>
                 </entry>
             </row>
+            <row>
+                <entry>
+                    <para><literal>-m</literal>, <literal>--mode=</literal>
+                        <literal>MODE</literal></para>
+                </entry>
+                <entry>
+                  <para>Lockahead request mode <literal>{READ,WRITE}</literal>.
+                  Request a lock with this mode.</para>
+                </entry>
+            </row>
           </tbody>
           </tgroup>
         </informaltable>
@@ -2215,6 +2241,18 @@ ldlm.namespaces.filter-<replaceable>fsname</replaceable>-*.
       random IO is a net benefit. Fetching that data into each client cache with
       fadvise() may not be, due to much more data being sent to the client.
       </para>
+      <para>
+      <literal>ladvise lockahead</literal> is different in that it attempts to
+      control LDLM locking behavior by explicitly requesting LDLM locks in
+      advance of use.  This does not directly affect caching behavior, instead
+      it is used in special cases to avoid pathological results (lock exchange)
+      from the normal LDLM locking behavior.
+      </para>
+      <para>
+      Note that the <literal>noexpand</literal> advice works on a specific
+      file descriptor, so using it via lfs has no effect.  It must be used
+      on a particular file descriptor which is used for i/o to have any effect.
+      </para>
       <para>The main difference between the Linux <literal>fadvise()</literal>
       system call and <literal>lfs ladvise</literal> is that
       <literal>fadvise()</literal> is only a client side mechanism that does
@@ -2233,6 +2271,17 @@ ldlm.namespaces.filter-<replaceable>fsname</replaceable>-*.
         cache of the file in the memory.</para>
         <screen>client1$ lfs ladvise -a dontneed -s 0 -e 1048576000 /mnt/lustre/file1
         </screen>
+        <para>The following example requests an LDLM read lock on the first
+       1 MiB of <literal>/mnt/lustre/file1</literal>.  This will attempt to
+       request a lock from the OST holding that region of the file.</para>
+        <screen>client1$ lfs ladvise -a lockahead -m READ -s 0 -e 1M /mnt/lustre/file1
+        </screen>
+        <para>The following example requests an LDLM write lock on
+       [3 MiB, 10 MiB] of <literal>/mnt/lustre/file1</literal>.  This will
+       attempt to request a lock from the OST holding that region of the
+       file.</para>
+        <screen>client1$ lfs ladvise -a lockahead -m WRITE -s 3M -e 10M /mnt/lustre/file1
+        </screen>
       </section>
   </section>
   <section condition="l29">
@@ -2274,10 +2323,9 @@ ldlm.namespaces.filter-<replaceable>fsname</replaceable>-*.
           16MB.  To temporarily change <literal>brw_size</literal>, the
           following command should be run on the OSS:</para>
           <screen>oss# lctl set_param obdfilter.<replaceable>fsname</replaceable>-OST*.brw_size=16</screen>
-          <para>To persistently change <literal>brw_size</literal>, one of the following
-          commands should be run on the OSS:</para>
+          <para>To persistently change <literal>brw_size</literal>, the
+          following command should be run:</para>
           <screen>oss# lctl set_param -P obdfilter.<replaceable>fsname</replaceable>-OST*.brw_size=16</screen>
-          <screen>oss# lctl conf_param <replaceable>fsname</replaceable>-OST*.obdfilter.brw_size=16</screen>
           <para>When a client connects to an OST target, it will fetch
           <literal>brw_size</literal> from the target and pick the maximum value
           of <literal>brw_size</literal> and its local setting for
@@ -2290,13 +2338,13 @@ ldlm.namespaces.filter-<replaceable>fsname</replaceable>-*.
           <screen>client$ lctl set_param osc.<replaceable>fsname</replaceable>-OST*.max_pages_per_rpc=16M</screen>
           <para>To persistently make this change, the following command should
           be run:</para>
-          <screen>client$ lctl conf_param <replaceable>fsname</replaceable>-OST*.osc.max_pages_per_rpc=16M</screen>
+          <screen>client$ lctl set_param -P obdfilter.<replaceable>fsname</replaceable>-OST*.osc.max_pages_per_rpc=16M</screen>
           <caution><para>The <literal>brw_size</literal> of an OST can be
           changed on the fly.  However, clients have to be remounted to
           renegotiate the new maximum RPC size.</para></caution>
       </section>
   </section>
-  <section xml:id="dbdoclet.50438272_80545">
+  <section xml:id="tuning_IO_small_files">
     <title>
     <indexterm>
       <primary>tuning</primary>
@@ -2336,7 +2384,7 @@ ldlm.namespaces.filter-<replaceable>fsname</replaceable>-*.
       </listitem>
     </itemizedlist>
   </section>
-  <section xml:id="dbdoclet.50438272_45406">
+  <section xml:id="write_vs_read_performance">
     <title>
     <indexterm>
       <primary>tuning</primary>
@@ -2363,3 +2411,6 @@ ldlm.namespaces.filter-<replaceable>fsname</replaceable>-*.
     client is more likely to become CPU-bound during reads than writes.</para>
   </section>
 </chapter>
+<!--
+  vim:expandtab:shiftwidth=2:tabstop=8:
+  -->