X-Git-Url: https://git.whamcloud.com/?a=blobdiff_plain;f=LustreTroubleshooting.xml;h=b82c5588c5ed709f3097a8a26e485859199e6909;hb=d2c7df42886ed80cf2e5a82d9a1521c0003dddf8;hp=69cffe48799f1eec7881bfe55592c47d166707f8;hpb=2486c888f37c9efc590bebccffe7eca9817268aa;p=doc%2Fmanual.git diff --git a/LustreTroubleshooting.xml b/LustreTroubleshooting.xml index 69cffe4..b82c558 100644 --- a/LustreTroubleshooting.xml +++ b/LustreTroubleshooting.xml @@ -1,4 +1,7 @@ - + + Lustre File System Troubleshooting This chapter provides information about troubleshooting a Lustre file system, submitting a bug to the Jira bug tracking system, and Lustre file system performance tips. It includes the @@ -201,7 +204,8 @@ Which server node it was communicating with, and so on. - Lustre logs are dumped to /proc/sys/lnet/debug_path. + Lustre logs are dumped to the pathname stored in the parameter + lnet.debug_path. Collect the first group of messages related to a problem, and any messages that precede "LBUG" or "assertion failure" errors. Messages that mention server nodes (OST or MDS) are specific to that server; you must collect similar messages from the relevant server console logs. Another Lustre debug log holds information for a short period of time for action by the Lustre software, which, in turn, depends on the processes on the Lustre node. Use the @@ -373,7 +377,7 @@ If the SCSI devices are inaccessible to the Lustre file system at the block device level, then ldiskfs remounts the device read-only to prevent file system corruption. This is a normal - behavior. The status in /proc/fs/lustre/health_check + behavior. The status in the parameter health_check also shows "not healthy" on the affected nodes. To determine what caused the "not healthy" condition: @@ -745,7 +749,8 @@ server now claims 791)! Lustre or kernel stack traces showing processes stuck in "try_to_free_pages" - For information on determining the MDS memory and OSS memory requirements, see . + For information on determining the MDS memory and OSS memory + requirements, see .
Setting SCSI I/O Sizes @@ -764,3 +769,6 @@ server now claims 791)!
+