X-Git-Url: https://git.whamcloud.com/?p=fs%2Flustre-release.git;a=blobdiff_plain;f=lustre%2FChangeLog;h=05fe2eae7c4bdf27cb8d0bcd0cb05a8b64e11284;hp=49629d65fc241374126ab8d80c73a80520aaab1c;hb=52aa92420fe9829603bfb98dcda31bf22d995e42;hpb=04ec54ff56b83a9114f7a25fbd4aa5f65e68ef7a diff --git a/lustre/ChangeLog b/lustre/ChangeLog index 49629d6..05fe2ea 100644 --- a/lustre/ChangeLog +++ b/lustre/ChangeLog @@ -1,15 +1,100 @@ TBD Intel Corporation + * version 2.8.0 + * See https://wiki.hpdd.intel.com/display/PUB/Lustre+Support+Matrix + for currently supported client and server kernel versions. + * Server known to build on patched kernels: + 2.6.32-431.29.2.el6 (RHEL6.5) + 2.6.32-504.30.3.el6 (RHEL6.6) + 2.6.32-573.7.1.el6 (RHEL6.7) + 3.10.0-229.14.1.el7 (RHEL7.1) + 3.0.101-0.47.55 (SLES11 SP3) + 3.0.101-65 (SLES11 SP4) + * Client known to build on unpatched kernels: + 2.6.32-431.29.2.el6 (RHEL6.5) + 2.6.32-504.30.3.el6 (RHEL6.6) + 2.6.32-573.7.1.el6 (RHEL6.7) + 3.10.0-229.14.1.el7 (RHEL7.1) + 3.0.101-0.47.55 (SLES11 SP3) + 3.0.101-65 (SLES11 SP4) + 3.12.39-47 (SLES12) + * Recommended e2fsprogs version: 1.42.13.wc3 or newer + * Recommended ZFS / SPL version: 0.6.5.2 + * Tested with ZFS / SPL version: 0.6.5.2 + * NFS export disabled when stack size < 8192 (32-bit Lustre clients), + since the NFSv4 export of Lustre filesystem with 4K stack may cause a + stack overflow. For more information, please refer to bugzilla 17630. + * NFSv4 reexport to 32-bit NFS client nodes requires Lustre client on + the re-exporting nodes to be mounted with "32bitapi" mount option + +-------------------------------------------------------------------------------- + +03-10-2015 Intel Corporation + * version 2.7.0 + * See https://wiki.hpdd.intel.com/display/PUB/Lustre+Support+Matrix + for currently supported client and server kernel versions. + * Server known to build on patched kernels: + 2.6.32-431.29.2.el6 (RHEL6.5) + 2.6.32-504.8.1.el6 (RHEL6.6) + 3.0.101-0.46 (SLES11 SP3) + * Client known to build on unpatched kernels: + 2.6.32-431.29.2.el6 (RHEL6.5) + 2.6.32-504.8.1.el6 (RHEL6.6) + 3.10.0-123.20.1.el7 (RHEL7) + 3.0.101-0.46 (SLES11 SP3) + * Recommended e2fsprogs version: 1.42.9.wc1 or newer + * NFS export disabled when stack size < 8192 (32-bit Lustre clients), + since the NFSv4 export of Lustre filesystem with 4K stack may cause a + stack overflow. For more information, please refer to bugzilla 17630. + * NFSv4 reexport to 32-bit NFS client nodes requires Lustre client on + the re-exporting nodes to be mounted with "32bitapi" mount option + +Severity : enhancement +Jira : LU-6050 +Description: control OST-index in IDIF via ROCOMPAT flag. +Details : Introduce new flag OBD_ROCOMPAT_IDX_IN_IDIF that is stored in the + last_rcvd file. For new formatted OST device, it will be auto set; + for the case of upgrading from old OST device, you can enable it + via the lproc interface osd-ldiskfs.index_in_idif. With such flag + enabled, for new created OST-object, its IDIF-in-LMA will contain + the OST-index; for the existing OST-object, the OSD will convert + old format IDIF as new format IDIF with OST-index stored in the + LMA EA when accessing such OST-object or via OI scrub. Once such + flag is enabled, it cannot be reverted back, so the system cannot + be downgraded to the orignal incompatible version. + +-------------------------------------------------------------------------------- + +07-30-2014 Intel Corporation + * version 2.6.0 + * See https://wiki.hpdd.intel.com/display/PUB/Lustre+Support+Matrix + for currently supported client and server kernel versions. + * Server known to build on patched kernels: + 2.6.32-431.20.3.el6 (RHEL6.5) + 3.0.101-0.31 (SLES11 SP3) + * Client known to build on unpatched kernels: + 2.6.32-431.20.3.el6 (RHEL6.5) + 3.0.101-0.31 (SLES11 SP3) + 3.6.10-4.fc18 (FC18) + 3.10 (kernel.org) + * Recommended e2fsprogs version: 1.42.9.wc1 or newer + * NFS export disabled when stack size < 8192 (32-bit Lustre clients), + since the NFSv4 export of Lustre filesystem with 4K stack may cause a + stack overflow. For more information, please refer to bugzilla 17630. + * NFSv4 reexport to 32-bit NFS client nodes requires Lustre client on + the re-exporting nodes to be mounted with "32bitapi" mount option + +-------------------------------------------------------------------------------- + +10-31-2013 Intel Corporation * version 2.5.0 - * See http://wiki.whamcloud.com/display/PUB/Lustre+Support+Matrix - * Server support for kernels: - 2.6.32-358.6.1.el6 (RHEL6) - * Client support for unpatched kernels: - 2.6.32-358.6.1.el6 (RHEL6) - 2.6.27.19-5 (SLES11) - 2.6.18-348.3.1.el5 (RHEL5) - 2.6.16.60-0.69.1 (SLES10) - 3.6.10-2.fc17 (FC17) - * Recommended e2fsprogs version: 1.42.7.wc1 + * See https://wiki.hpdd.intel.com/display/PUB/Lustre+Support+Matrix + * Server builds for kernels: + 2.6.32-358.18.1.el6 (RHEL6.4) + * Client builds for unpatched kernels: + 2.6.32-358.18.1.el6 (RHEL6.4) + 3.0.92-0.8 (SLES11 SP3) + 3.6.10-4.fc18 (FC18) + * Recommended e2fsprogs version: 1.42.7.wc2 or newer * NFS export disabled when stack size < 8192 (32-bit Lustre clients), since the NFSv4 export of Lustre filesystem with 4K stack may cause a stack overflow. For more information, please refer to bugzilla 17630. @@ -20,16 +105,14 @@ TBD Intel Corporation 05-30-2013 Intel Corporation * version 2.4.0 - * See http://wiki.whamcloud.com/display/PUB/Lustre+Support+Matrix - * Server support for kernels: - 2.6.32-358.6.1.el6 (RHEL6) - * Client support for unpatched kernels: - 2.6.32-358.6.1.el6 (RHEL6) - 2.6.27.19-5 (SLES11) - 2.6.18-348.3.1.el5 (RHEL5) - 2.6.16.60-0.69.1 (SLES10) + * See https://wiki.hpdd.intel.com/display/PUB/Lustre+Support+Matrix + * Server builds for kernels: + 2.6.32-358.6.2.el6 (RHEL6.4) + * Client builds for unpatched kernels: + 2.6.32-358.6.2.el6 (RHEL6.4) + 3.0.74-0.6.6 (SLES11 SP2) 3.6.10-2.fc17 (FC17) - * Recommended e2fsprogs version: 1.42.7.wc1 + * Recommended e2fsprogs version: 1.42.7.wc1 or newer * NFS export disabled when stack size < 8192 (32-bit Lustre clients), since the NFSv4 export of Lustre filesystem with 4K stack may cause a stack overflow. For more information, please refer to bugzilla 17630. @@ -40,7 +123,7 @@ TBD Intel Corporation 9-30-2012 Whamcloud, Inc. * version 2.3.0 - * See http://wiki.whamcloud.com/display/PUB/Lustre+Support+Matrix + * See https://wiki.hpdd.intel.com/display/PUB/Lustre+Support+Matrix * Server support for kernels: 2.6.32-279.5.1.el6 (RHEL6) * Client support for unpatched kernels: @@ -79,7 +162,7 @@ Description: Update RHEL6.2 kernel to 2.6.32-220.7.1.el6 04-30-2012 Whamcloud, Inc. * version 2.2.0 - * See http://wiki.whamcloud.com/display/PUB/Lustre+Support+Matrix + * See https://wiki.hpdd.intel.com/display/PUB/Lustre+Support+Matrix * Server support for kernels: 2.6.32-220.4.2.el6 (RHEL6) * Client support for unpatched kernels: @@ -383,7 +466,7 @@ Description: Remove set_info(KEY_UNLINKED) from MDS/OSC Severity : enhancement Bugzilla : 19526 Description: correctly handle big reply message. -Details : send LNet event if reply is bigger then buffer and adjust this buffer +Details : send LNet event if reply is bigger than buffer and adjust this buffer correctly. Severity : normal @@ -2650,7 +2733,7 @@ Severity : normal Bugzilla : 19128 Description: Out or order replies might be lost on replay Details : In ptlrpc_retain_replayable_request if we cannot find retained - request with tid smaller then one currently being added, add it + request with tid smaller than one currently being added, add it to the start, not end of the list. -------------------------------------------------------------------------------- @@ -2863,13 +2946,13 @@ Details : When osc reconnect to OST, OST(filter)should clear grant info of Severity : normal Frequency : rare Bugzilla : 11662 -Description: Grant space more than avaiable left space sometimes. +Description: Grant space more than available left space sometimes. Details : When then OST is about to be full, if two bulk writing from different clients came to OST. Accord the avaliable space of the OST, the first req should be permitted, and the second one should be denied by ENOSPC. But if the seconde arrived before the first one is commited. The OST might wrongly permit second - writing, which will cause grant space > avaiable space. + writing, which will cause grant space > available space. Severity : normal Frequency : when client is evicted @@ -2998,11 +3081,11 @@ Details : decrease the amount of synchronous RPC between clients and servers Severity : enhancement Bugzilla : 4900 Description: Async OSC create to avoid the blocking unnecessarily. -Details : If a OST has no remain object, system will block on the creating - when need to create a new object on this OST. Now, ways use - pre-created objects when available, instead of blocking on an - empty osc while others are not empty. If we must block, we block - for the shortest possible period of time. +Details : If an OST has no remaining object, system will block on the + creation when it needs to create a new object on this OST. Now, + ways use pre-created objects when available, instead of blocking on + an empty osc while others are not empty. If we must block, we + block for the shortest possible period of time. Severity : major Bugzilla : 11710 @@ -3768,7 +3851,7 @@ Details : If one node attempts to overwrite an executable in use by Severity : enhancement Bugzilla : 4900 Description: Async OSC create to avoid the blocking unnecessarily. -Details : If a OST has no remain object, system will block on the creating +Details : If an OST has no remaining object, system will block on the creating when need to create a new object on this OST. Now, ways use pre-created objects when available, instead of blocking on an empty osc while others are not empty. If we must block, we block @@ -4676,7 +4759,7 @@ Frequency : common Bugzilla : 9489, 3273 Description: First write from each client to each OST was only 4kB in size, to initialize client writeback cache, which caused sub-optimal - RPCs and poor layout on disk for the first writen file. + RPCs and poor layout on disk for the first written file. Details : Clients now request an initial cache grant at (re)connect time and so that they can start streaming writes to the cache right away and always do full-sized RPCs if there is enough data.