Whamcloud - gitweb
LU-14711 osc: Do not attempt sending empty pages
[fs/lustre-release.git] / lustre / osc / osc_request.c
index 15f4397..4ca9be0 100644 (file)
  */
 /*
  * This file is part of Lustre, http://www.lustre.org/
- * Lustre is a trademark of Sun Microsystems, Inc.
  */
 
 #define DEBUG_SUBSYSTEM S_OSC
 
 #include <linux/workqueue.h>
+#include <libcfs/libcfs.h>
+#include <linux/falloc.h>
 #include <lprocfs_status.h>
-#include <lustre_debug.h>
 #include <lustre_dlm.h>
 #include <lustre_fid.h>
 #include <lustre_ha.h>
 #include <obd_cksum.h>
 #include <obd_class.h>
 #include <lustre_osc.h>
+#include <linux/falloc.h>
 
 #include "osc_internal.h"
+#include <lnet/lnet_rdma.h>
 
 atomic_t osc_pool_req_count;
 unsigned int osc_reqpool_maxreqcount;
@@ -236,10 +238,7 @@ int osc_setattr_async(struct obd_export *exp, struct obdo *oa,
                sa->sa_upcall = upcall;
                sa->sa_cookie = cookie;
 
-               if (rqset == PTLRPCD_SET)
-                       ptlrpcd_add_req(req);
-               else
-                       ptlrpc_set_add_req(rqset, req);
+               ptlrpc_set_add_req(rqset, req);
        }
 
        RETURN(0);
@@ -324,10 +323,7 @@ int osc_ladvise_base(struct obd_export *exp, struct obdo *oa,
        la->la_upcall = upcall;
        la->la_cookie = cookie;
 
-       if (rqset == PTLRPCD_SET)
-               ptlrpcd_add_req(req);
-       else
-               ptlrpc_set_add_req(rqset, req);
+       ptlrpc_set_add_req(rqset, req);
 
        RETURN(0);
 }
@@ -426,6 +422,65 @@ int osc_punch_send(struct obd_export *exp, struct obdo *oa,
 }
 EXPORT_SYMBOL(osc_punch_send);
 
+/**
+ * osc_fallocate_base() - Handles fallocate request.
+ *
+ * @exp:       Export structure
+ * @oa:                Attributes passed to OSS from client (obdo structure)
+ * @upcall:    Primary & supplementary group information
+ * @cookie:    Exclusive identifier
+ * @rqset:     Request list.
+ * @mode:      Operation done on given range.
+ *
+ * osc_fallocate_base() - Handles fallocate requests only. Only block
+ * allocation or standard preallocate operation is supported currently.
+ * Other mode flags is not supported yet. ftruncate(2) or truncate(2)
+ * is supported via SETATTR request.
+ *
+ * Return: Non-zero on failure and O on success.
+ */
+int osc_fallocate_base(struct obd_export *exp, struct obdo *oa,
+                      obd_enqueue_update_f upcall, void *cookie, int mode)
+{
+       struct ptlrpc_request *req;
+       struct osc_setattr_args *sa;
+       struct ost_body *body;
+       struct obd_import *imp = class_exp2cliimp(exp);
+       int rc;
+       ENTRY;
+
+       oa->o_falloc_mode = mode;
+       req = ptlrpc_request_alloc(class_exp2cliimp(exp),
+                                  &RQF_OST_FALLOCATE);
+       if (req == NULL)
+               RETURN(-ENOMEM);
+
+       rc = ptlrpc_request_pack(req, LUSTRE_OST_VERSION, OST_FALLOCATE);
+       if (rc != 0) {
+               ptlrpc_request_free(req);
+               RETURN(rc);
+       }
+
+       body = req_capsule_client_get(&req->rq_pill, &RMF_OST_BODY);
+       LASSERT(body);
+
+       lustre_set_wire_obdo(&imp->imp_connect_data, &body->oa, oa);
+
+       ptlrpc_request_set_replen(req);
+
+       req->rq_interpret_reply = osc_setattr_interpret;
+       BUILD_BUG_ON(sizeof(*sa) > sizeof(req->rq_async_args));
+       sa = ptlrpc_req_async_args(sa, req);
+       sa->sa_oa = oa;
+       sa->sa_upcall = upcall;
+       sa->sa_cookie = cookie;
+
+       ptlrpcd_add_req(req);
+
+       RETURN(0);
+}
+EXPORT_SYMBOL(osc_fallocate_base);
+
 static int osc_sync_interpret(const struct lu_env *env,
                              struct ptlrpc_request *req, void *args, int rc)
 {
@@ -499,10 +554,7 @@ int osc_sync_base(struct osc_object *obj, struct obdo *oa,
        fa->fa_upcall = upcall;
        fa->fa_cookie = cookie;
 
-       if (rqset == PTLRPCD_SET)
-               ptlrpcd_add_req(req);
-       else
-               ptlrpc_set_add_req(rqset, req);
+       ptlrpc_set_add_req(rqset, req);
 
        RETURN (0);
 }
@@ -640,7 +692,7 @@ static void osc_announce_cached(struct client_obd *cli, struct obdo *oa,
 
        oa->o_valid |= bits;
        spin_lock(&cli->cl_loi_list_lock);
-       if (OCD_HAS_FLAG(&cli->cl_import->imp_connect_data, GRANT_PARAM))
+       if (cli->cl_ocd_grant_param)
                oa->o_dirty = cli->cl_dirty_grant;
        else
                oa->o_dirty = cli->cl_dirty_pages << PAGE_SHIFT;
@@ -671,13 +723,12 @@ static void osc_announce_cached(struct client_obd *cli, struct obdo *oa,
                nrpages *= cli->cl_max_rpcs_in_flight + 1;
                nrpages = max(nrpages, cli->cl_dirty_max_pages);
                undirty = nrpages << PAGE_SHIFT;
-               if (OCD_HAS_FLAG(&cli->cl_import->imp_connect_data,
-                                GRANT_PARAM)) {
+               if (cli->cl_ocd_grant_param) {
                        int nrextents;
 
                        /* take extent tax into account when asking for more
                         * grant space */
-                       nrextents = (nrpages + cli->cl_max_extent_pages - 1)  /
+                       nrextents = (nrpages + cli->cl_max_extent_pages - 1) /
                                     cli->cl_max_extent_pages;
                        undirty += nrextents * cli->cl_grant_extent_tax;
                }
@@ -688,11 +739,20 @@ static void osc_announce_cached(struct client_obd *cli, struct obdo *oa,
                                    ~(PTLRPC_MAX_BRW_SIZE * 4UL));
         }
        oa->o_grant = cli->cl_avail_grant + cli->cl_reserved_grant;
-        oa->o_dropped = cli->cl_lost_grant;
-        cli->cl_lost_grant = 0;
+       /* o_dropped AKA o_misc is 32 bits, but cl_lost_grant is 64 bits */
+       if (cli->cl_lost_grant > INT_MAX) {
+               CDEBUG(D_CACHE,
+                     "%s: avoided o_dropped overflow: cl_lost_grant %lu\n",
+                     cli_name(cli), cli->cl_lost_grant);
+               oa->o_dropped = INT_MAX;
+       } else {
+               oa->o_dropped = cli->cl_lost_grant;
+       }
+       cli->cl_lost_grant -= oa->o_dropped;
        spin_unlock(&cli->cl_loi_list_lock);
-       CDEBUG(D_CACHE, "dirty: %llu undirty: %u dropped %u grant: %llu\n",
-               oa->o_dirty, oa->o_undirty, oa->o_dropped, oa->o_grant);
+       CDEBUG(D_CACHE, "%s: dirty: %llu undirty: %u dropped %u grant: %llu"
+              " cl_lost_grant %lu\n", cli_name(cli), oa->o_dirty,
+              oa->o_undirty, oa->o_dropped, oa->o_grant, cli->cl_lost_grant);
 }
 
 void osc_update_next_shrink(struct client_obd *cli)
@@ -964,12 +1024,19 @@ void osc_init_grant(struct client_obd *cli, struct obd_connect_data *ocd)
        spin_lock(&cli->cl_loi_list_lock);
        cli->cl_avail_grant = ocd->ocd_grant;
        if (cli->cl_import->imp_state != LUSTRE_IMP_EVICTED) {
-               cli->cl_avail_grant -= cli->cl_reserved_grant;
+               unsigned long consumed = cli->cl_reserved_grant;
+
                if (OCD_HAS_FLAG(ocd, GRANT_PARAM))
-                       cli->cl_avail_grant -= cli->cl_dirty_grant;
+                       consumed += cli->cl_dirty_grant;
                else
-                       cli->cl_avail_grant -=
-                                       cli->cl_dirty_pages << PAGE_SHIFT;
+                       consumed += cli->cl_dirty_pages << PAGE_SHIFT;
+               if (cli->cl_avail_grant < consumed) {
+                       CERROR("%s: granted %ld but already consumed %ld\n",
+                              cli_name(cli), cli->cl_avail_grant, consumed);
+                       cli->cl_avail_grant = 0;
+               } else {
+                       cli->cl_avail_grant -= consumed;
+               }
        }
 
        if (OCD_HAS_FLAG(ocd, GRANT_PARAM)) {
@@ -987,10 +1054,10 @@ void osc_init_grant(struct client_obd *cli, struct obd_connect_data *ocd)
                                             ~chunk_mask) & chunk_mask;
                /* determine maximum extent size, in #pages */
                size = (u64)ocd->ocd_grant_max_blks << ocd->ocd_grant_blkbits;
-               cli->cl_max_extent_pages = size >> PAGE_SHIFT;
-               if (cli->cl_max_extent_pages == 0)
-                       cli->cl_max_extent_pages = 1;
+               cli->cl_max_extent_pages = (size >> PAGE_SHIFT) ?: 1;
+               cli->cl_ocd_grant_param = 1;
        } else {
+               cli->cl_ocd_grant_param = 0;
                cli->cl_grant_extent_tax = 0;
                cli->cl_chunkbits = PAGE_SHIFT;
                cli->cl_max_extent_pages = DT_MAX_BRW_PAGES;
@@ -1111,7 +1178,7 @@ static int osc_checksum_bulk_t10pi(const char *obd_name, int nob,
                                   size_t pg_count, struct brw_page **pga,
                                   int opc, obd_dif_csum_fn *fn,
                                   int sector_size,
-                                  u32 *check_sum)
+                                  u32 *check_sum, bool resend)
 {
        struct ahash_request *req;
        /* Used Adler as the default checksum type on top of DIF tags */
@@ -1144,6 +1211,10 @@ static int osc_checksum_bulk_t10pi(const char *obd_name, int nob,
        buffer = kmap(__page);
        guard_start = (__u16 *)buffer;
        guard_number = PAGE_SIZE / sizeof(*guard_start);
+       CDEBUG(D_PAGE | (resend ? D_HA : 0),
+              "GRD tags per page=%u, resend=%u, bytes=%u, pages=%zu\n",
+              guard_number, resend, nob, pg_count);
+
        while (nob > 0 && pg_count > 0) {
                unsigned int count = pga[i]->count > nob ? nob : pga[i]->count;
 
@@ -1169,6 +1240,12 @@ static int osc_checksum_bulk_t10pi(const char *obd_name, int nob,
                                                  guard_number - used_number,
                                                  &used, sector_size,
                                                  fn);
+               if (unlikely(resend))
+                       CDEBUG(D_PAGE | D_HA,
+                              "pga[%u]: used %u off %llu+%u gen checksum: %*phN\n",
+                              i, used, pga[i]->off & ~PAGE_MASK, count,
+                              (int)(used * sizeof(*guard_start)),
+                              guard_start + used_number);
                if (rc)
                        break;
 
@@ -1207,7 +1284,7 @@ out:
 #else /* !CONFIG_CRC_T10DIF */
 #define obd_dif_ip_fn NULL
 #define obd_dif_crc_fn NULL
-#define osc_checksum_bulk_t10pi(name, nob, pgc, pga, opc, fn, ssize, csum \
+#define osc_checksum_bulk_t10pi(name, nob, pgc, pga, opc, fn, ssize, csum, re) \
        -EOPNOTSUPP
 #endif /* CONFIG_CRC_T10DIF */
 
@@ -1269,7 +1346,7 @@ static int osc_checksum_bulk_rw(const char *obd_name,
                                enum cksum_types cksum_type,
                                int nob, size_t pg_count,
                                struct brw_page **pga, int opc,
-                               u32 *check_sum)
+                               u32 *check_sum, bool resend)
 {
        obd_dif_csum_fn *fn = NULL;
        int sector_size = 0;
@@ -1280,7 +1357,8 @@ static int osc_checksum_bulk_rw(const char *obd_name,
 
        if (fn)
                rc = osc_checksum_bulk_t10pi(obd_name, nob, pg_count, pga,
-                                            opc, fn, sector_size, check_sum);
+                                            opc, fn, sector_size, check_sum,
+                                            resend);
        else
                rc = osc_checksum_bulk(nob, pg_count, pga, opc, cksum_type,
                                       check_sum);
@@ -1288,28 +1366,65 @@ static int osc_checksum_bulk_rw(const char *obd_name,
        RETURN(rc);
 }
 
+static inline void osc_release_bounce_pages(struct brw_page **pga,
+                                           u32 page_count)
+{
+#ifdef HAVE_LUSTRE_CRYPTO
+       int i;
+
+       for (i = 0; i < page_count; i++) {
+               /* Bounce pages allocated by a call to
+                * llcrypt_encrypt_pagecache_blocks() in osc_brw_prep_request()
+                * are identified thanks to the PageChecked flag.
+                */
+               if (PageChecked(pga[i]->pg))
+                       llcrypt_finalize_bounce_page(&pga[i]->pg);
+               pga[i]->count -= pga[i]->bp_count_diff;
+               pga[i]->off += pga[i]->bp_off_diff;
+       }
+#endif
+}
+
 static int
 osc_brw_prep_request(int cmd, struct client_obd *cli, struct obdo *oa,
                     u32 page_count, struct brw_page **pga,
                     struct ptlrpc_request **reqp, int resend)
 {
-        struct ptlrpc_request   *req;
-        struct ptlrpc_bulk_desc *desc;
-        struct ost_body         *body;
-        struct obd_ioobj        *ioobj;
-        struct niobuf_remote    *niobuf;
+       struct ptlrpc_request *req;
+       struct ptlrpc_bulk_desc *desc;
+       struct ost_body *body;
+       struct obd_ioobj *ioobj;
+       struct niobuf_remote *niobuf;
        int niocount, i, requested_nob, opc, rc, short_io_size = 0;
-        struct osc_brw_async_args *aa;
-        struct req_capsule      *pill;
-        struct brw_page *pg_prev;
+       struct osc_brw_async_args *aa;
+       struct req_capsule *pill;
+       struct brw_page *pg_prev;
        void *short_io_buf;
        const char *obd_name = cli->cl_import->imp_obd->obd_name;
+       struct inode *inode = NULL;
+       bool directio = false;
+       bool enable_checksum = true;
 
-        ENTRY;
-        if (OBD_FAIL_CHECK(OBD_FAIL_OSC_BRW_PREP_REQ))
-                RETURN(-ENOMEM); /* Recoverable */
-        if (OBD_FAIL_CHECK(OBD_FAIL_OSC_BRW_PREP_REQ2))
-                RETURN(-EINVAL); /* Fatal */
+       ENTRY;
+       if (pga[0]->pg) {
+               inode = page2inode(pga[0]->pg);
+               if (inode == NULL) {
+                       /* Try to get reference to inode from cl_page if we are
+                        * dealing with direct IO, as handled pages are not
+                        * actual page cache pages.
+                        */
+                       struct osc_async_page *oap = brw_page2oap(pga[0]);
+                       struct cl_page *clpage = oap2cl_page(oap);
+
+                       inode = clpage->cp_inode;
+                       if (inode)
+                               directio = true;
+               }
+       }
+       if (OBD_FAIL_CHECK(OBD_FAIL_OSC_BRW_PREP_REQ))
+               RETURN(-ENOMEM); /* Recoverable */
+       if (OBD_FAIL_CHECK(OBD_FAIL_OSC_BRW_PREP_REQ2))
+               RETURN(-EINVAL); /* Fatal */
 
        if ((cmd & OBD_BRW_WRITE) != 0) {
                opc = OST_WRITE;
@@ -1323,6 +1438,96 @@ osc_brw_prep_request(int cmd, struct client_obd *cli, struct obdo *oa,
         if (req == NULL)
                 RETURN(-ENOMEM);
 
+       if (opc == OST_WRITE && inode && IS_ENCRYPTED(inode)) {
+               for (i = 0; i < page_count; i++) {
+                       struct brw_page *pg = pga[i];
+                       struct page *data_page = NULL;
+                       bool retried = false;
+                       bool lockedbymyself;
+                       u32 nunits = (pg->off & ~PAGE_MASK) + pg->count;
+                       struct address_space *map_orig = NULL;
+                       pgoff_t index_orig;
+
+retry_encrypt:
+                       if (nunits & ~LUSTRE_ENCRYPTION_MASK)
+                               nunits = (nunits & LUSTRE_ENCRYPTION_MASK) +
+                                       LUSTRE_ENCRYPTION_UNIT_SIZE;
+                       /* The page can already be locked when we arrive here.
+                        * This is possible when cl_page_assume/vvp_page_assume
+                        * is stuck on wait_on_page_writeback with page lock
+                        * held. In this case there is no risk for the lock to
+                        * be released while we are doing our encryption
+                        * processing, because writeback against that page will
+                        * end in vvp_page_completion_write/cl_page_completion,
+                        * which means only once the page is fully processed.
+                        */
+                       lockedbymyself = trylock_page(pg->pg);
+                       if (directio) {
+                               map_orig = pg->pg->mapping;
+                               pg->pg->mapping = inode->i_mapping;
+                               index_orig = pg->pg->index;
+                               pg->pg->index = pg->off >> PAGE_SHIFT;
+                       }
+                       data_page =
+                               llcrypt_encrypt_pagecache_blocks(pg->pg,
+                                                                nunits, 0,
+                                                                GFP_NOFS);
+                       if (directio) {
+                               pg->pg->mapping = map_orig;
+                               pg->pg->index = index_orig;
+                       }
+                       if (lockedbymyself)
+                               unlock_page(pg->pg);
+                       if (IS_ERR(data_page)) {
+                               rc = PTR_ERR(data_page);
+                               if (rc == -ENOMEM && !retried) {
+                                       retried = true;
+                                       rc = 0;
+                                       goto retry_encrypt;
+                               }
+                               ptlrpc_request_free(req);
+                               RETURN(rc);
+                       }
+                       /* Set PageChecked flag on bounce page for
+                        * disambiguation in osc_release_bounce_pages().
+                        */
+                       SetPageChecked(data_page);
+                       pg->pg = data_page;
+                       /* there should be no gap in the middle of page array */
+                       if (i == page_count - 1) {
+                               struct osc_async_page *oap = brw_page2oap(pg);
+
+                               oa->o_size = oap->oap_count +
+                                       oap->oap_obj_off + oap->oap_page_off;
+                       }
+                       /* len is forced to nunits, and relative offset to 0
+                        * so store the old, clear text info
+                        */
+                       pg->bp_count_diff = nunits - pg->count;
+                       pg->count = nunits;
+                       pg->bp_off_diff = pg->off & ~PAGE_MASK;
+                       pg->off = pg->off & PAGE_MASK;
+               }
+       } else if (opc == OST_READ && inode && IS_ENCRYPTED(inode)) {
+               for (i = 0; i < page_count; i++) {
+                       struct brw_page *pg = pga[i];
+                       u32 nunits = (pg->off & ~PAGE_MASK) + pg->count;
+
+                       if (nunits & ~LUSTRE_ENCRYPTION_MASK)
+                               nunits = (nunits & LUSTRE_ENCRYPTION_MASK) +
+                                       LUSTRE_ENCRYPTION_UNIT_SIZE;
+                       /* count/off are forced to cover the whole encryption
+                        * unit size so that all encrypted data is stored on the
+                        * OST, so adjust bp_{count,off}_diff for the size of
+                        * the clear text.
+                        */
+                       pg->bp_count_diff = nunits - pg->count;
+                       pg->count = nunits;
+                       pg->bp_off_diff = pg->off & ~PAGE_MASK;
+                       pg->off = pg->off & PAGE_MASK;
+               }
+       }
+
         for (niocount = i = 1; i < page_count; i++) {
                 if (!can_merge_pages(pga[i - 1], pga[i]))
                         niocount++;
@@ -1334,14 +1539,30 @@ osc_brw_prep_request(int cmd, struct client_obd *cli, struct obdo *oa,
         req_capsule_set_size(pill, &RMF_NIOBUF_REMOTE, RCL_CLIENT,
                              niocount * sizeof(*niobuf));
 
-       for (i = 0; i < page_count; i++)
+       for (i = 0; i < page_count; i++) {
                short_io_size += pga[i]->count;
+               if (!inode || !IS_ENCRYPTED(inode)) {
+                       pga[i]->bp_count_diff = 0;
+                       pga[i]->bp_off_diff = 0;
+               }
+       }
+
+       if (lnet_is_rdma_only_page(pga[0]->pg)) {
+               enable_checksum = false;
+               short_io_size = 0;
+       }
 
        /* Check if read/write is small enough to be a short io. */
        if (short_io_size > cli->cl_max_short_io_bytes || niocount > 1 ||
            !imp_connect_shortio(cli->cl_import))
                short_io_size = 0;
 
+       /* If this is an empty RPC to old server, just ignore it */
+       if (!short_io_size && !pga[0]->pg) {
+               ptlrpc_request_free(req);
+               RETURN(-ENODATA);
+       }
+
        req_capsule_set_size(pill, &RMF_SHORT_IO, RCL_CLIENT,
                             opc == OST_READ ? 0 : short_io_size);
        if (opc == OST_READ)
@@ -1369,8 +1590,7 @@ osc_brw_prep_request(int cmd, struct client_obd *cli, struct obdo *oa,
        desc = ptlrpc_prep_bulk_imp(req, page_count,
                cli->cl_import->imp_connect_data.ocd_brw_size >> LNET_MTU_BITS,
                (opc == OST_WRITE ? PTLRPC_BULK_GET_SOURCE :
-                       PTLRPC_BULK_PUT_SINK) |
-                       PTLRPC_BULK_BUF_KIOV,
+                       PTLRPC_BULK_PUT_SINK),
                OST_BULK_PORTAL,
                &ptlrpc_bulk_kiov_pin_ops);
 
@@ -1487,10 +1707,12 @@ no_bulk:
         if (osc_should_shrink_grant(cli))
                 osc_shrink_grant_local(cli, &body->oa);
 
+       if (!cli->cl_checksum || sptlrpc_flavor_has_bulk(&req->rq_flvr))
+               enable_checksum = false;
+
         /* size[REQ_REC_OFF] still sizeof (*body) */
         if (opc == OST_WRITE) {
-                if (cli->cl_checksum &&
-                    !sptlrpc_flavor_has_bulk(&req->rq_flvr)) {
+                if (enable_checksum) {
                         /* store cl_cksum_type in a local variable since
                          * it can be changed via lprocfs */
                        enum cksum_types cksum_type = cli->cl_cksum_type;
@@ -1505,17 +1727,18 @@ no_bulk:
                        rc = osc_checksum_bulk_rw(obd_name, cksum_type,
                                                  requested_nob, page_count,
                                                  pga, OST_WRITE,
-                                                 &body->oa.o_cksum);
+                                                 &body->oa.o_cksum, resend);
                        if (rc < 0) {
-                               CDEBUG(D_PAGE, "failed to checksum, rc = %d\n",
+                               CDEBUG(D_PAGE, "failed to checksum: rc = %d\n",
                                       rc);
                                GOTO(out, rc);
                        }
-                        CDEBUG(D_PAGE, "checksum at write origin: %x\n",
-                               body->oa.o_cksum);
+                       CDEBUG(D_PAGE | (resend ? D_HA : 0),
+                              "checksum at write origin: %x (%x)\n",
+                              body->oa.o_cksum, cksum_type);
 
-                        /* save this in 'oa', too, for later checking */
-                        oa->o_valid |= OBD_MD_FLCKSUM | OBD_MD_FLFLAGS;
+                       /* save this in 'oa', too, for later checking */
+                       oa->o_valid |= OBD_MD_FLCKSUM | OBD_MD_FLFLAGS;
                        oa->o_flags |= obd_cksum_type_pack(obd_name,
                                                           cksum_type);
                 } else {
@@ -1528,8 +1751,7 @@ no_bulk:
                 req_capsule_set_size(pill, &RMF_RCS, RCL_SERVER,
                                      sizeof(__u32) * niocount);
         } else {
-                if (cli->cl_checksum &&
-                    !sptlrpc_flavor_has_bulk(&req->rq_flvr)) {
+                if (enable_checksum) {
                         if ((body->oa.o_valid & OBD_MD_FLFLAGS) == 0)
                                 body->oa.o_flags = 0;
                        body->oa.o_flags |= obd_cksum_type_pack(obd_name,
@@ -1581,15 +1803,15 @@ static void dump_all_bulk_pages(struct obdo *oa, __u32 page_count,
         * file/fid, not during the resends/retries. */
        snprintf(dbgcksum_file_name, sizeof(dbgcksum_file_name),
                 "%s-checksum_dump-osc-"DFID":[%llu-%llu]-%x-%x",
-                (strncmp(libcfs_debug_file_path_arr, "NONE", 4) != 0 ?
-                 libcfs_debug_file_path_arr :
-                 LIBCFS_DEBUG_FILE_PATH_DEFAULT),
+                (strncmp(libcfs_debug_file_path, "NONE", 4) != 0 ?
+                 libcfs_debug_file_path : LIBCFS_DEBUG_FILE_PATH_DEFAULT),
                 oa->o_valid & OBD_MD_FLFID ? oa->o_parent_seq : 0ULL,
                 oa->o_valid & OBD_MD_FLFID ? oa->o_parent_oid : 0,
                 oa->o_valid & OBD_MD_FLFID ? oa->o_parent_ver : 0,
                 pga[0]->off,
                 pga[page_count-1]->off + pga[page_count-1]->count - 1,
                 client_cksum, server_cksum);
+       CWARN("dumping checksum data to %s\n", dbgcksum_file_name);
        filp = filp_open(dbgcksum_file_name,
                         O_CREAT | O_EXCL | O_WRONLY | O_LARGEFILE, 0600);
        if (IS_ERR(filp)) {
@@ -1616,8 +1838,6 @@ static void dump_all_bulk_pages(struct obdo *oa, __u32 page_count,
                        }
                        len -= rc;
                        buf += rc;
-                       CDEBUG(D_INFO, "%s: wrote %d bytes\n",
-                              dbgcksum_file_name, rc);
                }
                kunmap(pga[i]->pg);
        }
@@ -1626,6 +1846,8 @@ static void dump_all_bulk_pages(struct obdo *oa, __u32 page_count,
        if (rc)
                CERROR("%s: sync returns %d\n", dbgcksum_file_name, rc);
        filp_close(filp, NULL);
+
+       libcfs_debug_dumplog();
 }
 
 static int
@@ -1678,7 +1900,7 @@ check_write_checksum(struct obdo *oa, const struct lnet_process_id *peer,
                rc = osc_checksum_bulk_t10pi(obd_name, aa->aa_requested_nob,
                                             aa->aa_page_count, aa->aa_ppga,
                                             OST_WRITE, fn, sector_size,
-                                            &new_cksum);
+                                            &new_cksum, true);
        else
                rc = osc_checksum_bulk(aa->aa_requested_nob, aa->aa_page_count,
                                       aa->aa_ppga, OST_WRITE, cksum_type,
@@ -1725,6 +1947,8 @@ static int osc_brw_fini_request(struct ptlrpc_request *req, int rc)
                &req->rq_import->imp_connection->c_peer;
        struct ost_body *body;
        u32 client_cksum = 0;
+       struct inode *inode;
+       unsigned int blockbits = 0, blocksize = 0;
 
        ENTRY;
 
@@ -1838,22 +2062,23 @@ static int osc_brw_fini_request(struct ptlrpc_request *req, int rc)
                }
        }
 
-        if (rc < aa->aa_requested_nob)
-                handle_short_read(rc, aa->aa_page_count, aa->aa_ppga);
+       if (rc < aa->aa_requested_nob)
+               handle_short_read(rc, aa->aa_page_count, aa->aa_ppga);
 
-        if (body->oa.o_valid & OBD_MD_FLCKSUM) {
-                static int cksum_counter;
-               u32        server_cksum = body->oa.o_cksum;
-               char      *via = "";
-               char      *router = "";
+       if (body->oa.o_valid & OBD_MD_FLCKSUM) {
+               static int cksum_counter;
+               u32 server_cksum = body->oa.o_cksum;
+               int nob = rc;
+               char *via = "";
+               char *router = "";
                enum cksum_types cksum_type;
                u32 o_flags = body->oa.o_valid & OBD_MD_FLFLAGS ?
                        body->oa.o_flags : 0;
 
                cksum_type = obd_cksum_type_unpack(o_flags);
-               rc = osc_checksum_bulk_rw(obd_name, cksum_type, rc,
+               rc = osc_checksum_bulk_rw(obd_name, cksum_type, nob,
                                          aa->aa_page_count, aa->aa_ppga,
-                                         OST_READ, &client_cksum);
+                                         OST_READ, &client_cksum, false);
                if (rc < 0)
                        GOTO(out, rc);
 
@@ -1865,8 +2090,12 @@ static int osc_brw_fini_request(struct ptlrpc_request *req, int rc)
 
                if (server_cksum != client_cksum) {
                        struct ost_body *clbody;
+                       __u32 client_cksum2;
                        u32 page_count = aa->aa_page_count;
 
+                       osc_checksum_bulk_rw(obd_name, cksum_type, nob,
+                                            page_count, aa->aa_ppga,
+                                            OST_READ, &client_cksum2, true);
                        clbody = req_capsule_client_get(&req->rq_pill,
                                                        &RMF_OST_BODY);
                        if (cli->cl_checksum_dump)
@@ -1876,7 +2105,7 @@ static int osc_brw_fini_request(struct ptlrpc_request *req, int rc)
 
                        LCONSOLE_ERROR_MSG(0x133, "%s: BAD READ CHECKSUM: from "
                                           "%s%s%s inode "DFID" object "DOSTID
-                                          " extent [%llu-%llu], client %x, "
+                                          " extent [%llu-%llu], client %x/%x, "
                                           "server %x, cksum_type %x\n",
                                           obd_name,
                                           libcfs_nid2str(peer->nid),
@@ -1891,8 +2120,8 @@ static int osc_brw_fini_request(struct ptlrpc_request *req, int rc)
                                           aa->aa_ppga[0]->off,
                                           aa->aa_ppga[page_count-1]->off +
                                           aa->aa_ppga[page_count-1]->count - 1,
-                                          client_cksum, server_cksum,
-                                          cksum_type);
+                                          client_cksum, client_cksum2,
+                                          server_cksum, cksum_type);
                        cksum_counter = 0;
                        aa->aa_oa->o_cksum = client_cksum;
                        rc = -EAGAIN;
@@ -1912,6 +2141,83 @@ static int osc_brw_fini_request(struct ptlrpc_request *req, int rc)
        } else {
                rc = 0;
        }
+
+       inode = page2inode(aa->aa_ppga[0]->pg);
+       if (inode == NULL) {
+               /* Try to get reference to inode from cl_page if we are
+                * dealing with direct IO, as handled pages are not
+                * actual page cache pages.
+                */
+               struct osc_async_page *oap = brw_page2oap(aa->aa_ppga[0]);
+
+               inode = oap2cl_page(oap)->cp_inode;
+               if (inode) {
+                       blockbits = inode->i_blkbits;
+                       blocksize = 1 << blockbits;
+               }
+       }
+       if (inode && IS_ENCRYPTED(inode)) {
+               int idx;
+
+               if (!llcrypt_has_encryption_key(inode)) {
+                       CDEBUG(D_SEC, "no enc key for ino %lu\n", inode->i_ino);
+                       GOTO(out, rc);
+               }
+               for (idx = 0; idx < aa->aa_page_count; idx++) {
+                       struct brw_page *pg = aa->aa_ppga[idx];
+                       unsigned int offs = 0;
+
+                       while (offs < PAGE_SIZE) {
+                               /* do not decrypt if page is all 0s */
+                               if (memchr_inv(page_address(pg->pg) + offs, 0,
+                                        LUSTRE_ENCRYPTION_UNIT_SIZE) == NULL) {
+                                       /* if page is empty forward info to
+                                        * upper layers (ll_io_zero_page) by
+                                        * clearing PagePrivate2
+                                        */
+                                       if (!offs)
+                                               ClearPagePrivate2(pg->pg);
+                                       break;
+                               }
+
+                               if (blockbits) {
+                                       /* This is direct IO case. Directly call
+                                        * decrypt function that takes inode as
+                                        * input parameter. Page does not need
+                                        * to be locked.
+                                        */
+                                       u64 lblk_num =
+                                               ((u64)(pg->off >> PAGE_SHIFT) <<
+                                                    (PAGE_SHIFT - blockbits)) +
+                                                      (offs >> blockbits);
+                                       unsigned int i;
+
+                                       for (i = offs;
+                                            i < offs +
+                                                   LUSTRE_ENCRYPTION_UNIT_SIZE;
+                                            i += blocksize, lblk_num++) {
+                                               rc =
+                                                 llcrypt_decrypt_block_inplace(
+                                                         inode, pg->pg,
+                                                         blocksize, i,
+                                                         lblk_num);
+                                               if (rc)
+                                                       break;
+                                       }
+                               } else {
+                                       rc = llcrypt_decrypt_pagecache_blocks(
+                                               pg->pg,
+                                               LUSTRE_ENCRYPTION_UNIT_SIZE,
+                                               offs);
+                               }
+                               if (rc)
+                                       GOTO(out, rc);
+
+                               offs += LUSTRE_ENCRYPTION_UNIT_SIZE;
+                       }
+               }
+       }
+
 out:
        if (rc >= 0)
                lustre_get_wire_obdo(&req->rq_import->imp_connect_data,
@@ -1940,16 +2246,12 @@ static int osc_brw_redo_request(struct ptlrpc_request *request,
                 RETURN(rc);
 
        list_for_each_entry(oap, &aa->aa_oaps, oap_rpc_item) {
-                if (oap->oap_request != NULL) {
-                        LASSERTF(request == oap->oap_request,
-                                 "request %p != oap_request %p\n",
-                                 request, oap->oap_request);
-                        if (oap->oap_interrupted) {
-                                ptlrpc_req_finished(new_req);
-                                RETURN(-EINTR);
-                        }
-                }
-        }
+               if (oap->oap_request != NULL) {
+                       LASSERTF(request == oap->oap_request,
+                                "request %p != oap_request %p\n",
+                                request, oap->oap_request);
+               }
+       }
        /*
         * New request takes over pga and oaps from old request.
         * Note that copying a list_head doesn't work, need to move it...
@@ -2025,8 +2327,8 @@ static void sort_brw_pages(struct brw_page **array, int num)
 
 static void osc_release_ppga(struct brw_page **ppga, size_t count)
 {
-        LASSERT(ppga != NULL);
-        OBD_FREE(ppga, sizeof(*ppga) * count);
+       LASSERT(ppga != NULL);
+       OBD_FREE_PTR_ARRAY_LARGE(ppga, count);
 }
 
 static int brw_interpret(const struct lu_env *env,
@@ -2042,6 +2344,10 @@ static int brw_interpret(const struct lu_env *env,
 
        rc = osc_brw_fini_request(req, rc);
        CDEBUG(D_INODE, "request %p aa %p rc %d\n", req, aa, rc);
+
+       /* restore clear text pages */
+       osc_release_bounce_pages(aa->aa_ppga, aa->aa_page_count);
+
        /*
         * When server returns -EINPROGRESS, client should always retry
         * regardless of the number of times the bulk was resent already.
@@ -2054,7 +2360,7 @@ static int brw_interpret(const struct lu_env *env,
                               req->rq_import->imp_obd->obd_name,
                               POSTID(&aa->aa_oa->o_oi), rc);
                } else if (rc == -EINPROGRESS ||
-                   client_should_resend(aa->aa_resends, aa->aa_cli)) {
+                          client_should_resend(aa->aa_resends, aa->aa_cli)) {
                        rc = osc_brw_redo_request(req, aa, rc);
                } else {
                        CERROR("%s: too many resent retries for object: "
@@ -2129,7 +2435,7 @@ static int brw_interpret(const struct lu_env *env,
        list_for_each_entry_safe(ext, tmp, &aa->aa_exts, oe_link) {
                list_del_init(&ext->oe_link);
                osc_extent_finish(env, ext, 1,
-                                 rc && req->rq_no_delay ? -EWOULDBLOCK : rc);
+                                 rc && req->rq_no_delay ? -EAGAIN : rc);
        }
        LASSERT(list_empty(&aa->aa_exts));
        LASSERT(list_empty(&aa->aa_oaps));
@@ -2192,11 +2498,11 @@ int osc_build_rpc(const struct lu_env *env, struct client_obd *cli,
        struct cl_req_attr              *crattr = NULL;
        loff_t                          starting_offset = OBD_OBJECT_EOF;
        loff_t                          ending_offset = 0;
-       int                             mpflag = 0;
+       /* '1' for consistency with code that checks !mpflag to restore */
+       int mpflag = 1;
        int                             mem_tight = 0;
        int                             page_count = 0;
        bool                            soft_sync = false;
-       bool                            interrupted = false;
        bool                            ndelay = false;
        int                             i;
        int                             grant = 0;
@@ -2220,9 +2526,9 @@ int osc_build_rpc(const struct lu_env *env, struct client_obd *cli,
 
        soft_sync = osc_over_unstable_soft_limit(cli);
        if (mem_tight)
-               mpflag = cfs_memory_pressure_get_and_set();
+               mpflag = memalloc_noreclaim_save();
 
-       OBD_ALLOC(pga, sizeof(*pga) * page_count);
+       OBD_ALLOC_PTR_ARRAY_LARGE(pga, page_count);
        if (pga == NULL)
                GOTO(out, rc = -ENOMEM);
 
@@ -2253,15 +2559,13 @@ int osc_build_rpc(const struct lu_env *env, struct client_obd *cli,
                        else
                                LASSERT(oap->oap_page_off + oap->oap_count ==
                                        PAGE_SIZE);
-                       if (oap->oap_interrupted)
-                               interrupted = true;
                }
                if (ext->oe_ndelay)
                        ndelay = true;
        }
 
        /* first page in the list */
-       oap = list_entry(rpc_list.next, typeof(*oap), oap_rpc_item);
+       oap = list_first_entry(&rpc_list, typeof(*oap), oap_rpc_item);
 
        crattr = &osc_env_info(env)->oti_req_attr;
        memset(crattr, 0, sizeof(*crattr));
@@ -2293,8 +2597,6 @@ int osc_build_rpc(const struct lu_env *env, struct client_obd *cli,
        req->rq_interpret_reply = brw_interpret;
        req->rq_memalloc = mem_tight != 0;
        oap->oap_request = ptlrpc_request_addref(req);
-       if (interrupted && !req->rq_intr)
-               ptlrpc_mark_interrupted(req);
        if (ndelay) {
                req->rq_no_resend = req->rq_no_delay = 1;
                /* probably set a shorter timeout value.
@@ -2346,21 +2648,24 @@ int osc_build_rpc(const struct lu_env *env, struct client_obd *cli,
        EXIT;
 
 out:
-       if (mem_tight != 0)
-               cfs_memory_pressure_restore(mpflag);
+       if (mem_tight)
+               memalloc_noreclaim_restore(mpflag);
 
        if (rc != 0) {
                LASSERT(req == NULL);
 
                if (oa)
                        OBD_SLAB_FREE_PTR(oa, osc_obdo_kmem);
-               if (pga)
-                       OBD_FREE(pga, sizeof(*pga) * page_count);
+               if (pga) {
+                       osc_release_bounce_pages(pga, page_count);
+                       osc_release_ppga(pga, page_count);
+               }
                /* this should happen rarely and is pretty bad, it makes the
-                * pending list not follow the dirty order */
-               while (!list_empty(ext_list)) {
-                       ext = list_entry(ext_list->next, struct osc_extent,
-                                        oe_link);
+                * pending list not follow the dirty order
+                */
+               while ((ext = list_first_entry_or_null(ext_list,
+                                                      struct osc_extent,
+                                                      oe_link)) != NULL) {
                        list_del_init(&ext->oe_link);
                        osc_extent_finish(env, ext, 0, rc);
                }
@@ -2368,6 +2673,34 @@ out:
        RETURN(rc);
 }
 
+/* This is to refresh our lock in face of no RPCs. */
+void osc_send_empty_rpc(struct osc_object *osc, pgoff_t start)
+{
+       struct ptlrpc_request *req;
+       struct obdo oa;
+       struct brw_page bpg = { .off = start, .count = 1};
+       struct brw_page *pga = &bpg;
+       int rc;
+
+       memset(&oa, 0, sizeof(oa));
+       oa.o_oi = osc->oo_oinfo->loi_oi;
+       oa.o_valid = OBD_MD_FLID | OBD_MD_FLGROUP | OBD_MD_FLFLAGS;
+       /* For updated servers - don't do a read */
+       oa.o_flags = OBD_FL_NORPC;
+
+       rc = osc_brw_prep_request(OBD_BRW_READ, osc_cli(osc), &oa, 1, &pga,
+                                 &req, 0);
+
+       /* If we succeeded we ship it off, if not there's no point in doing
+        * anything. Also no resends.
+        * No interpret callback, no commit callback.
+        */
+       if (!rc) {
+               req->rq_no_resend = 1;
+               ptlrpcd_add_req(req);
+       }
+}
+
 static int osc_set_lock_data(struct ldlm_lock *lock, void *data)
 {
         int set = 0;
@@ -2434,6 +2767,10 @@ int osc_enqueue_interpret(const struct lu_env *env, struct ptlrpc_request *req,
        struct ost_lvb *lvb = aa->oa_lvb;
        __u32 lvb_len = sizeof(*lvb);
        __u64 flags = 0;
+       struct ldlm_enqueue_info einfo = {
+               .ei_type = aa->oa_type,
+               .ei_mode = mode,
+       };
 
        ENTRY;
 
@@ -2463,9 +2800,8 @@ int osc_enqueue_interpret(const struct lu_env *env, struct ptlrpc_request *req,
        }
 
        /* Complete obtaining the lock procedure. */
-       rc = ldlm_cli_enqueue_fini(aa->oa_exp, req, aa->oa_type, 1,
-                                  aa->oa_mode, aa->oa_flags, lvb, lvb_len,
-                                  lockh, rc);
+       rc = ldlm_cli_enqueue_fini(aa->oa_exp, req, &einfo, 1, aa->oa_flags,
+                                  lvb, lvb_len, lockh, rc);
        /* Complete osc stuff. */
        rc = osc_enqueue_fini(req, aa->oa_upcall, aa->oa_cookie, lockh, mode,
                              aa->oa_flags, aa->oa_speculative, rc);
@@ -2477,8 +2813,6 @@ int osc_enqueue_interpret(const struct lu_env *env, struct ptlrpc_request *req,
        RETURN(rc);
 }
 
-struct ptlrpc_request_set *PTLRPCD_SET = (void *)1;
-
 /* When enqueuing asynchronously, locks are not ordered, we can obtain a lock
  * from the 2nd OSC before a lock from the 1st one. This does not deadlock with
  * other synchronous requests, however keeping some locks and trying to obtain
@@ -2530,7 +2864,7 @@ int osc_enqueue_base(struct obd_export *exp, struct ldlm_res_id *res_id,
        if (intent != 0)
                match_flags |= LDLM_FL_BLOCK_GRANTED;
        mode = ldlm_lock_match(obd->obd_namespace, match_flags, res_id,
-                              einfo->ei_type, policy, mode, &lockh, 0);
+                              einfo->ei_type, policy, mode, &lockh);
        if (mode) {
                struct ldlm_lock *matched;
 
@@ -2572,23 +2906,6 @@ int osc_enqueue_base(struct obd_export *exp, struct ldlm_res_id *res_id,
        if (*flags & (LDLM_FL_TEST_LOCK | LDLM_FL_MATCH_LOCK))
                RETURN(-ENOLCK);
 
-       if (intent) {
-               req = ptlrpc_request_alloc(class_exp2cliimp(exp),
-                                          &RQF_LDLM_ENQUEUE_LVB);
-               if (req == NULL)
-                       RETURN(-ENOMEM);
-
-               rc = ldlm_prep_enqueue_req(exp, req, NULL, 0);
-               if (rc) {
-                        ptlrpc_request_free(req);
-                        RETURN(rc);
-                }
-
-                req_capsule_set_size(&req->rq_pill, &RMF_DLM_LVB, RCL_SERVER,
-                                     sizeof *lvb);
-                ptlrpc_request_set_replen(req);
-        }
-
         /* users of osc_enqueue() can pass this flag for ldlm_lock_match() */
         *flags &= ~LDLM_FL_BLOCK_GRANTED;
 
@@ -2617,20 +2934,13 @@ int osc_enqueue_base(struct obd_export *exp, struct ldlm_res_id *res_id,
                        }
 
                        req->rq_interpret_reply = osc_enqueue_interpret;
-                       if (rqset == PTLRPCD_SET)
-                               ptlrpcd_add_req(req);
-                       else
-                               ptlrpc_set_add_req(rqset, req);
-               } else if (intent) {
-                       ptlrpc_req_finished(req);
+                       ptlrpc_set_add_req(rqset, req);
                }
                RETURN(rc);
        }
 
        rc = osc_enqueue_fini(req, upcall, cookie, &lockh, einfo->ei_mode,
                              flags, speculative, rc);
-       if (intent)
-               ptlrpc_req_finished(req);
 
        RETURN(rc);
 }
@@ -2639,7 +2949,7 @@ int osc_match_base(const struct lu_env *env, struct obd_export *exp,
                   struct ldlm_res_id *res_id, enum ldlm_type type,
                   union ldlm_policy_data *policy, enum ldlm_mode mode,
                   __u64 *flags, struct osc_object *obj,
-                  struct lustre_handle *lockh, int unref)
+                  struct lustre_handle *lockh, enum ldlm_match_flags match_flags)
 {
        struct obd_device *obd = exp->exp_obd;
        __u64 lflags = *flags;
@@ -2654,15 +2964,10 @@ int osc_match_base(const struct lu_env *env, struct obd_export *exp,
        policy->l_extent.start -= policy->l_extent.start & ~PAGE_MASK;
        policy->l_extent.end |= ~PAGE_MASK;
 
-        /* Next, search for already existing extent locks that will cover us */
-        /* If we're trying to read, we also search for an existing PW lock.  The
-         * VFS and page cache already protect us locally, so lots of readers/
-         * writers can share a single PW lock. */
-        rc = mode;
-        if (mode == LCK_PR)
-                rc |= LCK_PW;
-        rc = ldlm_lock_match(obd->obd_namespace, lflags,
-                             res_id, type, policy, rc, lockh, unref);
+       /* Next, search for already existing extent locks that will cover us */
+       rc = ldlm_lock_match_with_skip(obd->obd_namespace, lflags, 0,
+                                       res_id, type, policy, mode, lockh,
+                                       match_flags);
        if (rc == 0 || lflags & LDLM_FL_TEST_LOCK)
                RETURN(rc);
 
@@ -2785,19 +3090,17 @@ static int osc_statfs(const struct lu_env *env, struct obd_export *exp,
        struct obd_device     *obd = class_exp2obd(exp);
        struct obd_statfs     *msfs;
        struct ptlrpc_request *req;
-       struct obd_import     *imp = NULL;
+       struct obd_import     *imp, *imp0;
        int rc;
        ENTRY;
 
-
-        /*Since the request might also come from lprocfs, so we need
-         *sync this with client_disconnect_export Bug15684*/
-       down_read(&obd->u.cli.cl_sem);
-        if (obd->u.cli.cl_import)
-                imp = class_import_get(obd->u.cli.cl_import);
-       up_read(&obd->u.cli.cl_sem);
-        if (!imp)
-                RETURN(-ENODEV);
+       /*Since the request might also come from lprocfs, so we need
+        *sync this with client_disconnect_export Bug15684
+        */
+       with_imp_locked(obd, imp0, rc)
+               imp = class_import_get(imp0);
+       if (rc)
+               RETURN(rc);
 
        /* We could possibly pass max_age in the request (as an absolute
         * timestamp or a "seconds.usec ago") so the target can avoid doing
@@ -2870,7 +3173,7 @@ static int osc_iocontrol(unsigned int cmd, struct obd_export *exp, int len,
        default:
                rc = -ENOTTY;
                CDEBUG(D_INODE, "%s: unrecognised ioctl %#x by %s: rc = %d\n",
-                      obd->obd_name, cmd, current_comm(), rc);
+                      obd->obd_name, cmd, current->comm, rc);
                break;
        }
 
@@ -3367,21 +3670,28 @@ static const struct obd_ops osc_obd_ops = {
         .o_quotactl             = osc_quotactl,
 };
 
-static struct shrinker *osc_cache_shrinker;
 LIST_HEAD(osc_shrink_list);
 DEFINE_SPINLOCK(osc_shrink_lock);
 
-#ifndef HAVE_SHRINKER_COUNT
-static int osc_cache_shrink(SHRINKER_ARGS(sc, nr_to_scan, gfp_mask))
+#ifdef HAVE_SHRINKER_COUNT
+static struct shrinker osc_cache_shrinker = {
+       .count_objects  = osc_cache_shrink_count,
+       .scan_objects   = osc_cache_shrink_scan,
+       .seeks          = DEFAULT_SEEKS,
+};
+#else
+static int osc_cache_shrink(struct shrinker *shrinker,
+                           struct shrink_control *sc)
 {
-       struct shrink_control scv = {
-               .nr_to_scan = shrink_param(sc, nr_to_scan),
-               .gfp_mask   = shrink_param(sc, gfp_mask)
-       };
-       (void)osc_cache_shrink_scan(shrinker, &scv);
+       (void)osc_cache_shrink_scan(shrinker, sc);
 
-       return osc_cache_shrink_count(shrinker, &scv);
+       return osc_cache_shrink_count(shrinker, sc);
 }
+
+static struct shrinker osc_cache_shrinker = {
+       .shrink   = osc_cache_shrink,
+       .seeks    = DEFAULT_SEEKS,
+};
 #endif
 
 static int __init osc_init(void)
@@ -3389,8 +3699,6 @@ static int __init osc_init(void)
        unsigned int reqpool_size;
        unsigned int reqsize;
        int rc;
-       DEF_SHRINKER_VAR(osc_shvar, osc_cache_shrink,
-                        osc_cache_shrink_count, osc_cache_shrink_scan);
        ENTRY;
 
        /* print an address of _any_ initialized kernel symbol from this
@@ -3402,16 +3710,18 @@ static int __init osc_init(void)
        if (rc)
                RETURN(rc);
 
-       rc = class_register_type(&osc_obd_ops, NULL, true, NULL,
+       rc = class_register_type(&osc_obd_ops, NULL, true,
                                 LUSTRE_OSC_NAME, &osc_device_type);
        if (rc)
                GOTO(out_kmem, rc);
 
-       osc_cache_shrinker = set_shrinker(DEFAULT_SEEKS, &osc_shvar);
+       rc = register_shrinker(&osc_cache_shrinker);
+       if (rc)
+               GOTO(out_type, rc);
 
        /* This is obviously too much memory, only prevent overflow here */
        if (osc_reqpool_mem_max >= 1 << 12 || osc_reqpool_mem_max == 0)
-               GOTO(out_type, rc = -EINVAL);
+               GOTO(out_shrinker, rc = -EINVAL);
 
        reqpool_size = osc_reqpool_mem_max << 20;
 
@@ -3432,7 +3742,7 @@ static int __init osc_init(void)
                                          ptlrpc_add_rqs_to_pool);
 
        if (osc_rq_pool == NULL)
-               GOTO(out_type, rc = -ENOMEM);
+               GOTO(out_shrinker, rc = -ENOMEM);
 
        rc = osc_start_grant_work();
        if (rc != 0)
@@ -3442,6 +3752,8 @@ static int __init osc_init(void)
 
 out_req_pool:
        ptlrpc_free_rq_pool(osc_rq_pool);
+out_shrinker:
+       unregister_shrinker(&osc_cache_shrinker);
 out_type:
        class_unregister_type(LUSTRE_OSC_NAME);
 out_kmem:
@@ -3453,7 +3765,7 @@ out_kmem:
 static void __exit osc_exit(void)
 {
        osc_stop_grant_work();
-       remove_shrinker(osc_cache_shrinker);
+       unregister_shrinker(&osc_cache_shrinker);
        class_unregister_type(LUSTRE_OSC_NAME);
        lu_kmem_fini(osc_caches);
        ptlrpc_free_rq_pool(osc_rq_pool);