From: Shyam Prasad N <nspmangalore@gmail.com>
To: Rohith Surabattula <rohiths.msft@gmail.com>
Cc: David Howells <dhowells@redhat.com>,
Steve French <smfrench@gmail.com>,
jlayton@kernel.org, CIFS <linux-cifs@vger.kernel.org>,
linux-cachefs@redhat.com, linux-fsdevel@vger.kernel.org
Subject: Re: [RFC PATCH 3/7] cifs: Change the I/O paths to use an iterator rather than a page list
Date: Mon, 31 Jan 2022 11:18:07 +0530 [thread overview]
Message-ID: <CANT5p=o-MYJY6kmkgw5Y_e7zdWjEkNZ1f_0aAKKeCDHjEFLfvw@mail.gmail.com> (raw)
In-Reply-To: <CACdtm0bi4O36cif-iwarBb2oNOj-qjECr0iPAHK821E07u7p8A@mail.gmail.com>
Looks good to me.
Reviewed-by: Shyam Prasad N <sprasad@microsoft.com>
On Mon, Jan 31, 2022 at 10:36 AM Rohith Surabattula
<rohiths.msft@gmail.com> wrote:
>
> Hi David,
>
> After copying the buf to the XArray iterator, "got_bytes" field is not
> updated. As a result, the read of data which is less than page size
> failed.
> Below is the patch to fix the above issue.
>
> diff --git a/fs/cifs/smb2ops.c b/fs/cifs/smb2ops.c
> index e1649ac194db..5faf45672891 100644
> --- a/fs/cifs/smb2ops.c
> +++ b/fs/cifs/smb2ops.c
> @@ -4917,6 +4917,7 @@ handle_read_data(struct TCP_Server_Info *server,
> struct mid_q_entry *mid,
> length = copy_to_iter(buf + data_offset, data_len,
> &rdata->iter);
> if (length < 0)
> return length;
> + rdata->got_bytes = data_len;
> } else {
> /* read response payload cannot be in both buf and pages */
> WARN_ONCE(1, "buf can not contain only a part of read data");
>
> Regards,
> Rohith
>
> On Wed, Jan 26, 2022 at 1:21 AM David Howells <dhowells@redhat.com> wrote:
> >
> >
> > ---
> >
> > fs/cifs/cifsencrypt.c | 40 +++--
> > fs/cifs/cifsfs.c | 2
> > fs/cifs/cifsfs.h | 3
> > fs/cifs/cifsglob.h | 28 +---
> > fs/cifs/cifsproto.h | 10 +
> > fs/cifs/cifssmb.c | 224 +++++++++++++++++++-----------
> > fs/cifs/connect.c | 16 ++
> > fs/cifs/misc.c | 19 ---
> > fs/cifs/smb2ops.c | 365 ++++++++++++++++++++++++-------------------------
> > fs/cifs/smb2pdu.c | 12 --
> > fs/cifs/transport.c | 37 +----
> > 11 files changed, 379 insertions(+), 377 deletions(-)
> >
> > diff --git a/fs/cifs/cifsencrypt.c b/fs/cifs/cifsencrypt.c
> > index 0912d8bbbac1..69bbf3d6c4d4 100644
> > --- a/fs/cifs/cifsencrypt.c
> > +++ b/fs/cifs/cifsencrypt.c
> > @@ -24,12 +24,27 @@
> > #include "../smbfs_common/arc4.h"
> > #include <crypto/aead.h>
> >
> > +static ssize_t cifs_signature_scan(struct iov_iter *i, const void *p,
> > + size_t len, size_t off, void *priv)
> > +{
> > + struct shash_desc *shash = priv;
> > + int rc;
> > +
> > + rc = crypto_shash_update(shash, p, len);
> > + if (rc) {
> > + cifs_dbg(VFS, "%s: Could not update with payload\n", __func__);
> > + return rc;
> > + }
> > +
> > + return len;
> > +}
> > +
> > int __cifs_calc_signature(struct smb_rqst *rqst,
> > struct TCP_Server_Info *server, char *signature,
> > struct shash_desc *shash)
> > {
> > int i;
> > - int rc;
> > + ssize_t rc;
> > struct kvec *iov = rqst->rq_iov;
> > int n_vec = rqst->rq_nvec;
> > int is_smb2 = server->vals->header_preamble_size == 0;
> > @@ -62,25 +77,10 @@ int __cifs_calc_signature(struct smb_rqst *rqst,
> > }
> > }
> >
> > - /* now hash over the rq_pages array */
> > - for (i = 0; i < rqst->rq_npages; i++) {
> > - void *kaddr;
> > - unsigned int len, offset;
> > -
> > - rqst_page_get_length(rqst, i, &len, &offset);
> > -
> > - kaddr = (char *) kmap(rqst->rq_pages[i]) + offset;
> > -
> > - rc = crypto_shash_update(shash, kaddr, len);
> > - if (rc) {
> > - cifs_dbg(VFS, "%s: Could not update with payload\n",
> > - __func__);
> > - kunmap(rqst->rq_pages[i]);
> > - return rc;
> > - }
> > -
> > - kunmap(rqst->rq_pages[i]);
> > - }
> > + rc = iov_iter_scan(&rqst->rq_iter, iov_iter_count(&rqst->rq_iter),
> > + cifs_signature_scan, shash);
> > + if (rc < 0)
> > + return rc;
> >
> > rc = crypto_shash_final(shash, signature);
> > if (rc)
> > diff --git a/fs/cifs/cifsfs.c b/fs/cifs/cifsfs.c
> > index 199edac0cb59..a56cb9c8c5ff 100644
> > --- a/fs/cifs/cifsfs.c
> > +++ b/fs/cifs/cifsfs.c
> > @@ -935,7 +935,7 @@ cifs_loose_read_iter(struct kiocb *iocb, struct iov_iter *iter)
> > ssize_t rc;
> > struct inode *inode = file_inode(iocb->ki_filp);
> >
> > - if (iocb->ki_filp->f_flags & O_DIRECT)
> > + if (iocb->ki_flags & IOCB_DIRECT)
> > return cifs_user_readv(iocb, iter);
> >
> > rc = cifs_revalidate_mapping(inode);
> > diff --git a/fs/cifs/cifsfs.h b/fs/cifs/cifsfs.h
> > index 15a5c5db038b..1c77bbc0815f 100644
> > --- a/fs/cifs/cifsfs.h
> > +++ b/fs/cifs/cifsfs.h
> > @@ -110,6 +110,9 @@ extern int cifs_file_strict_mmap(struct file * , struct vm_area_struct *);
> > extern const struct file_operations cifs_dir_ops;
> > extern int cifs_dir_open(struct inode *inode, struct file *file);
> > extern int cifs_readdir(struct file *file, struct dir_context *ctx);
> > +extern void cifs_pages_written_back(struct inode *inode, loff_t start, unsigned int len);
> > +extern void cifs_pages_write_failed(struct inode *inode, loff_t start, unsigned int len);
> > +extern void cifs_pages_write_redirty(struct inode *inode, loff_t start, unsigned int len);
> >
> > /* Functions related to dir entries */
> > extern const struct dentry_operations cifs_dentry_ops;
> > diff --git a/fs/cifs/cifsglob.h b/fs/cifs/cifsglob.h
> > index 0a4085ced40f..3a4fed645636 100644
> > --- a/fs/cifs/cifsglob.h
> > +++ b/fs/cifs/cifsglob.h
> > @@ -191,11 +191,8 @@ struct cifs_cred {
> > struct smb_rqst {
> > struct kvec *rq_iov; /* array of kvecs */
> > unsigned int rq_nvec; /* number of kvecs in array */
> > - struct page **rq_pages; /* pointer to array of page ptrs */
> > - unsigned int rq_offset; /* the offset to the 1st page */
> > - unsigned int rq_npages; /* number pages in array */
> > - unsigned int rq_pagesz; /* page size to use */
> > - unsigned int rq_tailsz; /* length of last page */
> > + struct iov_iter rq_iter; /* Data iterator */
> > + struct xarray rq_buffer; /* Page buffer for encryption */
> > };
> >
> > struct mid_q_entry;
> > @@ -1323,28 +1320,18 @@ struct cifs_readdata {
> > struct address_space *mapping;
> > struct cifs_aio_ctx *ctx;
> > __u64 offset;
> > + ssize_t got_bytes;
> > unsigned int bytes;
> > - unsigned int got_bytes;
> > pid_t pid;
> > int result;
> > struct work_struct work;
> > - int (*read_into_pages)(struct TCP_Server_Info *server,
> > - struct cifs_readdata *rdata,
> > - unsigned int len);
> > - int (*copy_into_pages)(struct TCP_Server_Info *server,
> > - struct cifs_readdata *rdata,
> > - struct iov_iter *iter);
> > + struct iov_iter iter;
> > struct kvec iov[2];
> > struct TCP_Server_Info *server;
> > #ifdef CONFIG_CIFS_SMB_DIRECT
> > struct smbd_mr *mr;
> > #endif
> > - unsigned int pagesz;
> > - unsigned int page_offset;
> > - unsigned int tailsz;
> > struct cifs_credits credits;
> > - unsigned int nr_pages;
> > - struct page **pages;
> > };
> >
> > /* asynchronous write support */
> > @@ -1356,6 +1343,8 @@ struct cifs_writedata {
> > struct work_struct work;
> > struct cifsFileInfo *cfile;
> > struct cifs_aio_ctx *ctx;
> > + struct iov_iter iter;
> > + struct bio_vec *bv;
> > __u64 offset;
> > pid_t pid;
> > unsigned int bytes;
> > @@ -1364,12 +1353,7 @@ struct cifs_writedata {
> > #ifdef CONFIG_CIFS_SMB_DIRECT
> > struct smbd_mr *mr;
> > #endif
> > - unsigned int pagesz;
> > - unsigned int page_offset;
> > - unsigned int tailsz;
> > struct cifs_credits credits;
> > - unsigned int nr_pages;
> > - struct page **pages;
> > };
> >
> > /*
> > diff --git a/fs/cifs/cifsproto.h b/fs/cifs/cifsproto.h
> > index d3701295402d..1b143f0a03c0 100644
> > --- a/fs/cifs/cifsproto.h
> > +++ b/fs/cifs/cifsproto.h
> > @@ -242,6 +242,9 @@ extern int cifs_read_page_from_socket(struct TCP_Server_Info *server,
> > unsigned int page_offset,
> > unsigned int to_read);
> > extern int cifs_setup_cifs_sb(struct cifs_sb_info *cifs_sb);
> > +extern int cifs_read_iter_from_socket(struct TCP_Server_Info *server,
> > + struct iov_iter *iter,
> > + unsigned int to_read);
> > extern int cifs_match_super(struct super_block *, void *);
> > extern int cifs_mount(struct cifs_sb_info *cifs_sb, struct smb3_fs_context *ctx);
> > extern void cifs_umount(struct cifs_sb_info *);
> > @@ -575,10 +578,7 @@ int cifs_readv_receive(struct TCP_Server_Info *server, struct mid_q_entry *mid);
> > int cifs_async_writev(struct cifs_writedata *wdata,
> > void (*release)(struct kref *kref));
> > void cifs_writev_complete(struct work_struct *work);
> > -struct cifs_writedata *cifs_writedata_alloc(unsigned int nr_pages,
> > - work_func_t complete);
> > -struct cifs_writedata *cifs_writedata_direct_alloc(struct page **pages,
> > - work_func_t complete);
> > +struct cifs_writedata *cifs_writedata_alloc(work_func_t complete);
> > void cifs_writedata_release(struct kref *refcount);
> > int cifs_query_mf_symlink(unsigned int xid, struct cifs_tcon *tcon,
> > struct cifs_sb_info *cifs_sb,
> > @@ -602,8 +602,6 @@ int cifs_alloc_hash(const char *name, struct crypto_shash **shash,
> > struct sdesc **sdesc);
> > void cifs_free_hash(struct crypto_shash **shash, struct sdesc **sdesc);
> >
> > -extern void rqst_page_get_length(struct smb_rqst *rqst, unsigned int page,
> > - unsigned int *len, unsigned int *offset);
> > struct cifs_chan *
> > cifs_ses_find_chan(struct cifs_ses *ses, struct TCP_Server_Info *server);
> > int cifs_try_adding_channels(struct cifs_sb_info *cifs_sb, struct cifs_ses *ses);
> > diff --git a/fs/cifs/cifssmb.c b/fs/cifs/cifssmb.c
> > index 071e2f21a7db..38e7276352e2 100644
> > --- a/fs/cifs/cifssmb.c
> > +++ b/fs/cifs/cifssmb.c
> > @@ -24,6 +24,7 @@
> > #include <linux/task_io_accounting_ops.h>
> > #include <linux/uaccess.h>
> > #include "cifspdu.h"
> > +#include "cifsfs.h"
> > #include "cifsglob.h"
> > #include "cifsacl.h"
> > #include "cifsproto.h"
> > @@ -1388,11 +1389,11 @@ int
> > cifs_discard_remaining_data(struct TCP_Server_Info *server)
> > {
> > unsigned int rfclen = server->pdu_size;
> > - int remaining = rfclen + server->vals->header_preamble_size -
> > + size_t remaining = rfclen + server->vals->header_preamble_size -
> > server->total_read;
> >
> > while (remaining > 0) {
> > - int length;
> > + ssize_t length;
> >
> > length = cifs_discard_from_socket(server,
> > min_t(size_t, remaining,
> > @@ -1539,10 +1540,15 @@ cifs_readv_receive(struct TCP_Server_Info *server, struct mid_q_entry *mid)
> > return cifs_readv_discard(server, mid);
> > }
> >
> > - length = rdata->read_into_pages(server, rdata, data_len);
> > - if (length < 0)
> > - return length;
> > -
> > +#ifdef CONFIG_CIFS_SMB_DIRECT
> > + if (rdata->mr)
> > + length = data_len; /* An RDMA read is already done. */
> > + else
> > +#endif
> > + length = cifs_read_iter_from_socket(server, &rdata->iter,
> > + data_len);
> > + if (length > 0)
> > + rdata->got_bytes += length;
> > server->total_read += length;
> >
> > cifs_dbg(FYI, "total_read=%u buflen=%u remaining=%u\n",
> > @@ -1566,11 +1572,7 @@ cifs_readv_callback(struct mid_q_entry *mid)
> > struct TCP_Server_Info *server = tcon->ses->server;
> > struct smb_rqst rqst = { .rq_iov = rdata->iov,
> > .rq_nvec = 2,
> > - .rq_pages = rdata->pages,
> > - .rq_offset = rdata->page_offset,
> > - .rq_npages = rdata->nr_pages,
> > - .rq_pagesz = rdata->pagesz,
> > - .rq_tailsz = rdata->tailsz };
> > + .rq_iter = rdata->iter };
> > struct cifs_credits credits = { .value = 1, .instance = 0 };
> >
> > cifs_dbg(FYI, "%s: mid=%llu state=%d result=%d bytes=%u\n",
> > @@ -1925,10 +1927,93 @@ cifs_writedata_release(struct kref *refcount)
> > if (wdata->cfile)
> > cifsFileInfo_put(wdata->cfile);
> >
> > - kvfree(wdata->pages);
> > kfree(wdata);
> > }
> >
> > +/*
> > + * Completion of write to server.
> > + */
> > +void cifs_pages_written_back(struct inode *inode, loff_t start, unsigned int len)
> > +{
> > + struct address_space *mapping = inode->i_mapping;
> > + struct folio *folio;
> > + pgoff_t end;
> > +
> > + XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE);
> > +
> > + rcu_read_lock();
> > +
> > + end = (start + len - 1) / PAGE_SIZE;
> > + xas_for_each(&xas, folio, end) {
> > + if (!folio_test_writeback(folio)) {
> > + pr_err("bad %x @%llx page %lx %lx\n",
> > + len, start, folio_index(folio), end);
> > + BUG();
> > + }
> > +
> > + folio_detach_private(folio);
> > + folio_end_writeback(folio);
> > + }
> > +
> > + rcu_read_unlock();
> > +}
> > +
> > +/*
> > + * Failure of write to server.
> > + */
> > +void cifs_pages_write_failed(struct inode *inode, loff_t start, unsigned int len)
> > +{
> > + struct address_space *mapping = inode->i_mapping;
> > + struct folio *folio;
> > + pgoff_t end;
> > +
> > + XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE);
> > +
> > + rcu_read_lock();
> > +
> > + end = (start + len - 1) / PAGE_SIZE;
> > + xas_for_each(&xas, folio, end) {
> > + if (!folio_test_writeback(folio)) {
> > + pr_err("bad %x @%llx page %lx %lx\n",
> > + len, start, folio_index(folio), end);
> > + BUG();
> > + }
> > +
> > + folio_set_error(folio);
> > + folio_end_writeback(folio);
> > + }
> > +
> > + rcu_read_unlock();
> > +}
> > +
> > +/*
> > + * Redirty pages after a temporary failure.
> > + */
> > +void cifs_pages_write_redirty(struct inode *inode, loff_t start, unsigned int len)
> > +{
> > + struct address_space *mapping = inode->i_mapping;
> > + struct folio *folio;
> > + pgoff_t end;
> > +
> > + XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE);
> > +
> > + rcu_read_lock();
> > +
> > + end = (start + len - 1) / PAGE_SIZE;
> > + xas_for_each(&xas, folio, end) {
> > + if (!folio_test_writeback(folio)) {
> > + pr_err("bad %x @%llx page %lx %lx\n",
> > + len, start, folio_index(folio), end);
> > + BUG();
> > + }
> > +
> > + filemap_dirty_folio(folio->mapping, folio);
> > + folio_end_writeback(folio);
> > + }
> > +
> > + rcu_read_unlock();
> > +}
> > +
> > /*
> > * Write failed with a retryable error. Resend the write request. It's also
> > * possible that the page was redirtied so re-clean the page.
> > @@ -1936,51 +2021,56 @@ cifs_writedata_release(struct kref *refcount)
> > static void
> > cifs_writev_requeue(struct cifs_writedata *wdata)
> > {
> > - int i, rc = 0;
> > + int rc = 0;
> > struct inode *inode = d_inode(wdata->cfile->dentry);
> > struct TCP_Server_Info *server;
> > - unsigned int rest_len;
> > + unsigned int rest_len = wdata->bytes;
> > + loff_t fpos = wdata->offset;
> >
> > server = tlink_tcon(wdata->cfile->tlink)->ses->server;
> > - i = 0;
> > - rest_len = wdata->bytes;
> > do {
> > struct cifs_writedata *wdata2;
> > - unsigned int j, nr_pages, wsize, tailsz, cur_len;
> > + unsigned int wsize, cur_len;
> >
> > wsize = server->ops->wp_retry_size(inode);
> > if (wsize < rest_len) {
> > - nr_pages = wsize / PAGE_SIZE;
> > - if (!nr_pages) {
> > + if (wsize < PAGE_SIZE) {
> > rc = -ENOTSUPP;
> > break;
> > }
> > - cur_len = nr_pages * PAGE_SIZE;
> > - tailsz = PAGE_SIZE;
> > + cur_len = min(round_down(wsize, PAGE_SIZE), rest_len);
> > } else {
> > - nr_pages = DIV_ROUND_UP(rest_len, PAGE_SIZE);
> > cur_len = rest_len;
> > - tailsz = rest_len - (nr_pages - 1) * PAGE_SIZE;
> > }
> >
> > - wdata2 = cifs_writedata_alloc(nr_pages, cifs_writev_complete);
> > + wdata2 = cifs_writedata_alloc(cifs_writev_complete);
> > if (!wdata2) {
> > rc = -ENOMEM;
> > break;
> > }
> >
> > - for (j = 0; j < nr_pages; j++) {
> > - wdata2->pages[j] = wdata->pages[i + j];
> > - lock_page(wdata2->pages[j]);
> > - clear_page_dirty_for_io(wdata2->pages[j]);
> > - }
> > -
> > wdata2->sync_mode = wdata->sync_mode;
> > - wdata2->nr_pages = nr_pages;
> > - wdata2->offset = page_offset(wdata2->pages[0]);
> > - wdata2->pagesz = PAGE_SIZE;
> > - wdata2->tailsz = tailsz;
> > - wdata2->bytes = cur_len;
> > + wdata2->offset = fpos;
> > + wdata2->bytes = cur_len;
> > + wdata2->iter = wdata->iter;
> > +
> > + iov_iter_advance(&wdata2->iter, fpos - wdata->offset);
> > + iov_iter_truncate(&wdata2->iter, wdata2->bytes);
> > +
> > +#if 0
> > + if (iov_iter_is_xarray(&wdata2->iter)) {
> > + /* TODO: Check for pages having been redirtied and
> > + * clean them. We can do this by walking the xarray.
> > + * If it's not an xarray, then it's a DIO and we
> > + * shouldn't be mucking around with the page bits.
> > + */
> > + for (j = 0; j < nr_pages; j++) {
> > + wdata2->pages[j] = wdata->pages[i + j];
> > + lock_page(wdata2->pages[j]);
> > + clear_page_dirty_for_io(wdata2->pages[j]);
> > + }
> > + }
> > +#endif
> >
> > rc = cifs_get_writable_file(CIFS_I(inode), FIND_WR_ANY,
> > &wdata2->cfile);
> > @@ -1995,33 +2085,25 @@ cifs_writev_requeue(struct cifs_writedata *wdata)
> > cifs_writedata_release);
> > }
> >
> > - for (j = 0; j < nr_pages; j++) {
> > - unlock_page(wdata2->pages[j]);
> > - if (rc != 0 && !is_retryable_error(rc)) {
> > - SetPageError(wdata2->pages[j]);
> > - end_page_writeback(wdata2->pages[j]);
> > - put_page(wdata2->pages[j]);
> > - }
> > - }
> > + if (iov_iter_is_xarray(&wdata2->iter))
> > + cifs_pages_written_back(inode, wdata2->offset, wdata2->bytes);
> >
> > kref_put(&wdata2->refcount, cifs_writedata_release);
> > if (rc) {
> > if (is_retryable_error(rc))
> > continue;
> > - i += nr_pages;
> > + fpos += cur_len;
> > + rest_len -= cur_len;
> > break;
> > }
> >
> > + fpos += cur_len;
> > rest_len -= cur_len;
> > - i += nr_pages;
> > - } while (i < wdata->nr_pages);
> > + } while (rest_len > 0);
> >
> > - /* cleanup remaining pages from the original wdata */
> > - for (; i < wdata->nr_pages; i++) {
> > - SetPageError(wdata->pages[i]);
> > - end_page_writeback(wdata->pages[i]);
> > - put_page(wdata->pages[i]);
> > - }
> > + /* Clean up remaining pages from the original wdata */
> > + if (iov_iter_is_xarray(&wdata->iter))
> > + cifs_pages_written_back(inode, fpos, rest_len);
> >
> > if (rc != 0 && !is_retryable_error(rc))
> > mapping_set_error(inode->i_mapping, rc);
> > @@ -2034,7 +2116,6 @@ cifs_writev_complete(struct work_struct *work)
> > struct cifs_writedata *wdata = container_of(work,
> > struct cifs_writedata, work);
> > struct inode *inode = d_inode(wdata->cfile->dentry);
> > - int i = 0;
> >
> > if (wdata->result == 0) {
> > spin_lock(&inode->i_lock);
> > @@ -2045,40 +2126,25 @@ cifs_writev_complete(struct work_struct *work)
> > } else if (wdata->sync_mode == WB_SYNC_ALL && wdata->result == -EAGAIN)
> > return cifs_writev_requeue(wdata);
> >
> > - for (i = 0; i < wdata->nr_pages; i++) {
> > - struct page *page = wdata->pages[i];
> > - if (wdata->result == -EAGAIN)
> > - __set_page_dirty_nobuffers(page);
> > - else if (wdata->result < 0)
> > - SetPageError(page);
> > - end_page_writeback(page);
> > - cifs_readpage_to_fscache(inode, page);
> > - put_page(page);
> > - }
> > + if (wdata->result == -EAGAIN)
> > + cifs_pages_write_redirty(inode, wdata->offset, wdata->bytes);
> > + else if (wdata->result < 0)
> > + cifs_pages_write_failed(inode, wdata->offset, wdata->bytes);
> > + else
> > + cifs_pages_written_back(inode, wdata->offset, wdata->bytes);
> > +
> > if (wdata->result != -EAGAIN)
> > mapping_set_error(inode->i_mapping, wdata->result);
> > kref_put(&wdata->refcount, cifs_writedata_release);
> > }
> >
> > struct cifs_writedata *
> > -cifs_writedata_alloc(unsigned int nr_pages, work_func_t complete)
> > -{
> > - struct page **pages =
> > - kcalloc(nr_pages, sizeof(struct page *), GFP_NOFS);
> > - if (pages)
> > - return cifs_writedata_direct_alloc(pages, complete);
> > -
> > - return NULL;
> > -}
> > -
> > -struct cifs_writedata *
> > -cifs_writedata_direct_alloc(struct page **pages, work_func_t complete)
> > +cifs_writedata_alloc(work_func_t complete)
> > {
> > struct cifs_writedata *wdata;
> >
> > wdata = kzalloc(sizeof(*wdata), GFP_NOFS);
> > if (wdata != NULL) {
> > - wdata->pages = pages;
> > kref_init(&wdata->refcount);
> > INIT_LIST_HEAD(&wdata->list);
> > init_completion(&wdata->done);
> > @@ -2186,11 +2252,7 @@ cifs_async_writev(struct cifs_writedata *wdata,
> >
> > rqst.rq_iov = iov;
> > rqst.rq_nvec = 2;
> > - rqst.rq_pages = wdata->pages;
> > - rqst.rq_offset = wdata->page_offset;
> > - rqst.rq_npages = wdata->nr_pages;
> > - rqst.rq_pagesz = wdata->pagesz;
> > - rqst.rq_tailsz = wdata->tailsz;
> > + rqst.rq_iter = wdata->iter;
> >
> > cifs_dbg(FYI, "async write at %llu %u bytes\n",
> > wdata->offset, wdata->bytes);
> > diff --git a/fs/cifs/connect.c b/fs/cifs/connect.c
> > index ed210d774a21..d0851c9881b3 100644
> > --- a/fs/cifs/connect.c
> > +++ b/fs/cifs/connect.c
> > @@ -704,6 +704,22 @@ cifs_read_page_from_socket(struct TCP_Server_Info *server, struct page *page,
> > return cifs_readv_from_socket(server, &smb_msg);
> > }
> >
> > +int
> > +cifs_read_iter_from_socket(struct TCP_Server_Info *server, struct iov_iter *iter,
> > + unsigned int to_read)
> > +{
> > + struct msghdr smb_msg;
> > + int ret;
> > +
> > + smb_msg.msg_iter = *iter;
> > + if (smb_msg.msg_iter.count > to_read)
> > + smb_msg.msg_iter.count = to_read;
> > + ret = cifs_readv_from_socket(server, &smb_msg);
> > + if (ret > 0)
> > + iov_iter_advance(iter, ret);
> > + return ret;
> > +}
> > +
> > static bool
> > is_smb_response(struct TCP_Server_Info *server, unsigned char type)
> > {
> > diff --git a/fs/cifs/misc.c b/fs/cifs/misc.c
> > index 56598f7dbe00..f5fe5720456a 100644
> > --- a/fs/cifs/misc.c
> > +++ b/fs/cifs/misc.c
> > @@ -1122,25 +1122,6 @@ cifs_free_hash(struct crypto_shash **shash, struct sdesc **sdesc)
> > *shash = NULL;
> > }
> >
> > -/**
> > - * rqst_page_get_length - obtain the length and offset for a page in smb_rqst
> > - * @rqst: The request descriptor
> > - * @page: The index of the page to query
> > - * @len: Where to store the length for this page:
> > - * @offset: Where to store the offset for this page
> > - */
> > -void rqst_page_get_length(struct smb_rqst *rqst, unsigned int page,
> > - unsigned int *len, unsigned int *offset)
> > -{
> > - *len = rqst->rq_pagesz;
> > - *offset = (page == 0) ? rqst->rq_offset : 0;
> > -
> > - if (rqst->rq_npages == 1 || page == rqst->rq_npages-1)
> > - *len = rqst->rq_tailsz;
> > - else if (page == 0)
> > - *len = rqst->rq_pagesz - rqst->rq_offset;
> > -}
> > -
> > void extract_unc_hostname(const char *unc, const char **h, size_t *len)
> > {
> > const char *end;
> > diff --git a/fs/cifs/smb2ops.c b/fs/cifs/smb2ops.c
> > index af5d0830bc8a..e1649ac194db 100644
> > --- a/fs/cifs/smb2ops.c
> > +++ b/fs/cifs/smb2ops.c
> > @@ -4406,15 +4406,30 @@ fill_transform_hdr(struct smb2_transform_hdr *tr_hdr, unsigned int orig_len,
> > static inline void smb2_sg_set_buf(struct scatterlist *sg, const void *buf,
> > unsigned int buflen)
> > {
> > - void *addr;
> > + struct page *page;
> > +
> > /*
> > * VMAP_STACK (at least) puts stack into the vmalloc address space
> > */
> > if (is_vmalloc_addr(buf))
> > - addr = vmalloc_to_page(buf);
> > + page = vmalloc_to_page(buf);
> > else
> > - addr = virt_to_page(buf);
> > - sg_set_page(sg, addr, buflen, offset_in_page(buf));
> > + page = virt_to_page(buf);
> > + sg_set_page(sg, page, buflen, offset_in_page(buf));
> > +}
> > +
> > +struct cifs_init_sg_priv {
> > + struct scatterlist *sg;
> > + unsigned int idx;
> > +};
> > +
> > +static ssize_t cifs_init_sg_scan(struct iov_iter *i, const void *p,
> > + size_t len, size_t off, void *_priv)
> > +{
> > + struct cifs_init_sg_priv *priv = _priv;
> > +
> > + smb2_sg_set_buf(&priv->sg[priv->idx++], p, len);
> > + return len;
> > }
> >
> > /* Assumes the first rqst has a transform header as the first iov.
> > @@ -4426,43 +4441,46 @@ static inline void smb2_sg_set_buf(struct scatterlist *sg, const void *buf,
> > static struct scatterlist *
> > init_sg(int num_rqst, struct smb_rqst *rqst, u8 *sign)
> > {
> > + struct cifs_init_sg_priv priv;
> > unsigned int sg_len;
> > - struct scatterlist *sg;
> > unsigned int i;
> > unsigned int j;
> > - unsigned int idx = 0;
> > + ssize_t rc;
> > int skip;
> >
> > sg_len = 1;
> > - for (i = 0; i < num_rqst; i++)
> > - sg_len += rqst[i].rq_nvec + rqst[i].rq_npages;
> > + for (i = 0; i < num_rqst; i++) {
> > + unsigned int np = iov_iter_npages(&rqst[i].rq_iter, INT_MAX);
> > + sg_len += rqst[i].rq_nvec + np;
> > + }
> >
> > - sg = kmalloc_array(sg_len, sizeof(struct scatterlist), GFP_KERNEL);
> > - if (!sg)
> > + priv.idx = 0;
> > + priv.sg = kmalloc_array(sg_len, sizeof(struct scatterlist), GFP_KERNEL);
> > + if (!priv.sg)
> > return NULL;
> >
> > - sg_init_table(sg, sg_len);
> > + sg_init_table(priv.sg, sg_len);
> > for (i = 0; i < num_rqst; i++) {
> > + struct iov_iter *iter = &rqst[i].rq_iter;
> > + size_t count = iov_iter_count(iter);
> > +
> > for (j = 0; j < rqst[i].rq_nvec; j++) {
> > /*
> > * The first rqst has a transform header where the
> > * first 20 bytes are not part of the encrypted blob
> > */
> > skip = (i == 0) && (j == 0) ? 20 : 0;
> > - smb2_sg_set_buf(&sg[idx++],
> > + smb2_sg_set_buf(&priv.sg[priv.idx++],
> > rqst[i].rq_iov[j].iov_base + skip,
> > rqst[i].rq_iov[j].iov_len - skip);
> > - }
> > -
> > - for (j = 0; j < rqst[i].rq_npages; j++) {
> > - unsigned int len, offset;
> > -
> > - rqst_page_get_length(&rqst[i], j, &len, &offset);
> > - sg_set_page(&sg[idx++], rqst[i].rq_pages[j], len, offset);
> > }
> > +
> > + rc = iov_iter_scan(iter, count, cifs_init_sg_scan, &priv);
> > + iov_iter_revert(iter, count);
> > + WARN_ON(rc < 0);
> > }
> > - smb2_sg_set_buf(&sg[idx], sign, SMB2_SIGNATURE_SIZE);
> > - return sg;
> > + smb2_sg_set_buf(&priv.sg[priv.idx], sign, SMB2_SIGNATURE_SIZE);
> > + return priv.sg;
> > }
> >
> > static int
> > @@ -4599,18 +4617,30 @@ crypt_message(struct TCP_Server_Info *server, int num_rqst,
> > return rc;
> > }
> >
> > +/*
> > + * Clear a read buffer, discarding the folios which have XA_MARK_0 set.
> > + */
> > +static void cifs_clear_xarray_buffer(struct xarray *buffer)
> > +{
> > + struct folio *folio;
> > + XA_STATE(xas, buffer, 0);
> > +
> > + rcu_read_lock();
> > + xas_for_each_marked(&xas, folio, ULONG_MAX, XA_MARK_0) {
> > + folio_put(folio);
> > + }
> > + rcu_read_unlock();
> > + xa_destroy(buffer);
> > +}
> > +
> > void
> > smb3_free_compound_rqst(int num_rqst, struct smb_rqst *rqst)
> > {
> > - int i, j;
> > + int i;
> >
> > - for (i = 0; i < num_rqst; i++) {
> > - if (rqst[i].rq_pages) {
> > - for (j = rqst[i].rq_npages - 1; j >= 0; j--)
> > - put_page(rqst[i].rq_pages[j]);
> > - kfree(rqst[i].rq_pages);
> > - }
> > - }
> > + for (i = 0; i < num_rqst; i++)
> > + if (!xa_empty(&rqst[i].rq_buffer))
> > + cifs_clear_xarray_buffer(&rqst[i].rq_buffer);
> > }
> >
> > /*
> > @@ -4630,50 +4660,51 @@ static int
> > smb3_init_transform_rq(struct TCP_Server_Info *server, int num_rqst,
> > struct smb_rqst *new_rq, struct smb_rqst *old_rq)
> > {
> > - struct page **pages;
> > struct smb2_transform_hdr *tr_hdr = new_rq[0].rq_iov[0].iov_base;
> > - unsigned int npages;
> > + struct page *page;
> > unsigned int orig_len = 0;
> > int i, j;
> > int rc = -ENOMEM;
> >
> > for (i = 1; i < num_rqst; i++) {
> > - npages = old_rq[i - 1].rq_npages;
> > - pages = kmalloc_array(npages, sizeof(struct page *),
> > - GFP_KERNEL);
> > - if (!pages)
> > - goto err_free;
> > -
> > - new_rq[i].rq_pages = pages;
> > - new_rq[i].rq_npages = npages;
> > - new_rq[i].rq_offset = old_rq[i - 1].rq_offset;
> > - new_rq[i].rq_pagesz = old_rq[i - 1].rq_pagesz;
> > - new_rq[i].rq_tailsz = old_rq[i - 1].rq_tailsz;
> > - new_rq[i].rq_iov = old_rq[i - 1].rq_iov;
> > - new_rq[i].rq_nvec = old_rq[i - 1].rq_nvec;
> > -
> > - orig_len += smb_rqst_len(server, &old_rq[i - 1]);
> > -
> > - for (j = 0; j < npages; j++) {
> > - pages[j] = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
> > - if (!pages[j])
> > - goto err_free;
> > - }
> > -
> > - /* copy pages form the old */
> > - for (j = 0; j < npages; j++) {
> > - char *dst, *src;
> > - unsigned int offset, len;
> > -
> > - rqst_page_get_length(&new_rq[i], j, &len, &offset);
> > -
> > - dst = (char *) kmap(new_rq[i].rq_pages[j]) + offset;
> > - src = (char *) kmap(old_rq[i - 1].rq_pages[j]) + offset;
> > + struct smb_rqst *old = &old_rq[i - 1];
> > + struct smb_rqst *new = &new_rq[i];
> > + struct xarray *buffer = &new->rq_buffer;
> > + unsigned int npages;
> > + size_t size = iov_iter_count(&old->rq_iter), seg, copied = 0;
> > +
> > + xa_init(buffer);
> > +
> > + if (size > 0) {
> > + npages = DIV_ROUND_UP(size, PAGE_SIZE);
> > + for (j = 0; j < npages; j++) {
> > + void *o;
> > +
> > + rc = -ENOMEM;
> > + page = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
> > + if (!page)
> > + goto err_free;
> > + page->index = j;
> > + o = xa_store(buffer, j, page, GFP_KERNEL);
> > + if (xa_is_err(o)) {
> > + rc = xa_err(o);
> > + put_page(page);
> > + goto err_free;
> > + }
> >
> > - memcpy(dst, src, len);
> > - kunmap(new_rq[i].rq_pages[j]);
> > - kunmap(old_rq[i - 1].rq_pages[j]);
> > + seg = min(size - copied, PAGE_SIZE);
> > + if (copy_page_from_iter(page, 0, seg, &old->rq_iter) != seg) {
> > + rc = -EFAULT;
> > + goto err_free;
> > + }
> > + copied += seg;
> > + }
> > + iov_iter_xarray(&new->rq_iter, iov_iter_rw(&old->rq_iter),
> > + buffer, 0, size);
> > }
> > + new->rq_iov = old->rq_iov;
> > + new->rq_nvec = old->rq_nvec;
> > + orig_len += smb_rqst_len(server, new);
> > }
> >
> > /* fill the 1st iov with a transform header */
> > @@ -4701,12 +4732,12 @@ smb3_is_transform_hdr(void *buf)
> >
> > static int
> > decrypt_raw_data(struct TCP_Server_Info *server, char *buf,
> > - unsigned int buf_data_size, struct page **pages,
> > - unsigned int npages, unsigned int page_data_size,
> > + unsigned int buf_data_size, struct iov_iter *iter,
> > bool is_offloaded)
> > {
> > struct kvec iov[2];
> > struct smb_rqst rqst = {NULL};
> > + size_t iter_size = 0;
> > int rc;
> >
> > iov[0].iov_base = buf;
> > @@ -4716,10 +4747,10 @@ decrypt_raw_data(struct TCP_Server_Info *server, char *buf,
> >
> > rqst.rq_iov = iov;
> > rqst.rq_nvec = 2;
> > - rqst.rq_pages = pages;
> > - rqst.rq_npages = npages;
> > - rqst.rq_pagesz = PAGE_SIZE;
> > - rqst.rq_tailsz = (page_data_size % PAGE_SIZE) ? : PAGE_SIZE;
> > + if (iter) {
> > + rqst.rq_iter = *iter;
> > + iter_size = iov_iter_count(iter);
> > + }
> >
> > rc = crypt_message(server, 1, &rqst, 0);
> > cifs_dbg(FYI, "Decrypt message returned %d\n", rc);
> > @@ -4730,73 +4761,37 @@ decrypt_raw_data(struct TCP_Server_Info *server, char *buf,
> > memmove(buf, iov[1].iov_base, buf_data_size);
> >
> > if (!is_offloaded)
> > - server->total_read = buf_data_size + page_data_size;
> > + server->total_read = buf_data_size + iter_size;
> >
> > return rc;
> > }
> >
> > static int
> > -read_data_into_pages(struct TCP_Server_Info *server, struct page **pages,
> > - unsigned int npages, unsigned int len)
> > +cifs_copy_pages_to_iter(struct xarray *pages, unsigned int data_size,
> > + unsigned int skip, struct iov_iter *iter)
> > {
> > - int i;
> > - int length;
> > + struct page *page;
> > + unsigned long index;
> >
> > - for (i = 0; i < npages; i++) {
> > - struct page *page = pages[i];
> > - size_t n;
> > + xa_for_each(pages, index, page) {
> > + size_t n, len = min_t(unsigned int, PAGE_SIZE - skip, data_size);
> >
> > - n = len;
> > - if (len >= PAGE_SIZE) {
> > - /* enough data to fill the page */
> > - n = PAGE_SIZE;
> > - len -= n;
> > - } else {
> > - zero_user(page, len, PAGE_SIZE - len);
> > - len = 0;
> > + n = copy_page_to_iter(page, skip, len, iter);
> > + if (n != len) {
> > + cifs_dbg(VFS, "%s: something went wrong\n", __func__);
> > + return -EIO;
> > }
> > - length = cifs_read_page_from_socket(server, page, 0, n);
> > - if (length < 0)
> > - return length;
> > - server->total_read += length;
> > + data_size -= n;
> > + skip = 0;
> > }
> >
> > return 0;
> > }
> >
> > -static int
> > -init_read_bvec(struct page **pages, unsigned int npages, unsigned int data_size,
> > - unsigned int cur_off, struct bio_vec **page_vec)
> > -{
> > - struct bio_vec *bvec;
> > - int i;
> > -
> > - bvec = kcalloc(npages, sizeof(struct bio_vec), GFP_KERNEL);
> > - if (!bvec)
> > - return -ENOMEM;
> > -
> > - for (i = 0; i < npages; i++) {
> > - bvec[i].bv_page = pages[i];
> > - bvec[i].bv_offset = (i == 0) ? cur_off : 0;
> > - bvec[i].bv_len = min_t(unsigned int, PAGE_SIZE, data_size);
> > - data_size -= bvec[i].bv_len;
> > - }
> > -
> > - if (data_size != 0) {
> > - cifs_dbg(VFS, "%s: something went wrong\n", __func__);
> > - kfree(bvec);
> > - return -EIO;
> > - }
> > -
> > - *page_vec = bvec;
> > - return 0;
> > -}
> > -
> > static int
> > handle_read_data(struct TCP_Server_Info *server, struct mid_q_entry *mid,
> > - char *buf, unsigned int buf_len, struct page **pages,
> > - unsigned int npages, unsigned int page_data_size,
> > - bool is_offloaded)
> > + char *buf, unsigned int buf_len, struct xarray *pages,
> > + unsigned int pages_len, bool is_offloaded)
> > {
> > unsigned int data_offset;
> > unsigned int data_len;
> > @@ -4805,9 +4800,6 @@ handle_read_data(struct TCP_Server_Info *server, struct mid_q_entry *mid,
> > unsigned int pad_len;
> > struct cifs_readdata *rdata = mid->callback_data;
> > struct smb2_hdr *shdr = (struct smb2_hdr *)buf;
> > - struct bio_vec *bvec = NULL;
> > - struct iov_iter iter;
> > - struct kvec iov;
> > int length;
> > bool use_rdma_mr = false;
> >
> > @@ -4896,7 +4888,7 @@ handle_read_data(struct TCP_Server_Info *server, struct mid_q_entry *mid,
> > return 0;
> > }
> >
> > - if (data_len > page_data_size - pad_len) {
> > + if (data_len > pages_len - pad_len) {
> > /* data_len is corrupt -- discard frame */
> > rdata->result = -EIO;
> > if (is_offloaded)
> > @@ -4906,8 +4898,9 @@ handle_read_data(struct TCP_Server_Info *server, struct mid_q_entry *mid,
> > return 0;
> > }
> >
> > - rdata->result = init_read_bvec(pages, npages, page_data_size,
> > - cur_off, &bvec);
> > + /* Copy the data to the output I/O iterator. */
> > + rdata->result = cifs_copy_pages_to_iter(pages, pages_len,
> > + cur_off, &rdata->iter);
> > if (rdata->result != 0) {
> > if (is_offloaded)
> > mid->mid_state = MID_RESPONSE_MALFORMED;
> > @@ -4915,14 +4908,15 @@ handle_read_data(struct TCP_Server_Info *server, struct mid_q_entry *mid,
> > dequeue_mid(mid, rdata->result);
> > return 0;
> > }
> > + rdata->got_bytes = pages_len;
> >
> > - iov_iter_bvec(&iter, WRITE, bvec, npages, data_len);
> > } else if (buf_len >= data_offset + data_len) {
> > /* read response payload is in buf */
> > - WARN_ONCE(npages > 0, "read data can be either in buf or in pages");
> > - iov.iov_base = buf + data_offset;
> > - iov.iov_len = data_len;
> > - iov_iter_kvec(&iter, WRITE, &iov, 1, data_len);
> > + WARN_ONCE(pages && !xa_empty(pages),
> > + "read data can be either in buf or in pages");
> > + length = copy_to_iter(buf + data_offset, data_len, &rdata->iter);
> > + if (length < 0)
> > + return length;
> > } else {
> > /* read response payload cannot be in both buf and pages */
> > WARN_ONCE(1, "buf can not contain only a part of read data");
> > @@ -4934,13 +4928,6 @@ handle_read_data(struct TCP_Server_Info *server, struct mid_q_entry *mid,
> > return 0;
> > }
> >
> > - length = rdata->copy_into_pages(server, rdata, &iter);
> > -
> > - kfree(bvec);
> > -
> > - if (length < 0)
> > - return length;
> > -
> > if (is_offloaded)
> > mid->mid_state = MID_RESPONSE_RECEIVED;
> > else
> > @@ -4951,9 +4938,8 @@ handle_read_data(struct TCP_Server_Info *server, struct mid_q_entry *mid,
> > struct smb2_decrypt_work {
> > struct work_struct decrypt;
> > struct TCP_Server_Info *server;
> > - struct page **ppages;
> > + struct xarray buffer;
> > char *buf;
> > - unsigned int npages;
> > unsigned int len;
> > };
> >
> > @@ -4962,11 +4948,13 @@ static void smb2_decrypt_offload(struct work_struct *work)
> > {
> > struct smb2_decrypt_work *dw = container_of(work,
> > struct smb2_decrypt_work, decrypt);
> > - int i, rc;
> > + int rc;
> > struct mid_q_entry *mid;
> > + struct iov_iter iter;
> >
> > + iov_iter_xarray(&iter, READ, &dw->buffer, 0, dw->len);
> > rc = decrypt_raw_data(dw->server, dw->buf, dw->server->vals->read_rsp_size,
> > - dw->ppages, dw->npages, dw->len, true);
> > + &iter, true);
> > if (rc) {
> > cifs_dbg(VFS, "error decrypting rc=%d\n", rc);
> > goto free_pages;
> > @@ -4980,7 +4968,7 @@ static void smb2_decrypt_offload(struct work_struct *work)
> > mid->decrypted = true;
> > rc = handle_read_data(dw->server, mid, dw->buf,
> > dw->server->vals->read_rsp_size,
> > - dw->ppages, dw->npages, dw->len,
> > + &dw->buffer, dw->len,
> > true);
> > if (rc >= 0) {
> > #ifdef CONFIG_CIFS_STATS2
> > @@ -5012,10 +5000,7 @@ static void smb2_decrypt_offload(struct work_struct *work)
> > }
> >
> > free_pages:
> > - for (i = dw->npages-1; i >= 0; i--)
> > - put_page(dw->ppages[i]);
> > -
> > - kfree(dw->ppages);
> > + cifs_clear_xarray_buffer(&dw->buffer);
> > cifs_small_buf_release(dw->buf);
> > kfree(dw);
> > }
> > @@ -5025,47 +5010,66 @@ static int
> > receive_encrypted_read(struct TCP_Server_Info *server, struct mid_q_entry **mid,
> > int *num_mids)
> > {
> > + struct page *page;
> > char *buf = server->smallbuf;
> > struct smb2_transform_hdr *tr_hdr = (struct smb2_transform_hdr *)buf;
> > - unsigned int npages;
> > - struct page **pages;
> > - unsigned int len;
> > + struct iov_iter iter;
> > + unsigned int len, npages;
> > unsigned int buflen = server->pdu_size;
> > int rc;
> > int i = 0;
> > struct smb2_decrypt_work *dw;
> >
> > + dw = kzalloc(sizeof(struct smb2_decrypt_work), GFP_KERNEL);
> > + if (!dw)
> > + return -ENOMEM;
> > + xa_init(&dw->buffer);
> > + INIT_WORK(&dw->decrypt, smb2_decrypt_offload);
> > + dw->server = server;
> > +
> > *num_mids = 1;
> > len = min_t(unsigned int, buflen, server->vals->read_rsp_size +
> > sizeof(struct smb2_transform_hdr)) - HEADER_SIZE(server) + 1;
> >
> > rc = cifs_read_from_socket(server, buf + HEADER_SIZE(server) - 1, len);
> > if (rc < 0)
> > - return rc;
> > + goto free_dw;
> > server->total_read += rc;
> >
> > len = le32_to_cpu(tr_hdr->OriginalMessageSize) -
> > server->vals->read_rsp_size;
> > + dw->len = len;
> > npages = DIV_ROUND_UP(len, PAGE_SIZE);
> >
> > - pages = kmalloc_array(npages, sizeof(struct page *), GFP_KERNEL);
> > - if (!pages) {
> > - rc = -ENOMEM;
> > - goto discard_data;
> > - }
> > -
> > + rc = -ENOMEM;
> > for (; i < npages; i++) {
> > - pages[i] = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
> > - if (!pages[i]) {
> > - rc = -ENOMEM;
> > + void *old;
> > +
> > + page = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
> > + if (!page) {
> > + goto discard_data;
> > + }
> > + page->index = i;
> > + old = xa_store(&dw->buffer, i, page, GFP_KERNEL);
> > + if (xa_is_err(old)) {
> > + rc = xa_err(old);
> > + put_page(page);
> > goto discard_data;
> > }
> > }
> >
> > - /* read read data into pages */
> > - rc = read_data_into_pages(server, pages, npages, len);
> > - if (rc)
> > - goto free_pages;
> > + iov_iter_xarray(&iter, READ, &dw->buffer, 0, npages * PAGE_SIZE);
> > +
> > + /* Read the data into the buffer and clear excess bufferage. */
> > + rc = cifs_read_iter_from_socket(server, &iter, dw->len);
> > + if (rc < 0)
> > + goto discard_data;
> > +
> > + server->total_read += rc;
> > + if (rc < npages * PAGE_SIZE)
> > + iov_iter_zero(npages * PAGE_SIZE - rc, &iter);
> > + iov_iter_revert(&iter, npages * PAGE_SIZE);
> > + iov_iter_truncate(&iter, dw->len);
> >
> > rc = cifs_discard_remaining_data(server);
> > if (rc)
> > @@ -5078,39 +5082,28 @@ receive_encrypted_read(struct TCP_Server_Info *server, struct mid_q_entry **mid,
> >
> > if ((server->min_offload) && (server->in_flight > 1) &&
> > (server->pdu_size >= server->min_offload)) {
> > - dw = kmalloc(sizeof(struct smb2_decrypt_work), GFP_KERNEL);
> > - if (dw == NULL)
> > - goto non_offloaded_decrypt;
> > -
> > dw->buf = server->smallbuf;
> > server->smallbuf = (char *)cifs_small_buf_get();
> >
> > - INIT_WORK(&dw->decrypt, smb2_decrypt_offload);
> > -
> > - dw->npages = npages;
> > - dw->server = server;
> > - dw->ppages = pages;
> > - dw->len = len;
> > queue_work(decrypt_wq, &dw->decrypt);
> > *num_mids = 0; /* worker thread takes care of finding mid */
> > return -1;
> > }
> >
> > -non_offloaded_decrypt:
> > rc = decrypt_raw_data(server, buf, server->vals->read_rsp_size,
> > - pages, npages, len, false);
> > + &iter, false);
> > if (rc)
> > goto free_pages;
> >
> > *mid = smb2_find_mid(server, buf);
> > - if (*mid == NULL)
> > + if (*mid == NULL) {
> > cifs_dbg(FYI, "mid not found\n");
> > - else {
> > + } else {
> > cifs_dbg(FYI, "mid found\n");
> > (*mid)->decrypted = true;
> > rc = handle_read_data(server, *mid, buf,
> > server->vals->read_rsp_size,
> > - pages, npages, len, false);
> > + &dw->buffer, dw->len, false);
> > if (rc >= 0) {
> > if (server->ops->is_network_name_deleted) {
> > server->ops->is_network_name_deleted(buf,
> > @@ -5120,9 +5113,9 @@ receive_encrypted_read(struct TCP_Server_Info *server, struct mid_q_entry **mid,
> > }
> >
> > free_pages:
> > - for (i = i - 1; i >= 0; i--)
> > - put_page(pages[i]);
> > - kfree(pages);
> > + cifs_clear_xarray_buffer(&dw->buffer);
> > +free_dw:
> > + kfree(dw);
> > return rc;
> > discard_data:
> > cifs_discard_remaining_data(server);
> > @@ -5160,7 +5153,7 @@ receive_encrypted_standard(struct TCP_Server_Info *server,
> > server->total_read += length;
> >
> > buf_size = pdu_length - sizeof(struct smb2_transform_hdr);
> > - length = decrypt_raw_data(server, buf, buf_size, NULL, 0, 0, false);
> > + length = decrypt_raw_data(server, buf, buf_size, NULL, false);
> > if (length)
> > return length;
> >
> > @@ -5259,7 +5252,7 @@ smb3_handle_read_data(struct TCP_Server_Info *server, struct mid_q_entry *mid)
> > char *buf = server->large_buf ? server->bigbuf : server->smallbuf;
> >
> > return handle_read_data(server, mid, buf, server->pdu_size,
> > - NULL, 0, 0, false);
> > + NULL, 0, false);
> > }
> >
> > static int
> > diff --git a/fs/cifs/smb2pdu.c b/fs/cifs/smb2pdu.c
> > index 7e7909b1ae11..ebbea7526ee2 100644
> > --- a/fs/cifs/smb2pdu.c
> > +++ b/fs/cifs/smb2pdu.c
> > @@ -4118,11 +4118,7 @@ smb2_readv_callback(struct mid_q_entry *mid)
> > struct cifs_credits credits = { .value = 0, .instance = 0 };
> > struct smb_rqst rqst = { .rq_iov = &rdata->iov[1],
> > .rq_nvec = 1,
> > - .rq_pages = rdata->pages,
> > - .rq_offset = rdata->page_offset,
> > - .rq_npages = rdata->nr_pages,
> > - .rq_pagesz = rdata->pagesz,
> > - .rq_tailsz = rdata->tailsz };
> > + .rq_iter = rdata->iter };
> >
> > WARN_ONCE(rdata->server != mid->server,
> > "rdata server %p != mid server %p",
> > @@ -4522,11 +4518,7 @@ smb2_async_writev(struct cifs_writedata *wdata,
> >
> > rqst.rq_iov = iov;
> > rqst.rq_nvec = 1;
> > - rqst.rq_pages = wdata->pages;
> > - rqst.rq_offset = wdata->page_offset;
> > - rqst.rq_npages = wdata->nr_pages;
> > - rqst.rq_pagesz = wdata->pagesz;
> > - rqst.rq_tailsz = wdata->tailsz;
> > + rqst.rq_iter = wdata->iter;
> > #ifdef CONFIG_CIFS_SMB_DIRECT
> > if (wdata->mr) {
> > iov[0].iov_len += sizeof(struct smbd_buffer_descriptor_v1);
> > diff --git a/fs/cifs/transport.c b/fs/cifs/transport.c
> > index 8540f7c13eae..cb19c43c0009 100644
> > --- a/fs/cifs/transport.c
> > +++ b/fs/cifs/transport.c
> > @@ -276,26 +276,7 @@ smb_rqst_len(struct TCP_Server_Info *server, struct smb_rqst *rqst)
> > for (i = 0; i < nvec; i++)
> > buflen += iov[i].iov_len;
> >
> > - /*
> > - * Add in the page array if there is one. The caller needs to make
> > - * sure rq_offset and rq_tailsz are set correctly. If a buffer of
> > - * multiple pages ends at page boundary, rq_tailsz needs to be set to
> > - * PAGE_SIZE.
> > - */
> > - if (rqst->rq_npages) {
> > - if (rqst->rq_npages == 1)
> > - buflen += rqst->rq_tailsz;
> > - else {
> > - /*
> > - * If there is more than one page, calculate the
> > - * buffer length based on rq_offset and rq_tailsz
> > - */
> > - buflen += rqst->rq_pagesz * (rqst->rq_npages - 1) -
> > - rqst->rq_offset;
> > - buflen += rqst->rq_tailsz;
> > - }
> > - }
> > -
> > + buflen += iov_iter_count(&rqst->rq_iter);
> > return buflen;
> > }
> >
> > @@ -382,23 +363,15 @@ __smb_send_rqst(struct TCP_Server_Info *server, int num_rqst,
> >
> > total_len += sent;
> >
> > - /* now walk the page array and send each page in it */
> > - for (i = 0; i < rqst[j].rq_npages; i++) {
> > - struct bio_vec bvec;
> > -
> > - bvec.bv_page = rqst[j].rq_pages[i];
> > - rqst_page_get_length(&rqst[j], i, &bvec.bv_len,
> > - &bvec.bv_offset);
> > -
> > - iov_iter_bvec(&smb_msg.msg_iter, WRITE,
> > - &bvec, 1, bvec.bv_len);
> > + if (iov_iter_count(&rqst[j].rq_iter) > 0) {
> > + smb_msg.msg_iter = rqst[j].rq_iter;
> > rc = smb_send_kvec(server, &smb_msg, &sent);
> > if (rc < 0)
> > break;
> > -
> > total_len += sent;
> > }
> > - }
> > +
> > +}
> >
> > unmask:
> > sigprocmask(SIG_SETMASK, &oldmask, NULL);
> >
> >
--
Regards,
Shyam
next prev parent reply other threads:[~2022-01-31 5:48 UTC|newest]
Thread overview: 20+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-01-25 13:57 [RFC][RFC PATCH 0/7] cifs: In-progress conversion to use iov_iters and netfslib David Howells
2022-01-25 13:57 ` [RFC PATCH 1/7] cifs: Transition from ->readpages() to ->readahead() David Howells
2022-01-25 14:20 ` Matthew Wilcox
2022-01-25 14:57 ` David Howells
2022-01-25 13:57 ` [RFC PATCH 2/7] cifs: Miscellaneous bits David Howells
2022-01-25 13:57 ` [RFC PATCH 3/7] cifs: Change the I/O paths to use an iterator rather than a page list David Howells
2022-01-31 5:06 ` Rohith Surabattula
2022-01-31 5:48 ` Shyam Prasad N [this message]
2022-02-14 16:06 ` David Howells
2022-01-25 13:58 ` [RFC PATCH 4/7] cifs: Make cifs_writepages() hand an iterator down David Howells
2022-01-25 13:58 ` [RFC PATCH 5/7] cifs: Make cifs_readahead() pass " David Howells
2022-01-25 13:58 ` [RFC PATCH 6/7] cifs: Get direct I/O and unbuffered I/O working with iterators David Howells
2022-01-25 13:59 ` [RFC PATCH 7/7] cifs: Use netfslib to handle reads David Howells
2022-02-08 5:59 ` Rohith Surabattula
2022-02-14 16:33 ` David Howells
2022-02-28 14:14 ` Rohith Surabattula
2022-02-28 14:28 ` David Howells
2022-03-01 7:21 ` Rohith Surabattula
2022-03-01 9:39 ` David Howells
2022-03-02 13:51 ` Rohith Surabattula
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to='CANT5p=o-MYJY6kmkgw5Y_e7zdWjEkNZ1f_0aAKKeCDHjEFLfvw@mail.gmail.com' \
--to=nspmangalore@gmail.com \
--cc=dhowells@redhat.com \
--cc=jlayton@kernel.org \
--cc=linux-cachefs@redhat.com \
--cc=linux-cifs@vger.kernel.org \
--cc=linux-fsdevel@vger.kernel.org \
--cc=rohiths.msft@gmail.com \
--cc=smfrench@gmail.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).