From: Zach Brown Date: Fri, 14 Jun 2013 20:54:13 +0000 (-0500) Subject: iov_iter: add bvec support X-Git-Tag: next-20130919~60^2~25 X-Git-Url: https://git.karo-electronics.de/?a=commitdiff_plain;h=66daeabae8993c6a7544744eb61e83597153a9bd;p=karo-tx-linux.git iov_iter: add bvec support This adds a set of iov_iter_ops calls which work with memory which is specified by an array of bio_vec structs instead of an array of iovec structs. The big difference is that the pages referenced by the bio_vec elements are pinned. They don't need to be faulted in and we can always use kmap_atomic() to map them one at a time. Signed-off-by: Zach Brown Signed-off-by: Dave Kleikamp Tested-by: Sedat Dilek --- diff --git a/fs/iov-iter.c b/fs/iov-iter.c index 59f9556ac4d2..5624e367f760 100644 --- a/fs/iov-iter.c +++ b/fs/iov-iter.c @@ -5,6 +5,7 @@ #include #include #include +#include static size_t __iovec_copy_to_user(char *vaddr, const struct iovec *iov, size_t base, size_t bytes, int atomic) @@ -109,6 +110,134 @@ success: return copied; } +#ifdef CONFIG_BLOCK +/* + * As an easily verifiable first pass, we implement all the methods that + * copy data to and from bvec pages with one function. We implement it + * all with kmap_atomic(). + */ +static size_t bvec_copy_tofrom_page(struct iov_iter *iter, struct page *page, + unsigned long page_offset, size_t bytes, + int topage) +{ + struct bio_vec *bvec = (struct bio_vec *)iter->data; + size_t bvec_offset = iter->iov_offset; + size_t remaining = bytes; + void *bvec_map; + void *page_map; + size_t copy; + + page_map = kmap_atomic(page); + + BUG_ON(bytes > iter->count); + while (remaining) { + BUG_ON(bvec->bv_len == 0); + BUG_ON(bvec_offset >= bvec->bv_len); + copy = min(remaining, bvec->bv_len - bvec_offset); + bvec_map = kmap_atomic(bvec->bv_page); + if (topage) + memcpy(page_map + page_offset, + bvec_map + bvec->bv_offset + bvec_offset, + copy); + else + memcpy(bvec_map + bvec->bv_offset + bvec_offset, + page_map + page_offset, + copy); + kunmap_atomic(bvec_map); + remaining -= copy; + bvec_offset += copy; + page_offset += copy; + if (bvec_offset == bvec->bv_len) { + bvec_offset = 0; + bvec++; + } + } + + kunmap_atomic(page_map); + + return bytes; +} + +static size_t ii_bvec_copy_to_user_atomic(struct page *page, struct iov_iter *i, + unsigned long offset, size_t bytes) +{ + return bvec_copy_tofrom_page(i, page, offset, bytes, 0); +} +static size_t ii_bvec_copy_to_user(struct page *page, struct iov_iter *i, + unsigned long offset, size_t bytes, + int check_access) +{ + return bvec_copy_tofrom_page(i, page, offset, bytes, 0); +} +static size_t ii_bvec_copy_from_user_atomic(struct page *page, + struct iov_iter *i, + unsigned long offset, size_t bytes) +{ + return bvec_copy_tofrom_page(i, page, offset, bytes, 1); +} +static size_t ii_bvec_copy_from_user(struct page *page, struct iov_iter *i, + unsigned long offset, size_t bytes) +{ + return bvec_copy_tofrom_page(i, page, offset, bytes, 1); +} + +/* + * bio_vecs have a stricter structure than iovecs that might have + * come from userspace. There are no zero length bio_vec elements. + */ +static void ii_bvec_advance(struct iov_iter *i, size_t bytes) +{ + struct bio_vec *bvec = (struct bio_vec *)i->data; + size_t offset = i->iov_offset; + size_t delta; + + BUG_ON(i->count < bytes); + while (bytes) { + BUG_ON(bvec->bv_len == 0); + BUG_ON(bvec->bv_len <= offset); + delta = min(bytes, bvec->bv_len - offset); + offset += delta; + i->count -= delta; + bytes -= delta; + if (offset == bvec->bv_len) { + bvec++; + offset = 0; + } + } + + i->data = (unsigned long)bvec; + i->iov_offset = offset; +} + +/* + * pages pointed to by bio_vecs are always pinned. + */ +static int ii_bvec_fault_in_readable(struct iov_iter *i, size_t bytes) +{ + return 0; +} + +static size_t ii_bvec_single_seg_count(const struct iov_iter *i) +{ + const struct bio_vec *bvec = (struct bio_vec *)i->data; + if (i->nr_segs == 1) + return i->count; + else + return min(i->count, bvec->bv_len - i->iov_offset); +} + +struct iov_iter_ops ii_bvec_ops = { + .ii_copy_to_user_atomic = ii_bvec_copy_to_user_atomic, + .ii_copy_to_user = ii_bvec_copy_to_user, + .ii_copy_from_user_atomic = ii_bvec_copy_from_user_atomic, + .ii_copy_from_user = ii_bvec_copy_from_user, + .ii_advance = ii_bvec_advance, + .ii_fault_in_readable = ii_bvec_fault_in_readable, + .ii_single_seg_count = ii_bvec_single_seg_count, +}; +EXPORT_SYMBOL(ii_bvec_ops); +#endif /* CONFIG_BLOCK */ + static size_t __iovec_copy_from_user(char *vaddr, const struct iovec *iov, size_t base, size_t bytes, int atomic) { diff --git a/include/linux/fs.h b/include/linux/fs.h index 0137010e0303..ea607efed90c 100644 --- a/include/linux/fs.h +++ b/include/linux/fs.h @@ -350,6 +350,25 @@ static inline size_t iov_iter_single_seg_count(const struct iov_iter *i) return i->ops->ii_single_seg_count(i); } +#ifdef CONFIG_BLOCK +extern struct iov_iter_ops ii_bvec_ops; + +struct bio_vec; +static inline void iov_iter_init_bvec(struct iov_iter *i, + struct bio_vec *bvec, + unsigned long nr_segs, + size_t count, size_t written) +{ + i->ops = &ii_bvec_ops; + i->data = (unsigned long)bvec; + i->nr_segs = nr_segs; + i->iov_offset = 0; + i->count = count + written; + + iov_iter_advance(i, written); +} +#endif + extern struct iov_iter_ops ii_iovec_ops; static inline void iov_iter_init(struct iov_iter *i,