[PATCH v4 02/10] block: Add copy offload support infrastructure
kernel test robot
lkp at intel.com
Tue Apr 26 17:11:24 PDT 2022
Hi Nitesh,
Thank you for the patch! Perhaps something to improve:
[auto build test WARNING on next-20220422]
[cannot apply to axboe-block/for-next device-mapper-dm/for-next linus/master v5.18-rc4 v5.18-rc3 v5.18-rc2 v5.18-rc4]
[If your patch is applied to the wrong git tree, kindly drop us a note.
And when submitting patch, we suggest to use '--base' as documented in
https://git-scm.com/docs/git-format-patch]
url: https://github.com/intel-lab-lkp/linux/commits/Nitesh-Shetty/block-Introduce-queue-limits-for-copy-offload-support/20220426-201825
base: e7d6987e09a328d4a949701db40ef63fbb970670
config: hexagon-randconfig-r041-20220425 (https://download.01.org/0day-ci/archive/20220427/202204270754.pM0Ewhl5-lkp@intel.com/config)
compiler: clang version 15.0.0 (https://github.com/llvm/llvm-project 1cddcfdc3c683b393df1a5c9063252eb60e52818)
reproduce (this is a W=1 build):
wget https://raw.githubusercontent.com/intel/lkp-tests/master/sbin/make.cross -O ~/bin/make.cross
chmod +x ~/bin/make.cross
# https://github.com/intel-lab-lkp/linux/commit/3e91cba65ef73ba116953031d5548da7fd33a150
git remote add linux-review https://github.com/intel-lab-lkp/linux
git fetch --no-tags linux-review Nitesh-Shetty/block-Introduce-queue-limits-for-copy-offload-support/20220426-201825
git checkout 3e91cba65ef73ba116953031d5548da7fd33a150
# save the config file
mkdir build_dir && cp config build_dir/.config
COMPILER_INSTALL_PATH=$HOME/0day COMPILER=clang make.cross W=1 O=build_dir ARCH=hexagon SHELL=/bin/bash
If you fix the issue, kindly add following tag as appropriate
Reported-by: kernel test robot <lkp at intel.com>
All warnings (new ones prefixed by >>):
>> block/blk-lib.c:178:5: warning: no previous prototype for function 'blk_copy_offload' [-Wmissing-prototypes]
int blk_copy_offload(struct block_device *src_bdev, int nr_srcs,
^
block/blk-lib.c:178:1: note: declare 'static' if the function is not intended to be used outside of this translation unit
int blk_copy_offload(struct block_device *src_bdev, int nr_srcs,
^
static
1 warning generated.
vim +/blk_copy_offload +178 block/blk-lib.c
173
174 /*
175 * blk_copy_offload - Use device's native copy offload feature
176 * Go through user provide payload, prepare new payload based on device's copy offload limits.
177 */
> 178 int blk_copy_offload(struct block_device *src_bdev, int nr_srcs,
179 struct range_entry *rlist, struct block_device *dst_bdev, gfp_t gfp_mask)
180 {
181 struct request_queue *sq = bdev_get_queue(src_bdev);
182 struct request_queue *dq = bdev_get_queue(dst_bdev);
183 struct bio *read_bio, *write_bio;
184 struct copy_ctx *ctx;
185 struct cio *cio;
186 struct page *token;
187 sector_t src_blk, copy_len, dst_blk;
188 sector_t remaining, max_copy_len = LONG_MAX;
189 unsigned long flags;
190 int ri = 0, ret = 0;
191
192 cio = kzalloc(sizeof(struct cio), GFP_KERNEL);
193 if (!cio)
194 return -ENOMEM;
195 cio->rlist = rlist;
196 spin_lock_init(&cio->lock);
197
198 max_copy_len = min_t(sector_t, sq->limits.max_copy_sectors, dq->limits.max_copy_sectors);
199 max_copy_len = min3(max_copy_len, (sector_t)sq->limits.max_copy_range_sectors,
200 (sector_t)dq->limits.max_copy_range_sectors) << SECTOR_SHIFT;
201
202 for (ri = 0; ri < nr_srcs; ri++) {
203 cio->rlist[ri].comp_len = rlist[ri].len;
204 src_blk = rlist[ri].src;
205 dst_blk = rlist[ri].dst;
206 for (remaining = rlist[ri].len; remaining > 0; remaining -= copy_len) {
207 copy_len = min(remaining, max_copy_len);
208
209 token = alloc_page(gfp_mask);
210 if (unlikely(!token)) {
211 ret = -ENOMEM;
212 goto err_token;
213 }
214
215 ctx = kzalloc(sizeof(struct copy_ctx), gfp_mask);
216 if (!ctx) {
217 ret = -ENOMEM;
218 goto err_ctx;
219 }
220 ctx->cio = cio;
221 ctx->range_idx = ri;
222 ctx->start_sec = dst_blk;
223
224 read_bio = bio_alloc(src_bdev, 1, REQ_OP_READ | REQ_COPY | REQ_NOMERGE,
225 gfp_mask);
226 if (!read_bio) {
227 ret = -ENOMEM;
228 goto err_read_bio;
229 }
230 read_bio->bi_iter.bi_sector = src_blk >> SECTOR_SHIFT;
231 __bio_add_page(read_bio, token, PAGE_SIZE, 0);
232 /*__bio_add_page increases bi_size by len, so overwrite it with copy len*/
233 read_bio->bi_iter.bi_size = copy_len;
234 ret = submit_bio_wait(read_bio);
235 bio_put(read_bio);
236 if (ret)
237 goto err_read_bio;
238
239 write_bio = bio_alloc(dst_bdev, 1, REQ_OP_WRITE | REQ_COPY | REQ_NOMERGE,
240 gfp_mask);
241 if (!write_bio) {
242 ret = -ENOMEM;
243 goto err_read_bio;
244 }
245 write_bio->bi_iter.bi_sector = dst_blk >> SECTOR_SHIFT;
246 __bio_add_page(write_bio, token, PAGE_SIZE, 0);
247 /*__bio_add_page increases bi_size by len, so overwrite it with copy len*/
248 write_bio->bi_iter.bi_size = copy_len;
249 write_bio->bi_end_io = bio_copy_end_io;
250 write_bio->bi_private = ctx;
251
252 spin_lock_irqsave(&cio->lock, flags);
253 ++cio->refcount;
254 spin_unlock_irqrestore(&cio->lock, flags);
255
256 submit_bio(write_bio);
257 src_blk += copy_len;
258 dst_blk += copy_len;
259 }
260 }
261
262 /* Wait for completion of all IO's*/
263 return cio_await_completion(cio);
264
265 err_read_bio:
266 kfree(ctx);
267 err_ctx:
268 __free_page(token);
269 err_token:
270 rlist[ri].comp_len = min_t(sector_t, rlist[ri].comp_len, (rlist[ri].len - remaining));
271
272 cio->io_err = ret;
273 return cio_await_completion(cio);
274 }
275
--
0-DAY CI Kernel Test Service
https://01.org/lkp
More information about the Linux-nvme
mailing list