Message ID | 20240306102846.1020868-4-lizhijian@fujitsu.com (mailing list archive) |
---|---|
State | Changes Requested, archived |
Delegated to: | Ira Weiny |
Headers | show |
Series | device backed vmemmap crash dump support | expand |
Hi Li, kernel test robot noticed the following build errors: [auto build test ERROR on nvdimm/libnvdimm-for-next] [also build test ERROR on tip/x86/core linus/master v6.8-rc7] [cannot apply to akpm-mm/mm-everything nvdimm/dax-misc next-20240306] [If your patch is applied to the wrong git tree, kindly drop us a note. And when submitting patch, we suggest to use '--base' as documented in https://git-scm.com/docs/git-format-patch#_base_tree_information] url: https://github.com/intel-lab-lkp/linux/commits/Li-Zhijian/mm-memremap-register-unregister-altmap-region-to-a-separate-resource/20240306-183118 base: https://git.kernel.org/pub/scm/linux/kernel/git/nvdimm/nvdimm.git libnvdimm-for-next patch link: https://lore.kernel.org/r/20240306102846.1020868-4-lizhijian%40fujitsu.com patch subject: [PATCH v3 3/7] nvdimm: pmem: assign a parent resource for vmemmap region for the fsdax config: riscv-defconfig (https://download.01.org/0day-ci/archive/20240307/202403071558.83GY9NwT-lkp@intel.com/config) compiler: clang version 19.0.0git (https://github.com/llvm/llvm-project 325f51237252e6dab8e4e1ea1fa7acbb4faee1cd) reproduce (this is a W=1 build): (https://download.01.org/0day-ci/archive/20240307/202403071558.83GY9NwT-lkp@intel.com/reproduce) If you fix the issue in a separate patch/commit (i.e. not just a new version of the same patch/commit), kindly add following tags | Reported-by: kernel test robot <lkp@intel.com> | Closes: https://lore.kernel.org/oe-kbuild-all/202403071558.83GY9NwT-lkp@intel.com/ All errors (new ones prefixed by >>): In file included from drivers/nvdimm/pmem.c:10: In file included from include/linux/blkdev.h:9: In file included from include/linux/blk_types.h:10: In file included from include/linux/bvec.h:10: In file included from include/linux/highmem.h:8: In file included from include/linux/cacheflush.h:5: In file included from arch/riscv/include/asm/cacheflush.h:9: In file included from include/linux/mm.h:2188: include/linux/vmstat.h:522:36: warning: arithmetic between different enumeration types ('enum node_stat_item' and 'enum lru_list') [-Wenum-enum-conversion] 522 | return node_stat_name(NR_LRU_BASE + lru) + 3; // skip "nr_" | ~~~~~~~~~~~ ^ ~~~ >> drivers/nvdimm/pmem.c:501:2: error: call to undeclared function 'pgmap_parent_resource'; ISO C99 and later do not support implicit function declarations [-Wimplicit-function-declaration] 501 | pgmap_parent_resource(&pmem->pgmap, parent); | ^ 1 warning and 1 error generated. vim +/pgmap_parent_resource +501 drivers/nvdimm/pmem.c 448 449 static int pmem_attach_disk(struct device *dev, 450 struct nd_namespace_common *ndns) 451 { 452 struct nd_namespace_io *nsio = to_nd_namespace_io(&ndns->dev); 453 struct nd_region *nd_region = to_nd_region(dev->parent); 454 int nid = dev_to_node(dev), fua; 455 struct resource *res = &nsio->res, *parent; 456 struct range bb_range; 457 struct nd_pfn *nd_pfn = NULL; 458 struct dax_device *dax_dev; 459 struct nd_pfn_sb *pfn_sb; 460 struct pmem_device *pmem; 461 struct request_queue *q; 462 struct gendisk *disk; 463 void *addr; 464 int rc; 465 466 pmem = devm_kzalloc(dev, sizeof(*pmem), GFP_KERNEL); 467 if (!pmem) 468 return -ENOMEM; 469 470 rc = devm_namespace_enable(dev, ndns, nd_info_block_reserve()); 471 if (rc) 472 return rc; 473 474 /* while nsio_rw_bytes is active, parse a pfn info block if present */ 475 if (is_nd_pfn(dev)) { 476 nd_pfn = to_nd_pfn(dev); 477 rc = nvdimm_setup_pfn(nd_pfn, &pmem->pgmap); 478 if (rc) 479 return rc; 480 } 481 482 /* we're attaching a block device, disable raw namespace access */ 483 devm_namespace_disable(dev, ndns); 484 485 dev_set_drvdata(dev, pmem); 486 pmem->phys_addr = res->start; 487 pmem->size = resource_size(res); 488 fua = nvdimm_has_flush(nd_region); 489 if (!IS_ENABLED(CONFIG_ARCH_HAS_UACCESS_FLUSHCACHE) || fua < 0) { 490 dev_warn(dev, "unable to guarantee persistence of writes\n"); 491 fua = 0; 492 } 493 494 parent = devm_request_mem_region(dev, res->start, resource_size(res), 495 dev_name(&ndns->dev)); 496 if (!res) { 497 dev_warn(dev, "could not reserve region %pR\n", res); 498 return -EBUSY; 499 } 500 > 501 pgmap_parent_resource(&pmem->pgmap, parent); 502 503 disk = blk_alloc_disk(nid); 504 if (!disk) 505 return -ENOMEM; 506 q = disk->queue; 507 508 pmem->disk = disk; 509 pmem->pgmap.owner = pmem; 510 pmem->pfn_flags = PFN_DEV; 511 if (is_nd_pfn(dev)) { 512 pmem->pgmap.type = MEMORY_DEVICE_FS_DAX; 513 pmem->pgmap.ops = &fsdax_pagemap_ops; 514 addr = devm_memremap_pages(dev, &pmem->pgmap); 515 pfn_sb = nd_pfn->pfn_sb; 516 pmem->data_offset = le64_to_cpu(pfn_sb->dataoff); 517 pmem->pfn_pad = resource_size(res) - 518 range_len(&pmem->pgmap.range); 519 pmem->pfn_flags |= PFN_MAP; 520 bb_range = pmem->pgmap.range; 521 bb_range.start += pmem->data_offset; 522 } else if (pmem_should_map_pages(dev)) { 523 pmem->pgmap.range.start = res->start; 524 pmem->pgmap.range.end = res->end; 525 pmem->pgmap.nr_range = 1; 526 pmem->pgmap.type = MEMORY_DEVICE_FS_DAX; 527 pmem->pgmap.ops = &fsdax_pagemap_ops; 528 addr = devm_memremap_pages(dev, &pmem->pgmap); 529 pmem->pfn_flags |= PFN_MAP; 530 bb_range = pmem->pgmap.range; 531 } else { 532 addr = devm_memremap(dev, pmem->phys_addr, 533 pmem->size, ARCH_MEMREMAP_PMEM); 534 bb_range.start = res->start; 535 bb_range.end = res->end; 536 } 537 538 if (IS_ERR(addr)) { 539 rc = PTR_ERR(addr); 540 goto out; 541 } 542 pmem->virt_addr = addr; 543 544 blk_queue_write_cache(q, true, fua); 545 blk_queue_physical_block_size(q, PAGE_SIZE); 546 blk_queue_logical_block_size(q, pmem_sector_size(ndns)); 547 blk_queue_max_hw_sectors(q, UINT_MAX); 548 blk_queue_flag_set(QUEUE_FLAG_NONROT, q); 549 blk_queue_flag_set(QUEUE_FLAG_SYNCHRONOUS, q); 550 if (pmem->pfn_flags & PFN_MAP) 551 blk_queue_flag_set(QUEUE_FLAG_DAX, q); 552 553 disk->fops = &pmem_fops; 554 disk->private_data = pmem; 555 nvdimm_namespace_disk_name(ndns, disk->disk_name); 556 set_capacity(disk, (pmem->size - pmem->pfn_pad - pmem->data_offset) 557 / 512); 558 if (devm_init_badblocks(dev, &pmem->bb)) 559 return -ENOMEM; 560 nvdimm_badblocks_populate(nd_region, &pmem->bb, &bb_range); 561 disk->bb = &pmem->bb; 562 563 dax_dev = alloc_dax(pmem, &pmem_dax_ops); 564 if (IS_ERR(dax_dev)) { 565 rc = PTR_ERR(dax_dev); 566 goto out; 567 } 568 set_dax_nocache(dax_dev); 569 set_dax_nomc(dax_dev); 570 if (is_nvdimm_sync(nd_region)) 571 set_dax_synchronous(dax_dev); 572 rc = dax_add_host(dax_dev, disk); 573 if (rc) 574 goto out_cleanup_dax; 575 dax_write_cache(dax_dev, nvdimm_has_cache(nd_region)); 576 pmem->dax_dev = dax_dev; 577 578 rc = device_add_disk(dev, disk, pmem_attribute_groups); 579 if (rc) 580 goto out_remove_host; 581 if (devm_add_action_or_reset(dev, pmem_release_disk, pmem)) 582 return -ENOMEM; 583 584 nvdimm_check_and_set_ro(disk); 585 586 pmem->bb_state = sysfs_get_dirent(disk_to_dev(disk)->kobj.sd, 587 "badblocks"); 588 if (!pmem->bb_state) 589 dev_warn(dev, "'badblocks' notification disabled\n"); 590 return 0; 591 592 out_remove_host: 593 dax_remove_host(pmem->disk); 594 out_cleanup_dax: 595 kill_dax(pmem->dax_dev); 596 put_dax(pmem->dax_dev); 597 out: 598 put_disk(pmem->disk); 599 return rc; 600 } 601
Hi Li, kernel test robot noticed the following build errors: [auto build test ERROR on nvdimm/libnvdimm-for-next] [also build test ERROR on tip/x86/core linus/master v6.8-rc7] [cannot apply to akpm-mm/mm-everything nvdimm/dax-misc next-20240306] [If your patch is applied to the wrong git tree, kindly drop us a note. And when submitting patch, we suggest to use '--base' as documented in https://git-scm.com/docs/git-format-patch#_base_tree_information] url: https://github.com/intel-lab-lkp/linux/commits/Li-Zhijian/mm-memremap-register-unregister-altmap-region-to-a-separate-resource/20240306-183118 base: https://git.kernel.org/pub/scm/linux/kernel/git/nvdimm/nvdimm.git libnvdimm-for-next patch link: https://lore.kernel.org/r/20240306102846.1020868-4-lizhijian%40fujitsu.com patch subject: [PATCH v3 3/7] nvdimm: pmem: assign a parent resource for vmemmap region for the fsdax config: loongarch-allmodconfig (https://download.01.org/0day-ci/archive/20240307/202403071804.b9EgMxWo-lkp@intel.com/config) compiler: loongarch64-linux-gcc (GCC) 13.2.0 reproduce (this is a W=1 build): (https://download.01.org/0day-ci/archive/20240307/202403071804.b9EgMxWo-lkp@intel.com/reproduce) If you fix the issue in a separate patch/commit (i.e. not just a new version of the same patch/commit), kindly add following tags | Reported-by: kernel test robot <lkp@intel.com> | Closes: https://lore.kernel.org/oe-kbuild-all/202403071804.b9EgMxWo-lkp@intel.com/ All errors (new ones prefixed by >>): drivers/nvdimm/pmem.c: In function 'pmem_attach_disk': >> drivers/nvdimm/pmem.c:501:9: error: implicit declaration of function 'pgmap_parent_resource' [-Werror=implicit-function-declaration] 501 | pgmap_parent_resource(&pmem->pgmap, parent); | ^~~~~~~~~~~~~~~~~~~~~ cc1: some warnings being treated as errors vim +/pgmap_parent_resource +501 drivers/nvdimm/pmem.c 448 449 static int pmem_attach_disk(struct device *dev, 450 struct nd_namespace_common *ndns) 451 { 452 struct nd_namespace_io *nsio = to_nd_namespace_io(&ndns->dev); 453 struct nd_region *nd_region = to_nd_region(dev->parent); 454 int nid = dev_to_node(dev), fua; 455 struct resource *res = &nsio->res, *parent; 456 struct range bb_range; 457 struct nd_pfn *nd_pfn = NULL; 458 struct dax_device *dax_dev; 459 struct nd_pfn_sb *pfn_sb; 460 struct pmem_device *pmem; 461 struct request_queue *q; 462 struct gendisk *disk; 463 void *addr; 464 int rc; 465 466 pmem = devm_kzalloc(dev, sizeof(*pmem), GFP_KERNEL); 467 if (!pmem) 468 return -ENOMEM; 469 470 rc = devm_namespace_enable(dev, ndns, nd_info_block_reserve()); 471 if (rc) 472 return rc; 473 474 /* while nsio_rw_bytes is active, parse a pfn info block if present */ 475 if (is_nd_pfn(dev)) { 476 nd_pfn = to_nd_pfn(dev); 477 rc = nvdimm_setup_pfn(nd_pfn, &pmem->pgmap); 478 if (rc) 479 return rc; 480 } 481 482 /* we're attaching a block device, disable raw namespace access */ 483 devm_namespace_disable(dev, ndns); 484 485 dev_set_drvdata(dev, pmem); 486 pmem->phys_addr = res->start; 487 pmem->size = resource_size(res); 488 fua = nvdimm_has_flush(nd_region); 489 if (!IS_ENABLED(CONFIG_ARCH_HAS_UACCESS_FLUSHCACHE) || fua < 0) { 490 dev_warn(dev, "unable to guarantee persistence of writes\n"); 491 fua = 0; 492 } 493 494 parent = devm_request_mem_region(dev, res->start, resource_size(res), 495 dev_name(&ndns->dev)); 496 if (!res) { 497 dev_warn(dev, "could not reserve region %pR\n", res); 498 return -EBUSY; 499 } 500 > 501 pgmap_parent_resource(&pmem->pgmap, parent); 502 503 disk = blk_alloc_disk(nid); 504 if (!disk) 505 return -ENOMEM; 506 q = disk->queue; 507 508 pmem->disk = disk; 509 pmem->pgmap.owner = pmem; 510 pmem->pfn_flags = PFN_DEV; 511 if (is_nd_pfn(dev)) { 512 pmem->pgmap.type = MEMORY_DEVICE_FS_DAX; 513 pmem->pgmap.ops = &fsdax_pagemap_ops; 514 addr = devm_memremap_pages(dev, &pmem->pgmap); 515 pfn_sb = nd_pfn->pfn_sb; 516 pmem->data_offset = le64_to_cpu(pfn_sb->dataoff); 517 pmem->pfn_pad = resource_size(res) - 518 range_len(&pmem->pgmap.range); 519 pmem->pfn_flags |= PFN_MAP; 520 bb_range = pmem->pgmap.range; 521 bb_range.start += pmem->data_offset; 522 } else if (pmem_should_map_pages(dev)) { 523 pmem->pgmap.range.start = res->start; 524 pmem->pgmap.range.end = res->end; 525 pmem->pgmap.nr_range = 1; 526 pmem->pgmap.type = MEMORY_DEVICE_FS_DAX; 527 pmem->pgmap.ops = &fsdax_pagemap_ops; 528 addr = devm_memremap_pages(dev, &pmem->pgmap); 529 pmem->pfn_flags |= PFN_MAP; 530 bb_range = pmem->pgmap.range; 531 } else { 532 addr = devm_memremap(dev, pmem->phys_addr, 533 pmem->size, ARCH_MEMREMAP_PMEM); 534 bb_range.start = res->start; 535 bb_range.end = res->end; 536 } 537 538 if (IS_ERR(addr)) { 539 rc = PTR_ERR(addr); 540 goto out; 541 } 542 pmem->virt_addr = addr; 543 544 blk_queue_write_cache(q, true, fua); 545 blk_queue_physical_block_size(q, PAGE_SIZE); 546 blk_queue_logical_block_size(q, pmem_sector_size(ndns)); 547 blk_queue_max_hw_sectors(q, UINT_MAX); 548 blk_queue_flag_set(QUEUE_FLAG_NONROT, q); 549 blk_queue_flag_set(QUEUE_FLAG_SYNCHRONOUS, q); 550 if (pmem->pfn_flags & PFN_MAP) 551 blk_queue_flag_set(QUEUE_FLAG_DAX, q); 552 553 disk->fops = &pmem_fops; 554 disk->private_data = pmem; 555 nvdimm_namespace_disk_name(ndns, disk->disk_name); 556 set_capacity(disk, (pmem->size - pmem->pfn_pad - pmem->data_offset) 557 / 512); 558 if (devm_init_badblocks(dev, &pmem->bb)) 559 return -ENOMEM; 560 nvdimm_badblocks_populate(nd_region, &pmem->bb, &bb_range); 561 disk->bb = &pmem->bb; 562 563 dax_dev = alloc_dax(pmem, &pmem_dax_ops); 564 if (IS_ERR(dax_dev)) { 565 rc = PTR_ERR(dax_dev); 566 goto out; 567 } 568 set_dax_nocache(dax_dev); 569 set_dax_nomc(dax_dev); 570 if (is_nvdimm_sync(nd_region)) 571 set_dax_synchronous(dax_dev); 572 rc = dax_add_host(dax_dev, disk); 573 if (rc) 574 goto out_cleanup_dax; 575 dax_write_cache(dax_dev, nvdimm_has_cache(nd_region)); 576 pmem->dax_dev = dax_dev; 577 578 rc = device_add_disk(dev, disk, pmem_attribute_groups); 579 if (rc) 580 goto out_remove_host; 581 if (devm_add_action_or_reset(dev, pmem_release_disk, pmem)) 582 return -ENOMEM; 583 584 nvdimm_check_and_set_ro(disk); 585 586 pmem->bb_state = sysfs_get_dirent(disk_to_dev(disk)->kobj.sd, 587 "badblocks"); 588 if (!pmem->bb_state) 589 dev_warn(dev, "'badblocks' notification disabled\n"); 590 return 0; 591 592 out_remove_host: 593 dax_remove_host(pmem->disk); 594 out_cleanup_dax: 595 kill_dax(pmem->dax_dev); 596 put_dax(pmem->dax_dev); 597 out: 598 put_disk(pmem->disk); 599 return rc; 600 } 601
diff --git a/drivers/nvdimm/pmem.c b/drivers/nvdimm/pmem.c index 4e8fdcb3f1c8..b2640a3fb693 100644 --- a/drivers/nvdimm/pmem.c +++ b/drivers/nvdimm/pmem.c @@ -452,7 +452,7 @@ static int pmem_attach_disk(struct device *dev, struct nd_namespace_io *nsio = to_nd_namespace_io(&ndns->dev); struct nd_region *nd_region = to_nd_region(dev->parent); int nid = dev_to_node(dev), fua; - struct resource *res = &nsio->res; + struct resource *res = &nsio->res, *parent; struct range bb_range; struct nd_pfn *nd_pfn = NULL; struct dax_device *dax_dev; @@ -491,12 +491,15 @@ static int pmem_attach_disk(struct device *dev, fua = 0; } - if (!devm_request_mem_region(dev, res->start, resource_size(res), - dev_name(&ndns->dev))) { + parent = devm_request_mem_region(dev, res->start, resource_size(res), + dev_name(&ndns->dev)); + if (!res) { dev_warn(dev, "could not reserve region %pR\n", res); return -EBUSY; } + pgmap_parent_resource(&pmem->pgmap, parent); + disk = blk_alloc_disk(nid); if (!disk) return -ENOMEM;
When the pmem is configured as fsdax, set the vmemmap region as a child of the namespace region so that it can be registered as a separate resource later. CC: Dan Williams <dan.j.williams@intel.com> CC: Vishal Verma <vishal.l.verma@intel.com> CC: Dave Jiang <dave.jiang@intel.com> CC: Ira Weiny <ira.weiny@intel.com> CC: Baoquan He <bhe@redhat.com> CC: nvdimm@lists.linux.dev Signed-off-by: Li Zhijian <lizhijian@fujitsu.com> --- drivers/nvdimm/pmem.c | 9 ++++++--- 1 file changed, 6 insertions(+), 3 deletions(-)