Commit a12562bb authored by Dan Williams's avatar Dan Williams

cxl/mem: Merge cxl_dvsec_ranges() and cxl_hdm_decode_init()

In preparation for changing how the driver handles 'mem_enable' in the CXL
DVSEC control register. Merge the contents of cxl_hdm_decode_init() into
cxl_dvsec_ranges() and rename the combined function cxl_hdm_decode_init().
The possible cleanups and fixes that result from this merge are saved for a
follow-on change.
Reviewed-by: default avatarIra Weiny <ira.weiny@intel.com>
Link: https://lore.kernel.org/r/165291690027.1426646.10249756632415633752.stgit@dwillia2-xfhSigned-off-by: default avatarDan Williams <dan.j.williams@intel.com>
parent dd2d42ad
...@@ -175,12 +175,70 @@ static int wait_for_valid(struct cxl_dev_state *cxlds) ...@@ -175,12 +175,70 @@ static int wait_for_valid(struct cxl_dev_state *cxlds)
return -ETIMEDOUT; return -ETIMEDOUT;
} }
/* static bool __cxl_hdm_decode_init(struct cxl_dev_state *cxlds,
* Return positive number of non-zero ranges on success and a negative struct cxl_endpoint_dvsec_info *info)
* error code on failure. The cxl_mem driver depends on ranges == 0 to {
* init HDM operation. struct cxl_register_map map;
struct cxl_component_reg_map *cmap = &map.component_map;
bool global_enable, retval = false;
void __iomem *crb;
u32 global_ctrl;
/* map hdm decoder */
crb = ioremap(cxlds->component_reg_phys, CXL_COMPONENT_REG_BLOCK_SIZE);
if (!crb) {
dev_dbg(cxlds->dev, "Failed to map component registers\n");
return false;
}
cxl_probe_component_regs(cxlds->dev, crb, cmap);
if (!cmap->hdm_decoder.valid) {
dev_dbg(cxlds->dev, "Invalid HDM decoder registers\n");
goto out;
}
global_ctrl = readl(crb + cmap->hdm_decoder.offset +
CXL_HDM_DECODER_CTRL_OFFSET);
global_enable = global_ctrl & CXL_HDM_DECODER_ENABLE;
/*
* Per CXL 2.0 Section 8.1.3.8.3 and 8.1.3.8.4 DVSEC CXL Range 1 Base
* [High,Low] when HDM operation is enabled the range register values
* are ignored by the device, but the spec also recommends matching the
* DVSEC Range 1,2 to HDM Decoder Range 0,1. So, non-zero info->ranges
* are expected even though Linux does not require or maintain that
* match.
*/ */
int cxl_dvsec_ranges(struct cxl_dev_state *cxlds, if (!global_enable && info->mem_enabled && info->ranges)
goto out;
retval = true;
/*
* Permanently (for this boot at least) opt the device into HDM
* operation. Individual HDM decoders still need to be enabled after
* this point.
*/
if (!global_enable) {
dev_dbg(cxlds->dev, "Enabling HDM decode\n");
writel(global_ctrl | CXL_HDM_DECODER_ENABLE,
crb + cmap->hdm_decoder.offset +
CXL_HDM_DECODER_CTRL_OFFSET);
}
out:
iounmap(crb);
return retval;
}
/**
* cxl_hdm_decode_init() - Setup HDM decoding for the endpoint
* @cxlds: Device state
* @info: DVSEC Range cached enumeration
*
* Try to enable the endpoint's HDM Decoder Capability
*/
int cxl_hdm_decode_init(struct cxl_dev_state *cxlds,
struct cxl_endpoint_dvsec_info *info) struct cxl_endpoint_dvsec_info *info)
{ {
struct pci_dev *pdev = to_pci_dev(cxlds->dev); struct pci_dev *pdev = to_pci_dev(cxlds->dev);
...@@ -270,6 +328,16 @@ int cxl_dvsec_ranges(struct cxl_dev_state *cxlds, ...@@ -270,6 +328,16 @@ int cxl_dvsec_ranges(struct cxl_dev_state *cxlds,
info->ranges = ranges; info->ranges = ranges;
/*
* If DVSEC ranges are being used instead of HDM decoder registers there
* is no use in trying to manage those.
*/
if (!__cxl_hdm_decode_init(cxlds, info)) {
dev_err(dev,
"Legacy range registers configuration prevents HDM operation.\n");
return -EBUSY;
}
return 0; return 0;
} }
EXPORT_SYMBOL_NS_GPL(cxl_dvsec_ranges, CXL); EXPORT_SYMBOL_NS_GPL(cxl_hdm_decode_init, CXL);
...@@ -74,6 +74,6 @@ static inline resource_size_t cxl_regmap_to_base(struct pci_dev *pdev, ...@@ -74,6 +74,6 @@ static inline resource_size_t cxl_regmap_to_base(struct pci_dev *pdev,
int devm_cxl_port_enumerate_dports(struct cxl_port *port); int devm_cxl_port_enumerate_dports(struct cxl_port *port);
struct cxl_dev_state; struct cxl_dev_state;
struct cxl_endpoint_dvsec_info; struct cxl_endpoint_dvsec_info;
int cxl_dvsec_ranges(struct cxl_dev_state *cxlds, int cxl_hdm_decode_init(struct cxl_dev_state *cxlds,
struct cxl_endpoint_dvsec_info *info); struct cxl_endpoint_dvsec_info *info);
#endif /* __CXL_PCI_H__ */ #endif /* __CXL_PCI_H__ */
...@@ -46,74 +46,6 @@ static int create_endpoint(struct cxl_memdev *cxlmd, ...@@ -46,74 +46,6 @@ static int create_endpoint(struct cxl_memdev *cxlmd,
return cxl_endpoint_autoremove(cxlmd, endpoint); return cxl_endpoint_autoremove(cxlmd, endpoint);
} }
/**
* cxl_hdm_decode_init() - Setup HDM decoding for the endpoint
* @cxlds: Device state
*
* Additionally, enables global HDM decoding. Warning: don't call this outside
* of probe. Once probe is complete, the port driver owns all access to the HDM
* decoder registers.
*
* Returns: false if DVSEC Ranges are being used instead of HDM
* decoders, or if it can not be determined if DVSEC Ranges are in use.
* Otherwise, returns true.
*/
__mock bool cxl_hdm_decode_init(struct cxl_dev_state *cxlds,
struct cxl_endpoint_dvsec_info *info)
{
struct cxl_register_map map;
struct cxl_component_reg_map *cmap = &map.component_map;
bool global_enable, retval = false;
void __iomem *crb;
u32 global_ctrl;
/* map hdm decoder */
crb = ioremap(cxlds->component_reg_phys, CXL_COMPONENT_REG_BLOCK_SIZE);
if (!crb) {
dev_dbg(cxlds->dev, "Failed to map component registers\n");
return false;
}
cxl_probe_component_regs(cxlds->dev, crb, cmap);
if (!cmap->hdm_decoder.valid) {
dev_dbg(cxlds->dev, "Invalid HDM decoder registers\n");
goto out;
}
global_ctrl = readl(crb + cmap->hdm_decoder.offset +
CXL_HDM_DECODER_CTRL_OFFSET);
global_enable = global_ctrl & CXL_HDM_DECODER_ENABLE;
/*
* Per CXL 2.0 Section 8.1.3.8.3 and 8.1.3.8.4 DVSEC CXL Range 1 Base
* [High,Low] when HDM operation is enabled the range register values
* are ignored by the device, but the spec also recommends matching the
* DVSEC Range 1,2 to HDM Decoder Range 0,1. So, non-zero info->ranges
* are expected even though Linux does not require or maintain that
* match.
*/
if (!global_enable && info->mem_enabled && info->ranges)
goto out;
retval = true;
/*
* Permanently (for this boot at least) opt the device into HDM
* operation. Individual HDM decoders still need to be enabled after
* this point.
*/
if (!global_enable) {
dev_dbg(cxlds->dev, "Enabling HDM decode\n");
writel(global_ctrl | CXL_HDM_DECODER_ENABLE,
crb + cmap->hdm_decoder.offset +
CXL_HDM_DECODER_CTRL_OFFSET);
}
out:
iounmap(crb);
return retval;
}
static void enable_suspend(void *data) static void enable_suspend(void *data)
{ {
cxl_mem_active_dec(); cxl_mem_active_dec();
...@@ -163,7 +95,7 @@ static int cxl_mem_probe(struct device *dev) ...@@ -163,7 +95,7 @@ static int cxl_mem_probe(struct device *dev)
if (rc) if (rc)
return rc; return rc;
rc = cxl_dvsec_ranges(cxlds, &info); rc = cxl_hdm_decode_init(cxlds, &info);
if (rc) if (rc)
return rc; return rc;
...@@ -173,16 +105,6 @@ static int cxl_mem_probe(struct device *dev) ...@@ -173,16 +105,6 @@ static int cxl_mem_probe(struct device *dev)
return rc; return rc;
} }
/*
* If DVSEC ranges are being used instead of HDM decoder registers there
* is no use in trying to manage those.
*/
if (!cxl_hdm_decode_init(cxlds, &info)) {
dev_err(dev,
"Legacy range registers configuration prevents HDM operation.\n");
return -EBUSY;
}
/* /*
* The kernel may be operating out of CXL memory on this device, * The kernel may be operating out of CXL memory on this device,
* there is no spec defined way to determine whether this device * there is no spec defined way to determine whether this device
......
...@@ -9,7 +9,7 @@ ldflags-y += --wrap=devm_cxl_setup_hdm ...@@ -9,7 +9,7 @@ ldflags-y += --wrap=devm_cxl_setup_hdm
ldflags-y += --wrap=devm_cxl_add_passthrough_decoder ldflags-y += --wrap=devm_cxl_add_passthrough_decoder
ldflags-y += --wrap=devm_cxl_enumerate_decoders ldflags-y += --wrap=devm_cxl_enumerate_decoders
ldflags-y += --wrap=cxl_await_media_ready ldflags-y += --wrap=cxl_await_media_ready
ldflags-y += --wrap=cxl_dvsec_ranges ldflags-y += --wrap=cxl_hdm_decode_init
DRIVERS := ../../../drivers DRIVERS := ../../../drivers
CXL_SRC := $(DRIVERS)/cxl CXL_SRC := $(DRIVERS)/cxl
...@@ -36,7 +36,6 @@ cxl_port-y += config_check.o ...@@ -36,7 +36,6 @@ cxl_port-y += config_check.o
obj-m += cxl_mem.o obj-m += cxl_mem.o
cxl_mem-y := $(CXL_SRC)/mem.o cxl_mem-y := $(CXL_SRC)/mem.o
cxl_mem-y += mock_mem.o
cxl_mem-y += config_check.o cxl_mem-y += config_check.o
obj-m += cxl_core.o obj-m += cxl_core.o
......
// SPDX-License-Identifier: GPL-2.0-only
/* Copyright(c) 2022 Intel Corporation. All rights reserved. */
#include <linux/types.h>
struct cxl_dev_state;
bool cxl_hdm_decode_init(struct cxl_dev_state *cxlds)
{
return true;
}
...@@ -208,7 +208,7 @@ int __wrap_cxl_await_media_ready(struct cxl_dev_state *cxlds) ...@@ -208,7 +208,7 @@ int __wrap_cxl_await_media_ready(struct cxl_dev_state *cxlds)
} }
EXPORT_SYMBOL_NS_GPL(__wrap_cxl_await_media_ready, CXL); EXPORT_SYMBOL_NS_GPL(__wrap_cxl_await_media_ready, CXL);
int __wrap_cxl_dvsec_ranges(struct cxl_dev_state *cxlds, int __wrap_cxl_hdm_decode_init(struct cxl_dev_state *cxlds,
struct cxl_endpoint_dvsec_info *info) struct cxl_endpoint_dvsec_info *info)
{ {
int rc = 0, index; int rc = 0, index;
...@@ -217,12 +217,12 @@ int __wrap_cxl_dvsec_ranges(struct cxl_dev_state *cxlds, ...@@ -217,12 +217,12 @@ int __wrap_cxl_dvsec_ranges(struct cxl_dev_state *cxlds,
if (ops && ops->is_mock_dev(cxlds->dev)) if (ops && ops->is_mock_dev(cxlds->dev))
info->mem_enabled = 1; info->mem_enabled = 1;
else else
rc = cxl_dvsec_ranges(cxlds, info); rc = cxl_hdm_decode_init(cxlds, info);
put_cxl_mock_ops(index); put_cxl_mock_ops(index);
return rc; return rc;
} }
EXPORT_SYMBOL_NS_GPL(__wrap_cxl_dvsec_ranges, CXL); EXPORT_SYMBOL_NS_GPL(__wrap_cxl_hdm_decode_init, CXL);
MODULE_LICENSE("GPL v2"); MODULE_LICENSE("GPL v2");
MODULE_IMPORT_NS(ACPI); MODULE_IMPORT_NS(ACPI);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment