Commit 85e4b30a authored by Bjorn Helgaas's avatar Bjorn Helgaas

Merge branch 'pci/host-hv' into next

* pci/host-hv:
  PCI: hv: Use vPCI protocol version 1.2
  PCI: hv: Add vPCI version protocol negotiation
  PCI: hv: Temporary own CPU-number-to-vCPU-number infra
  PCI: hv: Use page allocation for hbus structure
  PCI: hv: Fix comment formatting and use proper integer fields
parents 6f65daed 7dcf90e9
...@@ -152,6 +152,12 @@ ...@@ -152,6 +152,12 @@
*/ */
#define HV_X64_DEPRECATING_AEOI_RECOMMENDED (1 << 9) #define HV_X64_DEPRECATING_AEOI_RECOMMENDED (1 << 9)
/*
* HV_VP_SET available
*/
#define HV_X64_EX_PROCESSOR_MASKS_RECOMMENDED (1 << 11)
/* /*
* Crash notification flag. * Crash notification flag.
*/ */
......
...@@ -64,22 +64,39 @@ ...@@ -64,22 +64,39 @@
* major version. * major version.
*/ */
#define PCI_MAKE_VERSION(major, minor) ((u32)(((major) << 16) | (major))) #define PCI_MAKE_VERSION(major, minor) ((u32)(((major) << 16) | (minor)))
#define PCI_MAJOR_VERSION(version) ((u32)(version) >> 16) #define PCI_MAJOR_VERSION(version) ((u32)(version) >> 16)
#define PCI_MINOR_VERSION(version) ((u32)(version) & 0xff) #define PCI_MINOR_VERSION(version) ((u32)(version) & 0xff)
enum { enum pci_protocol_version_t {
PCI_PROTOCOL_VERSION_1_1 = PCI_MAKE_VERSION(1, 1), PCI_PROTOCOL_VERSION_1_1 = PCI_MAKE_VERSION(1, 1), /* Win10 */
PCI_PROTOCOL_VERSION_CURRENT = PCI_PROTOCOL_VERSION_1_1 PCI_PROTOCOL_VERSION_1_2 = PCI_MAKE_VERSION(1, 2), /* RS1 */
}; };
#define CPU_AFFINITY_ALL -1ULL #define CPU_AFFINITY_ALL -1ULL
/*
* Supported protocol versions in the order of probing - highest go
* first.
*/
static enum pci_protocol_version_t pci_protocol_versions[] = {
PCI_PROTOCOL_VERSION_1_2,
PCI_PROTOCOL_VERSION_1_1,
};
/*
* Protocol version negotiated by hv_pci_protocol_negotiation().
*/
static enum pci_protocol_version_t pci_protocol_version;
#define PCI_CONFIG_MMIO_LENGTH 0x2000 #define PCI_CONFIG_MMIO_LENGTH 0x2000
#define CFG_PAGE_OFFSET 0x1000 #define CFG_PAGE_OFFSET 0x1000
#define CFG_PAGE_SIZE (PCI_CONFIG_MMIO_LENGTH - CFG_PAGE_OFFSET) #define CFG_PAGE_SIZE (PCI_CONFIG_MMIO_LENGTH - CFG_PAGE_OFFSET)
#define MAX_SUPPORTED_MSI_MESSAGES 0x400 #define MAX_SUPPORTED_MSI_MESSAGES 0x400
#define STATUS_REVISION_MISMATCH 0xC0000059
/* /*
* Message Types * Message Types
*/ */
...@@ -109,6 +126,9 @@ enum pci_message_type { ...@@ -109,6 +126,9 @@ enum pci_message_type {
PCI_QUERY_PROTOCOL_VERSION = PCI_MESSAGE_BASE + 0x13, PCI_QUERY_PROTOCOL_VERSION = PCI_MESSAGE_BASE + 0x13,
PCI_CREATE_INTERRUPT_MESSAGE = PCI_MESSAGE_BASE + 0x14, PCI_CREATE_INTERRUPT_MESSAGE = PCI_MESSAGE_BASE + 0x14,
PCI_DELETE_INTERRUPT_MESSAGE = PCI_MESSAGE_BASE + 0x15, PCI_DELETE_INTERRUPT_MESSAGE = PCI_MESSAGE_BASE + 0x15,
PCI_RESOURCES_ASSIGNED2 = PCI_MESSAGE_BASE + 0x16,
PCI_CREATE_INTERRUPT_MESSAGE2 = PCI_MESSAGE_BASE + 0x17,
PCI_DELETE_INTERRUPT_MESSAGE2 = PCI_MESSAGE_BASE + 0x18, /* unused */
PCI_MESSAGE_MAXIMUM PCI_MESSAGE_MAXIMUM
}; };
...@@ -178,6 +198,30 @@ struct hv_msi_desc { ...@@ -178,6 +198,30 @@ struct hv_msi_desc {
u64 cpu_mask; u64 cpu_mask;
} __packed; } __packed;
/**
* struct hv_msi_desc2 - 1.2 version of hv_msi_desc
* @vector: IDT entry
* @delivery_mode: As defined in Intel's Programmer's
* Reference Manual, Volume 3, Chapter 8.
* @vector_count: Number of contiguous entries in the
* Interrupt Descriptor Table that are
* occupied by this Message-Signaled
* Interrupt. For "MSI", as first defined
* in PCI 2.2, this can be between 1 and
* 32. For "MSI-X," as first defined in PCI
* 3.0, this must be 1, as each MSI-X table
* entry would have its own descriptor.
* @processor_count: number of bits enabled in array.
* @processor_array: All the target virtual processors.
*/
struct hv_msi_desc2 {
u8 vector;
u8 delivery_mode;
u16 vector_count;
u16 processor_count;
u16 processor_array[32];
} __packed;
/** /**
* struct tran_int_desc * struct tran_int_desc
* @reserved: unused, padding * @reserved: unused, padding
...@@ -245,7 +289,7 @@ struct pci_packet { ...@@ -245,7 +289,7 @@ struct pci_packet {
struct pci_version_request { struct pci_version_request {
struct pci_message message_type; struct pci_message message_type;
enum pci_message_type protocol_version; u32 protocol_version;
} __packed; } __packed;
/* /*
...@@ -294,6 +338,14 @@ struct pci_resources_assigned { ...@@ -294,6 +338,14 @@ struct pci_resources_assigned {
u32 reserved[4]; u32 reserved[4];
} __packed; } __packed;
struct pci_resources_assigned2 {
struct pci_message message_type;
union win_slot_encoding wslot;
u8 memory_range[0x14][6]; /* not used here */
u32 msi_descriptor_count;
u8 reserved[70];
} __packed;
struct pci_create_interrupt { struct pci_create_interrupt {
struct pci_message message_type; struct pci_message message_type;
union win_slot_encoding wslot; union win_slot_encoding wslot;
...@@ -306,6 +358,12 @@ struct pci_create_int_response { ...@@ -306,6 +358,12 @@ struct pci_create_int_response {
struct tran_int_desc int_desc; struct tran_int_desc int_desc;
} __packed; } __packed;
struct pci_create_interrupt2 {
struct pci_message message_type;
union win_slot_encoding wslot;
struct hv_msi_desc2 int_desc;
} __packed;
struct pci_delete_interrupt { struct pci_delete_interrupt {
struct pci_message message_type; struct pci_message message_type;
union win_slot_encoding wslot; union win_slot_encoding wslot;
...@@ -331,17 +389,42 @@ static int pci_ring_size = (4 * PAGE_SIZE); ...@@ -331,17 +389,42 @@ static int pci_ring_size = (4 * PAGE_SIZE);
#define HV_PARTITION_ID_SELF ((u64)-1) #define HV_PARTITION_ID_SELF ((u64)-1)
#define HVCALL_RETARGET_INTERRUPT 0x7e #define HVCALL_RETARGET_INTERRUPT 0x7e
struct retarget_msi_interrupt { struct hv_interrupt_entry {
u64 partition_id; /* use "self" */
u64 device_id;
u32 source; /* 1 for MSI(-X) */ u32 source; /* 1 for MSI(-X) */
u32 reserved1; u32 reserved1;
u32 address; u32 address;
u32 data; u32 data;
u64 reserved2; };
#define HV_VP_SET_BANK_COUNT_MAX 5 /* current implementation limit */
struct hv_vp_set {
u64 format; /* 0 (HvGenericSetSparse4k) */
u64 valid_banks;
u64 masks[HV_VP_SET_BANK_COUNT_MAX];
};
/*
* flags for hv_device_interrupt_target.flags
*/
#define HV_DEVICE_INTERRUPT_TARGET_MULTICAST 1
#define HV_DEVICE_INTERRUPT_TARGET_PROCESSOR_SET 2
struct hv_device_interrupt_target {
u32 vector; u32 vector;
u32 flags; u32 flags;
union {
u64 vp_mask; u64 vp_mask;
struct hv_vp_set vp_set;
};
};
struct retarget_msi_interrupt {
u64 partition_id; /* use "self" */
u64 device_id;
struct hv_interrupt_entry int_entry;
u64 reserved2;
struct hv_device_interrupt_target int_target;
} __packed; } __packed;
/* /*
...@@ -382,7 +465,10 @@ struct hv_pcibus_device { ...@@ -382,7 +465,10 @@ struct hv_pcibus_device {
struct msi_domain_info msi_info; struct msi_domain_info msi_info;
struct msi_controller msi_chip; struct msi_controller msi_chip;
struct irq_domain *irq_domain; struct irq_domain *irq_domain;
/* hypercall arg, must not cross page boundary */
struct retarget_msi_interrupt retarget_msi_interrupt_params; struct retarget_msi_interrupt retarget_msi_interrupt_params;
spinlock_t retarget_msi_interrupt_lock; spinlock_t retarget_msi_interrupt_lock;
}; };
...@@ -476,6 +562,52 @@ static void put_pcichild(struct hv_pci_dev *hv_pcidev, ...@@ -476,6 +562,52 @@ static void put_pcichild(struct hv_pci_dev *hv_pcidev,
static void get_hvpcibus(struct hv_pcibus_device *hv_pcibus); static void get_hvpcibus(struct hv_pcibus_device *hv_pcibus);
static void put_hvpcibus(struct hv_pcibus_device *hv_pcibus); static void put_hvpcibus(struct hv_pcibus_device *hv_pcibus);
/*
* Temporary CPU to vCPU mapping to address transitioning
* vmbus_cpu_number_to_vp_number() being migrated to
* hv_cpu_number_to_vp_number() in a separate patch. Once that patch
* has been picked up in the main line, remove this code here and use
* the official code.
*/
static struct hv_tmpcpumap
{
bool initialized;
u32 vp_index[NR_CPUS];
} hv_tmpcpumap;
static void hv_tmpcpumap_init_cpu(void *_unused)
{
int cpu = smp_processor_id();
u64 vp_index;
hv_get_vp_index(vp_index);
hv_tmpcpumap.vp_index[cpu] = vp_index;
}
static void hv_tmpcpumap_init(void)
{
if (hv_tmpcpumap.initialized)
return;
memset(hv_tmpcpumap.vp_index, -1, sizeof(hv_tmpcpumap.vp_index));
on_each_cpu(hv_tmpcpumap_init_cpu, NULL, true);
hv_tmpcpumap.initialized = true;
}
/**
* hv_tmp_cpu_nr_to_vp_nr() - Convert Linux CPU nr to Hyper-V vCPU nr
*
* Remove once vmbus_cpu_number_to_vp_number() has been converted to
* hv_cpu_number_to_vp_number() and replace callers appropriately.
*/
static u32 hv_tmp_cpu_nr_to_vp_nr(int cpu)
{
return hv_tmpcpumap.vp_index[cpu];
}
/** /**
* devfn_to_wslot() - Convert from Linux PCI slot to Windows * devfn_to_wslot() - Convert from Linux PCI slot to Windows
* @devfn: The Linux representation of PCI slot * @devfn: The Linux representation of PCI slot
...@@ -786,8 +918,11 @@ static void hv_irq_unmask(struct irq_data *data) ...@@ -786,8 +918,11 @@ static void hv_irq_unmask(struct irq_data *data)
struct cpumask *dest; struct cpumask *dest;
struct pci_bus *pbus; struct pci_bus *pbus;
struct pci_dev *pdev; struct pci_dev *pdev;
int cpu;
unsigned long flags; unsigned long flags;
u32 var_size = 0;
int cpu_vmbus;
int cpu;
u64 res;
dest = irq_data_get_affinity_mask(data); dest = irq_data_get_affinity_mask(data);
pdev = msi_desc_to_pci_dev(msi_desc); pdev = msi_desc_to_pci_dev(msi_desc);
...@@ -799,23 +934,74 @@ static void hv_irq_unmask(struct irq_data *data) ...@@ -799,23 +934,74 @@ static void hv_irq_unmask(struct irq_data *data)
params = &hbus->retarget_msi_interrupt_params; params = &hbus->retarget_msi_interrupt_params;
memset(params, 0, sizeof(*params)); memset(params, 0, sizeof(*params));
params->partition_id = HV_PARTITION_ID_SELF; params->partition_id = HV_PARTITION_ID_SELF;
params->source = 1; /* MSI(-X) */ params->int_entry.source = 1; /* MSI(-X) */
params->address = msi_desc->msg.address_lo; params->int_entry.address = msi_desc->msg.address_lo;
params->data = msi_desc->msg.data; params->int_entry.data = msi_desc->msg.data;
params->device_id = (hbus->hdev->dev_instance.b[5] << 24) | params->device_id = (hbus->hdev->dev_instance.b[5] << 24) |
(hbus->hdev->dev_instance.b[4] << 16) | (hbus->hdev->dev_instance.b[4] << 16) |
(hbus->hdev->dev_instance.b[7] << 8) | (hbus->hdev->dev_instance.b[7] << 8) |
(hbus->hdev->dev_instance.b[6] & 0xf8) | (hbus->hdev->dev_instance.b[6] & 0xf8) |
PCI_FUNC(pdev->devfn); PCI_FUNC(pdev->devfn);
params->vector = cfg->vector; params->int_target.vector = cfg->vector;
/*
* Honoring apic->irq_delivery_mode set to dest_Fixed by
* setting the HV_DEVICE_INTERRUPT_TARGET_MULTICAST flag results in a
* spurious interrupt storm. Not doing so does not seem to have a
* negative effect (yet?).
*/
if (pci_protocol_version >= PCI_PROTOCOL_VERSION_1_2) {
/*
* PCI_PROTOCOL_VERSION_1_2 supports the VP_SET version of the
* HVCALL_RETARGET_INTERRUPT hypercall, which also coincides
* with >64 VP support.
* ms_hyperv.hints & HV_X64_EX_PROCESSOR_MASKS_RECOMMENDED
* is not sufficient for this hypercall.
*/
params->int_target.flags |=
HV_DEVICE_INTERRUPT_TARGET_PROCESSOR_SET;
params->int_target.vp_set.valid_banks =
(1ull << HV_VP_SET_BANK_COUNT_MAX) - 1;
for_each_cpu_and(cpu, dest, cpu_online_mask) /*
params->vp_mask |= (1ULL << vmbus_cpu_number_to_vp_number(cpu)); * var-sized hypercall, var-size starts after vp_mask (thus
* vp_set.format does not count, but vp_set.valid_banks does).
*/
var_size = 1 + HV_VP_SET_BANK_COUNT_MAX;
hv_do_hypercall(HVCALL_RETARGET_INTERRUPT, params, NULL); for_each_cpu_and(cpu, dest, cpu_online_mask) {
cpu_vmbus = hv_tmp_cpu_nr_to_vp_nr(cpu);
if (cpu_vmbus >= HV_VP_SET_BANK_COUNT_MAX * 64) {
dev_err(&hbus->hdev->device,
"too high CPU %d", cpu_vmbus);
res = 1;
goto exit_unlock;
}
params->int_target.vp_set.masks[cpu_vmbus / 64] |=
(1ULL << (cpu_vmbus & 63));
}
} else {
for_each_cpu_and(cpu, dest, cpu_online_mask) {
params->int_target.vp_mask |=
(1ULL << hv_tmp_cpu_nr_to_vp_nr(cpu));
}
}
res = hv_do_hypercall(HVCALL_RETARGET_INTERRUPT | (var_size << 17),
params, NULL);
exit_unlock:
spin_unlock_irqrestore(&hbus->retarget_msi_interrupt_lock, flags); spin_unlock_irqrestore(&hbus->retarget_msi_interrupt_lock, flags);
if (res) {
dev_err(&hbus->hdev->device,
"%s() failed: %#llx", __func__, res);
return;
}
pci_msi_unmask_irq(data); pci_msi_unmask_irq(data);
} }
...@@ -836,6 +1022,53 @@ static void hv_pci_compose_compl(void *context, struct pci_response *resp, ...@@ -836,6 +1022,53 @@ static void hv_pci_compose_compl(void *context, struct pci_response *resp,
complete(&comp_pkt->comp_pkt.host_event); complete(&comp_pkt->comp_pkt.host_event);
} }
static u32 hv_compose_msi_req_v1(
struct pci_create_interrupt *int_pkt, struct cpumask *affinity,
u32 slot, u8 vector)
{
int_pkt->message_type.type = PCI_CREATE_INTERRUPT_MESSAGE;
int_pkt->wslot.slot = slot;
int_pkt->int_desc.vector = vector;
int_pkt->int_desc.vector_count = 1;
int_pkt->int_desc.delivery_mode =
(apic->irq_delivery_mode == dest_LowestPrio) ?
dest_LowestPrio : dest_Fixed;
/*
* Create MSI w/ dummy vCPU set, overwritten by subsequent retarget in
* hv_irq_unmask().
*/
int_pkt->int_desc.cpu_mask = CPU_AFFINITY_ALL;
return sizeof(*int_pkt);
}
static u32 hv_compose_msi_req_v2(
struct pci_create_interrupt2 *int_pkt, struct cpumask *affinity,
u32 slot, u8 vector)
{
int cpu;
int_pkt->message_type.type = PCI_CREATE_INTERRUPT_MESSAGE2;
int_pkt->wslot.slot = slot;
int_pkt->int_desc.vector = vector;
int_pkt->int_desc.vector_count = 1;
int_pkt->int_desc.delivery_mode =
(apic->irq_delivery_mode == dest_LowestPrio) ?
dest_LowestPrio : dest_Fixed;
/*
* Create MSI w/ dummy vCPU set targeting just one vCPU, overwritten
* by subsequent retarget in hv_irq_unmask().
*/
cpu = cpumask_first_and(affinity, cpu_online_mask);
int_pkt->int_desc.processor_array[0] =
hv_tmp_cpu_nr_to_vp_nr(cpu);
int_pkt->int_desc.processor_count = 1;
return sizeof(*int_pkt);
}
/** /**
* hv_compose_msi_msg() - Supplies a valid MSI address/data * hv_compose_msi_msg() - Supplies a valid MSI address/data
* @data: Everything about this MSI * @data: Everything about this MSI
...@@ -854,15 +1087,17 @@ static void hv_compose_msi_msg(struct irq_data *data, struct msi_msg *msg) ...@@ -854,15 +1087,17 @@ static void hv_compose_msi_msg(struct irq_data *data, struct msi_msg *msg)
struct hv_pci_dev *hpdev; struct hv_pci_dev *hpdev;
struct pci_bus *pbus; struct pci_bus *pbus;
struct pci_dev *pdev; struct pci_dev *pdev;
struct pci_create_interrupt *int_pkt;
struct compose_comp_ctxt comp; struct compose_comp_ctxt comp;
struct tran_int_desc *int_desc; struct tran_int_desc *int_desc;
struct cpumask *affinity;
struct { struct {
struct pci_packet pkt; struct pci_packet pci_pkt;
u8 buffer[sizeof(struct pci_create_interrupt)]; union {
} ctxt; struct pci_create_interrupt v1;
int cpu; struct pci_create_interrupt2 v2;
} int_pkts;
} __packed ctxt;
u32 size;
int ret; int ret;
pdev = msi_desc_to_pci_dev(irq_data_get_msi_desc(data)); pdev = msi_desc_to_pci_dev(irq_data_get_msi_desc(data));
...@@ -885,36 +1120,44 @@ static void hv_compose_msi_msg(struct irq_data *data, struct msi_msg *msg) ...@@ -885,36 +1120,44 @@ static void hv_compose_msi_msg(struct irq_data *data, struct msi_msg *msg)
memset(&ctxt, 0, sizeof(ctxt)); memset(&ctxt, 0, sizeof(ctxt));
init_completion(&comp.comp_pkt.host_event); init_completion(&comp.comp_pkt.host_event);
ctxt.pkt.completion_func = hv_pci_compose_compl; ctxt.pci_pkt.completion_func = hv_pci_compose_compl;
ctxt.pkt.compl_ctxt = &comp; ctxt.pci_pkt.compl_ctxt = &comp;
int_pkt = (struct pci_create_interrupt *)&ctxt.pkt.message;
int_pkt->message_type.type = PCI_CREATE_INTERRUPT_MESSAGE; switch (pci_protocol_version) {
int_pkt->wslot.slot = hpdev->desc.win_slot.slot; case PCI_PROTOCOL_VERSION_1_1:
int_pkt->int_desc.vector = cfg->vector; size = hv_compose_msi_req_v1(&ctxt.int_pkts.v1,
int_pkt->int_desc.vector_count = 1; irq_data_get_affinity_mask(data),
int_pkt->int_desc.delivery_mode = hpdev->desc.win_slot.slot,
(apic->irq_delivery_mode == dest_LowestPrio) ? 1 : 0; cfg->vector);
break;
/* case PCI_PROTOCOL_VERSION_1_2:
* This bit doesn't have to work on machines with more than 64 size = hv_compose_msi_req_v2(&ctxt.int_pkts.v2,
* processors because Hyper-V only supports 64 in a guest. irq_data_get_affinity_mask(data),
hpdev->desc.win_slot.slot,
cfg->vector);
break;
default:
/* As we only negotiate protocol versions known to this driver,
* this path should never hit. However, this is it not a hot
* path so we print a message to aid future updates.
*/ */
affinity = irq_data_get_affinity_mask(data); dev_err(&hbus->hdev->device,
if (cpumask_weight(affinity) >= 32) { "Unexpected vPCI protocol, update driver.");
int_pkt->int_desc.cpu_mask = CPU_AFFINITY_ALL; goto free_int_desc;
} else {
for_each_cpu_and(cpu, affinity, cpu_online_mask) {
int_pkt->int_desc.cpu_mask |=
(1ULL << vmbus_cpu_number_to_vp_number(cpu));
}
} }
ret = vmbus_sendpacket(hpdev->hbus->hdev->channel, int_pkt, ret = vmbus_sendpacket(hpdev->hbus->hdev->channel, &ctxt.int_pkts,
sizeof(*int_pkt), (unsigned long)&ctxt.pkt, size, (unsigned long)&ctxt.pci_pkt,
VM_PKT_DATA_INBAND, VM_PKT_DATA_INBAND,
VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED); VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED);
if (ret) if (ret) {
dev_err(&hbus->hdev->device,
"Sending request for interrupt failed: 0x%x",
comp.comp_pkt.completion_status);
goto free_int_desc; goto free_int_desc;
}
wait_for_completion(&comp.comp_pkt.host_event); wait_for_completion(&comp.comp_pkt.host_event);
...@@ -1513,7 +1756,7 @@ static void pci_devices_present_work(struct work_struct *work) ...@@ -1513,7 +1756,7 @@ static void pci_devices_present_work(struct work_struct *work)
put_pcichild(hpdev, hv_pcidev_ref_initial); put_pcichild(hpdev, hv_pcidev_ref_initial);
} }
switch(hbus->state) { switch (hbus->state) {
case hv_pcibus_installed: case hv_pcibus_installed:
/* /*
* Tell the core to rescan bus * Tell the core to rescan bus
...@@ -1800,6 +2043,7 @@ static int hv_pci_protocol_negotiation(struct hv_device *hdev) ...@@ -1800,6 +2043,7 @@ static int hv_pci_protocol_negotiation(struct hv_device *hdev)
struct hv_pci_compl comp_pkt; struct hv_pci_compl comp_pkt;
struct pci_packet *pkt; struct pci_packet *pkt;
int ret; int ret;
int i;
/* /*
* Initiate the handshake with the host and negotiate * Initiate the handshake with the host and negotiate
...@@ -1816,26 +2060,44 @@ static int hv_pci_protocol_negotiation(struct hv_device *hdev) ...@@ -1816,26 +2060,44 @@ static int hv_pci_protocol_negotiation(struct hv_device *hdev)
pkt->compl_ctxt = &comp_pkt; pkt->compl_ctxt = &comp_pkt;
version_req = (struct pci_version_request *)&pkt->message; version_req = (struct pci_version_request *)&pkt->message;
version_req->message_type.type = PCI_QUERY_PROTOCOL_VERSION; version_req->message_type.type = PCI_QUERY_PROTOCOL_VERSION;
version_req->protocol_version = PCI_PROTOCOL_VERSION_CURRENT;
for (i = 0; i < ARRAY_SIZE(pci_protocol_versions); i++) {
version_req->protocol_version = pci_protocol_versions[i];
ret = vmbus_sendpacket(hdev->channel, version_req, ret = vmbus_sendpacket(hdev->channel, version_req,
sizeof(struct pci_version_request), sizeof(struct pci_version_request),
(unsigned long)pkt, VM_PKT_DATA_INBAND, (unsigned long)pkt, VM_PKT_DATA_INBAND,
VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED); VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED);
if (ret) if (ret) {
dev_err(&hdev->device,
"PCI Pass-through VSP failed sending version reqquest: %#x",
ret);
goto exit; goto exit;
}
wait_for_completion(&comp_pkt.host_event); wait_for_completion(&comp_pkt.host_event);
if (comp_pkt.completion_status < 0) { if (comp_pkt.completion_status >= 0) {
pci_protocol_version = pci_protocol_versions[i];
dev_info(&hdev->device,
"PCI VMBus probing: Using version %#x\n",
pci_protocol_version);
goto exit;
}
if (comp_pkt.completion_status != STATUS_REVISION_MISMATCH) {
dev_err(&hdev->device, dev_err(&hdev->device,
"PCI Pass-through VSP failed version request %x\n", "PCI Pass-through VSP failed version request: %#x",
comp_pkt.completion_status); comp_pkt.completion_status);
ret = -EPROTO; ret = -EPROTO;
goto exit; goto exit;
} }
ret = 0; reinit_completion(&comp_pkt.host_event);
}
dev_err(&hdev->device,
"PCI pass-through VSP failed to find supported version");
ret = -EPROTO;
exit: exit:
kfree(pkt); kfree(pkt);
...@@ -2094,13 +2356,18 @@ static int hv_send_resources_allocated(struct hv_device *hdev) ...@@ -2094,13 +2356,18 @@ static int hv_send_resources_allocated(struct hv_device *hdev)
{ {
struct hv_pcibus_device *hbus = hv_get_drvdata(hdev); struct hv_pcibus_device *hbus = hv_get_drvdata(hdev);
struct pci_resources_assigned *res_assigned; struct pci_resources_assigned *res_assigned;
struct pci_resources_assigned2 *res_assigned2;
struct hv_pci_compl comp_pkt; struct hv_pci_compl comp_pkt;
struct hv_pci_dev *hpdev; struct hv_pci_dev *hpdev;
struct pci_packet *pkt; struct pci_packet *pkt;
size_t size_res;
u32 wslot; u32 wslot;
int ret; int ret;
pkt = kmalloc(sizeof(*pkt) + sizeof(*res_assigned), GFP_KERNEL); size_res = (pci_protocol_version < PCI_PROTOCOL_VERSION_1_2)
? sizeof(*res_assigned) : sizeof(*res_assigned2);
pkt = kmalloc(sizeof(*pkt) + size_res, GFP_KERNEL);
if (!pkt) if (!pkt)
return -ENOMEM; return -ENOMEM;
...@@ -2111,20 +2378,28 @@ static int hv_send_resources_allocated(struct hv_device *hdev) ...@@ -2111,20 +2378,28 @@ static int hv_send_resources_allocated(struct hv_device *hdev)
if (!hpdev) if (!hpdev)
continue; continue;
memset(pkt, 0, sizeof(*pkt) + sizeof(*res_assigned)); memset(pkt, 0, sizeof(*pkt) + size_res);
init_completion(&comp_pkt.host_event); init_completion(&comp_pkt.host_event);
pkt->completion_func = hv_pci_generic_compl; pkt->completion_func = hv_pci_generic_compl;
pkt->compl_ctxt = &comp_pkt; pkt->compl_ctxt = &comp_pkt;
res_assigned = (struct pci_resources_assigned *)&pkt->message;
res_assigned->message_type.type = PCI_RESOURCES_ASSIGNED;
res_assigned->wslot.slot = hpdev->desc.win_slot.slot;
if (pci_protocol_version < PCI_PROTOCOL_VERSION_1_2) {
res_assigned =
(struct pci_resources_assigned *)&pkt->message;
res_assigned->message_type.type =
PCI_RESOURCES_ASSIGNED;
res_assigned->wslot.slot = hpdev->desc.win_slot.slot;
} else {
res_assigned2 =
(struct pci_resources_assigned2 *)&pkt->message;
res_assigned2->message_type.type =
PCI_RESOURCES_ASSIGNED2;
res_assigned2->wslot.slot = hpdev->desc.win_slot.slot;
}
put_pcichild(hpdev, hv_pcidev_ref_by_slot); put_pcichild(hpdev, hv_pcidev_ref_by_slot);
ret = vmbus_sendpacket( ret = vmbus_sendpacket(hdev->channel, &pkt->message,
hdev->channel, &pkt->message, size_res, (unsigned long)pkt,
sizeof(*res_assigned),
(unsigned long)pkt,
VM_PKT_DATA_INBAND, VM_PKT_DATA_INBAND,
VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED); VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED);
if (ret) if (ret)
...@@ -2204,11 +2479,19 @@ static int hv_pci_probe(struct hv_device *hdev, ...@@ -2204,11 +2479,19 @@ static int hv_pci_probe(struct hv_device *hdev,
struct hv_pcibus_device *hbus; struct hv_pcibus_device *hbus;
int ret; int ret;
hbus = kzalloc(sizeof(*hbus), GFP_KERNEL); /*
* hv_pcibus_device contains the hypercall arguments for retargeting in
* hv_irq_unmask(). Those must not cross a page boundary.
*/
BUILD_BUG_ON(sizeof(*hbus) > PAGE_SIZE);
hbus = (struct hv_pcibus_device *)get_zeroed_page(GFP_KERNEL);
if (!hbus) if (!hbus)
return -ENOMEM; return -ENOMEM;
hbus->state = hv_pcibus_init; hbus->state = hv_pcibus_init;
hv_tmpcpumap_init();
/* /*
* The PCI bus "domain" is what is called "segment" in ACPI and * The PCI bus "domain" is what is called "segment" in ACPI and
* other specs. Pull it from the instance ID, to get something * other specs. Pull it from the instance ID, to get something
...@@ -2308,7 +2591,7 @@ static int hv_pci_probe(struct hv_device *hdev, ...@@ -2308,7 +2591,7 @@ static int hv_pci_probe(struct hv_device *hdev,
close: close:
vmbus_close(hdev->channel); vmbus_close(hdev->channel);
free_bus: free_bus:
kfree(hbus); free_page((unsigned long)hbus);
return ret; return ret;
} }
...@@ -2386,7 +2669,7 @@ static int hv_pci_remove(struct hv_device *hdev) ...@@ -2386,7 +2669,7 @@ static int hv_pci_remove(struct hv_device *hdev)
irq_domain_free_fwnode(hbus->sysdata.fwnode); irq_domain_free_fwnode(hbus->sysdata.fwnode);
put_hvpcibus(hbus); put_hvpcibus(hbus);
wait_for_completion(&hbus->remove_event); wait_for_completion(&hbus->remove_event);
kfree(hbus); free_page((unsigned long)hbus);
return 0; return 0;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment