Commit bc3c5e08 authored by Matt Roper's avatar Matt Roper

drm/i915/sseu: Don't try to store EU mask internally in UAPI format

Storing the EU mask internally in the same format the I915_QUERY
topology queries use makes the final copy_to_user() a bit simpler, but
makes the rest of the driver's SSEU more complicated and harder to
follow.  Let's switch to an internal representation that's more natural:
Xe_HP platforms will be a simple array of u16 masks, whereas pre-Xe_HP
platforms will be a two-dimensional array, indexed by [slice][subslice].
We'll convert to the uapi format only when the query uapi is called.

v2:
 - Drop has_common_ss_eumask.  We waste some space repeating identical
   EU masks for every single DSS, but the code is simpler without it.
   (Tvrtko)

v3:
 - Mask down EUs passed to sseu_set_eus at the callsite rather than
   inside the function.  (Tvrtko)
 - Eliminate sseu->eu_stride and calculate it when needed.  (Tvrtko)

Cc: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com>
Signed-off-by: default avatarMatt Roper <matthew.d.roper@intel.com>
Acked-by: default avatarTvrtko Ursulin <tvrtko.ursulin@intel.com>
Reviewed-by: default avatarBalasubramani Vivekanandan <balasubramani.vivekanandan@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20220601150725.521468-5-matthew.d.roper@intel.com
parent 4cfd1665
...@@ -19,8 +19,6 @@ void intel_sseu_set_info(struct sseu_dev_info *sseu, u8 max_slices, ...@@ -19,8 +19,6 @@ void intel_sseu_set_info(struct sseu_dev_info *sseu, u8 max_slices,
sseu->ss_stride = GEN_SSEU_STRIDE(sseu->max_subslices); sseu->ss_stride = GEN_SSEU_STRIDE(sseu->max_subslices);
GEM_BUG_ON(sseu->ss_stride > GEN_MAX_SUBSLICE_STRIDE); GEM_BUG_ON(sseu->ss_stride > GEN_MAX_SUBSLICE_STRIDE);
sseu->eu_stride = GEN_SSEU_STRIDE(sseu->max_eus_per_subslice);
GEM_BUG_ON(sseu->eu_stride > GEN_MAX_EU_STRIDE);
} }
unsigned int unsigned int
...@@ -78,47 +76,77 @@ intel_sseu_subslices_per_slice(const struct sseu_dev_info *sseu, u8 slice) ...@@ -78,47 +76,77 @@ intel_sseu_subslices_per_slice(const struct sseu_dev_info *sseu, u8 slice)
return hweight32(intel_sseu_get_subslices(sseu, slice)); return hweight32(intel_sseu_get_subslices(sseu, slice));
} }
static int sseu_eu_idx(const struct sseu_dev_info *sseu, int slice,
int subslice)
{
int slice_stride = sseu->max_subslices * sseu->eu_stride;
return slice * slice_stride + subslice * sseu->eu_stride;
}
static u16 sseu_get_eus(const struct sseu_dev_info *sseu, int slice, static u16 sseu_get_eus(const struct sseu_dev_info *sseu, int slice,
int subslice) int subslice)
{ {
int i, offset = sseu_eu_idx(sseu, slice, subslice); if (sseu->has_xehp_dss) {
u16 eu_mask = 0; WARN_ON(slice > 0);
return sseu->eu_mask.xehp[subslice];
for (i = 0; i < sseu->eu_stride; i++) } else {
eu_mask |= return sseu->eu_mask.hsw[slice][subslice];
((u16)sseu->eu_mask[offset + i]) << (i * BITS_PER_BYTE); }
return eu_mask;
} }
static void sseu_set_eus(struct sseu_dev_info *sseu, int slice, int subslice, static void sseu_set_eus(struct sseu_dev_info *sseu, int slice, int subslice,
u16 eu_mask) u16 eu_mask)
{ {
int i, offset = sseu_eu_idx(sseu, slice, subslice); GEM_WARN_ON(eu_mask && __fls(eu_mask) >= sseu->max_eus_per_subslice);
if (sseu->has_xehp_dss) {
for (i = 0; i < sseu->eu_stride; i++) GEM_WARN_ON(slice > 0);
sseu->eu_mask[offset + i] = sseu->eu_mask.xehp[subslice] = eu_mask;
(eu_mask >> (BITS_PER_BYTE * i)) & 0xff; } else {
sseu->eu_mask.hsw[slice][subslice] = eu_mask;
}
} }
static u16 compute_eu_total(const struct sseu_dev_info *sseu) static u16 compute_eu_total(const struct sseu_dev_info *sseu)
{ {
u16 i, total = 0; int s, ss, total = 0;
for (i = 0; i < ARRAY_SIZE(sseu->eu_mask); i++) for (s = 0; s < sseu->max_slices; s++)
total += hweight8(sseu->eu_mask[i]); for (ss = 0; ss < sseu->max_subslices; ss++)
if (sseu->has_xehp_dss)
total += hweight16(sseu->eu_mask.xehp[ss]);
else
total += hweight16(sseu->eu_mask.hsw[s][ss]);
return total; return total;
} }
/**
* intel_sseu_copy_eumask_to_user - Copy EU mask into a userspace buffer
* @to: Pointer to userspace buffer to copy to
* @sseu: SSEU structure containing EU mask to copy
*
* Copies the EU mask to a userspace buffer in the format expected by
* the query ioctl's topology queries.
*
* Returns the result of the copy_to_user() operation.
*/
int intel_sseu_copy_eumask_to_user(void __user *to,
const struct sseu_dev_info *sseu)
{
u8 eu_mask[GEN_SS_MASK_SIZE * GEN_MAX_EU_STRIDE] = {};
int eu_stride = GEN_SSEU_STRIDE(sseu->max_eus_per_subslice);
int len = sseu->max_slices * sseu->max_subslices * eu_stride;
int s, ss, i;
for (s = 0; s < sseu->max_slices; s++) {
for (ss = 0; ss < sseu->max_subslices; ss++) {
int uapi_offset =
s * sseu->max_subslices * eu_stride +
ss * eu_stride;
u16 mask = sseu_get_eus(sseu, s, ss);
for (i = 0; i < eu_stride; i++)
eu_mask[uapi_offset + i] =
(mask >> (BITS_PER_BYTE * i)) & 0xff;
}
}
return copy_to_user(to, eu_mask, len);
}
static void gen11_compute_sseu_info(struct sseu_dev_info *sseu, static void gen11_compute_sseu_info(struct sseu_dev_info *sseu,
u32 g_ss_en, u32 c_ss_en, u16 eu_en) u32 g_ss_en, u32 c_ss_en, u16 eu_en)
{ {
...@@ -278,7 +306,7 @@ static void cherryview_sseu_info_init(struct intel_gt *gt) ...@@ -278,7 +306,7 @@ static void cherryview_sseu_info_init(struct intel_gt *gt)
CHV_FGT_EU_DIS_SS0_R1_SHIFT) << 4); CHV_FGT_EU_DIS_SS0_R1_SHIFT) << 4);
subslice_mask |= BIT(0); subslice_mask |= BIT(0);
sseu_set_eus(sseu, 0, 0, ~disabled_mask); sseu_set_eus(sseu, 0, 0, ~disabled_mask & 0xFF);
} }
if (!(fuse & CHV_FGT_DISABLE_SS1)) { if (!(fuse & CHV_FGT_DISABLE_SS1)) {
...@@ -289,7 +317,7 @@ static void cherryview_sseu_info_init(struct intel_gt *gt) ...@@ -289,7 +317,7 @@ static void cherryview_sseu_info_init(struct intel_gt *gt)
CHV_FGT_EU_DIS_SS1_R1_SHIFT) << 4); CHV_FGT_EU_DIS_SS1_R1_SHIFT) << 4);
subslice_mask |= BIT(1); subslice_mask |= BIT(1);
sseu_set_eus(sseu, 0, 1, ~disabled_mask); sseu_set_eus(sseu, 0, 1, ~disabled_mask & 0xFF);
} }
intel_sseu_set_subslices(sseu, 0, sseu->subslice_mask, subslice_mask); intel_sseu_set_subslices(sseu, 0, sseu->subslice_mask, subslice_mask);
...@@ -362,7 +390,7 @@ static void gen9_sseu_info_init(struct intel_gt *gt) ...@@ -362,7 +390,7 @@ static void gen9_sseu_info_init(struct intel_gt *gt)
eu_disabled_mask = (eu_disable >> (ss * 8)) & eu_mask; eu_disabled_mask = (eu_disable >> (ss * 8)) & eu_mask;
sseu_set_eus(sseu, s, ss, ~eu_disabled_mask); sseu_set_eus(sseu, s, ss, ~eu_disabled_mask & eu_mask);
eu_per_ss = sseu->max_eus_per_subslice - eu_per_ss = sseu->max_eus_per_subslice -
hweight8(eu_disabled_mask); hweight8(eu_disabled_mask);
...@@ -475,7 +503,7 @@ static void bdw_sseu_info_init(struct intel_gt *gt) ...@@ -475,7 +503,7 @@ static void bdw_sseu_info_init(struct intel_gt *gt)
eu_disabled_mask = eu_disabled_mask =
eu_disable[s] >> (ss * sseu->max_eus_per_subslice); eu_disable[s] >> (ss * sseu->max_eus_per_subslice);
sseu_set_eus(sseu, s, ss, ~eu_disabled_mask); sseu_set_eus(sseu, s, ss, ~eu_disabled_mask & 0xFF);
n_disabled = hweight8(eu_disabled_mask); n_disabled = hweight8(eu_disabled_mask);
......
...@@ -57,7 +57,11 @@ struct sseu_dev_info { ...@@ -57,7 +57,11 @@ struct sseu_dev_info {
u8 subslice_mask[GEN_SS_MASK_SIZE]; u8 subslice_mask[GEN_SS_MASK_SIZE];
u8 geometry_subslice_mask[GEN_SS_MASK_SIZE]; u8 geometry_subslice_mask[GEN_SS_MASK_SIZE];
u8 compute_subslice_mask[GEN_SS_MASK_SIZE]; u8 compute_subslice_mask[GEN_SS_MASK_SIZE];
u8 eu_mask[GEN_SS_MASK_SIZE * GEN_MAX_EU_STRIDE]; union {
u16 hsw[GEN_MAX_HSW_SLICES][GEN_MAX_SS_PER_HSW_SLICE];
u16 xehp[GEN_MAX_DSS];
} eu_mask;
u16 eu_total; u16 eu_total;
u8 eu_per_subslice; u8 eu_per_subslice;
u8 min_eu_in_pool; u8 min_eu_in_pool;
...@@ -78,7 +82,6 @@ struct sseu_dev_info { ...@@ -78,7 +82,6 @@ struct sseu_dev_info {
u8 max_eus_per_subslice; u8 max_eus_per_subslice;
u8 ss_stride; u8 ss_stride;
u8 eu_stride;
}; };
/* /*
...@@ -150,4 +153,7 @@ void intel_sseu_print_topology(struct drm_i915_private *i915, ...@@ -150,4 +153,7 @@ void intel_sseu_print_topology(struct drm_i915_private *i915,
u16 intel_slicemask_from_dssmask(u64 dss_mask, int dss_per_slice); u16 intel_slicemask_from_dssmask(u64 dss_mask, int dss_per_slice);
int intel_sseu_copy_eumask_to_user(void __user *to,
const struct sseu_dev_info *sseu);
#endif /* __INTEL_SSEU_H__ */ #endif /* __INTEL_SSEU_H__ */
...@@ -35,6 +35,7 @@ static int fill_topology_info(const struct sseu_dev_info *sseu, ...@@ -35,6 +35,7 @@ static int fill_topology_info(const struct sseu_dev_info *sseu,
{ {
struct drm_i915_query_topology_info topo; struct drm_i915_query_topology_info topo;
u32 slice_length, subslice_length, eu_length, total_length; u32 slice_length, subslice_length, eu_length, total_length;
int eu_stride = GEN_SSEU_STRIDE(sseu->max_eus_per_subslice);
int ret; int ret;
BUILD_BUG_ON(sizeof(u8) != sizeof(sseu->slice_mask)); BUILD_BUG_ON(sizeof(u8) != sizeof(sseu->slice_mask));
...@@ -44,7 +45,7 @@ static int fill_topology_info(const struct sseu_dev_info *sseu, ...@@ -44,7 +45,7 @@ static int fill_topology_info(const struct sseu_dev_info *sseu,
slice_length = sizeof(sseu->slice_mask); slice_length = sizeof(sseu->slice_mask);
subslice_length = sseu->max_slices * sseu->ss_stride; subslice_length = sseu->max_slices * sseu->ss_stride;
eu_length = sseu->max_slices * sseu->max_subslices * sseu->eu_stride; eu_length = sseu->max_slices * sseu->max_subslices * eu_stride;
total_length = sizeof(topo) + slice_length + subslice_length + total_length = sizeof(topo) + slice_length + subslice_length +
eu_length; eu_length;
...@@ -61,7 +62,7 @@ static int fill_topology_info(const struct sseu_dev_info *sseu, ...@@ -61,7 +62,7 @@ static int fill_topology_info(const struct sseu_dev_info *sseu,
topo.subslice_offset = slice_length; topo.subslice_offset = slice_length;
topo.subslice_stride = sseu->ss_stride; topo.subslice_stride = sseu->ss_stride;
topo.eu_offset = slice_length + subslice_length; topo.eu_offset = slice_length + subslice_length;
topo.eu_stride = sseu->eu_stride; topo.eu_stride = eu_stride;
if (copy_to_user(u64_to_user_ptr(query_item->data_ptr), if (copy_to_user(u64_to_user_ptr(query_item->data_ptr),
&topo, sizeof(topo))) &topo, sizeof(topo)))
...@@ -76,10 +77,10 @@ static int fill_topology_info(const struct sseu_dev_info *sseu, ...@@ -76,10 +77,10 @@ static int fill_topology_info(const struct sseu_dev_info *sseu,
subslice_mask, subslice_length)) subslice_mask, subslice_length))
return -EFAULT; return -EFAULT;
if (copy_to_user(u64_to_user_ptr(query_item->data_ptr + if (intel_sseu_copy_eumask_to_user(u64_to_user_ptr(query_item->data_ptr +
sizeof(topo) + sizeof(topo) +
slice_length + subslice_length), slice_length + subslice_length),
sseu->eu_mask, eu_length)) sseu))
return -EFAULT; return -EFAULT;
return total_length; return total_length;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment