Merge tag 'LA.UM.9.14.r1-26000-LAHAINA.QSSI15.0' of https://git.codelinaro.org/clo/la/kernel/msm-5.4 into android13-5.4-lahaina
LA.UM.9.14.r1-26000-LAHAINA.QSSI15.0 # By Prashanth K (6) and others # Via Gerrit - the friendly Code Review server (5) and others * tag 'clo/msm-5.4/LA.UM.9.14.r1-26000-LAHAINA.QSSI15.0': FROMGIT: media: venus: hfi: add a check to handle OOB in sfr region FROMGIT: media: venus: hfi: add check to handle incorrect queue size FROMGIT: media: venus: hfi_parser: refactor hfi packet parsing logic FROMGIT: media: venus: hfi_parser: add check to avoid out of bound access UPSTREAM: usb: dwc3: host: Set XHCI_SG_TRB_CACHE_SIZE_QUIRK UPSTREAM: usb: host: xhci-plat: Add support for XHCI_SG_TRB_CACHE_SIZE_QUIRK UPSTREAM: usb: xhci: Add error handling in xhci_map_urb_for_dma UPSTREAM: usb: xhci: Use temporary buffer to consolidate SG UPSTREAM: usb: xhci: Set quirk for XHCI_SG_TRB_CACHE_SIZE_QUIRK defconfig: Enable RTL8152 ETH-USB driver msm: mhi_dev: Breaking memory for event request in smaller chunks msm: eva: Validating the SFR buffer size before accessing msm: eva: Copy back the validated size to avoid security issue Change-Id: Ibd883e18a8a410fb23eb3cda97e88b77c34cdbd7
This commit is contained in:
1
arch/arm/configs/vendor/sdxlemur.config
vendored
1
arch/arm/configs/vendor/sdxlemur.config
vendored
@@ -250,6 +250,7 @@ CONFIG_USB_CONFIGFS_F_UAC2=y
|
||||
CONFIG_USB_CONFIGFS_F_CCID=y
|
||||
CONFIG_USB_G_QTI=y
|
||||
CONFIG_USB_BAM=y
|
||||
CONFIG_USB_RTL8152=y
|
||||
CONFIG_USB_NET_SMSC75XX=y
|
||||
CONFIG_USB_NET_SMSC95XX=y
|
||||
CONFIG_ARM_PSCI=y
|
||||
|
||||
@@ -19,6 +19,8 @@ static void init_codecs(struct venus_core *core)
|
||||
struct venus_caps *caps = core->caps, *cap;
|
||||
unsigned long bit;
|
||||
|
||||
core->codecs_count = 0;
|
||||
|
||||
if (hweight_long(core->dec_codecs) + hweight_long(core->enc_codecs) > MAX_CODEC_NUM)
|
||||
return;
|
||||
|
||||
@@ -62,7 +64,7 @@ fill_buf_mode(struct venus_caps *cap, const void *data, unsigned int num)
|
||||
cap->cap_bufs_mode_dynamic = true;
|
||||
}
|
||||
|
||||
static void
|
||||
static int
|
||||
parse_alloc_mode(struct venus_core *core, u32 codecs, u32 domain, void *data)
|
||||
{
|
||||
struct hfi_buffer_alloc_mode_supported *mode = data;
|
||||
@@ -70,7 +72,7 @@ parse_alloc_mode(struct venus_core *core, u32 codecs, u32 domain, void *data)
|
||||
u32 *type;
|
||||
|
||||
if (num_entries > MAX_ALLOC_MODE_ENTRIES)
|
||||
return;
|
||||
return -EINVAL;
|
||||
|
||||
type = mode->data;
|
||||
|
||||
@@ -82,6 +84,8 @@ parse_alloc_mode(struct venus_core *core, u32 codecs, u32 domain, void *data)
|
||||
|
||||
type++;
|
||||
}
|
||||
|
||||
return sizeof(*mode);
|
||||
}
|
||||
|
||||
static void fill_profile_level(struct venus_caps *cap, const void *data,
|
||||
@@ -96,7 +100,7 @@ static void fill_profile_level(struct venus_caps *cap, const void *data,
|
||||
cap->num_pl += num;
|
||||
}
|
||||
|
||||
static void
|
||||
static int
|
||||
parse_profile_level(struct venus_core *core, u32 codecs, u32 domain, void *data)
|
||||
{
|
||||
struct hfi_profile_level_supported *pl = data;
|
||||
@@ -104,12 +108,14 @@ parse_profile_level(struct venus_core *core, u32 codecs, u32 domain, void *data)
|
||||
struct hfi_profile_level pl_arr[HFI_MAX_PROFILE_COUNT] = {};
|
||||
|
||||
if (pl->profile_count > HFI_MAX_PROFILE_COUNT)
|
||||
return;
|
||||
return -EINVAL;
|
||||
|
||||
memcpy(pl_arr, proflevel, pl->profile_count * sizeof(*proflevel));
|
||||
|
||||
for_each_codec(core->caps, ARRAY_SIZE(core->caps), codecs, domain,
|
||||
fill_profile_level, pl_arr, pl->profile_count);
|
||||
|
||||
return pl->profile_count * sizeof(*proflevel) + sizeof(u32);
|
||||
}
|
||||
|
||||
static void
|
||||
@@ -124,7 +130,7 @@ fill_caps(struct venus_caps *cap, const void *data, unsigned int num)
|
||||
cap->num_caps += num;
|
||||
}
|
||||
|
||||
static void
|
||||
static int
|
||||
parse_caps(struct venus_core *core, u32 codecs, u32 domain, void *data)
|
||||
{
|
||||
struct hfi_capabilities *caps = data;
|
||||
@@ -133,12 +139,14 @@ parse_caps(struct venus_core *core, u32 codecs, u32 domain, void *data)
|
||||
struct hfi_capability caps_arr[MAX_CAP_ENTRIES] = {};
|
||||
|
||||
if (num_caps > MAX_CAP_ENTRIES)
|
||||
return;
|
||||
return -EINVAL;
|
||||
|
||||
memcpy(caps_arr, cap, num_caps * sizeof(*cap));
|
||||
|
||||
for_each_codec(core->caps, ARRAY_SIZE(core->caps), codecs, domain,
|
||||
fill_caps, caps_arr, num_caps);
|
||||
|
||||
return sizeof(*caps);
|
||||
}
|
||||
|
||||
static void fill_raw_fmts(struct venus_caps *cap, const void *fmts,
|
||||
@@ -153,7 +161,7 @@ static void fill_raw_fmts(struct venus_caps *cap, const void *fmts,
|
||||
cap->num_fmts += num_fmts;
|
||||
}
|
||||
|
||||
static void
|
||||
static int
|
||||
parse_raw_formats(struct venus_core *core, u32 codecs, u32 domain, void *data)
|
||||
{
|
||||
struct hfi_uncompressed_format_supported *fmt = data;
|
||||
@@ -162,7 +170,8 @@ parse_raw_formats(struct venus_core *core, u32 codecs, u32 domain, void *data)
|
||||
struct raw_formats rawfmts[MAX_FMT_ENTRIES] = {};
|
||||
u32 entries = fmt->format_entries;
|
||||
unsigned int i = 0;
|
||||
u32 num_planes;
|
||||
u32 num_planes = 0;
|
||||
u32 size;
|
||||
|
||||
while (entries) {
|
||||
num_planes = pinfo->num_planes;
|
||||
@@ -172,7 +181,7 @@ parse_raw_formats(struct venus_core *core, u32 codecs, u32 domain, void *data)
|
||||
i++;
|
||||
|
||||
if (i >= MAX_FMT_ENTRIES)
|
||||
return;
|
||||
return -EINVAL;
|
||||
|
||||
if (pinfo->num_planes > MAX_PLANES)
|
||||
break;
|
||||
@@ -184,9 +193,13 @@ parse_raw_formats(struct venus_core *core, u32 codecs, u32 domain, void *data)
|
||||
|
||||
for_each_codec(core->caps, ARRAY_SIZE(core->caps), codecs, domain,
|
||||
fill_raw_fmts, rawfmts, i);
|
||||
size = fmt->format_entries * (sizeof(*constr) * num_planes + 2 * sizeof(u32))
|
||||
+ 2 * sizeof(u32);
|
||||
|
||||
return size;
|
||||
}
|
||||
|
||||
static void parse_codecs(struct venus_core *core, void *data)
|
||||
static int parse_codecs(struct venus_core *core, void *data)
|
||||
{
|
||||
struct hfi_codec_supported *codecs = data;
|
||||
|
||||
@@ -198,21 +211,27 @@ static void parse_codecs(struct venus_core *core, void *data)
|
||||
core->dec_codecs &= ~HFI_VIDEO_CODEC_SPARK;
|
||||
core->enc_codecs &= ~HFI_VIDEO_CODEC_HEVC;
|
||||
}
|
||||
|
||||
return sizeof(*codecs);
|
||||
}
|
||||
|
||||
static void parse_max_sessions(struct venus_core *core, const void *data)
|
||||
static int parse_max_sessions(struct venus_core *core, const void *data)
|
||||
{
|
||||
const struct hfi_max_sessions_supported *sessions = data;
|
||||
|
||||
core->max_sessions_supported = sessions->max_sessions;
|
||||
|
||||
return sizeof(*sessions);
|
||||
}
|
||||
|
||||
static void parse_codecs_mask(u32 *codecs, u32 *domain, void *data)
|
||||
static int parse_codecs_mask(u32 *codecs, u32 *domain, void *data)
|
||||
{
|
||||
struct hfi_codec_mask_supported *mask = data;
|
||||
|
||||
*codecs = mask->codecs;
|
||||
*domain = mask->video_domains;
|
||||
|
||||
return sizeof(*mask);
|
||||
}
|
||||
|
||||
static void parser_init(struct venus_inst *inst, u32 *codecs, u32 *domain)
|
||||
@@ -246,46 +265,76 @@ static void parser_fini(struct venus_inst *inst, u32 codecs, u32 domain)
|
||||
u32 hfi_parser(struct venus_core *core, struct venus_inst *inst, void *buf,
|
||||
u32 size)
|
||||
{
|
||||
unsigned int words_count = size >> 2;
|
||||
u32 *word = buf, *data, codecs = 0, domain = 0;
|
||||
u32 *words = buf, *payload, codecs = 0, domain = 0;
|
||||
u32 *frame_size = buf + size;
|
||||
u32 rem_bytes = size;
|
||||
int ret;
|
||||
|
||||
if (size % 4)
|
||||
return HFI_ERR_SYS_INSUFFICIENT_RESOURCES;
|
||||
|
||||
parser_init(inst, &codecs, &domain);
|
||||
|
||||
while (words_count) {
|
||||
data = word + 1;
|
||||
while (words < frame_size) {
|
||||
payload = words + 1;
|
||||
|
||||
switch (*word) {
|
||||
switch (*words) {
|
||||
case HFI_PROPERTY_PARAM_CODEC_SUPPORTED:
|
||||
parse_codecs(core, data);
|
||||
if (rem_bytes <= sizeof(struct hfi_codec_supported))
|
||||
return HFI_ERR_SYS_INSUFFICIENT_RESOURCES;
|
||||
|
||||
ret = parse_codecs(core, payload);
|
||||
if (ret < 0)
|
||||
return HFI_ERR_SYS_INSUFFICIENT_RESOURCES;
|
||||
|
||||
init_codecs(core);
|
||||
break;
|
||||
case HFI_PROPERTY_PARAM_MAX_SESSIONS_SUPPORTED:
|
||||
parse_max_sessions(core, data);
|
||||
if (rem_bytes <= sizeof(struct hfi_max_sessions_supported))
|
||||
return HFI_ERR_SYS_INSUFFICIENT_RESOURCES;
|
||||
|
||||
ret = parse_max_sessions(core, payload);
|
||||
break;
|
||||
case HFI_PROPERTY_PARAM_CODEC_MASK_SUPPORTED:
|
||||
parse_codecs_mask(&codecs, &domain, data);
|
||||
if (rem_bytes <= sizeof(struct hfi_codec_mask_supported))
|
||||
return HFI_ERR_SYS_INSUFFICIENT_RESOURCES;
|
||||
|
||||
ret = parse_codecs_mask(&codecs, &domain, payload);
|
||||
break;
|
||||
case HFI_PROPERTY_PARAM_UNCOMPRESSED_FORMAT_SUPPORTED:
|
||||
parse_raw_formats(core, codecs, domain, data);
|
||||
if (rem_bytes <= sizeof(struct hfi_uncompressed_format_supported))
|
||||
return HFI_ERR_SYS_INSUFFICIENT_RESOURCES;
|
||||
|
||||
ret = parse_raw_formats(core, codecs, domain, payload);
|
||||
break;
|
||||
case HFI_PROPERTY_PARAM_CAPABILITY_SUPPORTED:
|
||||
parse_caps(core, codecs, domain, data);
|
||||
if (rem_bytes <= sizeof(struct hfi_capabilities))
|
||||
return HFI_ERR_SYS_INSUFFICIENT_RESOURCES;
|
||||
|
||||
ret = parse_caps(core, codecs, domain, payload);
|
||||
break;
|
||||
case HFI_PROPERTY_PARAM_PROFILE_LEVEL_SUPPORTED:
|
||||
parse_profile_level(core, codecs, domain, data);
|
||||
if (rem_bytes <= sizeof(struct hfi_profile_level_supported))
|
||||
return HFI_ERR_SYS_INSUFFICIENT_RESOURCES;
|
||||
|
||||
ret = parse_profile_level(core, codecs, domain, payload);
|
||||
break;
|
||||
case HFI_PROPERTY_PARAM_BUFFER_ALLOC_MODE_SUPPORTED:
|
||||
parse_alloc_mode(core, codecs, domain, data);
|
||||
if (rem_bytes <= sizeof(struct hfi_buffer_alloc_mode_supported))
|
||||
return HFI_ERR_SYS_INSUFFICIENT_RESOURCES;
|
||||
|
||||
ret = parse_alloc_mode(core, codecs, domain, payload);
|
||||
break;
|
||||
default:
|
||||
ret = sizeof(u32);
|
||||
break;
|
||||
}
|
||||
|
||||
word++;
|
||||
words_count--;
|
||||
if (ret < 0)
|
||||
return HFI_ERR_SYS_INSUFFICIENT_RESOURCES;
|
||||
|
||||
words += ret / sizeof(u32);
|
||||
rem_bytes -= ret;
|
||||
}
|
||||
|
||||
parser_fini(inst, codecs, domain);
|
||||
|
||||
@@ -188,6 +188,9 @@ static int venus_write_queue(struct venus_hfi_device *hdev,
|
||||
/* ensure rd/wr indices's are read from memory */
|
||||
rmb();
|
||||
|
||||
if (qsize > IFACEQ_QUEUE_SIZE / 4)
|
||||
return -EINVAL;
|
||||
|
||||
if (wr_idx >= rd_idx)
|
||||
empty_space = qsize - (wr_idx - rd_idx);
|
||||
else
|
||||
@@ -256,6 +259,9 @@ static int venus_read_queue(struct venus_hfi_device *hdev,
|
||||
wr_idx = qhdr->write_idx;
|
||||
qsize = qhdr->q_size;
|
||||
|
||||
if (qsize > IFACEQ_QUEUE_SIZE / 4)
|
||||
return -EINVAL;
|
||||
|
||||
/* make sure data is valid before using it */
|
||||
rmb();
|
||||
|
||||
@@ -970,18 +976,26 @@ static void venus_sfr_print(struct venus_hfi_device *hdev)
|
||||
{
|
||||
struct device *dev = hdev->core->dev;
|
||||
struct hfi_sfr *sfr = hdev->sfr.kva;
|
||||
u32 size;
|
||||
void *p;
|
||||
|
||||
if (!sfr)
|
||||
return;
|
||||
|
||||
p = memchr(sfr->data, '\0', sfr->buf_size);
|
||||
size = sfr->buf_size;
|
||||
if (!size)
|
||||
return;
|
||||
|
||||
if (size > ALIGNED_SFR_SIZE)
|
||||
size = ALIGNED_SFR_SIZE;
|
||||
|
||||
p = memchr(sfr->data, '\0', size);
|
||||
/*
|
||||
* SFR isn't guaranteed to be NULL terminated since SYS_ERROR indicates
|
||||
* that Venus is in the process of crashing.
|
||||
*/
|
||||
if (!p)
|
||||
sfr->data[sfr->buf_size - 1] = '\0';
|
||||
sfr->data[size - 1] = '\0';
|
||||
|
||||
dev_err_ratelimited(dev, "SFR message from FW: %s\n", sfr->data);
|
||||
}
|
||||
|
||||
@@ -3189,6 +3189,7 @@ static int mhi_dev_alloc_evt_buf_evt_req(struct mhi_dev *mhi,
|
||||
{
|
||||
int rc;
|
||||
uint32_t size, i;
|
||||
struct event_req *req, *tmp;
|
||||
|
||||
size = mhi_dev_get_evt_ring_size(mhi, ch->ch_id);
|
||||
|
||||
@@ -3207,7 +3208,14 @@ static int mhi_dev_alloc_evt_buf_evt_req(struct mhi_dev *mhi,
|
||||
* they were allocated with a different size
|
||||
*/
|
||||
if (ch->evt_buf_size) {
|
||||
kfree(ch->ereqs);
|
||||
list_for_each_entry_safe(req, tmp, &ch->event_req_buffers, list) {
|
||||
list_del(&req->list);
|
||||
kfree(req);
|
||||
}
|
||||
list_for_each_entry_safe(req, tmp, &ch->flush_event_req_buffers, list) {
|
||||
list_del(&req->list);
|
||||
kfree(req);
|
||||
}
|
||||
kfree(ch->tr_events);
|
||||
}
|
||||
/*
|
||||
@@ -3221,14 +3229,8 @@ static int mhi_dev_alloc_evt_buf_evt_req(struct mhi_dev *mhi,
|
||||
mhi_log(MHI_MSG_INFO,
|
||||
"ch_id:%d evt buf size is %d\n", ch->ch_id, ch->evt_buf_size);
|
||||
|
||||
/* Allocate event requests */
|
||||
ch->ereqs = kcalloc(ch->evt_req_size, sizeof(*ch->ereqs), GFP_KERNEL);
|
||||
if (!ch->ereqs) {
|
||||
mhi_log(MHI_MSG_ERROR,
|
||||
"Failed to alloc ereqs for ch_id:%d\n", ch->ch_id);
|
||||
rc = -ENOMEM;
|
||||
goto free_ereqs;
|
||||
}
|
||||
INIT_LIST_HEAD(&ch->event_req_buffers);
|
||||
INIT_LIST_HEAD(&ch->flush_event_req_buffers);
|
||||
|
||||
/* Allocate buffers to queue transfer completion events */
|
||||
ch->tr_events = kcalloc(ch->evt_buf_size, sizeof(*ch->tr_events),
|
||||
@@ -3241,11 +3243,13 @@ static int mhi_dev_alloc_evt_buf_evt_req(struct mhi_dev *mhi,
|
||||
goto free_ereqs;
|
||||
}
|
||||
|
||||
/* Organize event flush requests into a linked list */
|
||||
INIT_LIST_HEAD(&ch->event_req_buffers);
|
||||
INIT_LIST_HEAD(&ch->flush_event_req_buffers);
|
||||
for (i = 0; i < ch->evt_req_size; ++i)
|
||||
list_add_tail(&ch->ereqs[i].list, &ch->event_req_buffers);
|
||||
/* Allocate event requests */
|
||||
for (i = 0; i < ch->evt_req_size; ++i) {
|
||||
req = kzalloc(sizeof(struct event_req), GFP_KERNEL);
|
||||
if (!req)
|
||||
goto free_ereqs;
|
||||
list_add_tail(&req->list, &ch->event_req_buffers);
|
||||
}
|
||||
|
||||
ch->curr_ereq =
|
||||
container_of(ch->event_req_buffers.next,
|
||||
@@ -3263,8 +3267,13 @@ static int mhi_dev_alloc_evt_buf_evt_req(struct mhi_dev *mhi,
|
||||
return 0;
|
||||
|
||||
free_ereqs:
|
||||
kfree(ch->ereqs);
|
||||
ch->ereqs = NULL;
|
||||
if (!list_empty(&ch->event_req_buffers)) {
|
||||
list_for_each_entry_safe(req, tmp, &ch->event_req_buffers, list) {
|
||||
list_del(&req->list);
|
||||
kfree(req);
|
||||
}
|
||||
}
|
||||
kfree(ch->tr_events);
|
||||
ch->evt_buf_size = 0;
|
||||
ch->evt_req_size = 0;
|
||||
|
||||
|
||||
@@ -43,7 +43,7 @@ out:
|
||||
|
||||
int dwc3_host_init(struct dwc3 *dwc)
|
||||
{
|
||||
struct property_entry props[6];
|
||||
struct property_entry props[7];
|
||||
struct platform_device *xhci;
|
||||
int ret, irq;
|
||||
struct resource *res;
|
||||
@@ -88,6 +88,8 @@ int dwc3_host_init(struct dwc3 *dwc)
|
||||
|
||||
memset(props, 0, sizeof(struct property_entry) * ARRAY_SIZE(props));
|
||||
|
||||
props[prop_idx++] = PROPERTY_ENTRY_BOOL("xhci-sg-trb-cache-size-quirk");
|
||||
|
||||
if (dwc->usb3_lpm_capable)
|
||||
props[prop_idx++] = PROPERTY_ENTRY_BOOL("usb3-lpm-capable");
|
||||
|
||||
|
||||
@@ -309,6 +309,9 @@ static int xhci_plat_probe(struct platform_device *pdev)
|
||||
if (device_property_read_bool(tmpdev, "quirk-broken-port-ped"))
|
||||
xhci->quirks |= XHCI_BROKEN_PORT_PED;
|
||||
|
||||
if (device_property_read_bool(tmpdev, "xhci-sg-trb-cache-size-quirk"))
|
||||
xhci->quirks |= XHCI_SG_TRB_CACHE_SIZE_QUIRK;
|
||||
|
||||
device_property_read_u32(tmpdev, "imod-interval-ns",
|
||||
&xhci->imod_interval);
|
||||
}
|
||||
@@ -336,6 +339,9 @@ static int xhci_plat_probe(struct platform_device *pdev)
|
||||
if ((xhci->quirks & XHCI_SKIP_PHY_INIT) || (priv && (priv->quirks & XHCI_SKIP_PHY_INIT)))
|
||||
hcd->skip_phy_initialization = 1;
|
||||
|
||||
if (priv && (priv->quirks & XHCI_SG_TRB_CACHE_SIZE_QUIRK))
|
||||
xhci->quirks |= XHCI_SG_TRB_CACHE_SIZE_QUIRK;
|
||||
|
||||
ret = usb_add_hcd(hcd, irq, IRQF_SHARED);
|
||||
if (ret)
|
||||
goto disable_usb_phy;
|
||||
|
||||
@@ -3380,7 +3380,7 @@ int xhci_queue_bulk_tx(struct xhci_hcd *xhci, gfp_t mem_flags,
|
||||
|
||||
full_len = urb->transfer_buffer_length;
|
||||
/* If we have scatter/gather list, we use it. */
|
||||
if (urb->num_sgs) {
|
||||
if (urb->num_sgs && !(urb->transfer_flags & URB_DMA_MAP_SINGLE)) {
|
||||
num_sgs = urb->num_mapped_sgs;
|
||||
sg = urb->sg;
|
||||
addr = (u64) sg_dma_address(sg);
|
||||
|
||||
@@ -1349,6 +1349,112 @@ EXPORT_SYMBOL_GPL(xhci_resume);
|
||||
|
||||
/*-------------------------------------------------------------------------*/
|
||||
|
||||
static int xhci_map_temp_buffer(struct usb_hcd *hcd, struct urb *urb)
|
||||
{
|
||||
void *temp;
|
||||
int ret = 0;
|
||||
unsigned int buf_len;
|
||||
enum dma_data_direction dir;
|
||||
|
||||
dir = usb_urb_dir_in(urb) ? DMA_FROM_DEVICE : DMA_TO_DEVICE;
|
||||
buf_len = urb->transfer_buffer_length;
|
||||
|
||||
temp = kzalloc_node(buf_len, GFP_ATOMIC,
|
||||
dev_to_node(hcd->self.sysdev));
|
||||
if (!temp)
|
||||
return -ENOMEM;
|
||||
|
||||
if (usb_urb_dir_out(urb))
|
||||
sg_pcopy_to_buffer(urb->sg, urb->num_sgs,
|
||||
temp, buf_len, 0);
|
||||
|
||||
urb->transfer_buffer = temp;
|
||||
urb->transfer_dma = dma_map_single(hcd->self.sysdev,
|
||||
urb->transfer_buffer,
|
||||
urb->transfer_buffer_length,
|
||||
dir);
|
||||
|
||||
if (dma_mapping_error(hcd->self.sysdev,
|
||||
urb->transfer_dma)) {
|
||||
ret = -EAGAIN;
|
||||
kfree(temp);
|
||||
} else {
|
||||
urb->transfer_flags |= URB_DMA_MAP_SINGLE;
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static bool xhci_urb_temp_buffer_required(struct usb_hcd *hcd,
|
||||
struct urb *urb)
|
||||
{
|
||||
bool ret = false;
|
||||
unsigned int i;
|
||||
unsigned int len = 0;
|
||||
unsigned int trb_size;
|
||||
unsigned int max_pkt;
|
||||
struct scatterlist *sg;
|
||||
struct scatterlist *tail_sg;
|
||||
|
||||
tail_sg = urb->sg;
|
||||
max_pkt = usb_endpoint_maxp(&urb->ep->desc);
|
||||
|
||||
if (!urb->num_sgs)
|
||||
return ret;
|
||||
|
||||
if (urb->dev->speed >= USB_SPEED_SUPER)
|
||||
trb_size = TRB_CACHE_SIZE_SS;
|
||||
else
|
||||
trb_size = TRB_CACHE_SIZE_HS;
|
||||
|
||||
if (urb->transfer_buffer_length != 0 &&
|
||||
!(urb->transfer_flags & URB_NO_TRANSFER_DMA_MAP)) {
|
||||
for_each_sg(urb->sg, sg, urb->num_sgs, i) {
|
||||
len = len + sg->length;
|
||||
if (i > trb_size - 2) {
|
||||
len = len - tail_sg->length;
|
||||
if (len < max_pkt) {
|
||||
ret = true;
|
||||
break;
|
||||
}
|
||||
|
||||
tail_sg = sg_next(tail_sg);
|
||||
}
|
||||
}
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void xhci_unmap_temp_buf(struct usb_hcd *hcd, struct urb *urb)
|
||||
{
|
||||
unsigned int len;
|
||||
unsigned int buf_len;
|
||||
enum dma_data_direction dir;
|
||||
|
||||
dir = usb_urb_dir_in(urb) ? DMA_FROM_DEVICE : DMA_TO_DEVICE;
|
||||
|
||||
buf_len = urb->transfer_buffer_length;
|
||||
|
||||
if (IS_ENABLED(CONFIG_HAS_DMA) &&
|
||||
(urb->transfer_flags & URB_DMA_MAP_SINGLE))
|
||||
dma_unmap_single(hcd->self.sysdev,
|
||||
urb->transfer_dma,
|
||||
urb->transfer_buffer_length,
|
||||
dir);
|
||||
|
||||
if (usb_urb_dir_in(urb))
|
||||
len = sg_pcopy_from_buffer(urb->sg, urb->num_sgs,
|
||||
urb->transfer_buffer,
|
||||
buf_len,
|
||||
0);
|
||||
|
||||
urb->transfer_flags &= ~URB_DMA_MAP_SINGLE;
|
||||
kfree(urb->transfer_buffer);
|
||||
urb->transfer_buffer = NULL;
|
||||
}
|
||||
|
||||
|
||||
|
||||
/*
|
||||
* Bypass the DMA mapping if URB is suitable for Immediate Transfer (IDT),
|
||||
* we'll copy the actual data into the TRB address register. This is limited to
|
||||
@@ -1358,12 +1464,37 @@ EXPORT_SYMBOL_GPL(xhci_resume);
|
||||
static int xhci_map_urb_for_dma(struct usb_hcd *hcd, struct urb *urb,
|
||||
gfp_t mem_flags)
|
||||
{
|
||||
struct xhci_hcd *xhci;
|
||||
|
||||
xhci = hcd_to_xhci(hcd);
|
||||
|
||||
if (xhci_urb_suitable_for_idt(urb))
|
||||
return 0;
|
||||
|
||||
if (xhci->quirks & XHCI_SG_TRB_CACHE_SIZE_QUIRK) {
|
||||
if (xhci_urb_temp_buffer_required(hcd, urb))
|
||||
return xhci_map_temp_buffer(hcd, urb);
|
||||
}
|
||||
return usb_hcd_map_urb_for_dma(hcd, urb, mem_flags);
|
||||
}
|
||||
|
||||
static void xhci_unmap_urb_for_dma(struct usb_hcd *hcd, struct urb *urb)
|
||||
{
|
||||
struct xhci_hcd *xhci;
|
||||
bool unmap_temp_buf = false;
|
||||
|
||||
xhci = hcd_to_xhci(hcd);
|
||||
|
||||
if (urb->num_sgs && (urb->transfer_flags & URB_DMA_MAP_SINGLE))
|
||||
unmap_temp_buf = true;
|
||||
|
||||
if ((xhci->quirks & XHCI_SG_TRB_CACHE_SIZE_QUIRK) && unmap_temp_buf)
|
||||
xhci_unmap_temp_buf(hcd, urb);
|
||||
else
|
||||
usb_hcd_unmap_urb_for_dma(hcd, urb);
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* xhci_get_endpoint_index - Used for passing endpoint bitmasks between the core and
|
||||
* HCDs. Find the index for an endpoint given its descriptor. Use the return
|
||||
@@ -5598,6 +5729,7 @@ static const struct hc_driver xhci_hc_driver = {
|
||||
* managing i/o requests and associated device resources
|
||||
*/
|
||||
.map_urb_for_dma = xhci_map_urb_for_dma,
|
||||
.unmap_urb_for_dma = xhci_unmap_urb_for_dma,
|
||||
.urb_enqueue = xhci_urb_enqueue,
|
||||
.urb_dequeue = xhci_urb_dequeue,
|
||||
.alloc_dev = xhci_alloc_dev,
|
||||
|
||||
@@ -1338,6 +1338,10 @@ enum xhci_setup_dev {
|
||||
#define TRB_SIA (1<<31)
|
||||
#define TRB_FRAME_ID(p) (((p) & 0x7ff) << 20)
|
||||
|
||||
/* TRB cache size for xHC with TRB cache */
|
||||
#define TRB_CACHE_SIZE_HS 8
|
||||
#define TRB_CACHE_SIZE_SS 16
|
||||
|
||||
struct xhci_generic_trb {
|
||||
__le32 field[4];
|
||||
};
|
||||
@@ -1891,6 +1895,7 @@ struct xhci_hcd {
|
||||
#define XHCI_SNPS_BROKEN_SUSPEND BIT_ULL(35)
|
||||
#define XHCI_SKIP_PHY_INIT BIT_ULL(37)
|
||||
#define XHCI_DISABLE_SPARSE BIT_ULL(38)
|
||||
#define XHCI_SG_TRB_CACHE_SIZE_QUIRK BIT_ULL(39)
|
||||
#define XHCI_NO_SOFT_RETRY BIT_ULL(40)
|
||||
|
||||
unsigned int num_active_eps;
|
||||
|
||||
Reference in New Issue
Block a user