Files
kernel_nothing_sm7325/drivers/usb/gadget/function/f_gsi.c
Prashanth K 63a32bf361 usb: gadget: f_gsi: bail out if opts is null
Currently, functions gsi_inst_clean & gsi_free_inst utilises
gsi_opts without any check, however there is a possibility
that the opts structure could become NULL. In such case, due
to lack of if checks can result in NULL pointer dereference.

Change-Id: I548690e2eee377b5292f258972ae7e38417f3085
Signed-off-by: Prashanth K <quic_prashk@quicinc.com>
2024-08-28 22:14:17 -07:00

4118 lines
113 KiB
C

// SPDX-License-Identifier: GPL-2.0-only
/*
* Copyright (c) 2015-2021, The Linux Foundation. All rights reserved.
*/
#include <linux/module.h>
#include "f_gsi.h"
#include "rndis.h"
static struct usb_gsi_debugfs {
struct dentry *debugfs_root;
bool qti_packet_debug;
} debugfs;
static struct workqueue_struct *ipa_usb_wq;
static struct gsi_inst_status {
struct mutex gsi_lock;
bool inst_exist;
struct gsi_opts *opts;
} inst_status[IPA_USB_MAX_TETH_PROT_SIZE];
#define MAX_CDEV_INSTANCES 4
static int major;
static struct class *gsi_class;
static DEFINE_IDA(gsi_ida);
/* Deregister misc device and free instance structures */
static void gsi_inst_clean(struct gsi_opts *opts);
static void gsi_rndis_ipa_reset_trigger(struct gsi_data_port *d_port);
static int gsi_ctrl_send_notification(struct f_gsi *gsi);
static struct gsi_ctrl_pkt *gsi_ctrl_pkt_alloc(unsigned int len, gfp_t flags);
static void gsi_ctrl_pkt_free(struct gsi_ctrl_pkt *pkt);
static inline bool usb_gsi_remote_wakeup_allowed(struct usb_function *f)
{
bool remote_wakeup_allowed = false;
struct f_gsi *gsi = func_to_gsi(f);
if (f->config->cdev->gadget->speed >= USB_SPEED_SUPER)
remote_wakeup_allowed =
gsi->func_is_suspended ? gsi->func_wakeup_allowed : false;
else
remote_wakeup_allowed =
usb_get_remote_wakeup_status(f->config->cdev->gadget);
log_event_dbg("%s: remote_wakeup_allowed:%s", __func__,
(remote_wakeup_allowed ? "true" : "false"));
return remote_wakeup_allowed;
}
static void post_event(struct gsi_data_port *port, u8 event)
{
unsigned long flags;
struct f_gsi *gsi = d_port_to_gsi(port);
spin_lock_irqsave(&port->evt_q.q_lock, flags);
port->evt_q.tail++;
/* Check for wraparound and make room */
port->evt_q.tail = port->evt_q.tail % MAXQUEUELEN;
/* Check for overflow */
if (port->evt_q.tail == port->evt_q.head) {
log_event_err("%s: event queue overflow error", __func__);
spin_unlock_irqrestore(&port->evt_q.q_lock, flags);
return;
}
/* Add event to queue */
port->evt_q.event[port->evt_q.tail] = event;
spin_unlock_irqrestore(&port->evt_q.q_lock, flags);
}
static void __maybe_unused post_event_to_evt_queue(struct gsi_data_port *port,
u8 event)
{
post_event(port, event);
queue_delayed_work(port->ipa_usb_wq, &port->usb_ipa_w, 0);
}
static u8 read_event(struct gsi_data_port *port)
{
u8 event;
unsigned long flags;
struct f_gsi *gsi = d_port_to_gsi(port);
spin_lock_irqsave(&port->evt_q.q_lock, flags);
if (port->evt_q.head == port->evt_q.tail) {
log_event_dbg("%s: event queue empty", __func__);
spin_unlock_irqrestore(&port->evt_q.q_lock, flags);
return EVT_NONE;
}
port->evt_q.head++;
/* Check for wraparound and make room */
port->evt_q.head = port->evt_q.head % MAXQUEUELEN;
event = port->evt_q.event[port->evt_q.head];
spin_unlock_irqrestore(&port->evt_q.q_lock, flags);
return event;
}
static u8 peek_event(struct gsi_data_port *port)
{
u8 event;
unsigned long flags;
u8 peek_index = 0;
struct f_gsi *gsi = d_port_to_gsi(port);
spin_lock_irqsave(&port->evt_q.q_lock, flags);
if (port->evt_q.head == port->evt_q.tail) {
log_event_dbg("%s: event queue empty", __func__);
spin_unlock_irqrestore(&port->evt_q.q_lock, flags);
return EVT_NONE;
}
peek_index = (port->evt_q.head + 1) % MAXQUEUELEN;
event = port->evt_q.event[peek_index];
spin_unlock_irqrestore(&port->evt_q.q_lock, flags);
return event;
}
static void __maybe_unused reset_event_queue(struct gsi_data_port *port)
{
unsigned long flags;
spin_lock_irqsave(&port->evt_q.q_lock, flags);
port->evt_q.head = port->evt_q.tail = MAXQUEUELEN - 1;
memset(&port->evt_q.event[0], EVT_NONE, MAXQUEUELEN);
spin_unlock_irqrestore(&port->evt_q.q_lock, flags);
}
static int gsi_wakeup_host(struct f_gsi *gsi)
{
int ret;
struct usb_gadget *gadget;
struct usb_function *func;
func = &gsi->function;
gadget = gsi->function.config->cdev->gadget;
log_event_dbg("Entering %s", __func__);
if (!gadget) {
log_event_err("FAILED: d_port->cdev->gadget == NULL");
return -ENODEV;
}
gsi->rwake_inprogress = true;
/*
* In Super-Speed mode, remote wakeup is not allowed for suspended
* functions which have been disallowed by the host to issue Function
* Remote Wakeup.
* Note - We deviate here from the USB 3.0 spec and allow
* non-suspended functions to issue remote-wakeup even if they were not
* allowed to do so by the host. This is done in order to support non
* fully USB 3.0 compatible hosts.
*/
if ((gadget->speed >= USB_SPEED_SUPER) && (gsi->func_is_suspended)) {
log_event_dbg("%s: Calling usb_func_wakeup", __func__);
ret = usb_func_wakeup(func);
if (ret == -EAGAIN)
gsi->func_wakeup_pending = true;
} else {
log_event_dbg("%s: Calling usb_gadget_wakeup", __func__);
ret = usb_gadget_wakeup(gadget);
}
if ((ret == -EBUSY) || (ret == -EAGAIN))
log_event_dbg("RW delayed due to LPM exit.");
else if (ret)
log_event_err("wakeup failed. ret=%d.", ret);
if (ret)
gsi->rwake_inprogress = false;
return ret;
}
static void gsi_rw_timer_func(struct timer_list *t)
{
struct f_gsi *gsi = from_timer(gsi, t, gsi_rw_timer);
if (!atomic_read(&gsi->connected)) {
log_event_dbg("%s: gsi not connected.. bail-out\n", __func__);
gsi->debugfs_rw_timer_enable = 0;
return;
}
log_event_dbg("%s: calling gsi_wakeup_host\n", __func__);
gsi_wakeup_host(gsi);
if (gsi->debugfs_rw_timer_enable) {
log_event_dbg("%s: re-arm the timer\n", __func__);
mod_timer(&gsi->gsi_rw_timer,
jiffies + msecs_to_jiffies(gsi->gsi_rw_timer_interval));
}
}
static struct f_gsi *get_connected_gsi(void)
{
struct f_gsi *connected_gsi;
int i;
for (i = 0; i < IPA_USB_MAX_TETH_PROT_SIZE; i++) {
if (inst_status[i].opts) {
connected_gsi = inst_status[i].opts->gsi;
if (connected_gsi && atomic_read(&connected_gsi->connected))
return connected_gsi;
}
}
return NULL;
}
#define DEFAULT_RW_TIMER_INTERVAL 500 /* in ms */
static ssize_t usb_gsi_rw_write(struct file *file,
const char __user *ubuf, size_t count, loff_t *ppos)
{
struct f_gsi *gsi = NULL;
struct usb_function *func;
struct usb_gadget *gadget;
u8 input;
int i;
int ret;
if (ubuf == NULL) {
log_event_dbg("%s: buffer is Null.\n", __func__);
goto err;
}
ret = kstrtou8_from_user(ubuf, count, 0, &input);
if (ret) {
log_event_err("%s: Invalid value. err:%d\n", __func__, ret);
goto err;
}
for (i = 0; i < IPA_USB_MAX_TETH_PROT_SIZE; i++) {
gsi = NULL;
if (inst_status[i].opts)
gsi = inst_status[i].opts->gsi;
if (gsi && atomic_read(&gsi->connected)) {
func = &gsi->function;
gadget = func->config->cdev->gadget;
gsi->debugfs_rw_timer_enable = !!input;
if (gadget->speed >= USB_SPEED_SUPER &&
!gsi->func_is_suspended) {
gsi->debugfs_rw_timer_enable = 0;
del_timer_sync(&gsi->gsi_rw_timer);
continue;
}
if (gsi->debugfs_rw_timer_enable) {
mod_timer(&gsi->gsi_rw_timer, jiffies +
msecs_to_jiffies(gsi->gsi_rw_timer_interval));
log_event_dbg("%s: timer initialized\n",
__func__);
} else {
del_timer_sync(&gsi->gsi_rw_timer);
log_event_dbg("%s: timer deleted\n", __func__);
}
if (gadget->speed < USB_SPEED_SUPER)
break;
}
}
err:
return count;
}
static int usb_gsi_rw_show(struct seq_file *s, void *unused)
{
struct f_gsi *gsi;
int i;
u8 enable = 0;
for (i = 0; i < IPA_USB_MAX_TETH_PROT_SIZE; i++) {
gsi = NULL;
if (inst_status[i].opts)
gsi = inst_status[i].opts->gsi;
if (gsi && atomic_read(&gsi->connected))
enable |= gsi->debugfs_rw_timer_enable;
}
seq_printf(s, "%d\n", enable);
return 0;
}
static int usb_gsi_rw_open(struct inode *inode, struct file *f)
{
return single_open(f, usb_gsi_rw_show, inode->i_private);
}
static const struct file_operations fops_usb_gsi_rw = {
.open = usb_gsi_rw_open,
.read = seq_read,
.write = usb_gsi_rw_write,
.owner = THIS_MODULE,
.llseek = seq_lseek,
.release = seq_release,
};
static ssize_t usb_gsi_rw_timer_write(struct file *file,
const char __user *ubuf, size_t count, loff_t *ppos)
{
struct f_gsi *gsi = NULL;
u16 timer_val;
int i;
int ret;
if (ubuf == NULL) {
log_event_dbg("%s: buffer is NULL.\n", __func__);
goto err;
}
ret = kstrtou16_from_user(ubuf, count, 0, &timer_val);
if (ret) {
log_event_err("%s: Invalid value. err:%d\n", __func__, ret);
goto err;
}
if (timer_val <= 0 || timer_val > 10000) {
log_event_err("%s: value must be > 0 and < 10000.\n", __func__);
goto err;
}
for (i = 0; i < IPA_USB_MAX_TETH_PROT_SIZE; i++) {
gsi = NULL;
if (inst_status[i].opts)
gsi = inst_status[i].opts->gsi;
if (gsi && atomic_read(&gsi->connected))
gsi->gsi_rw_timer_interval = timer_val;
}
err:
return count;
}
static int usb_gsi_rw_timer_show(struct seq_file *s, void *unused)
{
struct f_gsi *gsi;
gsi = get_connected_gsi();
if (!gsi) {
log_event_dbg("%s: gsi not connected\n", __func__);
return 0;
}
seq_printf(s, "%ums\n", gsi->gsi_rw_timer_interval);
return 0;
}
static int usb_gsi_rw_timer_open(struct inode *inode, struct file *f)
{
return single_open(f, usb_gsi_rw_timer_show, inode->i_private);
}
static const struct file_operations fops_usb_gsi_rw_timer = {
.open = usb_gsi_rw_timer_open,
.read = seq_read,
.write = usb_gsi_rw_timer_write,
.owner = THIS_MODULE,
.llseek = seq_lseek,
.release = seq_release,
};
static int usb_gsi_debugfs_init(void)
{
debugfs.debugfs_root = debugfs_create_dir("usb_gsi", NULL);
if (!debugfs.debugfs_root)
return -ENOMEM;
debugfs_create_file("remote_wakeup_enable", 0600,
debugfs.debugfs_root,
inst_status, &fops_usb_gsi_rw);
debugfs_create_file("remote_wakeup_interval", 0600,
debugfs.debugfs_root,
inst_status,
&fops_usb_gsi_rw_timer);
debugfs_create_bool("log_ctrl_packets", 0644, debugfs.debugfs_root,
&debugfs.qti_packet_debug);
return 0;
}
static void usb_gsi_debugfs_exit(void)
{
debugfs_remove_recursive(debugfs.debugfs_root);
}
/*
* Callback for when when network interface is up
* and userspace is ready to answer DHCP requests, or remote wakeup
*/
static int ipa_usb_notify_cb(enum ipa_usb_notify_event event,
void *driver_data)
{
struct f_gsi *gsi = driver_data;
unsigned long flags;
struct gsi_ctrl_pkt *cpkt_notify_connect, *cpkt_notify_speed;
if (!gsi) {
log_event_err("%s: invalid driver data", __func__);
return -EINVAL;
}
spin_lock_irqsave(&gsi->d_port.lock, flags);
switch (event) {
case IPA_USB_DEVICE_READY:
if (gsi->d_port.net_ready_trigger) {
spin_unlock_irqrestore(&gsi->d_port.lock, flags);
log_event_dbg("%s: Already triggered", __func__);
return 1;
}
log_event_err("%s: Set net_ready_trigger", __func__);
gsi->d_port.net_ready_trigger = true;
if (gsi->prot_id == IPA_USB_ECM) {
cpkt_notify_connect = gsi_ctrl_pkt_alloc(0, GFP_ATOMIC);
if (IS_ERR(cpkt_notify_connect)) {
spin_unlock_irqrestore(&gsi->d_port.lock,
flags);
log_event_dbg("%s: err cpkt_notify_connect\n",
__func__);
return -ENOMEM;
}
cpkt_notify_connect->type = GSI_CTRL_NOTIFY_CONNECT;
cpkt_notify_speed = gsi_ctrl_pkt_alloc(0, GFP_ATOMIC);
if (IS_ERR(cpkt_notify_speed)) {
spin_unlock_irqrestore(&gsi->d_port.lock,
flags);
gsi_ctrl_pkt_free(cpkt_notify_connect);
log_event_dbg("%s: err cpkt_notify_speed\n",
__func__);
return -ENOMEM;
}
cpkt_notify_speed->type = GSI_CTRL_NOTIFY_SPEED;
spin_lock(&gsi->c_port.lock);
list_add_tail(&cpkt_notify_connect->list,
&gsi->c_port.cpkt_resp_q);
list_add_tail(&cpkt_notify_speed->list,
&gsi->c_port.cpkt_resp_q);
spin_unlock(&gsi->c_port.lock);
gsi_ctrl_send_notification(gsi);
}
/*
* Do not post EVT_CONNECTED for RNDIS.
* Data path for RNDIS is enabled on EVT_HOST_READY.
*/
if (gsi->prot_id != IPA_USB_RNDIS) {
post_event(&gsi->d_port, EVT_IPA_READY);
queue_delayed_work(gsi->d_port.ipa_usb_wq,
&gsi->d_port.usb_ipa_w, 0);
}
break;
case IPA_USB_REMOTE_WAKEUP:
if (!gsi->rwake_inprogress)
gsi_wakeup_host(gsi);
break;
case IPA_USB_SUSPEND_COMPLETED:
post_event(&gsi->d_port, EVT_IPA_SUSPEND);
queue_delayed_work(gsi->d_port.ipa_usb_wq,
&gsi->d_port.usb_ipa_w, 0);
break;
}
spin_unlock_irqrestore(&gsi->d_port.lock, flags);
return 1;
}
static int ipa_connect_channels(struct gsi_data_port *d_port)
{
int ret;
struct f_gsi *gsi = d_port_to_gsi(d_port);
struct ipa_usb_xdci_chan_params *in_params =
&d_port->ipa_in_channel_params;
struct ipa_usb_xdci_chan_params *out_params =
&d_port->ipa_out_channel_params;
struct ipa_usb_xdci_connect_params *conn_params =
&d_port->ipa_conn_pms;
struct usb_composite_dev *cdev = gsi->function.config->cdev;
struct gsi_channel_info gsi_channel_info;
struct ipa_req_chan_out_params ipa_in_channel_out_params;
struct ipa_req_chan_out_params ipa_out_channel_out_params;
if (gsi->prot_id == IPA_USB_RMNET) {
d_port->in_request.use_tcm_mem = gsi->rmnet_use_tcm_mem;
/* override needed for moving from LLCC TCM to DDR memory */
d_port->in_request.buf_len = GSI_IN_RMNET_BUFF_SIZE;
d_port->in_request.num_bufs = GSI_NUM_IN_RMNET_BUFFERS;
}
ret = usb_gsi_ep_op(d_port->in_ep, &d_port->in_request,
GSI_EP_OP_PREPARE_TRBS);
if (ret) {
log_event_err("%s: GSI_EP_OP_PREPARE_TRBS failed: %d\n",
__func__, ret);
return ret;
}
log_event_dbg("IN: num_bufs:=%zu, buf_len=%zu\n",
d_port->in_request.num_bufs, d_port->in_request.buf_len);
ret = usb_gsi_ep_op(d_port->in_ep, &d_port->in_request,
GSI_EP_OP_STARTXFER);
if (ret) {
log_event_err("%s: GSI_EP_OP_STARTXFER failed: %d\n",
__func__, ret);
goto free_trb_ep_in;
}
d_port->in_xfer_rsc_index = usb_gsi_ep_op(d_port->in_ep, NULL,
GSI_EP_OP_GET_XFER_IDX);
memset(in_params, 0x0, sizeof(*in_params));
gsi_channel_info.ch_req = &d_port->in_request;
usb_gsi_ep_op(d_port->in_ep, (void *)&gsi_channel_info,
GSI_EP_OP_GET_CH_INFO);
log_event_dbg("%s: USB GSI IN OPS Completed", __func__);
in_params->client =
(gsi->prot_id != IPA_USB_DIAG) ? IPA_CLIENT_USB_CONS :
IPA_CLIENT_USB_DPL_CONS;
in_params->ipa_ep_cfg.mode.mode = IPA_BASIC;
in_params->teth_prot = gsi->prot_id;
in_params->gevntcount_low_addr =
gsi_channel_info.gevntcount_low_addr;
in_params->gevntcount_hi_addr =
gsi_channel_info.gevntcount_hi_addr;
in_params->dir = GSI_CHAN_DIR_FROM_GSI;
in_params->xfer_ring_len = gsi_channel_info.xfer_ring_len;
in_params->xfer_scratch.last_trb_addr_iova =
gsi_channel_info.last_trb_addr;
in_params->xfer_ring_base_addr_iova =
gsi_channel_info.xfer_ring_base_addr;
in_params->data_buff_base_len = d_port->in_request.buf_len *
d_port->in_request.num_bufs;
in_params->data_buff_base_addr_iova = d_port->in_request.dma;
in_params->sgt_xfer_rings = &d_port->in_request.sgt_trb_xfer_ring;
in_params->sgt_data_buff = &d_port->in_request.sgt_data_buff;
log_event_dbg("%s(): IN: sgt_xfer_rings:%pK sgt_data_buff:%pK\n",
__func__, in_params->sgt_xfer_rings, in_params->sgt_data_buff);
in_params->xfer_scratch.const_buffer_size =
gsi_channel_info.const_buffer_size;
in_params->xfer_scratch.depcmd_low_addr =
gsi_channel_info.depcmd_low_addr;
in_params->xfer_scratch.depcmd_hi_addr =
gsi_channel_info.depcmd_hi_addr;
if (d_port->out_ep) {
log_event_dbg("OUT: num_bufs:=%zu, buf_len=%zu\n",
d_port->out_request.num_bufs,
d_port->out_request.buf_len);
ret = usb_gsi_ep_op(d_port->out_ep, &d_port->out_request,
GSI_EP_OP_PREPARE_TRBS);
if (ret) {
log_event_err("%s: GSI_EP_OP_PREPARE_TRBS failed: %d\n",
__func__, ret);
goto end_xfer_ep_in;
}
ret = usb_gsi_ep_op(d_port->out_ep, &d_port->out_request,
GSI_EP_OP_STARTXFER);
if (ret) {
log_event_err("%s: GSI_EP_OP_STARTXFER failed: %d\n",
__func__, ret);
goto free_trb_ep_out;
}
d_port->out_xfer_rsc_index =
usb_gsi_ep_op(d_port->out_ep,
NULL, GSI_EP_OP_GET_XFER_IDX);
memset(out_params, 0x0, sizeof(*out_params));
gsi_channel_info.ch_req = &d_port->out_request;
usb_gsi_ep_op(d_port->out_ep, (void *)&gsi_channel_info,
GSI_EP_OP_GET_CH_INFO);
log_event_dbg("%s: USB GSI OUT OPS Completed", __func__);
out_params->client = IPA_CLIENT_USB_PROD;
out_params->ipa_ep_cfg.mode.mode = IPA_BASIC;
out_params->teth_prot = gsi->prot_id;
out_params->gevntcount_low_addr =
gsi_channel_info.gevntcount_low_addr;
out_params->gevntcount_hi_addr =
gsi_channel_info.gevntcount_hi_addr;
out_params->dir = GSI_CHAN_DIR_TO_GSI;
out_params->xfer_ring_len =
gsi_channel_info.xfer_ring_len;
out_params->xfer_ring_base_addr_iova =
gsi_channel_info.xfer_ring_base_addr;
out_params->data_buff_base_len = d_port->out_request.buf_len *
d_port->out_request.num_bufs;
out_params->data_buff_base_addr_iova =
d_port->out_request.dma;
out_params->sgt_xfer_rings =
&d_port->out_request.sgt_trb_xfer_ring;
out_params->sgt_data_buff = &d_port->out_request.sgt_data_buff;
log_event_dbg("%s(): OUT: sgt_xfer_rings:%pK sgt_data_buff:%pK\n",
__func__, out_params->sgt_xfer_rings,
out_params->sgt_data_buff);
out_params->xfer_scratch.last_trb_addr_iova =
gsi_channel_info.last_trb_addr;
out_params->xfer_scratch.const_buffer_size =
gsi_channel_info.const_buffer_size;
out_params->xfer_scratch.depcmd_low_addr =
gsi_channel_info.depcmd_low_addr;
out_params->xfer_scratch.depcmd_hi_addr =
gsi_channel_info.depcmd_hi_addr;
}
/* Populate connection params */
conn_params->max_pkt_size =
(cdev->gadget->speed >= USB_SPEED_SUPER) ?
IPA_USB_SUPER_SPEED_1024B : IPA_USB_HIGH_SPEED_512B;
conn_params->ipa_to_usb_xferrscidx =
d_port->in_xfer_rsc_index;
conn_params->usb_to_ipa_xferrscidx =
d_port->out_xfer_rsc_index;
conn_params->usb_to_ipa_xferrscidx_valid =
(gsi->prot_id != IPA_USB_DIAG) ? true : false;
conn_params->ipa_to_usb_xferrscidx_valid = true;
conn_params->teth_prot = gsi->prot_id;
conn_params->teth_prot_params.max_xfer_size_bytes_to_dev = 23700;
conn_params->teth_prot_params.max_xfer_size_bytes_to_dev
= d_port->out_aggr_size;
conn_params->teth_prot_params.max_xfer_size_bytes_to_host
= d_port->in_aggr_size;
conn_params->teth_prot_params.max_packet_number_to_dev =
DEFAULT_MAX_PKT_PER_XFER;
conn_params->max_supported_bandwidth_mbps =
(cdev->gadget->speed >= USB_SPEED_SUPER) ? 3600 : 400;
memset(&ipa_in_channel_out_params, 0x0,
sizeof(ipa_in_channel_out_params));
memset(&ipa_out_channel_out_params, 0x0,
sizeof(ipa_out_channel_out_params));
log_event_dbg("%s: Calling xdci_connect", __func__);
ret = ipa_usb_xdci_connect(out_params, in_params,
&ipa_out_channel_out_params,
&ipa_in_channel_out_params,
conn_params);
if (ret) {
log_event_err("%s: IPA connect failed %d", __func__, ret);
goto end_xfer_ep_out;
}
log_event_dbg("%s: xdci_connect done", __func__);
log_event_dbg("%s: IN CH HDL %x", __func__,
ipa_in_channel_out_params.clnt_hdl);
log_event_dbg("%s: IN CH DBL addr %x", __func__,
ipa_in_channel_out_params.db_reg_phs_addr_lsb);
log_event_dbg("%s: OUT CH HDL %x", __func__,
ipa_out_channel_out_params.clnt_hdl);
log_event_dbg("%s: OUT CH DBL addr %x", __func__,
ipa_out_channel_out_params.db_reg_phs_addr_lsb);
d_port->in_channel_handle = ipa_in_channel_out_params.clnt_hdl;
d_port->in_request.db_reg_phs_addr_lsb =
ipa_in_channel_out_params.db_reg_phs_addr_lsb;
d_port->in_request.db_reg_phs_addr_msb =
ipa_in_channel_out_params.db_reg_phs_addr_msb;
if (gsi->prot_id != IPA_USB_DIAG) {
d_port->out_channel_handle =
ipa_out_channel_out_params.clnt_hdl;
d_port->out_request.db_reg_phs_addr_lsb =
ipa_out_channel_out_params.db_reg_phs_addr_lsb;
d_port->out_request.db_reg_phs_addr_msb =
ipa_out_channel_out_params.db_reg_phs_addr_msb;
}
return ret;
end_xfer_ep_out:
if (d_port->out_ep)
usb_gsi_ep_op(d_port->out_ep, NULL,
GSI_EP_OP_ENDXFER);
free_trb_ep_out:
if (d_port->out_ep)
usb_gsi_ep_op(d_port->out_ep, &d_port->out_request,
GSI_EP_OP_FREE_TRBS);
end_xfer_ep_in:
usb_gsi_ep_op(d_port->in_ep, NULL,
GSI_EP_OP_ENDXFER);
free_trb_ep_in:
usb_gsi_ep_op(d_port->in_ep, &d_port->in_request,
GSI_EP_OP_FREE_TRBS);
return ret;
}
static void ipa_data_path_enable(struct gsi_data_port *d_port)
{
struct f_gsi *gsi = d_port_to_gsi(d_port);
bool block_db = false;
log_event_dbg("IN: db_reg_phs_addr_lsb = %x",
gsi->d_port.in_request.db_reg_phs_addr_lsb);
usb_gsi_ep_op(gsi->d_port.in_ep,
&gsi->d_port.in_request,
GSI_EP_OP_STORE_DBL_INFO);
if (gsi->d_port.out_ep) {
log_event_dbg("OUT: db_reg_phs_addr_lsb = %x",
gsi->d_port.out_request.db_reg_phs_addr_lsb);
usb_gsi_ep_op(gsi->d_port.out_ep,
&gsi->d_port.out_request,
GSI_EP_OP_STORE_DBL_INFO);
}
usb_gsi_ep_op(gsi->d_port.in_ep, &gsi->d_port.in_request,
GSI_EP_OP_ENABLE_GSI);
/* Unblock doorbell to GSI */
usb_gsi_ep_op(d_port->in_ep, (void *)&block_db,
GSI_EP_OP_SET_CLR_BLOCK_DBL);
usb_gsi_ep_op(gsi->d_port.in_ep, &gsi->d_port.in_request,
GSI_EP_OP_RING_DB);
if (gsi->d_port.out_ep)
usb_gsi_ep_op(gsi->d_port.out_ep, &gsi->d_port.out_request,
GSI_EP_OP_RING_DB);
}
static void ipa_data_path_disable(struct gsi_data_port *d_port)
{
struct f_gsi *gsi = d_port_to_gsi(d_port);
bool block_db = true;
log_event_dbg("%s: Disable eps", __func__);
if (gsi->d_port.in_ep)
usb_gsi_ep_op(gsi->d_port.in_ep,
&gsi->d_port.in_request, GSI_EP_OP_DISABLE);
if (gsi->d_port.out_ep)
usb_gsi_ep_op(gsi->d_port.out_ep,
&gsi->d_port.out_request, GSI_EP_OP_DISABLE);
if (gsi->d_port.in_ep)
usb_gsi_ep_op(d_port->in_ep, (void *)&block_db,
GSI_EP_OP_SET_CLR_BLOCK_DBL);
}
static void ipa_disconnect_channel(struct gsi_data_port *d_port)
{
int ret;
struct f_gsi *gsi = d_port_to_gsi(d_port);
log_event_dbg("%s: Calling xdci_disconnect", __func__);
ret = ipa_usb_xdci_disconnect(gsi->d_port.out_channel_handle,
gsi->d_port.in_channel_handle, gsi->prot_id);
if (ret)
log_event_err("%s: IPA disconnect failed %d",
__func__, ret);
log_event_dbg("%s: xdci_disconnect done", __func__);
/* invalidate channel handles*/
gsi->d_port.in_channel_handle = -EINVAL;
gsi->d_port.out_channel_handle = -EINVAL;
if (gsi->d_port.in_ep)
usb_gsi_ep_op(gsi->d_port.in_ep, &gsi->d_port.in_request,
GSI_EP_OP_FREE_TRBS);
if (gsi->d_port.out_ep)
usb_gsi_ep_op(gsi->d_port.out_ep, &gsi->d_port.out_request,
GSI_EP_OP_FREE_TRBS);
}
static int ipa_suspend_work_handler(struct gsi_data_port *d_port)
{
int ret = 0;
bool block_db, f_suspend;
struct f_gsi *gsi = d_port_to_gsi(d_port);
struct usb_function *f = &gsi->function;
f_suspend = gsi->func_wakeup_allowed;
log_event_dbg("%s: f_suspend:%d", __func__, f_suspend);
if (!usb_gsi_ep_op(gsi->d_port.in_ep, (void *) &f_suspend,
GSI_EP_OP_CHECK_FOR_SUSPEND)) {
ret = -EFAULT;
block_db = false;
usb_gsi_ep_op(d_port->in_ep, (void *)&block_db,
GSI_EP_OP_SET_CLR_BLOCK_DBL);
goto done;
}
/*
* Ensure that the DBL is blocked before suspend.
*/
block_db = true;
usb_gsi_ep_op(gsi->d_port.in_ep, (void *)&block_db,
GSI_EP_OP_SET_CLR_BLOCK_DBL);
log_event_dbg("%s: Calling xdci_suspend", __func__);
ret = ipa_usb_xdci_suspend(gsi->d_port.out_channel_handle,
gsi->d_port.in_channel_handle, gsi->prot_id,
usb_gsi_remote_wakeup_allowed(f));
if (!ret) {
d_port->sm_state = STATE_SUSPENDED;
log_event_dbg("%s: STATE SUSPENDED", __func__);
goto done;
}
if (ret == -EFAULT) {
block_db = false;
usb_gsi_ep_op(d_port->in_ep, (void *)&block_db,
GSI_EP_OP_SET_CLR_BLOCK_DBL);
gsi_wakeup_host(gsi);
} else if (ret == -EINPROGRESS) {
d_port->sm_state = STATE_SUSPEND_IN_PROGRESS;
} else {
log_event_err("%s: Error %d for %d", __func__, ret,
gsi->prot_id);
}
done:
log_event_dbg("%s: xdci_suspend ret %d", __func__, ret);
return ret;
}
static void ipa_resume_work_handler(struct gsi_data_port *d_port)
{
bool block_db;
struct f_gsi *gsi = d_port_to_gsi(d_port);
int ret;
log_event_dbg("%s: Calling xdci_resume", __func__);
ret = ipa_usb_xdci_resume(gsi->d_port.out_channel_handle,
gsi->d_port.in_channel_handle,
gsi->prot_id);
if (ret)
log_event_dbg("%s: xdci_resume ret %d", __func__, ret);
log_event_dbg("%s: xdci_resume done", __func__);
block_db = false;
usb_gsi_ep_op(d_port->in_ep, (void *)&block_db,
GSI_EP_OP_SET_CLR_BLOCK_DBL);
}
static int gsi_ep_enable(struct f_gsi *gsi)
{
int ret;
if (gsi->d_port.in_ep && !gsi->d_port.in_ep->desc) {
ret = config_ep_by_speed(gsi->d_port.gadget, &gsi->function,
gsi->d_port.in_ep);
if (ret)
return ret;
log_event_dbg("%s: Enable IN ep", __func__);
ret = usb_gsi_ep_op(gsi->d_port.in_ep,
&gsi->d_port.in_request, GSI_EP_OP_CONFIG);
if (ret)
return ret;
}
if (gsi->d_port.out_ep && !gsi->d_port.out_ep->desc) {
ret = config_ep_by_speed(gsi->d_port.gadget, &gsi->function,
gsi->d_port.out_ep);
if (ret)
return ret;
log_event_dbg("%s: Enable OUT ep", __func__);
ret = usb_gsi_ep_op(gsi->d_port.out_ep,
&gsi->d_port.out_request, GSI_EP_OP_CONFIG);
if (ret) {
if (gsi->d_port.in_ep)
usb_gsi_ep_op(gsi->d_port.in_ep,
&gsi->d_port.in_request,
GSI_EP_OP_DISABLE);
return ret;
}
}
return 0;
}
static void ipa_work_handler(struct work_struct *w)
{
struct gsi_data_port *d_port = container_of(w, struct gsi_data_port,
usb_ipa_w.work);
u8 event;
int ret = 0;
struct usb_gadget *gadget = d_port->gadget;
struct device *dev;
struct device *gad_dev;
struct f_gsi *gsi = d_port_to_gsi(d_port);
bool block_db;
event = read_event(d_port);
log_event_dbg("%s: event = %x sm_state %x", __func__,
event, d_port->sm_state);
if (gadget) {
dev = &gadget->dev;
if (!dev || !dev->parent) {
log_event_err("%s(): dev or dev->parent is NULL.\n",
__func__);
return;
}
gad_dev = dev->parent;
} else {
log_event_err("%s(): gadget is NULL.\n", __func__);
return;
}
switch (d_port->sm_state) {
case STATE_UNINITIALIZED:
break;
case STATE_INITIALIZED:
if (event == EVT_SET_ALT) {
if (!atomic_read(&gsi->connected)) {
log_event_err("USB cable not connected\n");
break;
}
usb_gadget_autopm_get(d_port->gadget);
log_event_dbg("%s: get = %d", __func__,
atomic_read(&gad_dev->power.usage_count));
/* Configure EPs for GSI */
ret = gsi_ep_enable(gsi);
if (ret) {
log_event_err("%s:ep enable err %d", __func__,
ret);
usb_composite_setup_continue(gsi->d_port.cdev);
usb_gadget_autopm_put_async(d_port->gadget);
break;
}
usb_composite_setup_continue(gsi->d_port.cdev);
/*
* Skip rest for RNDIS and wait for EVT_HOST_READY
* which is invoked when the host sends the
* GEN_CURRENT_PACKET_FILTER message.
*/
if (gsi->prot_id == IPA_USB_RNDIS) {
d_port->sm_state = STATE_HOST_NRDY;
usb_gadget_autopm_put_async(d_port->gadget);
break;
}
ret = ipa_connect_channels(d_port);
if (ret) {
log_event_err("%s: ipa_connect_channels failed\n",
__func__);
ipa_data_path_disable(d_port);
usb_gadget_autopm_put_async(d_port->gadget);
d_port->sm_state = STATE_INITIALIZED;
break;
}
d_port->sm_state = STATE_WAIT_FOR_IPA_RDY;
log_event_dbg("%s: ST_INIT_EVT_SET_ALT",
__func__);
}
break;
case STATE_WAIT_FOR_IPA_RDY:
if (event == EVT_IPA_READY) {
if (peek_event(d_port) == EVT_SUSPEND) {
log_event_dbg("%s: ST_WAIT_IPARDY_EVT_SUSPEND",
__func__);
break;
}
ipa_data_path_enable(d_port);
d_port->sm_state = STATE_CONNECTED;
log_event_dbg("%s: ST_WAIT_IPARDY_EVT_IPARDY %d",
__func__, __LINE__);
} else if (event == EVT_SUSPEND) {
if (peek_event(d_port) == EVT_DISCONNECTED) {
read_event(d_port);
ipa_data_path_disable(d_port);
ipa_disconnect_channel(d_port);
d_port->sm_state = STATE_INITIALIZED;
usb_gadget_autopm_put_async(d_port->gadget);
log_event_dbg("%s: ST_WAIT_IPARDY_EVT_SUS_DIS",
__func__);
log_event_dbg("%s: put_async1 = %d", __func__,
atomic_read(
&gad_dev->power.usage_count));
break;
}
ret = ipa_suspend_work_handler(d_port);
if (!ret) {
usb_gadget_autopm_put_async(d_port->gadget);
log_event_dbg("%s: ST_WAIT_IPARDY_EVT_SUS",
__func__);
log_event_dbg("%s: put_async2 = %d", __func__,
atomic_read(
&gad_dev->power.usage_count));
}
} else if (event == EVT_DISCONNECTED) {
ipa_data_path_disable(d_port);
ipa_disconnect_channel(d_port);
d_port->sm_state = STATE_INITIALIZED;
usb_gadget_autopm_put_async(d_port->gadget);
log_event_dbg("%s: ST_WAIT_IPARDY_EVT_DIS",
__func__);
log_event_dbg("%s: put_async3 = %d",
__func__, atomic_read(
&gad_dev->power.usage_count));
}
break;
case STATE_CONNECTED:
if (event == EVT_DISCONNECTED || event == EVT_HOST_NRDY) {
if (peek_event(d_port) == EVT_HOST_READY) {
read_event(d_port);
log_event_dbg("%s: NO_OP NRDY_RDY", __func__);
break;
}
if (event == EVT_HOST_NRDY) {
log_event_dbg("%s: ST_CON_HOST_NRDY\n",
__func__);
block_db = true;
/* stop USB ringing doorbell to GSI(OUT_EP) */
usb_gsi_ep_op(d_port->in_ep, (void *)&block_db,
GSI_EP_OP_SET_CLR_BLOCK_DBL);
gsi_rndis_ipa_reset_trigger(d_port);
usb_gsi_ep_op(d_port->in_ep, NULL,
GSI_EP_OP_ENDXFER);
usb_gsi_ep_op(d_port->out_ep, NULL,
GSI_EP_OP_ENDXFER);
/*
* don't disable endpoints for RNDIS flow
* control enable
*/
ipa_disconnect_channel(d_port);
d_port->sm_state = STATE_HOST_NRDY;
log_event_dbg("%s: ST_CON_EVT_HNRDY", __func__);
} else {
ipa_data_path_disable(d_port);
ipa_disconnect_channel(d_port);
d_port->sm_state = STATE_INITIALIZED;
log_event_dbg("%s: ST_CON_EVT_DIS", __func__);
}
usb_gadget_autopm_put_async(d_port->gadget);
log_event_dbg("%s: put_async4 = %d",
__func__, atomic_read(
&gad_dev->power.usage_count));
} else if (event == EVT_SUSPEND) {
if (peek_event(d_port) == EVT_DISCONNECTED) {
read_event(d_port);
ipa_data_path_disable(d_port);
ipa_disconnect_channel(d_port);
d_port->sm_state = STATE_INITIALIZED;
usb_gadget_autopm_put_async(d_port->gadget);
log_event_dbg("%s: ST_CON_EVT_SUS_DIS",
__func__);
log_event_dbg("%s: put_async5 = %d",
__func__, atomic_read(
&gad_dev->power.usage_count));
break;
}
ret = ipa_suspend_work_handler(d_port);
if (!ret) {
usb_gadget_autopm_put_async(d_port->gadget);
log_event_dbg("%s: ST_CON_EVT_SUS",
__func__);
log_event_dbg("%s: put_async6 = %d",
__func__, atomic_read(
&gad_dev->power.usage_count));
}
}
break;
case STATE_HOST_NRDY:
if (event == EVT_DISCONNECTED) {
usb_gadget_autopm_get(d_port->gadget);
ipa_data_path_disable(d_port);
d_port->sm_state = STATE_INITIALIZED;
usb_gadget_autopm_put_async(d_port->gadget);
log_event_dbg("%s: ST_HOST_NRDY_EVT_DIS", __func__);
log_event_dbg("%s: put_async = %d",
__func__, atomic_read(
&gad_dev->power.usage_count));
} else if (event == EVT_HOST_READY) {
/*
* When in a composition such as RNDIS + ADB,
* RNDIS host sends a GEN_CURRENT_PACKET_FILTER msg
* to enable/disable flow control eg. during RNDIS
* adaptor disable/enable from device manager.
* In the case of the msg to disable flow control,
* connect IPA channels and enable data path.
* EVT_HOST_READY is posted to the state machine
* in the handler for this msg.
*/
usb_gadget_autopm_get(d_port->gadget);
log_event_dbg("%s: get = %d", __func__,
atomic_read(&gad_dev->power.usage_count));
ret = ipa_connect_channels(d_port);
if (ret) {
log_event_err("%s: ipa_connect_channels failed\n",
__func__);
usb_gadget_autopm_put_async(d_port->gadget);
break;
}
ipa_data_path_enable(d_port);
d_port->sm_state = STATE_CONNECTED;
log_event_dbg("%s: ST_HOST_NRDY_EVT_HRDY_", __func__);
}
break;
case STATE_SUSPEND_IN_PROGRESS:
if (event == EVT_IPA_SUSPEND) {
d_port->sm_state = STATE_SUSPENDED;
usb_gadget_autopm_put_async(d_port->gadget);
log_event_dbg("%s: ST_SUS_IN_PROG_EVT_IPA_SUS",
__func__);
log_event_dbg("%s: put_async6 = %d",
__func__, atomic_read(
&gad_dev->power.usage_count));
} else if (event == EVT_RESUMED) {
ipa_resume_work_handler(d_port);
d_port->sm_state = STATE_CONNECTED;
/*
* Increment usage count here to disallow gadget
* parent suspend. This counter will decrement
* after IPA disconnect is done in disconnect work
* (due to cable disconnect) or in suspended state.
*/
usb_gadget_autopm_get_noresume(d_port->gadget);
log_event_dbg("%s: ST_SUS_IN_PROG_EVT_RES", __func__);
log_event_dbg("%s: get_nores1 = %d", __func__,
atomic_read(
&gad_dev->power.usage_count));
} else if (event == EVT_DISCONNECTED) {
ipa_data_path_disable(d_port);
ipa_disconnect_channel(d_port);
d_port->sm_state = STATE_INITIALIZED;
usb_gadget_autopm_put_async(d_port->gadget);
log_event_dbg("%s: ST_SUS_IN_PROG_EVT_DIS", __func__);
log_event_dbg("%s: put_async7 = %d", __func__,
atomic_read(
&gad_dev->power.usage_count));
}
break;
case STATE_SUSPENDED:
if (event == EVT_RESUMED) {
usb_gadget_autopm_get(d_port->gadget);
log_event_dbg("%s: ST_SUS_EVT_RES", __func__);
log_event_dbg("%s: get = %d", __func__,
atomic_read(&gad_dev->power.usage_count));
ipa_resume_work_handler(d_port);
d_port->sm_state = STATE_CONNECTED;
} else if (event == EVT_DISCONNECTED) {
usb_gadget_autopm_get(d_port->gadget);
ipa_data_path_disable(d_port);
ipa_disconnect_channel(d_port);
d_port->sm_state = STATE_INITIALIZED;
log_event_dbg("%s: ST_SUS_EVT_DIS", __func__);
usb_gadget_autopm_put_async(d_port->gadget);
}
break;
default:
log_event_dbg("%s: Invalid state to SM", __func__);
}
if (peek_event(d_port) != EVT_NONE) {
log_event_dbg("%s: New events to process", __func__);
queue_delayed_work(d_port->ipa_usb_wq, &d_port->usb_ipa_w, 0);
}
}
static struct gsi_ctrl_pkt *gsi_ctrl_pkt_alloc(unsigned int len, gfp_t flags)
{
struct gsi_ctrl_pkt *pkt;
pkt = kzalloc(sizeof(struct gsi_ctrl_pkt), flags);
if (!pkt)
return ERR_PTR(-ENOMEM);
pkt->buf = kmalloc(len, flags);
if (!pkt->buf) {
kfree(pkt);
return ERR_PTR(-ENOMEM);
}
pkt->len = len;
return pkt;
}
static void gsi_ctrl_pkt_free(struct gsi_ctrl_pkt *pkt)
{
if (pkt) {
kfree(pkt->buf);
kfree(pkt);
}
}
static void gsi_ctrl_clear_cpkt_queues(struct f_gsi *gsi, bool skip_req_q)
{
struct gsi_ctrl_pkt *cpkt = NULL;
struct list_head *act, *tmp;
unsigned long flags;
spin_lock_irqsave(&gsi->c_port.lock, flags);
if (skip_req_q)
goto clean_resp_q;
list_for_each_safe(act, tmp, &gsi->c_port.cpkt_req_q) {
cpkt = list_entry(act, struct gsi_ctrl_pkt, list);
list_del(&cpkt->list);
gsi_ctrl_pkt_free(cpkt);
}
clean_resp_q:
list_for_each_safe(act, tmp, &gsi->c_port.cpkt_resp_q) {
cpkt = list_entry(act, struct gsi_ctrl_pkt, list);
list_del(&cpkt->list);
gsi_ctrl_pkt_free(cpkt);
}
spin_unlock_irqrestore(&gsi->c_port.lock, flags);
}
static int gsi_ctrl_send_cpkt_tomodem(struct f_gsi *gsi, void *buf, size_t len)
{
unsigned long flags;
struct gsi_ctrl_port *c_port = &gsi->c_port;
struct gsi_ctrl_pkt *cpkt;
spin_lock_irqsave(&c_port->lock, flags);
/* drop cpkt if port is not open */
if (!gsi->c_port.is_open) {
log_event_dbg("%s: ctrl device %s is not open",
__func__, gsi->c_port.name);
c_port->cpkt_drop_cnt++;
spin_unlock_irqrestore(&c_port->lock, flags);
return -ENODEV;
}
cpkt = gsi_ctrl_pkt_alloc(len, GFP_ATOMIC);
if (IS_ERR(cpkt)) {
log_event_err("%s: Reset func pkt allocation failed", __func__);
spin_unlock_irqrestore(&c_port->lock, flags);
return -ENOMEM;
}
memcpy(cpkt->buf, buf, len);
cpkt->len = len;
list_add_tail(&cpkt->list, &c_port->cpkt_req_q);
c_port->host_to_modem++;
spin_unlock_irqrestore(&c_port->lock, flags);
log_event_dbg("%s: Wake up read queue", __func__);
wake_up(&c_port->read_wq);
return 0;
}
static int gsi_ctrl_dev_open(struct inode *ip, struct file *fp)
{
struct gsi_ctrl_port *c_port = container_of(ip->i_cdev,
struct gsi_ctrl_port, cdev);
struct f_gsi *gsi;
struct gsi_inst_status *inst_cur;
if (!c_port) {
pr_err_ratelimited("%s: gsi ctrl port %pK\n", __func__, c_port);
return -ENODEV;
}
gsi = container_of(c_port, struct f_gsi, c_port);
inst_cur = &inst_status[gsi->prot_id];
log_event_dbg("%s: open ctrl dev %s", __func__, c_port->name);
mutex_lock(&inst_cur->gsi_lock);
fp->private_data = &gsi->prot_id;
if (!inst_cur->inst_exist) {
mutex_unlock(&inst_cur->gsi_lock);
log_event_err("%s: [prot_id = %d], GSI instance freed already\n",
__func__, gsi->prot_id);
return -ENODEV;
}
if (c_port->is_open) {
mutex_unlock(&inst_cur->gsi_lock);
log_event_err("%s: Already opened\n", __func__);
return -EBUSY;
}
c_port->is_open = true;
mutex_unlock(&inst_cur->gsi_lock);
return 0;
}
static int gsi_ctrl_dev_release(struct inode *ip, struct file *fp)
{
enum ipa_usb_teth_prot prot_id =
*(enum ipa_usb_teth_prot *)(fp->private_data);
struct gsi_inst_status *inst_cur = &inst_status[prot_id];
struct f_gsi *gsi;
mutex_lock(&inst_cur->gsi_lock);
if (unlikely(!inst_cur->inst_exist)) {
if (inst_cur->opts) {
/* GSI instance clean up */
gsi_inst_clean(inst_cur->opts);
inst_cur->opts = NULL;
}
mutex_unlock(&inst_cur->gsi_lock);
pr_err_ratelimited("%s: prot_id:%d: delayed free memory\n",
__func__, prot_id);
return -ENODEV;
}
inst_cur->opts->gsi->c_port.is_open = false;
gsi = inst_cur->opts->gsi;
mutex_unlock(&inst_cur->gsi_lock);
log_event_dbg("close ctrl dev %s\n",
inst_cur->opts->gsi->c_port.name);
return 0;
}
static ssize_t
gsi_ctrl_dev_read(struct file *fp, char __user *buf, size_t count, loff_t *pos)
{
struct gsi_ctrl_port *c_port;
struct gsi_ctrl_pkt *cpkt = NULL;
enum ipa_usb_teth_prot prot_id =
*(enum ipa_usb_teth_prot *)(fp->private_data);
struct gsi_inst_status *inst_cur = &inst_status[prot_id];
struct f_gsi *gsi;
unsigned long flags;
int ret = 0;
pr_debug("%s: Enter %zu\n", __func__, count);
mutex_lock(&inst_cur->gsi_lock);
if (unlikely(!inst_cur->inst_exist)) {
mutex_unlock(&inst_cur->gsi_lock);
pr_err_ratelimited("%s: free_inst is called and being freed\n",
__func__);
return -ENODEV;
}
mutex_unlock(&inst_cur->gsi_lock);
gsi = inst_cur->opts->gsi;
c_port = &inst_cur->opts->gsi->c_port;
if (!c_port) {
log_event_err("%s: gsi ctrl port %pK", __func__, c_port);
return -ENODEV;
}
if (count > GSI_MAX_CTRL_PKT_SIZE) {
log_event_err("Large buff size %zu, should be %d",
count, GSI_MAX_CTRL_PKT_SIZE);
return -EINVAL;
}
/* block until a new packet is available */
spin_lock_irqsave(&c_port->lock, flags);
while (list_empty(&c_port->cpkt_req_q)) {
log_event_dbg("Requests list is empty. Wait.");
spin_unlock_irqrestore(&c_port->lock, flags);
ret = wait_event_interruptible(c_port->read_wq,
!list_empty(&c_port->cpkt_req_q));
if (ret < 0) {
log_event_err("Waiting failed");
return -ERESTARTSYS;
}
log_event_dbg("Received request packet");
spin_lock_irqsave(&c_port->lock, flags);
}
cpkt = list_first_entry(&c_port->cpkt_req_q, struct gsi_ctrl_pkt,
list);
list_del(&cpkt->list);
spin_unlock_irqrestore(&c_port->lock, flags);
if (cpkt->len > count) {
log_event_err("cpkt size large:%d > buf size:%zu",
cpkt->len, count);
gsi_ctrl_pkt_free(cpkt);
return -ENOMEM;
}
log_event_dbg("%s: cpkt size:%d", __func__, cpkt->len);
if (debugfs.qti_packet_debug)
print_hex_dump(KERN_DEBUG, "READ:", DUMP_PREFIX_OFFSET, 16, 1,
cpkt->buf, min_t(int, 30, cpkt->len), false);
ret = copy_to_user(buf, cpkt->buf, cpkt->len);
if (ret) {
log_event_err("copy_to_user failed: err %d", ret);
ret = -EFAULT;
} else {
log_event_dbg("%s: copied %d bytes to user", __func__,
cpkt->len);
ret = cpkt->len;
c_port->copied_to_modem++;
}
gsi_ctrl_pkt_free(cpkt);
log_event_dbg("%s: Exit %zu", __func__, count);
return ret;
}
static ssize_t gsi_ctrl_dev_write(struct file *fp, const char __user *buf,
size_t count, loff_t *pos)
{
int ret = 0;
unsigned long flags;
struct gsi_ctrl_pkt *cpkt;
struct gsi_ctrl_port *c_port;
enum ipa_usb_teth_prot prot_id =
*(enum ipa_usb_teth_prot *)(fp->private_data);
struct gsi_inst_status *inst_cur = &inst_status[prot_id];
struct f_gsi *gsi;
if (prot_id == IPA_USB_DIAG)
return -EINVAL;
pr_debug("Enter %zu\n", count);
mutex_lock(&inst_cur->gsi_lock);
if (unlikely(!inst_cur->inst_exist)) {
mutex_unlock(&inst_cur->gsi_lock);
pr_err_ratelimited("%s: free_inst is called and being freed\n",
__func__);
return -ENODEV;
}
mutex_unlock(&inst_cur->gsi_lock);
gsi = inst_cur->opts->gsi;
c_port = &gsi->c_port;
if (!count || count > GSI_MAX_CTRL_PKT_SIZE) {
log_event_err("error: ctrl pkt length %zu", count);
return -EINVAL;
}
if (!atomic_read(&gsi->connected)) {
log_event_err("USB cable not connected\n");
return -ECONNRESET;
}
if (gsi->func_is_suspended && !gsi->func_wakeup_allowed) {
c_port->cpkt_drop_cnt++;
log_event_err("drop ctrl pkt of len %zu", count);
return -ENOTSUPP;
}
cpkt = gsi_ctrl_pkt_alloc(count, GFP_KERNEL);
if (IS_ERR(cpkt)) {
log_event_err("failed to allocate ctrl pkt");
return -ENOMEM;
}
ret = copy_from_user(cpkt->buf, buf, count);
if (ret) {
log_event_err("copy_from_user failed err:%d", ret);
gsi_ctrl_pkt_free(cpkt);
return ret;
}
cpkt->type = GSI_CTRL_NOTIFY_RESPONSE_AVAILABLE;
c_port->copied_from_modem++;
if (debugfs.qti_packet_debug)
print_hex_dump(KERN_DEBUG, "WRITE:", DUMP_PREFIX_OFFSET, 16, 1,
cpkt->buf, min_t(int, 30, count), false);
spin_lock_irqsave(&c_port->lock, flags);
list_add_tail(&cpkt->list, &c_port->cpkt_resp_q);
spin_unlock_irqrestore(&c_port->lock, flags);
if (!gsi_ctrl_send_notification(gsi))
c_port->modem_to_host++;
log_event_dbg("Exit %zu", count);
return ret ? ret : count;
}
static long gsi_ctrl_dev_ioctl(struct file *fp, unsigned int cmd,
unsigned long arg)
{
struct gsi_ctrl_port *c_port;
struct f_gsi *gsi;
struct gsi_ctrl_pkt *cpkt;
struct ep_info info = { {0}, {0} };
struct data_buf_info data_info = {0};
enum ipa_usb_teth_prot prot_id =
*(enum ipa_usb_teth_prot *)(fp->private_data);
struct gsi_inst_status *inst_cur = &inst_status[prot_id];
int val, ret = 0;
unsigned long flags;
mutex_lock(&inst_cur->gsi_lock);
if (unlikely(!inst_cur->inst_exist)) {
mutex_unlock(&inst_cur->gsi_lock);
pr_err_ratelimited("%s: free_inst is called and being freed\n",
__func__);
return -ENODEV;
}
mutex_unlock(&inst_cur->gsi_lock);
gsi = inst_cur->opts->gsi;
c_port = &gsi->c_port;
if (!atomic_read(&gsi->connected) && cmd != QTI_CTRL_GET_LINE_STATE
&& cmd != GSI_MBIM_GPS_USB_STATUS) {
log_event_err("%s:cmd %u failed, USB not connected\n",
__func__, cmd);
return -ECONNRESET;
}
switch (cmd) {
case QTI_CTRL_MODEM_OFFLINE:
if (gsi->prot_id == IPA_USB_DIAG) {
log_event_dbg("%s:Modem Offline not handled", __func__);
goto exit_ioctl;
}
atomic_set(&c_port->ctrl_online, 0);
gsi_ctrl_clear_cpkt_queues(gsi, true);
cpkt = gsi_ctrl_pkt_alloc(0, GFP_KERNEL);
if (IS_ERR(cpkt)) {
log_event_err("%s: err allocating cpkt\n", __func__);
return -ENOMEM;
}
cpkt->type = GSI_CTRL_NOTIFY_OFFLINE;
spin_lock_irqsave(&c_port->lock, flags);
list_add_tail(&cpkt->list, &c_port->cpkt_resp_q);
spin_unlock_irqrestore(&c_port->lock, flags);
gsi_ctrl_send_notification(gsi);
break;
case QTI_CTRL_MODEM_ONLINE:
if (gsi->prot_id == IPA_USB_DIAG) {
log_event_dbg("%s:Modem Online not handled", __func__);
goto exit_ioctl;
}
atomic_set(&c_port->ctrl_online, 1);
break;
case QTI_CTRL_GET_LINE_STATE:
case GSI_MBIM_GPS_USB_STATUS:
val = atomic_read(&gsi->connected);
if (gsi->prot_id == IPA_USB_RMNET)
val = gsi->rmnet_dtr_status;
ret = copy_to_user((void __user *)arg, &val, sizeof(val));
if (ret) {
log_event_err("copy_to_user fail LINE_STATE");
ret = -EFAULT;
}
log_event_dbg("%s: Sent line_state: %d for prot id:%d",
__func__, val, gsi->prot_id);
break;
case QTI_CTRL_EP_LOOKUP:
case GSI_MBIM_EP_LOOKUP:
log_event_dbg("%s: EP_LOOKUP for prot id:%d", __func__,
gsi->prot_id);
if (gsi->prot_id == IPA_USB_DIAG &&
(gsi->d_port.in_channel_handle == -EINVAL)) {
ret = -EAGAIN;
break;
}
if (gsi->prot_id != IPA_USB_GPS) {
if (gsi->d_port.in_channel_handle == -EINVAL &&
gsi->d_port.out_channel_handle == -EINVAL) {
ret = -EAGAIN;
break;
}
info.ph_ep_info.ep_type = GSI_MBIM_DATA_EP_TYPE_HSUSB;
info.ph_ep_info.peripheral_iface_id = gsi->data_id;
} else {
info.ph_ep_info.ep_type = GSI_MBIM_DATA_EP_TYPE_HSUSB;
info.ph_ep_info.peripheral_iface_id = gsi->ctrl_id;
}
log_event_dbg("%s: prot id :%d ep_type:%d intf:%d",
__func__, gsi->prot_id, info.ph_ep_info.ep_type,
info.ph_ep_info.peripheral_iface_id);
if (gsi->prot_id != IPA_USB_GPS) {
info.ipa_ep_pair.cons_pipe_num =
(gsi->prot_id == IPA_USB_DIAG) ? -1 :
gsi->d_port.out_channel_handle;
info.ipa_ep_pair.prod_pipe_num =
gsi->d_port.in_channel_handle;
log_event_dbg("%s: ipa_cons_idx:%d ipa_prod_idx:%d",
__func__,
info.ipa_ep_pair.cons_pipe_num,
info.ipa_ep_pair.prod_pipe_num);
}
ret = copy_to_user((void __user *)arg, &info,
sizeof(info));
if (ret) {
log_event_err("copy_to_user fail MBIM");
ret = -EFAULT;
}
break;
case GSI_MBIM_GET_NTB_SIZE:
ret = copy_to_user((void __user *)arg,
&gsi->d_port.ntb_info.ntb_input_size,
sizeof(gsi->d_port.ntb_info.ntb_input_size));
if (ret) {
log_event_err("copy_to_user failNTB_SIZE");
ret = -EFAULT;
}
log_event_dbg("Sent NTB size %d",
gsi->d_port.ntb_info.ntb_input_size);
break;
case GSI_MBIM_GET_DATAGRAM_COUNT:
ret = copy_to_user((void __user *)arg,
&gsi->d_port.ntb_info.ntb_max_datagrams,
sizeof(gsi->d_port.ntb_info.ntb_max_datagrams));
if (ret) {
log_event_err("copy_to_user fail DATAGRAM");
ret = -EFAULT;
}
log_event_dbg("Sent NTB datagrams count %d",
gsi->d_port.ntb_info.ntb_max_datagrams);
break;
case QTI_CTRL_DATA_BUF_INFO:
if (gsi->d_port.out_ep) {
data_info.epout_buf_len =
gsi->d_port.out_request.buf_len;
data_info.epout_total_buf_len =
gsi->d_port.out_request.buf_len *
gsi->d_port.out_request.num_bufs;
log_event_dbg("prot id :%d OUT: buf_len:%u total_len: %u",
gsi->prot_id, data_info.epout_buf_len,
data_info.epout_total_buf_len);
}
if (gsi->d_port.in_ep) {
data_info.epin_buf_len =
gsi->d_port.in_request.buf_len;
data_info.epin_total_buf_len =
gsi->d_port.in_request.buf_len *
gsi->d_port.in_request.num_bufs;
log_event_dbg("prot id :%d IN: buf_len:%u total_len:%u\n",
gsi->prot_id, data_info.epin_buf_len,
data_info.epin_total_buf_len);
}
ret = copy_to_user((void __user *)arg, &data_info,
sizeof(data_info));
if (ret) {
log_event_err("QTI_CTRL_DATA_BUF_INFO: copy_to_user failed");
ret = -EFAULT;
}
break;
default:
log_event_err("wrong parameter");
ret = -EINVAL;
}
exit_ioctl:
return ret;
}
static __poll_t gsi_ctrl_dev_poll(struct file *fp, poll_table *wait)
{
struct gsi_ctrl_port *c_port;
enum ipa_usb_teth_prot prot_id =
*(enum ipa_usb_teth_prot *)(fp->private_data);
struct gsi_inst_status *inst_cur = &inst_status[prot_id];
struct f_gsi *gsi;
unsigned long flags;
__poll_t mask = 0;
mutex_lock(&inst_cur->gsi_lock);
if (unlikely(!inst_cur->inst_exist)) {
mutex_unlock(&inst_cur->gsi_lock);
pr_err_ratelimited("%s: free_inst is called and being freed\n",
__func__);
mask = EPOLLHUP;
goto out;
}
mutex_unlock(&inst_cur->gsi_lock);
gsi = inst_cur->opts->gsi;
c_port = &inst_cur->opts->gsi->c_port;
if (!c_port) {
log_event_err("%s: gsi ctrl port %pK", __func__, c_port);
mask = EPOLLHUP;
goto out;
}
poll_wait(fp, &c_port->read_wq, wait);
spin_lock_irqsave(&c_port->lock, flags);
if (!list_empty(&c_port->cpkt_req_q)) {
mask = EPOLLIN | EPOLLRDNORM;
log_event_dbg("%s sets POLLIN for %s", __func__, c_port->name);
}
spin_unlock_irqrestore(&c_port->lock, flags);
out:
return mask;
}
/* file operations for rmnet/mbim/dpl devices */
static const struct file_operations gsi_ctrl_dev_fops = {
.owner = THIS_MODULE,
.open = gsi_ctrl_dev_open,
.release = gsi_ctrl_dev_release,
.read = gsi_ctrl_dev_read,
.write = gsi_ctrl_dev_write,
.unlocked_ioctl = gsi_ctrl_dev_ioctl,
#ifdef CONFIG_COMPAT
.compat_ioctl = gsi_ctrl_dev_ioctl,
#endif
.poll = gsi_ctrl_dev_poll,
};
/* peak (theoretical) bulk transfer rate in bits-per-second */
static unsigned int gsi_xfer_bitrate(struct usb_gadget *g)
{
if (gadget_is_superspeed(g) && g->speed >= USB_SPEED_SUPER)
return 13 * 1024 * 8 * 1000 * 8;
else if (gadget_is_dualspeed(g) && g->speed == USB_SPEED_HIGH)
return 13 * 512 * 8 * 1000 * 8;
else
return 19 * 64 * 1 * 1000 * 8;
}
static void gsi_uevent_work(struct work_struct *w)
{
struct gsi_ctrl_port *c_port =
container_of(w, struct gsi_ctrl_port, uevent_work);
if (c_port->dev)
kobject_uevent(&c_port->dev->kobj, KOBJ_CHANGE);
}
static int gsi_function_ctrl_port_init(struct f_gsi *gsi)
{
int ret;
int sz = GSI_CTRL_NAME_LEN;
int minor;
struct device *dev;
bool ctrl_dev_create = true;
if (!gsi) {
log_event_err("%s: gsi prot ctx is NULL", __func__);
return -EINVAL;
}
INIT_LIST_HEAD(&gsi->c_port.cpkt_req_q);
INIT_LIST_HEAD(&gsi->c_port.cpkt_resp_q);
spin_lock_init(&gsi->c_port.lock);
init_waitqueue_head(&gsi->c_port.read_wq);
INIT_WORK(&gsi->c_port.uevent_work, gsi_uevent_work);
gsi->c_port.dev = NULL;
gsi->c_port.uevent_wq = NULL;
if (gsi->prot_id == IPA_USB_RMNET)
strlcat(gsi->c_port.name, GSI_RMNET_CTRL_NAME, sz);
else if (gsi->prot_id == IPA_USB_MBIM)
strlcat(gsi->c_port.name, GSI_MBIM_CTRL_NAME, sz);
else if (gsi->prot_id == IPA_USB_DIAG)
strlcat(gsi->c_port.name, GSI_DPL_CTRL_NAME, sz);
else if (gsi->prot_id == IPA_USB_GPS)
strlcat(gsi->c_port.name, GSI_GPS_CTRL_NAME, sz);
else
ctrl_dev_create = false;
if (!ctrl_dev_create)
return 0;
gsi->c_port.uevent_wq = alloc_workqueue(gsi->c_port.name,
WQ_UNBOUND | WQ_MEM_RECLAIM |
WQ_FREEZABLE, 1);
minor = ida_simple_get(&gsi_ida, 0, MAX_CDEV_INSTANCES, GFP_KERNEL);
if (minor < 0) {
pr_err("%s: No more minor numbers left! rc:%d\n", __func__,
minor);
return minor;
}
cdev_init(&gsi->c_port.cdev, &gsi_ctrl_dev_fops);
ret = cdev_add(&gsi->c_port.cdev, MKDEV(major, minor), 1);
if (ret) {
log_event_err("%s: Failed to add cdev for (%s)\n", __func__,
gsi->c_port.name);
goto err_cdev_add;
}
dev = device_create(gsi_class, NULL, MKDEV(major, minor), &gsi->c_port,
gsi->c_port.name);
if (IS_ERR(dev)) {
log_event_err("%s: device_create failed for (%s)\n", __func__,
gsi->c_port.name);
ret = PTR_ERR(dev);
goto err_create_dev;
}
gsi->c_port.dev = dev;
return 0;
err_create_dev:
cdev_del(&gsi->c_port.cdev);
err_cdev_add:
ida_simple_remove(&gsi_ida, minor);
return ret;
}
static struct net_device *gsi_rndis_get_netdev(const char *netname)
{
struct net_device *net_dev;
net_dev = dev_get_by_name(&init_net, netname);
if (!net_dev)
return ERR_PTR(-EINVAL);
/*
* Decrement net_dev refcount as it was incremented in
* dev_get_by_name().
*/
dev_put(net_dev);
return net_dev;
}
static void gsi_rndis_open(struct f_gsi *gsi)
{
struct usb_composite_dev *cdev = gsi->function.config->cdev;
log_event_dbg("%s", __func__);
rndis_set_param_medium(gsi->params, RNDIS_MEDIUM_802_3,
gsi_xfer_bitrate(cdev->gadget) / 100);
rndis_signal_connect(gsi->params);
}
static void gsi_rndis_ipa_reset_trigger(struct gsi_data_port *d_port)
{
unsigned long flags;
struct f_gsi *gsi = d_port_to_gsi(d_port);
log_event_dbg("%s: setting net_ready_trigger\n", __func__);
spin_lock_irqsave(&d_port->lock, flags);
d_port->net_ready_trigger = false;
spin_unlock_irqrestore(&d_port->lock, flags);
}
static void gsi_rndis_flow_ctrl_enable(bool enable, struct rndis_params *param)
{
struct f_gsi *gsi = param->v;
struct gsi_data_port *d_port;
if (!gsi) {
pr_err("%s: gsi prot ctx is %pK\n", __func__, gsi);
return;
}
d_port = &gsi->d_port;
if (enable) {
log_event_dbg("%s: posting HOST_NRDY\n", __func__);
post_event(d_port, EVT_HOST_NRDY);
} else {
log_event_dbg("%s: posting HOST_READY\n", __func__);
post_event(d_port, EVT_HOST_READY);
/*
* If host supports flow control with RNDIS_MSG_INIT then
* mark the flag to true. This flag will be used further to
* enable the flow control on resume path.
*/
gsi->host_supports_flow_control = true;
}
queue_delayed_work(gsi->d_port.ipa_usb_wq, &gsi->d_port.usb_ipa_w, 0);
}
static int queue_notification_request(struct f_gsi *gsi)
{
int ret;
unsigned long flags;
struct usb_function *func = &gsi->function;
struct usb_request *req = gsi->c_port.notify_req;
struct usb_ep *ep = gsi->c_port.notify;
struct usb_gadget *gadget = func->config->cdev->gadget;
if (gsi->c_port.is_suspended) {
/*For remote wakeup, queue the req from gsi_resume*/
spin_lock_irqsave(&gsi->c_port.lock, flags);
gsi->c_port.notify_req_queued = false;
spin_unlock_irqrestore(&gsi->c_port.lock, flags);
if (gsi->rwake_inprogress) {
log_event_dbg("%s remote-wakeup in progress\n",
__func__);
return -EBUSY;
}
if (!usb_gsi_remote_wakeup_allowed(func)) {
log_event_dbg("%s remote-wakeup not capable\n",
__func__);
return -EOPNOTSUPP;
}
log_event_dbg("%s wakeup host\n", __func__);
if (gadget->speed >= USB_SPEED_SUPER
&& gsi->func_is_suspended) {
ret = usb_func_wakeup(func);
if (ret == -EAGAIN)
gsi->func_wakeup_pending = true;
} else
ret = usb_gadget_wakeup(gadget);
gsi->rwake_inprogress = true;
return ret;
}
ret = usb_ep_queue(ep, req, GFP_ATOMIC);
if (ret < 0) {
spin_lock_irqsave(&gsi->c_port.lock, flags);
gsi->c_port.notify_req_queued = false;
spin_unlock_irqrestore(&gsi->c_port.lock, flags);
}
log_event_dbg("%s: ret:%d req_queued:%d",
__func__, ret, gsi->c_port.notify_req_queued);
return ret;
}
static int gsi_ctrl_send_notification(struct f_gsi *gsi)
{
__le32 *data;
struct usb_cdc_notification *event;
struct usb_request *req = gsi->c_port.notify_req;
struct usb_composite_dev *cdev;
struct gsi_ctrl_pkt *cpkt;
unsigned long flags;
bool del_free_cpkt = false;
if (!atomic_read(&gsi->connected)) {
log_event_dbg("%s: cable disconnect", __func__);
return -ENODEV;
}
spin_lock_irqsave(&gsi->c_port.lock, flags);
if (list_empty(&gsi->c_port.cpkt_resp_q)) {
spin_unlock_irqrestore(&gsi->c_port.lock, flags);
log_event_dbg("%s: cpkt_resp_q is empty\n", __func__);
return 0;
}
log_event_dbg("%s: notify_req_queued:%d\n",
__func__, gsi->c_port.notify_req_queued);
if (gsi->c_port.notify_req_queued) {
spin_unlock_irqrestore(&gsi->c_port.lock, flags);
log_event_dbg("%s: notify_req is already queued.\n", __func__);
return 0;
}
cpkt = list_first_entry(&gsi->c_port.cpkt_resp_q,
struct gsi_ctrl_pkt, list);
log_event_dbg("%s: cpkt->type:%d\n", __func__, cpkt->type);
event = req->buf;
cdev = gsi->function.config->cdev;
switch (cpkt->type) {
case GSI_CTRL_NOTIFY_CONNECT:
del_free_cpkt = true;
event->bNotificationType = USB_CDC_NOTIFY_NETWORK_CONNECTION;
event->wValue = cpu_to_le16(1);
event->wLength = cpu_to_le16(0);
break;
case GSI_CTRL_NOTIFY_SPEED:
del_free_cpkt = true;
event->bNotificationType = USB_CDC_NOTIFY_SPEED_CHANGE;
event->wValue = cpu_to_le16(0);
event->wLength = cpu_to_le16(8);
/* SPEED_CHANGE data is up/down speeds in bits/sec */
data = req->buf + sizeof(*event);
data[0] = cpu_to_le32(gsi_xfer_bitrate(cdev->gadget));
data[1] = data[0];
log_event_dbg("notify speed %d",
gsi_xfer_bitrate(cdev->gadget));
break;
case GSI_CTRL_NOTIFY_OFFLINE:
del_free_cpkt = true;
event->bNotificationType = USB_CDC_NOTIFY_NETWORK_CONNECTION;
event->wValue = cpu_to_le16(0);
event->wLength = cpu_to_le16(0);
break;
case GSI_CTRL_NOTIFY_RESPONSE_AVAILABLE:
event->bNotificationType = USB_CDC_NOTIFY_RESPONSE_AVAILABLE;
event->wValue = cpu_to_le16(0);
event->wLength = cpu_to_le16(0);
if (gsi->prot_id == IPA_USB_RNDIS) {
data = req->buf;
data[0] = cpu_to_le32(1);
data[1] = cpu_to_le32(0);
/*
* we need to free dummy packet for RNDIS as sending
* notification about response available multiple time,
* RNDIS host driver doesn't like. All SEND/GET
* ENCAPSULATED response is one-to-one for RNDIS case
* and host expects to have below sequence:
* ep0: USB_CDC_SEND_ENCAPSULATED_COMMAND
* int_ep: device->host: RESPONSE_AVAILABLE
* ep0: USB_GET_SEND_ENCAPSULATED_COMMAND
* For RMNET case: host ignores multiple notification.
*/
del_free_cpkt = true;
}
break;
default:
spin_unlock_irqrestore(&gsi->c_port.lock, flags);
log_event_err("%s:unknown notify state", __func__);
WARN_ON(1);
return -EINVAL;
}
/*
* Delete and free cpkt related to non NOTIFY_RESPONSE_AVAILABLE
* notification whereas NOTIFY_RESPONSE_AVAILABLE related cpkt is
* deleted from USB_CDC_GET_ENCAPSULATED_RESPONSE setup request
*/
if (del_free_cpkt) {
list_del(&cpkt->list);
gsi_ctrl_pkt_free(cpkt);
}
gsi->c_port.notify_req_queued = true;
spin_unlock_irqrestore(&gsi->c_port.lock, flags);
log_event_dbg("send Notify type %02x", event->bNotificationType);
return queue_notification_request(gsi);
}
static void gsi_ctrl_notify_resp_complete(struct usb_ep *ep,
struct usb_request *req)
{
struct f_gsi *gsi = req->context;
struct usb_cdc_notification *event = req->buf;
int status = req->status;
unsigned long flags;
spin_lock_irqsave(&gsi->c_port.lock, flags);
gsi->c_port.notify_req_queued = false;
spin_unlock_irqrestore(&gsi->c_port.lock, flags);
switch (status) {
case -ECONNRESET:
case -ESHUTDOWN:
/* connection gone */
log_event_dbg("ESHUTDOWN/ECONNRESET, connection gone");
gsi_ctrl_clear_cpkt_queues(gsi, false);
gsi_ctrl_send_cpkt_tomodem(gsi, NULL, 0);
break;
default:
log_event_err("Unknown event %02x --> %d",
event->bNotificationType, req->status);
/* FALLTHROUGH */
case 0:
break;
}
}
static void gsi_rndis_response_available(void *_rndis)
{
struct f_gsi *gsi = _rndis;
struct gsi_ctrl_pkt *cpkt;
unsigned long flags;
cpkt = gsi_ctrl_pkt_alloc(0, GFP_ATOMIC);
if (IS_ERR(cpkt)) {
log_event_err("%s: err allocating cpkt\n", __func__);
return;
}
cpkt->type = GSI_CTRL_NOTIFY_RESPONSE_AVAILABLE;
spin_lock_irqsave(&gsi->c_port.lock, flags);
list_add_tail(&cpkt->list, &gsi->c_port.cpkt_resp_q);
spin_unlock_irqrestore(&gsi->c_port.lock, flags);
gsi_ctrl_send_notification(gsi);
}
static void gsi_rndis_command_complete(struct usb_ep *ep,
struct usb_request *req)
{
struct f_gsi *gsi = req->context;
struct usb_composite_dev *cdev = gsi->function.config->cdev;
int status;
u32 MsgType;
if (!req->buf || !gsi->params)
return;
MsgType = get_unaligned_le32((__le32 *)req->buf);
/* intercept halt message to perform flow control */
if (MsgType == RNDIS_MSG_HALT) {
log_event_dbg("RNDIS_MSG_HALT, state:%d\n",
gsi->params->state);
if (gsi->params->state == RNDIS_DATA_INITIALIZED)
gsi_rndis_flow_ctrl_enable(true, gsi->params);
gsi->params->state = RNDIS_UNINITIALIZED;
return;
}
status = rndis_msg_parser(gsi->params, (u8 *) req->buf);
if (status < 0)
log_event_err("RNDIS command error %d, %d/%d",
status, req->actual, req->length);
if (MsgType == RNDIS_MSG_INIT) {
rndis_init_msg_type *buf = (rndis_init_msg_type *)req->buf;
log_event_dbg("RNDIS host major:%d minor:%d version\n",
le32_to_cpu(buf->MajorVersion),
le32_to_cpu(buf->MinorVersion));
/* honor host dl aggr size */
gsi->d_port.in_aggr_size = le32_to_cpu(buf->MaxTransferSize);
log_event_dbg("RNDIS host DL MaxTransferSize:%d\n",
le32_to_cpu(buf->MaxTransferSize));
} else if (MsgType == RNDIS_MSG_SET) {
rndis_set_msg_type *buf = (rndis_set_msg_type *)req->buf;
if (le32_to_cpu(buf->OID) ==
RNDIS_OID_GEN_CURRENT_PACKET_FILTER)
gsi_rndis_flow_ctrl_enable(!(*gsi->params->filter),
gsi->params);
}
cdev->setup_pending = false;
}
static void
gsi_ctrl_set_ntb_cmd_complete(struct usb_ep *ep, struct usb_request *req)
{
/* now for SET_NTB_INPUT_SIZE only */
unsigned int in_size = 0;
struct f_gsi *gsi = req->context;
struct gsi_ntb_info *ntb = NULL;
log_event_dbg("dev:%pK", gsi);
req->context = NULL;
if (req->status || req->actual != req->length) {
log_event_err("Bad control-OUT transfer");
goto invalid;
}
if (req->length == 4) {
in_size = get_unaligned_le32(req->buf);
if (in_size < USB_CDC_NCM_NTB_MIN_IN_SIZE ||
in_size > le32_to_cpu(mbim_gsi_ntb_parameters.dwNtbInMaxSize))
goto invalid;
} else if (req->length == 8) {
ntb = (struct gsi_ntb_info *)req->buf;
in_size = get_unaligned_le32(&(ntb->ntb_input_size));
if (in_size < USB_CDC_NCM_NTB_MIN_IN_SIZE ||
in_size > le32_to_cpu(mbim_gsi_ntb_parameters.dwNtbInMaxSize))
goto invalid;
gsi->d_port.ntb_info.ntb_max_datagrams =
get_unaligned_le16(&(ntb->ntb_max_datagrams));
} else {
goto invalid;
}
log_event_dbg("Set NTB INPUT SIZE %d", in_size);
gsi->d_port.ntb_info.ntb_input_size = in_size;
return;
invalid:
log_event_err("Illegal NTB INPUT SIZE %d from host", in_size);
usb_ep_set_halt(ep);
}
static void gsi_ctrl_cmd_complete(struct usb_ep *ep, struct usb_request *req)
{
struct f_gsi *gsi = req->context;
struct usb_composite_dev *cdev = gsi->function.config->cdev;
gsi_ctrl_send_cpkt_tomodem(gsi, req->buf, req->actual);
cdev->setup_pending = false;
}
static void gsi_ctrl_reset_cmd_complete(struct usb_ep *ep,
struct usb_request *req)
{
struct f_gsi *gsi = req->context;
gsi_ctrl_send_cpkt_tomodem(gsi, req->buf, 0);
}
static void gsi_ctrl_send_response_complete(struct usb_ep *ep,
struct usb_request *req)
{
struct f_gsi *gsi = req->context;
gsi_ctrl_send_notification(gsi);
}
static int
gsi_setup(struct usb_function *f, const struct usb_ctrlrequest *ctrl)
{
struct f_gsi *gsi = func_to_gsi(f);
struct usb_composite_dev *cdev = f->config->cdev;
struct usb_request *req = cdev->req;
int id, value = -EOPNOTSUPP;
u16 w_index = le16_to_cpu(ctrl->wIndex);
u16 w_value = le16_to_cpu(ctrl->wValue);
u16 w_length = le16_to_cpu(ctrl->wLength);
struct gsi_ctrl_pkt *cpkt;
u8 *buf;
u32 n;
bool line_state;
if (!atomic_read(&gsi->connected)) {
log_event_dbg("usb cable is not connected");
return -ENOTCONN;
}
/* rmnet and dpl does not have ctrl_id */
if (gsi->ctrl_id == -ENODEV)
id = gsi->data_id;
else
id = gsi->ctrl_id;
/* composite driver infrastructure handles everything except
* CDC class messages; interface activation uses set_alt().
*/
switch ((ctrl->bRequestType << 8) | ctrl->bRequest) {
case ((USB_DIR_OUT | USB_TYPE_CLASS | USB_RECIP_INTERFACE) << 8)
| USB_CDC_RESET_FUNCTION:
log_event_dbg("USB_CDC_RESET_FUNCTION");
value = 0;
req->complete = gsi_ctrl_reset_cmd_complete;
req->context = gsi;
break;
case ((USB_DIR_OUT | USB_TYPE_CLASS | USB_RECIP_INTERFACE) << 8)
| USB_CDC_SEND_ENCAPSULATED_COMMAND:
log_event_dbg("USB_CDC_SEND_ENCAPSULATED_COMMAND");
if (w_value || w_index != id)
goto invalid;
/* read the request; process it later */
value = w_length;
req->context = gsi;
if (gsi->prot_id == IPA_USB_RNDIS)
req->complete = gsi_rndis_command_complete;
else
req->complete = gsi_ctrl_cmd_complete;
/* later, rndis_response_available() sends a notification */
break;
case ((USB_DIR_IN | USB_TYPE_CLASS | USB_RECIP_INTERFACE) << 8)
| USB_CDC_GET_ENCAPSULATED_RESPONSE:
log_event_dbg("USB_CDC_GET_ENCAPSULATED_RESPONSE");
if (w_value || w_index != id)
goto invalid;
if (gsi->prot_id == IPA_USB_RNDIS) {
rndis_init_cmplt_type *res;
/* return the result */
buf = rndis_get_next_response(gsi->params, &n);
if (!buf)
break;
res = (rndis_init_cmplt_type *)buf;
if (le32_to_cpu(res->MessageType) == RNDIS_MSG_INIT_C) {
log_event_dbg("%s: max_pkt_per_xfer : %d",
__func__, DEFAULT_MAX_PKT_PER_XFER);
res->MaxPacketsPerTransfer =
cpu_to_le32(DEFAULT_MAX_PKT_PER_XFER);
res->MaxTransferSize = cpu_to_le32(
le32_to_cpu(res->MaxTransferSize)
* DEFAULT_MAX_PKT_PER_XFER);
/* In case of aggregated packets QC device
* will request aliment to 4 (2^2).
*/
log_event_dbg("%s: pkt_alignment_factor : %d",
__func__, DEFAULT_PKT_ALIGNMENT_FACTOR);
res->PacketAlignmentFactor =
cpu_to_le32(
DEFAULT_PKT_ALIGNMENT_FACTOR);
}
memcpy(req->buf, buf, n);
rndis_free_response(gsi->params, buf);
value = n;
break;
}
spin_lock(&gsi->c_port.lock);
if (list_empty(&gsi->c_port.cpkt_resp_q)) {
log_event_dbg("ctrl resp queue empty");
spin_unlock(&gsi->c_port.lock);
break;
}
cpkt = list_first_entry(&gsi->c_port.cpkt_resp_q,
struct gsi_ctrl_pkt, list);
list_del(&cpkt->list);
gsi->c_port.get_encap_cnt++;
spin_unlock(&gsi->c_port.lock);
value = min_t(unsigned int, w_length, cpkt->len);
memcpy(req->buf, cpkt->buf, value);
gsi_ctrl_pkt_free(cpkt);
req->complete = gsi_ctrl_send_response_complete;
req->context = gsi;
log_event_dbg("copied encap_resp %d bytes",
value);
break;
case ((USB_DIR_OUT | USB_TYPE_CLASS | USB_RECIP_INTERFACE) << 8)
| USB_CDC_REQ_SET_CONTROL_LINE_STATE:
line_state = (w_value & GSI_CTRL_DTR ? true : false);
if (gsi->prot_id == IPA_USB_RMNET)
gsi->rmnet_dtr_status = line_state;
log_event_dbg("%s: USB_CDC_REQ_SET_CONTROL_LINE_STATE DTR:%d\n",
__func__, line_state);
if (gsi->c_port.uevent_wq)
queue_work(gsi->c_port.uevent_wq,
&gsi->c_port.uevent_work);
gsi_ctrl_send_cpkt_tomodem(gsi, NULL, 0);
value = 0;
break;
case ((USB_DIR_OUT | USB_TYPE_CLASS | USB_RECIP_INTERFACE) << 8)
| USB_CDC_SET_ETHERNET_PACKET_FILTER:
/* see 6.2.30: no data, wIndex = interface,
* wValue = packet filter bitmap
*/
if (w_length != 0 || w_index != id)
goto invalid;
log_event_dbg("packet filter %02x", w_value);
/* REVISIT locking of cdc_filter. This assumes the UDC
* driver won't have a concurrent packet TX irq running on
* another CPU; or that if it does, this write is atomic...
*/
gsi->d_port.cdc_filter = w_value;
value = 0;
break;
case ((USB_DIR_IN | USB_TYPE_CLASS | USB_RECIP_INTERFACE) << 8)
| USB_CDC_GET_NTB_PARAMETERS:
log_event_dbg("USB_CDC_GET_NTB_PARAMETERS");
if (w_length == 0 || w_value != 0 || w_index != id)
break;
value = w_length > sizeof(mbim_gsi_ntb_parameters) ?
sizeof(mbim_gsi_ntb_parameters) : w_length;
memcpy(req->buf, &mbim_gsi_ntb_parameters, value);
break;
case ((USB_DIR_IN | USB_TYPE_CLASS | USB_RECIP_INTERFACE) << 8)
| USB_CDC_GET_NTB_INPUT_SIZE:
log_event_dbg("USB_CDC_GET_NTB_INPUT_SIZE");
if (w_length < 4 || w_value != 0 || w_index != id)
break;
put_unaligned_le32(gsi->d_port.ntb_info.ntb_input_size,
req->buf);
value = 4;
log_event_dbg("Reply to host INPUT SIZE %d",
gsi->d_port.ntb_info.ntb_input_size);
break;
case ((USB_DIR_OUT | USB_TYPE_CLASS | USB_RECIP_INTERFACE) << 8)
| USB_CDC_SET_NTB_INPUT_SIZE:
log_event_dbg("USB_CDC_SET_NTB_INPUT_SIZE");
if (w_length != 4 && w_length != 8) {
log_event_err("wrong NTB length %d", w_length);
break;
}
if (w_value != 0 || w_index != id)
break;
req->complete = gsi_ctrl_set_ntb_cmd_complete;
req->length = w_length;
req->context = gsi;
value = req->length;
break;
default:
invalid:
log_event_err("inval ctrl req%02x.%02x v%04x i%04x l%d",
ctrl->bRequestType, ctrl->bRequest,
w_value, w_index, w_length);
}
/* respond with data transfer or status phase? */
if (value >= 0) {
log_event_dbg("req%02x.%02x v%04x i%04x l%d",
ctrl->bRequestType, ctrl->bRequest,
w_value, w_index, w_length);
req->zero = (value < w_length);
req->length = value;
value = usb_ep_queue(cdev->gadget->ep0, req, GFP_ATOMIC);
if (value < 0)
log_event_err("response on err %d", value);
else
cdev->setup_pending = true;
}
/* device either stalls (value < 0) or reports success */
return value;
}
/*
* Because the data interface supports multiple altsettings,
* function *MUST* implement a get_alt() method.
*/
static int gsi_get_alt(struct usb_function *f, unsigned int intf)
{
struct f_gsi *gsi = func_to_gsi(f);
/* RNDIS, RMNET and DPL only support alt 0*/
if (intf == gsi->ctrl_id || gsi->prot_id == IPA_USB_RNDIS ||
gsi->prot_id == IPA_USB_RMNET ||
gsi->prot_id == IPA_USB_DIAG)
return 0;
else if (intf == gsi->data_id)
return gsi->data_interface_up;
return -EINVAL;
}
static int gsi_set_alt(struct usb_function *f, unsigned int intf,
unsigned int alt)
{
struct f_gsi *gsi = func_to_gsi(f);
struct usb_composite_dev *cdev = f->config->cdev;
struct net_device *net;
int ret = 0;
log_event_dbg("intf=%u, alt=%u", intf, alt);
/* Control interface has only altsetting 0 */
if (intf == gsi->ctrl_id || gsi->prot_id == IPA_USB_RMNET) {
if (alt != 0)
goto fail;
if (!gsi->c_port.notify)
goto fail;
if (gsi->c_port.notify->driver_data) {
log_event_dbg("reset gsi control %d", intf);
usb_ep_disable(gsi->c_port.notify);
}
ret = config_ep_by_speed(cdev->gadget, f,
gsi->c_port.notify);
if (ret) {
gsi->c_port.notify->desc = NULL;
log_event_err("Config-fail notify ep %s: err %d",
gsi->c_port.notify->name, ret);
goto fail;
}
ret = usb_ep_enable(gsi->c_port.notify);
if (ret) {
log_event_err("usb ep#%s enable failed, err#%d",
gsi->c_port.notify->name, ret);
goto fail;
}
gsi->c_port.notify->driver_data = gsi;
}
/* Data interface has two altsettings, 0 and 1 */
if (intf == gsi->data_id) {
gsi->d_port.net_ready_trigger = false;
/* for rndis and rmnet alt is always 0 update alt accordingly */
if (gsi->prot_id == IPA_USB_RNDIS ||
gsi->prot_id == IPA_USB_RMNET ||
gsi->prot_id == IPA_USB_DIAG) {
if (gsi->d_port.in_ep &&
!gsi->d_port.in_ep->driver_data)
alt = 1;
else
alt = 0;
}
if (alt > 1)
goto notify_ep_disable;
if (gsi->data_interface_up == alt)
return 0;
if (gsi->d_port.in_ep && gsi->d_port.in_ep->driver_data)
gsi->d_port.ntb_info.ntb_input_size =
MBIM_NTB_DEFAULT_IN_SIZE;
if (alt == 1) {
if (gsi->d_port.in_ep) {
if (gsi->prot_id == IPA_USB_DIAG)
gsi->d_port.in_request.ep_intr_num = 3;
else
gsi->d_port.in_request.ep_intr_num = 2;
}
if (gsi->d_port.out_ep)
gsi->d_port.out_request.ep_intr_num = 1;
gsi->d_port.gadget = cdev->gadget;
gsi->d_port.cdev = cdev;
if (gsi->prot_id == IPA_USB_RNDIS) {
gsi_rndis_open(gsi);
net = gsi_rndis_get_netdev("rndis0");
if (IS_ERR(net))
goto notify_ep_disable;
log_event_dbg("RNDIS RX/TX early activation");
gsi->d_port.cdc_filter = 0;
rndis_set_param_dev(gsi->params, net,
&gsi->d_port.cdc_filter);
}
if (gsi->prot_id == IPA_USB_ECM)
gsi->d_port.cdc_filter = DEFAULT_FILTER;
post_event(&gsi->d_port, EVT_SET_ALT);
/*
* delay until delayed status is returned to
* composite layer.
*/
queue_delayed_work(gsi->d_port.ipa_usb_wq,
&gsi->d_port.usb_ipa_w,
msecs_to_jiffies(1));
ret = USB_GADGET_DELAYED_STATUS;
}
if (alt == 0 && ((gsi->d_port.in_ep &&
!gsi->d_port.in_ep->driver_data) ||
(gsi->d_port.out_ep &&
!gsi->d_port.out_ep->driver_data))) {
post_event(&gsi->d_port, EVT_DISCONNECTED);
queue_delayed_work(gsi->d_port.ipa_usb_wq,
&gsi->d_port.usb_ipa_w, 0);
log_event_dbg("%s: Disconnecting\n", __func__);
}
gsi->data_interface_up = alt;
log_event_dbg("DATA_INTERFACE id = %d, status = %d",
gsi->data_id, gsi->data_interface_up);
}
gsi->c_port.is_suspended = false;
gsi->func_wakeup_pending = false;
atomic_set(&gsi->connected, 1);
/* send 0 len pkt to qti to notify state change */
if (gsi->prot_id == IPA_USB_DIAG ||
gsi->prot_id == IPA_USB_GPS ||
gsi->prot_id == IPA_USB_MBIM)
gsi_ctrl_send_cpkt_tomodem(gsi, NULL, 0);
if (gsi->c_port.uevent_wq)
queue_work(gsi->c_port.uevent_wq, &gsi->c_port.uevent_work);
return ret;
notify_ep_disable:
if (gsi->c_port.notify && gsi->c_port.notify->driver_data)
usb_ep_disable(gsi->c_port.notify);
fail:
return -EINVAL;
}
static void gsi_disable(struct usb_function *f)
{
struct f_gsi *gsi = func_to_gsi(f);
atomic_set(&gsi->connected, 0);
del_timer(&gsi->gsi_rw_timer);
gsi->debugfs_rw_timer_enable = 0;
if (gsi->prot_id == IPA_USB_RNDIS)
rndis_uninit(gsi->params);
if (gsi->prot_id == IPA_USB_RMNET)
gsi->rmnet_dtr_status = false;
/* Disable Control Path */
if (gsi->c_port.notify &&
gsi->c_port.notify->driver_data) {
usb_ep_disable(gsi->c_port.notify);
gsi->c_port.notify->driver_data = NULL;
}
gsi_ctrl_clear_cpkt_queues(gsi, false);
if (gsi->c_port.uevent_wq)
queue_work(gsi->c_port.uevent_wq, &gsi->c_port.uevent_work);
/* send 0 len pkt to qti/qbi/gps to notify state change */
gsi_ctrl_send_cpkt_tomodem(gsi, NULL, 0);
gsi->c_port.notify_req_queued = false;
gsi->func_is_suspended = false;
gsi->func_wakeup_allowed = false;
gsi->func_wakeup_pending = false;
/* Disable Data Path - only if it was initialized already (alt=1) */
if (!gsi->data_interface_up) {
log_event_dbg("%s: data intf is closed", __func__);
return;
}
gsi->data_interface_up = false;
gsi->host_supports_flow_control = false;
log_event_dbg("%s deactivated", gsi->function.name);
gsi->d_port.net_ready_trigger = false;
post_event(&gsi->d_port, EVT_DISCONNECTED);
queue_delayed_work(gsi->d_port.ipa_usb_wq, &gsi->d_port.usb_ipa_w, 0);
}
static void gsi_suspend(struct usb_function *f)
{
bool block_db;
struct f_gsi *gsi = func_to_gsi(f);
/* Check if function is already suspended in gsi_func_suspend()
* Or func_suspend would have bailed out earlier if func_remote_wakeup
* wasn't enabled.
*/
if (gsi->func_is_suspended && (gsi->d_port.sm_state == STATE_SUSPENDED ||
gsi->d_port.sm_state == STATE_SUSPEND_IN_PROGRESS)) {
log_event_dbg("%s: func already suspended, return\n", __func__);
return;
}
gsi->c_port.is_suspended = true;
/*
* GPS doesn't use any data interface, hence bail out as there is no
* GSI specific handling needed.
*/
if (gsi->prot_id == IPA_USB_GPS) {
log_event_dbg("%s: suspend done\n", __func__);
return;
}
block_db = true;
usb_gsi_ep_op(gsi->d_port.in_ep, (void *)&block_db,
GSI_EP_OP_SET_CLR_BLOCK_DBL);
post_event(&gsi->d_port, EVT_SUSPEND);
queue_delayed_work(gsi->d_port.ipa_usb_wq, &gsi->d_port.usb_ipa_w, 0);
log_event_dbg("gsi suspended");
/*
* If host suspended bus without receiving notification request then
* initiate remote-wakeup. As driver won't be able to do it later since
* notification request is already queued.
*/
if (gsi->c_port.notify_req_queued && usb_gsi_remote_wakeup_allowed(f)) {
mod_timer(&gsi->gsi_rw_timer, jiffies + msecs_to_jiffies(2000));
log_event_dbg("%s: pending response, arm rw_timer\n", __func__);
}
}
static void gsi_resume(struct usb_function *f)
{
struct f_gsi *gsi = func_to_gsi(f);
struct usb_composite_dev *cdev = f->config->cdev;
log_event_dbg("%s", __func__);
/*
* If the function is in USB3 Function Suspend state, resume is
* canceled. In this case resume is done by a Function Resume request.
*/
if ((cdev->gadget->speed >= USB_SPEED_SUPER) &&
gsi->func_is_suspended) {
if (gsi->func_wakeup_pending) {
usb_func_wakeup(&gsi->function);
gsi->func_wakeup_pending = false;
}
return;
}
/* Keep timer enabled if user enabled using debugfs */
if (!gsi->debugfs_rw_timer_enable)
del_timer(&gsi->gsi_rw_timer);
if (gsi->c_port.notify && !gsi->c_port.notify->desc)
config_ep_by_speed(cdev->gadget, f, gsi->c_port.notify);
gsi->c_port.is_suspended = false;
/* Check any pending cpkt, and queue immediately on resume */
gsi_ctrl_send_notification(gsi);
gsi->rwake_inprogress = false;
if (gsi->prot_id == IPA_USB_GPS) {
log_event_dbg("%s: resume done\n", __func__);
return;
}
/*
* Linux host does not send RNDIS_MSG_INIT or non-zero
* RNDIS_MESSAGE_PACKET_FILTER after performing bus resume.
* Check whether host supports flow_control are not. If yes
* Trigger state machine explicitly on resume.
*/
if (gsi->prot_id == IPA_USB_RNDIS &&
!usb_gsi_remote_wakeup_allowed(f) &&
gsi->host_supports_flow_control && gsi->params) {
if (gsi->params->state != RNDIS_DATA_INITIALIZED)
gsi_rndis_flow_ctrl_enable(false, gsi->params);
gsi->params->state = RNDIS_DATA_INITIALIZED;
}
post_event(&gsi->d_port, EVT_RESUMED);
queue_delayed_work(gsi->d_port.ipa_usb_wq, &gsi->d_port.usb_ipa_w, 0);
log_event_dbg("%s: completed", __func__);
}
static int gsi_get_status(struct usb_function *f)
{
#ifdef CONFIG_USB_FUNC_WAKEUP_SUPPORTED
struct f_gsi *gsi = func_to_gsi(f);
/* Disable function remote wake-up for DPL interface */
if (gsi->prot_id == IPA_USB_DIAG)
return 0;
return (gsi->func_wakeup_allowed ? USB_INTRF_STAT_FUNC_RW : 0) |
USB_INTRF_STAT_FUNC_RW_CAP;
#else
return 0;
#endif
}
static int gsi_func_suspend(struct usb_function *f, u8 options)
{
bool func_wakeup_allowed;
struct f_gsi *gsi = func_to_gsi(f);
log_event_dbg("func susp %u cmd for %s",
options, f->name ? f->name : "");
func_wakeup_allowed = !!(options & (USB_INTRF_FUNC_SUSPEND_RW >> 8));
if (options & (USB_INTRF_FUNC_SUSPEND_LP >> 8)) {
gsi->func_wakeup_allowed = func_wakeup_allowed;
if (!gsi->func_is_suspended) {
gsi_suspend(f);
gsi->func_is_suspended = true;
}
} else {
if (gsi->func_is_suspended) {
gsi->func_is_suspended = false;
gsi_resume(f);
}
gsi->func_wakeup_allowed = func_wakeup_allowed;
}
return 0;
}
static int gsi_update_function_bind_params(struct f_gsi *gsi,
struct usb_composite_dev *cdev,
struct gsi_function_bind_info *info)
{
struct usb_ep *ep;
struct usb_cdc_notification *event;
struct usb_function *f = &gsi->function;
int status;
if (info->ctrl_str_idx >= 0 && info->ctrl_desc) {
/* ctrl interface label */
status = usb_string_id(cdev);
if (status < 0)
return status;
info->string_defs[info->ctrl_str_idx].id = status;
info->ctrl_desc->iInterface = status;
}
if (info->data_str_idx >= 0 && info->data_desc) {
/* data interface label */
status = usb_string_id(cdev);
if (status < 0)
return status;
info->string_defs[info->data_str_idx].id = status;
info->data_desc->iInterface = status;
}
if (info->iad_str_idx >= 0 && info->iad_desc) {
/* IAD iFunction label */
status = usb_string_id(cdev);
if (status < 0)
return status;
info->string_defs[info->iad_str_idx].id = status;
info->iad_desc->iFunction = status;
}
if (info->mac_str_idx >= 0 && info->cdc_eth_desc) {
/* IAD iFunction label */
status = usb_string_id(cdev);
if (status < 0)
return status;
info->string_defs[info->mac_str_idx].id = status;
info->cdc_eth_desc->iMACAddress = status;
}
if (info->ctrl_desc)
info->ctrl_desc->bInterfaceNumber = gsi->ctrl_id;
if (info->iad_desc)
info->iad_desc->bFirstInterface = gsi->ctrl_id;
if (info->union_desc) {
info->union_desc->bMasterInterface0 = gsi->ctrl_id;
info->union_desc->bSlaveInterface0 = gsi->data_id;
}
if (info->data_desc)
info->data_desc->bInterfaceNumber = gsi->data_id;
if (info->data_nop_desc)
info->data_nop_desc->bInterfaceNumber = gsi->data_id;
/* allocate instance-specific endpoints */
if (info->fs_in_desc) {
ep = usb_ep_autoconfig_by_name(cdev->gadget,
info->fs_in_desc, info->in_epname);
if (!ep)
goto fail;
gsi->d_port.in_ep = ep;
msm_ep_config(gsi->d_port.in_ep, NULL, 0);
ep->driver_data = cdev; /* claim */
}
if (info->fs_out_desc) {
ep = usb_ep_autoconfig_by_name(cdev->gadget,
info->fs_out_desc, info->out_epname);
if (!ep)
goto fail;
gsi->d_port.out_ep = ep;
msm_ep_config(gsi->d_port.out_ep, NULL, 0);
ep->driver_data = cdev; /* claim */
}
if (info->fs_notify_desc) {
ep = usb_ep_autoconfig(cdev->gadget, info->fs_notify_desc);
if (!ep)
goto fail;
gsi->c_port.notify = ep;
ep->driver_data = cdev; /* claim */
/* allocate notification request and buffer */
gsi->c_port.notify_req = usb_ep_alloc_request(ep, GFP_KERNEL);
if (!gsi->c_port.notify_req)
goto fail;
gsi->c_port.notify_req->buf =
kmalloc(info->notify_buf_len, GFP_KERNEL);
if (!gsi->c_port.notify_req->buf)
goto fail;
gsi->c_port.notify_req->length = info->notify_buf_len;
gsi->c_port.notify_req->context = gsi;
gsi->c_port.notify_req->complete =
gsi_ctrl_notify_resp_complete;
event = gsi->c_port.notify_req->buf;
event->bmRequestType = USB_DIR_IN | USB_TYPE_CLASS
| USB_RECIP_INTERFACE;
if (gsi->ctrl_id == -ENODEV)
event->wIndex = cpu_to_le16(gsi->data_id);
else
event->wIndex = cpu_to_le16(gsi->ctrl_id);
event->wLength = cpu_to_le16(0);
}
gsi->d_port.in_request.buf_len = info->in_req_buf_len;
gsi->d_port.in_request.num_bufs = info->in_req_num_buf;
if (gsi->d_port.out_ep) {
gsi->d_port.out_request.buf_len = info->out_req_buf_len;
gsi->d_port.out_request.num_bufs = info->out_req_num_buf;
}
/* Initialize event queue */
spin_lock_init(&gsi->d_port.evt_q.q_lock);
gsi->d_port.evt_q.head = gsi->d_port.evt_q.tail = MAXQUEUELEN - 1;
if (info->fs_in_desc) {
info->hs_in_desc->bEndpointAddress =
info->fs_in_desc->bEndpointAddress;
info->ss_in_desc->bEndpointAddress =
info->fs_in_desc->bEndpointAddress;
}
if (info->fs_out_desc) {
info->hs_out_desc->bEndpointAddress =
info->fs_out_desc->bEndpointAddress;
info->ss_out_desc->bEndpointAddress =
info->fs_out_desc->bEndpointAddress;
}
if (info->fs_notify_desc) {
info->hs_notify_desc->bEndpointAddress =
info->fs_notify_desc->bEndpointAddress;
info->ss_notify_desc->bEndpointAddress =
info->fs_notify_desc->bEndpointAddress;
}
status = usb_assign_descriptors(f, info->fs_desc_hdr, info->hs_desc_hdr,
info->ss_desc_hdr, info->ss_desc_hdr);
if (status)
goto fail;
return 0;
fail:
if (gsi->c_port.notify_req) {
kfree(gsi->c_port.notify_req->buf);
usb_ep_free_request(gsi->c_port.notify, gsi->c_port.notify_req);
}
/* we might as well release our claims on endpoints */
if (gsi->c_port.notify)
gsi->c_port.notify->driver_data = NULL;
if (gsi->d_port.out_ep && gsi->d_port.out_ep->desc)
gsi->d_port.out_ep->driver_data = NULL;
if (gsi->d_port.in_ep && gsi->d_port.in_ep->desc)
gsi->d_port.in_ep->driver_data = NULL;
log_event_err("%s: bind failed for %s", __func__, f->name);
return -ENOMEM;
}
static void ipa_ready_callback(void *user_data)
{
struct f_gsi *gsi = user_data;
log_event_info("%s: ipa is ready\n", __func__);
gsi->d_port.ipa_ready = true;
wake_up_interruptible(&gsi->d_port.wait_for_ipa_ready);
}
static void gsi_get_ether_addr(const char *str, u8 *dev_addr)
{
if (str) {
unsigned int i;
for (i = 0; i < ETH_ALEN; i++) {
unsigned char num;
if ((*str == '.') || (*str == ':'))
str++;
num = hex_to_bin(*str++) << 4;
num |= hex_to_bin(*str++);
dev_addr[i] = num;
}
if (is_valid_ether_addr(dev_addr))
return;
}
random_ether_addr(dev_addr);
}
void rmnet_gsi_update_in_buffer_mem_type(struct usb_function *f, bool use_tcm)
{
struct f_gsi *gsi = func_to_gsi(f);
if (gsi && gsi->prot_id == IPA_USB_RMNET)
gsi->rmnet_use_tcm_mem = use_tcm;
}
EXPORT_SYMBOL(rmnet_gsi_update_in_buffer_mem_type);
static int gsi_bind(struct usb_configuration *c, struct usb_function *f)
{
struct usb_composite_dev *cdev = c->cdev;
struct gsi_function_bind_info info = {0};
struct f_gsi *gsi = func_to_gsi(f);
struct rndis_params *params;
struct usb_os_desc *descs[1];
struct gsi_opts *opts;
int status;
__u8 class;
__u8 subclass;
__u8 proto;
if (gsi->prot_id == IPA_USB_RMNET ||
gsi->prot_id == IPA_USB_DIAG)
gsi->ctrl_id = -ENODEV;
else {
status = gsi->ctrl_id = usb_interface_id(c, f);
if (status < 0)
goto fail;
}
if (gsi->prot_id != IPA_USB_GPS) {
status = gsi->data_id = usb_interface_id(c, f);
if (status < 0)
goto fail;
}
switch (gsi->prot_id) {
case IPA_USB_RNDIS:
info.string_defs = rndis_gsi_string_defs;
info.ctrl_desc = &rndis_gsi_control_intf;
info.ctrl_str_idx = 0;
info.data_desc = &rndis_gsi_data_intf;
info.data_str_idx = 1;
info.iad_desc = &rndis_gsi_iad_descriptor;
info.iad_str_idx = 2;
info.union_desc = &rndis_gsi_union_desc;
info.fs_in_desc = &rndis_gsi_fs_in_desc;
info.fs_out_desc = &rndis_gsi_fs_out_desc;
info.fs_notify_desc = &rndis_gsi_fs_notify_desc;
info.hs_in_desc = &rndis_gsi_hs_in_desc;
info.hs_out_desc = &rndis_gsi_hs_out_desc;
info.hs_notify_desc = &rndis_gsi_hs_notify_desc;
info.ss_in_desc = &rndis_gsi_ss_in_desc;
info.ss_out_desc = &rndis_gsi_ss_out_desc;
info.ss_notify_desc = &rndis_gsi_ss_notify_desc;
info.fs_desc_hdr = gsi_eth_fs_function;
info.hs_desc_hdr = gsi_eth_hs_function;
info.ss_desc_hdr = gsi_eth_ss_function;
info.in_epname = "gsi-epin";
info.out_epname = "gsi-epout";
info.in_req_buf_len = GSI_IN_RNDIS_BUFF_SIZE;
gsi->d_port.in_aggr_size = GSI_IN_RNDIS_AGGR_SIZE;
info.in_req_num_buf = GSI_NUM_IN_RNDIS_BUFFERS;
gsi->d_port.out_aggr_size = GSI_OUT_AGGR_SIZE;
info.out_req_buf_len = GSI_OUT_AGGR_SIZE;
info.out_req_num_buf = GSI_NUM_OUT_BUFFERS;
info.notify_buf_len = sizeof(struct usb_cdc_notification);
params = rndis_register(gsi_rndis_response_available, gsi);
if (IS_ERR(params))
goto fail;
gsi->params = params;
rndis_set_param_medium(gsi->params, RNDIS_MEDIUM_802_3, 0);
/* export host's Ethernet address in CDC format */
gsi_get_ether_addr(gsi->dev_addr,
gsi->d_port.ipa_init_params.device_ethaddr);
gsi_get_ether_addr(gsi->host_addr,
gsi->d_port.ipa_init_params.host_ethaddr);
log_event_dbg("setting host_ethaddr=%pM, device_ethaddr = %pM",
gsi->d_port.ipa_init_params.host_ethaddr,
gsi->d_port.ipa_init_params.device_ethaddr);
memcpy(gsi->ethaddr, &gsi->d_port.ipa_init_params.host_ethaddr,
ETH_ALEN);
rndis_set_host_mac(gsi->params, gsi->ethaddr);
if (gsi->manufacturer && gsi->vendorID &&
rndis_set_param_vendor(gsi->params, gsi->vendorID,
gsi->manufacturer))
goto dereg_rndis;
/* Windows7/Windows10 automatically loads RNDIS drivers for
* class drivers which represents MISC_ACTIVE_SYNC,
* MISC_RNDIS_OVER_ETHERNET & WIRELESS_CONTROLLER_REMOTE_NDIS.
* All the codes listed below are from
* http://www.usb.org/developers/defined_class and its unknown
* why windows loads rndis class driver for some of them.
* Note that, Windows loads NDIS6 stack automatically for
* MISC_RNDIS_OVER_ETHERNET. Windows loads NDIS5 stack for
* MISC_ACTIVE_SYNC and WIRELESS_CONTROLLER_REMOTE_NDIS.
* For other class codes, NDIS stack can be selected using
* customized INF file but that defeats the purpose as its
* expected to load drivers automatically for known class
* drivers published by usbif.
* Linux rndis host driver supports MISC_ACTIVE_SYNC and
* WIRELESS_CONTROLLER_REMOTE_NDIS as of now.
* Default to rndis over ethernet which loads NDIS6 drivers
* for windows7/windows10 to avoid data stall issues
*/
if (gsi->rndis_id == RNDIS_ID_UNKNOWN)
gsi->rndis_id = MISC_RNDIS_OVER_ETHERNET;
switch (gsi->rndis_id) {
default:
/* fall throug */
case WIRELESS_CONTROLLER_REMOTE_NDIS:
class = USB_CLASS_WIRELESS_CONTROLLER;
subclass = 0x01;
proto = 0x03;
break;
case MISC_ACTIVE_SYNC:
class = USB_CLASS_MISC;
subclass = 0x01;
proto = 0x01;
break;
case MISC_RNDIS_OVER_ETHERNET:
class = USB_CLASS_MISC;
subclass = 0x04;
proto = 0x01;
break;
case MISC_RNDIS_OVER_WIFI:
class = USB_CLASS_MISC;
subclass = 0x04;
proto = 0x02;
break;
case MISC_RNDIS_OVER_WIMAX:
class = USB_CLASS_MISC;
subclass = 0x04;
proto = 0x03;
break;
case MISC_RNDIS_OVER_WWAN:
class = USB_CLASS_MISC;
subclass = 0x04;
proto = 0x04;
break;
case MISC_RNDIS_FOR_IPV4:
class = USB_CLASS_MISC;
subclass = 0x04;
proto = 0x05;
break;
case MISC_RNDIS_FOR_IPV6:
class = USB_CLASS_MISC;
subclass = 0x04;
proto = 0x06;
break;
case MISC_RNDIS_FOR_GPRS:
class = USB_CLASS_MISC;
subclass = 0x04;
proto = 0x07;
break;
}
info.iad_desc->bFunctionClass = class;
info.iad_desc->bFunctionSubClass = subclass;
info.iad_desc->bFunctionProtocol = proto;
info.ctrl_desc->bInterfaceClass = class;
info.ctrl_desc->bInterfaceSubClass = subclass;
info.ctrl_desc->bInterfaceProtocol = proto;
break;
case IPA_USB_MBIM:
info.string_defs = mbim_gsi_string_defs;
info.ctrl_desc = &mbim_gsi_control_intf;
info.ctrl_str_idx = 0;
info.data_desc = &mbim_gsi_data_intf;
info.data_str_idx = 1;
info.data_nop_desc = &mbim_gsi_data_nop_intf;
info.iad_desc = &mbim_gsi_iad_desc;
info.iad_str_idx = -1;
info.union_desc = &mbim_gsi_union_desc;
info.fs_in_desc = &mbim_gsi_fs_in_desc;
info.fs_out_desc = &mbim_gsi_fs_out_desc;
info.fs_notify_desc = &mbim_gsi_fs_notify_desc;
info.hs_in_desc = &mbim_gsi_hs_in_desc;
info.hs_out_desc = &mbim_gsi_hs_out_desc;
info.hs_notify_desc = &mbim_gsi_hs_notify_desc;
info.ss_in_desc = &mbim_gsi_ss_in_desc;
info.ss_out_desc = &mbim_gsi_ss_out_desc;
info.ss_notify_desc = &mbim_gsi_ss_notify_desc;
info.fs_desc_hdr = mbim_gsi_fs_function;
info.hs_desc_hdr = mbim_gsi_hs_function;
info.ss_desc_hdr = mbim_gsi_ss_function;
info.in_epname = "gsi-epin";
info.out_epname = "gsi-epout";
gsi->d_port.in_aggr_size = GSI_IN_MBIM_AGGR_SIZE;
info.in_req_buf_len = GSI_IN_MBIM_AGGR_SIZE;
info.in_req_num_buf = GSI_NUM_IN_BUFFERS;
gsi->d_port.out_aggr_size = GSI_OUT_MBIM_AGGR_SIZE;
info.out_req_buf_len = GSI_OUT_MBIM_BUF_LEN;
info.out_req_num_buf = GSI_NUM_OUT_BUFFERS;
info.notify_buf_len = sizeof(struct usb_cdc_notification);
mbim_gsi_desc.wMaxSegmentSize = cpu_to_le16(0x800);
if (cdev->use_os_string) {
f->os_desc_table = kzalloc(sizeof(*f->os_desc_table),
GFP_KERNEL);
if (!f->os_desc_table)
return -ENOMEM;
opts = container_of(f->fi, struct gsi_opts, func_inst);
f->os_desc_n = 1;
f->os_desc_table[0].os_desc = &opts->os_desc;
f->os_desc_table[0].if_id = gsi->data_id;
opts->os_desc.ext_compat_id = opts->ext_compat_id;
descs[0] = &opts->os_desc;
snprintf(sub_compatible_id, sizeof(sub_compatible_id),
"%u", c->bConfigurationValue);
memcpy(descs[0]->ext_compat_id, compatible_id,
strlen(compatible_id));
memcpy(descs[0]->ext_compat_id + 8, sub_compatible_id,
strlen(sub_compatible_id));
}
break;
case IPA_USB_RMNET:
info.string_defs = rmnet_gsi_string_defs;
info.data_desc = &rmnet_gsi_interface_desc;
info.data_str_idx = 0;
info.fs_in_desc = &rmnet_gsi_fs_in_desc;
info.fs_out_desc = &rmnet_gsi_fs_out_desc;
info.fs_notify_desc = &rmnet_gsi_fs_notify_desc;
info.hs_in_desc = &rmnet_gsi_hs_in_desc;
info.hs_out_desc = &rmnet_gsi_hs_out_desc;
info.hs_notify_desc = &rmnet_gsi_hs_notify_desc;
info.ss_in_desc = &rmnet_gsi_ss_in_desc;
info.ss_out_desc = &rmnet_gsi_ss_out_desc;
info.ss_notify_desc = &rmnet_gsi_ss_notify_desc;
info.fs_desc_hdr = rmnet_gsi_fs_function;
info.hs_desc_hdr = rmnet_gsi_hs_function;
info.ss_desc_hdr = rmnet_gsi_ss_function;
info.in_epname = "gsi-epin";
info.out_epname = "gsi-epout";
gsi->d_port.in_aggr_size = GSI_IN_RMNET_AGGR_SIZE;
info.in_req_buf_len = GSI_IN_RMNET_BUFF_SIZE;
info.in_req_num_buf = GSI_NUM_IN_RMNET_BUFFERS;
info.out_req_buf_len = GSI_OUT_RMNET_BUF_LEN;
gsi->d_port.out_aggr_size = GSI_OUT_AGGR_SIZE;
info.out_req_num_buf = GSI_NUM_OUT_BUFFERS;
info.notify_buf_len = sizeof(struct usb_cdc_notification);
break;
case IPA_USB_ECM:
info.string_defs = ecm_gsi_string_defs;
info.ctrl_desc = &ecm_gsi_control_intf;
info.ctrl_str_idx = 0;
info.data_desc = &ecm_gsi_data_intf;
info.data_str_idx = 2;
info.data_nop_desc = &ecm_gsi_data_nop_intf;
info.cdc_eth_desc = &ecm_gsi_desc;
info.mac_str_idx = 1;
info.union_desc = &ecm_gsi_union_desc;
info.fs_in_desc = &ecm_gsi_fs_in_desc;
info.fs_out_desc = &ecm_gsi_fs_out_desc;
info.fs_notify_desc = &ecm_gsi_fs_notify_desc;
info.hs_in_desc = &ecm_gsi_hs_in_desc;
info.hs_out_desc = &ecm_gsi_hs_out_desc;
info.hs_notify_desc = &ecm_gsi_hs_notify_desc;
info.ss_in_desc = &ecm_gsi_ss_in_desc;
info.ss_out_desc = &ecm_gsi_ss_out_desc;
info.ss_notify_desc = &ecm_gsi_ss_notify_desc;
info.fs_desc_hdr = ecm_gsi_fs_function;
info.hs_desc_hdr = ecm_gsi_hs_function;
info.ss_desc_hdr = ecm_gsi_ss_function;
info.in_epname = "gsi-epin";
info.out_epname = "gsi-epout";
gsi->d_port.in_aggr_size = GSI_ECM_AGGR_SIZE;
info.in_req_buf_len = GSI_IN_BUFF_SIZE;
info.in_req_num_buf = GSI_NUM_IN_BUFFERS;
gsi->d_port.out_aggr_size = GSI_ECM_AGGR_SIZE;
info.out_req_buf_len = GSI_OUT_ECM_BUF_LEN;
info.out_req_num_buf = GSI_NUM_OUT_BUFFERS;
info.notify_buf_len = GSI_CTRL_NOTIFY_BUFF_LEN;
/* export host's Ethernet address in CDC format */
gsi_get_ether_addr(gsi->dev_addr,
gsi->d_port.ipa_init_params.device_ethaddr);
gsi_get_ether_addr(gsi->host_addr,
gsi->d_port.ipa_init_params.host_ethaddr);
log_event_dbg("setting host_ethaddr=%pM, device_ethaddr = %pM",
gsi->d_port.ipa_init_params.host_ethaddr,
gsi->d_port.ipa_init_params.device_ethaddr);
snprintf(gsi->ethaddr, sizeof(gsi->ethaddr),
"%02X%02X%02X%02X%02X%02X",
gsi->d_port.ipa_init_params.host_ethaddr[0],
gsi->d_port.ipa_init_params.host_ethaddr[1],
gsi->d_port.ipa_init_params.host_ethaddr[2],
gsi->d_port.ipa_init_params.host_ethaddr[3],
gsi->d_port.ipa_init_params.host_ethaddr[4],
gsi->d_port.ipa_init_params.host_ethaddr[5]);
info.string_defs[1].s = gsi->ethaddr;
break;
case IPA_USB_DIAG:
info.string_defs = qdss_gsi_string_defs;
info.data_desc = &qdss_gsi_data_intf_desc;
info.data_str_idx = 0;
info.fs_in_desc = &qdss_gsi_fs_data_desc;
info.hs_in_desc = &qdss_gsi_hs_data_desc;
info.ss_in_desc = &qdss_gsi_ss_data_desc;
info.fs_desc_hdr = qdss_gsi_fs_data_only_desc;
info.hs_desc_hdr = qdss_gsi_hs_data_only_desc;
info.ss_desc_hdr = qdss_gsi_ss_data_only_desc;
info.in_epname = "gsi-epin";
info.out_epname = "";
info.in_req_buf_len = GSI_IN_DPL_BUFF_SIZE;
info.in_req_num_buf = GSI_NUM_IN_DPL_BUFFERS;
info.notify_buf_len = sizeof(struct usb_cdc_notification);
break;
case IPA_USB_GPS:
info.string_defs = gps_string_defs;
info.ctrl_str_idx = 0;
info.ctrl_desc = &gps_interface_desc;
info.fs_notify_desc = &gps_fs_notify_desc;
info.hs_notify_desc = &gps_hs_notify_desc;
info.ss_notify_desc = &gps_ss_notify_desc;
info.fs_desc_hdr = gps_fs_function;
info.hs_desc_hdr = gps_hs_function;
info.ss_desc_hdr = gps_ss_function;
info.notify_buf_len = sizeof(struct usb_cdc_notification);
break;
default:
log_event_err("%s: Invalid prot id %d", __func__,
gsi->prot_id);
return -EINVAL;
}
status = gsi_update_function_bind_params(gsi, cdev, &info);
if (status)
goto dereg_rndis;
if (gsi->prot_id == IPA_USB_GPS)
goto skip_ipa_init;
status = ipa_register_ipa_ready_cb(ipa_ready_callback, gsi);
if (!status) {
log_event_info("%s: ipa is not ready", __func__);
status = wait_event_interruptible_timeout(
gsi->d_port.wait_for_ipa_ready, gsi->d_port.ipa_ready,
msecs_to_jiffies(GSI_IPA_READY_TIMEOUT));
if (!status) {
log_event_err("%s: ipa ready timeout", __func__);
status = -ETIMEDOUT;
goto dereg_rndis;
}
}
gsi->d_port.ipa_usb_notify_cb = ipa_usb_notify_cb;
status = ipa_usb_init_teth_prot(gsi->prot_id,
&gsi->d_port.ipa_init_params, gsi->d_port.ipa_usb_notify_cb,
gsi);
if (status) {
log_event_err("%s: failed to init teth prot(%d) with err:%d",
__func__, gsi->prot_id, status);
goto dereg_rndis;
}
gsi->d_port.sm_state = STATE_INITIALIZED;
skip_ipa_init:
DBG(cdev, "%s: %s speed IN/%s OUT/%s NOTIFY/%s\n",
f->name,
gadget_is_superspeed(c->cdev->gadget) ? "super" :
gadget_is_dualspeed(c->cdev->gadget) ? "dual" : "full",
(gsi->d_port.in_ep == NULL ? "NULL" :
gsi->d_port.in_ep->name),
(gsi->d_port.out_ep == NULL ? "NULL" :
gsi->d_port.out_ep->name),
(gsi->c_port.notify == NULL ? "NULL" :
gsi->c_port.notify->name));
return 0;
dereg_rndis:
rndis_deregister(gsi->params);
kfree(f->os_desc_table);
fail:
return status;
}
static void gsi_unbind(struct usb_configuration *c, struct usb_function *f)
{
struct f_gsi *gsi = func_to_gsi(f);
log_event_dbg("%s:id:%d: dwq start", __func__, gsi->prot_id);
/*
* Use drain_workqueue to accomplish below conditions:
* 1. Make sure that any running work completed
* 2. Make sure to wait until all pending work completed i.e. workqueue
* is not having any pending work.
* Above conditions are making sure that ipa_usb_deinit_teth_prot()
* with ipa driver shall not fail due to unexpected state.
*/
drain_workqueue(gsi->d_port.ipa_usb_wq);
log_event_dbg("%s:id:%d: dwq end", __func__, gsi->prot_id);
ipa_usb_deinit_teth_prot(gsi->prot_id);
/* Reset string ids */
rndis_gsi_string_defs[0].id = 0;
ecm_gsi_string_defs[0].id = 0;
rmnet_gsi_string_defs[0].id = 0;
mbim_gsi_string_defs[0].id = 0;
qdss_gsi_string_defs[0].id = 0;
gsi->func_wakeup_pending = false;
if (gsi->prot_id == IPA_USB_RNDIS) {
gsi->d_port.sm_state = STATE_UNINITIALIZED;
rndis_deregister(gsi->params);
}
if (gsi->prot_id == IPA_USB_MBIM) {
kfree(f->os_desc_table);
f->os_desc_table = NULL;
f->os_desc_n = 0;
}
usb_free_all_descriptors(f);
if (gsi->c_port.notify) {
kfree(gsi->c_port.notify_req->buf);
usb_ep_free_request(gsi->c_port.notify, gsi->c_port.notify_req);
}
}
static void gsi_free_func(struct usb_function *f)
{
pr_debug("%s\n", __func__);
}
static int gsi_bind_config(struct f_gsi *gsi)
{
int status = 0;
enum ipa_usb_teth_prot prot_id = gsi->prot_id;
log_event_dbg("%s: prot id %d", __func__, prot_id);
switch (prot_id) {
case IPA_USB_RNDIS:
gsi->function.name = "rndis";
gsi->function.strings = rndis_gsi_strings;
break;
case IPA_USB_ECM:
gsi->function.name = "cdc_ethernet";
gsi->function.strings = ecm_gsi_strings;
break;
case IPA_USB_RMNET:
gsi->function.name = "rmnet";
gsi->function.strings = rmnet_gsi_strings;
break;
case IPA_USB_MBIM:
gsi->function.name = "mbim";
gsi->function.strings = mbim_gsi_strings;
break;
case IPA_USB_DIAG:
gsi->function.name = "dpl";
gsi->function.strings = qdss_gsi_strings;
break;
case IPA_USB_GPS:
gsi->function.name = "gps";
gsi->function.strings = gps_strings;
break;
default:
log_event_err("%s: invalid prot id %d", __func__, prot_id);
return -EINVAL;
}
/* descriptors are per-instance copies */
gsi->function.bind = gsi_bind;
gsi->function.unbind = gsi_unbind;
gsi->function.set_alt = gsi_set_alt;
gsi->function.get_alt = gsi_get_alt;
gsi->function.setup = gsi_setup;
gsi->function.disable = gsi_disable;
gsi->function.free_func = gsi_free_func;
gsi->function.suspend = gsi_suspend;
gsi->function.get_status = gsi_get_status;
gsi->function.func_suspend = gsi_func_suspend;
gsi->function.resume = gsi_resume;
return status;
}
static struct f_gsi *gsi_function_init(enum ipa_usb_teth_prot prot_id)
{
struct f_gsi *gsi;
int ret = 0;
if (prot_id >= IPA_USB_MAX_TETH_PROT_SIZE) {
pr_err("%s: invalid prot id %d\n", __func__, prot_id);
ret = -EINVAL;
goto error;
}
gsi = kzalloc(sizeof(*gsi), GFP_KERNEL);
if (!gsi) {
ret = -ENOMEM;
goto error;
}
spin_lock_init(&gsi->d_port.lock);
init_waitqueue_head(&gsi->d_port.wait_for_ipa_ready);
INIT_DELAYED_WORK(&gsi->d_port.usb_ipa_w, ipa_work_handler);
gsi->d_port.in_channel_handle = -EINVAL;
gsi->d_port.out_channel_handle = -EINVAL;
gsi->prot_id = prot_id;
gsi->d_port.ipa_usb_wq = ipa_usb_wq;
ret = gsi_function_ctrl_port_init(gsi);
if (ret) {
kfree(gsi);
goto error;
}
gsi->gsi_rw_timer_interval = DEFAULT_RW_TIMER_INTERVAL;
timer_setup(&gsi->gsi_rw_timer, gsi_rw_timer_func, 0);
return gsi;
error:
return ERR_PTR(ret);
}
static void gsi_opts_release(struct config_item *item)
{
struct gsi_opts *opts = to_gsi_opts(item);
struct f_gsi *gsi;
gsi = opts->gsi;
log_event_dbg("%s: releasing %s instance\n",
__func__, gsi->function.name);
usb_put_function_instance(&opts->func_inst);
}
static struct configfs_item_operations gsi_item_ops = {
.release = gsi_opts_release,
};
static ssize_t gsi_info_show(struct config_item *item, char *page)
{
struct ipa_usb_xdci_chan_params *ipa_chnl_params;
struct ipa_usb_xdci_connect_params *con_pms;
struct f_gsi *gsi = to_gsi_opts(item)->gsi;
int ret, j = 0;
unsigned int len = 0;
char *buf;
buf = kzalloc(PAGE_SIZE, GFP_KERNEL);
if (!buf)
return -ENOMEM;
if (gsi && atomic_read(&gsi->connected)) {
len += scnprintf(buf + len, PAGE_SIZE - len,
"Info: Prot_id:%d\n", gsi->prot_id);
ipa_chnl_params = &gsi->d_port.ipa_in_channel_params;
con_pms = &gsi->d_port.ipa_conn_pms;
len += scnprintf(buf + len, PAGE_SIZE - len, "%55s\n",
"==================================================");
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10s\n", "Ctrl Name: ", gsi->c_port.name);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "Ctrl Online: ",
gsi->c_port.ctrl_online.counter);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "Ctrl Open: ",
gsi->c_port.is_open);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "Ctrl Host to Modem: ",
gsi->c_port.host_to_modem);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "Ctrl Modem to Host: ",
gsi->c_port.modem_to_host);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "Ctrl Cpd to Modem: ",
gsi->c_port.copied_to_modem);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "Ctrl Cpd From Modem: ",
gsi->c_port.copied_from_modem);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "Ctrl Pkt Drops: ",
gsi->c_port.cpkt_drop_cnt);
len += scnprintf(buf + len, PAGE_SIZE - len, "%25s\n",
"==============");
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "Protocol ID: ", gsi->prot_id);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "SM State: ", gsi->d_port.sm_state);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "IN XferRscIndex: ",
gsi->d_port.in_xfer_rsc_index);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10d\n", "IN Chnl Hdl: ",
gsi->d_port.in_channel_handle);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10x\n", "IN Chnl Dbl Addr: ",
gsi->d_port.in_request.db_reg_phs_addr_lsb);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "IN TRB Ring Len: ",
ipa_chnl_params->xfer_ring_len);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10x\n", "IN TRB Base Addr: ", (unsigned int)
ipa_chnl_params->xfer_ring_base_addr_iova);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10x\n", "GEVENTCNTLO IN Addr: ",
ipa_chnl_params->gevntcount_low_addr);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10x\n", "DEPCMDLO IN Addr: ",
ipa_chnl_params->xfer_scratch.depcmd_low_addr);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10x\n", "IN LastTRB Addr Off: ",
ipa_chnl_params->xfer_scratch.last_trb_addr_iova);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "IN Buffer Size: ",
ipa_chnl_params->xfer_scratch.const_buffer_size);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "IN/DL Aggr Size: ",
con_pms->teth_prot_params.max_xfer_size_bytes_to_host);
ipa_chnl_params = &gsi->d_port.ipa_out_channel_params;
len += scnprintf(buf + len, PAGE_SIZE - len, "%25s\n",
"==============");
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "OUT XferRscIndex: ",
gsi->d_port.out_xfer_rsc_index);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10d\n", "OUT Channel Hdl: ",
gsi->d_port.out_channel_handle);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10x\n", "OUT Channel Dbl Addr: ",
gsi->d_port.out_request.db_reg_phs_addr_lsb);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "OUT TRB Ring Len: ",
ipa_chnl_params->xfer_ring_len);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10x\n", "OUT TRB Base Addr: ", (unsigned int)
ipa_chnl_params->xfer_ring_base_addr_iova);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10x\n", "GEVENTCNTLO OUT Addr: ",
ipa_chnl_params->gevntcount_low_addr);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10x\n", "DEPCMDLO OUT Addr: ",
ipa_chnl_params->xfer_scratch.depcmd_low_addr);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10x\n", "OUT LastTRB Addr Off: ",
ipa_chnl_params->xfer_scratch.last_trb_addr_iova);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "OUT Buffer Size: ",
ipa_chnl_params->xfer_scratch.const_buffer_size);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "OUT/UL Aggr Size: ",
con_pms->teth_prot_params.max_xfer_size_bytes_to_dev);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "OUT/UL Packets to dev: ",
con_pms->teth_prot_params.max_packet_number_to_dev);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "Net_ready_trigger:",
gsi->d_port.net_ready_trigger);
len += scnprintf(buf + len, PAGE_SIZE - len, "%25s\n",
"USB Bus Events");
for (j = 0; j < MAXQUEUELEN; j++)
len += scnprintf(buf + len, PAGE_SIZE - len,
"%d\t", gsi->d_port.evt_q.event[j]);
len += scnprintf(buf + len, PAGE_SIZE - len, "\n");
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "Eventq head: ",
gsi->d_port.evt_q.head);
len += scnprintf(buf + len, PAGE_SIZE - len,
"%25s %10u\n", "Eventq tail: ",
gsi->d_port.evt_q.tail);
}
if (len > PAGE_SIZE)
len = PAGE_SIZE;
ret = scnprintf(page, len, buf);
kfree(buf);
return ret;
}
CONFIGFS_ATTR_RO(gsi_, info);
static struct configfs_attribute *gsi_attrs[] = {
&gsi_attr_info,
NULL,
};
static struct config_item_type gsi_func_type = {
.ct_item_ops = &gsi_item_ops,
.ct_attrs = gsi_attrs,
.ct_owner = THIS_MODULE,
};
static ssize_t gsi_rndis_class_id_show(struct config_item *item, char *page)
{
struct f_gsi *gsi = to_gsi_opts(item)->gsi;
return snprintf(page, PAGE_SIZE, "%d\n", gsi->rndis_id);
}
static ssize_t gsi_rndis_class_id_store(struct config_item *item,
const char *page, size_t len)
{
struct f_gsi *gsi = to_gsi_opts(item)->gsi;
u8 id;
if (kstrtou8(page, 0, &id))
return -EINVAL;
if (id > RNDIS_ID_UNKNOWN && id < RNDIS_ID_MAX)
gsi->rndis_id = id;
else
return -EINVAL;
return len;
}
CONFIGFS_ATTR(gsi_, rndis_class_id);
static ssize_t gsi_host_addr_show(struct config_item *item, char *page)
{
struct f_gsi *gsi = to_gsi_opts(item)->gsi;
return scnprintf(page, PAGE_SIZE, "%s\n", gsi->host_addr);
}
static ssize_t gsi_host_addr_store(struct config_item *item,
const char *page, size_t len)
{
struct f_gsi *gsi = to_gsi_opts(item)->gsi;
if (len > GSI_MAX_MAC_ADDR_LEN)
return -EINVAL;
strlcpy(gsi->host_addr, page, GSI_MAX_MAC_ADDR_LEN);
return len;
}
CONFIGFS_ATTR(gsi_, host_addr);
static ssize_t gsi_dev_addr_show(struct config_item *item, char *page)
{
struct f_gsi *gsi = to_gsi_opts(item)->gsi;
return scnprintf(page, PAGE_SIZE, "%s\n", gsi->dev_addr);
}
static ssize_t gsi_dev_addr_store(struct config_item *item,
const char *page, size_t len)
{
struct f_gsi *gsi = to_gsi_opts(item)->gsi;
if (len > GSI_MAX_MAC_ADDR_LEN)
return -EINVAL;
strlcpy(gsi->dev_addr, page, GSI_MAX_MAC_ADDR_LEN);
return len;
}
CONFIGFS_ATTR(gsi_, dev_addr);
static struct configfs_attribute *gsi_rndis_attrs[] = {
&gsi_attr_info,
&gsi_attr_rndis_class_id,
&gsi_attr_host_addr,
&gsi_attr_dev_addr,
NULL,
};
static struct config_item_type gsi_func_rndis_type = {
.ct_item_ops = &gsi_item_ops,
.ct_attrs = gsi_rndis_attrs,
.ct_owner = THIS_MODULE,
};
static struct configfs_attribute *gsi_ecm_attrs[] = {
&gsi_attr_host_addr,
&gsi_attr_dev_addr,
NULL,
};
static struct config_item_type gsi_func_ecm_type = {
.ct_item_ops = &gsi_item_ops,
.ct_attrs = gsi_ecm_attrs,
.ct_owner = THIS_MODULE,
};
static ssize_t gsi_rmnet_use_tcm_mem_show(struct config_item *item, char *page)
{
struct f_gsi *gsi = to_gsi_opts(item)->gsi;
return scnprintf(page, PAGE_SIZE, "%s\n",
gsi->rmnet_use_tcm_mem ? "Enabled" : "Disabled");
}
static ssize_t gsi_rmnet_use_tcm_mem_store(struct config_item *item,
const char *page, size_t len)
{
struct f_gsi *gsi = to_gsi_opts(item)->gsi;
bool enable;
if (kstrtobool(page, &enable))
return -EINVAL;
gsi->rmnet_use_tcm_mem = enable;
return len;
}
CONFIGFS_ATTR(gsi_, rmnet_use_tcm_mem);
static struct configfs_attribute *gsi_rmnet_attrs[] = {
&gsi_attr_info,
&gsi_attr_rmnet_use_tcm_mem,
NULL,
};
static struct config_item_type gsi_func_rmnet_type = {
.ct_item_ops = &gsi_item_ops,
.ct_attrs = gsi_rmnet_attrs,
.ct_owner = THIS_MODULE,
};
static void gsi_inst_clean(struct gsi_opts *opts)
{
if (!opts)
return;
if (opts->gsi->c_port.cdev.dev) {
struct cdev *cdev = &opts->gsi->c_port.cdev;
struct f_gsi *gsi = opts->gsi;
int minor = MINOR(cdev->dev);
if (gsi->c_port.dev)
dev_set_drvdata(gsi->c_port.dev, NULL);
if (gsi->c_port.uevent_wq) {
cancel_work_sync(&gsi->c_port.uevent_work);
destroy_workqueue(gsi->c_port.uevent_wq);
gsi->c_port.uevent_wq = NULL;
}
device_destroy(gsi_class, cdev->dev);
cdev_del(cdev);
cdev->dev = 0;
gsi->c_port.dev = NULL;
ida_simple_remove(&gsi_ida, minor);
}
kfree(opts->gsi);
kfree(opts);
}
static int gsi_set_inst_name(struct usb_function_instance *fi,
const char *name)
{
int prot_id, name_len;
struct f_gsi *gsi;
char gsi_inst_name[MAX_INST_NAME_LEN + sizeof("gsi.") + 1];
void *ipc_log_ctxt;
struct gsi_opts *opts, *opts_prev;
struct usb_os_desc *descs[1];
char *names[1];
opts = container_of(fi, struct gsi_opts, func_inst);
name_len = strlen(name) + 1;
if (name_len > MAX_INST_NAME_LEN)
return -ENAMETOOLONG;
prot_id = name_to_prot_id(name);
if (prot_id < 0) {
pr_err("%s: failed to find prot id for %s instance\n",
__func__, name);
return -EINVAL;
}
mutex_lock(&inst_status[prot_id].gsi_lock);
opts_prev = inst_status[prot_id].opts;
if (opts_prev) {
mutex_unlock(&inst_status[prot_id].gsi_lock);
pr_err("%s: prot_id = %d, prev inst do not freed yet\n",
__func__, prot_id);
return -EBUSY;
}
mutex_unlock(&inst_status[prot_id].gsi_lock);
if (prot_id == IPA_USB_RMNET)
config_group_init_type_name(&opts->func_inst.group,
fi->group.cg_item.ci_name,
&gsi_func_rmnet_type);
if (prot_id == IPA_USB_RNDIS)
config_group_init_type_name(&opts->func_inst.group,
fi->group.cg_item.ci_name,
&gsi_func_rndis_type);
if (prot_id == IPA_USB_ECM)
config_group_init_type_name(&opts->func_inst.group,
fi->group.cg_item.ci_name,
&gsi_func_ecm_type);
if (prot_id == IPA_USB_MBIM) {
opts->os_desc.ext_compat_id = opts->ext_compat_id;
INIT_LIST_HEAD(&opts->os_desc.ext_prop);
descs[0] = &opts->os_desc;
names[0] = "MBIM";
opts->interf_group = usb_os_desc_prepare_interf_dir(
&opts->func_inst.group, 1,
descs, names, THIS_MODULE);
}
gsi = gsi_function_init(prot_id);
if (IS_ERR(gsi))
return PTR_ERR(gsi);
opts->gsi = gsi;
/*
* create instance name with prefixing "gsi." to differentiate
* ipc log debugfs entry
*/
snprintf(gsi_inst_name, sizeof(gsi_inst_name), "gsi.%s", name);
ipc_log_ctxt = ipc_log_context_create(NUM_LOG_PAGES, gsi_inst_name, 0);
if (!ipc_log_ctxt)
pr_err("%s: Err allocating ipc_log_ctxt for prot:%s\n",
__func__, gsi_inst_name);
opts->gsi->ipc_log_ctxt = ipc_log_ctxt;
/* Set instance status */
mutex_lock(&inst_status[prot_id].gsi_lock);
inst_status[prot_id].inst_exist = true;
inst_status[prot_id].opts = opts;
mutex_unlock(&inst_status[prot_id].gsi_lock);
return 0;
}
static void gsi_free_inst(struct usb_function_instance *f)
{
struct gsi_opts *opts = container_of(f, struct gsi_opts, func_inst);
enum ipa_usb_teth_prot prot_id;
struct f_gsi *gsi;
if (!opts || !opts->gsi)
return;
prot_id = opts->gsi->prot_id;
gsi = opts->gsi;
mutex_lock(&inst_status[prot_id].gsi_lock);
if (opts->gsi->c_port.is_open) {
/* Mark instance exist as false */
inst_status[prot_id].inst_exist = false;
mutex_unlock(&inst_status[prot_id].gsi_lock);
log_event_err(
"%s: [prot_id = %d] Dev is open, free mem when dev close\n",
__func__, prot_id);
return;
}
ipc_log_context_destroy(opts->gsi->ipc_log_ctxt);
if (opts && opts->interf_group)
kfree(opts->interf_group);
/* Clear instance status */
gsi_inst_clean(opts);
inst_status[prot_id].inst_exist = false;
inst_status[prot_id].opts = NULL;
mutex_unlock(&inst_status[prot_id].gsi_lock);
}
static struct usb_function_instance *gsi_alloc_inst(void)
{
struct gsi_opts *opts;
opts = kzalloc(sizeof(*opts), GFP_KERNEL);
if (!opts)
return ERR_PTR(-ENOMEM);
opts->func_inst.set_inst_name = gsi_set_inst_name;
opts->func_inst.free_func_inst = gsi_free_inst;
config_group_init_type_name(&opts->func_inst.group, "",
&gsi_func_type);
return &opts->func_inst;
}
static struct usb_function *gsi_alloc(struct usb_function_instance *fi)
{
struct gsi_opts *opts;
int ret;
opts = container_of(fi, struct gsi_opts, func_inst);
ret = gsi_bind_config(opts->gsi);
if (ret)
return ERR_PTR(ret);
return &opts->gsi->function;
}
DECLARE_USB_FUNCTION(gsi, gsi_alloc_inst, gsi_alloc);
MODULE_LICENSE("GPL v2");
MODULE_DESCRIPTION("GSI function driver");
static int usb_gsi_uevent(struct device *dev, struct kobj_uevent_env *env)
{
struct gsi_ctrl_port *c_port = dev_get_drvdata(dev);
struct f_gsi *gsi;
char *str = "undefined";
if (!c_port) {
dev_dbg(dev, "%s: gsi is not initialized\n", __func__);
add_uevent_var(env, "STATE=%s", str);
return 0;
}
gsi = c_port_to_gsi(c_port);
switch (gsi->prot_id) {
case IPA_USB_RMNET:
str = gsi->rmnet_dtr_status ? "connected" : "disconnected";
break;
case IPA_USB_MBIM:
case IPA_USB_DIAG:
case IPA_USB_GPS:
str = atomic_read(&gsi->connected) ?
"connected" : "disconnected";
break;
default:
return 0;
}
add_uevent_var(env, "STATE=%s", str);
log_event_dbg("%s:STATE=%s\n", c_port->name, str);
return 0;
}
static int fgsi_init(void)
{
int i;
int ret;
dev_t dev;
ipa_usb_wq = alloc_workqueue("k_ipa_usb",
WQ_UNBOUND | WQ_MEM_RECLAIM | WQ_FREEZABLE, 1);
if (!ipa_usb_wq) {
pr_err("%s(): Failed to create workqueue\n", __func__);
return -ENOMEM;
}
for (i = 0; i < IPA_USB_MAX_TETH_PROT_SIZE; i++)
mutex_init(&inst_status[i].gsi_lock);
gsi_class = class_create(THIS_MODULE, "gsi_usb");
if (IS_ERR(gsi_class)) {
ret = PTR_ERR(gsi_class);
gsi_class = NULL;
pr_err("%s: class_create() failed:%d\n", __func__, ret);
return ret;
}
gsi_class->dev_uevent = usb_gsi_uevent;
ret = alloc_chrdev_region(&dev, 0, MAX_CDEV_INSTANCES, "gsi_usb");
if (ret) {
pr_err("%s: alloc_chrdev_region() failed:%d\n", __func__, ret);
class_destroy(gsi_class);
gsi_class = NULL;
return ret;
}
major = MAJOR(dev);
usb_gsi_debugfs_init();
return usb_function_register(&gsiusb_func);
}
module_init(fgsi_init);
static void __exit fgsi_exit(void)
{
if (ipa_usb_wq)
destroy_workqueue(ipa_usb_wq);
usb_function_unregister(&gsiusb_func);
if (major) {
unregister_chrdev_region(MKDEV(major, 0), MAX_CDEV_INSTANCES);
major = 0;
}
class_destroy(gsi_class);
usb_gsi_debugfs_exit();
}
module_exit(fgsi_exit);