Merge 5.15.106 into android13-5.15-lts
Changes in 5.15.106 fsverity: don't drop pagecache at end of FS_IOC_ENABLE_VERITY usb: dwc3: gadget: move cmd_endtransfer to extra function usb: dwc3: gadget: Add 1ms delay after end transfer command without IOC kernel: kcsan: kcsan_test: build without structleak plugin kcsan: avoid passing -g for test ksmbd: don't terminate inactive sessions after a few seconds bus: imx-weim: fix branch condition evaluates to a garbage value xfrm: Zero padding when dumping algos and encap ASoC: codecs: tx-macro: Fix for KASAN: slab-out-of-bounds md: avoid signed overflow in slot_store() x86/PVH: obtain VGA console info in Dom0 net: hsr: Don't log netdev_err message on unknown prp dst node ALSA: asihpi: check pao in control_message() ALSA: hda/ca0132: fixup buffer overrun at tuning_ctl_set() fbdev: tgafb: Fix potential divide by zero sched_getaffinity: don't assume 'cpumask_size()' is fully initialized fbdev: nvidia: Fix potential divide by zero fbdev: intelfb: Fix potential divide by zero fbdev: lxfb: Fix potential divide by zero fbdev: au1200fb: Fix potential divide by zero tools/power turbostat: Fix /dev/cpu_dma_latency warnings tools/power turbostat: fix decoding of HWP_STATUS tracing: Fix wrong return in kprobe_event_gen_test.c ca8210: Fix unsigned mac_len comparison with zero in ca8210_skb_tx() mips: bmips: BCM6358: disable RAC flush for TP1 ALSA: usb-audio: Fix recursive locking at XRUN during syncing platform/x86: think-lmi: add missing type attribute platform/x86: think-lmi: use correct possible_values delimiters platform/x86: think-lmi: only display possible_values if available platform/x86: think-lmi: Add possible_values for ThinkStation mtd: rawnand: meson: invalidate cache on polling ECC bit SUNRPC: fix shutdown of NFS TCP client socket sfc: ef10: don't overwrite offload features at NIC reset scsi: megaraid_sas: Fix crash after a double completion scsi: mpt3sas: Don't print sense pool info twice ptp_qoriq: fix memory leak in probe() net: dsa: microchip: ksz8863_smi: fix bulk access r8169: fix RTL8168H and RTL8107E rx crc error regulator: Handle deferred clk net/net_failover: fix txq exceeding warning net: stmmac: don't reject VLANs when IFF_PROMISC is set drm/i915/tc: Fix the ICL PHY ownership check in TC-cold state platform/x86/intel/pmc: Alder Lake PCH slp_s0_residency fix can: bcm: bcm_tx_setup(): fix KMSAN uninit-value in vfs_write s390/vfio-ap: fix memory leak in vfio_ap device driver loop: suppress uevents while reconfiguring the device loop: LOOP_CONFIGURE: send uevents for partitions net: mvpp2: classifier flow fix fragmentation flags net: mvpp2: parser fix QinQ net: mvpp2: parser fix PPPoE smsc911x: avoid PHY being resumed when interface is not up ice: add profile conflict check for AVF FDIR ice: fix invalid check for empty list in ice_sched_assoc_vsi_to_agg() ALSA: ymfpci: Create card with device-managed snd_devm_card_new() ALSA: ymfpci: Fix BUG_ON in probe function net: ipa: compute DMA pool size properly i40e: fix registers dump after run ethtool adapter self test bnxt_en: Fix reporting of test result in ethtool selftest bnxt_en: Fix typo in PCI id to device description string mapping bnxt_en: Add missing 200G link speed reporting net: dsa: mv88e6xxx: Enable IGMP snooping on user ports only net: ethernet: mtk_eth_soc: fix flow block refcounting logic pinctrl: ocelot: Fix alt mode for ocelot iommu/vt-d: Allow zero SAGAW if second-stage not supported Input: alps - fix compatibility with -funsigned-char Input: focaltech - use explicitly signed char type cifs: prevent infinite recursion in CIFSGetDFSRefer() cifs: fix DFS traversal oops without CONFIG_CIFS_DFS_UPCALL Input: goodix - add Lenovo Yoga Book X90F to nine_bytes_report DMI table btrfs: fix race between quota disable and quota assign ioctls btrfs: scan device in non-exclusive mode zonefs: Always invalidate last cached page on append write can: j1939: prevent deadlock by moving j1939_sk_errqueue() xen/netback: don't do grant copy across page boundary net: phy: dp83869: fix default value for tx-/rx-internal-delay pinctrl: amd: Disable and mask interrupts on resume pinctrl: at91-pio4: fix domain name assignment powerpc: Don't try to copy PPR for task with NULL pt_regs NFSv4: Fix hangs when recovering open state after a server reboot ALSA: hda/conexant: Partial revert of a quirk for Lenovo ALSA: usb-audio: Fix regression on detection of Roland VS-100 ALSA: hda/realtek: Add quirks for some Clevo laptops ALSA: hda/realtek: Add quirk for Lenovo ZhaoYang CF4620Z xtensa: fix KASAN report for show_stack rcu: Fix rcu_torture_read ftrace event drm/etnaviv: fix reference leak when mmaping imported buffer drm/amd/display: Add DSC Support for Synaptics Cascaded MST Hub KVM: arm64: Disable interrupts while walking userspace PTs s390/uaccess: add missing earlyclobber annotations to __clear_user() KVM: VMX: Move preemption timer <=> hrtimer dance to common x86 KVM: x86: Inject #GP on x2APIC WRMSR that sets reserved bits 63:32 KVM: x86: Purge "highest ISR" cache when updating APICv state zonefs: Fix error message in zonefs_file_dio_append() selftests/bpf: Test btf dump for struct with padding only fields libbpf: Fix BTF-to-C converter's padding logic selftests/bpf: Add few corner cases to test padding handling of btf_dump libbpf: Fix btf_dump's packed struct determination hsr: ratelimit only when errors are printed x86/PVH: avoid 32-bit build warning when obtaining VGA console info Linux 5.15.106 Change-Id: I41744dcf497a25cba8f5a51fa273ac2330876fb1 Signed-off-by: Greg Kroah-Hartman <gregkh@google.com>
This commit is contained in:
2
Makefile
2
Makefile
@@ -1,7 +1,7 @@
|
||||
# SPDX-License-Identifier: GPL-2.0
|
||||
VERSION = 5
|
||||
PATCHLEVEL = 15
|
||||
SUBLEVEL = 105
|
||||
SUBLEVEL = 106
|
||||
EXTRAVERSION =
|
||||
NAME = Trick or Treat
|
||||
|
||||
|
||||
@@ -609,14 +609,33 @@ static int get_user_mapping_size(struct kvm *kvm, u64 addr)
|
||||
CONFIG_PGTABLE_LEVELS),
|
||||
.mm_ops = &kvm_user_mm_ops,
|
||||
};
|
||||
unsigned long flags;
|
||||
kvm_pte_t pte = 0; /* Keep GCC quiet... */
|
||||
u32 level = ~0;
|
||||
int ret;
|
||||
|
||||
/*
|
||||
* Disable IRQs so that we hazard against a concurrent
|
||||
* teardown of the userspace page tables (which relies on
|
||||
* IPI-ing threads).
|
||||
*/
|
||||
local_irq_save(flags);
|
||||
ret = kvm_pgtable_get_leaf(&pgt, addr, &pte, &level);
|
||||
VM_BUG_ON(ret);
|
||||
VM_BUG_ON(level >= KVM_PGTABLE_MAX_LEVELS);
|
||||
VM_BUG_ON(!(pte & PTE_VALID));
|
||||
local_irq_restore(flags);
|
||||
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
/*
|
||||
* Not seeing an error, but not updating level? Something went
|
||||
* deeply wrong...
|
||||
*/
|
||||
if (WARN_ON(level >= KVM_PGTABLE_MAX_LEVELS))
|
||||
return -EFAULT;
|
||||
|
||||
/* Oops, the userspace PTs are gone... Replay the fault */
|
||||
if (!kvm_pte_valid(pte))
|
||||
return -EAGAIN;
|
||||
|
||||
return BIT(ARM64_HW_PGTABLE_LEVEL_SHIFT(level));
|
||||
}
|
||||
@@ -1030,7 +1049,7 @@ static bool fault_supports_stage2_huge_mapping(struct kvm_memory_slot *memslot,
|
||||
*
|
||||
* Returns the size of the mapping.
|
||||
*/
|
||||
static unsigned long
|
||||
static long
|
||||
transparent_hugepage_adjust(struct kvm *kvm, struct kvm_memory_slot *memslot,
|
||||
unsigned long hva, kvm_pfn_t *pfnp,
|
||||
phys_addr_t *ipap)
|
||||
@@ -1042,8 +1061,15 @@ transparent_hugepage_adjust(struct kvm *kvm, struct kvm_memory_slot *memslot,
|
||||
* sure that the HVA and IPA are sufficiently aligned and that the
|
||||
* block map is contained within the memslot.
|
||||
*/
|
||||
if (fault_supports_stage2_huge_mapping(memslot, hva, PMD_SIZE) &&
|
||||
get_user_mapping_size(kvm, hva) >= PMD_SIZE) {
|
||||
if (fault_supports_stage2_huge_mapping(memslot, hva, PMD_SIZE)) {
|
||||
int sz = get_user_mapping_size(kvm, hva);
|
||||
|
||||
if (sz < 0)
|
||||
return sz;
|
||||
|
||||
if (sz < PMD_SIZE)
|
||||
return PAGE_SIZE;
|
||||
|
||||
/*
|
||||
* The address we faulted on is backed by a transparent huge
|
||||
* page. However, because we map the compound huge page and
|
||||
@@ -1255,7 +1281,7 @@ static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
|
||||
kvm_pfn_t pfn;
|
||||
bool logging_active = memslot_is_logging(memslot);
|
||||
unsigned long fault_level = kvm_vcpu_trap_get_fault_level(vcpu);
|
||||
unsigned long vma_pagesize, fault_granule;
|
||||
long vma_pagesize, fault_granule;
|
||||
enum kvm_pgtable_prot prot = KVM_PGTABLE_PROT_R;
|
||||
struct kvm_pgtable *pgt;
|
||||
|
||||
@@ -1402,6 +1428,11 @@ static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
|
||||
vma_pagesize = transparent_hugepage_adjust(kvm, memslot,
|
||||
hva, &pfn,
|
||||
&fault_ipa);
|
||||
|
||||
if (vma_pagesize < 0) {
|
||||
ret = vma_pagesize;
|
||||
goto out_unlock;
|
||||
}
|
||||
}
|
||||
|
||||
if (fault_status != FSC_PERM && !device && kvm_has_mte(kvm)) {
|
||||
|
||||
@@ -64,6 +64,8 @@ phys_addr_t dma_to_phys(struct device *dev, dma_addr_t dma_addr)
|
||||
return dma_addr;
|
||||
}
|
||||
|
||||
bool bmips_rac_flush_disable;
|
||||
|
||||
void arch_sync_dma_for_cpu_all(void)
|
||||
{
|
||||
void __iomem *cbr = BMIPS_GET_CBR();
|
||||
@@ -74,6 +76,9 @@ void arch_sync_dma_for_cpu_all(void)
|
||||
boot_cpu_type() != CPU_BMIPS4380)
|
||||
return;
|
||||
|
||||
if (unlikely(bmips_rac_flush_disable))
|
||||
return;
|
||||
|
||||
/* Flush stale data out of the readahead cache */
|
||||
cfg = __raw_readl(cbr + BMIPS_RAC_CONFIG);
|
||||
__raw_writel(cfg | 0x100, cbr + BMIPS_RAC_CONFIG);
|
||||
|
||||
@@ -34,6 +34,8 @@
|
||||
#define REG_BCM6328_OTP ((void __iomem *)CKSEG1ADDR(0x1000062c))
|
||||
#define BCM6328_TP1_DISABLED BIT(9)
|
||||
|
||||
extern bool bmips_rac_flush_disable;
|
||||
|
||||
static const unsigned long kbase = VMLINUX_LOAD_ADDRESS & 0xfff00000;
|
||||
|
||||
struct bmips_quirk {
|
||||
@@ -103,6 +105,12 @@ static void bcm6358_quirks(void)
|
||||
* disable SMP for now
|
||||
*/
|
||||
bmips_smp_enabled = 0;
|
||||
|
||||
/*
|
||||
* RAC flush causes kernel panics on BCM6358 when booting from TP1
|
||||
* because the bootloader is not initializing it properly.
|
||||
*/
|
||||
bmips_rac_flush_disable = !!(read_c0_brcm_cmt_local() & (1 << 31));
|
||||
}
|
||||
|
||||
static void bcm6368_quirks(void)
|
||||
|
||||
@@ -290,6 +290,9 @@ static int gpr_set(struct task_struct *target, const struct user_regset *regset,
|
||||
static int ppr_get(struct task_struct *target, const struct user_regset *regset,
|
||||
struct membuf to)
|
||||
{
|
||||
if (!target->thread.regs)
|
||||
return -EINVAL;
|
||||
|
||||
return membuf_write(&to, &target->thread.regs->ppr, sizeof(u64));
|
||||
}
|
||||
|
||||
@@ -297,6 +300,9 @@ static int ppr_set(struct task_struct *target, const struct user_regset *regset,
|
||||
unsigned int pos, unsigned int count, const void *kbuf,
|
||||
const void __user *ubuf)
|
||||
{
|
||||
if (!target->thread.regs)
|
||||
return -EINVAL;
|
||||
|
||||
return user_regset_copyin(&pos, &count, &kbuf, &ubuf,
|
||||
&target->thread.regs->ppr, 0, sizeof(u64));
|
||||
}
|
||||
|
||||
@@ -227,7 +227,7 @@ static inline unsigned long clear_user_mvcos(void __user *to, unsigned long size
|
||||
"4: slgr %0,%0\n"
|
||||
"5:\n"
|
||||
EX_TABLE(0b,2b) EX_TABLE(6b,2b) EX_TABLE(3b,5b) EX_TABLE(7b,5b)
|
||||
: "+a" (size), "+a" (to), "+a" (tmp1), "=a" (tmp2)
|
||||
: "+&a" (size), "+&a" (to), "+a" (tmp1), "=&a" (tmp2)
|
||||
: "a" (empty_zero_page), [spec] "K" (0x81UL)
|
||||
: "cc", "memory", "0");
|
||||
return size;
|
||||
|
||||
@@ -2316,6 +2316,7 @@ void kvm_apic_update_apicv(struct kvm_vcpu *vcpu)
|
||||
apic->irr_pending = (apic_search_irr(apic) != -1);
|
||||
apic->isr_count = count_vectors(apic->regs + APIC_ISR);
|
||||
}
|
||||
apic->highest_isr_cache = -1;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(kvm_apic_update_apicv);
|
||||
|
||||
@@ -2368,7 +2369,6 @@ void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event)
|
||||
kvm_lapic_set_reg(apic, APIC_TMR + 0x10 * i, 0);
|
||||
}
|
||||
kvm_apic_update_apicv(vcpu);
|
||||
apic->highest_isr_cache = -1;
|
||||
update_divide_count(apic);
|
||||
atomic_set(&apic->lapic_timer.pending, 0);
|
||||
|
||||
@@ -2638,7 +2638,6 @@ int kvm_apic_set_state(struct kvm_vcpu *vcpu, struct kvm_lapic_state *s)
|
||||
__start_apic_timer(apic, APIC_TMCCT);
|
||||
kvm_lapic_set_reg(apic, APIC_TMCCT, 0);
|
||||
kvm_apic_update_apicv(vcpu);
|
||||
apic->highest_isr_cache = -1;
|
||||
if (vcpu->arch.apicv_active) {
|
||||
static_call(kvm_x86_apicv_post_state_restore)(vcpu);
|
||||
static_call(kvm_x86_hwapic_irr_update)(vcpu,
|
||||
@@ -2802,6 +2801,10 @@ int kvm_x2apic_msr_write(struct kvm_vcpu *vcpu, u32 msr, u64 data)
|
||||
/* if this is ICR write vector before command */
|
||||
if (reg == APIC_ICR)
|
||||
kvm_lapic_reg_write(apic, APIC_ICR2, (u32)(data >> 32));
|
||||
else if (data >> 32)
|
||||
/* Bits 63:32 are reserved in all other registers. */
|
||||
return 1;
|
||||
|
||||
return kvm_lapic_reg_write(apic, reg, (u32)data);
|
||||
}
|
||||
|
||||
@@ -2836,6 +2839,10 @@ int kvm_hv_vapic_msr_write(struct kvm_vcpu *vcpu, u32 reg, u64 data)
|
||||
/* if this is ICR write vector before command */
|
||||
if (reg == APIC_ICR)
|
||||
kvm_lapic_reg_write(apic, APIC_ICR2, (u32)(data >> 32));
|
||||
else if (data >> 32)
|
||||
/* Bits 63:32 are reserved in all other registers. */
|
||||
return 1;
|
||||
|
||||
return kvm_lapic_reg_write(apic, reg, (u32)data);
|
||||
}
|
||||
|
||||
|
||||
@@ -7597,17 +7597,11 @@ static int vmx_pre_block(struct kvm_vcpu *vcpu)
|
||||
if (pi_pre_block(vcpu))
|
||||
return 1;
|
||||
|
||||
if (kvm_lapic_hv_timer_in_use(vcpu))
|
||||
kvm_lapic_switch_to_sw_timer(vcpu);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void vmx_post_block(struct kvm_vcpu *vcpu)
|
||||
{
|
||||
if (kvm_x86_ops.set_hv_timer)
|
||||
kvm_lapic_switch_to_hv_timer(vcpu);
|
||||
|
||||
pi_post_block(vcpu);
|
||||
}
|
||||
|
||||
|
||||
@@ -10043,12 +10043,28 @@ out:
|
||||
|
||||
static inline int vcpu_block(struct kvm *kvm, struct kvm_vcpu *vcpu)
|
||||
{
|
||||
bool hv_timer;
|
||||
|
||||
if (!kvm_arch_vcpu_runnable(vcpu) &&
|
||||
(!kvm_x86_ops.pre_block || static_call(kvm_x86_pre_block)(vcpu) == 0)) {
|
||||
/*
|
||||
* Switch to the software timer before halt-polling/blocking as
|
||||
* the guest's timer may be a break event for the vCPU, and the
|
||||
* hypervisor timer runs only when the CPU is in guest mode.
|
||||
* Switch before halt-polling so that KVM recognizes an expired
|
||||
* timer before blocking.
|
||||
*/
|
||||
hv_timer = kvm_lapic_hv_timer_in_use(vcpu);
|
||||
if (hv_timer)
|
||||
kvm_lapic_switch_to_sw_timer(vcpu);
|
||||
|
||||
srcu_read_unlock(&kvm->srcu, vcpu->srcu_idx);
|
||||
kvm_vcpu_block(vcpu);
|
||||
vcpu->srcu_idx = srcu_read_lock(&kvm->srcu);
|
||||
|
||||
if (hv_timer)
|
||||
kvm_lapic_switch_to_hv_timer(vcpu);
|
||||
|
||||
if (kvm_x86_ops.post_block)
|
||||
static_call(kvm_x86_post_block)(vcpu);
|
||||
|
||||
@@ -10287,6 +10303,11 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
|
||||
r = -EINTR;
|
||||
goto out;
|
||||
}
|
||||
/*
|
||||
* It should be impossible for the hypervisor timer to be in
|
||||
* use before KVM has ever run the vCPU.
|
||||
*/
|
||||
WARN_ON_ONCE(kvm_lapic_hv_timer_in_use(vcpu));
|
||||
kvm_vcpu_block(vcpu);
|
||||
if (kvm_apic_accept_events(vcpu) < 0) {
|
||||
r = 0;
|
||||
|
||||
@@ -45,7 +45,7 @@ obj-$(CONFIG_PARAVIRT_SPINLOCKS)+= spinlock.o
|
||||
|
||||
obj-$(CONFIG_XEN_DEBUG_FS) += debugfs.o
|
||||
|
||||
obj-$(CONFIG_XEN_PV_DOM0) += vga.o
|
||||
obj-$(CONFIG_XEN_DOM0) += vga.o
|
||||
|
||||
obj-$(CONFIG_SWIOTLB_XEN) += pci-swiotlb-xen.o
|
||||
|
||||
|
||||
@@ -1353,7 +1353,8 @@ asmlinkage __visible void __init xen_start_kernel(void)
|
||||
|
||||
x86_platform.set_legacy_features =
|
||||
xen_dom0_set_legacy_features;
|
||||
xen_init_vga(info, xen_start_info->console.dom0.info_size);
|
||||
xen_init_vga(info, xen_start_info->console.dom0.info_size,
|
||||
&boot_params.screen_info);
|
||||
xen_start_info->console.domU.mfn = 0;
|
||||
xen_start_info->console.domU.evtchn = 0;
|
||||
|
||||
|
||||
@@ -43,6 +43,19 @@ void __init xen_pvh_init(struct boot_params *boot_params)
|
||||
x86_init.oem.banner = xen_banner;
|
||||
|
||||
xen_efi_init(boot_params);
|
||||
|
||||
if (xen_initial_domain()) {
|
||||
struct xen_platform_op op = {
|
||||
.cmd = XENPF_get_dom0_console,
|
||||
};
|
||||
int ret = HYPERVISOR_platform_op(&op);
|
||||
|
||||
if (ret > 0)
|
||||
xen_init_vga(&op.u.dom0_console,
|
||||
min(ret * sizeof(char),
|
||||
sizeof(op.u.dom0_console)),
|
||||
&boot_params->screen_info);
|
||||
}
|
||||
}
|
||||
|
||||
void __init mem_map_via_hcall(struct boot_params *boot_params_p)
|
||||
|
||||
@@ -9,10 +9,9 @@
|
||||
|
||||
#include "xen-ops.h"
|
||||
|
||||
void __init xen_init_vga(const struct dom0_vga_console_info *info, size_t size)
|
||||
void __init xen_init_vga(const struct dom0_vga_console_info *info, size_t size,
|
||||
struct screen_info *screen_info)
|
||||
{
|
||||
struct screen_info *screen_info = &boot_params.screen_info;
|
||||
|
||||
/* This is drawn from a dump from vgacon:startup in
|
||||
* standard Linux. */
|
||||
screen_info->orig_video_mode = 3;
|
||||
|
||||
@@ -110,11 +110,12 @@ static inline void xen_uninit_lock_cpu(int cpu)
|
||||
|
||||
struct dom0_vga_console_info;
|
||||
|
||||
#ifdef CONFIG_XEN_PV_DOM0
|
||||
void __init xen_init_vga(const struct dom0_vga_console_info *, size_t size);
|
||||
#ifdef CONFIG_XEN_DOM0
|
||||
void __init xen_init_vga(const struct dom0_vga_console_info *, size_t size,
|
||||
struct screen_info *);
|
||||
#else
|
||||
static inline void __init xen_init_vga(const struct dom0_vga_console_info *info,
|
||||
size_t size)
|
||||
size_t size, struct screen_info *si)
|
||||
{
|
||||
}
|
||||
#endif
|
||||
|
||||
@@ -510,7 +510,7 @@ static size_t kstack_depth_to_print = CONFIG_PRINT_STACK_DEPTH;
|
||||
|
||||
void show_stack(struct task_struct *task, unsigned long *sp, const char *loglvl)
|
||||
{
|
||||
size_t len;
|
||||
size_t len, off = 0;
|
||||
|
||||
if (!sp)
|
||||
sp = stack_pointer(task);
|
||||
@@ -519,9 +519,17 @@ void show_stack(struct task_struct *task, unsigned long *sp, const char *loglvl)
|
||||
kstack_depth_to_print * STACK_DUMP_ENTRY_SIZE);
|
||||
|
||||
printk("%sStack:\n", loglvl);
|
||||
print_hex_dump(loglvl, " ", DUMP_PREFIX_NONE,
|
||||
STACK_DUMP_LINE_SIZE, STACK_DUMP_ENTRY_SIZE,
|
||||
sp, len, false);
|
||||
while (off < len) {
|
||||
u8 line[STACK_DUMP_LINE_SIZE];
|
||||
size_t line_len = len - off > STACK_DUMP_LINE_SIZE ?
|
||||
STACK_DUMP_LINE_SIZE : len - off;
|
||||
|
||||
__memcpy(line, (u8 *)sp + off, line_len);
|
||||
print_hex_dump(loglvl, " ", DUMP_PREFIX_NONE,
|
||||
STACK_DUMP_LINE_SIZE, STACK_DUMP_ENTRY_SIZE,
|
||||
line, line_len, false);
|
||||
off += STACK_DUMP_LINE_SIZE;
|
||||
}
|
||||
show_trace(task, sp, loglvl);
|
||||
}
|
||||
|
||||
|
||||
@@ -739,6 +739,10 @@ static int loop_change_fd(struct loop_device *lo, struct block_device *bdev,
|
||||
|
||||
if (!file)
|
||||
return -EBADF;
|
||||
|
||||
/* suppress uevents while reconfiguring the device */
|
||||
dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 1);
|
||||
|
||||
is_loop = is_loop_device(file);
|
||||
error = loop_global_lock_killable(lo, is_loop);
|
||||
if (error)
|
||||
@@ -793,13 +797,18 @@ static int loop_change_fd(struct loop_device *lo, struct block_device *bdev,
|
||||
fput(old_file);
|
||||
if (partscan)
|
||||
loop_reread_partitions(lo);
|
||||
return 0;
|
||||
|
||||
error = 0;
|
||||
done:
|
||||
/* enable and uncork uevent now that we are done */
|
||||
dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 0);
|
||||
return error;
|
||||
|
||||
out_err:
|
||||
loop_global_unlock(lo, is_loop);
|
||||
out_putf:
|
||||
fput(file);
|
||||
return error;
|
||||
goto done;
|
||||
}
|
||||
|
||||
/* loop sysfs attributes */
|
||||
@@ -1259,6 +1268,9 @@ static int loop_configure(struct loop_device *lo, fmode_t mode,
|
||||
goto out_unlock;
|
||||
}
|
||||
|
||||
/* suppress uevents while reconfiguring the device */
|
||||
dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 1);
|
||||
|
||||
disk_force_media_change(lo->lo_disk, DISK_EVENT_MEDIA_CHANGE);
|
||||
set_disk_ro(lo->lo_disk, (lo->lo_flags & LO_FLAGS_READ_ONLY) != 0);
|
||||
|
||||
@@ -1307,11 +1319,16 @@ static int loop_configure(struct loop_device *lo, fmode_t mode,
|
||||
if (partscan)
|
||||
lo->lo_disk->flags &= ~GENHD_FL_NO_PART_SCAN;
|
||||
|
||||
/* enable and uncork uevent now that we are done */
|
||||
dev_set_uevent_suppress(disk_to_dev(lo->lo_disk), 0);
|
||||
|
||||
loop_global_unlock(lo, is_loop);
|
||||
if (partscan)
|
||||
loop_reread_partitions(lo);
|
||||
|
||||
if (!(mode & FMODE_EXCL))
|
||||
bd_abort_claiming(bdev, loop_configure);
|
||||
|
||||
return 0;
|
||||
|
||||
out_unlock:
|
||||
|
||||
@@ -192,8 +192,8 @@ static int weim_parse_dt(struct platform_device *pdev, void __iomem *base)
|
||||
const struct of_device_id *of_id = of_match_device(weim_id_table,
|
||||
&pdev->dev);
|
||||
const struct imx_weim_devtype *devtype = of_id->data;
|
||||
int ret = 0, have_child = 0;
|
||||
struct device_node *child;
|
||||
int ret, have_child = 0;
|
||||
struct cs_timing_state ts = {};
|
||||
u32 reg;
|
||||
|
||||
|
||||
@@ -184,6 +184,21 @@ static bool needs_dsc_aux_workaround(struct dc_link *link)
|
||||
return false;
|
||||
}
|
||||
|
||||
bool is_synaptics_cascaded_panamera(struct dc_link *link, struct drm_dp_mst_port *port)
|
||||
{
|
||||
u8 branch_vendor_data[4] = { 0 }; // Vendor data 0x50C ~ 0x50F
|
||||
|
||||
if (drm_dp_dpcd_read(port->mgr->aux, DP_BRANCH_VENDOR_SPECIFIC_START, &branch_vendor_data, 4) == 4) {
|
||||
if (link->dpcd_caps.branch_dev_id == DP_BRANCH_DEVICE_ID_90CC24 &&
|
||||
IS_SYNAPTICS_CASCADED_PANAMERA(link->dpcd_caps.branch_dev_name, branch_vendor_data)) {
|
||||
DRM_INFO("Synaptics Cascaded MST hub\n");
|
||||
return true;
|
||||
}
|
||||
}
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
static bool validate_dsc_caps_on_connector(struct amdgpu_dm_connector *aconnector)
|
||||
{
|
||||
struct dc_sink *dc_sink = aconnector->dc_sink;
|
||||
@@ -207,6 +222,10 @@ static bool validate_dsc_caps_on_connector(struct amdgpu_dm_connector *aconnecto
|
||||
needs_dsc_aux_workaround(aconnector->dc_link))
|
||||
aconnector->dsc_aux = &aconnector->mst_port->dm_dp_aux.aux;
|
||||
|
||||
/* synaptics cascaded MST hub case */
|
||||
if (!aconnector->dsc_aux && is_synaptics_cascaded_panamera(aconnector->dc_link, port))
|
||||
aconnector->dsc_aux = port->mgr->aux;
|
||||
|
||||
if (!aconnector->dsc_aux)
|
||||
return false;
|
||||
|
||||
|
||||
@@ -26,6 +26,18 @@
|
||||
#ifndef __DAL_AMDGPU_DM_MST_TYPES_H__
|
||||
#define __DAL_AMDGPU_DM_MST_TYPES_H__
|
||||
|
||||
#define DP_BRANCH_VENDOR_SPECIFIC_START 0x50C
|
||||
|
||||
/**
|
||||
* Panamera MST Hub detection
|
||||
* Offset DPCD 050Eh == 0x5A indicates cascaded MST hub case
|
||||
* Check from beginning of branch device vendor specific field (050Ch)
|
||||
*/
|
||||
#define IS_SYNAPTICS_PANAMERA(branchDevName) (((int)branchDevName[4] & 0xF0) == 0x50 ? 1 : 0)
|
||||
#define BRANCH_HW_REVISION_PANAMERA_A2 0x10
|
||||
#define SYNAPTICS_CASCADED_HUB_ID 0x5A
|
||||
#define IS_SYNAPTICS_CASCADED_PANAMERA(devName, data) ((IS_SYNAPTICS_PANAMERA(devName) && ((int)data[2] == SYNAPTICS_CASCADED_HUB_ID)) ? 1 : 0)
|
||||
|
||||
struct amdgpu_display_manager;
|
||||
struct amdgpu_dm_connector;
|
||||
|
||||
|
||||
@@ -88,7 +88,15 @@ static void *etnaviv_gem_prime_vmap_impl(struct etnaviv_gem_object *etnaviv_obj)
|
||||
static int etnaviv_gem_prime_mmap_obj(struct etnaviv_gem_object *etnaviv_obj,
|
||||
struct vm_area_struct *vma)
|
||||
{
|
||||
return dma_buf_mmap(etnaviv_obj->base.dma_buf, vma, 0);
|
||||
int ret;
|
||||
|
||||
ret = dma_buf_mmap(etnaviv_obj->base.dma_buf, vma, 0);
|
||||
if (!ret) {
|
||||
/* Drop the reference acquired by drm_gem_mmap_obj(). */
|
||||
drm_gem_object_put(&etnaviv_obj->base);
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static const struct etnaviv_gem_ops etnaviv_gem_prime_ops = {
|
||||
|
||||
@@ -386,9 +386,9 @@ static bool icl_tc_phy_is_owned(struct intel_digital_port *dig_port)
|
||||
PORT_TX_DFLEXDPCSSS(dig_port->tc_phy_fia));
|
||||
if (val == 0xffffffff) {
|
||||
drm_dbg_kms(&i915->drm,
|
||||
"Port %s: PHY in TCCOLD, assume safe mode\n",
|
||||
"Port %s: PHY in TCCOLD, assume not owned\n",
|
||||
dig_port->tc_port_name);
|
||||
return true;
|
||||
return false;
|
||||
}
|
||||
|
||||
return val & DP_PHY_MODE_STATUS_NOT_SAFE(dig_port->tc_phy_fia_idx);
|
||||
|
||||
@@ -852,8 +852,8 @@ static void alps_process_packet_v6(struct psmouse *psmouse)
|
||||
x = y = z = 0;
|
||||
|
||||
/* Divide 4 since trackpoint's speed is too fast */
|
||||
input_report_rel(dev2, REL_X, (char)x / 4);
|
||||
input_report_rel(dev2, REL_Y, -((char)y / 4));
|
||||
input_report_rel(dev2, REL_X, (s8)x / 4);
|
||||
input_report_rel(dev2, REL_Y, -((s8)y / 4));
|
||||
|
||||
psmouse_report_standard_buttons(dev2, packet[3]);
|
||||
|
||||
@@ -1104,8 +1104,8 @@ static void alps_process_trackstick_packet_v7(struct psmouse *psmouse)
|
||||
((packet[3] & 0x20) << 1);
|
||||
z = (packet[5] & 0x3f) | ((packet[3] & 0x80) >> 1);
|
||||
|
||||
input_report_rel(dev2, REL_X, (char)x);
|
||||
input_report_rel(dev2, REL_Y, -((char)y));
|
||||
input_report_rel(dev2, REL_X, (s8)x);
|
||||
input_report_rel(dev2, REL_Y, -((s8)y));
|
||||
input_report_abs(dev2, ABS_PRESSURE, z);
|
||||
|
||||
psmouse_report_standard_buttons(dev2, packet[1]);
|
||||
@@ -2294,20 +2294,20 @@ static int alps_get_v3_v7_resolution(struct psmouse *psmouse, int reg_pitch)
|
||||
if (reg < 0)
|
||||
return reg;
|
||||
|
||||
x_pitch = (char)(reg << 4) >> 4; /* sign extend lower 4 bits */
|
||||
x_pitch = (s8)(reg << 4) >> 4; /* sign extend lower 4 bits */
|
||||
x_pitch = 50 + 2 * x_pitch; /* In 0.1 mm units */
|
||||
|
||||
y_pitch = (char)reg >> 4; /* sign extend upper 4 bits */
|
||||
y_pitch = (s8)reg >> 4; /* sign extend upper 4 bits */
|
||||
y_pitch = 36 + 2 * y_pitch; /* In 0.1 mm units */
|
||||
|
||||
reg = alps_command_mode_read_reg(psmouse, reg_pitch + 1);
|
||||
if (reg < 0)
|
||||
return reg;
|
||||
|
||||
x_electrode = (char)(reg << 4) >> 4; /* sign extend lower 4 bits */
|
||||
x_electrode = (s8)(reg << 4) >> 4; /* sign extend lower 4 bits */
|
||||
x_electrode = 17 + x_electrode;
|
||||
|
||||
y_electrode = (char)reg >> 4; /* sign extend upper 4 bits */
|
||||
y_electrode = (s8)reg >> 4; /* sign extend upper 4 bits */
|
||||
y_electrode = 13 + y_electrode;
|
||||
|
||||
x_phys = x_pitch * (x_electrode - 1); /* In 0.1 mm units */
|
||||
|
||||
@@ -202,8 +202,8 @@ static void focaltech_process_rel_packet(struct psmouse *psmouse,
|
||||
state->pressed = packet[0] >> 7;
|
||||
finger1 = ((packet[0] >> 4) & 0x7) - 1;
|
||||
if (finger1 < FOC_MAX_FINGERS) {
|
||||
state->fingers[finger1].x += (char)packet[1];
|
||||
state->fingers[finger1].y += (char)packet[2];
|
||||
state->fingers[finger1].x += (s8)packet[1];
|
||||
state->fingers[finger1].y += (s8)packet[2];
|
||||
} else {
|
||||
psmouse_err(psmouse, "First finger in rel packet invalid: %d\n",
|
||||
finger1);
|
||||
@@ -218,8 +218,8 @@ static void focaltech_process_rel_packet(struct psmouse *psmouse,
|
||||
*/
|
||||
finger2 = ((packet[3] >> 4) & 0x7) - 1;
|
||||
if (finger2 < FOC_MAX_FINGERS) {
|
||||
state->fingers[finger2].x += (char)packet[4];
|
||||
state->fingers[finger2].y += (char)packet[5];
|
||||
state->fingers[finger2].x += (s8)packet[4];
|
||||
state->fingers[finger2].y += (s8)packet[5];
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -123,10 +123,18 @@ static const unsigned long goodix_irq_flags[] = {
|
||||
static const struct dmi_system_id nine_bytes_report[] = {
|
||||
#if defined(CONFIG_DMI) && defined(CONFIG_X86)
|
||||
{
|
||||
.ident = "Lenovo YogaBook",
|
||||
/* YB1-X91L/F and YB1-X90L/F */
|
||||
/* Lenovo Yoga Book X90F / X90L */
|
||||
.matches = {
|
||||
DMI_MATCH(DMI_PRODUCT_NAME, "Lenovo YB1-X9")
|
||||
DMI_EXACT_MATCH(DMI_SYS_VENDOR, "Intel Corporation"),
|
||||
DMI_EXACT_MATCH(DMI_PRODUCT_NAME, "CHERRYVIEW D1 PLATFORM"),
|
||||
DMI_EXACT_MATCH(DMI_PRODUCT_VERSION, "YETI-11"),
|
||||
}
|
||||
},
|
||||
{
|
||||
/* Lenovo Yoga Book X91F / X91L */
|
||||
.matches = {
|
||||
/* Non exact match to match F + L versions */
|
||||
DMI_MATCH(DMI_PRODUCT_NAME, "Lenovo YB1-X91"),
|
||||
}
|
||||
},
|
||||
#endif
|
||||
|
||||
@@ -1080,7 +1080,8 @@ static int alloc_iommu(struct dmar_drhd_unit *drhd)
|
||||
}
|
||||
|
||||
err = -EINVAL;
|
||||
if (cap_sagaw(iommu->cap) == 0) {
|
||||
if (!cap_sagaw(iommu->cap) &&
|
||||
(!ecap_smts(iommu->ecap) || ecap_slts(iommu->ecap))) {
|
||||
pr_info("%s: No supported address widths. Not attempting DMA translation.\n",
|
||||
iommu->name);
|
||||
drhd->ignored = 1;
|
||||
|
||||
@@ -3189,6 +3189,9 @@ slot_store(struct md_rdev *rdev, const char *buf, size_t len)
|
||||
err = kstrtouint(buf, 10, (unsigned int *)&slot);
|
||||
if (err < 0)
|
||||
return err;
|
||||
if (slot < 0)
|
||||
/* overflow */
|
||||
return -ENOSPC;
|
||||
}
|
||||
if (rdev->mddev->pers && slot == -1) {
|
||||
/* Setting 'slot' on an active array requires also
|
||||
|
||||
@@ -172,6 +172,7 @@ struct meson_nfc {
|
||||
|
||||
dma_addr_t daddr;
|
||||
dma_addr_t iaddr;
|
||||
u32 info_bytes;
|
||||
|
||||
unsigned long assigned_cs;
|
||||
};
|
||||
@@ -499,6 +500,7 @@ static int meson_nfc_dma_buffer_setup(struct nand_chip *nand, void *databuf,
|
||||
nfc->daddr, datalen, dir);
|
||||
return ret;
|
||||
}
|
||||
nfc->info_bytes = infolen;
|
||||
cmd = GENCMDIADDRL(NFC_CMD_AIL, nfc->iaddr);
|
||||
writel(cmd, nfc->reg_base + NFC_REG_CMD);
|
||||
|
||||
@@ -516,8 +518,10 @@ static void meson_nfc_dma_buffer_release(struct nand_chip *nand,
|
||||
struct meson_nfc *nfc = nand_get_controller_data(nand);
|
||||
|
||||
dma_unmap_single(nfc->dev, nfc->daddr, datalen, dir);
|
||||
if (infolen)
|
||||
if (infolen) {
|
||||
dma_unmap_single(nfc->dev, nfc->iaddr, infolen, dir);
|
||||
nfc->info_bytes = 0;
|
||||
}
|
||||
}
|
||||
|
||||
static int meson_nfc_read_buf(struct nand_chip *nand, u8 *buf, int len)
|
||||
@@ -706,6 +710,8 @@ static void meson_nfc_check_ecc_pages_valid(struct meson_nfc *nfc,
|
||||
usleep_range(10, 15);
|
||||
/* info is updated by nfc dma engine*/
|
||||
smp_rmb();
|
||||
dma_sync_single_for_cpu(nfc->dev, nfc->iaddr, nfc->info_bytes,
|
||||
DMA_FROM_DEVICE);
|
||||
ret = *info & ECC_COMPLETE;
|
||||
} while (!ret);
|
||||
}
|
||||
|
||||
@@ -86,22 +86,16 @@ static const struct regmap_bus regmap_smi[] = {
|
||||
{
|
||||
.read = ksz8863_mdio_read,
|
||||
.write = ksz8863_mdio_write,
|
||||
.max_raw_read = 1,
|
||||
.max_raw_write = 1,
|
||||
},
|
||||
{
|
||||
.read = ksz8863_mdio_read,
|
||||
.write = ksz8863_mdio_write,
|
||||
.val_format_endian_default = REGMAP_ENDIAN_BIG,
|
||||
.max_raw_read = 2,
|
||||
.max_raw_write = 2,
|
||||
},
|
||||
{
|
||||
.read = ksz8863_mdio_read,
|
||||
.write = ksz8863_mdio_write,
|
||||
.val_format_endian_default = REGMAP_ENDIAN_BIG,
|
||||
.max_raw_read = 4,
|
||||
.max_raw_write = 4,
|
||||
}
|
||||
};
|
||||
|
||||
@@ -112,7 +106,6 @@ static const struct regmap_config ksz8863_regmap_config[] = {
|
||||
.pad_bits = 24,
|
||||
.val_bits = 8,
|
||||
.cache_type = REGCACHE_NONE,
|
||||
.use_single_read = 1,
|
||||
.lock = ksz_regmap_lock,
|
||||
.unlock = ksz_regmap_unlock,
|
||||
},
|
||||
@@ -122,7 +115,6 @@ static const struct regmap_config ksz8863_regmap_config[] = {
|
||||
.pad_bits = 24,
|
||||
.val_bits = 16,
|
||||
.cache_type = REGCACHE_NONE,
|
||||
.use_single_read = 1,
|
||||
.lock = ksz_regmap_lock,
|
||||
.unlock = ksz_regmap_unlock,
|
||||
},
|
||||
@@ -132,7 +124,6 @@ static const struct regmap_config ksz8863_regmap_config[] = {
|
||||
.pad_bits = 24,
|
||||
.val_bits = 32,
|
||||
.cache_type = REGCACHE_NONE,
|
||||
.use_single_read = 1,
|
||||
.lock = ksz_regmap_lock,
|
||||
.unlock = ksz_regmap_unlock,
|
||||
}
|
||||
|
||||
@@ -2912,9 +2912,14 @@ static int mv88e6xxx_setup_port(struct mv88e6xxx_chip *chip, int port)
|
||||
* If this is the upstream port for this switch, enable
|
||||
* forwarding of unknown unicasts and multicasts.
|
||||
*/
|
||||
reg = MV88E6XXX_PORT_CTL0_IGMP_MLD_SNOOP |
|
||||
MV88E6185_PORT_CTL0_USE_TAG | MV88E6185_PORT_CTL0_USE_IP |
|
||||
reg = MV88E6185_PORT_CTL0_USE_TAG | MV88E6185_PORT_CTL0_USE_IP |
|
||||
MV88E6XXX_PORT_CTL0_STATE_FORWARDING;
|
||||
/* Forward any IPv4 IGMP or IPv6 MLD frames received
|
||||
* by a USER port to the CPU port to allow snooping.
|
||||
*/
|
||||
if (dsa_is_user_port(ds, port))
|
||||
reg |= MV88E6XXX_PORT_CTL0_IGMP_MLD_SNOOP;
|
||||
|
||||
err = mv88e6xxx_port_write(chip, port, MV88E6XXX_PORT_CTL0, reg);
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
@@ -223,12 +223,12 @@ static const struct pci_device_id bnxt_pci_tbl[] = {
|
||||
{ PCI_VDEVICE(BROADCOM, 0x1750), .driver_data = BCM57508 },
|
||||
{ PCI_VDEVICE(BROADCOM, 0x1751), .driver_data = BCM57504 },
|
||||
{ PCI_VDEVICE(BROADCOM, 0x1752), .driver_data = BCM57502 },
|
||||
{ PCI_VDEVICE(BROADCOM, 0x1800), .driver_data = BCM57508_NPAR },
|
||||
{ PCI_VDEVICE(BROADCOM, 0x1800), .driver_data = BCM57502_NPAR },
|
||||
{ PCI_VDEVICE(BROADCOM, 0x1801), .driver_data = BCM57504_NPAR },
|
||||
{ PCI_VDEVICE(BROADCOM, 0x1802), .driver_data = BCM57502_NPAR },
|
||||
{ PCI_VDEVICE(BROADCOM, 0x1803), .driver_data = BCM57508_NPAR },
|
||||
{ PCI_VDEVICE(BROADCOM, 0x1802), .driver_data = BCM57508_NPAR },
|
||||
{ PCI_VDEVICE(BROADCOM, 0x1803), .driver_data = BCM57502_NPAR },
|
||||
{ PCI_VDEVICE(BROADCOM, 0x1804), .driver_data = BCM57504_NPAR },
|
||||
{ PCI_VDEVICE(BROADCOM, 0x1805), .driver_data = BCM57502_NPAR },
|
||||
{ PCI_VDEVICE(BROADCOM, 0x1805), .driver_data = BCM57508_NPAR },
|
||||
{ PCI_VDEVICE(BROADCOM, 0xd802), .driver_data = BCM58802 },
|
||||
{ PCI_VDEVICE(BROADCOM, 0xd804), .driver_data = BCM58804 },
|
||||
#ifdef CONFIG_BNXT_SRIOV
|
||||
|
||||
@@ -1202,6 +1202,7 @@ struct bnxt_link_info {
|
||||
#define BNXT_LINK_SPEED_40GB PORT_PHY_QCFG_RESP_LINK_SPEED_40GB
|
||||
#define BNXT_LINK_SPEED_50GB PORT_PHY_QCFG_RESP_LINK_SPEED_50GB
|
||||
#define BNXT_LINK_SPEED_100GB PORT_PHY_QCFG_RESP_LINK_SPEED_100GB
|
||||
#define BNXT_LINK_SPEED_200GB PORT_PHY_QCFG_RESP_LINK_SPEED_200GB
|
||||
u16 support_speeds;
|
||||
u16 support_pam4_speeds;
|
||||
u16 auto_link_speeds; /* fw adv setting */
|
||||
|
||||
@@ -1670,6 +1670,8 @@ u32 bnxt_fw_to_ethtool_speed(u16 fw_link_speed)
|
||||
return SPEED_50000;
|
||||
case BNXT_LINK_SPEED_100GB:
|
||||
return SPEED_100000;
|
||||
case BNXT_LINK_SPEED_200GB:
|
||||
return SPEED_200000;
|
||||
default:
|
||||
return SPEED_UNKNOWN;
|
||||
}
|
||||
@@ -3504,6 +3506,7 @@ static void bnxt_self_test(struct net_device *dev, struct ethtool_test *etest,
|
||||
bnxt_ulp_stop(bp);
|
||||
rc = bnxt_close_nic(bp, true, false);
|
||||
if (rc) {
|
||||
etest->flags |= ETH_TEST_FL_FAILED;
|
||||
bnxt_ulp_start(bp, rc);
|
||||
return;
|
||||
}
|
||||
|
||||
@@ -44,7 +44,7 @@ static i40e_status i40e_diag_reg_pattern_test(struct i40e_hw *hw,
|
||||
return 0;
|
||||
}
|
||||
|
||||
struct i40e_diag_reg_test_info i40e_reg_list[] = {
|
||||
const struct i40e_diag_reg_test_info i40e_reg_list[] = {
|
||||
/* offset mask elements stride */
|
||||
{I40E_QTX_CTL(0), 0x0000FFBF, 1,
|
||||
I40E_QTX_CTL(1) - I40E_QTX_CTL(0)},
|
||||
@@ -78,27 +78,28 @@ i40e_status i40e_diag_reg_test(struct i40e_hw *hw)
|
||||
{
|
||||
i40e_status ret_code = 0;
|
||||
u32 reg, mask;
|
||||
u32 elements;
|
||||
u32 i, j;
|
||||
|
||||
for (i = 0; i40e_reg_list[i].offset != 0 &&
|
||||
!ret_code; i++) {
|
||||
|
||||
elements = i40e_reg_list[i].elements;
|
||||
/* set actual reg range for dynamically allocated resources */
|
||||
if (i40e_reg_list[i].offset == I40E_QTX_CTL(0) &&
|
||||
hw->func_caps.num_tx_qp != 0)
|
||||
i40e_reg_list[i].elements = hw->func_caps.num_tx_qp;
|
||||
elements = hw->func_caps.num_tx_qp;
|
||||
if ((i40e_reg_list[i].offset == I40E_PFINT_ITRN(0, 0) ||
|
||||
i40e_reg_list[i].offset == I40E_PFINT_ITRN(1, 0) ||
|
||||
i40e_reg_list[i].offset == I40E_PFINT_ITRN(2, 0) ||
|
||||
i40e_reg_list[i].offset == I40E_QINT_TQCTL(0) ||
|
||||
i40e_reg_list[i].offset == I40E_QINT_RQCTL(0)) &&
|
||||
hw->func_caps.num_msix_vectors != 0)
|
||||
i40e_reg_list[i].elements =
|
||||
hw->func_caps.num_msix_vectors - 1;
|
||||
elements = hw->func_caps.num_msix_vectors - 1;
|
||||
|
||||
/* test register access */
|
||||
mask = i40e_reg_list[i].mask;
|
||||
for (j = 0; j < i40e_reg_list[i].elements && !ret_code; j++) {
|
||||
for (j = 0; j < elements && !ret_code; j++) {
|
||||
reg = i40e_reg_list[i].offset +
|
||||
(j * i40e_reg_list[i].stride);
|
||||
ret_code = i40e_diag_reg_pattern_test(hw, reg, mask);
|
||||
|
||||
@@ -20,7 +20,7 @@ struct i40e_diag_reg_test_info {
|
||||
u32 stride; /* bytes between each element */
|
||||
};
|
||||
|
||||
extern struct i40e_diag_reg_test_info i40e_reg_list[];
|
||||
extern const struct i40e_diag_reg_test_info i40e_reg_list[];
|
||||
|
||||
i40e_status i40e_diag_reg_test(struct i40e_hw *hw);
|
||||
i40e_status i40e_diag_eeprom_test(struct i40e_hw *hw);
|
||||
|
||||
@@ -2758,7 +2758,7 @@ static enum ice_status
|
||||
ice_sched_assoc_vsi_to_agg(struct ice_port_info *pi, u32 agg_id,
|
||||
u16 vsi_handle, unsigned long *tc_bitmap)
|
||||
{
|
||||
struct ice_sched_agg_vsi_info *agg_vsi_info, *old_agg_vsi_info = NULL;
|
||||
struct ice_sched_agg_vsi_info *agg_vsi_info, *iter, *old_agg_vsi_info = NULL;
|
||||
struct ice_sched_agg_info *agg_info, *old_agg_info;
|
||||
enum ice_status status = 0;
|
||||
struct ice_hw *hw = pi->hw;
|
||||
@@ -2776,11 +2776,13 @@ ice_sched_assoc_vsi_to_agg(struct ice_port_info *pi, u32 agg_id,
|
||||
if (old_agg_info && old_agg_info != agg_info) {
|
||||
struct ice_sched_agg_vsi_info *vtmp;
|
||||
|
||||
list_for_each_entry_safe(old_agg_vsi_info, vtmp,
|
||||
list_for_each_entry_safe(iter, vtmp,
|
||||
&old_agg_info->agg_vsi_list,
|
||||
list_entry)
|
||||
if (old_agg_vsi_info->vsi_handle == vsi_handle)
|
||||
if (iter->vsi_handle == vsi_handle) {
|
||||
old_agg_vsi_info = iter;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
/* check if entry already exist */
|
||||
|
||||
@@ -731,6 +731,72 @@ static void ice_vc_fdir_rem_prof_all(struct ice_vf *vf)
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* ice_vc_fdir_has_prof_conflict
|
||||
* @vf: pointer to the VF structure
|
||||
* @conf: FDIR configuration for each filter
|
||||
*
|
||||
* Check if @conf has conflicting profile with existing profiles
|
||||
*
|
||||
* Return: true on success, and false on error.
|
||||
*/
|
||||
static bool
|
||||
ice_vc_fdir_has_prof_conflict(struct ice_vf *vf,
|
||||
struct virtchnl_fdir_fltr_conf *conf)
|
||||
{
|
||||
struct ice_fdir_fltr *desc;
|
||||
|
||||
list_for_each_entry(desc, &vf->fdir.fdir_rule_list, fltr_node) {
|
||||
struct virtchnl_fdir_fltr_conf *existing_conf;
|
||||
enum ice_fltr_ptype flow_type_a, flow_type_b;
|
||||
struct ice_fdir_fltr *a, *b;
|
||||
|
||||
existing_conf = to_fltr_conf_from_desc(desc);
|
||||
a = &existing_conf->input;
|
||||
b = &conf->input;
|
||||
flow_type_a = a->flow_type;
|
||||
flow_type_b = b->flow_type;
|
||||
|
||||
/* No need to compare two rules with different tunnel types or
|
||||
* with the same protocol type.
|
||||
*/
|
||||
if (existing_conf->ttype != conf->ttype ||
|
||||
flow_type_a == flow_type_b)
|
||||
continue;
|
||||
|
||||
switch (flow_type_a) {
|
||||
case ICE_FLTR_PTYPE_NONF_IPV4_UDP:
|
||||
case ICE_FLTR_PTYPE_NONF_IPV4_TCP:
|
||||
case ICE_FLTR_PTYPE_NONF_IPV4_SCTP:
|
||||
if (flow_type_b == ICE_FLTR_PTYPE_NONF_IPV4_OTHER)
|
||||
return true;
|
||||
break;
|
||||
case ICE_FLTR_PTYPE_NONF_IPV4_OTHER:
|
||||
if (flow_type_b == ICE_FLTR_PTYPE_NONF_IPV4_UDP ||
|
||||
flow_type_b == ICE_FLTR_PTYPE_NONF_IPV4_TCP ||
|
||||
flow_type_b == ICE_FLTR_PTYPE_NONF_IPV4_SCTP)
|
||||
return true;
|
||||
break;
|
||||
case ICE_FLTR_PTYPE_NONF_IPV6_UDP:
|
||||
case ICE_FLTR_PTYPE_NONF_IPV6_TCP:
|
||||
case ICE_FLTR_PTYPE_NONF_IPV6_SCTP:
|
||||
if (flow_type_b == ICE_FLTR_PTYPE_NONF_IPV6_OTHER)
|
||||
return true;
|
||||
break;
|
||||
case ICE_FLTR_PTYPE_NONF_IPV6_OTHER:
|
||||
if (flow_type_b == ICE_FLTR_PTYPE_NONF_IPV6_UDP ||
|
||||
flow_type_b == ICE_FLTR_PTYPE_NONF_IPV6_TCP ||
|
||||
flow_type_b == ICE_FLTR_PTYPE_NONF_IPV6_SCTP)
|
||||
return true;
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
/**
|
||||
* ice_vc_fdir_write_flow_prof
|
||||
* @vf: pointer to the VF structure
|
||||
@@ -871,6 +937,13 @@ ice_vc_fdir_config_input_set(struct ice_vf *vf, struct virtchnl_fdir_add *fltr,
|
||||
enum ice_fltr_ptype flow;
|
||||
int ret;
|
||||
|
||||
ret = ice_vc_fdir_has_prof_conflict(vf, conf);
|
||||
if (ret) {
|
||||
dev_dbg(dev, "Found flow profile conflict for VF %d\n",
|
||||
vf->vf_id);
|
||||
return ret;
|
||||
}
|
||||
|
||||
flow = input->flow_type;
|
||||
ret = ice_vc_fdir_alloc_prof(vf, flow);
|
||||
if (ret) {
|
||||
|
||||
@@ -62,35 +62,38 @@ static const struct mvpp2_cls_flow cls_flows[MVPP2_N_PRS_FLOWS] = {
|
||||
MVPP2_DEF_FLOW(MVPP22_FLOW_TCP4, MVPP2_FL_IP4_TCP_FRAG_UNTAG,
|
||||
MVPP22_CLS_HEK_IP4_2T,
|
||||
MVPP2_PRS_RI_VLAN_NONE | MVPP2_PRS_RI_L3_IP4 |
|
||||
MVPP2_PRS_RI_L4_TCP,
|
||||
MVPP2_PRS_RI_IP_FRAG_TRUE | MVPP2_PRS_RI_L4_TCP,
|
||||
MVPP2_PRS_IP_MASK | MVPP2_PRS_RI_VLAN_MASK),
|
||||
|
||||
MVPP2_DEF_FLOW(MVPP22_FLOW_TCP4, MVPP2_FL_IP4_TCP_FRAG_UNTAG,
|
||||
MVPP22_CLS_HEK_IP4_2T,
|
||||
MVPP2_PRS_RI_VLAN_NONE | MVPP2_PRS_RI_L3_IP4_OPT |
|
||||
MVPP2_PRS_RI_L4_TCP,
|
||||
MVPP2_PRS_RI_IP_FRAG_TRUE | MVPP2_PRS_RI_L4_TCP,
|
||||
MVPP2_PRS_IP_MASK | MVPP2_PRS_RI_VLAN_MASK),
|
||||
|
||||
MVPP2_DEF_FLOW(MVPP22_FLOW_TCP4, MVPP2_FL_IP4_TCP_FRAG_UNTAG,
|
||||
MVPP22_CLS_HEK_IP4_2T,
|
||||
MVPP2_PRS_RI_VLAN_NONE | MVPP2_PRS_RI_L3_IP4_OTHER |
|
||||
MVPP2_PRS_RI_L4_TCP,
|
||||
MVPP2_PRS_RI_IP_FRAG_TRUE | MVPP2_PRS_RI_L4_TCP,
|
||||
MVPP2_PRS_IP_MASK | MVPP2_PRS_RI_VLAN_MASK),
|
||||
|
||||
/* TCP over IPv4 flows, fragmented, with vlan tag */
|
||||
MVPP2_DEF_FLOW(MVPP22_FLOW_TCP4, MVPP2_FL_IP4_TCP_FRAG_TAG,
|
||||
MVPP22_CLS_HEK_IP4_2T | MVPP22_CLS_HEK_TAGGED,
|
||||
MVPP2_PRS_RI_L3_IP4 | MVPP2_PRS_RI_L4_TCP,
|
||||
MVPP2_PRS_RI_L3_IP4 | MVPP2_PRS_RI_IP_FRAG_TRUE |
|
||||
MVPP2_PRS_RI_L4_TCP,
|
||||
MVPP2_PRS_IP_MASK),
|
||||
|
||||
MVPP2_DEF_FLOW(MVPP22_FLOW_TCP4, MVPP2_FL_IP4_TCP_FRAG_TAG,
|
||||
MVPP22_CLS_HEK_IP4_2T | MVPP22_CLS_HEK_TAGGED,
|
||||
MVPP2_PRS_RI_L3_IP4_OPT | MVPP2_PRS_RI_L4_TCP,
|
||||
MVPP2_PRS_RI_L3_IP4_OPT | MVPP2_PRS_RI_IP_FRAG_TRUE |
|
||||
MVPP2_PRS_RI_L4_TCP,
|
||||
MVPP2_PRS_IP_MASK),
|
||||
|
||||
MVPP2_DEF_FLOW(MVPP22_FLOW_TCP4, MVPP2_FL_IP4_TCP_FRAG_TAG,
|
||||
MVPP22_CLS_HEK_IP4_2T | MVPP22_CLS_HEK_TAGGED,
|
||||
MVPP2_PRS_RI_L3_IP4_OTHER | MVPP2_PRS_RI_L4_TCP,
|
||||
MVPP2_PRS_RI_L3_IP4_OTHER | MVPP2_PRS_RI_IP_FRAG_TRUE |
|
||||
MVPP2_PRS_RI_L4_TCP,
|
||||
MVPP2_PRS_IP_MASK),
|
||||
|
||||
/* UDP over IPv4 flows, Not fragmented, no vlan tag */
|
||||
@@ -132,35 +135,38 @@ static const struct mvpp2_cls_flow cls_flows[MVPP2_N_PRS_FLOWS] = {
|
||||
MVPP2_DEF_FLOW(MVPP22_FLOW_UDP4, MVPP2_FL_IP4_UDP_FRAG_UNTAG,
|
||||
MVPP22_CLS_HEK_IP4_2T,
|
||||
MVPP2_PRS_RI_VLAN_NONE | MVPP2_PRS_RI_L3_IP4 |
|
||||
MVPP2_PRS_RI_L4_UDP,
|
||||
MVPP2_PRS_RI_IP_FRAG_TRUE | MVPP2_PRS_RI_L4_UDP,
|
||||
MVPP2_PRS_IP_MASK | MVPP2_PRS_RI_VLAN_MASK),
|
||||
|
||||
MVPP2_DEF_FLOW(MVPP22_FLOW_UDP4, MVPP2_FL_IP4_UDP_FRAG_UNTAG,
|
||||
MVPP22_CLS_HEK_IP4_2T,
|
||||
MVPP2_PRS_RI_VLAN_NONE | MVPP2_PRS_RI_L3_IP4_OPT |
|
||||
MVPP2_PRS_RI_L4_UDP,
|
||||
MVPP2_PRS_RI_IP_FRAG_TRUE | MVPP2_PRS_RI_L4_UDP,
|
||||
MVPP2_PRS_IP_MASK | MVPP2_PRS_RI_VLAN_MASK),
|
||||
|
||||
MVPP2_DEF_FLOW(MVPP22_FLOW_UDP4, MVPP2_FL_IP4_UDP_FRAG_UNTAG,
|
||||
MVPP22_CLS_HEK_IP4_2T,
|
||||
MVPP2_PRS_RI_VLAN_NONE | MVPP2_PRS_RI_L3_IP4_OTHER |
|
||||
MVPP2_PRS_RI_L4_UDP,
|
||||
MVPP2_PRS_RI_IP_FRAG_TRUE | MVPP2_PRS_RI_L4_UDP,
|
||||
MVPP2_PRS_IP_MASK | MVPP2_PRS_RI_VLAN_MASK),
|
||||
|
||||
/* UDP over IPv4 flows, fragmented, with vlan tag */
|
||||
MVPP2_DEF_FLOW(MVPP22_FLOW_UDP4, MVPP2_FL_IP4_UDP_FRAG_TAG,
|
||||
MVPP22_CLS_HEK_IP4_2T | MVPP22_CLS_HEK_TAGGED,
|
||||
MVPP2_PRS_RI_L3_IP4 | MVPP2_PRS_RI_L4_UDP,
|
||||
MVPP2_PRS_RI_L3_IP4 | MVPP2_PRS_RI_IP_FRAG_TRUE |
|
||||
MVPP2_PRS_RI_L4_UDP,
|
||||
MVPP2_PRS_IP_MASK),
|
||||
|
||||
MVPP2_DEF_FLOW(MVPP22_FLOW_UDP4, MVPP2_FL_IP4_UDP_FRAG_TAG,
|
||||
MVPP22_CLS_HEK_IP4_2T | MVPP22_CLS_HEK_TAGGED,
|
||||
MVPP2_PRS_RI_L3_IP4_OPT | MVPP2_PRS_RI_L4_UDP,
|
||||
MVPP2_PRS_RI_L3_IP4_OPT | MVPP2_PRS_RI_IP_FRAG_TRUE |
|
||||
MVPP2_PRS_RI_L4_UDP,
|
||||
MVPP2_PRS_IP_MASK),
|
||||
|
||||
MVPP2_DEF_FLOW(MVPP22_FLOW_UDP4, MVPP2_FL_IP4_UDP_FRAG_TAG,
|
||||
MVPP22_CLS_HEK_IP4_2T | MVPP22_CLS_HEK_TAGGED,
|
||||
MVPP2_PRS_RI_L3_IP4_OTHER | MVPP2_PRS_RI_L4_UDP,
|
||||
MVPP2_PRS_RI_L3_IP4_OTHER | MVPP2_PRS_RI_IP_FRAG_TRUE |
|
||||
MVPP2_PRS_RI_L4_UDP,
|
||||
MVPP2_PRS_IP_MASK),
|
||||
|
||||
/* TCP over IPv6 flows, not fragmented, no vlan tag */
|
||||
|
||||
@@ -1539,8 +1539,8 @@ static int mvpp2_prs_vlan_init(struct platform_device *pdev, struct mvpp2 *priv)
|
||||
if (!priv->prs_double_vlans)
|
||||
return -ENOMEM;
|
||||
|
||||
/* Double VLAN: 0x8100, 0x88A8 */
|
||||
err = mvpp2_prs_double_vlan_add(priv, ETH_P_8021Q, ETH_P_8021AD,
|
||||
/* Double VLAN: 0x88A8, 0x8100 */
|
||||
err = mvpp2_prs_double_vlan_add(priv, ETH_P_8021AD, ETH_P_8021Q,
|
||||
MVPP2_PRS_PORT_MASK);
|
||||
if (err)
|
||||
return err;
|
||||
@@ -1607,59 +1607,45 @@ static int mvpp2_prs_vlan_init(struct platform_device *pdev, struct mvpp2 *priv)
|
||||
static int mvpp2_prs_pppoe_init(struct mvpp2 *priv)
|
||||
{
|
||||
struct mvpp2_prs_entry pe;
|
||||
int tid;
|
||||
int tid, ihl;
|
||||
|
||||
/* IPv4 over PPPoE with options */
|
||||
tid = mvpp2_prs_tcam_first_free(priv, MVPP2_PE_FIRST_FREE_TID,
|
||||
MVPP2_PE_LAST_FREE_TID);
|
||||
if (tid < 0)
|
||||
return tid;
|
||||
/* IPv4 over PPPoE with header length >= 5 */
|
||||
for (ihl = MVPP2_PRS_IPV4_IHL_MIN; ihl <= MVPP2_PRS_IPV4_IHL_MAX; ihl++) {
|
||||
tid = mvpp2_prs_tcam_first_free(priv, MVPP2_PE_FIRST_FREE_TID,
|
||||
MVPP2_PE_LAST_FREE_TID);
|
||||
if (tid < 0)
|
||||
return tid;
|
||||
|
||||
memset(&pe, 0, sizeof(pe));
|
||||
mvpp2_prs_tcam_lu_set(&pe, MVPP2_PRS_LU_PPPOE);
|
||||
pe.index = tid;
|
||||
memset(&pe, 0, sizeof(pe));
|
||||
mvpp2_prs_tcam_lu_set(&pe, MVPP2_PRS_LU_PPPOE);
|
||||
pe.index = tid;
|
||||
|
||||
mvpp2_prs_match_etype(&pe, 0, PPP_IP);
|
||||
mvpp2_prs_match_etype(&pe, 0, PPP_IP);
|
||||
mvpp2_prs_tcam_data_byte_set(&pe, MVPP2_ETH_TYPE_LEN,
|
||||
MVPP2_PRS_IPV4_HEAD | ihl,
|
||||
MVPP2_PRS_IPV4_HEAD_MASK |
|
||||
MVPP2_PRS_IPV4_IHL_MASK);
|
||||
|
||||
mvpp2_prs_sram_next_lu_set(&pe, MVPP2_PRS_LU_IP4);
|
||||
mvpp2_prs_sram_ri_update(&pe, MVPP2_PRS_RI_L3_IP4_OPT,
|
||||
MVPP2_PRS_RI_L3_PROTO_MASK);
|
||||
/* goto ipv4 dest-address (skip eth_type + IP-header-size - 4) */
|
||||
mvpp2_prs_sram_shift_set(&pe, MVPP2_ETH_TYPE_LEN +
|
||||
sizeof(struct iphdr) - 4,
|
||||
MVPP2_PRS_SRAM_OP_SEL_SHIFT_ADD);
|
||||
/* Set L3 offset */
|
||||
mvpp2_prs_sram_offset_set(&pe, MVPP2_PRS_SRAM_UDF_TYPE_L3,
|
||||
MVPP2_ETH_TYPE_LEN,
|
||||
MVPP2_PRS_SRAM_OP_SEL_UDF_ADD);
|
||||
mvpp2_prs_sram_next_lu_set(&pe, MVPP2_PRS_LU_IP4);
|
||||
mvpp2_prs_sram_ri_update(&pe, MVPP2_PRS_RI_L3_IP4,
|
||||
MVPP2_PRS_RI_L3_PROTO_MASK);
|
||||
/* goto ipv4 dst-address (skip eth_type + IP-header-size - 4) */
|
||||
mvpp2_prs_sram_shift_set(&pe, MVPP2_ETH_TYPE_LEN +
|
||||
sizeof(struct iphdr) - 4,
|
||||
MVPP2_PRS_SRAM_OP_SEL_SHIFT_ADD);
|
||||
/* Set L3 offset */
|
||||
mvpp2_prs_sram_offset_set(&pe, MVPP2_PRS_SRAM_UDF_TYPE_L3,
|
||||
MVPP2_ETH_TYPE_LEN,
|
||||
MVPP2_PRS_SRAM_OP_SEL_UDF_ADD);
|
||||
/* Set L4 offset */
|
||||
mvpp2_prs_sram_offset_set(&pe, MVPP2_PRS_SRAM_UDF_TYPE_L4,
|
||||
MVPP2_ETH_TYPE_LEN + (ihl * 4),
|
||||
MVPP2_PRS_SRAM_OP_SEL_UDF_ADD);
|
||||
|
||||
/* Update shadow table and hw entry */
|
||||
mvpp2_prs_shadow_set(priv, pe.index, MVPP2_PRS_LU_PPPOE);
|
||||
mvpp2_prs_hw_write(priv, &pe);
|
||||
|
||||
/* IPv4 over PPPoE without options */
|
||||
tid = mvpp2_prs_tcam_first_free(priv, MVPP2_PE_FIRST_FREE_TID,
|
||||
MVPP2_PE_LAST_FREE_TID);
|
||||
if (tid < 0)
|
||||
return tid;
|
||||
|
||||
pe.index = tid;
|
||||
|
||||
mvpp2_prs_tcam_data_byte_set(&pe, MVPP2_ETH_TYPE_LEN,
|
||||
MVPP2_PRS_IPV4_HEAD |
|
||||
MVPP2_PRS_IPV4_IHL_MIN,
|
||||
MVPP2_PRS_IPV4_HEAD_MASK |
|
||||
MVPP2_PRS_IPV4_IHL_MASK);
|
||||
|
||||
/* Clear ri before updating */
|
||||
pe.sram[MVPP2_PRS_SRAM_RI_WORD] = 0x0;
|
||||
pe.sram[MVPP2_PRS_SRAM_RI_CTRL_WORD] = 0x0;
|
||||
mvpp2_prs_sram_ri_update(&pe, MVPP2_PRS_RI_L3_IP4,
|
||||
MVPP2_PRS_RI_L3_PROTO_MASK);
|
||||
|
||||
/* Update shadow table and hw entry */
|
||||
mvpp2_prs_shadow_set(priv, pe.index, MVPP2_PRS_LU_PPPOE);
|
||||
mvpp2_prs_hw_write(priv, &pe);
|
||||
/* Update shadow table and hw entry */
|
||||
mvpp2_prs_shadow_set(priv, pe.index, MVPP2_PRS_LU_PPPOE);
|
||||
mvpp2_prs_hw_write(priv, &pe);
|
||||
}
|
||||
|
||||
/* IPv6 over PPPoE */
|
||||
tid = mvpp2_prs_tcam_first_free(priv, MVPP2_PE_FIRST_FREE_TID,
|
||||
|
||||
@@ -462,6 +462,7 @@ mtk_eth_setup_tc_block(struct net_device *dev, struct flow_block_offload *f)
|
||||
if (IS_ERR(block_cb))
|
||||
return PTR_ERR(block_cb);
|
||||
|
||||
flow_block_cb_incref(block_cb);
|
||||
flow_block_cb_add(block_cb, f);
|
||||
list_add_tail(&block_cb->driver_list, &block_cb_list);
|
||||
return 0;
|
||||
@@ -470,7 +471,7 @@ mtk_eth_setup_tc_block(struct net_device *dev, struct flow_block_offload *f)
|
||||
if (!block_cb)
|
||||
return -ENOENT;
|
||||
|
||||
if (flow_block_cb_decref(block_cb)) {
|
||||
if (!flow_block_cb_decref(block_cb)) {
|
||||
flow_block_cb_remove(block_cb, f);
|
||||
list_del(&block_cb->driver_list);
|
||||
}
|
||||
|
||||
@@ -970,6 +970,9 @@ static void rtl8168h_2_hw_phy_config(struct rtl8169_private *tp,
|
||||
/* disable phy pfm mode */
|
||||
phy_modify_paged(phydev, 0x0a44, 0x11, BIT(7), 0);
|
||||
|
||||
/* disable 10m pll off */
|
||||
phy_modify_paged(phydev, 0x0a43, 0x10, BIT(0), 0);
|
||||
|
||||
rtl8168g_disable_aldps(phydev);
|
||||
rtl8168g_config_eee_phy(phydev);
|
||||
}
|
||||
|
||||
@@ -1304,7 +1304,8 @@ static void efx_ef10_fini_nic(struct efx_nic *efx)
|
||||
static int efx_ef10_init_nic(struct efx_nic *efx)
|
||||
{
|
||||
struct efx_ef10_nic_data *nic_data = efx->nic_data;
|
||||
netdev_features_t hw_enc_features = 0;
|
||||
struct net_device *net_dev = efx->net_dev;
|
||||
netdev_features_t tun_feats, tso_feats;
|
||||
int rc;
|
||||
|
||||
if (nic_data->must_check_datapath_caps) {
|
||||
@@ -1349,20 +1350,30 @@ static int efx_ef10_init_nic(struct efx_nic *efx)
|
||||
nic_data->must_restore_piobufs = false;
|
||||
}
|
||||
|
||||
/* add encapsulated checksum offload features */
|
||||
/* encap features might change during reset if fw variant changed */
|
||||
if (efx_has_cap(efx, VXLAN_NVGRE) && !efx_ef10_is_vf(efx))
|
||||
hw_enc_features |= NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM;
|
||||
/* add encapsulated TSO features */
|
||||
net_dev->hw_enc_features |= NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM;
|
||||
else
|
||||
net_dev->hw_enc_features &= ~(NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM);
|
||||
|
||||
tun_feats = NETIF_F_GSO_UDP_TUNNEL | NETIF_F_GSO_GRE |
|
||||
NETIF_F_GSO_UDP_TUNNEL_CSUM | NETIF_F_GSO_GRE_CSUM;
|
||||
tso_feats = NETIF_F_TSO | NETIF_F_TSO6;
|
||||
|
||||
if (efx_has_cap(efx, TX_TSO_V2_ENCAP)) {
|
||||
netdev_features_t encap_tso_features;
|
||||
|
||||
encap_tso_features = NETIF_F_GSO_UDP_TUNNEL | NETIF_F_GSO_GRE |
|
||||
NETIF_F_GSO_UDP_TUNNEL_CSUM | NETIF_F_GSO_GRE_CSUM;
|
||||
|
||||
hw_enc_features |= encap_tso_features | NETIF_F_TSO;
|
||||
efx->net_dev->features |= encap_tso_features;
|
||||
/* If this is first nic_init, or if it is a reset and a new fw
|
||||
* variant has added new features, enable them by default.
|
||||
* If the features are not new, maintain their current value.
|
||||
*/
|
||||
if (!(net_dev->hw_features & tun_feats))
|
||||
net_dev->features |= tun_feats;
|
||||
net_dev->hw_enc_features |= tun_feats | tso_feats;
|
||||
net_dev->hw_features |= tun_feats;
|
||||
} else {
|
||||
net_dev->hw_enc_features &= ~(tun_feats | tso_feats);
|
||||
net_dev->hw_features &= ~tun_feats;
|
||||
net_dev->features &= ~tun_feats;
|
||||
}
|
||||
efx->net_dev->hw_enc_features = hw_enc_features;
|
||||
|
||||
/* don't fail init if RSS setup doesn't work */
|
||||
rc = efx->type->rx_push_rss_config(efx, false,
|
||||
@@ -3993,7 +4004,10 @@ static unsigned int ef10_check_caps(const struct efx_nic *efx,
|
||||
NETIF_F_HW_VLAN_CTAG_FILTER | \
|
||||
NETIF_F_IPV6_CSUM | \
|
||||
NETIF_F_RXHASH | \
|
||||
NETIF_F_NTUPLE)
|
||||
NETIF_F_NTUPLE | \
|
||||
NETIF_F_SG | \
|
||||
NETIF_F_RXCSUM | \
|
||||
NETIF_F_RXALL)
|
||||
|
||||
const struct efx_nic_type efx_hunt_a0_vf_nic_type = {
|
||||
.is_vf = true,
|
||||
|
||||
@@ -1003,21 +1003,18 @@ static int efx_pci_probe_post_io(struct efx_nic *efx)
|
||||
}
|
||||
|
||||
/* Determine netdevice features */
|
||||
net_dev->features |= (efx->type->offload_features | NETIF_F_SG |
|
||||
NETIF_F_TSO | NETIF_F_RXCSUM | NETIF_F_RXALL);
|
||||
if (efx->type->offload_features & (NETIF_F_IPV6_CSUM | NETIF_F_HW_CSUM)) {
|
||||
net_dev->features |= NETIF_F_TSO6;
|
||||
if (efx_has_cap(efx, TX_TSO_V2_ENCAP))
|
||||
net_dev->hw_enc_features |= NETIF_F_TSO6;
|
||||
}
|
||||
/* Check whether device supports TSO */
|
||||
if (!efx->type->tso_versions || !efx->type->tso_versions(efx))
|
||||
net_dev->features &= ~NETIF_F_ALL_TSO;
|
||||
net_dev->features |= efx->type->offload_features;
|
||||
|
||||
/* Add TSO features */
|
||||
if (efx->type->tso_versions && efx->type->tso_versions(efx))
|
||||
net_dev->features |= NETIF_F_TSO | NETIF_F_TSO6;
|
||||
|
||||
/* Mask for features that also apply to VLAN devices */
|
||||
net_dev->vlan_features |= (NETIF_F_HW_CSUM | NETIF_F_SG |
|
||||
NETIF_F_HIGHDMA | NETIF_F_ALL_TSO |
|
||||
NETIF_F_RXCSUM);
|
||||
|
||||
/* Determine user configurable features */
|
||||
net_dev->hw_features |= net_dev->features & ~efx->fixed_features;
|
||||
|
||||
/* Disable receiving frames with bad FCS, by default. */
|
||||
|
||||
@@ -1037,8 +1037,6 @@ static int smsc911x_mii_probe(struct net_device *dev)
|
||||
return ret;
|
||||
}
|
||||
|
||||
/* Indicate that the MAC is responsible for managing PHY PM */
|
||||
phydev->mac_managed_pm = true;
|
||||
phy_attached_info(phydev);
|
||||
|
||||
phy_set_max_speed(phydev, SPEED_100);
|
||||
@@ -1066,6 +1064,7 @@ static int smsc911x_mii_init(struct platform_device *pdev,
|
||||
struct net_device *dev)
|
||||
{
|
||||
struct smsc911x_data *pdata = netdev_priv(dev);
|
||||
struct phy_device *phydev;
|
||||
int err = -ENXIO;
|
||||
|
||||
pdata->mii_bus = mdiobus_alloc();
|
||||
@@ -1108,6 +1107,10 @@ static int smsc911x_mii_init(struct platform_device *pdev,
|
||||
goto err_out_free_bus_2;
|
||||
}
|
||||
|
||||
phydev = phy_find_first(pdata->mii_bus);
|
||||
if (phydev)
|
||||
phydev->mac_managed_pm = true;
|
||||
|
||||
return 0;
|
||||
|
||||
err_out_free_bus_2:
|
||||
|
||||
@@ -530,7 +530,6 @@ struct mac_device_info {
|
||||
unsigned int xlgmac;
|
||||
unsigned int num_vlan;
|
||||
u32 vlan_filter[32];
|
||||
unsigned int promisc;
|
||||
bool vlan_fail_q_en;
|
||||
u8 vlan_fail_q;
|
||||
};
|
||||
|
||||
@@ -478,12 +478,6 @@ static int dwmac4_add_hw_vlan_rx_fltr(struct net_device *dev,
|
||||
if (vid > 4095)
|
||||
return -EINVAL;
|
||||
|
||||
if (hw->promisc) {
|
||||
netdev_err(dev,
|
||||
"Adding VLAN in promisc mode not supported\n");
|
||||
return -EPERM;
|
||||
}
|
||||
|
||||
/* Single Rx VLAN Filter */
|
||||
if (hw->num_vlan == 1) {
|
||||
/* For single VLAN filter, VID 0 means VLAN promiscuous */
|
||||
@@ -533,12 +527,6 @@ static int dwmac4_del_hw_vlan_rx_fltr(struct net_device *dev,
|
||||
{
|
||||
int i, ret = 0;
|
||||
|
||||
if (hw->promisc) {
|
||||
netdev_err(dev,
|
||||
"Deleting VLAN in promisc mode not supported\n");
|
||||
return -EPERM;
|
||||
}
|
||||
|
||||
/* Single Rx VLAN Filter */
|
||||
if (hw->num_vlan == 1) {
|
||||
if ((hw->vlan_filter[0] & GMAC_VLAN_TAG_VID) == vid) {
|
||||
@@ -563,39 +551,6 @@ static int dwmac4_del_hw_vlan_rx_fltr(struct net_device *dev,
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void dwmac4_vlan_promisc_enable(struct net_device *dev,
|
||||
struct mac_device_info *hw)
|
||||
{
|
||||
void __iomem *ioaddr = hw->pcsr;
|
||||
u32 value;
|
||||
u32 hash;
|
||||
u32 val;
|
||||
int i;
|
||||
|
||||
/* Single Rx VLAN Filter */
|
||||
if (hw->num_vlan == 1) {
|
||||
dwmac4_write_single_vlan(dev, 0);
|
||||
return;
|
||||
}
|
||||
|
||||
/* Extended Rx VLAN Filter Enable */
|
||||
for (i = 0; i < hw->num_vlan; i++) {
|
||||
if (hw->vlan_filter[i] & GMAC_VLAN_TAG_DATA_VEN) {
|
||||
val = hw->vlan_filter[i] & ~GMAC_VLAN_TAG_DATA_VEN;
|
||||
dwmac4_write_vlan_filter(dev, hw, i, val);
|
||||
}
|
||||
}
|
||||
|
||||
hash = readl(ioaddr + GMAC_VLAN_HASH_TABLE);
|
||||
if (hash & GMAC_VLAN_VLHT) {
|
||||
value = readl(ioaddr + GMAC_VLAN_TAG);
|
||||
if (value & GMAC_VLAN_VTHM) {
|
||||
value &= ~GMAC_VLAN_VTHM;
|
||||
writel(value, ioaddr + GMAC_VLAN_TAG);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
static void dwmac4_restore_hw_vlan_rx_fltr(struct net_device *dev,
|
||||
struct mac_device_info *hw)
|
||||
{
|
||||
@@ -715,22 +670,12 @@ static void dwmac4_set_filter(struct mac_device_info *hw,
|
||||
}
|
||||
|
||||
/* VLAN filtering */
|
||||
if (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)
|
||||
if (dev->flags & IFF_PROMISC && !hw->vlan_fail_q_en)
|
||||
value &= ~GMAC_PACKET_FILTER_VTFE;
|
||||
else if (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)
|
||||
value |= GMAC_PACKET_FILTER_VTFE;
|
||||
|
||||
writel(value, ioaddr + GMAC_PACKET_FILTER);
|
||||
|
||||
if (dev->flags & IFF_PROMISC && !hw->vlan_fail_q_en) {
|
||||
if (!hw->promisc) {
|
||||
hw->promisc = 1;
|
||||
dwmac4_vlan_promisc_enable(dev, hw);
|
||||
}
|
||||
} else {
|
||||
if (hw->promisc) {
|
||||
hw->promisc = 0;
|
||||
dwmac4_restore_hw_vlan_rx_fltr(dev, hw);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
static void dwmac4_flow_ctrl(struct mac_device_info *hw, unsigned int duplex,
|
||||
|
||||
@@ -1945,10 +1945,9 @@ static int ca8210_skb_tx(
|
||||
struct ca8210_priv *priv
|
||||
)
|
||||
{
|
||||
int status;
|
||||
struct ieee802154_hdr header = { };
|
||||
struct secspec secspec;
|
||||
unsigned int mac_len;
|
||||
int mac_len, status;
|
||||
|
||||
dev_dbg(&priv->spi->dev, "%s called\n", __func__);
|
||||
|
||||
|
||||
@@ -155,7 +155,7 @@ int gsi_trans_pool_init_dma(struct device *dev, struct gsi_trans_pool *pool,
|
||||
* gsi_trans_pool_exit_dma() can assume the total allocated
|
||||
* size is exactly (count * size).
|
||||
*/
|
||||
total_size = get_order(total_size) << PAGE_SHIFT;
|
||||
total_size = PAGE_SIZE << get_order(total_size);
|
||||
|
||||
virt = dma_alloc_coherent(dev, total_size, &addr, GFP_KERNEL);
|
||||
if (!virt)
|
||||
|
||||
@@ -130,14 +130,10 @@ static u16 net_failover_select_queue(struct net_device *dev,
|
||||
txq = ops->ndo_select_queue(primary_dev, skb, sb_dev);
|
||||
else
|
||||
txq = netdev_pick_tx(primary_dev, skb, NULL);
|
||||
|
||||
qdisc_skb_cb(skb)->slave_dev_queue_mapping = skb->queue_mapping;
|
||||
|
||||
return txq;
|
||||
} else {
|
||||
txq = skb_rx_queue_recorded(skb) ? skb_get_rx_queue(skb) : 0;
|
||||
}
|
||||
|
||||
txq = skb_rx_queue_recorded(skb) ? skb_get_rx_queue(skb) : 0;
|
||||
|
||||
/* Save the original txq to restore before passing to the driver */
|
||||
qdisc_skb_cb(skb)->slave_dev_queue_mapping = skb->queue_mapping;
|
||||
|
||||
|
||||
@@ -587,15 +587,13 @@ static int dp83869_of_init(struct phy_device *phydev)
|
||||
&dp83869_internal_delay[0],
|
||||
delay_size, true);
|
||||
if (dp83869->rx_int_delay < 0)
|
||||
dp83869->rx_int_delay =
|
||||
dp83869_internal_delay[DP83869_CLK_DELAY_DEF];
|
||||
dp83869->rx_int_delay = DP83869_CLK_DELAY_DEF;
|
||||
|
||||
dp83869->tx_int_delay = phy_get_internal_delay(phydev, dev,
|
||||
&dp83869_internal_delay[0],
|
||||
delay_size, false);
|
||||
if (dp83869->tx_int_delay < 0)
|
||||
dp83869->tx_int_delay =
|
||||
dp83869_internal_delay[DP83869_CLK_DELAY_DEF];
|
||||
dp83869->tx_int_delay = DP83869_CLK_DELAY_DEF;
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
@@ -166,7 +166,7 @@ struct xenvif_queue { /* Per-queue data for xenvif */
|
||||
struct pending_tx_info pending_tx_info[MAX_PENDING_REQS];
|
||||
grant_handle_t grant_tx_handle[MAX_PENDING_REQS];
|
||||
|
||||
struct gnttab_copy tx_copy_ops[MAX_PENDING_REQS];
|
||||
struct gnttab_copy tx_copy_ops[2 * MAX_PENDING_REQS];
|
||||
struct gnttab_map_grant_ref tx_map_ops[MAX_PENDING_REQS];
|
||||
struct gnttab_unmap_grant_ref tx_unmap_ops[MAX_PENDING_REQS];
|
||||
/* passed to gnttab_[un]map_refs with pages under (un)mapping */
|
||||
|
||||
@@ -334,6 +334,7 @@ static int xenvif_count_requests(struct xenvif_queue *queue,
|
||||
struct xenvif_tx_cb {
|
||||
u16 copy_pending_idx[XEN_NETBK_LEGACY_SLOTS_MAX + 1];
|
||||
u8 copy_count;
|
||||
u32 split_mask;
|
||||
};
|
||||
|
||||
#define XENVIF_TX_CB(skb) ((struct xenvif_tx_cb *)(skb)->cb)
|
||||
@@ -361,6 +362,8 @@ static inline struct sk_buff *xenvif_alloc_skb(unsigned int size)
|
||||
struct sk_buff *skb =
|
||||
alloc_skb(size + NET_SKB_PAD + NET_IP_ALIGN,
|
||||
GFP_ATOMIC | __GFP_NOWARN);
|
||||
|
||||
BUILD_BUG_ON(sizeof(*XENVIF_TX_CB(skb)) > sizeof(skb->cb));
|
||||
if (unlikely(skb == NULL))
|
||||
return NULL;
|
||||
|
||||
@@ -396,11 +399,13 @@ static void xenvif_get_requests(struct xenvif_queue *queue,
|
||||
nr_slots = shinfo->nr_frags + 1;
|
||||
|
||||
copy_count(skb) = 0;
|
||||
XENVIF_TX_CB(skb)->split_mask = 0;
|
||||
|
||||
/* Create copy ops for exactly data_len bytes into the skb head. */
|
||||
__skb_put(skb, data_len);
|
||||
while (data_len > 0) {
|
||||
int amount = data_len > txp->size ? txp->size : data_len;
|
||||
bool split = false;
|
||||
|
||||
cop->source.u.ref = txp->gref;
|
||||
cop->source.domid = queue->vif->domid;
|
||||
@@ -413,6 +418,13 @@ static void xenvif_get_requests(struct xenvif_queue *queue,
|
||||
cop->dest.u.gmfn = virt_to_gfn(skb->data + skb_headlen(skb)
|
||||
- data_len);
|
||||
|
||||
/* Don't cross local page boundary! */
|
||||
if (cop->dest.offset + amount > XEN_PAGE_SIZE) {
|
||||
amount = XEN_PAGE_SIZE - cop->dest.offset;
|
||||
XENVIF_TX_CB(skb)->split_mask |= 1U << copy_count(skb);
|
||||
split = true;
|
||||
}
|
||||
|
||||
cop->len = amount;
|
||||
cop->flags = GNTCOPY_source_gref;
|
||||
|
||||
@@ -420,7 +432,8 @@ static void xenvif_get_requests(struct xenvif_queue *queue,
|
||||
pending_idx = queue->pending_ring[index];
|
||||
callback_param(queue, pending_idx).ctx = NULL;
|
||||
copy_pending_idx(skb, copy_count(skb)) = pending_idx;
|
||||
copy_count(skb)++;
|
||||
if (!split)
|
||||
copy_count(skb)++;
|
||||
|
||||
cop++;
|
||||
data_len -= amount;
|
||||
@@ -441,7 +454,8 @@ static void xenvif_get_requests(struct xenvif_queue *queue,
|
||||
nr_slots--;
|
||||
} else {
|
||||
/* The copy op partially covered the tx_request.
|
||||
* The remainder will be mapped.
|
||||
* The remainder will be mapped or copied in the next
|
||||
* iteration.
|
||||
*/
|
||||
txp->offset += amount;
|
||||
txp->size -= amount;
|
||||
@@ -539,6 +553,13 @@ static int xenvif_tx_check_gop(struct xenvif_queue *queue,
|
||||
pending_idx = copy_pending_idx(skb, i);
|
||||
|
||||
newerr = (*gopp_copy)->status;
|
||||
|
||||
/* Split copies need to be handled together. */
|
||||
if (XENVIF_TX_CB(skb)->split_mask & (1U << i)) {
|
||||
(*gopp_copy)++;
|
||||
if (!newerr)
|
||||
newerr = (*gopp_copy)->status;
|
||||
}
|
||||
if (likely(!newerr)) {
|
||||
/* The first frag might still have this slot mapped */
|
||||
if (i < copy_count(skb) - 1 || !sharedslot)
|
||||
|
||||
@@ -863,32 +863,34 @@ static const struct pinconf_ops amd_pinconf_ops = {
|
||||
.pin_config_group_set = amd_pinconf_group_set,
|
||||
};
|
||||
|
||||
static void amd_gpio_irq_init(struct amd_gpio *gpio_dev)
|
||||
static void amd_gpio_irq_init_pin(struct amd_gpio *gpio_dev, int pin)
|
||||
{
|
||||
struct pinctrl_desc *desc = gpio_dev->pctrl->desc;
|
||||
const struct pin_desc *pd;
|
||||
unsigned long flags;
|
||||
u32 pin_reg, mask;
|
||||
int i;
|
||||
|
||||
mask = BIT(WAKE_CNTRL_OFF_S0I3) | BIT(WAKE_CNTRL_OFF_S3) |
|
||||
BIT(INTERRUPT_MASK_OFF) | BIT(INTERRUPT_ENABLE_OFF) |
|
||||
BIT(WAKE_CNTRL_OFF_S4);
|
||||
|
||||
for (i = 0; i < desc->npins; i++) {
|
||||
int pin = desc->pins[i].number;
|
||||
const struct pin_desc *pd = pin_desc_get(gpio_dev->pctrl, pin);
|
||||
pd = pin_desc_get(gpio_dev->pctrl, pin);
|
||||
if (!pd)
|
||||
return;
|
||||
|
||||
if (!pd)
|
||||
continue;
|
||||
raw_spin_lock_irqsave(&gpio_dev->lock, flags);
|
||||
pin_reg = readl(gpio_dev->base + pin * 4);
|
||||
pin_reg &= ~mask;
|
||||
writel(pin_reg, gpio_dev->base + pin * 4);
|
||||
raw_spin_unlock_irqrestore(&gpio_dev->lock, flags);
|
||||
}
|
||||
|
||||
raw_spin_lock_irqsave(&gpio_dev->lock, flags);
|
||||
static void amd_gpio_irq_init(struct amd_gpio *gpio_dev)
|
||||
{
|
||||
struct pinctrl_desc *desc = gpio_dev->pctrl->desc;
|
||||
int i;
|
||||
|
||||
pin_reg = readl(gpio_dev->base + i * 4);
|
||||
pin_reg &= ~mask;
|
||||
writel(pin_reg, gpio_dev->base + i * 4);
|
||||
|
||||
raw_spin_unlock_irqrestore(&gpio_dev->lock, flags);
|
||||
}
|
||||
for (i = 0; i < desc->npins; i++)
|
||||
amd_gpio_irq_init_pin(gpio_dev, i);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_PM_SLEEP
|
||||
@@ -941,8 +943,10 @@ static int amd_gpio_resume(struct device *dev)
|
||||
for (i = 0; i < desc->npins; i++) {
|
||||
int pin = desc->pins[i].number;
|
||||
|
||||
if (!amd_gpio_should_save(gpio_dev, pin))
|
||||
if (!amd_gpio_should_save(gpio_dev, pin)) {
|
||||
amd_gpio_irq_init_pin(gpio_dev, pin);
|
||||
continue;
|
||||
}
|
||||
|
||||
raw_spin_lock_irqsave(&gpio_dev->lock, flags);
|
||||
gpio_dev->saved_regs[i] |= readl(gpio_dev->base + pin * 4) & PIN_IRQ_PENDING;
|
||||
|
||||
@@ -1184,7 +1184,6 @@ static int atmel_pinctrl_probe(struct platform_device *pdev)
|
||||
dev_err(dev, "can't add the irq domain\n");
|
||||
return -ENODEV;
|
||||
}
|
||||
atmel_pioctrl->irq_domain->name = "atmel gpio";
|
||||
|
||||
for (i = 0; i < atmel_pioctrl->npins; i++) {
|
||||
int irq = irq_create_mapping(atmel_pioctrl->irq_domain, i);
|
||||
|
||||
@@ -739,7 +739,7 @@ static int ocelot_pinmux_set_mux(struct pinctrl_dev *pctldev,
|
||||
regmap_update_bits(info->map, REG_ALT(0, info, pin->pin),
|
||||
BIT(p), f << p);
|
||||
regmap_update_bits(info->map, REG_ALT(1, info, pin->pin),
|
||||
BIT(p), f << (p - 1));
|
||||
BIT(p), (f >> 1) << p);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
@@ -958,7 +958,18 @@ static inline void pmc_core_reg_write(struct pmc_dev *pmcdev, int reg_offset,
|
||||
|
||||
static inline u64 pmc_core_adjust_slp_s0_step(struct pmc_dev *pmcdev, u32 value)
|
||||
{
|
||||
return (u64)value * pmcdev->map->slp_s0_res_counter_step;
|
||||
/*
|
||||
* ADL PCH does not have the SLP_S0 counter and LPM Residency counters are
|
||||
* used as a workaround which uses 30.5 usec tick. All other client
|
||||
* programs have the legacy SLP_S0 residency counter that is using the 122
|
||||
* usec tick.
|
||||
*/
|
||||
const int lpm_adj_x2 = pmcdev->map->lpm_res_counter_step_x2;
|
||||
|
||||
if (pmcdev->map == &adl_reg_map)
|
||||
return (u64)value * GET_X2_COUNTER((u64)lpm_adj_x2);
|
||||
else
|
||||
return (u64)value * pmcdev->map->slp_s0_res_counter_step;
|
||||
}
|
||||
|
||||
static int set_etr3(struct pmc_dev *pmcdev)
|
||||
|
||||
@@ -525,12 +525,23 @@ static ssize_t possible_values_show(struct kobject *kobj, struct kobj_attribute
|
||||
{
|
||||
struct tlmi_attr_setting *setting = to_tlmi_attr_setting(kobj);
|
||||
|
||||
if (!tlmi_priv.can_get_bios_selections)
|
||||
return -EOPNOTSUPP;
|
||||
|
||||
return sysfs_emit(buf, "%s\n", setting->possible_values);
|
||||
}
|
||||
|
||||
static ssize_t type_show(struct kobject *kobj, struct kobj_attribute *attr,
|
||||
char *buf)
|
||||
{
|
||||
struct tlmi_attr_setting *setting = to_tlmi_attr_setting(kobj);
|
||||
|
||||
if (setting->possible_values) {
|
||||
/* Figure out what setting type is as BIOS does not return this */
|
||||
if (strchr(setting->possible_values, ';'))
|
||||
return sysfs_emit(buf, "enumeration\n");
|
||||
}
|
||||
/* Anything else is going to be a string */
|
||||
return sysfs_emit(buf, "string\n");
|
||||
}
|
||||
|
||||
static ssize_t current_value_store(struct kobject *kobj,
|
||||
struct kobj_attribute *attr,
|
||||
const char *buf, size_t count)
|
||||
@@ -601,14 +612,30 @@ static struct kobj_attribute attr_possible_values = __ATTR_RO(possible_values);
|
||||
|
||||
static struct kobj_attribute attr_current_val = __ATTR_RW_MODE(current_value, 0600);
|
||||
|
||||
static struct kobj_attribute attr_type = __ATTR_RO(type);
|
||||
|
||||
static umode_t attr_is_visible(struct kobject *kobj,
|
||||
struct attribute *attr, int n)
|
||||
{
|
||||
struct tlmi_attr_setting *setting = to_tlmi_attr_setting(kobj);
|
||||
|
||||
/* We don't want to display possible_values attributes if not available */
|
||||
if ((attr == &attr_possible_values.attr) && (!setting->possible_values))
|
||||
return 0;
|
||||
|
||||
return attr->mode;
|
||||
}
|
||||
|
||||
static struct attribute *tlmi_attrs[] = {
|
||||
&attr_displ_name.attr,
|
||||
&attr_current_val.attr,
|
||||
&attr_possible_values.attr,
|
||||
&attr_type.attr,
|
||||
NULL
|
||||
};
|
||||
|
||||
static const struct attribute_group tlmi_attr_group = {
|
||||
.is_visible = attr_is_visible,
|
||||
.attrs = tlmi_attrs,
|
||||
};
|
||||
|
||||
@@ -916,7 +943,34 @@ static int tlmi_analyze(void)
|
||||
if (ret || !setting->possible_values)
|
||||
pr_info("Error retrieving possible values for %d : %s\n",
|
||||
i, setting->display_name);
|
||||
} else {
|
||||
/*
|
||||
* Older Thinkstations don't support the bios_selections API.
|
||||
* Instead they store this as a [Optional:Option1,Option2] section of the
|
||||
* name string.
|
||||
* Try and pull that out if it's available.
|
||||
*/
|
||||
char *item, *optstart, *optend;
|
||||
|
||||
if (!tlmi_setting(setting->index, &item, LENOVO_BIOS_SETTING_GUID)) {
|
||||
optstart = strstr(item, "[Optional:");
|
||||
if (optstart) {
|
||||
optstart += strlen("[Optional:");
|
||||
optend = strstr(optstart, "]");
|
||||
if (optend)
|
||||
setting->possible_values =
|
||||
kstrndup(optstart, optend - optstart,
|
||||
GFP_KERNEL);
|
||||
}
|
||||
}
|
||||
}
|
||||
/*
|
||||
* firmware-attributes requires that possible_values are separated by ';' but
|
||||
* Lenovo FW uses ','. Replace appropriately.
|
||||
*/
|
||||
if (setting->possible_values)
|
||||
strreplace(setting->possible_values, ',', ';');
|
||||
|
||||
kobject_init(&setting->kobj, &tlmi_attr_setting_ktype);
|
||||
tlmi_priv.setting[i] = setting;
|
||||
kfree(item);
|
||||
|
||||
@@ -601,7 +601,7 @@ static int ptp_qoriq_probe(struct platform_device *dev)
|
||||
return 0;
|
||||
|
||||
no_clock:
|
||||
iounmap(ptp_qoriq->base);
|
||||
iounmap(base);
|
||||
no_ioremap:
|
||||
release_resource(ptp_qoriq->rsrc);
|
||||
no_resource:
|
||||
|
||||
@@ -215,7 +215,7 @@ static int reg_fixed_voltage_probe(struct platform_device *pdev)
|
||||
drvdata->enable_clock = devm_clk_get(dev, NULL);
|
||||
if (IS_ERR(drvdata->enable_clock)) {
|
||||
dev_err(dev, "Can't get enable-clock from devicetree\n");
|
||||
return -ENOENT;
|
||||
return PTR_ERR(drvdata->enable_clock);
|
||||
}
|
||||
} else if (drvtype && drvtype->has_performance_state) {
|
||||
drvdata->desc.ops = &fixed_voltage_domain_ops;
|
||||
|
||||
@@ -86,8 +86,9 @@ static struct ap_driver vfio_ap_drv = {
|
||||
|
||||
static void vfio_ap_matrix_dev_release(struct device *dev)
|
||||
{
|
||||
struct ap_matrix_dev *matrix_dev = dev_get_drvdata(dev);
|
||||
struct ap_matrix_dev *matrix_dev;
|
||||
|
||||
matrix_dev = container_of(dev, struct ap_matrix_dev, device);
|
||||
kfree(matrix_dev);
|
||||
}
|
||||
|
||||
|
||||
@@ -4737,7 +4737,7 @@ int megasas_task_abort_fusion(struct scsi_cmnd *scmd)
|
||||
devhandle = megasas_get_tm_devhandle(scmd->device);
|
||||
|
||||
if (devhandle == (u16)ULONG_MAX) {
|
||||
ret = SUCCESS;
|
||||
ret = FAILED;
|
||||
sdev_printk(KERN_INFO, scmd->device,
|
||||
"task abort issued for invalid devhandle\n");
|
||||
mutex_unlock(&instance->reset_mutex);
|
||||
@@ -4807,7 +4807,7 @@ int megasas_reset_target_fusion(struct scsi_cmnd *scmd)
|
||||
devhandle = megasas_get_tm_devhandle(scmd->device);
|
||||
|
||||
if (devhandle == (u16)ULONG_MAX) {
|
||||
ret = SUCCESS;
|
||||
ret = FAILED;
|
||||
sdev_printk(KERN_INFO, scmd->device,
|
||||
"target reset issued for invalid devhandle\n");
|
||||
mutex_unlock(&instance->reset_mutex);
|
||||
|
||||
@@ -6456,11 +6456,6 @@ _base_allocate_memory_pools(struct MPT3SAS_ADAPTER *ioc)
|
||||
else if (rc == -EAGAIN)
|
||||
goto try_32bit_dma;
|
||||
total_sz += sense_sz;
|
||||
ioc_info(ioc,
|
||||
"sense pool(0x%p)- dma(0x%llx): depth(%d),"
|
||||
"element_size(%d), pool_size(%d kB)\n",
|
||||
ioc->sense, (unsigned long long)ioc->sense_dma, ioc->scsiio_depth,
|
||||
SCSI_SENSE_BUFFERSIZE, sz / 1024);
|
||||
/* reply pool, 4 byte align */
|
||||
sz = ioc->reply_free_queue_depth * ioc->reply_sz;
|
||||
rc = _base_allocate_reply_pool(ioc, sz);
|
||||
|
||||
@@ -1040,6 +1040,9 @@ static int au1200fb_fb_check_var(struct fb_var_screeninfo *var,
|
||||
u32 pixclock;
|
||||
int screen_size, plane;
|
||||
|
||||
if (!var->pixclock)
|
||||
return -EINVAL;
|
||||
|
||||
plane = fbdev->plane;
|
||||
|
||||
/* Make sure that the mode respect all LCD controller and
|
||||
|
||||
@@ -234,6 +234,9 @@ static void get_modedb(struct fb_videomode **modedb, unsigned int *size)
|
||||
|
||||
static int lxfb_check_var(struct fb_var_screeninfo *var, struct fb_info *info)
|
||||
{
|
||||
if (!var->pixclock)
|
||||
return -EINVAL;
|
||||
|
||||
if (var->xres > 1920 || var->yres > 1440)
|
||||
return -EINVAL;
|
||||
|
||||
|
||||
@@ -1214,6 +1214,9 @@ static int intelfb_check_var(struct fb_var_screeninfo *var,
|
||||
|
||||
dinfo = GET_DINFO(info);
|
||||
|
||||
if (!var->pixclock)
|
||||
return -EINVAL;
|
||||
|
||||
/* update the pitch */
|
||||
if (intelfbhw_validate_mode(dinfo, var) != 0)
|
||||
return -EINVAL;
|
||||
|
||||
@@ -763,6 +763,8 @@ static int nvidiafb_check_var(struct fb_var_screeninfo *var,
|
||||
int pitch, err = 0;
|
||||
|
||||
NVTRACE_ENTER();
|
||||
if (!var->pixclock)
|
||||
return -EINVAL;
|
||||
|
||||
var->transp.offset = 0;
|
||||
var->transp.length = 0;
|
||||
|
||||
@@ -166,6 +166,9 @@ tgafb_check_var(struct fb_var_screeninfo *var, struct fb_info *info)
|
||||
{
|
||||
struct tga_par *par = (struct tga_par *)info->par;
|
||||
|
||||
if (!var->pixclock)
|
||||
return -EINVAL;
|
||||
|
||||
if (par->tga_type == TGA_TYPE_8PLANE) {
|
||||
if (var->bits_per_pixel != 8)
|
||||
return -EINVAL;
|
||||
|
||||
@@ -4267,7 +4267,9 @@ static long btrfs_ioctl_qgroup_assign(struct file *file, void __user *arg)
|
||||
}
|
||||
|
||||
/* update qgroup status and info */
|
||||
mutex_lock(&fs_info->qgroup_ioctl_lock);
|
||||
err = btrfs_run_qgroups(trans);
|
||||
mutex_unlock(&fs_info->qgroup_ioctl_lock);
|
||||
if (err < 0)
|
||||
btrfs_handle_fs_error(fs_info, err,
|
||||
"failed to update qgroup status and info");
|
||||
|
||||
@@ -2752,13 +2752,22 @@ cleanup:
|
||||
}
|
||||
|
||||
/*
|
||||
* called from commit_transaction. Writes all changed qgroups to disk.
|
||||
* Writes all changed qgroups to disk.
|
||||
* Called by the transaction commit path and the qgroup assign ioctl.
|
||||
*/
|
||||
int btrfs_run_qgroups(struct btrfs_trans_handle *trans)
|
||||
{
|
||||
struct btrfs_fs_info *fs_info = trans->fs_info;
|
||||
int ret = 0;
|
||||
|
||||
/*
|
||||
* In case we are called from the qgroup assign ioctl, assert that we
|
||||
* are holding the qgroup_ioctl_lock, otherwise we can race with a quota
|
||||
* disable operation (ioctl) and access a freed quota root.
|
||||
*/
|
||||
if (trans->transaction->state != TRANS_STATE_COMMIT_DOING)
|
||||
lockdep_assert_held(&fs_info->qgroup_ioctl_lock);
|
||||
|
||||
if (!fs_info->quota_root)
|
||||
return ret;
|
||||
|
||||
|
||||
@@ -1406,8 +1406,17 @@ struct btrfs_device *btrfs_scan_one_device(const char *path, fmode_t flags,
|
||||
* So, we need to add a special mount option to scan for
|
||||
* later supers, using BTRFS_SUPER_MIRROR_MAX instead
|
||||
*/
|
||||
flags |= FMODE_EXCL;
|
||||
|
||||
/*
|
||||
* Avoid using flag |= FMODE_EXCL here, as the systemd-udev may
|
||||
* initiate the device scan which may race with the user's mount
|
||||
* or mkfs command, resulting in failure.
|
||||
* Since the device scan is solely for reading purposes, there is
|
||||
* no need for FMODE_EXCL. Additionally, the devices are read again
|
||||
* during the mount process. It is ok to get some inconsistent
|
||||
* values temporarily, as the device paths of the fsid are the only
|
||||
* required information for assembling the volume.
|
||||
*/
|
||||
bdev = blkdev_get_by_path(path, flags, holder);
|
||||
if (IS_ERR(bdev))
|
||||
return ERR_CAST(bdev);
|
||||
|
||||
@@ -118,7 +118,10 @@ extern const struct dentry_operations cifs_ci_dentry_ops;
|
||||
#ifdef CONFIG_CIFS_DFS_UPCALL
|
||||
extern struct vfsmount *cifs_dfs_d_automount(struct path *path);
|
||||
#else
|
||||
#define cifs_dfs_d_automount NULL
|
||||
static inline struct vfsmount *cifs_dfs_d_automount(struct path *path)
|
||||
{
|
||||
return ERR_PTR(-EREMOTE);
|
||||
}
|
||||
#endif
|
||||
|
||||
/* Functions related to symlinks */
|
||||
|
||||
@@ -4751,8 +4751,13 @@ CIFSGetDFSRefer(const unsigned int xid, struct cifs_ses *ses,
|
||||
return -ENODEV;
|
||||
|
||||
getDFSRetry:
|
||||
rc = smb_init(SMB_COM_TRANSACTION2, 15, ses->tcon_ipc, (void **) &pSMB,
|
||||
(void **) &pSMBr);
|
||||
/*
|
||||
* Use smb_init_no_reconnect() instead of smb_init() as
|
||||
* CIFSGetDFSRefer() may be called from cifs_reconnect_tcon() and thus
|
||||
* causing an infinite recursion.
|
||||
*/
|
||||
rc = smb_init_no_reconnect(SMB_COM_TRANSACTION2, 15, ses->tcon_ipc,
|
||||
(void **)&pSMB, (void **)&pSMBr);
|
||||
if (rc)
|
||||
return rc;
|
||||
|
||||
|
||||
@@ -292,7 +292,7 @@ int ksmbd_conn_handler_loop(void *p)
|
||||
kvfree(conn->request_buf);
|
||||
conn->request_buf = NULL;
|
||||
|
||||
size = t->ops->read(t, hdr_buf, sizeof(hdr_buf));
|
||||
size = t->ops->read(t, hdr_buf, sizeof(hdr_buf), -1);
|
||||
if (size != sizeof(hdr_buf))
|
||||
break;
|
||||
|
||||
@@ -335,7 +335,7 @@ int ksmbd_conn_handler_loop(void *p)
|
||||
* We already read 4 bytes to find out PDU size, now
|
||||
* read in PDU
|
||||
*/
|
||||
size = t->ops->read(t, conn->request_buf + 4, pdu_size);
|
||||
size = t->ops->read(t, conn->request_buf + 4, pdu_size, 2);
|
||||
if (size < 0) {
|
||||
pr_err("sock_read failed: %d\n", size);
|
||||
break;
|
||||
|
||||
@@ -122,7 +122,8 @@ struct ksmbd_conn_ops {
|
||||
struct ksmbd_transport_ops {
|
||||
int (*prepare)(struct ksmbd_transport *t);
|
||||
void (*disconnect)(struct ksmbd_transport *t);
|
||||
int (*read)(struct ksmbd_transport *t, char *buf, unsigned int size);
|
||||
int (*read)(struct ksmbd_transport *t, char *buf,
|
||||
unsigned int size, int max_retries);
|
||||
int (*writev)(struct ksmbd_transport *t, struct kvec *iovs, int niov,
|
||||
int size, bool need_invalidate_rkey,
|
||||
unsigned int remote_key);
|
||||
|
||||
@@ -663,7 +663,7 @@ static int smb_direct_post_recv(struct smb_direct_transport *t,
|
||||
}
|
||||
|
||||
static int smb_direct_read(struct ksmbd_transport *t, char *buf,
|
||||
unsigned int size)
|
||||
unsigned int size, int unused)
|
||||
{
|
||||
struct smb_direct_recvmsg *recvmsg;
|
||||
struct smb_direct_data_transfer *data_transfer;
|
||||
|
||||
@@ -291,16 +291,18 @@ static int ksmbd_tcp_run_kthread(struct interface *iface)
|
||||
|
||||
/**
|
||||
* ksmbd_tcp_readv() - read data from socket in given iovec
|
||||
* @t: TCP transport instance
|
||||
* @iov_orig: base IO vector
|
||||
* @nr_segs: number of segments in base iov
|
||||
* @to_read: number of bytes to read from socket
|
||||
* @t: TCP transport instance
|
||||
* @iov_orig: base IO vector
|
||||
* @nr_segs: number of segments in base iov
|
||||
* @to_read: number of bytes to read from socket
|
||||
* @max_retries: maximum retry count
|
||||
*
|
||||
* Return: on success return number of bytes read from socket,
|
||||
* otherwise return error number
|
||||
*/
|
||||
static int ksmbd_tcp_readv(struct tcp_transport *t, struct kvec *iov_orig,
|
||||
unsigned int nr_segs, unsigned int to_read)
|
||||
unsigned int nr_segs, unsigned int to_read,
|
||||
int max_retries)
|
||||
{
|
||||
int length = 0;
|
||||
int total_read;
|
||||
@@ -308,7 +310,6 @@ static int ksmbd_tcp_readv(struct tcp_transport *t, struct kvec *iov_orig,
|
||||
struct msghdr ksmbd_msg;
|
||||
struct kvec *iov;
|
||||
struct ksmbd_conn *conn = KSMBD_TRANS(t)->conn;
|
||||
int max_retry = 2;
|
||||
|
||||
iov = get_conn_iovec(t, nr_segs);
|
||||
if (!iov)
|
||||
@@ -335,14 +336,23 @@ static int ksmbd_tcp_readv(struct tcp_transport *t, struct kvec *iov_orig,
|
||||
} else if (conn->status == KSMBD_SESS_NEED_RECONNECT) {
|
||||
total_read = -EAGAIN;
|
||||
break;
|
||||
} else if ((length == -ERESTARTSYS || length == -EAGAIN) &&
|
||||
max_retry) {
|
||||
} else if (length == -ERESTARTSYS || length == -EAGAIN) {
|
||||
/*
|
||||
* If max_retries is negative, Allow unlimited
|
||||
* retries to keep connection with inactive sessions.
|
||||
*/
|
||||
if (max_retries == 0) {
|
||||
total_read = length;
|
||||
break;
|
||||
} else if (max_retries > 0) {
|
||||
max_retries--;
|
||||
}
|
||||
|
||||
usleep_range(1000, 2000);
|
||||
length = 0;
|
||||
max_retry--;
|
||||
continue;
|
||||
} else if (length <= 0) {
|
||||
total_read = -EAGAIN;
|
||||
total_read = length;
|
||||
break;
|
||||
}
|
||||
}
|
||||
@@ -358,14 +368,15 @@ static int ksmbd_tcp_readv(struct tcp_transport *t, struct kvec *iov_orig,
|
||||
* Return: on success return number of bytes read from socket,
|
||||
* otherwise return error number
|
||||
*/
|
||||
static int ksmbd_tcp_read(struct ksmbd_transport *t, char *buf, unsigned int to_read)
|
||||
static int ksmbd_tcp_read(struct ksmbd_transport *t, char *buf,
|
||||
unsigned int to_read, int max_retries)
|
||||
{
|
||||
struct kvec iov;
|
||||
|
||||
iov.iov_base = buf;
|
||||
iov.iov_len = to_read;
|
||||
|
||||
return ksmbd_tcp_readv(TCP_TRANS(t), &iov, 1, to_read);
|
||||
return ksmbd_tcp_readv(TCP_TRANS(t), &iov, 1, to_read, max_retries);
|
||||
}
|
||||
|
||||
static int ksmbd_tcp_writev(struct ksmbd_transport *t, struct kvec *iov,
|
||||
|
||||
@@ -1994,8 +1994,7 @@ _nfs4_opendata_reclaim_to_nfs4_state(struct nfs4_opendata *data)
|
||||
if (!data->rpc_done) {
|
||||
if (data->rpc_status)
|
||||
return ERR_PTR(data->rpc_status);
|
||||
/* cached opens have already been processed */
|
||||
goto update;
|
||||
return nfs4_try_open_cached(data);
|
||||
}
|
||||
|
||||
ret = nfs_refresh_inode(inode, &data->f_attr);
|
||||
@@ -2004,7 +2003,7 @@ _nfs4_opendata_reclaim_to_nfs4_state(struct nfs4_opendata *data)
|
||||
|
||||
if (data->o_res.delegation_type != 0)
|
||||
nfs4_opendata_check_deleg(data, state);
|
||||
update:
|
||||
|
||||
if (!update_open_stateid(state, &data->o_res.stateid,
|
||||
NULL, data->o_arg.fmode))
|
||||
return ERR_PTR(-EAGAIN);
|
||||
|
||||
@@ -391,25 +391,27 @@ int fsverity_ioctl_enable(struct file *filp, const void __user *uarg)
|
||||
goto out_drop_write;
|
||||
|
||||
err = enable_verity(filp, &arg);
|
||||
if (err)
|
||||
goto out_allow_write_access;
|
||||
|
||||
/*
|
||||
* Some pages of the file may have been evicted from pagecache after
|
||||
* being used in the Merkle tree construction, then read into pagecache
|
||||
* again by another process reading from the file concurrently. Since
|
||||
* these pages didn't undergo verification against the file digest which
|
||||
* fs-verity now claims to be enforcing, we have to wipe the pagecache
|
||||
* to ensure that all future reads are verified.
|
||||
* We no longer drop the inode's pagecache after enabling verity. This
|
||||
* used to be done to try to avoid a race condition where pages could be
|
||||
* evicted after being used in the Merkle tree construction, then
|
||||
* re-instantiated by a concurrent read. Such pages are unverified, and
|
||||
* the backing storage could have filled them with different content, so
|
||||
* they shouldn't be used to fulfill reads once verity is enabled.
|
||||
*
|
||||
* But, dropping the pagecache has a big performance impact, and it
|
||||
* doesn't fully solve the race condition anyway. So for those reasons,
|
||||
* and also because this race condition isn't very important relatively
|
||||
* speaking (especially for small-ish files, where the chance of a page
|
||||
* being used, evicted, *and* re-instantiated all while enabling verity
|
||||
* is quite small), we no longer drop the inode's pagecache.
|
||||
*/
|
||||
filemap_write_and_wait(inode->i_mapping);
|
||||
invalidate_inode_pages2(inode->i_mapping);
|
||||
|
||||
/*
|
||||
* allow_write_access() is needed to pair with deny_write_access().
|
||||
* Regardless, the filesystem won't allow writing to verity files.
|
||||
*/
|
||||
out_allow_write_access:
|
||||
allow_write_access(filp);
|
||||
out_drop_write:
|
||||
mnt_drop_write_file(filp);
|
||||
|
||||
@@ -736,6 +736,7 @@ static ssize_t zonefs_file_dio_append(struct kiocb *iocb, struct iov_iter *from)
|
||||
struct zonefs_inode_info *zi = ZONEFS_I(inode);
|
||||
struct block_device *bdev = inode->i_sb->s_bdev;
|
||||
unsigned int max = bdev_max_zone_append_sectors(bdev);
|
||||
pgoff_t start, end;
|
||||
struct bio *bio;
|
||||
ssize_t size;
|
||||
int nr_pages;
|
||||
@@ -744,6 +745,19 @@ static ssize_t zonefs_file_dio_append(struct kiocb *iocb, struct iov_iter *from)
|
||||
max = ALIGN_DOWN(max << SECTOR_SHIFT, inode->i_sb->s_blocksize);
|
||||
iov_iter_truncate(from, max);
|
||||
|
||||
/*
|
||||
* If the inode block size (zone write granularity) is smaller than the
|
||||
* page size, we may be appending data belonging to the last page of the
|
||||
* inode straddling inode->i_size, with that page already cached due to
|
||||
* a buffered read or readahead. So make sure to invalidate that page.
|
||||
* This will always be a no-op for the case where the block size is
|
||||
* equal to the page size.
|
||||
*/
|
||||
start = iocb->ki_pos >> PAGE_SHIFT;
|
||||
end = (iocb->ki_pos + iov_iter_count(from) - 1) >> PAGE_SHIFT;
|
||||
if (invalidate_inode_pages2_range(inode->i_mapping, start, end))
|
||||
return -EBUSY;
|
||||
|
||||
nr_pages = iov_iter_npages(from, BIO_MAX_VECS);
|
||||
if (!nr_pages)
|
||||
return 0;
|
||||
@@ -782,7 +796,7 @@ static ssize_t zonefs_file_dio_append(struct kiocb *iocb, struct iov_iter *from)
|
||||
if (bio->bi_iter.bi_sector != wpsector) {
|
||||
zonefs_warn(inode->i_sb,
|
||||
"Corrupted write pointer %llu for zone at %llu\n",
|
||||
wpsector, zi->i_zsector);
|
||||
bio->bi_iter.bi_sector, zi->i_zsector);
|
||||
ret = -EIO;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -768,7 +768,7 @@ TRACE_EVENT_RCU(rcu_torture_read,
|
||||
TP_ARGS(rcutorturename, rhp, secs, c_old, c),
|
||||
|
||||
TP_STRUCT__entry(
|
||||
__field(char, rcutorturename[RCUTORTURENAME_LEN])
|
||||
__array(char, rcutorturename, RCUTORTURENAME_LEN)
|
||||
__field(struct rcu_head *, rhp)
|
||||
__field(unsigned long, secs)
|
||||
__field(unsigned long, c_old)
|
||||
|
||||
@@ -500,6 +500,8 @@ struct xenpf_symdata {
|
||||
};
|
||||
DEFINE_GUEST_HANDLE_STRUCT(xenpf_symdata);
|
||||
|
||||
#define XENPF_get_dom0_console 64
|
||||
|
||||
struct xen_platform_op {
|
||||
uint32_t cmd;
|
||||
uint32_t interface_version; /* XENPF_INTERFACE_VERSION */
|
||||
@@ -523,6 +525,7 @@ struct xen_platform_op {
|
||||
struct xenpf_mem_hotadd mem_add;
|
||||
struct xenpf_core_parking core_parking;
|
||||
struct xenpf_symdata symdata;
|
||||
struct dom0_vga_console_info dom0_console;
|
||||
uint8_t pad[128];
|
||||
} u;
|
||||
};
|
||||
|
||||
@@ -152,7 +152,7 @@ COMPAT_SYSCALL_DEFINE3(sched_getaffinity, compat_pid_t, pid, unsigned int, len,
|
||||
if (len & (sizeof(compat_ulong_t)-1))
|
||||
return -EINVAL;
|
||||
|
||||
if (!alloc_cpumask_var(&mask, GFP_KERNEL))
|
||||
if (!zalloc_cpumask_var(&mask, GFP_KERNEL))
|
||||
return -ENOMEM;
|
||||
|
||||
ret = sched_getaffinity(pid, mask);
|
||||
|
||||
@@ -13,5 +13,6 @@ CFLAGS_core.o := $(call cc-option,-fno-conserve-stack) \
|
||||
obj-y := core.o debugfs.o report.o
|
||||
obj-$(CONFIG_KCSAN_SELFTEST) += selftest.o
|
||||
|
||||
CFLAGS_kcsan_test.o := $(CFLAGS_KCSAN) -g -fno-omit-frame-pointer
|
||||
CFLAGS_kcsan_test.o := $(CFLAGS_KCSAN) -fno-omit-frame-pointer
|
||||
CFLAGS_kcsan_test.o += $(DISABLE_STRUCTLEAK_PLUGIN)
|
||||
obj-$(CONFIG_KCSAN_KUNIT_TEST) += kcsan_test.o
|
||||
|
||||
@@ -8327,14 +8327,14 @@ SYSCALL_DEFINE3(sched_getaffinity, pid_t, pid, unsigned int, len,
|
||||
if (len & (sizeof(unsigned long)-1))
|
||||
return -EINVAL;
|
||||
|
||||
if (!alloc_cpumask_var(&mask, GFP_KERNEL))
|
||||
if (!zalloc_cpumask_var(&mask, GFP_KERNEL))
|
||||
return -ENOMEM;
|
||||
|
||||
ret = sched_getaffinity(pid, mask);
|
||||
if (ret == 0) {
|
||||
unsigned int retlen = min(len, cpumask_size());
|
||||
|
||||
if (copy_to_user(user_mask_ptr, mask, retlen))
|
||||
if (copy_to_user(user_mask_ptr, cpumask_bits(mask), retlen))
|
||||
ret = -EFAULT;
|
||||
else
|
||||
ret = retlen;
|
||||
|
||||
@@ -146,7 +146,7 @@ static int __init test_gen_kprobe_cmd(void)
|
||||
if (trace_event_file_is_valid(gen_kprobe_test))
|
||||
gen_kprobe_test = NULL;
|
||||
/* We got an error after creating the event, delete it */
|
||||
ret = kprobe_event_delete("gen_kprobe_test");
|
||||
kprobe_event_delete("gen_kprobe_test");
|
||||
goto out;
|
||||
}
|
||||
|
||||
@@ -211,7 +211,7 @@ static int __init test_gen_kretprobe_cmd(void)
|
||||
if (trace_event_file_is_valid(gen_kretprobe_test))
|
||||
gen_kretprobe_test = NULL;
|
||||
/* We got an error after creating the event, delete it */
|
||||
ret = kprobe_event_delete("gen_kretprobe_test");
|
||||
kprobe_event_delete("gen_kretprobe_test");
|
||||
goto out;
|
||||
}
|
||||
|
||||
|
||||
@@ -936,6 +936,8 @@ static int bcm_tx_setup(struct bcm_msg_head *msg_head, struct msghdr *msg,
|
||||
|
||||
cf = op->frames + op->cfsiz * i;
|
||||
err = memcpy_from_msg((u8 *)cf, msg, op->cfsiz);
|
||||
if (err < 0)
|
||||
goto free_op;
|
||||
|
||||
if (op->flags & CAN_FD_FRAME) {
|
||||
if (cf->len > 64)
|
||||
@@ -945,12 +947,8 @@ static int bcm_tx_setup(struct bcm_msg_head *msg_head, struct msghdr *msg,
|
||||
err = -EINVAL;
|
||||
}
|
||||
|
||||
if (err < 0) {
|
||||
if (op->frames != &op->sframe)
|
||||
kfree(op->frames);
|
||||
kfree(op);
|
||||
return err;
|
||||
}
|
||||
if (err < 0)
|
||||
goto free_op;
|
||||
|
||||
if (msg_head->flags & TX_CP_CAN_ID) {
|
||||
/* copy can_id into frame */
|
||||
@@ -1021,6 +1019,12 @@ static int bcm_tx_setup(struct bcm_msg_head *msg_head, struct msghdr *msg,
|
||||
bcm_tx_start_timer(op);
|
||||
|
||||
return msg_head->nframes * op->cfsiz + MHSIZ;
|
||||
|
||||
free_op:
|
||||
if (op->frames != &op->sframe)
|
||||
kfree(op->frames);
|
||||
kfree(op);
|
||||
return err;
|
||||
}
|
||||
|
||||
/*
|
||||
|
||||
@@ -1124,8 +1124,6 @@ static void __j1939_session_cancel(struct j1939_session *session,
|
||||
|
||||
if (session->sk)
|
||||
j1939_sk_send_loop_abort(session->sk, session->err);
|
||||
else
|
||||
j1939_sk_errqueue(session, J1939_ERRQUEUE_RX_ABORT);
|
||||
}
|
||||
|
||||
static void j1939_session_cancel(struct j1939_session *session,
|
||||
@@ -1140,6 +1138,9 @@ static void j1939_session_cancel(struct j1939_session *session,
|
||||
}
|
||||
|
||||
j1939_session_list_unlock(session->priv);
|
||||
|
||||
if (!session->sk)
|
||||
j1939_sk_errqueue(session, J1939_ERRQUEUE_RX_ABORT);
|
||||
}
|
||||
|
||||
static enum hrtimer_restart j1939_tp_txtimer(struct hrtimer *hrtimer)
|
||||
@@ -1253,6 +1254,9 @@ static enum hrtimer_restart j1939_tp_rxtimer(struct hrtimer *hrtimer)
|
||||
__j1939_session_cancel(session, J1939_XTP_ABORT_TIMEOUT);
|
||||
}
|
||||
j1939_session_list_unlock(session->priv);
|
||||
|
||||
if (!session->sk)
|
||||
j1939_sk_errqueue(session, J1939_ERRQUEUE_RX_ABORT);
|
||||
}
|
||||
|
||||
j1939_session_put(session);
|
||||
|
||||
@@ -385,7 +385,7 @@ void hsr_addr_subst_dest(struct hsr_node *node_src, struct sk_buff *skb,
|
||||
node_dst = find_node_by_addr_A(&port->hsr->node_db,
|
||||
eth_hdr(skb)->h_dest);
|
||||
if (!node_dst) {
|
||||
if (net_ratelimit())
|
||||
if (port->hsr->prot_version != PRP_V1 && net_ratelimit())
|
||||
netdev_err(skb->dev, "%s: Unknown node\n", __func__);
|
||||
return;
|
||||
}
|
||||
|
||||
@@ -2140,6 +2140,7 @@ static void xs_tcp_shutdown(struct rpc_xprt *xprt)
|
||||
switch (skst) {
|
||||
case TCP_FIN_WAIT1:
|
||||
case TCP_FIN_WAIT2:
|
||||
case TCP_LAST_ACK:
|
||||
break;
|
||||
case TCP_ESTABLISHED:
|
||||
case TCP_CLOSE_WAIT:
|
||||
|
||||
@@ -906,7 +906,9 @@ static int copy_to_user_aead(struct xfrm_algo_aead *aead, struct sk_buff *skb)
|
||||
return -EMSGSIZE;
|
||||
|
||||
ap = nla_data(nla);
|
||||
memcpy(ap, aead, sizeof(*aead));
|
||||
strscpy_pad(ap->alg_name, aead->alg_name, sizeof(ap->alg_name));
|
||||
ap->alg_key_len = aead->alg_key_len;
|
||||
ap->alg_icv_len = aead->alg_icv_len;
|
||||
|
||||
if (redact_secret && aead->alg_key_len)
|
||||
memset(ap->alg_key, 0, (aead->alg_key_len + 7) / 8);
|
||||
@@ -926,7 +928,8 @@ static int copy_to_user_ealg(struct xfrm_algo *ealg, struct sk_buff *skb)
|
||||
return -EMSGSIZE;
|
||||
|
||||
ap = nla_data(nla);
|
||||
memcpy(ap, ealg, sizeof(*ealg));
|
||||
strscpy_pad(ap->alg_name, ealg->alg_name, sizeof(ap->alg_name));
|
||||
ap->alg_key_len = ealg->alg_key_len;
|
||||
|
||||
if (redact_secret && ealg->alg_key_len)
|
||||
memset(ap->alg_key, 0, (ealg->alg_key_len + 7) / 8);
|
||||
@@ -937,6 +940,40 @@ static int copy_to_user_ealg(struct xfrm_algo *ealg, struct sk_buff *skb)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int copy_to_user_calg(struct xfrm_algo *calg, struct sk_buff *skb)
|
||||
{
|
||||
struct nlattr *nla = nla_reserve(skb, XFRMA_ALG_COMP, sizeof(*calg));
|
||||
struct xfrm_algo *ap;
|
||||
|
||||
if (!nla)
|
||||
return -EMSGSIZE;
|
||||
|
||||
ap = nla_data(nla);
|
||||
strscpy_pad(ap->alg_name, calg->alg_name, sizeof(ap->alg_name));
|
||||
ap->alg_key_len = 0;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int copy_to_user_encap(struct xfrm_encap_tmpl *ep, struct sk_buff *skb)
|
||||
{
|
||||
struct nlattr *nla = nla_reserve(skb, XFRMA_ENCAP, sizeof(*ep));
|
||||
struct xfrm_encap_tmpl *uep;
|
||||
|
||||
if (!nla)
|
||||
return -EMSGSIZE;
|
||||
|
||||
uep = nla_data(nla);
|
||||
memset(uep, 0, sizeof(*uep));
|
||||
|
||||
uep->encap_type = ep->encap_type;
|
||||
uep->encap_sport = ep->encap_sport;
|
||||
uep->encap_dport = ep->encap_dport;
|
||||
uep->encap_oa = ep->encap_oa;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int xfrm_smark_put(struct sk_buff *skb, struct xfrm_mark *m)
|
||||
{
|
||||
int ret = 0;
|
||||
@@ -992,12 +1029,12 @@ static int copy_to_user_state_extra(struct xfrm_state *x,
|
||||
goto out;
|
||||
}
|
||||
if (x->calg) {
|
||||
ret = nla_put(skb, XFRMA_ALG_COMP, sizeof(*(x->calg)), x->calg);
|
||||
ret = copy_to_user_calg(x->calg, skb);
|
||||
if (ret)
|
||||
goto out;
|
||||
}
|
||||
if (x->encap) {
|
||||
ret = nla_put(skb, XFRMA_ENCAP, sizeof(*x->encap), x->encap);
|
||||
ret = copy_to_user_encap(x->encap, skb);
|
||||
if (ret)
|
||||
goto out;
|
||||
}
|
||||
|
||||
@@ -2137,6 +2137,8 @@ int pcm_lib_apply_appl_ptr(struct snd_pcm_substream *substream,
|
||||
ret = substream->ops->ack(substream);
|
||||
if (ret < 0) {
|
||||
runtime->control->appl_ptr = old_appl_ptr;
|
||||
if (ret == -EPIPE)
|
||||
__snd_pcm_xrun(substream);
|
||||
return ret;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -430,7 +430,7 @@ void HPI_6205(struct hpi_message *phm, struct hpi_response *phr)
|
||||
pao = hpi_find_adapter(phm->adapter_index);
|
||||
} else {
|
||||
/* subsys messages don't address an adapter */
|
||||
_HPI_6205(NULL, phm, phr);
|
||||
phr->error = HPI_ERROR_INVALID_OBJ_INDEX;
|
||||
return;
|
||||
}
|
||||
|
||||
|
||||
@@ -4231,8 +4231,10 @@ static int tuning_ctl_set(struct hda_codec *codec, hda_nid_t nid,
|
||||
|
||||
for (i = 0; i < TUNING_CTLS_COUNT; i++)
|
||||
if (nid == ca0132_tuning_ctls[i].nid)
|
||||
break;
|
||||
goto found;
|
||||
|
||||
return -EINVAL;
|
||||
found:
|
||||
snd_hda_power_up(codec);
|
||||
dspio_set_param(codec, ca0132_tuning_ctls[i].mid, 0x20,
|
||||
ca0132_tuning_ctls[i].req,
|
||||
|
||||
@@ -980,7 +980,10 @@ static const struct snd_pci_quirk cxt5066_fixups[] = {
|
||||
SND_PCI_QUIRK(0x17aa, 0x3905, "Lenovo G50-30", CXT_FIXUP_STEREO_DMIC),
|
||||
SND_PCI_QUIRK(0x17aa, 0x390b, "Lenovo G50-80", CXT_FIXUP_STEREO_DMIC),
|
||||
SND_PCI_QUIRK(0x17aa, 0x3975, "Lenovo U300s", CXT_FIXUP_STEREO_DMIC),
|
||||
SND_PCI_QUIRK(0x17aa, 0x3977, "Lenovo IdeaPad U310", CXT_PINCFG_LENOVO_NOTEBOOK),
|
||||
/* NOTE: we'd need to extend the quirk for 17aa:3977 as the same
|
||||
* PCI SSID is used on multiple Lenovo models
|
||||
*/
|
||||
SND_PCI_QUIRK(0x17aa, 0x3977, "Lenovo IdeaPad U310", CXT_FIXUP_STEREO_DMIC),
|
||||
SND_PCI_QUIRK(0x17aa, 0x3978, "Lenovo G50-70", CXT_FIXUP_STEREO_DMIC),
|
||||
SND_PCI_QUIRK(0x17aa, 0x397b, "Lenovo S205", CXT_FIXUP_STEREO_DMIC),
|
||||
SND_PCI_QUIRK_VENDOR(0x17aa, "Thinkpad", CXT_FIXUP_THINKPAD_ACPI),
|
||||
@@ -1003,6 +1006,7 @@ static const struct hda_model_fixup cxt5066_fixup_models[] = {
|
||||
{ .id = CXT_FIXUP_MUTE_LED_GPIO, .name = "mute-led-gpio" },
|
||||
{ .id = CXT_FIXUP_HP_ZBOOK_MUTE_LED, .name = "hp-zbook-mute-led" },
|
||||
{ .id = CXT_FIXUP_HP_MIC_NO_PRESENCE, .name = "hp-mic-fix" },
|
||||
{ .id = CXT_PINCFG_LENOVO_NOTEBOOK, .name = "lenovo-20149" },
|
||||
{}
|
||||
};
|
||||
|
||||
|
||||
@@ -2625,6 +2625,7 @@ static const struct snd_pci_quirk alc882_fixup_tbl[] = {
|
||||
SND_PCI_QUIRK(0x1558, 0x65e5, "Clevo PC50D[PRS](?:-D|-G)?", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
|
||||
SND_PCI_QUIRK(0x1558, 0x65f1, "Clevo PC50HS", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
|
||||
SND_PCI_QUIRK(0x1558, 0x65f5, "Clevo PD50PN[NRT]", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
|
||||
SND_PCI_QUIRK(0x1558, 0x66a2, "Clevo PE60RNE", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
|
||||
SND_PCI_QUIRK(0x1558, 0x67d1, "Clevo PB71[ER][CDF]", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
|
||||
SND_PCI_QUIRK(0x1558, 0x67e1, "Clevo PB71[DE][CDF]", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
|
||||
SND_PCI_QUIRK(0x1558, 0x67e5, "Clevo PC70D[PRS](?:-D|-G)?", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
|
||||
@@ -2645,6 +2646,7 @@ static const struct snd_pci_quirk alc882_fixup_tbl[] = {
|
||||
SND_PCI_QUIRK(0x1558, 0x96e1, "Clevo P960[ER][CDFN]-K", ALC1220_FIXUP_CLEVO_P950),
|
||||
SND_PCI_QUIRK(0x1558, 0x97e1, "Clevo P970[ER][CDFN]", ALC1220_FIXUP_CLEVO_P950),
|
||||
SND_PCI_QUIRK(0x1558, 0x97e2, "Clevo P970RC-M", ALC1220_FIXUP_CLEVO_P950),
|
||||
SND_PCI_QUIRK(0x1558, 0xd502, "Clevo PD50SNE", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
|
||||
SND_PCI_QUIRK_VENDOR(0x1558, "Clevo laptop", ALC882_FIXUP_EAPD),
|
||||
SND_PCI_QUIRK(0x161f, 0x2054, "Medion laptop", ALC883_FIXUP_EAPD),
|
||||
SND_PCI_QUIRK(0x17aa, 0x3a0d, "Lenovo Y530", ALC882_FIXUP_LENOVO_Y530),
|
||||
@@ -9201,6 +9203,7 @@ static const struct snd_pci_quirk alc269_fixup_tbl[] = {
|
||||
SND_PCI_QUIRK(0x1558, 0x5101, "Clevo S510WU", ALC293_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1558, 0x5157, "Clevo W517GU1", ALC293_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1558, 0x51a1, "Clevo NS50MU", ALC293_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1558, 0x5630, "Clevo NP50RNJS", ALC256_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1558, 0x70a1, "Clevo NB70T[HJK]", ALC293_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1558, 0x70b3, "Clevo NK70SB", ALC293_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1558, 0x70f2, "Clevo NH79EPY", ALC293_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
@@ -9235,6 +9238,7 @@ static const struct snd_pci_quirk alc269_fixup_tbl[] = {
|
||||
SND_PCI_QUIRK(0x1558, 0x971d, "Clevo N970T[CDF]", ALC293_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1558, 0xa500, "Clevo NL5[03]RU", ALC293_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1558, 0xa600, "Clevo NL50NU", ALC293_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1558, 0xa671, "Clevo NP70SN[CDE]", ALC256_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1558, 0xb018, "Clevo NP50D[BE]", ALC293_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1558, 0xb019, "Clevo NH77D[BE]Q", ALC293_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1558, 0xb022, "Clevo NH77D[DC][QW]", ALC293_FIXUP_SYSTEM76_MIC_NO_PRESENCE),
|
||||
@@ -9328,6 +9332,7 @@ static const struct snd_pci_quirk alc269_fixup_tbl[] = {
|
||||
SND_PCI_QUIRK(0x17aa, 0x511e, "Thinkpad", ALC298_FIXUP_TPT470_DOCK),
|
||||
SND_PCI_QUIRK(0x17aa, 0x511f, "Thinkpad", ALC298_FIXUP_TPT470_DOCK),
|
||||
SND_PCI_QUIRK(0x17aa, 0x9e54, "LENOVO NB", ALC269_FIXUP_LENOVO_EAPD),
|
||||
SND_PCI_QUIRK(0x17aa, 0x9e56, "Lenovo ZhaoYang CF4620Z", ALC286_FIXUP_SONY_MIC_NO_PRESENCE),
|
||||
SND_PCI_QUIRK(0x1849, 0x1233, "ASRock NUC Box 1100", ALC233_FIXUP_NO_AUDIO_JACK),
|
||||
SND_PCI_QUIRK(0x1849, 0xa233, "Positivo Master C6300", ALC269_FIXUP_HEADSET_MIC),
|
||||
SND_PCI_QUIRK(0x19e5, 0x3204, "Huawei MACH-WX9", ALC256_FIXUP_HUAWEI_MACH_WX9_PINS),
|
||||
|
||||
@@ -170,7 +170,7 @@ static int snd_card_ymfpci_probe(struct pci_dev *pci,
|
||||
return -ENOENT;
|
||||
}
|
||||
|
||||
err = snd_card_new(&pci->dev, index[dev], id[dev], THIS_MODULE,
|
||||
err = snd_devm_card_new(&pci->dev, index[dev], id[dev], THIS_MODULE,
|
||||
sizeof(*chip), &card);
|
||||
if (err < 0)
|
||||
return err;
|
||||
|
||||
@@ -2165,7 +2165,7 @@ static int snd_ymfpci_memalloc(struct snd_ymfpci *chip)
|
||||
chip->work_base = ptr;
|
||||
chip->work_base_addr = ptr_addr;
|
||||
|
||||
snd_BUG_ON(ptr + chip->work_size !=
|
||||
snd_BUG_ON(ptr + PAGE_ALIGN(chip->work_size) !=
|
||||
chip->work_ptr->area + chip->work_ptr->bytes);
|
||||
|
||||
snd_ymfpci_writel(chip, YDSXGR_PLAYCTRLBASE, chip->bank_base_playback_addr);
|
||||
|
||||
@@ -239,7 +239,7 @@ enum {
|
||||
|
||||
struct tx_mute_work {
|
||||
struct tx_macro *tx;
|
||||
u32 decimator;
|
||||
u8 decimator;
|
||||
struct delayed_work dwork;
|
||||
};
|
||||
|
||||
@@ -632,7 +632,7 @@ exit:
|
||||
return 0;
|
||||
}
|
||||
|
||||
static bool is_amic_enabled(struct snd_soc_component *component, int decimator)
|
||||
static bool is_amic_enabled(struct snd_soc_component *component, u8 decimator)
|
||||
{
|
||||
u16 adc_mux_reg, adc_reg, adc_n;
|
||||
|
||||
@@ -843,7 +843,7 @@ static int tx_macro_enable_dec(struct snd_soc_dapm_widget *w,
|
||||
struct snd_kcontrol *kcontrol, int event)
|
||||
{
|
||||
struct snd_soc_component *component = snd_soc_dapm_to_component(w->dapm);
|
||||
unsigned int decimator;
|
||||
u8 decimator;
|
||||
u16 tx_vol_ctl_reg, dec_cfg_reg, hpf_gate_reg, tx_gain_ctl_reg;
|
||||
u8 hpf_cut_off_freq;
|
||||
int hpf_delay = TX_MACRO_DMIC_HPF_DELAY_MS;
|
||||
@@ -1058,7 +1058,8 @@ static int tx_macro_hw_params(struct snd_pcm_substream *substream,
|
||||
struct snd_soc_dai *dai)
|
||||
{
|
||||
struct snd_soc_component *component = dai->component;
|
||||
u32 decimator, sample_rate;
|
||||
u32 sample_rate;
|
||||
u8 decimator;
|
||||
int tx_fs_rate;
|
||||
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
|
||||
|
||||
@@ -1122,7 +1123,7 @@ static int tx_macro_digital_mute(struct snd_soc_dai *dai, int mute, int stream)
|
||||
{
|
||||
struct snd_soc_component *component = dai->component;
|
||||
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
|
||||
u16 decimator;
|
||||
u8 decimator;
|
||||
|
||||
decimator = tx->active_decimator[dai->id];
|
||||
|
||||
|
||||
@@ -446,8 +446,8 @@ static void push_back_to_ready_list(struct snd_usb_endpoint *ep,
|
||||
* This function is used both for implicit feedback endpoints and in low-
|
||||
* latency playback mode.
|
||||
*/
|
||||
void snd_usb_queue_pending_output_urbs(struct snd_usb_endpoint *ep,
|
||||
bool in_stream_lock)
|
||||
int snd_usb_queue_pending_output_urbs(struct snd_usb_endpoint *ep,
|
||||
bool in_stream_lock)
|
||||
{
|
||||
bool implicit_fb = snd_usb_endpoint_implicit_feedback_sink(ep);
|
||||
|
||||
@@ -471,7 +471,7 @@ void snd_usb_queue_pending_output_urbs(struct snd_usb_endpoint *ep,
|
||||
spin_unlock_irqrestore(&ep->lock, flags);
|
||||
|
||||
if (ctx == NULL)
|
||||
return;
|
||||
break;
|
||||
|
||||
/* copy over the length information */
|
||||
if (implicit_fb) {
|
||||
@@ -486,11 +486,14 @@ void snd_usb_queue_pending_output_urbs(struct snd_usb_endpoint *ep,
|
||||
break;
|
||||
if (err < 0) {
|
||||
/* push back to ready list again for -EAGAIN */
|
||||
if (err == -EAGAIN)
|
||||
if (err == -EAGAIN) {
|
||||
push_back_to_ready_list(ep, ctx);
|
||||
else
|
||||
break;
|
||||
}
|
||||
|
||||
if (!in_stream_lock)
|
||||
notify_xrun(ep);
|
||||
return;
|
||||
return -EPIPE;
|
||||
}
|
||||
|
||||
err = usb_submit_urb(ctx->urb, GFP_ATOMIC);
|
||||
@@ -498,13 +501,16 @@ void snd_usb_queue_pending_output_urbs(struct snd_usb_endpoint *ep,
|
||||
usb_audio_err(ep->chip,
|
||||
"Unable to submit urb #%d: %d at %s\n",
|
||||
ctx->index, err, __func__);
|
||||
notify_xrun(ep);
|
||||
return;
|
||||
if (!in_stream_lock)
|
||||
notify_xrun(ep);
|
||||
return -EPIPE;
|
||||
}
|
||||
|
||||
set_bit(ctx->index, &ep->active_mask);
|
||||
atomic_inc(&ep->submitted_urbs);
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
|
||||
@@ -49,7 +49,7 @@ int snd_usb_endpoint_implicit_feedback_sink(struct snd_usb_endpoint *ep);
|
||||
int snd_usb_endpoint_next_packet_size(struct snd_usb_endpoint *ep,
|
||||
struct snd_urb_ctx *ctx, int idx,
|
||||
unsigned int avail);
|
||||
void snd_usb_queue_pending_output_urbs(struct snd_usb_endpoint *ep,
|
||||
bool in_stream_lock);
|
||||
int snd_usb_queue_pending_output_urbs(struct snd_usb_endpoint *ep,
|
||||
bool in_stream_lock);
|
||||
|
||||
#endif /* __USBAUDIO_ENDPOINT_H */
|
||||
|
||||
@@ -39,8 +39,12 @@ static u64 parse_audio_format_i_type(struct snd_usb_audio *chip,
|
||||
case UAC_VERSION_1:
|
||||
default: {
|
||||
struct uac_format_type_i_discrete_descriptor *fmt = _fmt;
|
||||
if (format >= 64)
|
||||
return 0; /* invalid format */
|
||||
if (format >= 64) {
|
||||
usb_audio_info(chip,
|
||||
"%u:%d: invalid format type 0x%llx is detected, processed as PCM\n",
|
||||
fp->iface, fp->altsetting, format);
|
||||
format = UAC_FORMAT_TYPE_I_PCM;
|
||||
}
|
||||
sample_width = fmt->bBitResolution;
|
||||
sample_bytes = fmt->bSubframeSize;
|
||||
format = 1ULL << format;
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user